diff --git "a/runs/big_run/exp_c/run_20260408_124138/metrics.jsonl" "b/runs/big_run/exp_c/run_20260408_124138/metrics.jsonl" new file mode 100644--- /dev/null +++ "b/runs/big_run/exp_c/run_20260408_124138/metrics.jsonl" @@ -0,0 +1,3132 @@ +{"stage": "pretraining", "type": "config", "timestamp": "2026-04-08T12:41:39.587580+00:00", "device": "auto", "train_bin": "data/exp_c/train.bin", "val_bin": "data/exp_c/val.bin", "num_workers": 4, "pin_memory": true, "batch_size": 32, "gradient_accumulation_steps": 1, "learning_rate": 0.001305, "min_lr": 0.0001305, "weight_decay": 0.063, "beta1": 0.9, "beta2": 0.9978, "grad_clip": 1.786, "warmup_steps": 305, "max_steps": 3052, "lr_schedule": "wsd", "lr_decay_ratio": 0.179, "z_loss_coeff": 0.000275, "save_every": 250, "eval_every": 100, "checkpoint_dir": "checkpoints", "keep_last_checkpoints": 10, "keep_best_checkpoints": 10, "early_stopping_patience": 0, "early_stopping_min_delta": 0.001, "runs_dir": "runs/big_run/exp_c", "log_every": 10, "hf_upload": {"repo_id": "ParrotLabs/Preprocessed", "repo_type": "dataset", "path_in_repo": "", "private": null}, "compile": true} +{"stage": "pretraining", "type": "checkpoint_policy", "timestamp": "2026-04-08T12:41:39.587653+00:00", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints", "keep_last": 10, "keep_best": 10} +{"stage": "pretraining", "type": "model_architecture", "timestamp": "2026-04-08T12:41:41.309242+00:00", "vocab_size": 50258, "context_length": 1024, "n_layers": 14, "n_heads": 6, "d_model": 384, "d_ff": 768, "dropout": 0.005, "bias": false, "total_params": 39966592, "total_params_non_embedding": 39966592, "trainable_params": 39966592, "non_trainable_params": 0, "params_size_mb": 159.87} +{"stage": "pretraining", "type": "initial_validation", "timestamp": "2026-04-08T12:41:49.710046+00:00", "val_loss": 10.893890810012817, "val_ppl": 53846.399032417314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:27.986456+00:00", "epoch": 0, "step": 1, "train_loss": 10.928349494934082, "perplexity": 55734.21414819104, "lr": 4.278688524590164e-06, "grad_norm": 6.81222, "tokens_per_sec": 857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:28.291473+00:00", "epoch": 0, "step": 2, "train_loss": 10.888795852661133, "perplexity": 53572.75162868924, "lr": 8.557377049180328e-06, "grad_norm": 7.121474, "tokens_per_sec": 97699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:28.595463+00:00", "epoch": 0, "step": 3, "train_loss": 10.816173553466797, "perplexity": 49820.08791497173, "lr": 1.2836065573770493e-05, "grad_norm": 6.74162, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:28.900174+00:00", "epoch": 0, "step": 4, "train_loss": 10.699760437011719, "perplexity": 44345.23038179414, "lr": 1.7114754098360656e-05, "grad_norm": 6.671484, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:29.204751+00:00", "epoch": 0, "step": 5, "train_loss": 10.59575366973877, "perplexity": 39964.77298660201, "lr": 2.1393442622950817e-05, "grad_norm": 5.892272, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:29.507941+00:00", "epoch": 0, "step": 6, "train_loss": 10.481679916381836, "perplexity": 35656.256786716775, "lr": 2.5672131147540985e-05, "grad_norm": 4.270808, "tokens_per_sec": 108078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:29.812211+00:00", "epoch": 0, "step": 7, "train_loss": 10.387381553649902, "perplexity": 32447.5933219992, "lr": 2.9950819672131143e-05, "grad_norm": 2.962524, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:30.116559+00:00", "epoch": 0, "step": 8, "train_loss": 10.344070434570312, "perplexity": 31072.250549437747, "lr": 3.422950819672131e-05, "grad_norm": 2.357156, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:30.421608+00:00", "epoch": 0, "step": 9, "train_loss": 10.299345970153809, "perplexity": 29713.17919050884, "lr": 3.850819672131147e-05, "grad_norm": 1.987438, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:30.725641+00:00", "epoch": 0, "step": 10, "train_loss": 10.26159954071045, "perplexity": 28612.516493669707, "lr": 4.2786885245901634e-05, "grad_norm": 1.827039, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:31.028368+00:00", "epoch": 0, "step": 11, "train_loss": 10.2267484664917, "perplexity": 27632.515786633823, "lr": 4.70655737704918e-05, "grad_norm": 1.777457, "tokens_per_sec": 108160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:31.332548+00:00", "epoch": 0, "step": 12, "train_loss": 10.205941200256348, "perplexity": 27063.499044470027, "lr": 5.134426229508197e-05, "grad_norm": 1.739928, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:31.637040+00:00", "epoch": 0, "step": 13, "train_loss": 10.1965970993042, "perplexity": 26811.792793008786, "lr": 5.562295081967213e-05, "grad_norm": 1.732584, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:31.941353+00:00", "epoch": 0, "step": 14, "train_loss": 10.168211936950684, "perplexity": 26061.43558778457, "lr": 5.990163934426229e-05, "grad_norm": 1.716922, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:32.244676+00:00", "epoch": 0, "step": 15, "train_loss": 10.116865158081055, "perplexity": 24757.039655541073, "lr": 6.418032786885245e-05, "grad_norm": 1.72396, "tokens_per_sec": 108029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:32.547581+00:00", "epoch": 0, "step": 16, "train_loss": 10.090808868408203, "perplexity": 24120.294689518192, "lr": 6.845901639344262e-05, "grad_norm": 1.722242, "tokens_per_sec": 108180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:32.853421+00:00", "epoch": 0, "step": 17, "train_loss": 10.055009841918945, "perplexity": 23272.08475081854, "lr": 7.273770491803279e-05, "grad_norm": 1.701774, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:33.157392+00:00", "epoch": 0, "step": 18, "train_loss": 10.027606964111328, "perplexity": 22643.021097371628, "lr": 7.701639344262295e-05, "grad_norm": 1.707363, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:33.461548+00:00", "epoch": 0, "step": 19, "train_loss": 9.961505889892578, "perplexity": 21194.68856520139, "lr": 8.12950819672131e-05, "grad_norm": 1.702539, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:33.765581+00:00", "epoch": 0, "step": 20, "train_loss": 9.895317077636719, "perplexity": 19837.25625420823, "lr": 8.557377049180327e-05, "grad_norm": 1.694505, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:34.069302+00:00", "epoch": 0, "step": 21, "train_loss": 9.883737564086914, "perplexity": 19608.875298674786, "lr": 8.985245901639344e-05, "grad_norm": 1.709326, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:34.372477+00:00", "epoch": 0, "step": 22, "train_loss": 9.774530410766602, "perplexity": 17580.232752433047, "lr": 9.41311475409836e-05, "grad_norm": 1.713408, "tokens_per_sec": 108083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:34.676111+00:00", "epoch": 0, "step": 23, "train_loss": 9.698229789733887, "perplexity": 16288.747153944674, "lr": 9.840983606557377e-05, "grad_norm": 1.687083, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:34.980403+00:00", "epoch": 0, "step": 24, "train_loss": 9.676630973815918, "perplexity": 15940.701717704298, "lr": 0.00010268852459016394, "grad_norm": 1.676338, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:35.284517+00:00", "epoch": 0, "step": 25, "train_loss": 9.632057189941406, "perplexity": 15245.767308072278, "lr": 0.0001069672131147541, "grad_norm": 1.661994, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:35.587170+00:00", "epoch": 0, "step": 26, "train_loss": 9.547453880310059, "perplexity": 14008.980683198568, "lr": 0.00011124590163934426, "grad_norm": 1.648618, "tokens_per_sec": 108270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:35.890330+00:00", "epoch": 0, "step": 27, "train_loss": 9.51267147064209, "perplexity": 13530.091320735459, "lr": 0.0001155245901639344, "grad_norm": 1.644843, "tokens_per_sec": 108088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:36.193422+00:00", "epoch": 0, "step": 28, "train_loss": 9.420205116271973, "perplexity": 12335.112092459849, "lr": 0.00011980327868852457, "grad_norm": 1.650969, "tokens_per_sec": 108112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:36.496939+00:00", "epoch": 0, "step": 29, "train_loss": 9.34987735748291, "perplexity": 11497.41328697338, "lr": 0.00012408196721311476, "grad_norm": 1.804533, "tokens_per_sec": 107961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:36.802121+00:00", "epoch": 0, "step": 30, "train_loss": 9.291102409362793, "perplexity": 10841.128875337989, "lr": 0.0001283606557377049, "grad_norm": 2.049283, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:37.105579+00:00", "epoch": 0, "step": 31, "train_loss": 9.223809242248535, "perplexity": 10135.599841128409, "lr": 0.00013263934426229506, "grad_norm": 1.817086, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:37.408243+00:00", "epoch": 0, "step": 32, "train_loss": 9.189762115478516, "perplexity": 9796.320309030265, "lr": 0.00013691803278688525, "grad_norm": 1.631995, "tokens_per_sec": 108266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:37.712009+00:00", "epoch": 0, "step": 33, "train_loss": 9.143067359924316, "perplexity": 9349.399166254512, "lr": 0.0001411967213114754, "grad_norm": 1.800343, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:38.015719+00:00", "epoch": 0, "step": 34, "train_loss": 8.985447883605957, "perplexity": 7986.020731545474, "lr": 0.00014547540983606558, "grad_norm": 1.594725, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:38.321126+00:00", "epoch": 0, "step": 35, "train_loss": 8.97286319732666, "perplexity": 7886.148911869284, "lr": 0.00014975409836065574, "grad_norm": 1.696308, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:38.624503+00:00", "epoch": 0, "step": 36, "train_loss": 8.916935920715332, "perplexity": 7457.204721110262, "lr": 0.0001540327868852459, "grad_norm": 1.539556, "tokens_per_sec": 108010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:38.928238+00:00", "epoch": 0, "step": 37, "train_loss": 8.867230415344238, "perplexity": 7095.601864223357, "lr": 0.00015831147540983607, "grad_norm": 1.661441, "tokens_per_sec": 107884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:39.232212+00:00", "epoch": 0, "step": 38, "train_loss": 8.7997407913208, "perplexity": 6632.524575506849, "lr": 0.0001625901639344262, "grad_norm": 1.552393, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:39.536325+00:00", "epoch": 0, "step": 39, "train_loss": 8.744155883789062, "perplexity": 6273.915270443303, "lr": 0.00016686885245901638, "grad_norm": 1.590785, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:39.841345+00:00", "epoch": 0, "step": 40, "train_loss": 8.684138298034668, "perplexity": 5908.446983005011, "lr": 0.00017114754098360654, "grad_norm": 1.54248, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:40.145252+00:00", "epoch": 0, "step": 41, "train_loss": 8.700611114501953, "perplexity": 6006.581805125918, "lr": 0.00017542622950819672, "grad_norm": 1.532118, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:40.449614+00:00", "epoch": 0, "step": 42, "train_loss": 8.542961120605469, "perplexity": 5130.5139616650895, "lr": 0.00017970491803278687, "grad_norm": 1.524809, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:40.754009+00:00", "epoch": 0, "step": 43, "train_loss": 8.504578590393066, "perplexity": 4937.323147761803, "lr": 0.00018398360655737706, "grad_norm": 1.504514, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:41.057584+00:00", "epoch": 0, "step": 44, "train_loss": 8.409912109375, "perplexity": 4491.365745258359, "lr": 0.0001882622950819672, "grad_norm": 1.475675, "tokens_per_sec": 107940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:41.363763+00:00", "epoch": 0, "step": 45, "train_loss": 8.364470481872559, "perplexity": 4291.838529936916, "lr": 0.00019254098360655736, "grad_norm": 1.456848, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:41.668479+00:00", "epoch": 0, "step": 46, "train_loss": 8.218618392944336, "perplexity": 3709.3739433709684, "lr": 0.00019681967213114755, "grad_norm": 1.477675, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:41.972560+00:00", "epoch": 0, "step": 47, "train_loss": 8.231513023376465, "perplexity": 3757.514660851729, "lr": 0.0002010983606557377, "grad_norm": 1.414211, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:42.276515+00:00", "epoch": 0, "step": 48, "train_loss": 8.14714241027832, "perplexity": 3453.49627626319, "lr": 0.00020537704918032788, "grad_norm": 1.386684, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:42.581448+00:00", "epoch": 0, "step": 49, "train_loss": 8.101913452148438, "perplexity": 3300.7779171655197, "lr": 0.00020965573770491804, "grad_norm": 1.395689, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:42.887748+00:00", "epoch": 0, "step": 50, "train_loss": 7.991215705871582, "perplexity": 2954.8870503314442, "lr": 0.0002139344262295082, "grad_norm": 1.377186, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:43.192949+00:00", "epoch": 0, "step": 51, "train_loss": 7.911733150482178, "perplexity": 2729.1163391325454, "lr": 0.00021821311475409837, "grad_norm": 1.349724, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:43.497498+00:00", "epoch": 0, "step": 52, "train_loss": 7.960239410400391, "perplexity": 2864.75872344584, "lr": 0.00022249180327868853, "grad_norm": 1.304432, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:43.802242+00:00", "epoch": 0, "step": 53, "train_loss": 7.817339897155762, "perplexity": 2483.290805248987, "lr": 0.0002267704918032787, "grad_norm": 1.260145, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:44.106800+00:00", "epoch": 0, "step": 54, "train_loss": 7.817830562591553, "perplexity": 2484.5095691923593, "lr": 0.0002310491803278688, "grad_norm": 1.259531, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:44.412920+00:00", "epoch": 0, "step": 55, "train_loss": 7.836222171783447, "perplexity": 2530.6264804085213, "lr": 0.000235327868852459, "grad_norm": 1.207905, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:44.717459+00:00", "epoch": 0, "step": 56, "train_loss": 7.709974765777588, "perplexity": 2230.48597289638, "lr": 0.00023960655737704915, "grad_norm": 1.157749, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:45.024398+00:00", "epoch": 0, "step": 57, "train_loss": 7.68793249130249, "perplexity": 2181.8588835158826, "lr": 0.00024388524590163933, "grad_norm": 1.109547, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:45.329516+00:00", "epoch": 0, "step": 58, "train_loss": 7.570980548858643, "perplexity": 1941.0426358718378, "lr": 0.0002481639344262295, "grad_norm": 1.076965, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:45.633834+00:00", "epoch": 0, "step": 59, "train_loss": 7.596461296081543, "perplexity": 1991.137367845125, "lr": 0.00025244262295081964, "grad_norm": 1.015529, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:45.938802+00:00", "epoch": 0, "step": 60, "train_loss": 7.55000638961792, "perplexity": 1900.7548763981977, "lr": 0.0002567213114754098, "grad_norm": 0.993215, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:46.245232+00:00", "epoch": 0, "step": 61, "train_loss": 7.562587738037109, "perplexity": 1924.8200041975242, "lr": 0.000261, "grad_norm": 1.149355, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:46.549086+00:00", "epoch": 0, "step": 62, "train_loss": 7.5320940017700195, "perplexity": 1867.010937344267, "lr": 0.00026527868852459013, "grad_norm": 0.87329, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:46.853729+00:00", "epoch": 0, "step": 63, "train_loss": 7.452426910400391, "perplexity": 1724.0421681364178, "lr": 0.0002695573770491803, "grad_norm": 0.820261, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:47.158165+00:00", "epoch": 0, "step": 64, "train_loss": 7.4336090087890625, "perplexity": 1691.9026597100237, "lr": 0.0002738360655737705, "grad_norm": 0.81194, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:47.462658+00:00", "epoch": 0, "step": 65, "train_loss": 7.409358024597168, "perplexity": 1651.3658702311827, "lr": 0.0002781147540983607, "grad_norm": 1.053301, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:47.767495+00:00", "epoch": 0, "step": 66, "train_loss": 7.412396430969238, "perplexity": 1656.391021173255, "lr": 0.0002823934426229508, "grad_norm": 1.561312, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:48.073358+00:00", "epoch": 0, "step": 67, "train_loss": 7.366036415100098, "perplexity": 1581.3535161628056, "lr": 0.000286672131147541, "grad_norm": 1.660176, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:48.378386+00:00", "epoch": 0, "step": 68, "train_loss": 7.236823558807373, "perplexity": 1389.672738960015, "lr": 0.00029095081967213117, "grad_norm": 1.57048, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:48.683091+00:00", "epoch": 0, "step": 69, "train_loss": 7.297084808349609, "perplexity": 1475.9908535355862, "lr": 0.0002952295081967213, "grad_norm": 1.119119, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:48.988055+00:00", "epoch": 0, "step": 70, "train_loss": 7.29571533203125, "perplexity": 1473.970902468825, "lr": 0.0002995081967213115, "grad_norm": 0.804197, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:49.293375+00:00", "epoch": 0, "step": 71, "train_loss": 7.257749080657959, "perplexity": 1419.0587527798202, "lr": 0.00030378688524590166, "grad_norm": 0.599939, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:49.599267+00:00", "epoch": 0, "step": 72, "train_loss": 7.335907936096191, "perplexity": 1534.4203032645175, "lr": 0.0003080655737704918, "grad_norm": 0.95677, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:49.904493+00:00", "epoch": 0, "step": 73, "train_loss": 7.261074542999268, "perplexity": 1423.7856343996068, "lr": 0.00031234426229508197, "grad_norm": 0.773565, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:50.210345+00:00", "epoch": 0, "step": 74, "train_loss": 7.044577121734619, "perplexity": 1146.6238509731666, "lr": 0.00031662295081967215, "grad_norm": 0.729418, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:50.515880+00:00", "epoch": 0, "step": 75, "train_loss": 7.133010387420654, "perplexity": 1252.6422350387559, "lr": 0.0003209016393442623, "grad_norm": 0.772406, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:50.820998+00:00", "epoch": 0, "step": 76, "train_loss": 7.197504997253418, "perplexity": 1336.093046497509, "lr": 0.0003251803278688524, "grad_norm": 0.6893, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:51.126446+00:00", "epoch": 0, "step": 77, "train_loss": 7.101305961608887, "perplexity": 1213.550890939732, "lr": 0.0003294590163934426, "grad_norm": 0.523308, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:51.432567+00:00", "epoch": 0, "step": 78, "train_loss": 7.104081153869629, "perplexity": 1216.923405503938, "lr": 0.00033373770491803277, "grad_norm": 0.915879, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:51.737982+00:00", "epoch": 0, "step": 79, "train_loss": 7.060811519622803, "perplexity": 1165.390519409139, "lr": 0.00033801639344262295, "grad_norm": 1.019759, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:52.042851+00:00", "epoch": 0, "step": 80, "train_loss": 7.040471076965332, "perplexity": 1141.9254147035185, "lr": 0.0003422950819672131, "grad_norm": 0.657174, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:52.348139+00:00", "epoch": 0, "step": 81, "train_loss": 7.018161773681641, "perplexity": 1116.7319238159273, "lr": 0.00034657377049180326, "grad_norm": 0.653193, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:52.653803+00:00", "epoch": 0, "step": 82, "train_loss": 7.020668029785156, "perplexity": 1119.534250222954, "lr": 0.00035085245901639344, "grad_norm": 0.929799, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:52.960239+00:00", "epoch": 0, "step": 83, "train_loss": 7.121338844299316, "perplexity": 1238.106956745798, "lr": 0.00035513114754098357, "grad_norm": 0.490916, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:53.266838+00:00", "epoch": 0, "step": 84, "train_loss": 6.953841209411621, "perplexity": 1047.1643903767022, "lr": 0.00035940983606557375, "grad_norm": 1.351784, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:53.572655+00:00", "epoch": 0, "step": 85, "train_loss": 7.1016130447387695, "perplexity": 1213.9236091704051, "lr": 0.00036368852459016393, "grad_norm": 1.148849, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:53.878044+00:00", "epoch": 0, "step": 86, "train_loss": 6.988787651062012, "perplexity": 1084.406000476228, "lr": 0.0003679672131147541, "grad_norm": 0.673396, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:54.183938+00:00", "epoch": 0, "step": 87, "train_loss": 7.038981914520264, "perplexity": 1140.2261678022296, "lr": 0.00037224590163934424, "grad_norm": 1.36537, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:54.489307+00:00", "epoch": 0, "step": 88, "train_loss": 7.053116798400879, "perplexity": 1156.4575765756535, "lr": 0.0003765245901639344, "grad_norm": 1.151841, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:54.795901+00:00", "epoch": 0, "step": 89, "train_loss": 6.949464797973633, "perplexity": 1042.5915817058617, "lr": 0.0003808032786885246, "grad_norm": 1.035206, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:55.101612+00:00", "epoch": 0, "step": 90, "train_loss": 6.90904426574707, "perplexity": 1001.2898178654275, "lr": 0.00038508196721311473, "grad_norm": 0.702775, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:55.407522+00:00", "epoch": 0, "step": 91, "train_loss": 6.881433010101318, "perplexity": 974.0211423203165, "lr": 0.0003893606557377049, "grad_norm": 1.247231, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:55.713244+00:00", "epoch": 0, "step": 92, "train_loss": 6.922801971435547, "perplexity": 1015.1604638417787, "lr": 0.0003936393442622951, "grad_norm": 0.997952, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:56.018849+00:00", "epoch": 0, "step": 93, "train_loss": 7.010868549346924, "perplexity": 1108.6169754136145, "lr": 0.0003979180327868853, "grad_norm": 0.815323, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:56.324755+00:00", "epoch": 0, "step": 94, "train_loss": 6.9590325355529785, "perplexity": 1052.6146971704168, "lr": 0.0004021967213114754, "grad_norm": 0.653986, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:56.631068+00:00", "epoch": 0, "step": 95, "train_loss": 6.901897430419922, "perplexity": 994.1592751802614, "lr": 0.0004064754098360656, "grad_norm": 0.629168, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:56.936052+00:00", "epoch": 0, "step": 96, "train_loss": 6.876598834991455, "perplexity": 969.3239163117092, "lr": 0.00041075409836065577, "grad_norm": 0.587599, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:57.242335+00:00", "epoch": 0, "step": 97, "train_loss": 6.872303485870361, "perplexity": 965.1692609152166, "lr": 0.0004150327868852459, "grad_norm": 0.507508, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:57.547008+00:00", "epoch": 0, "step": 98, "train_loss": 6.741440773010254, "perplexity": 846.7798749782628, "lr": 0.0004193114754098361, "grad_norm": 0.517, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:57.852434+00:00", "epoch": 0, "step": 99, "train_loss": 6.807248115539551, "perplexity": 904.3786339251459, "lr": 0.00042359016393442626, "grad_norm": 0.403713, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:42:58.159751+00:00", "epoch": 0, "step": 100, "train_loss": 6.835635185241699, "perplexity": 930.4191515853602, "lr": 0.0004278688524590164, "grad_norm": 0.436401, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:43:42.622597+00:00", "step": 100, "epoch": 0, "val_loss": 6.763040566444397, "val_ppl": 865.2691083305112, "eval_train_loss": 6.835635185241699, "eval_train_ppl": 930.4191515853602} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:43:43.623213+00:00", "step": 100, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_6p7630_epoch_0000_step_0000100.pt", "val_loss": 6.763040566444397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:45.183684+00:00", "epoch": 0, "step": 101, "train_loss": 6.692477703094482, "perplexity": 806.3175947896177, "lr": 0.00043214754098360657, "grad_norm": 0.399606, "tokens_per_sec": 697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:45.489754+00:00", "epoch": 0, "step": 102, "train_loss": 6.840106964111328, "perplexity": 934.5890968751381, "lr": 0.00043642622950819675, "grad_norm": 0.429574, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:45.796118+00:00", "epoch": 0, "step": 103, "train_loss": 6.665544509887695, "perplexity": 784.8907294084912, "lr": 0.00044070491803278693, "grad_norm": 0.495855, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:46.101108+00:00", "epoch": 0, "step": 104, "train_loss": 6.679967403411865, "perplexity": 796.2931551528422, "lr": 0.00044498360655737706, "grad_norm": 0.6014, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:46.405822+00:00", "epoch": 0, "step": 105, "train_loss": 6.848873615264893, "perplexity": 942.8183321727364, "lr": 0.00044926229508196724, "grad_norm": 0.614842, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:46.711033+00:00", "epoch": 0, "step": 106, "train_loss": 6.773763656616211, "perplexity": 874.5973916295515, "lr": 0.0004535409836065574, "grad_norm": 0.660604, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:47.016763+00:00", "epoch": 0, "step": 107, "train_loss": 6.699807643890381, "perplexity": 812.2495689724807, "lr": 0.0004578196721311475, "grad_norm": 0.870204, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:47.321406+00:00", "epoch": 0, "step": 108, "train_loss": 6.590205192565918, "perplexity": 727.9302204452359, "lr": 0.0004620983606557376, "grad_norm": 1.035473, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:47.627790+00:00", "epoch": 0, "step": 109, "train_loss": 6.571826457977295, "perplexity": 714.6739737809897, "lr": 0.0004663770491803278, "grad_norm": 1.04769, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:47.933470+00:00", "epoch": 0, "step": 110, "train_loss": 6.745247840881348, "perplexity": 850.0097677435601, "lr": 0.000470655737704918, "grad_norm": 0.982306, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:48.239844+00:00", "epoch": 0, "step": 111, "train_loss": 6.6020097732543945, "perplexity": 736.5740494818411, "lr": 0.00047493442622950817, "grad_norm": 1.059809, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:48.544962+00:00", "epoch": 0, "step": 112, "train_loss": 6.608227729797363, "perplexity": 741.1683035459689, "lr": 0.0004792131147540983, "grad_norm": 0.945058, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:48.851168+00:00", "epoch": 0, "step": 113, "train_loss": 6.623857021331787, "perplexity": 752.8432368496331, "lr": 0.0004834918032786885, "grad_norm": 0.636143, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:49.157192+00:00", "epoch": 0, "step": 114, "train_loss": 6.539892196655273, "perplexity": 692.2119512504509, "lr": 0.00048777049180327866, "grad_norm": 0.742277, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:49.461921+00:00", "epoch": 0, "step": 115, "train_loss": 6.595344543457031, "perplexity": 731.6809391464811, "lr": 0.0004920491803278688, "grad_norm": 0.636401, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:49.767040+00:00", "epoch": 0, "step": 116, "train_loss": 6.553494453430176, "perplexity": 701.6919243018162, "lr": 0.000496327868852459, "grad_norm": 0.944121, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:50.072412+00:00", "epoch": 0, "step": 117, "train_loss": 6.648255348205566, "perplexity": 771.4372613955128, "lr": 0.0005006065573770491, "grad_norm": 1.062477, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:50.378441+00:00", "epoch": 0, "step": 118, "train_loss": 6.4962334632873535, "perplexity": 662.6410648615762, "lr": 0.0005048852459016393, "grad_norm": 1.286787, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:50.686228+00:00", "epoch": 0, "step": 119, "train_loss": 6.638172149658203, "perplexity": 763.6977912024577, "lr": 0.0005091639344262295, "grad_norm": 1.177764, "tokens_per_sec": 106464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:50.993814+00:00", "epoch": 0, "step": 120, "train_loss": 6.564681529998779, "perplexity": 709.5858783933369, "lr": 0.0005134426229508196, "grad_norm": 0.801928, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:51.301242+00:00", "epoch": 0, "step": 121, "train_loss": 6.5399394035339355, "perplexity": 692.2446291873476, "lr": 0.0005177213114754098, "grad_norm": 0.704334, "tokens_per_sec": 106534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:51.606463+00:00", "epoch": 0, "step": 122, "train_loss": 6.566887855529785, "perplexity": 711.1531841908884, "lr": 0.000522, "grad_norm": 0.886783, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:51.913045+00:00", "epoch": 0, "step": 123, "train_loss": 6.525075912475586, "perplexity": 682.0315463886861, "lr": 0.0005262786885245901, "grad_norm": 0.714637, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:52.219076+00:00", "epoch": 0, "step": 124, "train_loss": 6.492428779602051, "perplexity": 660.1247152056027, "lr": 0.0005305573770491803, "grad_norm": 0.84265, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:52.525856+00:00", "epoch": 0, "step": 125, "train_loss": 6.507431983947754, "perplexity": 670.1033699203518, "lr": 0.0005348360655737705, "grad_norm": 1.197786, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:52.831963+00:00", "epoch": 0, "step": 126, "train_loss": 6.532027244567871, "perplexity": 686.7890906426422, "lr": 0.0005391147540983606, "grad_norm": 0.964281, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:53.138038+00:00", "epoch": 0, "step": 127, "train_loss": 6.384025573730469, "perplexity": 592.3072915229643, "lr": 0.0005433934426229507, "grad_norm": 0.548385, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:53.443896+00:00", "epoch": 0, "step": 128, "train_loss": 6.4506001472473145, "perplexity": 633.0821213167695, "lr": 0.000547672131147541, "grad_norm": 0.566593, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:53.751221+00:00", "epoch": 0, "step": 129, "train_loss": 6.377052307128906, "perplexity": 588.1913423532167, "lr": 0.0005519508196721311, "grad_norm": 0.639072, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:54.058456+00:00", "epoch": 0, "step": 130, "train_loss": 6.496126174926758, "perplexity": 662.5699750016893, "lr": 0.0005562295081967213, "grad_norm": 0.639211, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:54.365621+00:00", "epoch": 0, "step": 131, "train_loss": 6.527984619140625, "perplexity": 684.0182640822013, "lr": 0.0005605081967213115, "grad_norm": 0.603716, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:54.671043+00:00", "epoch": 0, "step": 132, "train_loss": 6.4189019203186035, "perplexity": 613.3292598221856, "lr": 0.0005647868852459016, "grad_norm": 0.69622, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:54.976293+00:00", "epoch": 0, "step": 133, "train_loss": 6.503991603851318, "perplexity": 667.8019208237926, "lr": 0.0005690655737704918, "grad_norm": 1.131654, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:55.281792+00:00", "epoch": 0, "step": 134, "train_loss": 6.441305160522461, "perplexity": 627.2248949942405, "lr": 0.000573344262295082, "grad_norm": 1.526806, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:55.589957+00:00", "epoch": 0, "step": 135, "train_loss": 6.454780101776123, "perplexity": 635.7339141225357, "lr": 0.0005776229508196721, "grad_norm": 1.260026, "tokens_per_sec": 106333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:55.897404+00:00", "epoch": 0, "step": 136, "train_loss": 6.486633777618408, "perplexity": 656.3103539615217, "lr": 0.0005819016393442623, "grad_norm": 0.906714, "tokens_per_sec": 106581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:56.204001+00:00", "epoch": 0, "step": 137, "train_loss": 6.399447917938232, "perplexity": 601.5128617253921, "lr": 0.0005861803278688525, "grad_norm": 0.803191, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:56.510973+00:00", "epoch": 0, "step": 138, "train_loss": 6.285918235778809, "perplexity": 536.9571175742507, "lr": 0.0005904590163934426, "grad_norm": 0.598439, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:56.816359+00:00", "epoch": 0, "step": 139, "train_loss": 6.344669818878174, "perplexity": 569.4493372504527, "lr": 0.0005947377049180328, "grad_norm": 0.594781, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:57.122668+00:00", "epoch": 0, "step": 140, "train_loss": 6.2378387451171875, "perplexity": 511.7512899062855, "lr": 0.000599016393442623, "grad_norm": 0.843845, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:57.430232+00:00", "epoch": 0, "step": 141, "train_loss": 6.269724369049072, "perplexity": 528.3317332284879, "lr": 0.0006032950819672131, "grad_norm": 1.02954, "tokens_per_sec": 106479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:57.737728+00:00", "epoch": 0, "step": 142, "train_loss": 6.373065948486328, "perplexity": 585.8512679988042, "lr": 0.0006075737704918033, "grad_norm": 0.916866, "tokens_per_sec": 106564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:58.044526+00:00", "epoch": 0, "step": 143, "train_loss": 6.3388991355896, "perplexity": 566.1726888193726, "lr": 0.0006118524590163934, "grad_norm": 0.922427, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:58.351528+00:00", "epoch": 0, "step": 144, "train_loss": 6.2556023597717285, "perplexity": 520.9230633687833, "lr": 0.0006161311475409836, "grad_norm": 0.68267, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:58.657739+00:00", "epoch": 0, "step": 145, "train_loss": 6.251300811767578, "perplexity": 518.6871003041887, "lr": 0.0006204098360655738, "grad_norm": 0.588814, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:58.964631+00:00", "epoch": 0, "step": 146, "train_loss": 6.245206832885742, "perplexity": 515.5358436665895, "lr": 0.0006246885245901639, "grad_norm": 0.511014, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:59.271841+00:00", "epoch": 0, "step": 147, "train_loss": 6.236759662628174, "perplexity": 511.199365889944, "lr": 0.0006289672131147542, "grad_norm": 0.610496, "tokens_per_sec": 106663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:59.579852+00:00", "epoch": 0, "step": 148, "train_loss": 6.337179183959961, "perplexity": 565.199736136209, "lr": 0.0006332459016393443, "grad_norm": 0.476116, "tokens_per_sec": 106386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:43:59.886433+00:00", "epoch": 0, "step": 149, "train_loss": 6.252040863037109, "perplexity": 519.0710974225103, "lr": 0.0006375245901639344, "grad_norm": 0.513332, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:00.192159+00:00", "epoch": 0, "step": 150, "train_loss": 6.2533769607543945, "perplexity": 519.7650906490275, "lr": 0.0006418032786885245, "grad_norm": 0.538112, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:00.497664+00:00", "epoch": 0, "step": 151, "train_loss": 6.223135948181152, "perplexity": 504.2821577269995, "lr": 0.0006460819672131147, "grad_norm": 0.472401, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:00.804087+00:00", "epoch": 0, "step": 152, "train_loss": 6.209692001342773, "perplexity": 497.54798357404724, "lr": 0.0006503606557377048, "grad_norm": 0.612077, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:01.111611+00:00", "epoch": 0, "step": 153, "train_loss": 6.288570404052734, "perplexity": 538.3831083538242, "lr": 0.000654639344262295, "grad_norm": 0.899483, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:01.418064+00:00", "epoch": 0, "step": 154, "train_loss": 6.286209583282471, "perplexity": 537.1135814816072, "lr": 0.0006589180327868852, "grad_norm": 1.053477, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:01.724932+00:00", "epoch": 0, "step": 155, "train_loss": 6.214741230010986, "perplexity": 500.0665702255989, "lr": 0.0006631967213114754, "grad_norm": 0.826278, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:02.030488+00:00", "epoch": 0, "step": 156, "train_loss": 6.320268630981445, "perplexity": 555.7222566172217, "lr": 0.0006674754098360655, "grad_norm": 1.418608, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:02.336774+00:00", "epoch": 0, "step": 157, "train_loss": 6.243610382080078, "perplexity": 514.7134726657043, "lr": 0.0006717540983606557, "grad_norm": 1.312616, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:02.644540+00:00", "epoch": 0, "step": 158, "train_loss": 6.1830735206604, "perplexity": 484.478725747268, "lr": 0.0006760327868852459, "grad_norm": 0.484063, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:02.952444+00:00", "epoch": 0, "step": 159, "train_loss": 6.176769733428955, "perplexity": 481.4342807902242, "lr": 0.000680311475409836, "grad_norm": 0.852984, "tokens_per_sec": 106422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:03.260251+00:00", "epoch": 0, "step": 160, "train_loss": 6.2197957038879395, "perplexity": 502.6005421943402, "lr": 0.0006845901639344261, "grad_norm": 0.61687, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:03.566260+00:00", "epoch": 0, "step": 161, "train_loss": 6.2529730796813965, "perplexity": 519.555209752823, "lr": 0.0006888688524590164, "grad_norm": 0.746061, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:03.872755+00:00", "epoch": 0, "step": 162, "train_loss": 6.159362316131592, "perplexity": 473.12627362576933, "lr": 0.0006931475409836065, "grad_norm": 0.671157, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:04.180065+00:00", "epoch": 0, "step": 163, "train_loss": 6.200905799865723, "perplexity": 493.195575313301, "lr": 0.0006974262295081966, "grad_norm": 0.682541, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:04.489088+00:00", "epoch": 0, "step": 164, "train_loss": 6.130842208862305, "perplexity": 459.8232648745322, "lr": 0.0007017049180327869, "grad_norm": 0.694484, "tokens_per_sec": 106038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:04.797756+00:00", "epoch": 0, "step": 165, "train_loss": 6.20051383972168, "perplexity": 493.0023001851078, "lr": 0.000705983606557377, "grad_norm": 0.668683, "tokens_per_sec": 106160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:05.104622+00:00", "epoch": 0, "step": 166, "train_loss": 6.186433792114258, "perplexity": 486.1094440730901, "lr": 0.0007102622950819671, "grad_norm": 0.573667, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:05.411695+00:00", "epoch": 0, "step": 167, "train_loss": 6.120718002319336, "perplexity": 455.1914056808169, "lr": 0.0007145409836065574, "grad_norm": 0.592993, "tokens_per_sec": 106711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:05.718346+00:00", "epoch": 0, "step": 168, "train_loss": 6.109969615936279, "perplexity": 450.3250322552756, "lr": 0.0007188196721311475, "grad_norm": 0.792989, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:06.025968+00:00", "epoch": 0, "step": 169, "train_loss": 6.115706443786621, "perplexity": 452.91589400023895, "lr": 0.0007230983606557376, "grad_norm": 1.469369, "tokens_per_sec": 106520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:06.333608+00:00", "epoch": 0, "step": 170, "train_loss": 6.254404544830322, "perplexity": 520.2994674909462, "lr": 0.0007273770491803279, "grad_norm": 1.257268, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:06.642805+00:00", "epoch": 0, "step": 171, "train_loss": 6.130882740020752, "perplexity": 459.8419024218367, "lr": 0.000731655737704918, "grad_norm": 0.762353, "tokens_per_sec": 105927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:06.951011+00:00", "epoch": 0, "step": 172, "train_loss": 6.184006214141846, "perplexity": 484.93080669037306, "lr": 0.0007359344262295082, "grad_norm": 0.781248, "tokens_per_sec": 106318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:07.257952+00:00", "epoch": 0, "step": 173, "train_loss": 6.0647382736206055, "perplexity": 430.4100132747903, "lr": 0.0007402131147540983, "grad_norm": 0.545532, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:07.565270+00:00", "epoch": 0, "step": 174, "train_loss": 6.025798797607422, "perplexity": 413.9721899798889, "lr": 0.0007444918032786885, "grad_norm": 0.60806, "tokens_per_sec": 106626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:07.872884+00:00", "epoch": 0, "step": 175, "train_loss": 6.061284065246582, "perplexity": 428.92585217921095, "lr": 0.0007487704918032787, "grad_norm": 0.562246, "tokens_per_sec": 106524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:08.180881+00:00", "epoch": 0, "step": 176, "train_loss": 6.245929718017578, "perplexity": 515.9086515953243, "lr": 0.0007530491803278688, "grad_norm": 0.6244, "tokens_per_sec": 106391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:08.488610+00:00", "epoch": 0, "step": 177, "train_loss": 6.003958702087402, "perplexity": 405.02901320641746, "lr": 0.000757327868852459, "grad_norm": 0.574322, "tokens_per_sec": 106482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:08.796235+00:00", "epoch": 0, "step": 178, "train_loss": 5.928620338439941, "perplexity": 375.6359057354582, "lr": 0.0007616065573770492, "grad_norm": 0.667976, "tokens_per_sec": 106520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:09.102784+00:00", "epoch": 0, "step": 179, "train_loss": 6.191844463348389, "perplexity": 488.7467508246576, "lr": 0.0007658852459016393, "grad_norm": 0.841627, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:09.410006+00:00", "epoch": 0, "step": 180, "train_loss": 6.069311618804932, "perplexity": 432.38293482320097, "lr": 0.0007701639344262295, "grad_norm": 1.030391, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:09.718464+00:00", "epoch": 0, "step": 181, "train_loss": 6.111030101776123, "perplexity": 450.80284888941617, "lr": 0.0007744426229508197, "grad_norm": 0.80024, "tokens_per_sec": 106181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:10.025819+00:00", "epoch": 0, "step": 182, "train_loss": 6.057037830352783, "perplexity": 427.1083936675183, "lr": 0.0007787213114754098, "grad_norm": 0.810014, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:10.333662+00:00", "epoch": 0, "step": 183, "train_loss": 6.077256679534912, "perplexity": 435.83192658063916, "lr": 0.000783, "grad_norm": 0.824669, "tokens_per_sec": 106444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:10.641950+00:00", "epoch": 0, "step": 184, "train_loss": 6.0830888748168945, "perplexity": 438.3812102205416, "lr": 0.0007872786885245902, "grad_norm": 0.863518, "tokens_per_sec": 106292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:10.949028+00:00", "epoch": 0, "step": 185, "train_loss": 6.067134380340576, "perplexity": 431.4425581497827, "lr": 0.0007915573770491803, "grad_norm": 0.7718, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:11.257492+00:00", "epoch": 0, "step": 186, "train_loss": 5.914393901824951, "perplexity": 370.3297784664491, "lr": 0.0007958360655737705, "grad_norm": 0.631914, "tokens_per_sec": 106232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:11.566128+00:00", "epoch": 0, "step": 187, "train_loss": 5.957542896270752, "perplexity": 386.6588950689694, "lr": 0.0008001147540983607, "grad_norm": 0.588753, "tokens_per_sec": 106169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:11.874034+00:00", "epoch": 0, "step": 188, "train_loss": 6.008262634277344, "perplexity": 406.7759873466147, "lr": 0.0008043934426229508, "grad_norm": 0.465035, "tokens_per_sec": 106421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:12.183466+00:00", "epoch": 0, "step": 189, "train_loss": 6.014820575714111, "perplexity": 409.45236662698545, "lr": 0.000808672131147541, "grad_norm": 0.436524, "tokens_per_sec": 105898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:12.490150+00:00", "epoch": 0, "step": 190, "train_loss": 5.909595966339111, "perplexity": 368.55721580192824, "lr": 0.0008129508196721312, "grad_norm": 0.464837, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:12.797079+00:00", "epoch": 0, "step": 191, "train_loss": 5.851454257965088, "perplexity": 347.7397161974854, "lr": 0.0008172295081967213, "grad_norm": 0.428763, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:13.105545+00:00", "epoch": 0, "step": 192, "train_loss": 5.954270362854004, "perplexity": 385.39560911415157, "lr": 0.0008215081967213115, "grad_norm": 0.362081, "tokens_per_sec": 106228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:13.414209+00:00", "epoch": 0, "step": 193, "train_loss": 5.92501163482666, "perplexity": 374.28279005032294, "lr": 0.0008257868852459017, "grad_norm": 0.405718, "tokens_per_sec": 106160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:13.722031+00:00", "epoch": 0, "step": 194, "train_loss": 5.964443683624268, "perplexity": 389.33637361190773, "lr": 0.0008300655737704918, "grad_norm": 0.434922, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:14.028960+00:00", "epoch": 0, "step": 195, "train_loss": 5.979955196380615, "perplexity": 395.42265139247024, "lr": 0.000834344262295082, "grad_norm": 0.43459, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:14.336792+00:00", "epoch": 0, "step": 196, "train_loss": 5.938108444213867, "perplexity": 379.21694069587386, "lr": 0.0008386229508196722, "grad_norm": 0.640711, "tokens_per_sec": 106450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:14.644025+00:00", "epoch": 0, "step": 197, "train_loss": 5.956282138824463, "perplexity": 386.17171915769705, "lr": 0.0008429016393442623, "grad_norm": 0.738322, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:14.952311+00:00", "epoch": 0, "step": 198, "train_loss": 5.928311824798584, "perplexity": 375.52003480915494, "lr": 0.0008471803278688525, "grad_norm": 0.88543, "tokens_per_sec": 106291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:15.260573+00:00", "epoch": 0, "step": 199, "train_loss": 5.952460289001465, "perplexity": 384.6986455672493, "lr": 0.0008514590163934426, "grad_norm": 1.104285, "tokens_per_sec": 106300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:15.569720+00:00", "epoch": 0, "step": 200, "train_loss": 5.934554100036621, "perplexity": 377.87146572784667, "lr": 0.0008557377049180328, "grad_norm": 1.106892, "tokens_per_sec": 106050} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:44:21.068956+00:00", "step": 200, "epoch": 0, "val_loss": 5.90152792930603, "val_ppl": 365.59564562976936, "eval_train_loss": 5.934554100036621, "eval_train_ppl": 377.87146572784667} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:44:21.947998+00:00", "step": 200, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_5p9015_epoch_0000_step_0000200.pt", "val_loss": 5.90152792930603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:22.901632+00:00", "epoch": 0, "step": 201, "train_loss": 5.834853172302246, "perplexity": 342.01451315988976, "lr": 0.000860016393442623, "grad_norm": 1.015708, "tokens_per_sec": 4469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:23.208469+00:00", "epoch": 0, "step": 202, "train_loss": 5.896261215209961, "perplexity": 363.6752194962708, "lr": 0.0008642950819672131, "grad_norm": 0.743609, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:23.514673+00:00", "epoch": 0, "step": 203, "train_loss": 5.941699981689453, "perplexity": 380.5813612669126, "lr": 0.0008685737704918034, "grad_norm": 0.899736, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:23.822301+00:00", "epoch": 0, "step": 204, "train_loss": 5.842419147491455, "perplexity": 344.61200035176716, "lr": 0.0008728524590163935, "grad_norm": 0.829391, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:24.130571+00:00", "epoch": 0, "step": 205, "train_loss": 5.906120777130127, "perplexity": 367.27863268857993, "lr": 0.0008771311475409836, "grad_norm": 0.653005, "tokens_per_sec": 106296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:24.439357+00:00", "epoch": 0, "step": 206, "train_loss": 5.821360111236572, "perplexity": 337.43068493071513, "lr": 0.0008814098360655739, "grad_norm": 0.73668, "tokens_per_sec": 106120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:24.746892+00:00", "epoch": 0, "step": 207, "train_loss": 5.811469078063965, "perplexity": 334.10959839696307, "lr": 0.000885688524590164, "grad_norm": 0.802041, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:25.052249+00:00", "epoch": 0, "step": 208, "train_loss": 5.781966686248779, "perplexity": 324.39654970238865, "lr": 0.0008899672131147541, "grad_norm": 0.767099, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:25.359939+00:00", "epoch": 0, "step": 209, "train_loss": 5.917727470397949, "perplexity": 371.5663581440232, "lr": 0.0008942459016393444, "grad_norm": 0.628371, "tokens_per_sec": 106498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:25.668206+00:00", "epoch": 0, "step": 210, "train_loss": 5.77346658706665, "perplexity": 321.6508328203592, "lr": 0.0008985245901639345, "grad_norm": 0.548813, "tokens_per_sec": 106365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:25.977222+00:00", "epoch": 0, "step": 211, "train_loss": 5.875359535217285, "perplexity": 356.152687088136, "lr": 0.0009028032786885246, "grad_norm": 0.444238, "tokens_per_sec": 105973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:26.285091+00:00", "epoch": 0, "step": 212, "train_loss": 5.7417378425598145, "perplexity": 311.60546194816965, "lr": 0.0009070819672131148, "grad_norm": 0.380335, "tokens_per_sec": 106436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:26.592528+00:00", "epoch": 0, "step": 213, "train_loss": 5.738426208496094, "perplexity": 310.57524547744435, "lr": 0.000911360655737705, "grad_norm": 0.442569, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:26.901189+00:00", "epoch": 0, "step": 214, "train_loss": 5.82953405380249, "perplexity": 340.20012718063975, "lr": 0.000915639344262295, "grad_norm": 0.437864, "tokens_per_sec": 106166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:27.209819+00:00", "epoch": 0, "step": 215, "train_loss": 5.809223651885986, "perplexity": 333.360221608397, "lr": 0.0009199180327868851, "grad_norm": 0.380124, "tokens_per_sec": 106170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:27.517365+00:00", "epoch": 0, "step": 216, "train_loss": 5.665737152099609, "perplexity": 288.8007927804631, "lr": 0.0009241967213114752, "grad_norm": 0.470343, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:27.825120+00:00", "epoch": 0, "step": 217, "train_loss": 5.767750263214111, "perplexity": 319.81741768230324, "lr": 0.0009284754098360655, "grad_norm": 0.46127, "tokens_per_sec": 106475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:28.132080+00:00", "epoch": 0, "step": 218, "train_loss": 5.688125133514404, "perplexity": 295.3393793344206, "lr": 0.0009327540983606556, "grad_norm": 0.490056, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:28.438569+00:00", "epoch": 0, "step": 219, "train_loss": 5.71214485168457, "perplexity": 302.5192318109536, "lr": 0.0009370327868852458, "grad_norm": 0.667185, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:28.746978+00:00", "epoch": 0, "step": 220, "train_loss": 5.833141803741455, "perplexity": 341.42970083218177, "lr": 0.000941311475409836, "grad_norm": 0.93767, "tokens_per_sec": 106315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:29.055920+00:00", "epoch": 0, "step": 221, "train_loss": 5.654776096343994, "perplexity": 285.65251692280526, "lr": 0.0009455901639344261, "grad_norm": 0.609282, "tokens_per_sec": 106001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:29.363565+00:00", "epoch": 0, "step": 222, "train_loss": 5.726300239562988, "perplexity": 306.83196104210447, "lr": 0.0009498688524590163, "grad_norm": 0.546797, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:29.671825+00:00", "epoch": 0, "step": 223, "train_loss": 5.874211311340332, "perplexity": 355.74397875821603, "lr": 0.0009541475409836065, "grad_norm": 0.730733, "tokens_per_sec": 106300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:29.978446+00:00", "epoch": 0, "step": 224, "train_loss": 5.852284908294678, "perplexity": 348.0286863072995, "lr": 0.0009584262295081966, "grad_norm": 1.143116, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:30.286512+00:00", "epoch": 0, "step": 225, "train_loss": 5.727487564086914, "perplexity": 307.19648651636777, "lr": 0.0009627049180327868, "grad_norm": 1.207752, "tokens_per_sec": 106368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:30.593388+00:00", "epoch": 0, "step": 226, "train_loss": 5.922904968261719, "perplexity": 373.4951309695214, "lr": 0.000966983606557377, "grad_norm": 1.304963, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:30.901993+00:00", "epoch": 0, "step": 227, "train_loss": 5.7712578773498535, "perplexity": 320.9411834936708, "lr": 0.0009712622950819671, "grad_norm": 1.382043, "tokens_per_sec": 106180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:31.210308+00:00", "epoch": 0, "step": 228, "train_loss": 5.738298416137695, "perplexity": 310.53555887024123, "lr": 0.0009755409836065573, "grad_norm": 1.448253, "tokens_per_sec": 106281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:31.518393+00:00", "epoch": 0, "step": 229, "train_loss": 5.763561248779297, "perplexity": 318.4805000417752, "lr": 0.0009798196721311474, "grad_norm": 1.253211, "tokens_per_sec": 106365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:31.825988+00:00", "epoch": 0, "step": 230, "train_loss": 5.84033203125, "perplexity": 343.8935051012031, "lr": 0.0009840983606557376, "grad_norm": 1.207193, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:32.134396+00:00", "epoch": 0, "step": 231, "train_loss": 5.8533782958984375, "perplexity": 348.40942466803165, "lr": 0.0009883770491803277, "grad_norm": 1.343233, "tokens_per_sec": 106173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:32.442099+00:00", "epoch": 0, "step": 232, "train_loss": 5.879309177398682, "perplexity": 357.56214435776394, "lr": 0.000992655737704918, "grad_norm": 0.955129, "tokens_per_sec": 106492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:32.753127+00:00", "epoch": 0, "step": 233, "train_loss": 5.616541385650635, "perplexity": 274.9368364506083, "lr": 0.0009969344262295082, "grad_norm": 0.865716, "tokens_per_sec": 105355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:33.060940+00:00", "epoch": 0, "step": 234, "train_loss": 5.884246826171875, "perplexity": 359.33202657370396, "lr": 0.0010012131147540983, "grad_norm": 0.646221, "tokens_per_sec": 106454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:33.369193+00:00", "epoch": 0, "step": 235, "train_loss": 5.708376407623291, "perplexity": 301.381350376284, "lr": 0.0010054918032786884, "grad_norm": 0.546881, "tokens_per_sec": 106307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:33.676513+00:00", "epoch": 0, "step": 236, "train_loss": 5.637134552001953, "perplexity": 280.65735610515105, "lr": 0.0010097704918032786, "grad_norm": 0.517661, "tokens_per_sec": 106620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:33.985200+00:00", "epoch": 0, "step": 237, "train_loss": 5.778406620025635, "perplexity": 323.2437297781021, "lr": 0.0010140491803278687, "grad_norm": 0.397019, "tokens_per_sec": 106153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:34.293727+00:00", "epoch": 0, "step": 238, "train_loss": 5.720009803771973, "perplexity": 304.90791218989943, "lr": 0.001018327868852459, "grad_norm": 0.40621, "tokens_per_sec": 106208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:34.601846+00:00", "epoch": 0, "step": 239, "train_loss": 5.8281378746032715, "perplexity": 339.72547826407407, "lr": 0.0010226065573770492, "grad_norm": 0.347052, "tokens_per_sec": 106348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:34.908339+00:00", "epoch": 0, "step": 240, "train_loss": 5.60313081741333, "perplexity": 271.2743898779985, "lr": 0.0010268852459016393, "grad_norm": 0.319467, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:35.215478+00:00", "epoch": 0, "step": 241, "train_loss": 5.774834632873535, "perplexity": 322.0911670234846, "lr": 0.0010311639344262294, "grad_norm": 0.342826, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:35.523325+00:00", "epoch": 0, "step": 242, "train_loss": 5.774580001831055, "perplexity": 322.0091630546728, "lr": 0.0010354426229508195, "grad_norm": 0.307339, "tokens_per_sec": 106442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:35.831435+00:00", "epoch": 0, "step": 243, "train_loss": 5.535865783691406, "perplexity": 253.6272789691979, "lr": 0.0010397213114754099, "grad_norm": 0.319784, "tokens_per_sec": 106351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:36.139771+00:00", "epoch": 0, "step": 244, "train_loss": 5.575543403625488, "perplexity": 263.892917049721, "lr": 0.001044, "grad_norm": 0.293355, "tokens_per_sec": 106275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:36.447312+00:00", "epoch": 0, "step": 245, "train_loss": 5.575479984283447, "perplexity": 263.87618166523157, "lr": 0.0010482786885245901, "grad_norm": 0.337054, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:36.754042+00:00", "epoch": 0, "step": 246, "train_loss": 5.583226680755615, "perplexity": 265.928288613196, "lr": 0.0010525573770491803, "grad_norm": 0.363014, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:37.062895+00:00", "epoch": 0, "step": 247, "train_loss": 5.632284641265869, "perplexity": 279.29948841077487, "lr": 0.0010568360655737704, "grad_norm": 0.531211, "tokens_per_sec": 106098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:37.371003+00:00", "epoch": 0, "step": 248, "train_loss": 5.7280731201171875, "perplexity": 307.3764199469377, "lr": 0.0010611147540983605, "grad_norm": 0.771281, "tokens_per_sec": 106351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:37.679429+00:00", "epoch": 0, "step": 249, "train_loss": 5.640615463256836, "perplexity": 281.63600175594706, "lr": 0.0010653934426229509, "grad_norm": 0.795983, "tokens_per_sec": 106243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:37.988270+00:00", "epoch": 0, "step": 250, "train_loss": 5.503636360168457, "perplexity": 245.58334001780787, "lr": 0.001069672131147541, "grad_norm": 0.490846, "tokens_per_sec": 106166} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:44:38.863478+00:00", "step": 250, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0000_step_0000250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:39.825961+00:00", "epoch": 0, "step": 251, "train_loss": 5.6522040367126465, "perplexity": 284.91874567166343, "lr": 0.0010739508196721311, "grad_norm": 0.559173, "tokens_per_sec": 17829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:40.133632+00:00", "epoch": 0, "step": 252, "train_loss": 5.727445602416992, "perplexity": 307.1835963092486, "lr": 0.0010782295081967212, "grad_norm": 0.455799, "tokens_per_sec": 106505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:40.442024+00:00", "epoch": 0, "step": 253, "train_loss": 5.5841851234436035, "perplexity": 266.18328781848794, "lr": 0.0010825081967213114, "grad_norm": 0.494723, "tokens_per_sec": 106254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:40.750754+00:00", "epoch": 0, "step": 254, "train_loss": 5.628895282745361, "perplexity": 278.3544447605325, "lr": 0.0010867868852459015, "grad_norm": 0.431926, "tokens_per_sec": 106138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:41.058019+00:00", "epoch": 0, "step": 255, "train_loss": 5.589069843292236, "perplexity": 267.4866994164523, "lr": 0.0010910655737704918, "grad_norm": 0.475887, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:41.365634+00:00", "epoch": 0, "step": 256, "train_loss": 5.595016002655029, "perplexity": 269.0819560835306, "lr": 0.001095344262295082, "grad_norm": 0.457593, "tokens_per_sec": 106523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:41.673047+00:00", "epoch": 0, "step": 257, "train_loss": 5.512491226196289, "perplexity": 247.76760400273005, "lr": 0.001099622950819672, "grad_norm": 0.386681, "tokens_per_sec": 106593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:41.981201+00:00", "epoch": 0, "step": 258, "train_loss": 5.616719722747803, "perplexity": 274.9858722602475, "lr": 0.0011039016393442622, "grad_norm": 0.356749, "tokens_per_sec": 106336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:42.289441+00:00", "epoch": 0, "step": 259, "train_loss": 5.648358345031738, "perplexity": 283.82514020423196, "lr": 0.0011081803278688524, "grad_norm": 0.523052, "tokens_per_sec": 106306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:42.598173+00:00", "epoch": 0, "step": 260, "train_loss": 5.51555061340332, "perplexity": 248.5267817579398, "lr": 0.0011124590163934427, "grad_norm": 0.560691, "tokens_per_sec": 106189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:42.905931+00:00", "epoch": 0, "step": 261, "train_loss": 5.474902629852295, "perplexity": 238.62723063072394, "lr": 0.0011167377049180328, "grad_norm": 0.670428, "tokens_per_sec": 106422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:43.213754+00:00", "epoch": 0, "step": 262, "train_loss": 5.566740989685059, "perplexity": 261.58021596682875, "lr": 0.001121016393442623, "grad_norm": 0.680852, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:43.521095+00:00", "epoch": 0, "step": 263, "train_loss": 5.630058288574219, "perplexity": 278.6783609239601, "lr": 0.001125295081967213, "grad_norm": 0.781565, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:43.830640+00:00", "epoch": 0, "step": 264, "train_loss": 5.5345306396484375, "perplexity": 253.2888759771873, "lr": 0.0011295737704918032, "grad_norm": 0.654621, "tokens_per_sec": 105863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:44.601934+00:00", "epoch": 0, "step": 265, "train_loss": 5.515465259552002, "perplexity": 248.50556994522884, "lr": 0.0011338524590163933, "grad_norm": 0.722755, "tokens_per_sec": 42484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:44.909578+00:00", "epoch": 0, "step": 266, "train_loss": 5.511321067810059, "perplexity": 247.4778462273785, "lr": 0.0011381311475409837, "grad_norm": 0.732011, "tokens_per_sec": 106513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:45.219788+00:00", "epoch": 0, "step": 267, "train_loss": 5.515494346618652, "perplexity": 248.51279834843083, "lr": 0.0011424098360655738, "grad_norm": 0.704813, "tokens_per_sec": 105632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:45.528746+00:00", "epoch": 0, "step": 268, "train_loss": 5.472255706787109, "perplexity": 237.99643790817782, "lr": 0.001146688524590164, "grad_norm": 0.617889, "tokens_per_sec": 106059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:45.836489+00:00", "epoch": 0, "step": 269, "train_loss": 5.655430316925049, "perplexity": 285.83945782203045, "lr": 0.001150967213114754, "grad_norm": 0.521728, "tokens_per_sec": 106478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:46.145166+00:00", "epoch": 0, "step": 270, "train_loss": 5.555419921875, "perplexity": 258.63554844504034, "lr": 0.0011552459016393442, "grad_norm": 0.48058, "tokens_per_sec": 106215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:46.453304+00:00", "epoch": 0, "step": 271, "train_loss": 5.525728225708008, "perplexity": 251.06910644223373, "lr": 0.0011595245901639343, "grad_norm": 0.413623, "tokens_per_sec": 106283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:46.759681+00:00", "epoch": 0, "step": 272, "train_loss": 5.5736870765686035, "perplexity": 263.4034998873839, "lr": 0.0011638032786885247, "grad_norm": 0.364563, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:47.068005+00:00", "epoch": 0, "step": 273, "train_loss": 5.589925765991211, "perplexity": 267.71574536298573, "lr": 0.0011680819672131148, "grad_norm": 0.359914, "tokens_per_sec": 106278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:47.376521+00:00", "epoch": 0, "step": 274, "train_loss": 5.544418811798096, "perplexity": 255.80586368193312, "lr": 0.001172360655737705, "grad_norm": 0.346043, "tokens_per_sec": 106212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:47.686107+00:00", "epoch": 0, "step": 275, "train_loss": 5.46327543258667, "perplexity": 235.86873262457118, "lr": 0.001176639344262295, "grad_norm": 0.341592, "tokens_per_sec": 105846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:47.994279+00:00", "epoch": 0, "step": 276, "train_loss": 5.516322135925293, "perplexity": 248.71859975381648, "lr": 0.0011809180327868852, "grad_norm": 0.349874, "tokens_per_sec": 106330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:48.302295+00:00", "epoch": 0, "step": 277, "train_loss": 5.4711127281188965, "perplexity": 237.72456845662748, "lr": 0.0011851967213114755, "grad_norm": 0.458138, "tokens_per_sec": 106384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:48.610459+00:00", "epoch": 0, "step": 278, "train_loss": 5.550715923309326, "perplexity": 257.42178420688055, "lr": 0.0011894754098360656, "grad_norm": 0.5259, "tokens_per_sec": 106333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:48.918177+00:00", "epoch": 0, "step": 279, "train_loss": 5.488203525543213, "perplexity": 241.822388660695, "lr": 0.0011937540983606558, "grad_norm": 0.599541, "tokens_per_sec": 106487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:49.227038+00:00", "epoch": 0, "step": 280, "train_loss": 5.441058158874512, "perplexity": 230.6861569607302, "lr": 0.001198032786885246, "grad_norm": 0.497423, "tokens_per_sec": 106148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:49.535664+00:00", "epoch": 0, "step": 281, "train_loss": 5.480491638183594, "perplexity": 239.96465416500695, "lr": 0.001202311475409836, "grad_norm": 0.526671, "tokens_per_sec": 106119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:49.842647+00:00", "epoch": 0, "step": 282, "train_loss": 5.590507984161377, "perplexity": 267.87165971805655, "lr": 0.0012065901639344262, "grad_norm": 0.597377, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:50.149582+00:00", "epoch": 0, "step": 283, "train_loss": 5.434731960296631, "perplexity": 229.23139692478077, "lr": 0.0012108688524590165, "grad_norm": 0.531308, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:50.457793+00:00", "epoch": 0, "step": 284, "train_loss": 5.397531509399414, "perplexity": 220.86055055617183, "lr": 0.0012151475409836066, "grad_norm": 0.509032, "tokens_per_sec": 106317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:50.766181+00:00", "epoch": 0, "step": 285, "train_loss": 5.371874809265137, "perplexity": 215.26607244733296, "lr": 0.0012194262295081968, "grad_norm": 0.39037, "tokens_per_sec": 106256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:51.074391+00:00", "epoch": 0, "step": 286, "train_loss": 5.519662857055664, "perplexity": 249.5508886840808, "lr": 0.0012237049180327869, "grad_norm": 0.449346, "tokens_per_sec": 106318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:51.383676+00:00", "epoch": 0, "step": 287, "train_loss": 5.517578125, "perplexity": 249.03118385780147, "lr": 0.001227983606557377, "grad_norm": 0.606679, "tokens_per_sec": 105947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:51.691958+00:00", "epoch": 0, "step": 288, "train_loss": 5.501553058624268, "perplexity": 245.07224843011946, "lr": 0.0012322622950819671, "grad_norm": 1.104137, "tokens_per_sec": 106291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:51.999334+00:00", "epoch": 0, "step": 289, "train_loss": 5.41707181930542, "perplexity": 225.21867504286698, "lr": 0.0012365409836065575, "grad_norm": 0.984459, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:52.307082+00:00", "epoch": 0, "step": 290, "train_loss": 5.469424247741699, "perplexity": 237.32351386929398, "lr": 0.0012408196721311476, "grad_norm": 1.030979, "tokens_per_sec": 106533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:52.614349+00:00", "epoch": 0, "step": 291, "train_loss": 5.5643157958984375, "perplexity": 260.94660188138806, "lr": 0.0012450983606557377, "grad_norm": 1.267794, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:52.922683+00:00", "epoch": 0, "step": 292, "train_loss": 5.6050238609313965, "perplexity": 271.788410481556, "lr": 0.0012493770491803279, "grad_norm": 0.970604, "tokens_per_sec": 106275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:53.231109+00:00", "epoch": 0, "step": 293, "train_loss": 5.504805564880371, "perplexity": 245.87064514264736, "lr": 0.001253655737704918, "grad_norm": 0.85592, "tokens_per_sec": 106241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:53.538043+00:00", "epoch": 0, "step": 294, "train_loss": 5.471539497375488, "perplexity": 237.82604364568667, "lr": 0.0012579344262295083, "grad_norm": 0.869738, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:53.845514+00:00", "epoch": 0, "step": 295, "train_loss": 5.5340094566345215, "perplexity": 253.15690051208682, "lr": 0.0012622131147540985, "grad_norm": 0.808272, "tokens_per_sec": 106573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:54.152775+00:00", "epoch": 0, "step": 296, "train_loss": 5.506080627441406, "perplexity": 246.18434554844694, "lr": 0.0012664918032786886, "grad_norm": 0.745731, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:54.460787+00:00", "epoch": 0, "step": 297, "train_loss": 5.477371692657471, "perplexity": 239.21714421733444, "lr": 0.0012707704918032787, "grad_norm": 0.590527, "tokens_per_sec": 106385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:54.769308+00:00", "epoch": 0, "step": 298, "train_loss": 5.6320061683654785, "perplexity": 279.2217219005949, "lr": 0.0012750491803278688, "grad_norm": 0.551489, "tokens_per_sec": 106211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:55.077899+00:00", "epoch": 0, "step": 299, "train_loss": 5.367435455322266, "perplexity": 214.31254824156198, "lr": 0.001279327868852459, "grad_norm": 0.4939, "tokens_per_sec": 106187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:44:55.387286+00:00", "epoch": 0, "step": 300, "train_loss": 5.4289116859436035, "perplexity": 227.90108245736948, "lr": 0.001283606557377049, "grad_norm": 0.486955, "tokens_per_sec": 105966} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:45:00.918502+00:00", "step": 300, "epoch": 0, "val_loss": 5.399879360198975, "val_ppl": 221.379707389296, "eval_train_loss": 5.4289116859436035, "eval_train_ppl": 227.90108245736948} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:45:01.877355+00:00", "step": 300, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_5p3999_epoch_0000_step_0000300.pt", "val_loss": 5.399879360198975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:02.863207+00:00", "epoch": 0, "step": 301, "train_loss": 5.36288595199585, "perplexity": 213.33974714926458, "lr": 0.0012878852459016392, "grad_norm": 0.447, "tokens_per_sec": 4383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:03.168763+00:00", "epoch": 0, "step": 302, "train_loss": 5.453580379486084, "perplexity": 233.59302212929813, "lr": 0.0012921639344262294, "grad_norm": 0.389651, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:03.475826+00:00", "epoch": 0, "step": 303, "train_loss": 5.447923183441162, "perplexity": 232.2752715096083, "lr": 0.0012964426229508195, "grad_norm": 0.360798, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:03.782722+00:00", "epoch": 0, "step": 304, "train_loss": 5.45273494720459, "perplexity": 233.39561850511438, "lr": 0.0013007213114754096, "grad_norm": 0.327715, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:04.089091+00:00", "epoch": 0, "step": 305, "train_loss": 5.519608020782471, "perplexity": 249.53720461856827, "lr": 0.001305, "grad_norm": 0.399931, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:04.395009+00:00", "epoch": 0, "step": 306, "train_loss": 5.378608703613281, "perplexity": 216.72054306555404, "lr": 0.001305, "grad_norm": 0.368451, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:04.701355+00:00", "epoch": 0, "step": 307, "train_loss": 5.333858966827393, "perplexity": 207.23615052805036, "lr": 0.001305, "grad_norm": 0.384372, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:05.007508+00:00", "epoch": 0, "step": 308, "train_loss": 5.342485427856445, "perplexity": 209.03159814973063, "lr": 0.001305, "grad_norm": 0.447782, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:05.314947+00:00", "epoch": 0, "step": 309, "train_loss": 5.426701068878174, "perplexity": 227.39783688182754, "lr": 0.001305, "grad_norm": 0.44308, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:05.621134+00:00", "epoch": 0, "step": 310, "train_loss": 5.370649337768555, "perplexity": 215.00243158654192, "lr": 0.001305, "grad_norm": 0.35684, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:05.927276+00:00", "epoch": 0, "step": 311, "train_loss": 5.2692670822143555, "perplexity": 194.27352373301235, "lr": 0.001305, "grad_norm": 0.469733, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:06.233620+00:00", "epoch": 0, "step": 312, "train_loss": 5.376887321472168, "perplexity": 216.34780509738954, "lr": 0.001305, "grad_norm": 0.488961, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:06.540489+00:00", "epoch": 0, "step": 313, "train_loss": 5.353236198425293, "perplexity": 211.29097214749774, "lr": 0.001305, "grad_norm": 0.444753, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:06.847630+00:00", "epoch": 0, "step": 314, "train_loss": 5.36268424987793, "perplexity": 213.29672040986617, "lr": 0.001305, "grad_norm": 0.351337, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:07.155516+00:00", "epoch": 0, "step": 315, "train_loss": 5.45034122467041, "perplexity": 232.8376022881648, "lr": 0.001305, "grad_norm": 0.29728, "tokens_per_sec": 106427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:07.463283+00:00", "epoch": 0, "step": 316, "train_loss": 5.257079124450684, "perplexity": 191.92009709604596, "lr": 0.001305, "grad_norm": 0.325519, "tokens_per_sec": 106479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:07.770391+00:00", "epoch": 0, "step": 317, "train_loss": 5.262921333312988, "perplexity": 193.04461602693775, "lr": 0.001305, "grad_norm": 0.285428, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:08.075756+00:00", "epoch": 0, "step": 318, "train_loss": 5.232543468475342, "perplexity": 187.26850984290553, "lr": 0.001305, "grad_norm": 0.343397, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:08.382494+00:00", "epoch": 0, "step": 319, "train_loss": 5.2984747886657715, "perplexity": 200.03148690184932, "lr": 0.001305, "grad_norm": 0.359829, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:08.689132+00:00", "epoch": 0, "step": 320, "train_loss": 5.235490798950195, "perplexity": 187.82126620664567, "lr": 0.001305, "grad_norm": 0.365578, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:08.997320+00:00", "epoch": 0, "step": 321, "train_loss": 5.275899887084961, "perplexity": 195.5663850159132, "lr": 0.001305, "grad_norm": 0.392417, "tokens_per_sec": 106266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:09.304064+00:00", "epoch": 0, "step": 322, "train_loss": 5.244365215301514, "perplexity": 189.49548820633805, "lr": 0.001305, "grad_norm": 0.473576, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:09.610643+00:00", "epoch": 0, "step": 323, "train_loss": 5.266831874847412, "perplexity": 193.80100299332207, "lr": 0.001305, "grad_norm": 0.45305, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:09.916869+00:00", "epoch": 0, "step": 324, "train_loss": 5.275656223297119, "perplexity": 195.51873837488193, "lr": 0.001305, "grad_norm": 0.319955, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:10.225117+00:00", "epoch": 0, "step": 325, "train_loss": 5.151988506317139, "perplexity": 172.77471256132742, "lr": 0.001305, "grad_norm": 0.366734, "tokens_per_sec": 106303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:10.532569+00:00", "epoch": 0, "step": 326, "train_loss": 5.198917865753174, "perplexity": 181.0761870721593, "lr": 0.001305, "grad_norm": 0.395884, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:10.839927+00:00", "epoch": 0, "step": 327, "train_loss": 5.378361225128174, "perplexity": 216.666916029907, "lr": 0.001305, "grad_norm": 0.579627, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:11.146354+00:00", "epoch": 0, "step": 328, "train_loss": 5.180755138397217, "perplexity": 177.81703677961514, "lr": 0.001305, "grad_norm": 0.540569, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:11.452366+00:00", "epoch": 0, "step": 329, "train_loss": 5.238651752471924, "perplexity": 188.41589980899377, "lr": 0.001305, "grad_norm": 0.53004, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:11.758566+00:00", "epoch": 0, "step": 330, "train_loss": 5.161234378814697, "perplexity": 174.3795732604068, "lr": 0.001305, "grad_norm": 0.484369, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:12.066473+00:00", "epoch": 0, "step": 331, "train_loss": 5.202319145202637, "perplexity": 181.69312638276887, "lr": 0.001305, "grad_norm": 0.391619, "tokens_per_sec": 106339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:12.372946+00:00", "epoch": 0, "step": 332, "train_loss": 5.285796165466309, "perplexity": 197.51137259963397, "lr": 0.001305, "grad_norm": 0.478447, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:12.679596+00:00", "epoch": 0, "step": 333, "train_loss": 5.302188873291016, "perplexity": 200.77580214119308, "lr": 0.001305, "grad_norm": 0.458204, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:12.985565+00:00", "epoch": 0, "step": 334, "train_loss": 5.240784168243408, "perplexity": 188.81810953211286, "lr": 0.001305, "grad_norm": 0.607166, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:13.290685+00:00", "epoch": 0, "step": 335, "train_loss": 5.281068325042725, "perplexity": 196.57977430746533, "lr": 0.001305, "grad_norm": 0.506959, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:13.596265+00:00", "epoch": 0, "step": 336, "train_loss": 5.271549701690674, "perplexity": 194.71748276401854, "lr": 0.001305, "grad_norm": 0.483698, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:13.903689+00:00", "epoch": 0, "step": 337, "train_loss": 5.18636417388916, "perplexity": 178.81722126286115, "lr": 0.001305, "grad_norm": 0.468664, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:14.210980+00:00", "epoch": 0, "step": 338, "train_loss": 5.267280101776123, "perplexity": 193.8878892926097, "lr": 0.001305, "grad_norm": 0.406689, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:14.518218+00:00", "epoch": 0, "step": 339, "train_loss": 5.192841053009033, "perplexity": 179.97915758816495, "lr": 0.001305, "grad_norm": 0.425668, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:14.824552+00:00", "epoch": 0, "step": 340, "train_loss": 5.286187171936035, "perplexity": 197.58861592450077, "lr": 0.001305, "grad_norm": 0.444259, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:15.130504+00:00", "epoch": 0, "step": 341, "train_loss": 5.124847412109375, "perplexity": 168.14848227201824, "lr": 0.001305, "grad_norm": 0.474711, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:15.437101+00:00", "epoch": 0, "step": 342, "train_loss": 5.2699294090271, "perplexity": 194.4022389178407, "lr": 0.001305, "grad_norm": 0.403133, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:15.743860+00:00", "epoch": 0, "step": 343, "train_loss": 5.161493301391602, "perplexity": 174.4247299146612, "lr": 0.001305, "grad_norm": 0.348929, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:16.050990+00:00", "epoch": 0, "step": 344, "train_loss": 5.063320636749268, "perplexity": 158.11468699174569, "lr": 0.001305, "grad_norm": 0.344896, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:16.357368+00:00", "epoch": 0, "step": 345, "train_loss": 5.192440986633301, "perplexity": 179.90716838007225, "lr": 0.001305, "grad_norm": 0.31456, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:16.664000+00:00", "epoch": 0, "step": 346, "train_loss": 5.313162326812744, "perplexity": 202.9911387909114, "lr": 0.001305, "grad_norm": 0.309023, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:16.974396+00:00", "epoch": 0, "step": 347, "train_loss": 5.150176525115967, "perplexity": 172.46193149241577, "lr": 0.001305, "grad_norm": 0.313776, "tokens_per_sec": 105570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:17.283080+00:00", "epoch": 0, "step": 348, "train_loss": 5.219746112823486, "perplexity": 184.88723761265288, "lr": 0.001305, "grad_norm": 0.333393, "tokens_per_sec": 106152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:17.592993+00:00", "epoch": 0, "step": 349, "train_loss": 5.142439842224121, "perplexity": 171.13279639567398, "lr": 0.001305, "grad_norm": 0.313541, "tokens_per_sec": 105734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:17.900516+00:00", "epoch": 0, "step": 350, "train_loss": 4.992388725280762, "perplexity": 147.28783378939409, "lr": 0.001305, "grad_norm": 0.291407, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:18.207419+00:00", "epoch": 0, "step": 351, "train_loss": 5.223476409912109, "perplexity": 185.57820990155525, "lr": 0.001305, "grad_norm": 0.301832, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:18.514121+00:00", "epoch": 0, "step": 352, "train_loss": 5.150208473205566, "perplexity": 172.4674414096708, "lr": 0.001305, "grad_norm": 0.265163, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:18.820910+00:00", "epoch": 0, "step": 353, "train_loss": 5.037808895111084, "perplexity": 154.13192555078984, "lr": 0.001305, "grad_norm": 0.287188, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:19.127342+00:00", "epoch": 0, "step": 354, "train_loss": 5.199563026428223, "perplexity": 181.19304800023025, "lr": 0.001305, "grad_norm": 0.321272, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:19.435048+00:00", "epoch": 0, "step": 355, "train_loss": 4.993398189544678, "perplexity": 147.4365906638563, "lr": 0.001305, "grad_norm": 0.324413, "tokens_per_sec": 106491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:19.742561+00:00", "epoch": 0, "step": 356, "train_loss": 5.197236061096191, "perplexity": 180.77190823797315, "lr": 0.001305, "grad_norm": 0.327399, "tokens_per_sec": 106558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:20.049025+00:00", "epoch": 0, "step": 357, "train_loss": 5.014963150024414, "perplexity": 150.65058519324342, "lr": 0.001305, "grad_norm": 0.446962, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:20.355175+00:00", "epoch": 0, "step": 358, "train_loss": 5.163708686828613, "perplexity": 174.81157626984168, "lr": 0.001305, "grad_norm": 0.63247, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:20.661103+00:00", "epoch": 0, "step": 359, "train_loss": 5.030325889587402, "perplexity": 152.98286009395517, "lr": 0.001305, "grad_norm": 0.765551, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:20.969232+00:00", "epoch": 0, "step": 360, "train_loss": 5.242908954620361, "perplexity": 189.2197342111903, "lr": 0.001305, "grad_norm": 0.740414, "tokens_per_sec": 106405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:21.276189+00:00", "epoch": 0, "step": 361, "train_loss": 5.101500511169434, "perplexity": 164.26820874666, "lr": 0.001305, "grad_norm": 0.666641, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:21.582388+00:00", "epoch": 0, "step": 362, "train_loss": 5.150244235992432, "perplexity": 172.4736094363114, "lr": 0.001305, "grad_norm": 0.620108, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:21.889013+00:00", "epoch": 0, "step": 363, "train_loss": 5.184704780578613, "perplexity": 178.52073922027847, "lr": 0.001305, "grad_norm": 0.706583, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:22.195495+00:00", "epoch": 0, "step": 364, "train_loss": 5.10047721862793, "perplexity": 164.10020028940448, "lr": 0.001305, "grad_norm": 0.802473, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:22.502645+00:00", "epoch": 0, "step": 365, "train_loss": 5.107552528381348, "perplexity": 165.26537716999752, "lr": 0.001305, "grad_norm": 0.794059, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:22.809267+00:00", "epoch": 0, "step": 366, "train_loss": 5.221245765686035, "perplexity": 185.16471229363484, "lr": 0.001305, "grad_norm": 0.656189, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:23.116591+00:00", "epoch": 0, "step": 367, "train_loss": 5.223016738891602, "perplexity": 185.49292457952302, "lr": 0.001305, "grad_norm": 0.635198, "tokens_per_sec": 106623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:23.422468+00:00", "epoch": 0, "step": 368, "train_loss": 5.181191921234131, "perplexity": 177.89472117376138, "lr": 0.001305, "grad_norm": 0.614438, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:23.729168+00:00", "epoch": 0, "step": 369, "train_loss": 5.104794502258301, "perplexity": 164.81019892780273, "lr": 0.001305, "grad_norm": 0.629404, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:24.035655+00:00", "epoch": 0, "step": 370, "train_loss": 5.098393440246582, "perplexity": 163.75860786472475, "lr": 0.001305, "grad_norm": 0.449056, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:24.343404+00:00", "epoch": 0, "step": 371, "train_loss": 5.205357074737549, "perplexity": 182.24593657173997, "lr": 0.001305, "grad_norm": 0.407347, "tokens_per_sec": 106423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:24.650108+00:00", "epoch": 0, "step": 372, "train_loss": 5.142031669616699, "perplexity": 171.06295892979364, "lr": 0.001305, "grad_norm": 0.371803, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:24.958300+00:00", "epoch": 0, "step": 373, "train_loss": 5.002928733825684, "perplexity": 148.8484588692444, "lr": 0.001305, "grad_norm": 0.332746, "tokens_per_sec": 106324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:25.265060+00:00", "epoch": 0, "step": 374, "train_loss": 5.226521015167236, "perplexity": 186.1440832880825, "lr": 0.001305, "grad_norm": 0.328709, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:25.571485+00:00", "epoch": 0, "step": 375, "train_loss": 5.046693801879883, "perplexity": 155.50747510518605, "lr": 0.001305, "grad_norm": 0.33842, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:25.877998+00:00", "epoch": 0, "step": 376, "train_loss": 5.098759174346924, "perplexity": 163.81851092547367, "lr": 0.001305, "grad_norm": 0.364272, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:26.184651+00:00", "epoch": 0, "step": 377, "train_loss": 5.125592231750488, "perplexity": 168.27376921653436, "lr": 0.001305, "grad_norm": 0.372927, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:26.493362+00:00", "epoch": 0, "step": 378, "train_loss": 5.074852466583252, "perplexity": 159.94859247621957, "lr": 0.001305, "grad_norm": 0.360843, "tokens_per_sec": 106150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:26.800902+00:00", "epoch": 0, "step": 379, "train_loss": 5.030460357666016, "perplexity": 153.00343278836712, "lr": 0.001305, "grad_norm": 0.36993, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:27.107312+00:00", "epoch": 0, "step": 380, "train_loss": 5.049944877624512, "perplexity": 156.0138643945539, "lr": 0.001305, "grad_norm": 0.417045, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:27.413946+00:00", "epoch": 0, "step": 381, "train_loss": 5.115281581878662, "perplexity": 166.54767118743268, "lr": 0.001305, "grad_norm": 0.405244, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:27.720168+00:00", "epoch": 0, "step": 382, "train_loss": 5.0607123374938965, "perplexity": 157.70281394898262, "lr": 0.001305, "grad_norm": 0.430502, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:28.027956+00:00", "epoch": 0, "step": 383, "train_loss": 5.038773059844971, "perplexity": 154.28060578237202, "lr": 0.001305, "grad_norm": 0.391723, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:28.335358+00:00", "epoch": 0, "step": 384, "train_loss": 4.952756404876709, "perplexity": 141.56463608089965, "lr": 0.001305, "grad_norm": 0.388489, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:28.642785+00:00", "epoch": 0, "step": 385, "train_loss": 4.979583740234375, "perplexity": 145.41383912318236, "lr": 0.001305, "grad_norm": 0.350454, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:28.949455+00:00", "epoch": 0, "step": 386, "train_loss": 5.014503002166748, "perplexity": 150.58127959584456, "lr": 0.001305, "grad_norm": 0.283199, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:29.256464+00:00", "epoch": 0, "step": 387, "train_loss": 5.071586608886719, "perplexity": 159.42707519857333, "lr": 0.001305, "grad_norm": 0.308252, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:29.563343+00:00", "epoch": 0, "step": 388, "train_loss": 5.008343696594238, "perplexity": 149.65665393044952, "lr": 0.001305, "grad_norm": 0.303429, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:29.871005+00:00", "epoch": 0, "step": 389, "train_loss": 4.993686676025391, "perplexity": 147.4791302627798, "lr": 0.001305, "grad_norm": 0.254506, "tokens_per_sec": 106507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:30.178735+00:00", "epoch": 0, "step": 390, "train_loss": 4.90537691116333, "perplexity": 135.01378862198436, "lr": 0.001305, "grad_norm": 0.277676, "tokens_per_sec": 106542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:30.484871+00:00", "epoch": 0, "step": 391, "train_loss": 5.029779434204102, "perplexity": 152.89928462371662, "lr": 0.001305, "grad_norm": 0.257908, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:30.791296+00:00", "epoch": 0, "step": 392, "train_loss": 5.094427108764648, "perplexity": 163.1103733492202, "lr": 0.001305, "grad_norm": 0.280696, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:31.098381+00:00", "epoch": 0, "step": 393, "train_loss": 4.931886196136475, "perplexity": 138.64076955522154, "lr": 0.001305, "grad_norm": 0.263569, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:31.404986+00:00", "epoch": 0, "step": 394, "train_loss": 4.946139812469482, "perplexity": 140.63105256145465, "lr": 0.001305, "grad_norm": 0.264647, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:31.713018+00:00", "epoch": 0, "step": 395, "train_loss": 4.952951431274414, "perplexity": 141.59224761431835, "lr": 0.001305, "grad_norm": 0.260472, "tokens_per_sec": 106379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:32.020415+00:00", "epoch": 0, "step": 396, "train_loss": 5.021978378295898, "perplexity": 151.71114913719302, "lr": 0.001305, "grad_norm": 0.29915, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:32.327341+00:00", "epoch": 0, "step": 397, "train_loss": 4.908824920654297, "perplexity": 135.48012094369025, "lr": 0.001305, "grad_norm": 0.358016, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:32.633588+00:00", "epoch": 0, "step": 398, "train_loss": 5.005659580230713, "perplexity": 149.25549667394674, "lr": 0.001305, "grad_norm": 0.469135, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:32.939688+00:00", "epoch": 0, "step": 399, "train_loss": 4.991259574890137, "perplexity": 147.1216175335718, "lr": 0.001305, "grad_norm": 0.558569, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:33.246916+00:00", "epoch": 0, "step": 400, "train_loss": 5.014268398284912, "perplexity": 150.54595678671564, "lr": 0.001305, "grad_norm": 0.46197, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:45:38.788646+00:00", "step": 400, "epoch": 0, "val_loss": 4.954746580123901, "val_ppl": 141.84665505601237, "eval_train_loss": 5.014268398284912, "eval_train_ppl": 150.54595678671564} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:45:39.609246+00:00", "step": 400, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_4p9547_epoch_0000_step_0000400.pt", "val_loss": 4.954746580123901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:40.645421+00:00", "epoch": 0, "step": 401, "train_loss": 5.025571346282959, "perplexity": 152.2572228645759, "lr": 0.001305, "grad_norm": 0.597492, "tokens_per_sec": 4429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:40.952541+00:00", "epoch": 0, "step": 402, "train_loss": 5.019251346588135, "perplexity": 151.29799162593548, "lr": 0.001305, "grad_norm": 0.683651, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:41.259076+00:00", "epoch": 0, "step": 403, "train_loss": 5.05668830871582, "perplexity": 157.0694884019915, "lr": 0.001305, "grad_norm": 0.66167, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:41.564762+00:00", "epoch": 0, "step": 404, "train_loss": 4.917840480804443, "perplexity": 136.70707264104684, "lr": 0.001305, "grad_norm": 0.632408, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:41.871471+00:00", "epoch": 0, "step": 405, "train_loss": 5.023011684417725, "perplexity": 151.8679942168807, "lr": 0.001305, "grad_norm": 0.505644, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:42.178464+00:00", "epoch": 0, "step": 406, "train_loss": 5.091095924377441, "perplexity": 162.56792661678875, "lr": 0.001305, "grad_norm": 0.506181, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:42.485501+00:00", "epoch": 0, "step": 407, "train_loss": 4.973781108856201, "perplexity": 144.57249956980598, "lr": 0.001305, "grad_norm": 0.425197, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:42.792498+00:00", "epoch": 0, "step": 408, "train_loss": 4.921747207641602, "perplexity": 137.24219443729572, "lr": 0.001305, "grad_norm": 0.395876, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:43.098794+00:00", "epoch": 0, "step": 409, "train_loss": 4.985163688659668, "perplexity": 146.2275088521864, "lr": 0.001305, "grad_norm": 0.365149, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:43.404850+00:00", "epoch": 0, "step": 410, "train_loss": 4.9213409423828125, "perplexity": 137.18644902613457, "lr": 0.001305, "grad_norm": 0.311051, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:43.710835+00:00", "epoch": 0, "step": 411, "train_loss": 4.997461795806885, "perplexity": 148.03693387004563, "lr": 0.001305, "grad_norm": 0.317614, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:44.018769+00:00", "epoch": 0, "step": 412, "train_loss": 5.013829231262207, "perplexity": 150.47985648269258, "lr": 0.001305, "grad_norm": 0.294386, "tokens_per_sec": 106414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:44.325130+00:00", "epoch": 0, "step": 413, "train_loss": 4.871656894683838, "perplexity": 130.53702391470242, "lr": 0.001305, "grad_norm": 0.26348, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:44.631766+00:00", "epoch": 0, "step": 414, "train_loss": 4.9561591148376465, "perplexity": 142.04715995703594, "lr": 0.001305, "grad_norm": 0.296336, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:44.937328+00:00", "epoch": 0, "step": 415, "train_loss": 4.945432186126709, "perplexity": 140.53157352519457, "lr": 0.001305, "grad_norm": 0.269886, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:45.243599+00:00", "epoch": 0, "step": 416, "train_loss": 5.054083347320557, "perplexity": 156.6608609088257, "lr": 0.001305, "grad_norm": 0.239754, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:45.551539+00:00", "epoch": 0, "step": 417, "train_loss": 4.829905986785889, "perplexity": 125.19918972323231, "lr": 0.001305, "grad_norm": 0.257893, "tokens_per_sec": 106411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:45.859198+00:00", "epoch": 0, "step": 418, "train_loss": 4.989275932312012, "perplexity": 146.83007008743274, "lr": 0.001305, "grad_norm": 0.275699, "tokens_per_sec": 106509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:46.165826+00:00", "epoch": 0, "step": 419, "train_loss": 4.913086891174316, "perplexity": 136.05876543218315, "lr": 0.001305, "grad_norm": 0.279709, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:46.472440+00:00", "epoch": 0, "step": 420, "train_loss": 4.926255226135254, "perplexity": 137.86228141959154, "lr": 0.001305, "grad_norm": 0.288086, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:46.779008+00:00", "epoch": 0, "step": 421, "train_loss": 4.872122764587402, "perplexity": 130.59785135318145, "lr": 0.001305, "grad_norm": 0.323542, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:47.085828+00:00", "epoch": 0, "step": 422, "train_loss": 5.013446807861328, "perplexity": 150.4223204664734, "lr": 0.001305, "grad_norm": 0.413891, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:47.391743+00:00", "epoch": 0, "step": 423, "train_loss": 4.921759605407715, "perplexity": 137.24389594447067, "lr": 0.001305, "grad_norm": 0.608168, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:47.697962+00:00", "epoch": 0, "step": 424, "train_loss": 4.877079010009766, "perplexity": 131.24673303627668, "lr": 0.001305, "grad_norm": 0.586252, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:48.004716+00:00", "epoch": 0, "step": 425, "train_loss": 4.788869380950928, "perplexity": 120.16543050579132, "lr": 0.001305, "grad_norm": 0.678957, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:48.314055+00:00", "epoch": 0, "step": 426, "train_loss": 4.952899932861328, "perplexity": 141.5849560260152, "lr": 0.001305, "grad_norm": 0.594466, "tokens_per_sec": 105939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:48.624005+00:00", "epoch": 0, "step": 427, "train_loss": 5.035553455352783, "perplexity": 153.78468201851595, "lr": 0.001305, "grad_norm": 0.462339, "tokens_per_sec": 105716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:48.930884+00:00", "epoch": 0, "step": 428, "train_loss": 4.918889045715332, "perplexity": 136.85049406068177, "lr": 0.001305, "grad_norm": 0.422398, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:49.237096+00:00", "epoch": 0, "step": 429, "train_loss": 4.917476654052734, "perplexity": 136.65734399772614, "lr": 0.001305, "grad_norm": 0.423468, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:49.544626+00:00", "epoch": 0, "step": 430, "train_loss": 4.884323596954346, "perplexity": 132.20101391664778, "lr": 0.001305, "grad_norm": 0.402897, "tokens_per_sec": 106632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:49.850434+00:00", "epoch": 0, "step": 431, "train_loss": 4.941369533538818, "perplexity": 139.96180074244896, "lr": 0.001305, "grad_norm": 0.332204, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:50.158163+00:00", "epoch": 0, "step": 432, "train_loss": 4.852705955505371, "perplexity": 128.08651774934242, "lr": 0.001305, "grad_norm": 0.366765, "tokens_per_sec": 106482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:50.466040+00:00", "epoch": 0, "step": 433, "train_loss": 4.969701290130615, "perplexity": 143.9838715431511, "lr": 0.001305, "grad_norm": 0.309003, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:50.773098+00:00", "epoch": 0, "step": 434, "train_loss": 4.968593597412109, "perplexity": 143.82446995734702, "lr": 0.001305, "grad_norm": 0.272222, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:51.080211+00:00", "epoch": 0, "step": 435, "train_loss": 4.957848072052002, "perplexity": 142.28727424692647, "lr": 0.001305, "grad_norm": 0.316035, "tokens_per_sec": 106697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:51.386981+00:00", "epoch": 0, "step": 436, "train_loss": 5.019430160522461, "perplexity": 151.32504823405094, "lr": 0.001305, "grad_norm": 0.31127, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:51.693441+00:00", "epoch": 0, "step": 437, "train_loss": 4.809511661529541, "perplexity": 122.6716975786862, "lr": 0.001305, "grad_norm": 0.30253, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:52.000785+00:00", "epoch": 0, "step": 438, "train_loss": 4.832970142364502, "perplexity": 125.58340787032981, "lr": 0.001305, "grad_norm": 0.295356, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:52.307912+00:00", "epoch": 0, "step": 439, "train_loss": 4.9320831298828125, "perplexity": 138.66807528998729, "lr": 0.001305, "grad_norm": 0.289409, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:52.615005+00:00", "epoch": 0, "step": 440, "train_loss": 4.780111789703369, "perplexity": 119.11766542902289, "lr": 0.001305, "grad_norm": 0.306254, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:52.922259+00:00", "epoch": 0, "step": 441, "train_loss": 4.7993059158325195, "perplexity": 121.4261083239745, "lr": 0.001305, "grad_norm": 0.312255, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:53.227943+00:00", "epoch": 0, "step": 442, "train_loss": 4.823108196258545, "perplexity": 124.3509980382406, "lr": 0.001305, "grad_norm": 0.299842, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:53.534705+00:00", "epoch": 0, "step": 443, "train_loss": 4.825822353363037, "perplexity": 124.6889646227466, "lr": 0.001305, "grad_norm": 0.331237, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:53.842452+00:00", "epoch": 0, "step": 444, "train_loss": 4.892681121826172, "perplexity": 133.31051707057793, "lr": 0.001305, "grad_norm": 0.378698, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:54.150655+00:00", "epoch": 0, "step": 445, "train_loss": 4.822254180908203, "perplexity": 124.2448457113957, "lr": 0.001305, "grad_norm": 0.411253, "tokens_per_sec": 106322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:54.457544+00:00", "epoch": 0, "step": 446, "train_loss": 4.886268138885498, "perplexity": 132.45833443573648, "lr": 0.001305, "grad_norm": 0.329217, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:54.764119+00:00", "epoch": 0, "step": 447, "train_loss": 4.7670979499816895, "perplexity": 117.57753049408493, "lr": 0.001305, "grad_norm": 0.321746, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:55.070641+00:00", "epoch": 0, "step": 448, "train_loss": 4.869021415710449, "perplexity": 130.19344927366285, "lr": 0.001305, "grad_norm": 0.337203, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:55.378041+00:00", "epoch": 0, "step": 449, "train_loss": 4.7744011878967285, "perplexity": 118.43937045319888, "lr": 0.001305, "grad_norm": 0.308124, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:55.684818+00:00", "epoch": 0, "step": 450, "train_loss": 4.749296188354492, "perplexity": 115.5029635823706, "lr": 0.001305, "grad_norm": 0.341489, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:55.993248+00:00", "epoch": 0, "step": 451, "train_loss": 4.73323917388916, "perplexity": 113.66314138078477, "lr": 0.001305, "grad_norm": 0.304202, "tokens_per_sec": 106183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:56.299503+00:00", "epoch": 0, "step": 452, "train_loss": 4.739256858825684, "perplexity": 114.34919250396189, "lr": 0.001305, "grad_norm": 0.321471, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:56.606124+00:00", "epoch": 0, "step": 453, "train_loss": 4.830668926239014, "perplexity": 125.29474557159972, "lr": 0.001305, "grad_norm": 0.348692, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:56.913944+00:00", "epoch": 0, "step": 454, "train_loss": 4.930834770202637, "perplexity": 138.49507566121522, "lr": 0.001305, "grad_norm": 0.413673, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:57.222198+00:00", "epoch": 0, "step": 455, "train_loss": 4.866450786590576, "perplexity": 129.85920000126336, "lr": 0.001305, "grad_norm": 0.443211, "tokens_per_sec": 106302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:57.530133+00:00", "epoch": 0, "step": 456, "train_loss": 4.860389709472656, "perplexity": 129.07449386009108, "lr": 0.001305, "grad_norm": 0.456062, "tokens_per_sec": 106412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:57.838145+00:00", "epoch": 0, "step": 457, "train_loss": 4.847103595733643, "perplexity": 127.37093733608857, "lr": 0.001305, "grad_norm": 0.398342, "tokens_per_sec": 106386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:58.144026+00:00", "epoch": 0, "step": 458, "train_loss": 4.752488136291504, "perplexity": 115.87223205804541, "lr": 0.001305, "grad_norm": 0.350159, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:58.451000+00:00", "epoch": 0, "step": 459, "train_loss": 4.724153995513916, "perplexity": 112.63516819729067, "lr": 0.001305, "grad_norm": 0.362878, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:58.757753+00:00", "epoch": 0, "step": 460, "train_loss": 4.793679714202881, "perplexity": 120.74485877612965, "lr": 0.001305, "grad_norm": 0.379864, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:59.065911+00:00", "epoch": 0, "step": 461, "train_loss": 4.857424736022949, "perplexity": 128.6923582037475, "lr": 0.001305, "grad_norm": 0.404901, "tokens_per_sec": 106279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:59.373025+00:00", "epoch": 0, "step": 462, "train_loss": 4.85678768157959, "perplexity": 128.61040027373093, "lr": 0.001305, "grad_norm": 0.372685, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:59.679490+00:00", "epoch": 0, "step": 463, "train_loss": 4.8024797439575195, "perplexity": 121.81210614303068, "lr": 0.001305, "grad_norm": 0.366175, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:45:59.986825+00:00", "epoch": 0, "step": 464, "train_loss": 4.786595344543457, "perplexity": 119.89248040875502, "lr": 0.001305, "grad_norm": 0.338534, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:00.294647+00:00", "epoch": 0, "step": 465, "train_loss": 4.8612895011901855, "perplexity": 129.19068628726313, "lr": 0.001305, "grad_norm": 0.365216, "tokens_per_sec": 106455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:00.602404+00:00", "epoch": 0, "step": 466, "train_loss": 4.679369926452637, "perplexity": 107.7021908869484, "lr": 0.001305, "grad_norm": 0.446457, "tokens_per_sec": 106469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:00.910471+00:00", "epoch": 0, "step": 467, "train_loss": 4.907834529876709, "perplexity": 135.346009103888, "lr": 0.001305, "grad_norm": 0.508747, "tokens_per_sec": 106366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:01.217600+00:00", "epoch": 0, "step": 468, "train_loss": 4.794920921325684, "perplexity": 120.89482120284787, "lr": 0.001305, "grad_norm": 0.493624, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:01.524812+00:00", "epoch": 0, "step": 469, "train_loss": 4.833889961242676, "perplexity": 125.6989750019409, "lr": 0.001305, "grad_norm": 0.504229, "tokens_per_sec": 106670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:01.832402+00:00", "epoch": 0, "step": 470, "train_loss": 4.74026346206665, "perplexity": 114.46435472334595, "lr": 0.001305, "grad_norm": 0.660958, "tokens_per_sec": 106594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:02.138901+00:00", "epoch": 0, "step": 471, "train_loss": 4.665051460266113, "perplexity": 106.17104867181683, "lr": 0.001305, "grad_norm": 0.617933, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:02.447008+00:00", "epoch": 0, "step": 472, "train_loss": 4.846217155456543, "perplexity": 127.25808063485471, "lr": 0.001305, "grad_norm": 0.533759, "tokens_per_sec": 106352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:02.755139+00:00", "epoch": 0, "step": 473, "train_loss": 4.715400218963623, "perplexity": 111.65348807612263, "lr": 0.001305, "grad_norm": 0.514801, "tokens_per_sec": 106345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:03.061944+00:00", "epoch": 0, "step": 474, "train_loss": 4.734004497528076, "perplexity": 113.7501637656535, "lr": 0.001305, "grad_norm": 0.549608, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:03.369947+00:00", "epoch": 0, "step": 475, "train_loss": 4.776731014251709, "perplexity": 118.71563531956545, "lr": 0.001305, "grad_norm": 0.51558, "tokens_per_sec": 106390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:03.677229+00:00", "epoch": 0, "step": 476, "train_loss": 4.717794418334961, "perplexity": 111.92112905216304, "lr": 0.001305, "grad_norm": 0.405383, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:03.985157+00:00", "epoch": 0, "step": 477, "train_loss": 4.622624397277832, "perplexity": 101.76074261523908, "lr": 0.001305, "grad_norm": 0.366343, "tokens_per_sec": 106413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:04.293326+00:00", "epoch": 0, "step": 478, "train_loss": 4.815329551696777, "perplexity": 123.38746816020164, "lr": 0.001305, "grad_norm": 0.361542, "tokens_per_sec": 106335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:04.602497+00:00", "epoch": 0, "step": 479, "train_loss": 4.692245006561279, "perplexity": 109.09783042867035, "lr": 0.001305, "grad_norm": 0.325388, "tokens_per_sec": 105982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:04.910403+00:00", "epoch": 0, "step": 480, "train_loss": 4.741796970367432, "perplexity": 114.64002142020115, "lr": 0.001305, "grad_norm": 0.310007, "tokens_per_sec": 106498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:05.217346+00:00", "epoch": 0, "step": 481, "train_loss": 4.717226505279541, "perplexity": 111.8575856270752, "lr": 0.001305, "grad_norm": 0.322832, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:05.524383+00:00", "epoch": 0, "step": 482, "train_loss": 4.753361225128174, "perplexity": 115.97344298697212, "lr": 0.001305, "grad_norm": 0.353631, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:05.831067+00:00", "epoch": 0, "step": 483, "train_loss": 4.683182239532471, "perplexity": 108.11356901085662, "lr": 0.001305, "grad_norm": 0.368179, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:06.138829+00:00", "epoch": 0, "step": 484, "train_loss": 4.662221431732178, "perplexity": 105.87100633903408, "lr": 0.001305, "grad_norm": 0.375974, "tokens_per_sec": 106472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:06.447322+00:00", "epoch": 0, "step": 485, "train_loss": 4.765298843383789, "perplexity": 117.36618615571132, "lr": 0.001305, "grad_norm": 0.402178, "tokens_per_sec": 106220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:06.755360+00:00", "epoch": 0, "step": 486, "train_loss": 4.624719142913818, "perplexity": 101.97412890371784, "lr": 0.001305, "grad_norm": 0.379335, "tokens_per_sec": 106377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:07.062583+00:00", "epoch": 0, "step": 487, "train_loss": 4.657357692718506, "perplexity": 105.35732760735986, "lr": 0.001305, "grad_norm": 0.356104, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:07.369730+00:00", "epoch": 0, "step": 488, "train_loss": 4.594883441925049, "perplexity": 98.9765983535351, "lr": 0.001305, "grad_norm": 0.356004, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:07.677242+00:00", "epoch": 0, "step": 489, "train_loss": 4.692986011505127, "perplexity": 109.17870241995324, "lr": 0.001305, "grad_norm": 0.443096, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:07.985381+00:00", "epoch": 0, "step": 490, "train_loss": 4.797532558441162, "perplexity": 121.21096725465837, "lr": 0.001305, "grad_norm": 0.463477, "tokens_per_sec": 106399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:08.293005+00:00", "epoch": 0, "step": 491, "train_loss": 4.749615669250488, "perplexity": 115.53987046787461, "lr": 0.001305, "grad_norm": 0.492795, "tokens_per_sec": 106465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:08.600423+00:00", "epoch": 0, "step": 492, "train_loss": 4.675340175628662, "perplexity": 107.26905120311896, "lr": 0.001305, "grad_norm": 0.564373, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:08.908048+00:00", "epoch": 0, "step": 493, "train_loss": 4.643489360809326, "perplexity": 103.90628226488006, "lr": 0.001305, "grad_norm": 0.499915, "tokens_per_sec": 106519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:09.214547+00:00", "epoch": 0, "step": 494, "train_loss": 4.7002034187316895, "perplexity": 109.96954004141274, "lr": 0.001305, "grad_norm": 0.477185, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:09.522073+00:00", "epoch": 0, "step": 495, "train_loss": 4.695277214050293, "perplexity": 109.42913973256162, "lr": 0.001305, "grad_norm": 0.366093, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:09.829530+00:00", "epoch": 0, "step": 496, "train_loss": 4.715601444244385, "perplexity": 111.67595784127535, "lr": 0.001305, "grad_norm": 0.350677, "tokens_per_sec": 106577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:10.137919+00:00", "epoch": 0, "step": 497, "train_loss": 4.6563005447387695, "perplexity": 105.24600817225449, "lr": 0.001305, "grad_norm": 0.358888, "tokens_per_sec": 106256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:10.444692+00:00", "epoch": 0, "step": 498, "train_loss": 4.677318096160889, "perplexity": 107.4814308277374, "lr": 0.001305, "grad_norm": 0.371638, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:10.752653+00:00", "epoch": 0, "step": 499, "train_loss": 4.717730522155762, "perplexity": 111.9139779481115, "lr": 0.001305, "grad_norm": 0.368396, "tokens_per_sec": 106403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:11.061057+00:00", "epoch": 0, "step": 500, "train_loss": 4.638061046600342, "perplexity": 103.34377443235405, "lr": 0.001305, "grad_norm": 0.359388, "tokens_per_sec": 106315} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:46:16.769984+00:00", "step": 500, "epoch": 0, "val_loss": 4.6274735689163204, "val_ppl": 102.25539628321656, "eval_train_loss": 4.638061046600342, "eval_train_ppl": 103.34377443235405} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:46:17.623171+00:00", "step": 500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_4p6275_epoch_0000_step_0000500.pt", "val_loss": 4.6274735689163204} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:46:19.430808+00:00", "step": 500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0000_step_0000500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:20.405924+00:00", "epoch": 0, "step": 501, "train_loss": 4.809143543243408, "perplexity": 122.6265481942834, "lr": 0.001305, "grad_norm": 0.366817, "tokens_per_sec": 3506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:20.713687+00:00", "epoch": 0, "step": 502, "train_loss": 4.645999431610107, "perplexity": 104.16742199250523, "lr": 0.001305, "grad_norm": 0.390263, "tokens_per_sec": 106472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:21.021642+00:00", "epoch": 0, "step": 503, "train_loss": 4.785282135009766, "perplexity": 119.73513979367362, "lr": 0.001305, "grad_norm": 0.315064, "tokens_per_sec": 106405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:21.328228+00:00", "epoch": 0, "step": 504, "train_loss": 4.651116371154785, "perplexity": 104.70180643382079, "lr": 0.001305, "grad_norm": 0.289601, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:21.634528+00:00", "epoch": 0, "step": 505, "train_loss": 4.620810031890869, "perplexity": 101.57627883903506, "lr": 0.001305, "grad_norm": 0.30088, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:21.940781+00:00", "epoch": 0, "step": 506, "train_loss": 4.637204170227051, "perplexity": 103.25525952229972, "lr": 0.001305, "grad_norm": 0.3152, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:22.248074+00:00", "epoch": 0, "step": 507, "train_loss": 4.810358047485352, "perplexity": 122.77556913220343, "lr": 0.001305, "grad_norm": 0.325317, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:22.644335+00:00", "epoch": 0, "step": 508, "train_loss": 4.56306791305542, "perplexity": 95.87717191625748, "lr": 0.001305, "grad_norm": 0.305135, "tokens_per_sec": 82693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:22.951140+00:00", "epoch": 0, "step": 509, "train_loss": 4.608026027679443, "perplexity": 100.28599234919689, "lr": 0.001305, "grad_norm": 0.346369, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:23.257194+00:00", "epoch": 0, "step": 510, "train_loss": 4.612854957580566, "perplexity": 100.77143752335081, "lr": 0.001305, "grad_norm": 0.338676, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:23.565018+00:00", "epoch": 0, "step": 511, "train_loss": 4.551482677459717, "perplexity": 94.77282172186514, "lr": 0.001305, "grad_norm": 0.348874, "tokens_per_sec": 106395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:23.871459+00:00", "epoch": 0, "step": 512, "train_loss": 4.630467414855957, "perplexity": 102.56199190721276, "lr": 0.001305, "grad_norm": 0.359556, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:24.178131+00:00", "epoch": 0, "step": 513, "train_loss": 4.71079683303833, "perplexity": 111.14068520094591, "lr": 0.001305, "grad_norm": 0.342352, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:24.484960+00:00", "epoch": 0, "step": 514, "train_loss": 4.572125434875488, "perplexity": 96.7495262130816, "lr": 0.001305, "grad_norm": 0.359401, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:24.791336+00:00", "epoch": 0, "step": 515, "train_loss": 4.475518226623535, "perplexity": 87.84010970013422, "lr": 0.001305, "grad_norm": 0.335385, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:25.098423+00:00", "epoch": 0, "step": 516, "train_loss": 4.5372748374938965, "perplexity": 93.43582504587474, "lr": 0.001305, "grad_norm": 0.333239, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:25.405517+00:00", "epoch": 0, "step": 517, "train_loss": 4.694632053375244, "perplexity": 109.358563123976, "lr": 0.001305, "grad_norm": 0.303019, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:25.711895+00:00", "epoch": 0, "step": 518, "train_loss": 4.666201114654541, "perplexity": 106.29317887415623, "lr": 0.001305, "grad_norm": 0.31716, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:26.019040+00:00", "epoch": 0, "step": 519, "train_loss": 4.586929798126221, "perplexity": 98.19249611431836, "lr": 0.001305, "grad_norm": 0.272409, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:26.326242+00:00", "epoch": 0, "step": 520, "train_loss": 4.600167274475098, "perplexity": 99.50095822051375, "lr": 0.001305, "grad_norm": 0.296492, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:26.633658+00:00", "epoch": 0, "step": 521, "train_loss": 4.619762897491455, "perplexity": 101.46997049257533, "lr": 0.001305, "grad_norm": 0.35715, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:26.941468+00:00", "epoch": 0, "step": 522, "train_loss": 4.7053351402282715, "perplexity": 110.53532357413306, "lr": 0.001305, "grad_norm": 0.434658, "tokens_per_sec": 106456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:27.250831+00:00", "epoch": 0, "step": 523, "train_loss": 4.709986686706543, "perplexity": 111.05068144554693, "lr": 0.001305, "grad_norm": 0.462007, "tokens_per_sec": 105921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:27.557211+00:00", "epoch": 0, "step": 524, "train_loss": 4.614144325256348, "perplexity": 100.90145275824385, "lr": 0.001305, "grad_norm": 0.426878, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:27.863666+00:00", "epoch": 0, "step": 525, "train_loss": 4.691253662109375, "perplexity": 108.98973049075, "lr": 0.001305, "grad_norm": 0.49387, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:28.170389+00:00", "epoch": 0, "step": 526, "train_loss": 4.674332618713379, "perplexity": 107.16102595871308, "lr": 0.001305, "grad_norm": 0.491847, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:28.477624+00:00", "epoch": 0, "step": 527, "train_loss": 4.53831148147583, "perplexity": 93.53273495346325, "lr": 0.001305, "grad_norm": 0.472079, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:28.784411+00:00", "epoch": 0, "step": 528, "train_loss": 4.550067901611328, "perplexity": 94.63883422609491, "lr": 0.001305, "grad_norm": 0.493738, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:29.092535+00:00", "epoch": 0, "step": 529, "train_loss": 4.557272911071777, "perplexity": 95.32317028546633, "lr": 0.001305, "grad_norm": 0.494131, "tokens_per_sec": 106348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:29.399292+00:00", "epoch": 0, "step": 530, "train_loss": 4.698852062225342, "perplexity": 109.82103235399249, "lr": 0.001305, "grad_norm": 0.426954, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:29.705674+00:00", "epoch": 0, "step": 531, "train_loss": 4.659334182739258, "perplexity": 105.56577123951102, "lr": 0.001305, "grad_norm": 0.474624, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:30.012728+00:00", "epoch": 0, "step": 532, "train_loss": 4.5951008796691895, "perplexity": 98.99812194173923, "lr": 0.001305, "grad_norm": 0.446985, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:30.320489+00:00", "epoch": 0, "step": 533, "train_loss": 4.407360076904297, "perplexity": 82.05256487124848, "lr": 0.001305, "grad_norm": 0.365643, "tokens_per_sec": 106472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:30.627975+00:00", "epoch": 0, "step": 534, "train_loss": 4.5591721534729, "perplexity": 95.50438412231158, "lr": 0.001305, "grad_norm": 0.36647, "tokens_per_sec": 106568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:30.935178+00:00", "epoch": 0, "step": 535, "train_loss": 4.559741973876953, "perplexity": 95.55881997691837, "lr": 0.001305, "grad_norm": 0.334427, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:31.242445+00:00", "epoch": 0, "step": 536, "train_loss": 4.6320481300354, "perplexity": 102.72424140599335, "lr": 0.001305, "grad_norm": 0.363312, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:31.548814+00:00", "epoch": 0, "step": 537, "train_loss": 4.557892322540283, "perplexity": 95.38223284047908, "lr": 0.001305, "grad_norm": 0.347855, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:31.856447+00:00", "epoch": 0, "step": 538, "train_loss": 4.501190185546875, "perplexity": 90.12433217097562, "lr": 0.001305, "grad_norm": 0.349636, "tokens_per_sec": 106518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:32.163984+00:00", "epoch": 0, "step": 539, "train_loss": 4.461602210998535, "perplexity": 86.62619140413149, "lr": 0.001305, "grad_norm": 0.356241, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:32.471916+00:00", "epoch": 0, "step": 540, "train_loss": 4.459258556365967, "perplexity": 86.42340725017601, "lr": 0.001305, "grad_norm": 0.376769, "tokens_per_sec": 106473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:32.779297+00:00", "epoch": 0, "step": 541, "train_loss": 4.587060451507568, "perplexity": 98.20532613408314, "lr": 0.001305, "grad_norm": 0.351023, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:33.085227+00:00", "epoch": 0, "step": 542, "train_loss": 4.616785526275635, "perplexity": 101.16830602955315, "lr": 0.001305, "grad_norm": 0.354472, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:33.391867+00:00", "epoch": 0, "step": 543, "train_loss": 4.610724449157715, "perplexity": 100.5569716687022, "lr": 0.001305, "grad_norm": 0.352237, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:33.698876+00:00", "epoch": 0, "step": 544, "train_loss": 4.420979022979736, "perplexity": 83.17767836755235, "lr": 0.001305, "grad_norm": 0.343419, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:34.007645+00:00", "epoch": 0, "step": 545, "train_loss": 4.541170120239258, "perplexity": 93.80049378579969, "lr": 0.001305, "grad_norm": 0.316078, "tokens_per_sec": 106124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:34.315919+00:00", "epoch": 0, "step": 546, "train_loss": 4.604968547821045, "perplexity": 99.97983821605625, "lr": 0.001305, "grad_norm": 0.354499, "tokens_per_sec": 106295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:34.622878+00:00", "epoch": 0, "step": 547, "train_loss": 4.4388747215271, "perplexity": 84.67959990999881, "lr": 0.001305, "grad_norm": 0.373219, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:34.929488+00:00", "epoch": 0, "step": 548, "train_loss": 4.518756866455078, "perplexity": 91.72150499679145, "lr": 0.001305, "grad_norm": 0.347985, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:35.236161+00:00", "epoch": 0, "step": 549, "train_loss": 4.510162353515625, "perplexity": 90.93658118576543, "lr": 0.001305, "grad_norm": 0.320501, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:35.543851+00:00", "epoch": 0, "step": 550, "train_loss": 4.486406326293945, "perplexity": 88.80174727225877, "lr": 0.001305, "grad_norm": 0.391868, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:35.851704+00:00", "epoch": 0, "step": 551, "train_loss": 4.474673748016357, "perplexity": 87.76596191915422, "lr": 0.001305, "grad_norm": 0.43292, "tokens_per_sec": 106383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:36.160469+00:00", "epoch": 0, "step": 552, "train_loss": 4.521713733673096, "perplexity": 91.9931146670771, "lr": 0.001305, "grad_norm": 0.391238, "tokens_per_sec": 106127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:36.466737+00:00", "epoch": 0, "step": 553, "train_loss": 4.564466953277588, "perplexity": 96.0114018107726, "lr": 0.001305, "grad_norm": 0.402142, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:36.773516+00:00", "epoch": 0, "step": 554, "train_loss": 4.569301128387451, "perplexity": 96.47666140678506, "lr": 0.001305, "grad_norm": 0.400275, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:37.080912+00:00", "epoch": 0, "step": 555, "train_loss": 4.455203533172607, "perplexity": 86.0736679091336, "lr": 0.001305, "grad_norm": 0.408237, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:37.388251+00:00", "epoch": 0, "step": 556, "train_loss": 4.441575050354004, "perplexity": 84.9085716850947, "lr": 0.001305, "grad_norm": 0.335608, "tokens_per_sec": 106620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:37.697785+00:00", "epoch": 0, "step": 557, "train_loss": 4.42045259475708, "perplexity": 83.13390281351914, "lr": 0.001305, "grad_norm": 0.312326, "tokens_per_sec": 105862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:38.004757+00:00", "epoch": 0, "step": 558, "train_loss": 4.556873798370361, "perplexity": 95.28513318851533, "lr": 0.001305, "grad_norm": 0.306971, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:38.311715+00:00", "epoch": 0, "step": 559, "train_loss": 4.5966386795043945, "perplexity": 99.15047835415507, "lr": 0.001305, "grad_norm": 0.312111, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:38.618950+00:00", "epoch": 0, "step": 560, "train_loss": 4.5895843505859375, "perplexity": 98.45349951674972, "lr": 0.001305, "grad_norm": 0.328908, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:38.925493+00:00", "epoch": 0, "step": 561, "train_loss": 4.614415168762207, "perplexity": 100.92878496266314, "lr": 0.001305, "grad_norm": 0.399568, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:39.234215+00:00", "epoch": 0, "step": 562, "train_loss": 4.45527458190918, "perplexity": 86.07978355174245, "lr": 0.001305, "grad_norm": 0.460838, "tokens_per_sec": 106142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:39.543106+00:00", "epoch": 0, "step": 563, "train_loss": 4.604249477386475, "perplexity": 99.90797151204963, "lr": 0.001305, "grad_norm": 0.431324, "tokens_per_sec": 106082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:39.852260+00:00", "epoch": 0, "step": 564, "train_loss": 4.503125190734863, "perplexity": 90.29889205396378, "lr": 0.001305, "grad_norm": 0.423723, "tokens_per_sec": 105993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:40.159714+00:00", "epoch": 0, "step": 565, "train_loss": 4.5143842697143555, "perplexity": 91.3213194057095, "lr": 0.001305, "grad_norm": 0.48825, "tokens_per_sec": 106579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:40.465945+00:00", "epoch": 0, "step": 566, "train_loss": 4.486563205718994, "perplexity": 88.81567953212806, "lr": 0.001305, "grad_norm": 0.49165, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:40.773076+00:00", "epoch": 0, "step": 567, "train_loss": 4.564589977264404, "perplexity": 96.02321424279454, "lr": 0.001305, "grad_norm": 0.52617, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:41.081244+00:00", "epoch": 0, "step": 568, "train_loss": 4.560492515563965, "perplexity": 95.63056777636187, "lr": 0.001305, "grad_norm": 0.568376, "tokens_per_sec": 106332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:41.389951+00:00", "epoch": 0, "step": 569, "train_loss": 4.485826015472412, "perplexity": 88.75022960692073, "lr": 0.001305, "grad_norm": 0.527774, "tokens_per_sec": 106146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:41.697850+00:00", "epoch": 0, "step": 570, "train_loss": 4.555840969085693, "perplexity": 95.18677071712912, "lr": 0.001305, "grad_norm": 0.554459, "tokens_per_sec": 106479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:42.005525+00:00", "epoch": 0, "step": 571, "train_loss": 4.571557998657227, "perplexity": 96.69464260075598, "lr": 0.001305, "grad_norm": 0.470086, "tokens_per_sec": 106448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:42.312485+00:00", "epoch": 0, "step": 572, "train_loss": 4.43071985244751, "perplexity": 83.99185689748107, "lr": 0.001305, "grad_norm": 0.464383, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:42.620392+00:00", "epoch": 0, "step": 573, "train_loss": 4.495109558105469, "perplexity": 89.57798244161177, "lr": 0.001305, "grad_norm": 0.428242, "tokens_per_sec": 106422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:42.929557+00:00", "epoch": 0, "step": 574, "train_loss": 4.5096893310546875, "perplexity": 90.89357631228111, "lr": 0.001305, "grad_norm": 0.377975, "tokens_per_sec": 105989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:43.237701+00:00", "epoch": 0, "step": 575, "train_loss": 4.477972507476807, "perplexity": 88.05595876829098, "lr": 0.001305, "grad_norm": 0.426954, "tokens_per_sec": 106340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:43.545397+00:00", "epoch": 0, "step": 576, "train_loss": 4.566281318664551, "perplexity": 96.18575970160184, "lr": 0.001305, "grad_norm": 0.384947, "tokens_per_sec": 106494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:43.852821+00:00", "epoch": 0, "step": 577, "train_loss": 4.435923099517822, "perplexity": 84.43002624392248, "lr": 0.001305, "grad_norm": 0.311702, "tokens_per_sec": 106589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:44.160037+00:00", "epoch": 0, "step": 578, "train_loss": 4.435812950134277, "perplexity": 84.4207268407502, "lr": 0.001305, "grad_norm": 0.314687, "tokens_per_sec": 106663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:44.469214+00:00", "epoch": 0, "step": 579, "train_loss": 4.357483863830566, "perplexity": 78.06047632966958, "lr": 0.001305, "grad_norm": 0.271201, "tokens_per_sec": 105984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:44.777397+00:00", "epoch": 0, "step": 580, "train_loss": 4.462371349334717, "perplexity": 86.69284455832002, "lr": 0.001305, "grad_norm": 0.277211, "tokens_per_sec": 106383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:45.086466+00:00", "epoch": 0, "step": 581, "train_loss": 4.50649881362915, "perplexity": 90.60404090268027, "lr": 0.001305, "grad_norm": 0.274096, "tokens_per_sec": 105969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:45.393970+00:00", "epoch": 0, "step": 582, "train_loss": 4.469743728637695, "perplexity": 87.33433885535973, "lr": 0.001305, "grad_norm": 0.277021, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:45.701181+00:00", "epoch": 0, "step": 583, "train_loss": 4.510488986968994, "perplexity": 90.96628896682917, "lr": 0.001305, "grad_norm": 0.28535, "tokens_per_sec": 106663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:46.008725+00:00", "epoch": 0, "step": 584, "train_loss": 4.424190521240234, "perplexity": 83.44523273217358, "lr": 0.001305, "grad_norm": 0.298918, "tokens_per_sec": 106547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:46.316204+00:00", "epoch": 0, "step": 585, "train_loss": 4.620384693145752, "perplexity": 101.53308369899453, "lr": 0.001305, "grad_norm": 0.29959, "tokens_per_sec": 106570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:46.623658+00:00", "epoch": 0, "step": 586, "train_loss": 4.369845867156982, "perplexity": 79.03144941875706, "lr": 0.001305, "grad_norm": 0.316678, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:46.931822+00:00", "epoch": 0, "step": 587, "train_loss": 4.564305782318115, "perplexity": 95.9959288079553, "lr": 0.001305, "grad_norm": 0.31488, "tokens_per_sec": 106332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:47.239237+00:00", "epoch": 0, "step": 588, "train_loss": 4.482059955596924, "perplexity": 88.41661952034327, "lr": 0.001305, "grad_norm": 0.294088, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:47.545852+00:00", "epoch": 0, "step": 589, "train_loss": 4.422412395477295, "perplexity": 83.2969884516043, "lr": 0.001305, "grad_norm": 0.266261, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:47.854329+00:00", "epoch": 0, "step": 590, "train_loss": 4.473036289215088, "perplexity": 87.62236637034032, "lr": 0.001305, "grad_norm": 0.293932, "tokens_per_sec": 106286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:48.161954+00:00", "epoch": 0, "step": 591, "train_loss": 4.383162021636963, "perplexity": 80.09088254070936, "lr": 0.001305, "grad_norm": 0.281486, "tokens_per_sec": 106458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:48.470842+00:00", "epoch": 0, "step": 592, "train_loss": 4.4256181716918945, "perplexity": 83.56444843530011, "lr": 0.001305, "grad_norm": 0.309602, "tokens_per_sec": 106084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:48.778332+00:00", "epoch": 0, "step": 593, "train_loss": 4.409634590148926, "perplexity": 82.23940692361441, "lr": 0.001305, "grad_norm": 0.31828, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:49.085315+00:00", "epoch": 0, "step": 594, "train_loss": 4.434841632843018, "perplexity": 84.33876733981306, "lr": 0.001305, "grad_norm": 0.31369, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:49.392987+00:00", "epoch": 0, "step": 595, "train_loss": 4.520889759063721, "perplexity": 91.91734589641254, "lr": 0.001305, "grad_norm": 0.330407, "tokens_per_sec": 106494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:49.702075+00:00", "epoch": 0, "step": 596, "train_loss": 4.48704195022583, "perplexity": 88.85820973056208, "lr": 0.001305, "grad_norm": 0.340415, "tokens_per_sec": 106014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:50.010089+00:00", "epoch": 0, "step": 597, "train_loss": 4.362453460693359, "perplexity": 78.44937095232024, "lr": 0.001305, "grad_norm": 0.330787, "tokens_per_sec": 106385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:50.316975+00:00", "epoch": 0, "step": 598, "train_loss": 4.40612268447876, "perplexity": 81.95109644006186, "lr": 0.001305, "grad_norm": 0.321609, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:50.624528+00:00", "epoch": 0, "step": 599, "train_loss": 4.419227600097656, "perplexity": 83.03212657695586, "lr": 0.001305, "grad_norm": 0.32161, "tokens_per_sec": 106545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:50.932327+00:00", "epoch": 0, "step": 600, "train_loss": 4.5336198806762695, "perplexity": 93.09494447147937, "lr": 0.001305, "grad_norm": 0.279919, "tokens_per_sec": 106519} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:46:56.641182+00:00", "step": 600, "epoch": 0, "val_loss": 4.370036721229553, "val_ppl": 79.0465343322026, "eval_train_loss": 4.5336198806762695, "eval_train_ppl": 93.09494447147937} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:46:57.484908+00:00", "step": 600, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_4p3700_epoch_0000_step_0000600.pt", "val_loss": 4.370036721229553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:58.460312+00:00", "epoch": 0, "step": 601, "train_loss": 4.503144264221191, "perplexity": 90.3006143850722, "lr": 0.001305, "grad_norm": 0.303143, "tokens_per_sec": 4353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:58.766639+00:00", "epoch": 0, "step": 602, "train_loss": 4.497891426086426, "perplexity": 89.82752349697195, "lr": 0.001305, "grad_norm": 0.307167, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:59.072430+00:00", "epoch": 0, "step": 603, "train_loss": 4.604333400726318, "perplexity": 99.91635647453803, "lr": 0.001305, "grad_norm": 0.376348, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:59.380370+00:00", "epoch": 0, "step": 604, "train_loss": 4.496495723724365, "perplexity": 89.70223846089735, "lr": 0.001305, "grad_norm": 0.437026, "tokens_per_sec": 106413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:59.687224+00:00", "epoch": 0, "step": 605, "train_loss": 4.444781303405762, "perplexity": 85.18124695153992, "lr": 0.001305, "grad_norm": 0.511037, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:46:59.993631+00:00", "epoch": 0, "step": 606, "train_loss": 4.389227390289307, "perplexity": 80.57813947167297, "lr": 0.001305, "grad_norm": 0.543132, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:00.301486+00:00", "epoch": 0, "step": 607, "train_loss": 4.359407424926758, "perplexity": 78.21077493301746, "lr": 0.001305, "grad_norm": 0.463888, "tokens_per_sec": 106441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:00.608892+00:00", "epoch": 0, "step": 608, "train_loss": 4.400228977203369, "perplexity": 81.4695211925073, "lr": 0.001305, "grad_norm": 0.451654, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:00.915732+00:00", "epoch": 0, "step": 609, "train_loss": 4.51706075668335, "perplexity": 91.56606711300903, "lr": 0.001305, "grad_norm": 0.378443, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:01.222487+00:00", "epoch": 0, "step": 610, "train_loss": 4.478240013122559, "perplexity": 88.07951738529465, "lr": 0.001305, "grad_norm": 0.402905, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:01.530016+00:00", "epoch": 0, "step": 611, "train_loss": 4.36488151550293, "perplexity": 78.64008175922092, "lr": 0.001305, "grad_norm": 0.367458, "tokens_per_sec": 106491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:01.837812+00:00", "epoch": 0, "step": 612, "train_loss": 4.504290580749512, "perplexity": 90.40418682391702, "lr": 0.001305, "grad_norm": 0.428727, "tokens_per_sec": 106460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:02.145269+00:00", "epoch": 0, "step": 613, "train_loss": 4.4101996421813965, "perplexity": 82.28588959897411, "lr": 0.001305, "grad_norm": 0.472546, "tokens_per_sec": 106579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:02.451446+00:00", "epoch": 0, "step": 614, "train_loss": 4.472100257873535, "perplexity": 87.54038746259567, "lr": 0.001305, "grad_norm": 0.406207, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:02.758103+00:00", "epoch": 0, "step": 615, "train_loss": 4.5043721199035645, "perplexity": 90.41155860537378, "lr": 0.001305, "grad_norm": 0.388878, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:03.064253+00:00", "epoch": 0, "step": 616, "train_loss": 4.281935691833496, "perplexity": 72.38041066341897, "lr": 0.001305, "grad_norm": 0.345867, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:03.372392+00:00", "epoch": 0, "step": 617, "train_loss": 4.472506046295166, "perplexity": 87.57591754661388, "lr": 0.001305, "grad_norm": 0.377835, "tokens_per_sec": 106342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:03.680886+00:00", "epoch": 0, "step": 618, "train_loss": 4.412426471710205, "perplexity": 82.4693304176499, "lr": 0.001305, "grad_norm": 0.328305, "tokens_per_sec": 106219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:03.987858+00:00", "epoch": 0, "step": 619, "train_loss": 4.285565376281738, "perplexity": 72.64360608356503, "lr": 0.001305, "grad_norm": 0.35512, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:04.294090+00:00", "epoch": 0, "step": 620, "train_loss": 4.449769973754883, "perplexity": 85.60724982296678, "lr": 0.001305, "grad_norm": 0.343674, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:04.601921+00:00", "epoch": 0, "step": 621, "train_loss": 4.480615615844727, "perplexity": 88.28900806135104, "lr": 0.001305, "grad_norm": 0.361891, "tokens_per_sec": 106389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:04.907881+00:00", "epoch": 0, "step": 622, "train_loss": 4.479185581207275, "perplexity": 88.16284195416956, "lr": 0.001305, "grad_norm": 0.368165, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:05.216190+00:00", "epoch": 0, "step": 623, "train_loss": 4.34943151473999, "perplexity": 77.43443007831772, "lr": 0.001305, "grad_norm": 0.4085, "tokens_per_sec": 106284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:05.523997+00:00", "epoch": 0, "step": 624, "train_loss": 4.3905463218688965, "perplexity": 80.68448664135038, "lr": 0.001305, "grad_norm": 0.39197, "tokens_per_sec": 106455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:05.831208+00:00", "epoch": 0, "step": 625, "train_loss": 4.330820560455322, "perplexity": 76.00662902193133, "lr": 0.001305, "grad_norm": 0.352633, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:06.137539+00:00", "epoch": 0, "step": 626, "train_loss": 4.398430347442627, "perplexity": 81.32311938788678, "lr": 0.001305, "grad_norm": 0.33887, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:06.444998+00:00", "epoch": 0, "step": 627, "train_loss": 4.34576416015625, "perplexity": 77.15097065703331, "lr": 0.001305, "grad_norm": 0.316875, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:06.752130+00:00", "epoch": 0, "step": 628, "train_loss": 4.441486358642578, "perplexity": 84.9010413325021, "lr": 0.001305, "grad_norm": 0.332853, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:07.059920+00:00", "epoch": 0, "step": 629, "train_loss": 4.441801071166992, "perplexity": 84.92776495845456, "lr": 0.001305, "grad_norm": 0.368872, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:07.366500+00:00", "epoch": 0, "step": 630, "train_loss": 4.39078950881958, "perplexity": 80.70411044165353, "lr": 0.001305, "grad_norm": 0.4935, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:07.672890+00:00", "epoch": 0, "step": 631, "train_loss": 4.430431842803955, "perplexity": 83.9676699159237, "lr": 0.001305, "grad_norm": 0.533365, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:07.979208+00:00", "epoch": 0, "step": 632, "train_loss": 4.371322154998779, "perplexity": 79.14820875063891, "lr": 0.001305, "grad_norm": 0.417801, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:08.285849+00:00", "epoch": 0, "step": 633, "train_loss": 4.364176273345947, "perplexity": 78.58464101021035, "lr": 0.001305, "grad_norm": 0.389985, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:08.594235+00:00", "epoch": 0, "step": 634, "train_loss": 4.246647357940674, "perplexity": 69.87076755245901, "lr": 0.001305, "grad_norm": 0.312424, "tokens_per_sec": 106256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:08.902514+00:00", "epoch": 0, "step": 635, "train_loss": 4.487246990203857, "perplexity": 88.8764310839219, "lr": 0.001305, "grad_norm": 0.298346, "tokens_per_sec": 106293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:09.209611+00:00", "epoch": 0, "step": 636, "train_loss": 4.3214192390441895, "perplexity": 75.29541468876155, "lr": 0.001305, "grad_norm": 0.346655, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:09.517146+00:00", "epoch": 0, "step": 637, "train_loss": 4.338032245635986, "perplexity": 76.55674615364015, "lr": 0.001305, "grad_norm": 0.302186, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:09.823686+00:00", "epoch": 0, "step": 638, "train_loss": 4.284829616546631, "perplexity": 72.59017750090592, "lr": 0.001305, "grad_norm": 0.313621, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:10.131408+00:00", "epoch": 0, "step": 639, "train_loss": 4.338772296905518, "perplexity": 76.61342304013652, "lr": 0.001305, "grad_norm": 0.282107, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:10.440622+00:00", "epoch": 0, "step": 640, "train_loss": 4.315840244293213, "perplexity": 74.8765115810955, "lr": 0.001305, "grad_norm": 0.280138, "tokens_per_sec": 106050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:10.748711+00:00", "epoch": 0, "step": 641, "train_loss": 4.3860697746276855, "perplexity": 80.32410595766497, "lr": 0.001305, "grad_norm": 0.289293, "tokens_per_sec": 106281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:11.055124+00:00", "epoch": 0, "step": 642, "train_loss": 4.366186141967773, "perplexity": 78.74274464487883, "lr": 0.001305, "grad_norm": 0.328835, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:11.362661+00:00", "epoch": 0, "step": 643, "train_loss": 4.377898693084717, "perplexity": 79.6704453320415, "lr": 0.001305, "grad_norm": 0.312933, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:11.669328+00:00", "epoch": 0, "step": 644, "train_loss": 4.410965442657471, "perplexity": 82.3489283068587, "lr": 0.001305, "grad_norm": 0.289688, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:11.976692+00:00", "epoch": 0, "step": 645, "train_loss": 4.513492107391357, "perplexity": 91.23988229821045, "lr": 0.001305, "grad_norm": 0.348659, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:12.284123+00:00", "epoch": 0, "step": 646, "train_loss": 4.326445579528809, "perplexity": 75.67482781089205, "lr": 0.001305, "grad_norm": 0.387796, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:12.590995+00:00", "epoch": 0, "step": 647, "train_loss": 4.408900260925293, "perplexity": 82.17903829171422, "lr": 0.001305, "grad_norm": 0.343854, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:12.898917+00:00", "epoch": 0, "step": 648, "train_loss": 4.498479843139648, "perplexity": 89.88039509739649, "lr": 0.001305, "grad_norm": 0.328181, "tokens_per_sec": 106417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:13.205722+00:00", "epoch": 0, "step": 649, "train_loss": 4.297881126403809, "perplexity": 73.54379847812227, "lr": 0.001305, "grad_norm": 0.330633, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:13.513462+00:00", "epoch": 0, "step": 650, "train_loss": 4.331007480621338, "perplexity": 76.02083752153258, "lr": 0.001305, "grad_norm": 0.358628, "tokens_per_sec": 106537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:13.821832+00:00", "epoch": 0, "step": 651, "train_loss": 4.2421674728393555, "perplexity": 69.55845462722024, "lr": 0.001305, "grad_norm": 0.345335, "tokens_per_sec": 106207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:14.131068+00:00", "epoch": 0, "step": 652, "train_loss": 4.202871799468994, "perplexity": 66.87811606325234, "lr": 0.001305, "grad_norm": 0.346843, "tokens_per_sec": 105964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:14.438702+00:00", "epoch": 0, "step": 653, "train_loss": 4.357806205749512, "perplexity": 78.08564254924933, "lr": 0.001305, "grad_norm": 0.373475, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:14.746240+00:00", "epoch": 0, "step": 654, "train_loss": 4.396566867828369, "perplexity": 81.17171652462713, "lr": 0.001305, "grad_norm": 0.362904, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:15.054610+00:00", "epoch": 0, "step": 655, "train_loss": 4.3390302658081055, "perplexity": 76.63318947025421, "lr": 0.001305, "grad_norm": 0.369088, "tokens_per_sec": 106262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:15.362317+00:00", "epoch": 0, "step": 656, "train_loss": 4.35036563873291, "perplexity": 77.50679723201702, "lr": 0.001305, "grad_norm": 0.332086, "tokens_per_sec": 106492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:15.670809+00:00", "epoch": 0, "step": 657, "train_loss": 4.3852152824401855, "perplexity": 80.2554989529055, "lr": 0.001305, "grad_norm": 0.296722, "tokens_per_sec": 106220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:15.989740+00:00", "epoch": 0, "step": 658, "train_loss": 4.327958106994629, "perplexity": 75.78937467222643, "lr": 0.001305, "grad_norm": 0.303312, "tokens_per_sec": 102744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:16.299155+00:00", "epoch": 0, "step": 659, "train_loss": 4.303308486938477, "perplexity": 73.94403231176041, "lr": 0.001305, "grad_norm": 0.342037, "tokens_per_sec": 105903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:16.621834+00:00", "epoch": 0, "step": 660, "train_loss": 4.19517707824707, "perplexity": 66.36548242014776, "lr": 0.001305, "grad_norm": 0.338583, "tokens_per_sec": 101600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:16.930404+00:00", "epoch": 0, "step": 661, "train_loss": 4.406868934631348, "perplexity": 82.01227528280621, "lr": 0.001305, "grad_norm": 0.308502, "tokens_per_sec": 106138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:17.237689+00:00", "epoch": 0, "step": 662, "train_loss": 4.352273464202881, "perplexity": 77.6548078181611, "lr": 0.001305, "grad_norm": 0.329317, "tokens_per_sec": 106637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:17.559842+00:00", "epoch": 0, "step": 663, "train_loss": 4.310058116912842, "perplexity": 74.44481531746972, "lr": 0.001305, "grad_norm": 0.343968, "tokens_per_sec": 101717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:17.868110+00:00", "epoch": 0, "step": 664, "train_loss": 4.3552165031433105, "perplexity": 77.88368557435446, "lr": 0.001305, "grad_norm": 0.331723, "tokens_per_sec": 106297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:18.176835+00:00", "epoch": 0, "step": 665, "train_loss": 4.315164089202881, "perplexity": 74.82590055902196, "lr": 0.001305, "grad_norm": 0.290923, "tokens_per_sec": 106139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:18.496974+00:00", "epoch": 0, "step": 666, "train_loss": 4.3895487785339355, "perplexity": 80.60404050039337, "lr": 0.001305, "grad_norm": 0.296171, "tokens_per_sec": 102356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:18.818012+00:00", "epoch": 0, "step": 667, "train_loss": 4.257443904876709, "perplexity": 70.62921752676301, "lr": 0.001305, "grad_norm": 0.269432, "tokens_per_sec": 102070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:19.125387+00:00", "epoch": 0, "step": 668, "train_loss": 4.2055511474609375, "perplexity": 67.05754607961579, "lr": 0.001305, "grad_norm": 0.297778, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:19.445613+00:00", "epoch": 0, "step": 669, "train_loss": 4.328253746032715, "perplexity": 75.81178428246623, "lr": 0.001305, "grad_norm": 0.28794, "tokens_per_sec": 102328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:19.764598+00:00", "epoch": 0, "step": 670, "train_loss": 4.3219895362854, "perplexity": 75.33836770286294, "lr": 0.001305, "grad_norm": 0.279535, "tokens_per_sec": 102779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:20.087625+00:00", "epoch": 0, "step": 671, "train_loss": 4.2972283363342285, "perplexity": 73.4958054831761, "lr": 0.001305, "grad_norm": 0.262023, "tokens_per_sec": 101389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:20.393993+00:00", "epoch": 0, "step": 672, "train_loss": 4.315375804901123, "perplexity": 74.84174405390452, "lr": 0.001305, "grad_norm": 0.261114, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:20.703395+00:00", "epoch": 0, "step": 673, "train_loss": 4.2023468017578125, "perplexity": 66.8430144203383, "lr": 0.001305, "grad_norm": 0.286406, "tokens_per_sec": 105909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:21.022856+00:00", "epoch": 0, "step": 674, "train_loss": 4.326213836669922, "perplexity": 75.6572927418414, "lr": 0.001305, "grad_norm": 0.287952, "tokens_per_sec": 102573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:21.342678+00:00", "epoch": 0, "step": 675, "train_loss": 4.325343608856201, "perplexity": 75.59148230057801, "lr": 0.001305, "grad_norm": 0.315204, "tokens_per_sec": 102458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:21.650189+00:00", "epoch": 0, "step": 676, "train_loss": 4.3068766593933105, "perplexity": 74.20834865376366, "lr": 0.001305, "grad_norm": 0.350876, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:21.972471+00:00", "epoch": 0, "step": 677, "train_loss": 4.38119649887085, "perplexity": 79.93361669314952, "lr": 0.001305, "grad_norm": 0.332334, "tokens_per_sec": 101675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:22.279454+00:00", "epoch": 0, "step": 678, "train_loss": 4.329805850982666, "perplexity": 75.92954349180108, "lr": 0.001305, "grad_norm": 0.278356, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:22.597182+00:00", "epoch": 0, "step": 679, "train_loss": 4.221006870269775, "perplexity": 68.10201967943321, "lr": 0.001305, "grad_norm": 0.29738, "tokens_per_sec": 103133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:22.916209+00:00", "epoch": 0, "step": 680, "train_loss": 4.306238174438477, "perplexity": 74.16098286240684, "lr": 0.001305, "grad_norm": 0.37233, "tokens_per_sec": 102761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:23.224854+00:00", "epoch": 0, "step": 681, "train_loss": 4.421533107757568, "perplexity": 83.22377862353683, "lr": 0.001305, "grad_norm": 0.431192, "tokens_per_sec": 106116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:23.543722+00:00", "epoch": 0, "step": 682, "train_loss": 4.334879398345947, "perplexity": 76.31575452852185, "lr": 0.001305, "grad_norm": 0.401697, "tokens_per_sec": 102764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:23.851009+00:00", "epoch": 0, "step": 683, "train_loss": 4.274330139160156, "perplexity": 71.83200574236874, "lr": 0.001305, "grad_norm": 0.332873, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:24.158236+00:00", "epoch": 0, "step": 684, "train_loss": 4.348986625671387, "perplexity": 77.39998800886052, "lr": 0.001305, "grad_norm": 0.327805, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:24.464902+00:00", "epoch": 0, "step": 685, "train_loss": 4.2199482917785645, "perplexity": 68.02996648989937, "lr": 0.001305, "grad_norm": 0.322424, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:24.784351+00:00", "epoch": 0, "step": 686, "train_loss": 4.319935321807861, "perplexity": 75.18376538473949, "lr": 0.001305, "grad_norm": 0.285954, "tokens_per_sec": 102578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:25.092831+00:00", "epoch": 0, "step": 687, "train_loss": 4.125310897827148, "perplexity": 61.88704680813676, "lr": 0.001305, "grad_norm": 0.328082, "tokens_per_sec": 106224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:25.400993+00:00", "epoch": 0, "step": 688, "train_loss": 4.320957660675049, "perplexity": 75.26066797382866, "lr": 0.001305, "grad_norm": 0.320788, "tokens_per_sec": 106334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:25.722468+00:00", "epoch": 0, "step": 689, "train_loss": 4.180079936981201, "perplexity": 65.37107857183636, "lr": 0.001305, "grad_norm": 0.300744, "tokens_per_sec": 101930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:26.029771+00:00", "epoch": 0, "step": 690, "train_loss": 4.24464225769043, "perplexity": 69.73081002028697, "lr": 0.001305, "grad_norm": 0.315055, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:26.338281+00:00", "epoch": 0, "step": 691, "train_loss": 4.298783779144287, "perplexity": 73.6102129594611, "lr": 0.001305, "grad_norm": 0.375645, "tokens_per_sec": 106157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:26.645659+00:00", "epoch": 0, "step": 692, "train_loss": 4.3029255867004395, "perplexity": 73.91572454405816, "lr": 0.001305, "grad_norm": 0.411473, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:26.953220+00:00", "epoch": 0, "step": 693, "train_loss": 4.335562705993652, "perplexity": 76.3679194875596, "lr": 0.001305, "grad_norm": 0.428205, "tokens_per_sec": 106541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:27.261155+00:00", "epoch": 0, "step": 694, "train_loss": 4.2478814125061035, "perplexity": 69.95704511680901, "lr": 0.001305, "grad_norm": 0.402748, "tokens_per_sec": 106411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:27.581568+00:00", "epoch": 0, "step": 695, "train_loss": 4.215875625610352, "perplexity": 67.75346657505855, "lr": 0.001305, "grad_norm": 0.407088, "tokens_per_sec": 102268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:27.890077+00:00", "epoch": 0, "step": 696, "train_loss": 4.337821006774902, "perplexity": 76.54057610170616, "lr": 0.001305, "grad_norm": 0.457991, "tokens_per_sec": 106215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:28.196808+00:00", "epoch": 0, "step": 697, "train_loss": 4.255210876464844, "perplexity": 70.4716764396435, "lr": 0.001305, "grad_norm": 0.484043, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:28.503992+00:00", "epoch": 0, "step": 698, "train_loss": 4.330873489379883, "perplexity": 76.010652077532, "lr": 0.001305, "grad_norm": 0.569331, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:28.811291+00:00", "epoch": 0, "step": 699, "train_loss": 4.35929012298584, "perplexity": 78.20160119537626, "lr": 0.001305, "grad_norm": 0.537048, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:29.120594+00:00", "epoch": 0, "step": 700, "train_loss": 4.290926933288574, "perplexity": 73.03413490451148, "lr": 0.001305, "grad_norm": 0.425341, "tokens_per_sec": 105999} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:47:34.784584+00:00", "step": 700, "epoch": 0, "val_loss": 4.242300176620484, "val_ppl": 69.56768590965821, "eval_train_loss": 4.290926933288574, "eval_train_ppl": 73.03413490451148} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:47:35.693497+00:00", "step": 700, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_4p2423_epoch_0000_step_0000700.pt", "val_loss": 4.242300176620484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:36.693507+00:00", "epoch": 0, "step": 701, "train_loss": 4.204995632171631, "perplexity": 67.02030493247621, "lr": 0.001305, "grad_norm": 0.399886, "tokens_per_sec": 4327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:36.998784+00:00", "epoch": 0, "step": 702, "train_loss": 4.25576114654541, "perplexity": 70.51046556598466, "lr": 0.001305, "grad_norm": 0.372892, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:37.304490+00:00", "epoch": 0, "step": 703, "train_loss": 4.340172290802002, "perplexity": 76.72075648033109, "lr": 0.001305, "grad_norm": 0.336408, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:37.611434+00:00", "epoch": 0, "step": 704, "train_loss": 4.13670015335083, "perplexity": 62.59592332307079, "lr": 0.001305, "grad_norm": 0.322725, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:37.917541+00:00", "epoch": 0, "step": 705, "train_loss": 4.3241143226623535, "perplexity": 75.49861582644525, "lr": 0.001305, "grad_norm": 0.322593, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:38.222996+00:00", "epoch": 0, "step": 706, "train_loss": 4.281112194061279, "perplexity": 72.32083009208642, "lr": 0.001305, "grad_norm": 0.327788, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:38.532688+00:00", "epoch": 0, "step": 707, "train_loss": 4.3328094482421875, "perplexity": 76.15794810657277, "lr": 0.001305, "grad_norm": 0.301227, "tokens_per_sec": 105809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:38.839930+00:00", "epoch": 0, "step": 708, "train_loss": 4.24029541015625, "perplexity": 69.42835865192316, "lr": 0.001305, "grad_norm": 0.286186, "tokens_per_sec": 106652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:39.147493+00:00", "epoch": 0, "step": 709, "train_loss": 4.3453240394592285, "perplexity": 77.11702238926749, "lr": 0.001305, "grad_norm": 0.261246, "tokens_per_sec": 106540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:39.454781+00:00", "epoch": 0, "step": 710, "train_loss": 4.263826847076416, "perplexity": 71.08148159203168, "lr": 0.001305, "grad_norm": 0.289313, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:39.761528+00:00", "epoch": 0, "step": 711, "train_loss": 4.2009806632995605, "perplexity": 66.75175995498311, "lr": 0.001305, "grad_norm": 0.283403, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:40.067615+00:00", "epoch": 0, "step": 712, "train_loss": 4.314203262329102, "perplexity": 74.75404035104798, "lr": 0.001305, "grad_norm": 0.267092, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:40.375010+00:00", "epoch": 0, "step": 713, "train_loss": 4.2234272956848145, "perplexity": 68.26705518617712, "lr": 0.001305, "grad_norm": 0.278443, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:40.682186+00:00", "epoch": 0, "step": 714, "train_loss": 4.250495433807373, "perplexity": 70.14015354331528, "lr": 0.001305, "grad_norm": 0.281799, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:41.013230+00:00", "epoch": 0, "step": 715, "train_loss": 4.30029821395874, "perplexity": 73.72177528427808, "lr": 0.001305, "grad_norm": 0.285123, "tokens_per_sec": 98983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:41.319901+00:00", "epoch": 0, "step": 716, "train_loss": 4.236751556396484, "perplexity": 69.18275015941762, "lr": 0.001305, "grad_norm": 0.307607, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:41.626046+00:00", "epoch": 0, "step": 717, "train_loss": 4.366131782531738, "perplexity": 78.73846435002633, "lr": 0.001305, "grad_norm": 0.326242, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:41.931729+00:00", "epoch": 0, "step": 718, "train_loss": 4.333110809326172, "perplexity": 76.18090260699104, "lr": 0.001305, "grad_norm": 0.356944, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:42.238705+00:00", "epoch": 0, "step": 719, "train_loss": 4.216068267822266, "perplexity": 67.76652001000534, "lr": 0.001305, "grad_norm": 0.336725, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:42.545568+00:00", "epoch": 0, "step": 720, "train_loss": 4.3052520751953125, "perplexity": 74.08788881824773, "lr": 0.001305, "grad_norm": 0.313347, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:42.852651+00:00", "epoch": 0, "step": 721, "train_loss": 4.232157230377197, "perplexity": 68.8656310822958, "lr": 0.001305, "grad_norm": 0.322529, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:43.159017+00:00", "epoch": 0, "step": 722, "train_loss": 4.170650959014893, "perplexity": 64.75759292884092, "lr": 0.001305, "grad_norm": 0.306447, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:43.464457+00:00", "epoch": 0, "step": 723, "train_loss": 4.22348690032959, "perplexity": 68.27112434102041, "lr": 0.001305, "grad_norm": 0.33423, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:43.770784+00:00", "epoch": 0, "step": 724, "train_loss": 4.21588659286499, "perplexity": 67.75420964865386, "lr": 0.001305, "grad_norm": 0.303587, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:44.078386+00:00", "epoch": 0, "step": 725, "train_loss": 4.363953590393066, "perplexity": 78.5671434985702, "lr": 0.001305, "grad_norm": 0.331421, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:44.386478+00:00", "epoch": 0, "step": 726, "train_loss": 4.246249198913574, "perplexity": 69.84295341322039, "lr": 0.001305, "grad_norm": 0.294354, "tokens_per_sec": 106358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:44.692576+00:00", "epoch": 0, "step": 727, "train_loss": 4.189510822296143, "perplexity": 65.99050198129386, "lr": 0.001305, "grad_norm": 0.314344, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:44.998189+00:00", "epoch": 0, "step": 728, "train_loss": 4.2235002517700195, "perplexity": 68.27203586495517, "lr": 0.001305, "grad_norm": 0.339372, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:45.306056+00:00", "epoch": 0, "step": 729, "train_loss": 4.218882083892822, "perplexity": 67.95747105762958, "lr": 0.001305, "grad_norm": 0.336796, "tokens_per_sec": 106438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:45.612661+00:00", "epoch": 0, "step": 730, "train_loss": 4.271677494049072, "perplexity": 71.64171342409382, "lr": 0.001305, "grad_norm": 0.31246, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:45.922066+00:00", "epoch": 0, "step": 731, "train_loss": 4.2060136795043945, "perplexity": 67.0885695175472, "lr": 0.001305, "grad_norm": 0.293144, "tokens_per_sec": 105856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:46.229562+00:00", "epoch": 0, "step": 732, "train_loss": 4.382750034332275, "perplexity": 80.05789291000251, "lr": 0.001305, "grad_norm": 0.285077, "tokens_per_sec": 106562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:46.536148+00:00", "epoch": 0, "step": 733, "train_loss": 4.236140727996826, "perplexity": 69.14050427466042, "lr": 0.001305, "grad_norm": 0.277644, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:46.843328+00:00", "epoch": 0, "step": 734, "train_loss": 4.440560340881348, "perplexity": 84.82245785074667, "lr": 0.001305, "grad_norm": 0.302554, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:47.150549+00:00", "epoch": 0, "step": 735, "train_loss": 4.309920310974121, "perplexity": 74.43455708665081, "lr": 0.001305, "grad_norm": 0.292768, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:47.456978+00:00", "epoch": 0, "step": 736, "train_loss": 4.2863264083862305, "perplexity": 72.69891124181194, "lr": 0.001305, "grad_norm": 0.301976, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:47.764845+00:00", "epoch": 0, "step": 737, "train_loss": 4.1765031814575195, "perplexity": 65.13767985935849, "lr": 0.001305, "grad_norm": 0.34567, "tokens_per_sec": 106437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:48.072122+00:00", "epoch": 0, "step": 738, "train_loss": 4.240655899047852, "perplexity": 69.45339131570684, "lr": 0.001305, "grad_norm": 0.352515, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:48.378222+00:00", "epoch": 0, "step": 739, "train_loss": 4.235136032104492, "perplexity": 69.071073978029, "lr": 0.001305, "grad_norm": 0.336684, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:48.685400+00:00", "epoch": 0, "step": 740, "train_loss": 4.33643913269043, "perplexity": 76.43487970955279, "lr": 0.001305, "grad_norm": 0.349878, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:48.992723+00:00", "epoch": 0, "step": 741, "train_loss": 4.215754985809326, "perplexity": 67.745293303353, "lr": 0.001305, "grad_norm": 0.33275, "tokens_per_sec": 106564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:49.300706+00:00", "epoch": 0, "step": 742, "train_loss": 4.230221271514893, "perplexity": 68.73243902226794, "lr": 0.001305, "grad_norm": 0.333652, "tokens_per_sec": 106396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:49.609221+00:00", "epoch": 0, "step": 743, "train_loss": 4.256306171417236, "perplexity": 70.54890599798637, "lr": 0.001305, "grad_norm": 0.336396, "tokens_per_sec": 106213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:49.915697+00:00", "epoch": 0, "step": 744, "train_loss": 4.1041107177734375, "perplexity": 60.588840004438346, "lr": 0.001305, "grad_norm": 0.343012, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:50.223983+00:00", "epoch": 0, "step": 745, "train_loss": 4.2469048500061035, "perplexity": 69.88876103719961, "lr": 0.001305, "grad_norm": 0.335288, "tokens_per_sec": 106292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:50.531176+00:00", "epoch": 0, "step": 746, "train_loss": 4.207204341888428, "perplexity": 67.16849692750799, "lr": 0.001305, "grad_norm": 0.332137, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:50.838739+00:00", "epoch": 0, "step": 747, "train_loss": 4.2119293212890625, "perplexity": 67.48661765696055, "lr": 0.001305, "grad_norm": 0.326166, "tokens_per_sec": 106541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:51.146309+00:00", "epoch": 0, "step": 748, "train_loss": 4.251741409301758, "perplexity": 70.22760092313561, "lr": 0.001305, "grad_norm": 0.334327, "tokens_per_sec": 106538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:51.454927+00:00", "epoch": 0, "step": 749, "train_loss": 4.2719292640686035, "perplexity": 71.65975293049001, "lr": 0.001305, "grad_norm": 0.41238, "tokens_per_sec": 106177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:51.762697+00:00", "epoch": 0, "step": 750, "train_loss": 4.217758655548096, "perplexity": 67.88116857661753, "lr": 0.001305, "grad_norm": 0.462723, "tokens_per_sec": 106524} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:47:52.740082+00:00", "step": 750, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0000_step_0000750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:53.727988+00:00", "epoch": 0, "step": 751, "train_loss": 4.156276702880859, "perplexity": 63.83340884363246, "lr": 0.001305, "grad_norm": 0.495934, "tokens_per_sec": 16672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:54.034016+00:00", "epoch": 0, "step": 752, "train_loss": 4.2690653800964355, "perplexity": 71.45482130289257, "lr": 0.001305, "grad_norm": 0.51675, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:54.340960+00:00", "epoch": 0, "step": 753, "train_loss": 4.14788818359375, "perplexity": 63.300180686538994, "lr": 0.001305, "grad_norm": 0.551108, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:54.648867+00:00", "epoch": 0, "step": 754, "train_loss": 4.242062091827393, "perplexity": 69.5511248730963, "lr": 0.001305, "grad_norm": 0.483619, "tokens_per_sec": 106423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:54.955643+00:00", "epoch": 0, "step": 755, "train_loss": 4.196455955505371, "perplexity": 66.45041002076444, "lr": 0.001305, "grad_norm": 0.439046, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:55.262019+00:00", "epoch": 0, "step": 756, "train_loss": 4.247844219207764, "perplexity": 69.95444323194566, "lr": 0.001305, "grad_norm": 0.449913, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:55.568697+00:00", "epoch": 0, "step": 757, "train_loss": 4.255126953125, "perplexity": 70.46576246935584, "lr": 0.001305, "grad_norm": 0.410878, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:55.878167+00:00", "epoch": 0, "step": 758, "train_loss": 4.279656887054443, "perplexity": 72.21565762898669, "lr": 0.001305, "grad_norm": 0.371917, "tokens_per_sec": 105883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:56.186689+00:00", "epoch": 0, "step": 759, "train_loss": 4.181313991546631, "perplexity": 65.4517998467798, "lr": 0.001305, "grad_norm": 0.349582, "tokens_per_sec": 106227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:56.493683+00:00", "epoch": 0, "step": 760, "train_loss": 4.124716758728027, "perplexity": 61.85028821483898, "lr": 0.001305, "grad_norm": 0.274857, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:56.799599+00:00", "epoch": 0, "step": 761, "train_loss": 4.143040657043457, "perplexity": 62.99407390980733, "lr": 0.001305, "grad_norm": 0.326277, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:57.105956+00:00", "epoch": 0, "step": 762, "train_loss": 4.198065280914307, "perplexity": 66.55743645112271, "lr": 0.001305, "grad_norm": 0.302722, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:57.414001+00:00", "epoch": 0, "step": 763, "train_loss": 4.123015403747559, "perplexity": 61.745148384389886, "lr": 0.001305, "grad_norm": 0.3268, "tokens_per_sec": 106374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:57.722480+00:00", "epoch": 0, "step": 764, "train_loss": 4.175782680511475, "perplexity": 65.09076500252556, "lr": 0.001305, "grad_norm": 0.300183, "tokens_per_sec": 106224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:58.692084+00:00", "epoch": 0, "step": 765, "train_loss": 4.217983245849609, "perplexity": 67.89641574085375, "lr": 0.001305, "grad_norm": 0.285753, "tokens_per_sec": 33795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:58.998066+00:00", "epoch": 0, "step": 766, "train_loss": 4.1494245529174805, "perplexity": 63.39750788845157, "lr": 0.001305, "grad_norm": 0.269687, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:59.305937+00:00", "epoch": 0, "step": 767, "train_loss": 4.209436893463135, "perplexity": 67.31862157925266, "lr": 0.001305, "grad_norm": 0.274196, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:59.613176+00:00", "epoch": 0, "step": 768, "train_loss": 4.260558128356934, "perplexity": 70.8495155449606, "lr": 0.001305, "grad_norm": 0.265066, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:47:59.920783+00:00", "epoch": 0, "step": 769, "train_loss": 4.268294811248779, "perplexity": 71.39978165222524, "lr": 0.001305, "grad_norm": 0.243828, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:00.227844+00:00", "epoch": 0, "step": 770, "train_loss": 4.364726543426514, "perplexity": 78.62789568673604, "lr": 0.001305, "grad_norm": 0.25196, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:00.534400+00:00", "epoch": 0, "step": 771, "train_loss": 4.15635347366333, "perplexity": 63.83830957249115, "lr": 0.001305, "grad_norm": 0.260522, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:00.841481+00:00", "epoch": 0, "step": 772, "train_loss": 4.080338001251221, "perplexity": 59.165464471608786, "lr": 0.001305, "grad_norm": 0.246795, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:01.148687+00:00", "epoch": 0, "step": 773, "train_loss": 4.25658655166626, "perplexity": 70.56868929111113, "lr": 0.001305, "grad_norm": 0.265264, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:01.456425+00:00", "epoch": 0, "step": 774, "train_loss": 4.138047218322754, "perplexity": 62.68030091705598, "lr": 0.001305, "grad_norm": 0.272042, "tokens_per_sec": 106481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:01.763906+00:00", "epoch": 0, "step": 775, "train_loss": 4.078280448913574, "perplexity": 59.04385358516265, "lr": 0.001305, "grad_norm": 0.283746, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:02.070133+00:00", "epoch": 0, "step": 776, "train_loss": 4.3144354820251465, "perplexity": 74.77140172732517, "lr": 0.001305, "grad_norm": 0.266929, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:02.377591+00:00", "epoch": 0, "step": 777, "train_loss": 4.231217384338379, "perplexity": 68.80093839705634, "lr": 0.001305, "grad_norm": 0.24849, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:02.684707+00:00", "epoch": 0, "step": 778, "train_loss": 4.209725856781006, "perplexity": 67.33807700231515, "lr": 0.001305, "grad_norm": 0.269861, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:02.991860+00:00", "epoch": 0, "step": 779, "train_loss": 4.14448881149292, "perplexity": 63.08536514417894, "lr": 0.001305, "grad_norm": 0.26495, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:03.299130+00:00", "epoch": 0, "step": 780, "train_loss": 4.269397258758545, "perplexity": 71.47853956896334, "lr": 0.001305, "grad_norm": 0.273883, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:03.606051+00:00", "epoch": 0, "step": 781, "train_loss": 4.161509990692139, "perplexity": 64.16834308331157, "lr": 0.001305, "grad_norm": 0.285556, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:03.912220+00:00", "epoch": 0, "step": 782, "train_loss": 4.187073230743408, "perplexity": 65.82983998481878, "lr": 0.001305, "grad_norm": 0.290248, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:04.219532+00:00", "epoch": 0, "step": 783, "train_loss": 4.1804046630859375, "perplexity": 65.39230971450998, "lr": 0.001305, "grad_norm": 0.287201, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:04.526575+00:00", "epoch": 0, "step": 784, "train_loss": 4.164074420928955, "perplexity": 64.33310949825292, "lr": 0.001305, "grad_norm": 0.27042, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:04.833241+00:00", "epoch": 0, "step": 785, "train_loss": 4.188010215759277, "perplexity": 65.89155046486432, "lr": 0.001305, "grad_norm": 0.258435, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:05.140475+00:00", "epoch": 0, "step": 786, "train_loss": 4.160651206970215, "perplexity": 64.11326001041733, "lr": 0.001305, "grad_norm": 0.276587, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:05.446867+00:00", "epoch": 0, "step": 787, "train_loss": 4.224844455718994, "perplexity": 68.36386911264881, "lr": 0.001305, "grad_norm": 0.278095, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:05.754908+00:00", "epoch": 0, "step": 788, "train_loss": 4.2207865715026855, "perplexity": 68.08701854088834, "lr": 0.001305, "grad_norm": 0.280915, "tokens_per_sec": 106377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:06.062168+00:00", "epoch": 0, "step": 789, "train_loss": 4.132242679595947, "perplexity": 62.31752457620956, "lr": 0.001305, "grad_norm": 0.321019, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:06.369087+00:00", "epoch": 0, "step": 790, "train_loss": 4.248896598815918, "perplexity": 70.02810061246738, "lr": 0.001305, "grad_norm": 0.34259, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:06.675955+00:00", "epoch": 0, "step": 791, "train_loss": 4.10701847076416, "perplexity": 60.7652737737632, "lr": 0.001305, "grad_norm": 0.292431, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:06.982278+00:00", "epoch": 0, "step": 792, "train_loss": 4.094176769256592, "perplexity": 59.98993326665661, "lr": 0.001305, "grad_norm": 0.373135, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:07.289472+00:00", "epoch": 0, "step": 793, "train_loss": 4.24267053604126, "perplexity": 69.5934557292696, "lr": 0.001305, "grad_norm": 0.387105, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:07.596034+00:00", "epoch": 0, "step": 794, "train_loss": 4.204184055328369, "perplexity": 66.96593487069049, "lr": 0.001305, "grad_norm": 0.402761, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:07.903720+00:00", "epoch": 0, "step": 795, "train_loss": 4.206002712249756, "perplexity": 67.08783374415667, "lr": 0.001305, "grad_norm": 0.436702, "tokens_per_sec": 106500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:08.210908+00:00", "epoch": 0, "step": 796, "train_loss": 4.219716548919678, "perplexity": 68.01420285760018, "lr": 0.001305, "grad_norm": 0.399825, "tokens_per_sec": 106670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:08.517571+00:00", "epoch": 0, "step": 797, "train_loss": 4.192063331604004, "perplexity": 66.15915850893919, "lr": 0.001305, "grad_norm": 0.353447, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:08.823199+00:00", "epoch": 0, "step": 798, "train_loss": 4.139188289642334, "perplexity": 62.75186443250768, "lr": 0.001305, "grad_norm": 0.314178, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:09.130301+00:00", "epoch": 0, "step": 799, "train_loss": 4.073415756225586, "perplexity": 58.757320894171365, "lr": 0.001305, "grad_norm": 0.281202, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:09.437812+00:00", "epoch": 0, "step": 800, "train_loss": 4.231672286987305, "perplexity": 68.83224324597124, "lr": 0.001305, "grad_norm": 0.309311, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:48:15.139216+00:00", "step": 800, "epoch": 0, "val_loss": 4.119911205768585, "val_ppl": 61.553776401497316, "eval_train_loss": 4.231672286987305, "eval_train_ppl": 68.83224324597124} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:48:15.979356+00:00", "step": 800, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_4p1199_epoch_0000_step_0000800.pt", "val_loss": 4.119911205768585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:16.982795+00:00", "epoch": 0, "step": 801, "train_loss": 4.217668533325195, "perplexity": 67.87505125046998, "lr": 0.001305, "grad_norm": 0.304455, "tokens_per_sec": 4343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:17.287585+00:00", "epoch": 0, "step": 802, "train_loss": 4.139383792877197, "perplexity": 62.764133824311756, "lr": 0.001305, "grad_norm": 0.284973, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:17.593421+00:00", "epoch": 0, "step": 803, "train_loss": 4.193818092346191, "perplexity": 66.2753539207848, "lr": 0.001305, "grad_norm": 0.272515, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:17.898217+00:00", "epoch": 0, "step": 804, "train_loss": 4.232497692108154, "perplexity": 68.88908118596245, "lr": 0.001305, "grad_norm": 0.295597, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:18.203481+00:00", "epoch": 0, "step": 805, "train_loss": 4.145386219024658, "perplexity": 63.142003836195066, "lr": 0.001305, "grad_norm": 0.299733, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:18.508502+00:00", "epoch": 0, "step": 806, "train_loss": 4.20273494720459, "perplexity": 66.86896426786622, "lr": 0.001305, "grad_norm": 0.294984, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:18.816488+00:00", "epoch": 0, "step": 807, "train_loss": 4.27313232421875, "perplexity": 71.74601580291774, "lr": 0.001305, "grad_norm": 0.297911, "tokens_per_sec": 106396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:19.121526+00:00", "epoch": 0, "step": 808, "train_loss": 4.166321277618408, "perplexity": 64.4778192858806, "lr": 0.001305, "grad_norm": 0.327648, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:19.428164+00:00", "epoch": 0, "step": 809, "train_loss": 3.9949328899383545, "perplexity": 54.32219493563222, "lr": 0.001305, "grad_norm": 0.268067, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:19.733967+00:00", "epoch": 0, "step": 810, "train_loss": 4.137203216552734, "perplexity": 62.62742095066828, "lr": 0.001305, "grad_norm": 0.264783, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:20.040390+00:00", "epoch": 0, "step": 811, "train_loss": 4.006044864654541, "perplexity": 54.9291879922915, "lr": 0.001305, "grad_norm": 0.273046, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:20.348473+00:00", "epoch": 0, "step": 812, "train_loss": 4.1435699462890625, "perplexity": 63.027424821025384, "lr": 0.001305, "grad_norm": 0.264996, "tokens_per_sec": 106362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:20.654928+00:00", "epoch": 0, "step": 813, "train_loss": 4.159383296966553, "perplexity": 64.03202167901247, "lr": 0.001305, "grad_norm": 0.258126, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:20.961659+00:00", "epoch": 0, "step": 814, "train_loss": 4.22075080871582, "perplexity": 68.0845836028963, "lr": 0.001305, "grad_norm": 0.29583, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:21.272756+00:00", "epoch": 0, "step": 815, "train_loss": 4.148894309997559, "perplexity": 63.36390071952213, "lr": 0.001305, "grad_norm": 0.350713, "tokens_per_sec": 105331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:21.578950+00:00", "epoch": 0, "step": 816, "train_loss": 4.203671455383301, "perplexity": 66.93161693259236, "lr": 0.001305, "grad_norm": 0.376455, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:21.885821+00:00", "epoch": 0, "step": 817, "train_loss": 4.182326793670654, "perplexity": 65.5181231492072, "lr": 0.001305, "grad_norm": 0.386526, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:22.192697+00:00", "epoch": 0, "step": 818, "train_loss": 4.257269382476807, "perplexity": 70.6168922217692, "lr": 0.001305, "grad_norm": 0.43631, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:22.498962+00:00", "epoch": 0, "step": 819, "train_loss": 4.10830545425415, "perplexity": 60.843528023036974, "lr": 0.001305, "grad_norm": 0.544125, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:22.805971+00:00", "epoch": 0, "step": 820, "train_loss": 4.194492816925049, "perplexity": 66.32008662047654, "lr": 0.001305, "grad_norm": 0.534919, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:23.112440+00:00", "epoch": 0, "step": 821, "train_loss": 4.089612007141113, "perplexity": 59.71671754876745, "lr": 0.001305, "grad_norm": 0.485143, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:23.418096+00:00", "epoch": 0, "step": 822, "train_loss": 4.191339492797852, "perplexity": 66.11128727024831, "lr": 0.001305, "grad_norm": 0.424833, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:23.724823+00:00", "epoch": 0, "step": 823, "train_loss": 4.129315376281738, "perplexity": 62.13536902233841, "lr": 0.001305, "grad_norm": 0.410327, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:24.031453+00:00", "epoch": 0, "step": 824, "train_loss": 4.243502140045166, "perplexity": 69.65135399657851, "lr": 0.001305, "grad_norm": 0.342377, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:24.338303+00:00", "epoch": 0, "step": 825, "train_loss": 4.1021013259887695, "perplexity": 60.46721552363028, "lr": 0.001305, "grad_norm": 0.337199, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:24.644668+00:00", "epoch": 0, "step": 826, "train_loss": 4.047000885009766, "perplexity": 57.225573349940454, "lr": 0.001305, "grad_norm": 0.304193, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:24.950686+00:00", "epoch": 0, "step": 827, "train_loss": 4.177567481994629, "perplexity": 65.20704283199589, "lr": 0.001305, "grad_norm": 0.321129, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:25.257599+00:00", "epoch": 0, "step": 828, "train_loss": 4.1609015464782715, "perplexity": 64.1293121015407, "lr": 0.001305, "grad_norm": 0.30363, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:25.564480+00:00", "epoch": 0, "step": 829, "train_loss": 4.186491012573242, "perplexity": 65.79152381110912, "lr": 0.001305, "grad_norm": 0.275087, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:25.871408+00:00", "epoch": 0, "step": 830, "train_loss": 4.150505065917969, "perplexity": 63.46604674181634, "lr": 0.001305, "grad_norm": 0.266332, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:26.178498+00:00", "epoch": 0, "step": 831, "train_loss": 4.128983020782471, "perplexity": 62.11472142210662, "lr": 0.001305, "grad_norm": 0.27567, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:26.484290+00:00", "epoch": 0, "step": 832, "train_loss": 4.088652610778809, "perplexity": 59.659453021264454, "lr": 0.001305, "grad_norm": 0.239187, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:26.790110+00:00", "epoch": 0, "step": 833, "train_loss": 4.261098861694336, "perplexity": 70.88783659975675, "lr": 0.001305, "grad_norm": 0.251595, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:27.097966+00:00", "epoch": 0, "step": 834, "train_loss": 4.255890369415283, "perplexity": 70.51957771943782, "lr": 0.001305, "grad_norm": 0.242451, "tokens_per_sec": 106440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:27.405574+00:00", "epoch": 0, "step": 835, "train_loss": 4.032018661499023, "perplexity": 56.37459769118021, "lr": 0.001305, "grad_norm": 0.279674, "tokens_per_sec": 106526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:27.713372+00:00", "epoch": 0, "step": 836, "train_loss": 4.149231433868408, "perplexity": 63.38526580414249, "lr": 0.001305, "grad_norm": 0.301639, "tokens_per_sec": 106459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:28.020273+00:00", "epoch": 0, "step": 837, "train_loss": 4.2102532386779785, "perplexity": 67.37359925117528, "lr": 0.001305, "grad_norm": 0.312, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:28.325502+00:00", "epoch": 0, "step": 838, "train_loss": 4.204361438751221, "perplexity": 66.97781457103211, "lr": 0.001305, "grad_norm": 0.321968, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:28.631655+00:00", "epoch": 0, "step": 839, "train_loss": 4.195280075073242, "perplexity": 66.37231820623049, "lr": 0.001305, "grad_norm": 0.328527, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:28.940149+00:00", "epoch": 0, "step": 840, "train_loss": 4.237844944000244, "perplexity": 69.25843508977817, "lr": 0.001305, "grad_norm": 0.381488, "tokens_per_sec": 106279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:29.248217+00:00", "epoch": 0, "step": 841, "train_loss": 4.105917930603027, "perplexity": 60.69843593518186, "lr": 0.001305, "grad_norm": 0.359603, "tokens_per_sec": 106306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:29.555202+00:00", "epoch": 0, "step": 842, "train_loss": 4.141045570373535, "perplexity": 62.86852055921959, "lr": 0.001305, "grad_norm": 0.295751, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:29.861511+00:00", "epoch": 0, "step": 843, "train_loss": 4.206808090209961, "perplexity": 67.14188657040788, "lr": 0.001305, "grad_norm": 0.315044, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:30.168026+00:00", "epoch": 0, "step": 844, "train_loss": 4.111307144165039, "perplexity": 61.02643580613476, "lr": 0.001305, "grad_norm": 0.296622, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:30.475423+00:00", "epoch": 0, "step": 845, "train_loss": 4.245942115783691, "perplexity": 69.82150911324588, "lr": 0.001305, "grad_norm": 0.313477, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:30.782944+00:00", "epoch": 0, "step": 846, "train_loss": 4.094176769256592, "perplexity": 59.98993326665661, "lr": 0.001305, "grad_norm": 0.28115, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:31.090696+00:00", "epoch": 0, "step": 847, "train_loss": 4.102423191070557, "perplexity": 60.486680941351324, "lr": 0.001305, "grad_norm": 0.281181, "tokens_per_sec": 106476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:31.396752+00:00", "epoch": 0, "step": 848, "train_loss": 4.235560417175293, "perplexity": 69.1003929314725, "lr": 0.001305, "grad_norm": 0.264065, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:31.702510+00:00", "epoch": 0, "step": 849, "train_loss": 4.235175132751465, "perplexity": 69.07377475450936, "lr": 0.001305, "grad_norm": 0.283143, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:32.009129+00:00", "epoch": 0, "step": 850, "train_loss": 4.152897357940674, "perplexity": 63.618057814089816, "lr": 0.001305, "grad_norm": 0.290162, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:32.316176+00:00", "epoch": 0, "step": 851, "train_loss": 4.199290752410889, "perplexity": 66.63905069012705, "lr": 0.001305, "grad_norm": 0.27315, "tokens_per_sec": 106652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:32.623994+00:00", "epoch": 0, "step": 852, "train_loss": 4.054316520690918, "perplexity": 57.64574985138295, "lr": 0.001305, "grad_norm": 0.263272, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:32.932077+00:00", "epoch": 0, "step": 853, "train_loss": 4.043288707733154, "perplexity": 57.01353568111222, "lr": 0.001305, "grad_norm": 0.252769, "tokens_per_sec": 106362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:33.238364+00:00", "epoch": 0, "step": 854, "train_loss": 4.050142288208008, "perplexity": 57.40562460781353, "lr": 0.001305, "grad_norm": 0.26898, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:33.544214+00:00", "epoch": 0, "step": 855, "train_loss": 4.0295023918151855, "perplexity": 56.232922321542915, "lr": 0.001305, "grad_norm": 0.270184, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:33.851373+00:00", "epoch": 0, "step": 856, "train_loss": 4.208032131195068, "perplexity": 67.22412131046607, "lr": 0.001305, "grad_norm": 0.28833, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:34.157452+00:00", "epoch": 0, "step": 857, "train_loss": 4.101743698120117, "perplexity": 60.44559462856697, "lr": 0.001305, "grad_norm": 0.292081, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:34.464573+00:00", "epoch": 0, "step": 858, "train_loss": 4.067643165588379, "perplexity": 58.4191160319814, "lr": 0.001305, "grad_norm": 0.326488, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:34.770885+00:00", "epoch": 0, "step": 859, "train_loss": 4.095272064208984, "perplexity": 60.05567593499187, "lr": 0.001305, "grad_norm": 0.316598, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:35.076796+00:00", "epoch": 0, "step": 860, "train_loss": 4.126391887664795, "perplexity": 61.953982248565694, "lr": 0.001305, "grad_norm": 0.313929, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:35.382456+00:00", "epoch": 0, "step": 861, "train_loss": 4.102557182312012, "perplexity": 60.494786169824835, "lr": 0.001305, "grad_norm": 0.301696, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:35.689167+00:00", "epoch": 0, "step": 862, "train_loss": 4.04996919631958, "perplexity": 57.39568901975344, "lr": 0.001305, "grad_norm": 0.307777, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:35.996833+00:00", "epoch": 0, "step": 863, "train_loss": 4.155702114105225, "perplexity": 63.79674141875461, "lr": 0.001305, "grad_norm": 0.303971, "tokens_per_sec": 106505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:36.303457+00:00", "epoch": 0, "step": 864, "train_loss": 4.13775634765625, "perplexity": 62.66207170745127, "lr": 0.001305, "grad_norm": 0.35387, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:36.609371+00:00", "epoch": 0, "step": 865, "train_loss": 4.051745891571045, "perplexity": 57.49775431049943, "lr": 0.001305, "grad_norm": 0.352032, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:36.915330+00:00", "epoch": 0, "step": 866, "train_loss": 4.088910102844238, "perplexity": 59.67481683499049, "lr": 0.001305, "grad_norm": 0.326298, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:37.221846+00:00", "epoch": 0, "step": 867, "train_loss": 4.0647125244140625, "perplexity": 58.24816119117465, "lr": 0.001305, "grad_norm": 0.307819, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:37.529386+00:00", "epoch": 0, "step": 868, "train_loss": 4.093048095703125, "perplexity": 59.92226241184022, "lr": 0.001305, "grad_norm": 0.312419, "tokens_per_sec": 106549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:37.836745+00:00", "epoch": 0, "step": 869, "train_loss": 4.1323041915893555, "perplexity": 62.32135796926912, "lr": 0.001305, "grad_norm": 0.291149, "tokens_per_sec": 106611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:38.143772+00:00", "epoch": 0, "step": 870, "train_loss": 4.117459774017334, "perplexity": 61.40306632282787, "lr": 0.001305, "grad_norm": 0.306207, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:38.450373+00:00", "epoch": 0, "step": 871, "train_loss": 4.067562580108643, "perplexity": 58.41440848917254, "lr": 0.001305, "grad_norm": 0.302825, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:38.756888+00:00", "epoch": 0, "step": 872, "train_loss": 4.137299060821533, "perplexity": 62.633423717696864, "lr": 0.001305, "grad_norm": 0.29413, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:39.063779+00:00", "epoch": 0, "step": 873, "train_loss": 4.111248016357422, "perplexity": 61.022827553453936, "lr": 0.001305, "grad_norm": 0.280937, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:39.370915+00:00", "epoch": 0, "step": 874, "train_loss": 3.9925155639648438, "perplexity": 54.19103906998356, "lr": 0.001305, "grad_norm": 0.280059, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:39.678016+00:00", "epoch": 0, "step": 875, "train_loss": 4.020052433013916, "perplexity": 55.70402648022123, "lr": 0.001305, "grad_norm": 0.281591, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:39.984611+00:00", "epoch": 0, "step": 876, "train_loss": 4.192920684814453, "perplexity": 66.21590459813235, "lr": 0.001305, "grad_norm": 0.278519, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:40.290645+00:00", "epoch": 0, "step": 877, "train_loss": 4.2408576011657715, "perplexity": 69.46740162473894, "lr": 0.001305, "grad_norm": 0.266083, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:40.596598+00:00", "epoch": 0, "step": 878, "train_loss": 4.166318893432617, "perplexity": 64.47766555896328, "lr": 0.001305, "grad_norm": 0.275831, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:40.903776+00:00", "epoch": 0, "step": 879, "train_loss": 4.202420234680176, "perplexity": 66.84792307845312, "lr": 0.001305, "grad_norm": 0.290906, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:41.211674+00:00", "epoch": 0, "step": 880, "train_loss": 4.082845687866211, "perplexity": 59.31401910131346, "lr": 0.001305, "grad_norm": 0.284789, "tokens_per_sec": 106484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:41.519374+00:00", "epoch": 0, "step": 881, "train_loss": 3.938641309738159, "perplexity": 51.34878678655258, "lr": 0.001305, "grad_norm": 0.230652, "tokens_per_sec": 106434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:41.825808+00:00", "epoch": 0, "step": 882, "train_loss": 4.094690322875977, "perplexity": 60.02074922615884, "lr": 0.001305, "grad_norm": 0.248598, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:42.133447+00:00", "epoch": 0, "step": 883, "train_loss": 4.11872673034668, "perplexity": 61.480910628601926, "lr": 0.001305, "grad_norm": 0.270875, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:42.439664+00:00", "epoch": 0, "step": 884, "train_loss": 4.090692520141602, "perplexity": 59.7812771108673, "lr": 0.001305, "grad_norm": 0.264259, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:42.746190+00:00", "epoch": 0, "step": 885, "train_loss": 4.07682466506958, "perplexity": 58.95796103269071, "lr": 0.001305, "grad_norm": 0.253471, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:43.053337+00:00", "epoch": 0, "step": 886, "train_loss": 4.00441837310791, "perplexity": 54.83991874987279, "lr": 0.001305, "grad_norm": 0.294242, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:43.360514+00:00", "epoch": 0, "step": 887, "train_loss": 4.1065673828125, "perplexity": 60.73786947222433, "lr": 0.001305, "grad_norm": 0.317486, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:43.666997+00:00", "epoch": 0, "step": 888, "train_loss": 4.0952863693237305, "perplexity": 60.05653504447209, "lr": 0.001305, "grad_norm": 0.28975, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:43.974118+00:00", "epoch": 0, "step": 889, "train_loss": 4.152863502502441, "perplexity": 63.615904033321826, "lr": 0.001305, "grad_norm": 0.298581, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:44.279988+00:00", "epoch": 0, "step": 890, "train_loss": 4.031144618988037, "perplexity": 56.32534542368167, "lr": 0.001305, "grad_norm": 0.344908, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:44.586491+00:00", "epoch": 0, "step": 891, "train_loss": 4.152022838592529, "perplexity": 63.56244691158932, "lr": 0.001305, "grad_norm": 0.384873, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:44.892983+00:00", "epoch": 0, "step": 892, "train_loss": 4.071722984313965, "perplexity": 58.657942288146636, "lr": 0.001305, "grad_norm": 0.371737, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:45.200137+00:00", "epoch": 0, "step": 893, "train_loss": 4.097768783569336, "perplexity": 60.20580544143305, "lr": 0.001305, "grad_norm": 0.318059, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:45.506459+00:00", "epoch": 0, "step": 894, "train_loss": 4.108514785766602, "perplexity": 60.856265823945954, "lr": 0.001305, "grad_norm": 0.381865, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:45.812340+00:00", "epoch": 0, "step": 895, "train_loss": 4.100915908813477, "perplexity": 60.395579115709296, "lr": 0.001305, "grad_norm": 0.374329, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:46.118506+00:00", "epoch": 0, "step": 896, "train_loss": 4.218764305114746, "perplexity": 67.94946758105635, "lr": 0.001305, "grad_norm": 0.314496, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:46.426168+00:00", "epoch": 0, "step": 897, "train_loss": 4.157021999359131, "perplexity": 63.88100139151584, "lr": 0.001305, "grad_norm": 0.327666, "tokens_per_sec": 106507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:46.732893+00:00", "epoch": 0, "step": 898, "train_loss": 4.15403938293457, "perplexity": 63.690752727980104, "lr": 0.001305, "grad_norm": 0.295295, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:47.040203+00:00", "epoch": 0, "step": 899, "train_loss": 4.0106520652771, "perplexity": 55.18284164937246, "lr": 0.001305, "grad_norm": 0.26443, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:47.347543+00:00", "epoch": 0, "step": 900, "train_loss": 4.085553169250488, "perplexity": 59.47482829957859, "lr": 0.001305, "grad_norm": 0.289804, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:48:53.018829+00:00", "step": 900, "epoch": 0, "val_loss": 4.040788054466248, "val_ppl": 56.87114270878083, "eval_train_loss": 4.085553169250488, "eval_train_ppl": 59.47482829957859} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:48:53.817249+00:00", "step": 900, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_4p0408_epoch_0000_step_0000900.pt", "val_loss": 4.040788054466248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:54.783864+00:00", "epoch": 0, "step": 901, "train_loss": 4.007937431335449, "perplexity": 55.033243578303654, "lr": 0.001305, "grad_norm": 0.272976, "tokens_per_sec": 4406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:55.090455+00:00", "epoch": 0, "step": 902, "train_loss": 4.006047248840332, "perplexity": 54.92931895383715, "lr": 0.001305, "grad_norm": 0.256868, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:55.395853+00:00", "epoch": 0, "step": 903, "train_loss": 4.054203033447266, "perplexity": 57.639208165330025, "lr": 0.001305, "grad_norm": 0.26142, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:55.701534+00:00", "epoch": 0, "step": 904, "train_loss": 4.0910844802856445, "perplexity": 59.80471358163593, "lr": 0.001305, "grad_norm": 0.276399, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:56.007027+00:00", "epoch": 0, "step": 905, "train_loss": 4.157680511474609, "perplexity": 63.92308165854325, "lr": 0.001305, "grad_norm": 0.262317, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:56.313635+00:00", "epoch": 0, "step": 906, "train_loss": 4.014095306396484, "perplexity": 55.373176975985785, "lr": 0.001305, "grad_norm": 0.252685, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:56.620229+00:00", "epoch": 0, "step": 907, "train_loss": 4.123371124267578, "perplexity": 61.76711630767076, "lr": 0.001305, "grad_norm": 0.246493, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:56.925732+00:00", "epoch": 0, "step": 908, "train_loss": 4.074232578277588, "perplexity": 58.80533477632358, "lr": 0.001305, "grad_norm": 0.245893, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:57.230909+00:00", "epoch": 0, "step": 909, "train_loss": 4.068615913391113, "perplexity": 58.47597074690553, "lr": 0.001305, "grad_norm": 0.253266, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:57.536414+00:00", "epoch": 0, "step": 910, "train_loss": 4.065176010131836, "perplexity": 58.2751646393462, "lr": 0.001305, "grad_norm": 0.24602, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:57.842586+00:00", "epoch": 0, "step": 911, "train_loss": 4.146656513214111, "perplexity": 63.22226372281983, "lr": 0.001305, "grad_norm": 0.238531, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:58.149544+00:00", "epoch": 0, "step": 912, "train_loss": 4.110022068023682, "perplexity": 60.94806255806942, "lr": 0.001305, "grad_norm": 0.251354, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:58.455816+00:00", "epoch": 0, "step": 913, "train_loss": 4.062336444854736, "perplexity": 58.10992322324305, "lr": 0.001305, "grad_norm": 0.298969, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:58.761729+00:00", "epoch": 0, "step": 914, "train_loss": 4.080039978027344, "perplexity": 59.147834416358215, "lr": 0.001305, "grad_norm": 0.309661, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:59.068314+00:00", "epoch": 0, "step": 915, "train_loss": 4.094246864318848, "perplexity": 59.9941384121419, "lr": 0.001305, "grad_norm": 0.322373, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:59.374980+00:00", "epoch": 0, "step": 916, "train_loss": 4.016757965087891, "perplexity": 55.52081331229329, "lr": 0.001305, "grad_norm": 0.358799, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:59.683183+00:00", "epoch": 0, "step": 917, "train_loss": 4.060696125030518, "perplexity": 58.01468249817579, "lr": 0.001305, "grad_norm": 0.402166, "tokens_per_sec": 106323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:48:59.990197+00:00", "epoch": 0, "step": 918, "train_loss": 4.0666399002075195, "perplexity": 58.360535546086496, "lr": 0.001305, "grad_norm": 0.430376, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:00.297063+00:00", "epoch": 0, "step": 919, "train_loss": 4.163175582885742, "perplexity": 64.27531043188894, "lr": 0.001305, "grad_norm": 0.420314, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:00.603041+00:00", "epoch": 0, "step": 920, "train_loss": 4.1638336181640625, "perplexity": 64.31761977267084, "lr": 0.001305, "grad_norm": 0.376889, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:00.910157+00:00", "epoch": 0, "step": 921, "train_loss": 4.017813682556152, "perplexity": 55.579458555716265, "lr": 0.001305, "grad_norm": 0.284242, "tokens_per_sec": 106641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:01.216184+00:00", "epoch": 0, "step": 922, "train_loss": 4.019052505493164, "perplexity": 55.6483543298233, "lr": 0.001305, "grad_norm": 0.307114, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:01.523233+00:00", "epoch": 0, "step": 923, "train_loss": 4.170483589172363, "perplexity": 64.74675536767559, "lr": 0.001305, "grad_norm": 0.323721, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:01.828676+00:00", "epoch": 0, "step": 924, "train_loss": 4.131250381469727, "perplexity": 62.255717683845155, "lr": 0.001305, "grad_norm": 0.317346, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:02.135545+00:00", "epoch": 0, "step": 925, "train_loss": 4.140369892120361, "perplexity": 62.82605601486866, "lr": 0.001305, "grad_norm": 0.315671, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:02.443045+00:00", "epoch": 0, "step": 926, "train_loss": 4.03220272064209, "perplexity": 56.38497490630333, "lr": 0.001305, "grad_norm": 0.291477, "tokens_per_sec": 106562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:02.751368+00:00", "epoch": 0, "step": 927, "train_loss": 4.131292819976807, "perplexity": 62.25835977962376, "lr": 0.001305, "grad_norm": 0.277564, "tokens_per_sec": 106279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:03.059381+00:00", "epoch": 0, "step": 928, "train_loss": 4.124638557434082, "perplexity": 61.8454516313857, "lr": 0.001305, "grad_norm": 0.289914, "tokens_per_sec": 106387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:03.365730+00:00", "epoch": 0, "step": 929, "train_loss": 4.053549289703369, "perplexity": 57.60153920785302, "lr": 0.001305, "grad_norm": 0.267315, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:03.672333+00:00", "epoch": 0, "step": 930, "train_loss": 4.007662296295166, "perplexity": 55.018104087412546, "lr": 0.001305, "grad_norm": 0.263564, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:03.979936+00:00", "epoch": 0, "step": 931, "train_loss": 4.181377410888672, "perplexity": 65.45595088848826, "lr": 0.001305, "grad_norm": 0.253086, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:04.287791+00:00", "epoch": 0, "step": 932, "train_loss": 4.021162509918213, "perplexity": 55.76589656741731, "lr": 0.001305, "grad_norm": 0.27465, "tokens_per_sec": 106440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:04.594989+00:00", "epoch": 0, "step": 933, "train_loss": 4.128565311431885, "perplexity": 62.08878094032764, "lr": 0.001305, "grad_norm": 0.286585, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:04.901664+00:00", "epoch": 0, "step": 934, "train_loss": 4.048813343048096, "perplexity": 57.329386350291244, "lr": 0.001305, "grad_norm": 0.286156, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:05.208289+00:00", "epoch": 0, "step": 935, "train_loss": 4.113967418670654, "perplexity": 61.18899901298329, "lr": 0.001305, "grad_norm": 0.290818, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:05.515217+00:00", "epoch": 0, "step": 936, "train_loss": 4.20292854309082, "perplexity": 66.88191107744908, "lr": 0.001305, "grad_norm": 0.276818, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:05.822218+00:00", "epoch": 0, "step": 937, "train_loss": 4.088744640350342, "perplexity": 59.664943707812796, "lr": 0.001305, "grad_norm": 0.285841, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:06.130052+00:00", "epoch": 0, "step": 938, "train_loss": 4.005232334136963, "perplexity": 54.88457447811737, "lr": 0.001305, "grad_norm": 0.292768, "tokens_per_sec": 106447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:06.436420+00:00", "epoch": 0, "step": 939, "train_loss": 4.097341060638428, "perplexity": 60.18005954431975, "lr": 0.001305, "grad_norm": 0.28723, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:06.743359+00:00", "epoch": 0, "step": 940, "train_loss": 4.096302509307861, "perplexity": 60.11759190694512, "lr": 0.001305, "grad_norm": 0.292493, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:07.050401+00:00", "epoch": 0, "step": 941, "train_loss": 4.017369270324707, "perplexity": 55.55476385223757, "lr": 0.001305, "grad_norm": 0.319457, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:07.359214+00:00", "epoch": 0, "step": 942, "train_loss": 4.04896879196167, "perplexity": 57.338298833815266, "lr": 0.001305, "grad_norm": 0.28734, "tokens_per_sec": 106109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:07.666369+00:00", "epoch": 0, "step": 943, "train_loss": 3.9244415760040283, "perplexity": 50.624800060871905, "lr": 0.001305, "grad_norm": 0.297579, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:07.972894+00:00", "epoch": 0, "step": 944, "train_loss": 4.107227802276611, "perplexity": 60.77799519187716, "lr": 0.001305, "grad_norm": 0.296569, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:08.279600+00:00", "epoch": 0, "step": 945, "train_loss": 4.108156204223633, "perplexity": 60.834447802251205, "lr": 0.001305, "grad_norm": 0.291428, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:08.586741+00:00", "epoch": 0, "step": 946, "train_loss": 4.109009742736816, "perplexity": 60.886394512483875, "lr": 0.001305, "grad_norm": 0.275504, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:08.893679+00:00", "epoch": 0, "step": 947, "train_loss": 4.139857769012451, "perplexity": 62.793889577095655, "lr": 0.001305, "grad_norm": 0.272563, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:09.200979+00:00", "epoch": 0, "step": 948, "train_loss": 4.139309406280518, "perplexity": 62.759465187647166, "lr": 0.001305, "grad_norm": 0.297482, "tokens_per_sec": 106633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:09.508352+00:00", "epoch": 0, "step": 949, "train_loss": 4.0066609382629395, "perplexity": 54.96303884158108, "lr": 0.001305, "grad_norm": 0.305176, "tokens_per_sec": 106608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:09.814498+00:00", "epoch": 0, "step": 950, "train_loss": 4.001777648925781, "perplexity": 54.695292693091666, "lr": 0.001305, "grad_norm": 0.29069, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:10.121231+00:00", "epoch": 0, "step": 951, "train_loss": 4.118349552154541, "perplexity": 61.457725742571554, "lr": 0.001305, "grad_norm": 0.285494, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:10.427973+00:00", "epoch": 0, "step": 952, "train_loss": 3.9263129234313965, "perplexity": 50.71962534808277, "lr": 0.001305, "grad_norm": 0.300835, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:10.736925+00:00", "epoch": 0, "step": 953, "train_loss": 3.915898561477661, "perplexity": 50.19415378640176, "lr": 0.001305, "grad_norm": 0.300391, "tokens_per_sec": 106061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:11.043802+00:00", "epoch": 0, "step": 954, "train_loss": 4.066866397857666, "perplexity": 58.37375556734433, "lr": 0.001305, "grad_norm": 0.350501, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:11.350006+00:00", "epoch": 0, "step": 955, "train_loss": 4.1171040534973145, "perplexity": 61.381227876566726, "lr": 0.001305, "grad_norm": 0.345496, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:11.656181+00:00", "epoch": 0, "step": 956, "train_loss": 4.0186052322387695, "perplexity": 55.62346987477095, "lr": 0.001305, "grad_norm": 0.332937, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:11.962742+00:00", "epoch": 0, "step": 957, "train_loss": 4.009480953216553, "perplexity": 55.11825418494274, "lr": 0.001305, "grad_norm": 0.365768, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:12.270263+00:00", "epoch": 0, "step": 958, "train_loss": 4.115842819213867, "perplexity": 61.30386056701608, "lr": 0.001305, "grad_norm": 0.35058, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:12.577920+00:00", "epoch": 0, "step": 959, "train_loss": 3.98490047454834, "perplexity": 53.779936739033076, "lr": 0.001305, "grad_norm": 0.398895, "tokens_per_sec": 106510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:12.884165+00:00", "epoch": 0, "step": 960, "train_loss": 4.007937908172607, "perplexity": 55.03326982020538, "lr": 0.001305, "grad_norm": 0.390084, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:13.190269+00:00", "epoch": 0, "step": 961, "train_loss": 4.067270278930664, "perplexity": 58.39733638398002, "lr": 0.001305, "grad_norm": 0.364494, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:13.496885+00:00", "epoch": 0, "step": 962, "train_loss": 4.072297096252441, "perplexity": 58.691628181912364, "lr": 0.001305, "grad_norm": 0.385789, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:13.804475+00:00", "epoch": 0, "step": 963, "train_loss": 4.049406051635742, "perplexity": 57.36337604190156, "lr": 0.001305, "grad_norm": 0.361747, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:14.112406+00:00", "epoch": 0, "step": 964, "train_loss": 4.099294185638428, "perplexity": 60.29771358223744, "lr": 0.001305, "grad_norm": 0.291396, "tokens_per_sec": 106414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:14.420423+00:00", "epoch": 0, "step": 965, "train_loss": 3.992286205291748, "perplexity": 54.17861131043049, "lr": 0.001305, "grad_norm": 0.28905, "tokens_per_sec": 106384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:14.727360+00:00", "epoch": 0, "step": 966, "train_loss": 4.11545467376709, "perplexity": 61.28007036999413, "lr": 0.001305, "grad_norm": 0.276869, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:15.033242+00:00", "epoch": 0, "step": 967, "train_loss": 4.091034889221191, "perplexity": 59.801747875766964, "lr": 0.001305, "grad_norm": 0.272283, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:15.339348+00:00", "epoch": 0, "step": 968, "train_loss": 4.062241077423096, "perplexity": 58.104381693357354, "lr": 0.001305, "grad_norm": 0.275515, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:15.646748+00:00", "epoch": 0, "step": 969, "train_loss": 4.049890041351318, "perplexity": 57.39114604561259, "lr": 0.001305, "grad_norm": 0.266396, "tokens_per_sec": 106597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:15.954126+00:00", "epoch": 0, "step": 970, "train_loss": 4.067681312561035, "perplexity": 58.42134458690932, "lr": 0.001305, "grad_norm": 0.248255, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:16.261562+00:00", "epoch": 0, "step": 971, "train_loss": 4.004470348358154, "perplexity": 54.84276914244747, "lr": 0.001305, "grad_norm": 0.237375, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:16.568505+00:00", "epoch": 0, "step": 972, "train_loss": 4.026582717895508, "perplexity": 56.06897997037743, "lr": 0.001305, "grad_norm": 0.256859, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:16.875433+00:00", "epoch": 0, "step": 973, "train_loss": 4.0810933113098145, "perplexity": 59.2101696230455, "lr": 0.001305, "grad_norm": 0.2452, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:17.181761+00:00", "epoch": 0, "step": 974, "train_loss": 4.065805435180664, "perplexity": 58.31185603372478, "lr": 0.001305, "grad_norm": 0.239672, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:17.489105+00:00", "epoch": 0, "step": 975, "train_loss": 4.093668460845947, "perplexity": 59.95944762773289, "lr": 0.001305, "grad_norm": 0.268666, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:17.796005+00:00", "epoch": 0, "step": 976, "train_loss": 4.057061195373535, "perplexity": 57.80418600995571, "lr": 0.001305, "grad_norm": 0.312163, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:18.102442+00:00", "epoch": 0, "step": 977, "train_loss": 4.029116153717041, "perplexity": 56.21120721844235, "lr": 0.001305, "grad_norm": 0.344337, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:18.409007+00:00", "epoch": 0, "step": 978, "train_loss": 4.049776554107666, "perplexity": 57.384633252204296, "lr": 0.001305, "grad_norm": 0.299383, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:18.715839+00:00", "epoch": 0, "step": 979, "train_loss": 4.041460990905762, "perplexity": 56.90942625281758, "lr": 0.001305, "grad_norm": 0.230249, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:19.022236+00:00", "epoch": 0, "step": 980, "train_loss": 4.1012043952941895, "perplexity": 60.41300493721511, "lr": 0.001305, "grad_norm": 0.252761, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:19.330104+00:00", "epoch": 0, "step": 981, "train_loss": 4.073058605194092, "perplexity": 58.73633940340974, "lr": 0.001305, "grad_norm": 0.262563, "tokens_per_sec": 106378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:19.636870+00:00", "epoch": 0, "step": 982, "train_loss": 4.143953323364258, "perplexity": 63.051592723224175, "lr": 0.001305, "grad_norm": 0.29066, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:19.943245+00:00", "epoch": 0, "step": 983, "train_loss": 4.108935356140137, "perplexity": 60.881865549261164, "lr": 0.001305, "grad_norm": 0.272886, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:20.250190+00:00", "epoch": 0, "step": 984, "train_loss": 4.061037063598633, "perplexity": 58.03446531312704, "lr": 0.001305, "grad_norm": 0.273525, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:20.557584+00:00", "epoch": 0, "step": 985, "train_loss": 4.060108661651611, "perplexity": 57.98061100562618, "lr": 0.001305, "grad_norm": 0.280141, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:20.864164+00:00", "epoch": 0, "step": 986, "train_loss": 4.015854835510254, "perplexity": 55.470693459383966, "lr": 0.001305, "grad_norm": 0.250111, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:21.171532+00:00", "epoch": 0, "step": 987, "train_loss": 4.073485374450684, "perplexity": 58.76141161695629, "lr": 0.001305, "grad_norm": 0.291284, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:21.478907+00:00", "epoch": 0, "step": 988, "train_loss": 3.998455286026001, "perplexity": 54.51387661375519, "lr": 0.001305, "grad_norm": 0.294037, "tokens_per_sec": 106607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:21.785773+00:00", "epoch": 0, "step": 989, "train_loss": 4.0847296714782715, "perplexity": 59.4258710718216, "lr": 0.001305, "grad_norm": 0.319364, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:22.092679+00:00", "epoch": 0, "step": 990, "train_loss": 4.041431903839111, "perplexity": 56.90777094861722, "lr": 0.001305, "grad_norm": 0.295196, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:22.398847+00:00", "epoch": 0, "step": 991, "train_loss": 4.028039932250977, "perplexity": 56.150744052315076, "lr": 0.001305, "grad_norm": 0.276917, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:22.706947+00:00", "epoch": 0, "step": 992, "train_loss": 4.1257171630859375, "perplexity": 61.912194473189395, "lr": 0.001305, "grad_norm": 0.283649, "tokens_per_sec": 106359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:23.015159+00:00", "epoch": 0, "step": 993, "train_loss": 3.9891765117645264, "perplexity": 54.010394121034814, "lr": 0.001305, "grad_norm": 0.299996, "tokens_per_sec": 106312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:23.323199+00:00", "epoch": 0, "step": 994, "train_loss": 4.023223876953125, "perplexity": 55.88096911090007, "lr": 0.001305, "grad_norm": 0.285363, "tokens_per_sec": 106377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:23.630338+00:00", "epoch": 0, "step": 995, "train_loss": 4.078592777252197, "perplexity": 59.062297533992826, "lr": 0.001305, "grad_norm": 0.320836, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:23.937558+00:00", "epoch": 0, "step": 996, "train_loss": 3.996054172515869, "perplexity": 54.38313962811148, "lr": 0.001305, "grad_norm": 0.346096, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:24.244390+00:00", "epoch": 0, "step": 997, "train_loss": 3.98390531539917, "perplexity": 53.726443764371716, "lr": 0.001305, "grad_norm": 0.386709, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:24.552317+00:00", "epoch": 0, "step": 998, "train_loss": 3.9484307765960693, "perplexity": 51.853932552703924, "lr": 0.001305, "grad_norm": 0.358699, "tokens_per_sec": 106414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:24.860560+00:00", "epoch": 0, "step": 999, "train_loss": 4.009867191314697, "perplexity": 55.13954706640849, "lr": 0.001305, "grad_norm": 0.329507, "tokens_per_sec": 106306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:25.169468+00:00", "epoch": 0, "step": 1000, "train_loss": 4.070833206176758, "perplexity": 58.60577294653679, "lr": 0.001305, "grad_norm": 0.329844, "tokens_per_sec": 106149} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:49:30.846195+00:00", "step": 1000, "epoch": 0, "val_loss": 3.980747711658478, "val_ppl": 53.55706450159656, "eval_train_loss": 4.070833206176758, "eval_train_ppl": 58.60577294653679} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:49:31.701323+00:00", "step": 1000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p9807_epoch_0000_step_0001000.pt", "val_loss": 3.980747711658478} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:49:33.308441+00:00", "step": 1000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0000_step_0001000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:34.292940+00:00", "epoch": 0, "step": 1001, "train_loss": 4.155086040496826, "perplexity": 63.757450034499875, "lr": 0.001305, "grad_norm": 0.278317, "tokens_per_sec": 3592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:34.598976+00:00", "epoch": 0, "step": 1002, "train_loss": 4.0439863204956055, "perplexity": 57.053322927667736, "lr": 0.001305, "grad_norm": 0.272809, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:34.905016+00:00", "epoch": 0, "step": 1003, "train_loss": 4.091091156005859, "perplexity": 59.805112822503936, "lr": 0.001305, "grad_norm": 0.297302, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:35.209929+00:00", "epoch": 0, "step": 1004, "train_loss": 4.026014804840088, "perplexity": 56.03714670476732, "lr": 0.001305, "grad_norm": 0.300479, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:35.515338+00:00", "epoch": 0, "step": 1005, "train_loss": 4.026595592498779, "perplexity": 56.069701840897274, "lr": 0.001305, "grad_norm": 0.299368, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:35.821719+00:00", "epoch": 0, "step": 1006, "train_loss": 4.102006435394287, "perplexity": 60.46147802582414, "lr": 0.001305, "grad_norm": 0.313714, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:36.128627+00:00", "epoch": 0, "step": 1007, "train_loss": 4.086877346038818, "perplexity": 59.553635652621274, "lr": 0.001305, "grad_norm": 0.314596, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:36.435808+00:00", "epoch": 0, "step": 1008, "train_loss": 4.041133403778076, "perplexity": 56.89078651056969, "lr": 0.001305, "grad_norm": 0.293, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:36.743492+00:00", "epoch": 0, "step": 1009, "train_loss": 4.00779390335083, "perplexity": 55.02534533458886, "lr": 0.001305, "grad_norm": 0.309535, "tokens_per_sec": 106500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:37.293389+00:00", "epoch": 0, "step": 1010, "train_loss": 4.047861099243164, "perplexity": 57.27482078128436, "lr": 0.001305, "grad_norm": 0.291388, "tokens_per_sec": 59608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:37.599444+00:00", "epoch": 0, "step": 1011, "train_loss": 3.953580856323242, "perplexity": 52.121673290810655, "lr": 0.001305, "grad_norm": 0.298533, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:37.905432+00:00", "epoch": 0, "step": 1012, "train_loss": 3.961136817932129, "perplexity": 52.516994287502, "lr": 0.001305, "grad_norm": 0.274539, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:38.211557+00:00", "epoch": 0, "step": 1013, "train_loss": 4.059482097625732, "perplexity": 57.944293819280944, "lr": 0.001305, "grad_norm": 0.302105, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:38.517997+00:00", "epoch": 0, "step": 1014, "train_loss": 4.013208866119385, "perplexity": 55.32411371068328, "lr": 0.001305, "grad_norm": 0.266508, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:38.825336+00:00", "epoch": 0, "step": 1015, "train_loss": 4.084924221038818, "perplexity": 59.43743347361736, "lr": 0.001305, "grad_norm": 0.241153, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:39.132240+00:00", "epoch": 0, "step": 1016, "train_loss": 3.97373628616333, "perplexity": 53.182866497782825, "lr": 0.001305, "grad_norm": 0.302851, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:39.438011+00:00", "epoch": 0, "step": 1017, "train_loss": 4.058664798736572, "perplexity": 57.89695535977954, "lr": 0.001305, "grad_norm": 0.317935, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:39.744006+00:00", "epoch": 0, "step": 1018, "train_loss": 4.029726028442383, "perplexity": 56.24549946893134, "lr": 0.001305, "grad_norm": 0.290476, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:40.049719+00:00", "epoch": 0, "step": 1019, "train_loss": 4.018253803253174, "perplexity": 55.60392560958933, "lr": 0.001305, "grad_norm": 0.288901, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:40.356235+00:00", "epoch": 0, "step": 1020, "train_loss": 3.9430270195007324, "perplexity": 51.57448221762586, "lr": 0.001305, "grad_norm": 0.240588, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:40.663721+00:00", "epoch": 0, "step": 1021, "train_loss": 3.9643237590789795, "perplexity": 52.68462983793861, "lr": 0.001305, "grad_norm": 0.264182, "tokens_per_sec": 106502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:40.969763+00:00", "epoch": 0, "step": 1022, "train_loss": 3.954657793045044, "perplexity": 52.17783527080057, "lr": 0.001305, "grad_norm": 0.283218, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:41.276005+00:00", "epoch": 0, "step": 1023, "train_loss": 3.9242382049560547, "perplexity": 50.61450548907448, "lr": 0.001305, "grad_norm": 0.282577, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:41.581416+00:00", "epoch": 0, "step": 1024, "train_loss": 4.000959396362305, "perplexity": 54.650556434910456, "lr": 0.001305, "grad_norm": 0.263458, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:41.887105+00:00", "epoch": 0, "step": 1025, "train_loss": 3.9764201641082764, "perplexity": 53.325794535133895, "lr": 0.001305, "grad_norm": 0.281809, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:42.194028+00:00", "epoch": 0, "step": 1026, "train_loss": 4.031959533691406, "perplexity": 56.37126448335704, "lr": 0.001305, "grad_norm": 0.306926, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:42.500852+00:00", "epoch": 0, "step": 1027, "train_loss": 4.0895280838012695, "perplexity": 59.711706132676575, "lr": 0.001305, "grad_norm": 0.345113, "tokens_per_sec": 106797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:42.806424+00:00", "epoch": 0, "step": 1028, "train_loss": 3.9627885818481445, "perplexity": 52.6038116447939, "lr": 0.001305, "grad_norm": 0.304967, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:43.113767+00:00", "epoch": 0, "step": 1029, "train_loss": 4.053809642791748, "perplexity": 57.616537898874206, "lr": 0.001305, "grad_norm": 0.305744, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:43.420199+00:00", "epoch": 0, "step": 1030, "train_loss": 3.886479616165161, "perplexity": 48.73900415293887, "lr": 0.001305, "grad_norm": 0.313219, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:43.726552+00:00", "epoch": 0, "step": 1031, "train_loss": 4.136990547180176, "perplexity": 62.61410343251405, "lr": 0.001305, "grad_norm": 0.328729, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:44.033379+00:00", "epoch": 0, "step": 1032, "train_loss": 3.9958345890045166, "perplexity": 54.37119929835131, "lr": 0.001305, "grad_norm": 0.317212, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:44.339234+00:00", "epoch": 0, "step": 1033, "train_loss": 4.016068935394287, "perplexity": 55.48257099986551, "lr": 0.001305, "grad_norm": 0.300237, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:44.646141+00:00", "epoch": 0, "step": 1034, "train_loss": 4.107605457305908, "perplexity": 60.800952642157384, "lr": 0.001305, "grad_norm": 0.29584, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:44.953654+00:00", "epoch": 0, "step": 1035, "train_loss": 4.133735179901123, "perplexity": 62.41060294312475, "lr": 0.001305, "grad_norm": 0.266399, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:45.259497+00:00", "epoch": 0, "step": 1036, "train_loss": 4.144252300262451, "perplexity": 63.07044651112579, "lr": 0.001305, "grad_norm": 0.265602, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:45.565682+00:00", "epoch": 0, "step": 1037, "train_loss": 4.067138195037842, "perplexity": 58.38962354584203, "lr": 0.001305, "grad_norm": 0.251038, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:45.873692+00:00", "epoch": 0, "step": 1038, "train_loss": 3.8986971378326416, "perplexity": 49.33812643450173, "lr": 0.001305, "grad_norm": 0.247709, "tokens_per_sec": 106390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:46.181648+00:00", "epoch": 0, "step": 1039, "train_loss": 4.046926975250244, "perplexity": 57.221343977873545, "lr": 0.001305, "grad_norm": 0.260736, "tokens_per_sec": 106401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:46.488458+00:00", "epoch": 0, "step": 1040, "train_loss": 4.067603588104248, "perplexity": 58.41680399609629, "lr": 0.001305, "grad_norm": 0.274952, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:46.794739+00:00", "epoch": 0, "step": 1041, "train_loss": 4.0008416175842285, "perplexity": 54.64412013818939, "lr": 0.001305, "grad_norm": 0.269224, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:47.100235+00:00", "epoch": 0, "step": 1042, "train_loss": 4.018970012664795, "perplexity": 55.643763929020714, "lr": 0.001305, "grad_norm": 0.247121, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:47.407257+00:00", "epoch": 0, "step": 1043, "train_loss": 4.051864147186279, "perplexity": 57.50455414486139, "lr": 0.001305, "grad_norm": 0.229975, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:47.713683+00:00", "epoch": 0, "step": 1044, "train_loss": 4.014735698699951, "perplexity": 55.40864888909753, "lr": 0.001305, "grad_norm": 0.265427, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:48.021014+00:00", "epoch": 0, "step": 1045, "train_loss": 3.8929755687713623, "perplexity": 49.05664097390407, "lr": 0.001305, "grad_norm": 0.263513, "tokens_per_sec": 106626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:48.329304+00:00", "epoch": 0, "step": 1046, "train_loss": 3.9858593940734863, "perplexity": 53.83153210438129, "lr": 0.001305, "grad_norm": 0.256664, "tokens_per_sec": 106290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:48.635545+00:00", "epoch": 0, "step": 1047, "train_loss": 3.997727870941162, "perplexity": 54.474236816614386, "lr": 0.001305, "grad_norm": 0.266628, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:48.941041+00:00", "epoch": 0, "step": 1048, "train_loss": 4.085544109344482, "perplexity": 59.47428946566538, "lr": 0.001305, "grad_norm": 0.305298, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:49.247856+00:00", "epoch": 0, "step": 1049, "train_loss": 3.9629852771759033, "perplexity": 52.61415958642926, "lr": 0.001305, "grad_norm": 0.352983, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:49.554475+00:00", "epoch": 0, "step": 1050, "train_loss": 3.943528890609741, "perplexity": 51.600372456450984, "lr": 0.001305, "grad_norm": 0.32344, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:49.861387+00:00", "epoch": 0, "step": 1051, "train_loss": 4.025890350341797, "perplexity": 56.03017306374802, "lr": 0.001305, "grad_norm": 0.3038, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:50.167803+00:00", "epoch": 0, "step": 1052, "train_loss": 4.011784076690674, "perplexity": 55.24534462632012, "lr": 0.001305, "grad_norm": 0.310362, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:50.474079+00:00", "epoch": 0, "step": 1053, "train_loss": 3.998565196990967, "perplexity": 54.51986861582516, "lr": 0.001305, "grad_norm": 0.331227, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:50.780509+00:00", "epoch": 0, "step": 1054, "train_loss": 4.101650714874268, "perplexity": 60.4399744622758, "lr": 0.001305, "grad_norm": 0.356773, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:51.087176+00:00", "epoch": 0, "step": 1055, "train_loss": 3.964747667312622, "perplexity": 52.7069680206486, "lr": 0.001305, "grad_norm": 0.288196, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:51.393895+00:00", "epoch": 0, "step": 1056, "train_loss": 3.992335319519043, "perplexity": 54.181272316407004, "lr": 0.001305, "grad_norm": 0.333868, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:51.700242+00:00", "epoch": 0, "step": 1057, "train_loss": 4.010826587677002, "perplexity": 55.192473131760806, "lr": 0.001305, "grad_norm": 0.380411, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:52.006039+00:00", "epoch": 0, "step": 1058, "train_loss": 4.0076446533203125, "perplexity": 55.01713341294846, "lr": 0.001305, "grad_norm": 0.337527, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:52.311919+00:00", "epoch": 0, "step": 1059, "train_loss": 4.089902400970459, "perplexity": 59.734061433209014, "lr": 0.001305, "grad_norm": 0.2759, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:52.619771+00:00", "epoch": 0, "step": 1060, "train_loss": 4.035571575164795, "perplexity": 56.575248005270424, "lr": 0.001305, "grad_norm": 0.291826, "tokens_per_sec": 106498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:52.926533+00:00", "epoch": 0, "step": 1061, "train_loss": 4.029468536376953, "perplexity": 56.23101856354109, "lr": 0.001305, "grad_norm": 0.28587, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:53.234251+00:00", "epoch": 0, "step": 1062, "train_loss": 3.948094606399536, "perplexity": 51.83650373569522, "lr": 0.001305, "grad_norm": 0.25239, "tokens_per_sec": 106488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:53.540470+00:00", "epoch": 0, "step": 1063, "train_loss": 3.9785680770874023, "perplexity": 53.440456799537934, "lr": 0.001305, "grad_norm": 0.266161, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:53.846302+00:00", "epoch": 0, "step": 1064, "train_loss": 4.014606475830078, "perplexity": 55.40148928707441, "lr": 0.001305, "grad_norm": 0.257262, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:54.153274+00:00", "epoch": 0, "step": 1065, "train_loss": 4.058176040649414, "perplexity": 57.868664668840786, "lr": 0.001305, "grad_norm": 0.286178, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:54.459595+00:00", "epoch": 0, "step": 1066, "train_loss": 4.038111209869385, "perplexity": 56.719111071019306, "lr": 0.001305, "grad_norm": 0.287907, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:54.766282+00:00", "epoch": 0, "step": 1067, "train_loss": 3.8266379833221436, "perplexity": 45.90793524698252, "lr": 0.001305, "grad_norm": 0.302876, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:55.073292+00:00", "epoch": 0, "step": 1068, "train_loss": 4.045862674713135, "perplexity": 57.16047566757708, "lr": 0.001305, "grad_norm": 0.29974, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:55.379573+00:00", "epoch": 0, "step": 1069, "train_loss": 3.9972994327545166, "perplexity": 54.45090297228568, "lr": 0.001305, "grad_norm": 0.343594, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:55.686579+00:00", "epoch": 0, "step": 1070, "train_loss": 3.9680988788604736, "perplexity": 52.88389651735762, "lr": 0.001305, "grad_norm": 0.306631, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:55.992925+00:00", "epoch": 0, "step": 1071, "train_loss": 3.984431028366089, "perplexity": 53.75469587812568, "lr": 0.001305, "grad_norm": 0.273446, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:56.300363+00:00", "epoch": 0, "step": 1072, "train_loss": 3.9733877182006836, "perplexity": 53.16433188483379, "lr": 0.001305, "grad_norm": 0.29207, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:56.607670+00:00", "epoch": 0, "step": 1073, "train_loss": 4.129256725311279, "perplexity": 62.13172482951419, "lr": 0.001305, "grad_norm": 0.304818, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:56.913515+00:00", "epoch": 0, "step": 1074, "train_loss": 4.031503677368164, "perplexity": 56.34557314221819, "lr": 0.001305, "grad_norm": 0.267086, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:57.220661+00:00", "epoch": 0, "step": 1075, "train_loss": 3.9764370918273926, "perplexity": 53.32669722684567, "lr": 0.001305, "grad_norm": 0.256661, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:57.526471+00:00", "epoch": 0, "step": 1076, "train_loss": 3.9570324420928955, "perplexity": 52.30188654847622, "lr": 0.001305, "grad_norm": 0.259791, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:57.834156+00:00", "epoch": 0, "step": 1077, "train_loss": 3.9481492042541504, "perplexity": 51.839333974851684, "lr": 0.001305, "grad_norm": 0.291145, "tokens_per_sec": 106497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:58.140833+00:00", "epoch": 0, "step": 1078, "train_loss": 4.056660175323486, "perplexity": 57.78101001971794, "lr": 0.001305, "grad_norm": 0.276321, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:58.448355+00:00", "epoch": 0, "step": 1079, "train_loss": 3.939277172088623, "perplexity": 51.381447929707306, "lr": 0.001305, "grad_norm": 0.251327, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:58.754407+00:00", "epoch": 0, "step": 1080, "train_loss": 4.044550895690918, "perplexity": 57.08554291304929, "lr": 0.001305, "grad_norm": 0.281562, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:59.059765+00:00", "epoch": 0, "step": 1081, "train_loss": 3.8516030311584473, "perplexity": 47.06845498769288, "lr": 0.001305, "grad_norm": 0.279755, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:59.365091+00:00", "epoch": 0, "step": 1082, "train_loss": 3.9531493186950684, "perplexity": 52.099185680016376, "lr": 0.001305, "grad_norm": 0.280657, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:59.672008+00:00", "epoch": 0, "step": 1083, "train_loss": 3.9669435024261475, "perplexity": 52.82283099319291, "lr": 0.001305, "grad_norm": 0.296269, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:49:59.979438+00:00", "epoch": 0, "step": 1084, "train_loss": 3.9119789600372314, "perplexity": 49.99779777895347, "lr": 0.001305, "grad_norm": 0.268687, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:00.286610+00:00", "epoch": 0, "step": 1085, "train_loss": 4.018195629119873, "perplexity": 55.60069099349528, "lr": 0.001305, "grad_norm": 0.270581, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:00.592999+00:00", "epoch": 0, "step": 1086, "train_loss": 4.050235748291016, "perplexity": 57.41098999297525, "lr": 0.001305, "grad_norm": 0.242941, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:00.898777+00:00", "epoch": 0, "step": 1087, "train_loss": 3.887457847595215, "perplexity": 48.78670550634844, "lr": 0.001305, "grad_norm": 0.251651, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:01.205551+00:00", "epoch": 0, "step": 1088, "train_loss": 4.0039215087890625, "perplexity": 54.81267751915568, "lr": 0.001305, "grad_norm": 0.264875, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:01.513006+00:00", "epoch": 0, "step": 1089, "train_loss": 3.941650629043579, "perplexity": 51.503544422719884, "lr": 0.001305, "grad_norm": 0.262646, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:01.819940+00:00", "epoch": 0, "step": 1090, "train_loss": 3.9623470306396484, "perplexity": 52.58058949545199, "lr": 0.001305, "grad_norm": 0.262122, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:02.126483+00:00", "epoch": 0, "step": 1091, "train_loss": 3.9419496059417725, "perplexity": 51.51894509478522, "lr": 0.001305, "grad_norm": 0.260716, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:02.432863+00:00", "epoch": 0, "step": 1092, "train_loss": 3.9848034381866455, "perplexity": 53.774718382829015, "lr": 0.001305, "grad_norm": 0.253543, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:02.739741+00:00", "epoch": 0, "step": 1093, "train_loss": 3.929363250732422, "perplexity": 50.87457300637348, "lr": 0.001305, "grad_norm": 0.302297, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:03.046177+00:00", "epoch": 0, "step": 1094, "train_loss": 4.005411624908447, "perplexity": 54.89441565800815, "lr": 0.001305, "grad_norm": 0.352055, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:03.353464+00:00", "epoch": 0, "step": 1095, "train_loss": 4.072176933288574, "perplexity": 58.68457604562712, "lr": 0.001305, "grad_norm": 0.337208, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:03.659701+00:00", "epoch": 0, "step": 1096, "train_loss": 3.962031364440918, "perplexity": 52.56399420006366, "lr": 0.001305, "grad_norm": 0.287777, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:03.966930+00:00", "epoch": 0, "step": 1097, "train_loss": 4.002562999725342, "perplexity": 54.73826455672545, "lr": 0.001305, "grad_norm": 0.291489, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:04.273445+00:00", "epoch": 0, "step": 1098, "train_loss": 3.9610836505889893, "perplexity": 52.514202172671375, "lr": 0.001305, "grad_norm": 0.294237, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:04.580697+00:00", "epoch": 0, "step": 1099, "train_loss": 3.986543655395508, "perplexity": 53.86837954490695, "lr": 0.001305, "grad_norm": 0.321268, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:04.888896+00:00", "epoch": 0, "step": 1100, "train_loss": 3.93034029006958, "perplexity": 50.92430375595504, "lr": 0.001305, "grad_norm": 0.344016, "tokens_per_sec": 106375} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:50:10.595049+00:00", "step": 1100, "epoch": 0, "val_loss": 3.9379961609840395, "val_ppl": 51.31566986455081, "eval_train_loss": 3.93034029006958, "eval_train_ppl": 50.92430375595504} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:50:11.552353+00:00", "step": 1100, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p9380_epoch_0000_step_0001100.pt", "val_loss": 3.9379961609840395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:12.508946+00:00", "epoch": 0, "step": 1101, "train_loss": 3.9965498447418213, "perplexity": 54.410102521811936, "lr": 0.001305, "grad_norm": 0.389326, "tokens_per_sec": 4300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:12.814780+00:00", "epoch": 0, "step": 1102, "train_loss": 4.034045696258545, "perplexity": 56.48898685649386, "lr": 0.001305, "grad_norm": 0.350976, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:13.121979+00:00", "epoch": 0, "step": 1103, "train_loss": 3.860746145248413, "perplexity": 47.50078063128584, "lr": 0.001305, "grad_norm": 0.343828, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:13.428203+00:00", "epoch": 0, "step": 1104, "train_loss": 3.9585776329040527, "perplexity": 52.38276541352436, "lr": 0.001305, "grad_norm": 0.34558, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:13.734236+00:00", "epoch": 0, "step": 1105, "train_loss": 3.949089527130127, "perplexity": 51.88810261200273, "lr": 0.001305, "grad_norm": 0.372637, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:14.040011+00:00", "epoch": 0, "step": 1106, "train_loss": 3.989687919616699, "perplexity": 54.038022524776416, "lr": 0.001305, "grad_norm": 0.343436, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:14.346862+00:00", "epoch": 0, "step": 1107, "train_loss": 3.9623794555664062, "perplexity": 52.582294444856544, "lr": 0.001305, "grad_norm": 0.312497, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:14.654507+00:00", "epoch": 0, "step": 1108, "train_loss": 3.99497652053833, "perplexity": 54.324565097294666, "lr": 0.001305, "grad_norm": 0.296599, "tokens_per_sec": 106512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:14.959933+00:00", "epoch": 0, "step": 1109, "train_loss": 4.004568576812744, "perplexity": 54.84815652749869, "lr": 0.001305, "grad_norm": 0.29201, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:15.265578+00:00", "epoch": 0, "step": 1110, "train_loss": 3.895937204360962, "perplexity": 49.20214422516807, "lr": 0.001305, "grad_norm": 0.281561, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:15.572023+00:00", "epoch": 0, "step": 1111, "train_loss": 3.96746563911438, "perplexity": 52.850418932941736, "lr": 0.001305, "grad_norm": 0.261558, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:15.878439+00:00", "epoch": 0, "step": 1112, "train_loss": 3.892289161682129, "perplexity": 49.022979701755155, "lr": 0.001305, "grad_norm": 0.250359, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:16.184819+00:00", "epoch": 0, "step": 1113, "train_loss": 4.008264541625977, "perplexity": 55.05124846323024, "lr": 0.001305, "grad_norm": 0.230957, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:16.491481+00:00", "epoch": 0, "step": 1114, "train_loss": 4.061429500579834, "perplexity": 58.05724465293548, "lr": 0.001305, "grad_norm": 0.24045, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:16.819317+00:00", "epoch": 0, "step": 1115, "train_loss": 3.9638113975524902, "perplexity": 52.65764317462669, "lr": 0.001305, "grad_norm": 0.243966, "tokens_per_sec": 99953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:17.125651+00:00", "epoch": 0, "step": 1116, "train_loss": 3.9608774185180664, "perplexity": 52.503373176685955, "lr": 0.001305, "grad_norm": 0.263083, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:17.433011+00:00", "epoch": 0, "step": 1117, "train_loss": 4.105119705200195, "perplexity": 60.65000423398491, "lr": 0.001305, "grad_norm": 0.262903, "tokens_per_sec": 106612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:17.740757+00:00", "epoch": 0, "step": 1118, "train_loss": 3.9909884929656982, "perplexity": 54.10834865893966, "lr": 0.001305, "grad_norm": 0.27663, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:18.047614+00:00", "epoch": 0, "step": 1119, "train_loss": 3.990072250366211, "perplexity": 54.05879498996836, "lr": 0.001305, "grad_norm": 0.263256, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:18.353169+00:00", "epoch": 0, "step": 1120, "train_loss": 4.04335880279541, "perplexity": 57.0175321885113, "lr": 0.001305, "grad_norm": 0.286719, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:18.659910+00:00", "epoch": 0, "step": 1121, "train_loss": 3.9095284938812256, "perplexity": 49.875429858078405, "lr": 0.001305, "grad_norm": 0.286107, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:18.966284+00:00", "epoch": 0, "step": 1122, "train_loss": 3.9745371341705322, "perplexity": 53.22547494960447, "lr": 0.001305, "grad_norm": 0.25167, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:19.272689+00:00", "epoch": 0, "step": 1123, "train_loss": 3.9178216457366943, "perplexity": 50.290774248311344, "lr": 0.001305, "grad_norm": 0.259287, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:19.579787+00:00", "epoch": 0, "step": 1124, "train_loss": 4.069129943847656, "perplexity": 58.50603690364347, "lr": 0.001305, "grad_norm": 0.260192, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:19.885812+00:00", "epoch": 0, "step": 1125, "train_loss": 4.045229434967041, "perplexity": 57.12429084052325, "lr": 0.001305, "grad_norm": 0.268613, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:20.191397+00:00", "epoch": 0, "step": 1126, "train_loss": 3.894322156906128, "perplexity": 49.1227445617357, "lr": 0.001305, "grad_norm": 0.274628, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:20.497653+00:00", "epoch": 0, "step": 1127, "train_loss": 4.088395595550537, "perplexity": 59.64412160361594, "lr": 0.001305, "grad_norm": 0.306626, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:20.804427+00:00", "epoch": 0, "step": 1128, "train_loss": 3.887842893600464, "perplexity": 48.80549424944602, "lr": 0.001305, "grad_norm": 0.28279, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:21.112579+00:00", "epoch": 0, "step": 1129, "train_loss": 4.027771472930908, "perplexity": 56.13567188496788, "lr": 0.001305, "grad_norm": 0.260729, "tokens_per_sec": 106337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:21.419275+00:00", "epoch": 0, "step": 1130, "train_loss": 3.946058988571167, "perplexity": 51.73109175018901, "lr": 0.001305, "grad_norm": 0.270996, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:21.725860+00:00", "epoch": 0, "step": 1131, "train_loss": 3.964371919631958, "perplexity": 52.687167219945444, "lr": 0.001305, "grad_norm": 0.252979, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:22.031333+00:00", "epoch": 0, "step": 1132, "train_loss": 4.119678020477295, "perplexity": 61.53942463959203, "lr": 0.001305, "grad_norm": 0.257894, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:22.338098+00:00", "epoch": 0, "step": 1133, "train_loss": 4.06617546081543, "perplexity": 58.333436907760095, "lr": 0.001305, "grad_norm": 0.267227, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:22.646028+00:00", "epoch": 0, "step": 1134, "train_loss": 3.870678186416626, "perplexity": 47.97491098356691, "lr": 0.001305, "grad_norm": 0.259028, "tokens_per_sec": 106417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:22.953890+00:00", "epoch": 0, "step": 1135, "train_loss": 3.8695735931396484, "perplexity": 47.921947476383025, "lr": 0.001305, "grad_norm": 0.252162, "tokens_per_sec": 106435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:23.260326+00:00", "epoch": 0, "step": 1136, "train_loss": 4.0273847579956055, "perplexity": 56.11396757920679, "lr": 0.001305, "grad_norm": 0.285267, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:23.566678+00:00", "epoch": 0, "step": 1137, "train_loss": 3.9944961071014404, "perplexity": 54.298473114240196, "lr": 0.001305, "grad_norm": 0.275516, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:23.872934+00:00", "epoch": 0, "step": 1138, "train_loss": 3.9584505558013916, "perplexity": 52.37610918640222, "lr": 0.001305, "grad_norm": 0.301317, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:24.180548+00:00", "epoch": 0, "step": 1139, "train_loss": 3.8874027729034424, "perplexity": 48.78401866756917, "lr": 0.001305, "grad_norm": 0.308502, "tokens_per_sec": 106524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:24.487597+00:00", "epoch": 0, "step": 1140, "train_loss": 4.0211310386657715, "perplexity": 55.764141572424904, "lr": 0.001305, "grad_norm": 0.312706, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:24.794613+00:00", "epoch": 0, "step": 1141, "train_loss": 3.949732542037964, "perplexity": 51.921478164860666, "lr": 0.001305, "grad_norm": 0.323112, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:25.100653+00:00", "epoch": 0, "step": 1142, "train_loss": 3.899759531021118, "perplexity": 49.390570777281475, "lr": 0.001305, "grad_norm": 0.299387, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:25.407629+00:00", "epoch": 0, "step": 1143, "train_loss": 4.051936149597168, "perplexity": 57.50869476046226, "lr": 0.001305, "grad_norm": 0.274152, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:25.713429+00:00", "epoch": 0, "step": 1144, "train_loss": 4.013156890869141, "perplexity": 55.32123830075436, "lr": 0.001305, "grad_norm": 0.265103, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:26.020666+00:00", "epoch": 0, "step": 1145, "train_loss": 3.9819540977478027, "perplexity": 53.62171398747584, "lr": 0.001305, "grad_norm": 0.282866, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:26.327114+00:00", "epoch": 0, "step": 1146, "train_loss": 4.088577747344971, "perplexity": 59.65498687692793, "lr": 0.001305, "grad_norm": 0.307037, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:26.633680+00:00", "epoch": 0, "step": 1147, "train_loss": 3.872994899749756, "perplexity": 48.08618394350489, "lr": 0.001305, "grad_norm": 0.300391, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:26.940319+00:00", "epoch": 0, "step": 1148, "train_loss": 3.9038729667663574, "perplexity": 49.59415414308307, "lr": 0.001305, "grad_norm": 0.298479, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:27.246567+00:00", "epoch": 0, "step": 1149, "train_loss": 4.159337997436523, "perplexity": 64.02912112422096, "lr": 0.001305, "grad_norm": 0.287888, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:27.553632+00:00", "epoch": 0, "step": 1150, "train_loss": 3.8917150497436523, "perplexity": 48.9948431014006, "lr": 0.001305, "grad_norm": 0.283126, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:27.860931+00:00", "epoch": 0, "step": 1151, "train_loss": 3.9168031215667725, "perplexity": 50.23957795597181, "lr": 0.001305, "grad_norm": 0.276837, "tokens_per_sec": 106571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:28.168767+00:00", "epoch": 0, "step": 1152, "train_loss": 3.9612982273101807, "perplexity": 52.525471707036196, "lr": 0.001305, "grad_norm": 0.27014, "tokens_per_sec": 106446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:28.475075+00:00", "epoch": 0, "step": 1153, "train_loss": 3.9663350582122803, "perplexity": 52.790701022953485, "lr": 0.001305, "grad_norm": 0.246935, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:28.781350+00:00", "epoch": 0, "step": 1154, "train_loss": 3.9897286891937256, "perplexity": 54.04022567700858, "lr": 0.001305, "grad_norm": 0.266267, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:29.087582+00:00", "epoch": 0, "step": 1155, "train_loss": 3.8958230018615723, "perplexity": 49.196525538162376, "lr": 0.001305, "grad_norm": 0.281302, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:29.394639+00:00", "epoch": 0, "step": 1156, "train_loss": 3.9890856742858887, "perplexity": 54.005488175837854, "lr": 0.001305, "grad_norm": 0.293568, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:29.702354+00:00", "epoch": 0, "step": 1157, "train_loss": 3.8288025856018066, "perplexity": 46.007415296801234, "lr": 0.001305, "grad_norm": 0.269874, "tokens_per_sec": 106485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:30.010032+00:00", "epoch": 0, "step": 1158, "train_loss": 4.027138710021973, "perplexity": 56.10016254961086, "lr": 0.001305, "grad_norm": 0.250684, "tokens_per_sec": 106502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:30.317543+00:00", "epoch": 0, "step": 1159, "train_loss": 4.0090861320495605, "perplexity": 55.09649662695772, "lr": 0.001305, "grad_norm": 0.25708, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:30.623854+00:00", "epoch": 0, "step": 1160, "train_loss": 4.019067287445068, "perplexity": 55.64917692720034, "lr": 0.001305, "grad_norm": 0.249359, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:30.930361+00:00", "epoch": 0, "step": 1161, "train_loss": 3.982210159301758, "perplexity": 53.63544620495665, "lr": 0.001305, "grad_norm": 0.256113, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:31.237076+00:00", "epoch": 0, "step": 1162, "train_loss": 3.8314425945281982, "perplexity": 46.12903575277206, "lr": 0.001305, "grad_norm": 0.24411, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:31.544028+00:00", "epoch": 0, "step": 1163, "train_loss": 3.8672165870666504, "perplexity": 47.80912816528243, "lr": 0.001305, "grad_norm": 0.290233, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:31.851111+00:00", "epoch": 0, "step": 1164, "train_loss": 4.021851539611816, "perplexity": 55.804334166845635, "lr": 0.001305, "grad_norm": 0.311011, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:32.157485+00:00", "epoch": 0, "step": 1165, "train_loss": 3.91550612449646, "perplexity": 50.174459608830425, "lr": 0.001305, "grad_norm": 0.316769, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:32.463748+00:00", "epoch": 0, "step": 1166, "train_loss": 3.9342901706695557, "perplexity": 51.12584644818141, "lr": 0.001305, "grad_norm": 0.303341, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:32.770772+00:00", "epoch": 0, "step": 1167, "train_loss": 3.8870811462402344, "perplexity": 48.76833094935671, "lr": 0.001305, "grad_norm": 0.302049, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:33.077582+00:00", "epoch": 0, "step": 1168, "train_loss": 3.9064526557922363, "perplexity": 49.722256799741096, "lr": 0.001305, "grad_norm": 0.267914, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:33.383985+00:00", "epoch": 0, "step": 1169, "train_loss": 3.8931570053100586, "perplexity": 49.06554244854439, "lr": 0.001305, "grad_norm": 0.257162, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:33.691290+00:00", "epoch": 0, "step": 1170, "train_loss": 3.906771183013916, "perplexity": 49.73809721472292, "lr": 0.001305, "grad_norm": 0.248607, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:33.997986+00:00", "epoch": 0, "step": 1171, "train_loss": 3.9458844661712646, "perplexity": 51.72206430367593, "lr": 0.001305, "grad_norm": 0.248263, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:34.305027+00:00", "epoch": 0, "step": 1172, "train_loss": 3.964401960372925, "perplexity": 52.68875000526207, "lr": 0.001305, "grad_norm": 0.279488, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:34.611835+00:00", "epoch": 0, "step": 1173, "train_loss": 3.885810375213623, "perplexity": 48.70639692768478, "lr": 0.001305, "grad_norm": 0.307498, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:34.918444+00:00", "epoch": 0, "step": 1174, "train_loss": 3.9687111377716064, "perplexity": 52.91628506833351, "lr": 0.001305, "grad_norm": 0.312319, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:35.226078+00:00", "epoch": 0, "step": 1175, "train_loss": 3.9301915168762207, "perplexity": 50.91672814820331, "lr": 0.001305, "grad_norm": 0.271415, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:35.532287+00:00", "epoch": 0, "step": 1176, "train_loss": 3.972399950027466, "perplexity": 53.111843777156345, "lr": 0.001305, "grad_norm": 0.296931, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:35.838944+00:00", "epoch": 0, "step": 1177, "train_loss": 4.054068088531494, "perplexity": 57.631430572024165, "lr": 0.001305, "grad_norm": 0.305724, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:36.145671+00:00", "epoch": 0, "step": 1178, "train_loss": 3.8692543506622314, "perplexity": 47.90665119688901, "lr": 0.001305, "grad_norm": 0.300522, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:36.452707+00:00", "epoch": 0, "step": 1179, "train_loss": 3.9698290824890137, "perplexity": 52.97547562941331, "lr": 0.001305, "grad_norm": 0.326846, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:36.759667+00:00", "epoch": 0, "step": 1180, "train_loss": 3.8646037578582764, "perplexity": 47.68437413025193, "lr": 0.001305, "grad_norm": 0.261575, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:37.066971+00:00", "epoch": 0, "step": 1181, "train_loss": 3.9811034202575684, "perplexity": 53.57611859863399, "lr": 0.001305, "grad_norm": 0.261486, "tokens_per_sec": 106573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:37.373354+00:00", "epoch": 0, "step": 1182, "train_loss": 3.9760594367980957, "perplexity": 53.306561933778816, "lr": 0.001305, "grad_norm": 0.271483, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:37.680954+00:00", "epoch": 0, "step": 1183, "train_loss": 3.844043254852295, "perplexity": 46.713969600267575, "lr": 0.001305, "grad_norm": 0.28855, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:37.987170+00:00", "epoch": 0, "step": 1184, "train_loss": 3.8710269927978516, "perplexity": 47.99164785745122, "lr": 0.001305, "grad_norm": 0.341418, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:38.294768+00:00", "epoch": 0, "step": 1185, "train_loss": 3.9347312450408936, "perplexity": 51.14840172267375, "lr": 0.001305, "grad_norm": 0.378312, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:38.602230+00:00", "epoch": 0, "step": 1186, "train_loss": 3.903965950012207, "perplexity": 49.59876578290975, "lr": 0.001305, "grad_norm": 0.337171, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:38.908713+00:00", "epoch": 0, "step": 1187, "train_loss": 3.917018413543701, "perplexity": 50.250395298431755, "lr": 0.001305, "grad_norm": 0.327929, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:39.214916+00:00", "epoch": 0, "step": 1188, "train_loss": 4.039402484893799, "perplexity": 56.792398349385145, "lr": 0.001305, "grad_norm": 0.342646, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:39.522112+00:00", "epoch": 0, "step": 1189, "train_loss": 3.932455539703369, "perplexity": 51.032135376017465, "lr": 0.001305, "grad_norm": 0.33367, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:39.829705+00:00", "epoch": 0, "step": 1190, "train_loss": 3.9071450233459473, "perplexity": 49.75669479754695, "lr": 0.001305, "grad_norm": 0.297739, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:40.136789+00:00", "epoch": 0, "step": 1191, "train_loss": 3.9733285903930664, "perplexity": 53.16118848737803, "lr": 0.001305, "grad_norm": 0.294209, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:40.444129+00:00", "epoch": 0, "step": 1192, "train_loss": 3.92694354057312, "perplexity": 50.7516201004084, "lr": 0.001305, "grad_norm": 0.300342, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:40.751248+00:00", "epoch": 0, "step": 1193, "train_loss": 3.934605836868286, "perplexity": 51.14198769727593, "lr": 0.001305, "grad_norm": 0.314084, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:41.057939+00:00", "epoch": 0, "step": 1194, "train_loss": 3.954920768737793, "perplexity": 52.19155857754602, "lr": 0.001305, "grad_norm": 0.259552, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:41.365071+00:00", "epoch": 0, "step": 1195, "train_loss": 3.8797366619110107, "perplexity": 48.4114648099759, "lr": 0.001305, "grad_norm": 0.281064, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:41.671901+00:00", "epoch": 0, "step": 1196, "train_loss": 3.905799627304077, "perplexity": 49.68979734918178, "lr": 0.001305, "grad_norm": 0.312661, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:41.979761+00:00", "epoch": 0, "step": 1197, "train_loss": 3.8833987712860107, "perplexity": 48.58907790991513, "lr": 0.001305, "grad_norm": 0.331889, "tokens_per_sec": 106438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:42.287289+00:00", "epoch": 0, "step": 1198, "train_loss": 3.867893934249878, "perplexity": 47.84152251344945, "lr": 0.001305, "grad_norm": 0.303455, "tokens_per_sec": 106552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:42.593515+00:00", "epoch": 0, "step": 1199, "train_loss": 3.945671319961548, "perplexity": 51.71104111652788, "lr": 0.001305, "grad_norm": 0.310466, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:42.901141+00:00", "epoch": 0, "step": 1200, "train_loss": 3.929852247238159, "perplexity": 50.8994565782985, "lr": 0.001305, "grad_norm": 0.302409, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:50:48.586181+00:00", "step": 1200, "epoch": 0, "val_loss": 3.894532287120819, "val_ppl": 49.13306781917284, "eval_train_loss": 3.929852247238159, "eval_train_ppl": 50.8994565782985} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:50:49.533241+00:00", "step": 1200, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p8945_epoch_0000_step_0001200.pt", "val_loss": 3.894532287120819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:50.771390+00:00", "epoch": 0, "step": 1201, "train_loss": 3.914604902267456, "perplexity": 50.12926164026884, "lr": 0.001305, "grad_norm": 0.292437, "tokens_per_sec": 4163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:51.077531+00:00", "epoch": 0, "step": 1202, "train_loss": 3.798640727996826, "perplexity": 44.64046470141587, "lr": 0.001305, "grad_norm": 0.265953, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:51.384014+00:00", "epoch": 0, "step": 1203, "train_loss": 3.897697925567627, "perplexity": 49.28885179544714, "lr": 0.001305, "grad_norm": 0.257361, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:51.690378+00:00", "epoch": 0, "step": 1204, "train_loss": 3.940929412841797, "perplexity": 51.46641262367007, "lr": 0.001305, "grad_norm": 0.271345, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:51.996552+00:00", "epoch": 0, "step": 1205, "train_loss": 3.911940574645996, "perplexity": 49.99587863075868, "lr": 0.001305, "grad_norm": 0.261986, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:52.301731+00:00", "epoch": 0, "step": 1206, "train_loss": 3.847853660583496, "perplexity": 46.89230833344729, "lr": 0.001305, "grad_norm": 0.235561, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:52.607009+00:00", "epoch": 0, "step": 1207, "train_loss": 4.145885467529297, "perplexity": 63.173535257542724, "lr": 0.001305, "grad_norm": 0.233359, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:52.912772+00:00", "epoch": 0, "step": 1208, "train_loss": 3.9513981342315674, "perplexity": 52.00803023378821, "lr": 0.001305, "grad_norm": 0.238136, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:53.219560+00:00", "epoch": 0, "step": 1209, "train_loss": 3.9306557178497314, "perplexity": 50.94036922965968, "lr": 0.001305, "grad_norm": 0.269639, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:53.525424+00:00", "epoch": 0, "step": 1210, "train_loss": 3.8974151611328125, "perplexity": 49.274916631403784, "lr": 0.001305, "grad_norm": 0.275841, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:53.831996+00:00", "epoch": 0, "step": 1211, "train_loss": 3.9525599479675293, "perplexity": 52.068488991801246, "lr": 0.001305, "grad_norm": 0.27306, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:54.137671+00:00", "epoch": 0, "step": 1212, "train_loss": 3.8977904319763184, "perplexity": 49.29341154101485, "lr": 0.001305, "grad_norm": 0.287015, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:54.444000+00:00", "epoch": 0, "step": 1213, "train_loss": 3.9616503715515137, "perplexity": 52.543971506528884, "lr": 0.001305, "grad_norm": 0.279894, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:54.750877+00:00", "epoch": 0, "step": 1214, "train_loss": 3.9717209339141846, "perplexity": 53.075792220604654, "lr": 0.001305, "grad_norm": 0.258613, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:55.058982+00:00", "epoch": 0, "step": 1215, "train_loss": 3.9005372524261475, "perplexity": 49.42899782221239, "lr": 0.001305, "grad_norm": 0.273065, "tokens_per_sec": 106353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:55.365359+00:00", "epoch": 0, "step": 1216, "train_loss": 3.820981502532959, "perplexity": 45.64899094036138, "lr": 0.001305, "grad_norm": 0.256445, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:55.671285+00:00", "epoch": 0, "step": 1217, "train_loss": 3.9293320178985596, "perplexity": 50.872984074100515, "lr": 0.001305, "grad_norm": 0.250173, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:55.976458+00:00", "epoch": 0, "step": 1218, "train_loss": 3.936272382736206, "perplexity": 51.22728922526167, "lr": 0.001305, "grad_norm": 0.249053, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:56.282781+00:00", "epoch": 0, "step": 1219, "train_loss": 3.998000383377075, "perplexity": 54.48908374647842, "lr": 0.001305, "grad_norm": 0.23761, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:56.589621+00:00", "epoch": 0, "step": 1220, "train_loss": 3.8784351348876953, "perplexity": 48.348496966351284, "lr": 0.001305, "grad_norm": 0.254335, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:56.895841+00:00", "epoch": 0, "step": 1221, "train_loss": 3.825382709503174, "perplexity": 45.850344371505344, "lr": 0.001305, "grad_norm": 0.255384, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:57.202138+00:00", "epoch": 0, "step": 1222, "train_loss": 4.062953472137451, "perplexity": 58.14578969542627, "lr": 0.001305, "grad_norm": 0.270558, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:57.508871+00:00", "epoch": 0, "step": 1223, "train_loss": 3.970444679260254, "perplexity": 53.00809720100258, "lr": 0.001305, "grad_norm": 0.257925, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:57.816133+00:00", "epoch": 0, "step": 1224, "train_loss": 3.9536690711975098, "perplexity": 52.12627140047447, "lr": 0.001305, "grad_norm": 0.27883, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:58.121916+00:00", "epoch": 0, "step": 1225, "train_loss": 3.952338933944702, "perplexity": 52.05698239719286, "lr": 0.001305, "grad_norm": 0.3076, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:58.428596+00:00", "epoch": 0, "step": 1226, "train_loss": 4.047458648681641, "perplexity": 57.25177513517744, "lr": 0.001305, "grad_norm": 0.296498, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:58.734461+00:00", "epoch": 0, "step": 1227, "train_loss": 3.970705986022949, "perplexity": 53.02195038516533, "lr": 0.001305, "grad_norm": 0.268236, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:59.040358+00:00", "epoch": 0, "step": 1228, "train_loss": 3.88165020942688, "perplexity": 48.504191138025575, "lr": 0.001305, "grad_norm": 0.256741, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:59.346510+00:00", "epoch": 0, "step": 1229, "train_loss": 3.9924142360687256, "perplexity": 54.185548284195725, "lr": 0.001305, "grad_norm": 0.255962, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:59.652128+00:00", "epoch": 0, "step": 1230, "train_loss": 3.9109952449798584, "perplexity": 49.94863837582963, "lr": 0.001305, "grad_norm": 0.253934, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:50:59.959189+00:00", "epoch": 0, "step": 1231, "train_loss": 3.9951179027557373, "perplexity": 54.33224616773839, "lr": 0.001305, "grad_norm": 0.289875, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:00.266215+00:00", "epoch": 0, "step": 1232, "train_loss": 3.903773784637451, "perplexity": 49.58923553321698, "lr": 0.001305, "grad_norm": 0.256412, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:00.573207+00:00", "epoch": 0, "step": 1233, "train_loss": 4.036396503448486, "perplexity": 56.62193778271809, "lr": 0.001305, "grad_norm": 0.270265, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:00.879313+00:00", "epoch": 0, "step": 1234, "train_loss": 3.8029234409332275, "perplexity": 44.83205697158379, "lr": 0.001305, "grad_norm": 0.303881, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:01.185360+00:00", "epoch": 0, "step": 1235, "train_loss": 4.001848220825195, "perplexity": 54.69915277999125, "lr": 0.001305, "grad_norm": 0.262834, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:01.492395+00:00", "epoch": 0, "step": 1236, "train_loss": 3.9622445106506348, "perplexity": 52.57519921030529, "lr": 0.001305, "grad_norm": 0.284128, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:01.798760+00:00", "epoch": 0, "step": 1237, "train_loss": 3.9721250534057617, "perplexity": 53.097245517327615, "lr": 0.001305, "grad_norm": 0.266411, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:02.104820+00:00", "epoch": 0, "step": 1238, "train_loss": 3.9944674968719482, "perplexity": 54.296919644685985, "lr": 0.001305, "grad_norm": 0.266028, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:02.410118+00:00", "epoch": 0, "step": 1239, "train_loss": 3.9074461460113525, "perplexity": 49.77167992217325, "lr": 0.001305, "grad_norm": 0.267649, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:02.715868+00:00", "epoch": 0, "step": 1240, "train_loss": 3.933788299560547, "perplexity": 51.10019430049978, "lr": 0.001305, "grad_norm": 0.267779, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:03.021880+00:00", "epoch": 0, "step": 1241, "train_loss": 3.962116003036499, "perplexity": 52.56844333099233, "lr": 0.001305, "grad_norm": 0.298284, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:03.329061+00:00", "epoch": 0, "step": 1242, "train_loss": 4.009883880615234, "perplexity": 55.14046731456009, "lr": 0.001305, "grad_norm": 0.31732, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:03.636618+00:00", "epoch": 0, "step": 1243, "train_loss": 4.012483596801758, "perplexity": 55.284003375639905, "lr": 0.001305, "grad_norm": 0.306333, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:03.943998+00:00", "epoch": 0, "step": 1244, "train_loss": 3.889354705810547, "perplexity": 48.8793347940259, "lr": 0.001305, "grad_norm": 0.265869, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:04.250744+00:00", "epoch": 0, "step": 1245, "train_loss": 3.930076837539673, "perplexity": 50.91088938639915, "lr": 0.001305, "grad_norm": 0.28121, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:04.556344+00:00", "epoch": 0, "step": 1246, "train_loss": 3.9018607139587402, "perplexity": 49.49445850721914, "lr": 0.001305, "grad_norm": 0.298596, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:04.862201+00:00", "epoch": 0, "step": 1247, "train_loss": 3.8885042667388916, "perplexity": 48.83778356881755, "lr": 0.001305, "grad_norm": 0.261901, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:05.169133+00:00", "epoch": 0, "step": 1248, "train_loss": 3.951885461807251, "perplexity": 52.03338135772729, "lr": 0.001305, "grad_norm": 0.271742, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:05.475432+00:00", "epoch": 0, "step": 1249, "train_loss": 3.8910577297210693, "perplexity": 48.962648392299386, "lr": 0.001305, "grad_norm": 0.334583, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:05.781276+00:00", "epoch": 0, "step": 1250, "train_loss": 3.8986148834228516, "perplexity": 49.33406832293279, "lr": 0.001305, "grad_norm": 0.337857, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:51:06.789618+00:00", "step": 1250, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0000_step_0001250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:07.782865+00:00", "epoch": 0, "step": 1251, "train_loss": 3.983607769012451, "perplexity": 53.71046003322703, "lr": 0.001305, "grad_norm": 0.34464, "tokens_per_sec": 16370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:08.089588+00:00", "epoch": 0, "step": 1252, "train_loss": 3.9500608444213867, "perplexity": 51.93852690831145, "lr": 0.001305, "grad_norm": 0.32008, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:08.396368+00:00", "epoch": 0, "step": 1253, "train_loss": 3.8832955360412598, "perplexity": 48.5840620634754, "lr": 0.001305, "grad_norm": 0.30388, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:08.701877+00:00", "epoch": 0, "step": 1254, "train_loss": 3.878319263458252, "perplexity": 48.34289508145183, "lr": 0.001305, "grad_norm": 0.285803, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:09.008723+00:00", "epoch": 0, "step": 1255, "train_loss": 4.021191596984863, "perplexity": 55.76751865735839, "lr": 0.001305, "grad_norm": 0.285248, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:09.314585+00:00", "epoch": 0, "step": 1256, "train_loss": 3.874166250228882, "perplexity": 48.14254271960217, "lr": 0.001305, "grad_norm": 0.274863, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:09.620184+00:00", "epoch": 0, "step": 1257, "train_loss": 3.853470802307129, "perplexity": 47.15645024184892, "lr": 0.001305, "grad_norm": 0.271821, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:09.926232+00:00", "epoch": 0, "step": 1258, "train_loss": 3.957022190093994, "perplexity": 52.30135035234133, "lr": 0.001305, "grad_norm": 0.297807, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:10.233022+00:00", "epoch": 0, "step": 1259, "train_loss": 3.9242911338806152, "perplexity": 50.61718453131597, "lr": 0.001305, "grad_norm": 0.310386, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:10.540720+00:00", "epoch": 0, "step": 1260, "train_loss": 3.848013401031494, "perplexity": 46.89979953009576, "lr": 0.001305, "grad_norm": 0.318235, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:10.847270+00:00", "epoch": 0, "step": 1261, "train_loss": 3.9506423473358154, "perplexity": 51.96873809617535, "lr": 0.001305, "grad_norm": 0.295858, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:11.153286+00:00", "epoch": 0, "step": 1262, "train_loss": 3.859393358230591, "perplexity": 47.436565636310156, "lr": 0.001305, "grad_norm": 0.262001, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:11.459797+00:00", "epoch": 0, "step": 1263, "train_loss": 3.8529629707336426, "perplexity": 47.132508787150414, "lr": 0.001305, "grad_norm": 0.233029, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:11.766465+00:00", "epoch": 0, "step": 1264, "train_loss": 3.8370471000671387, "perplexity": 46.38829201211187, "lr": 0.001305, "grad_norm": 0.243216, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:12.523920+00:00", "epoch": 0, "step": 1265, "train_loss": 3.8973019123077393, "perplexity": 49.26933662096047, "lr": 0.001305, "grad_norm": 0.240173, "tokens_per_sec": 43261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:12.830038+00:00", "epoch": 0, "step": 1266, "train_loss": 4.009039402008057, "perplexity": 55.09392202553971, "lr": 0.001305, "grad_norm": 0.26336, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:13.136503+00:00", "epoch": 0, "step": 1267, "train_loss": 3.9000024795532227, "perplexity": 49.40257160168393, "lr": 0.001305, "grad_norm": 0.266375, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:13.442954+00:00", "epoch": 0, "step": 1268, "train_loss": 3.843909502029419, "perplexity": 46.70772189279937, "lr": 0.001305, "grad_norm": 0.261243, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:13.749177+00:00", "epoch": 0, "step": 1269, "train_loss": 3.84293270111084, "perplexity": 46.6621200227499, "lr": 0.001305, "grad_norm": 0.263865, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:14.056480+00:00", "epoch": 0, "step": 1270, "train_loss": 3.9597666263580322, "perplexity": 52.445085220281705, "lr": 0.001305, "grad_norm": 0.271966, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:14.362786+00:00", "epoch": 0, "step": 1271, "train_loss": 3.877955913543701, "perplexity": 48.32533288545927, "lr": 0.001305, "grad_norm": 0.263718, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:14.669482+00:00", "epoch": 0, "step": 1272, "train_loss": 3.898071765899658, "perplexity": 49.30728140081804, "lr": 0.001305, "grad_norm": 0.272647, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:14.976568+00:00", "epoch": 0, "step": 1273, "train_loss": 3.9280686378479004, "perplexity": 50.80875274373796, "lr": 0.001305, "grad_norm": 0.277886, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:15.283231+00:00", "epoch": 0, "step": 1274, "train_loss": 3.9982645511627197, "perplexity": 54.50347990849093, "lr": 0.001305, "grad_norm": 0.265865, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:15.590013+00:00", "epoch": 0, "step": 1275, "train_loss": 3.8364574909210205, "perplexity": 46.3609491124727, "lr": 0.001305, "grad_norm": 0.256579, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:15.896505+00:00", "epoch": 0, "step": 1276, "train_loss": 4.015187740325928, "perplexity": 55.4337015668345, "lr": 0.001305, "grad_norm": 0.263764, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:16.202547+00:00", "epoch": 0, "step": 1277, "train_loss": 3.8152077198028564, "perplexity": 45.38618301285234, "lr": 0.001305, "grad_norm": 0.259987, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:16.508925+00:00", "epoch": 0, "step": 1278, "train_loss": 3.958775281906128, "perplexity": 52.393119838071414, "lr": 0.001305, "grad_norm": 0.260626, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:16.815778+00:00", "epoch": 0, "step": 1279, "train_loss": 3.9012794494628906, "perplexity": 49.46569749543489, "lr": 0.001305, "grad_norm": 0.24062, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:17.122039+00:00", "epoch": 0, "step": 1280, "train_loss": 3.9995243549346924, "perplexity": 54.57218686762132, "lr": 0.001305, "grad_norm": 0.228449, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:17.427484+00:00", "epoch": 0, "step": 1281, "train_loss": 3.931875705718994, "perplexity": 51.00255378666542, "lr": 0.001305, "grad_norm": 0.241337, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:17.732738+00:00", "epoch": 0, "step": 1282, "train_loss": 3.9930777549743652, "perplexity": 54.22151335032591, "lr": 0.001305, "grad_norm": 0.251818, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:18.039037+00:00", "epoch": 0, "step": 1283, "train_loss": 3.9138107299804688, "perplexity": 50.08946617422675, "lr": 0.001305, "grad_norm": 0.247355, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:18.346379+00:00", "epoch": 0, "step": 1284, "train_loss": 3.9070487022399902, "perplexity": 49.7519024084831, "lr": 0.001305, "grad_norm": 0.223504, "tokens_per_sec": 106619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:18.653339+00:00", "epoch": 0, "step": 1285, "train_loss": 3.920520305633545, "perplexity": 50.426675236752644, "lr": 0.001305, "grad_norm": 0.250378, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:18.959798+00:00", "epoch": 0, "step": 1286, "train_loss": 3.875847816467285, "perplexity": 48.223565697730464, "lr": 0.001305, "grad_norm": 0.253471, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:19.267207+00:00", "epoch": 0, "step": 1287, "train_loss": 3.923142433166504, "perplexity": 50.55907391754012, "lr": 0.001305, "grad_norm": 0.263397, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:19.573872+00:00", "epoch": 0, "step": 1288, "train_loss": 3.937821388244629, "perplexity": 51.306702068039904, "lr": 0.001305, "grad_norm": 0.248371, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:19.880537+00:00", "epoch": 0, "step": 1289, "train_loss": 4.023877143859863, "perplexity": 55.917486225149155, "lr": 0.001305, "grad_norm": 0.246641, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:20.186927+00:00", "epoch": 0, "step": 1290, "train_loss": 3.809155225753784, "perplexity": 45.11231304448267, "lr": 0.001305, "grad_norm": 0.244668, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:20.492638+00:00", "epoch": 0, "step": 1291, "train_loss": 3.8849706649780273, "perplexity": 48.66551483460383, "lr": 0.001305, "grad_norm": 0.227646, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:20.798796+00:00", "epoch": 0, "step": 1292, "train_loss": 3.9459428787231445, "perplexity": 51.72508560968064, "lr": 0.001305, "grad_norm": 0.234288, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:21.105824+00:00", "epoch": 0, "step": 1293, "train_loss": 3.9156651496887207, "perplexity": 50.182439246381186, "lr": 0.001305, "grad_norm": 0.226688, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:21.412385+00:00", "epoch": 0, "step": 1294, "train_loss": 3.922607183456421, "perplexity": 50.532019428983176, "lr": 0.001305, "grad_norm": 0.228174, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:21.719252+00:00", "epoch": 0, "step": 1295, "train_loss": 3.9202542304992676, "perplexity": 50.41325973721243, "lr": 0.001305, "grad_norm": 0.244424, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:22.025727+00:00", "epoch": 0, "step": 1296, "train_loss": 3.860290288925171, "perplexity": 47.47913203477583, "lr": 0.001305, "grad_norm": 0.270019, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:22.332743+00:00", "epoch": 0, "step": 1297, "train_loss": 3.8986029624938965, "perplexity": 49.33348021851463, "lr": 0.001305, "grad_norm": 0.344054, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:22.639944+00:00", "epoch": 0, "step": 1298, "train_loss": 3.828907012939453, "perplexity": 46.012219979558424, "lr": 0.001305, "grad_norm": 0.379518, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:22.947548+00:00", "epoch": 0, "step": 1299, "train_loss": 3.8429038524627686, "perplexity": 46.660773903088064, "lr": 0.001305, "grad_norm": 0.361234, "tokens_per_sec": 106527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:23.254276+00:00", "epoch": 0, "step": 1300, "train_loss": 3.866915702819824, "perplexity": 47.79474531565795, "lr": 0.001305, "grad_norm": 0.331578, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:51:29.009799+00:00", "step": 1300, "epoch": 0, "val_loss": 3.8602791905403135, "val_ppl": 47.478605096019905, "eval_train_loss": 3.866915702819824, "eval_train_ppl": 47.79474531565795} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:51:29.945665+00:00", "step": 1300, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p8603_epoch_0000_step_0001300.pt", "val_loss": 3.8602791905403135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:30.938694+00:00", "epoch": 0, "step": 1301, "train_loss": 3.848067283630371, "perplexity": 46.90232668126538, "lr": 0.001305, "grad_norm": 0.291308, "tokens_per_sec": 4264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:31.243703+00:00", "epoch": 0, "step": 1302, "train_loss": 3.9440174102783203, "perplexity": 51.625586411557016, "lr": 0.001305, "grad_norm": 0.299187, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:31.551395+00:00", "epoch": 0, "step": 1303, "train_loss": 3.9702141284942627, "perplexity": 52.99587755226796, "lr": 0.001305, "grad_norm": 0.314121, "tokens_per_sec": 106495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:31.857267+00:00", "epoch": 0, "step": 1304, "train_loss": 3.863654136657715, "perplexity": 47.639113531255774, "lr": 0.001305, "grad_norm": 0.339352, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:32.162811+00:00", "epoch": 0, "step": 1305, "train_loss": 3.8824265003204346, "perplexity": 48.54185911866879, "lr": 0.001305, "grad_norm": 0.324867, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:32.468128+00:00", "epoch": 0, "step": 1306, "train_loss": 4.042934894561768, "perplexity": 56.993367109404474, "lr": 0.001305, "grad_norm": 0.313002, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:32.774978+00:00", "epoch": 0, "step": 1307, "train_loss": 3.940974473953247, "perplexity": 51.4687318096774, "lr": 0.001305, "grad_norm": 0.293353, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:33.081910+00:00", "epoch": 0, "step": 1308, "train_loss": 3.832843542098999, "perplexity": 46.193705402170394, "lr": 0.001305, "grad_norm": 0.276846, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:33.388425+00:00", "epoch": 0, "step": 1309, "train_loss": 3.8744113445281982, "perplexity": 48.15434362848601, "lr": 0.001305, "grad_norm": 0.286928, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:33.694115+00:00", "epoch": 0, "step": 1310, "train_loss": 3.8565807342529297, "perplexity": 47.30333187045853, "lr": 0.001305, "grad_norm": 0.308394, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:34.000470+00:00", "epoch": 0, "step": 1311, "train_loss": 3.9262828826904297, "perplexity": 50.71810171584139, "lr": 0.001305, "grad_norm": 0.327684, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:34.306068+00:00", "epoch": 0, "step": 1312, "train_loss": 3.9576568603515625, "perplexity": 52.33455499972709, "lr": 0.001305, "grad_norm": 0.335765, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:34.613148+00:00", "epoch": 0, "step": 1313, "train_loss": 3.889803886413574, "perplexity": 48.90129537486849, "lr": 0.001305, "grad_norm": 0.288507, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:34.919234+00:00", "epoch": 0, "step": 1314, "train_loss": 3.97721266746521, "perplexity": 53.368072156679894, "lr": 0.001305, "grad_norm": 0.289333, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:35.245367+00:00", "epoch": 0, "step": 1315, "train_loss": 3.8376386165618896, "perplexity": 46.415739569041385, "lr": 0.001305, "grad_norm": 0.284377, "tokens_per_sec": 100475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:35.551938+00:00", "epoch": 0, "step": 1316, "train_loss": 3.90511155128479, "perplexity": 49.65561875130784, "lr": 0.001305, "grad_norm": 0.280959, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:35.858236+00:00", "epoch": 0, "step": 1317, "train_loss": 3.9595258235931396, "perplexity": 52.4324578191732, "lr": 0.001305, "grad_norm": 0.263283, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:36.164452+00:00", "epoch": 0, "step": 1318, "train_loss": 3.989006519317627, "perplexity": 54.001213542316805, "lr": 0.001305, "grad_norm": 0.307529, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:36.471058+00:00", "epoch": 0, "step": 1319, "train_loss": 3.8348023891448975, "perplexity": 46.28428048794414, "lr": 0.001305, "grad_norm": 0.292656, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:36.777889+00:00", "epoch": 0, "step": 1320, "train_loss": 3.936459541320801, "perplexity": 51.236877749465016, "lr": 0.001305, "grad_norm": 0.275998, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:37.085232+00:00", "epoch": 0, "step": 1321, "train_loss": 3.7916345596313477, "perplexity": 44.3287991553265, "lr": 0.001305, "grad_norm": 0.278798, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:37.391194+00:00", "epoch": 0, "step": 1322, "train_loss": 3.8714516162872314, "perplexity": 48.01203056560761, "lr": 0.001305, "grad_norm": 0.252852, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:37.697930+00:00", "epoch": 0, "step": 1323, "train_loss": 4.006923675537109, "perplexity": 54.977481577825905, "lr": 0.001305, "grad_norm": 0.267419, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:38.003737+00:00", "epoch": 0, "step": 1324, "train_loss": 3.9097328186035156, "perplexity": 49.885621682618655, "lr": 0.001305, "grad_norm": 0.268821, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:38.311345+00:00", "epoch": 0, "step": 1325, "train_loss": 3.8648681640625, "perplexity": 47.69698384158598, "lr": 0.001305, "grad_norm": 0.249758, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:38.618422+00:00", "epoch": 0, "step": 1326, "train_loss": 3.8885691165924072, "perplexity": 48.84095079462397, "lr": 0.001305, "grad_norm": 0.246113, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:38.925620+00:00", "epoch": 0, "step": 1327, "train_loss": 4.02262544631958, "perplexity": 55.84753823118088, "lr": 0.001305, "grad_norm": 0.23262, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:39.231198+00:00", "epoch": 0, "step": 1328, "train_loss": 3.9173619747161865, "perplexity": 50.267662349132586, "lr": 0.001305, "grad_norm": 0.253927, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:39.537996+00:00", "epoch": 0, "step": 1329, "train_loss": 3.834462881088257, "perplexity": 46.26856926901628, "lr": 0.001305, "grad_norm": 0.267497, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:39.845688+00:00", "epoch": 0, "step": 1330, "train_loss": 3.9221885204315186, "perplexity": 50.51086796885071, "lr": 0.001305, "grad_norm": 0.301115, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:40.152652+00:00", "epoch": 0, "step": 1331, "train_loss": 4.048919677734375, "perplexity": 57.33548277672844, "lr": 0.001305, "grad_norm": 0.302545, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:40.460560+00:00", "epoch": 0, "step": 1332, "train_loss": 3.8934433460235596, "perplexity": 49.07959392263597, "lr": 0.001305, "grad_norm": 0.268694, "tokens_per_sec": 106422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:40.767408+00:00", "epoch": 0, "step": 1333, "train_loss": 3.9533584117889404, "perplexity": 52.110080398903975, "lr": 0.001305, "grad_norm": 0.235221, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:41.073976+00:00", "epoch": 0, "step": 1334, "train_loss": 3.9709763526916504, "perplexity": 53.036287691336526, "lr": 0.001305, "grad_norm": 0.249872, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:41.380271+00:00", "epoch": 0, "step": 1335, "train_loss": 3.853764533996582, "perplexity": 47.17030362013558, "lr": 0.001305, "grad_norm": 0.241672, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:41.687049+00:00", "epoch": 0, "step": 1336, "train_loss": 3.902029037475586, "perplexity": 49.50279028973724, "lr": 0.001305, "grad_norm": 0.248651, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:41.994261+00:00", "epoch": 0, "step": 1337, "train_loss": 3.9450631141662598, "perplexity": 51.679599724028215, "lr": 0.001305, "grad_norm": 0.247903, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:42.300885+00:00", "epoch": 0, "step": 1338, "train_loss": 3.9057059288024902, "perplexity": 49.68514170774272, "lr": 0.001305, "grad_norm": 0.244874, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:42.607773+00:00", "epoch": 0, "step": 1339, "train_loss": 3.8563122749328613, "perplexity": 47.29063455458027, "lr": 0.001305, "grad_norm": 0.248787, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:42.914249+00:00", "epoch": 0, "step": 1340, "train_loss": 3.9363675117492676, "perplexity": 51.232162658526256, "lr": 0.001305, "grad_norm": 0.225868, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:43.221791+00:00", "epoch": 0, "step": 1341, "train_loss": 3.897958278656006, "perplexity": 49.30168597087084, "lr": 0.001305, "grad_norm": 0.239625, "tokens_per_sec": 106473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:43.528288+00:00", "epoch": 0, "step": 1342, "train_loss": 3.9059081077575684, "perplexity": 49.69518801331766, "lr": 0.001305, "grad_norm": 0.249734, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:43.835923+00:00", "epoch": 0, "step": 1343, "train_loss": 3.839766263961792, "perplexity": 46.51460103052121, "lr": 0.001305, "grad_norm": 0.284323, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:44.142478+00:00", "epoch": 0, "step": 1344, "train_loss": 3.910651206970215, "perplexity": 49.9314571013736, "lr": 0.001305, "grad_norm": 0.280256, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:44.448656+00:00", "epoch": 0, "step": 1345, "train_loss": 3.7959096431732178, "perplexity": 44.51871413698673, "lr": 0.001305, "grad_norm": 0.244738, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:44.754947+00:00", "epoch": 0, "step": 1346, "train_loss": 3.833936929702759, "perplexity": 46.24424064930068, "lr": 0.001305, "grad_norm": 0.256177, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:45.062456+00:00", "epoch": 0, "step": 1347, "train_loss": 3.906038522720337, "perplexity": 49.70166943204, "lr": 0.001305, "grad_norm": 0.279768, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:45.369721+00:00", "epoch": 0, "step": 1348, "train_loss": 3.8877081871032715, "perplexity": 48.798920275060404, "lr": 0.001305, "grad_norm": 0.297499, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:45.677443+00:00", "epoch": 0, "step": 1349, "train_loss": 4.047285556793213, "perplexity": 57.241866174908516, "lr": 0.001305, "grad_norm": 0.263716, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:45.983735+00:00", "epoch": 0, "step": 1350, "train_loss": 3.8879284858703613, "perplexity": 48.809671801262816, "lr": 0.001305, "grad_norm": 0.276925, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:46.291194+00:00", "epoch": 0, "step": 1351, "train_loss": 3.9051177501678467, "perplexity": 49.655926561635624, "lr": 0.001305, "grad_norm": 0.292537, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:46.598364+00:00", "epoch": 0, "step": 1352, "train_loss": 3.9305100440979004, "perplexity": 50.932949095426856, "lr": 0.001305, "grad_norm": 0.283275, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:46.906225+00:00", "epoch": 0, "step": 1353, "train_loss": 3.9149107933044434, "perplexity": 50.14459807761516, "lr": 0.001305, "grad_norm": 0.290732, "tokens_per_sec": 106438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:47.214776+00:00", "epoch": 0, "step": 1354, "train_loss": 3.76183819770813, "perplexity": 43.02744628218469, "lr": 0.001305, "grad_norm": 0.259755, "tokens_per_sec": 106200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:47.536781+00:00", "epoch": 0, "step": 1355, "train_loss": 3.9039859771728516, "perplexity": 49.59975911530664, "lr": 0.001305, "grad_norm": 0.245057, "tokens_per_sec": 101762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:47.843861+00:00", "epoch": 0, "step": 1356, "train_loss": 3.8238768577575684, "perplexity": 45.78135250919215, "lr": 0.001305, "grad_norm": 0.251923, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:48.151590+00:00", "epoch": 0, "step": 1357, "train_loss": 3.9142675399780273, "perplexity": 50.11235277016244, "lr": 0.001305, "grad_norm": 0.253584, "tokens_per_sec": 106483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:48.458903+00:00", "epoch": 0, "step": 1358, "train_loss": 3.887559175491333, "perplexity": 48.7916492110393, "lr": 0.001305, "grad_norm": 0.257268, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:48.767311+00:00", "epoch": 0, "step": 1359, "train_loss": 3.858999490737915, "perplexity": 47.41788559411401, "lr": 0.001305, "grad_norm": 0.252832, "tokens_per_sec": 106248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:49.086642+00:00", "epoch": 0, "step": 1360, "train_loss": 3.9089603424072266, "perplexity": 49.84710110736097, "lr": 0.001305, "grad_norm": 0.242242, "tokens_per_sec": 102669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:49.395185+00:00", "epoch": 0, "step": 1361, "train_loss": 3.8403923511505127, "perplexity": 46.5437323447346, "lr": 0.001305, "grad_norm": 0.245363, "tokens_per_sec": 106146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:49.716038+00:00", "epoch": 0, "step": 1362, "train_loss": 3.9812073707580566, "perplexity": 53.581688152450475, "lr": 0.001305, "grad_norm": 0.273447, "tokens_per_sec": 102128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:50.024423+00:00", "epoch": 0, "step": 1363, "train_loss": 3.8464741706848145, "perplexity": 46.82766546511861, "lr": 0.001305, "grad_norm": 0.274585, "tokens_per_sec": 106256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:50.343800+00:00", "epoch": 0, "step": 1364, "train_loss": 3.938995599746704, "perplexity": 51.366982371728646, "lr": 0.001305, "grad_norm": 0.261524, "tokens_per_sec": 102600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:50.651298+00:00", "epoch": 0, "step": 1365, "train_loss": 3.8482048511505127, "perplexity": 46.90877936186525, "lr": 0.001305, "grad_norm": 0.250868, "tokens_per_sec": 106563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:50.967613+00:00", "epoch": 0, "step": 1366, "train_loss": 3.863177537918091, "perplexity": 47.61641419945611, "lr": 0.001305, "grad_norm": 0.264934, "tokens_per_sec": 103593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:51.275036+00:00", "epoch": 0, "step": 1367, "train_loss": 3.8596389293670654, "perplexity": 47.4482161180964, "lr": 0.001305, "grad_norm": 0.238694, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:51.595886+00:00", "epoch": 0, "step": 1368, "train_loss": 3.981682538986206, "perplexity": 53.60715451819591, "lr": 0.001305, "grad_norm": 0.232245, "tokens_per_sec": 102129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:51.904359+00:00", "epoch": 0, "step": 1369, "train_loss": 3.8929574489593506, "perplexity": 49.05575208484497, "lr": 0.001305, "grad_norm": 0.236807, "tokens_per_sec": 106227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:52.226755+00:00", "epoch": 0, "step": 1370, "train_loss": 3.900524377822876, "perplexity": 49.42836144757186, "lr": 0.001305, "grad_norm": 0.223178, "tokens_per_sec": 101698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:52.534281+00:00", "epoch": 0, "step": 1371, "train_loss": 3.783362865447998, "perplexity": 43.96363722117923, "lr": 0.001305, "grad_norm": 0.225542, "tokens_per_sec": 106488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:52.851681+00:00", "epoch": 0, "step": 1372, "train_loss": 3.837350368499756, "perplexity": 46.40236225014385, "lr": 0.001305, "grad_norm": 0.224666, "tokens_per_sec": 103239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:53.159999+00:00", "epoch": 0, "step": 1373, "train_loss": 3.785240888595581, "perplexity": 44.04627952732476, "lr": 0.001305, "grad_norm": 0.231336, "tokens_per_sec": 106281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:53.477827+00:00", "epoch": 0, "step": 1374, "train_loss": 3.889904737472534, "perplexity": 48.906227370985896, "lr": 0.001305, "grad_norm": 0.23474, "tokens_per_sec": 103099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:53.785767+00:00", "epoch": 0, "step": 1375, "train_loss": 3.8516628742218018, "perplexity": 47.07127179250898, "lr": 0.001305, "grad_norm": 0.231626, "tokens_per_sec": 106410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:54.107461+00:00", "epoch": 0, "step": 1376, "train_loss": 3.9027323722839355, "perplexity": 49.537619572146, "lr": 0.001305, "grad_norm": 0.265191, "tokens_per_sec": 101862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:54.415721+00:00", "epoch": 0, "step": 1377, "train_loss": 3.9251785278320312, "perplexity": 50.662121850310406, "lr": 0.001305, "grad_norm": 0.280302, "tokens_per_sec": 106300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:54.723849+00:00", "epoch": 0, "step": 1378, "train_loss": 3.9050631523132324, "perplexity": 49.65321552858544, "lr": 0.001305, "grad_norm": 0.282651, "tokens_per_sec": 106346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:55.031206+00:00", "epoch": 0, "step": 1379, "train_loss": 3.892186164855957, "perplexity": 49.01793075045383, "lr": 0.001305, "grad_norm": 0.251595, "tokens_per_sec": 106611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:55.338439+00:00", "epoch": 0, "step": 1380, "train_loss": 3.898893117904663, "perplexity": 49.3477966716297, "lr": 0.001305, "grad_norm": 0.246145, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:55.656904+00:00", "epoch": 0, "step": 1381, "train_loss": 3.9250049591064453, "perplexity": 50.653329253467355, "lr": 0.001305, "grad_norm": 0.29715, "tokens_per_sec": 102838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:55.964602+00:00", "epoch": 0, "step": 1382, "train_loss": 3.835597515106201, "perplexity": 46.32109695588541, "lr": 0.001305, "grad_norm": 0.340396, "tokens_per_sec": 106494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:56.285107+00:00", "epoch": 0, "step": 1383, "train_loss": 3.84291672706604, "perplexity": 46.661374645907564, "lr": 0.001305, "grad_norm": 0.343224, "tokens_per_sec": 102239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:56.593165+00:00", "epoch": 0, "step": 1384, "train_loss": 3.8930816650390625, "perplexity": 49.0618459765281, "lr": 0.001305, "grad_norm": 0.32758, "tokens_per_sec": 106371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:56.900551+00:00", "epoch": 0, "step": 1385, "train_loss": 3.8803462982177734, "perplexity": 48.44098719463205, "lr": 0.001305, "grad_norm": 0.315608, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:57.220029+00:00", "epoch": 0, "step": 1386, "train_loss": 3.862842082977295, "perplexity": 47.60044371688832, "lr": 0.001305, "grad_norm": 0.336148, "tokens_per_sec": 102567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:57.527637+00:00", "epoch": 0, "step": 1387, "train_loss": 3.959646701812744, "perplexity": 52.43879614439889, "lr": 0.001305, "grad_norm": 0.399225, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:57.835348+00:00", "epoch": 0, "step": 1388, "train_loss": 3.8484511375427246, "perplexity": 46.92033377868676, "lr": 0.001305, "grad_norm": 0.416908, "tokens_per_sec": 106490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:58.155068+00:00", "epoch": 0, "step": 1389, "train_loss": 3.991847038269043, "perplexity": 54.15482307489361, "lr": 0.001305, "grad_norm": 0.37197, "tokens_per_sec": 102489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:58.464834+00:00", "epoch": 0, "step": 1390, "train_loss": 3.886200189590454, "perplexity": 48.72538708252805, "lr": 0.001305, "grad_norm": 0.270808, "tokens_per_sec": 105839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:58.781803+00:00", "epoch": 0, "step": 1391, "train_loss": 3.869542121887207, "perplexity": 47.92043933640816, "lr": 0.001305, "grad_norm": 0.267994, "tokens_per_sec": 103326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:59.089246+00:00", "epoch": 0, "step": 1392, "train_loss": 3.9550111293792725, "perplexity": 52.196274853338544, "lr": 0.001305, "grad_norm": 0.266752, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:59.396448+00:00", "epoch": 0, "step": 1393, "train_loss": 3.8962485790252686, "perplexity": 49.21746691173374, "lr": 0.001305, "grad_norm": 0.251789, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:51:59.703244+00:00", "epoch": 0, "step": 1394, "train_loss": 3.833303451538086, "perplexity": 46.214955209430165, "lr": 0.001305, "grad_norm": 0.257023, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:00.020967+00:00", "epoch": 0, "step": 1395, "train_loss": 3.869717597961426, "perplexity": 47.92884896480046, "lr": 0.001305, "grad_norm": 0.265321, "tokens_per_sec": 103135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:00.328862+00:00", "epoch": 0, "step": 1396, "train_loss": 3.9187564849853516, "perplexity": 50.33781001994125, "lr": 0.001305, "grad_norm": 0.306503, "tokens_per_sec": 106425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:00.637009+00:00", "epoch": 0, "step": 1397, "train_loss": 3.914937734603882, "perplexity": 50.14594905644567, "lr": 0.001305, "grad_norm": 0.290518, "tokens_per_sec": 106339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:00.943987+00:00", "epoch": 0, "step": 1398, "train_loss": 3.8986899852752686, "perplexity": 49.33777354198377, "lr": 0.001305, "grad_norm": 0.235912, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:01.251314+00:00", "epoch": 0, "step": 1399, "train_loss": 4.113162994384766, "perplexity": 61.13979688849317, "lr": 0.001305, "grad_norm": 0.246404, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:01.570370+00:00", "epoch": 0, "step": 1400, "train_loss": 3.9330170154571533, "perplexity": 51.06079672826914, "lr": 0.001305, "grad_norm": 0.228393, "tokens_per_sec": 102756} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:52:07.245775+00:00", "step": 1400, "epoch": 0, "val_loss": 3.827421176433563, "val_ppl": 45.943904109071376, "eval_train_loss": 3.9330170154571533, "eval_train_ppl": 51.06079672826914} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:52:08.186256+00:00", "step": 1400, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p8274_epoch_0000_step_0001400.pt", "val_loss": 3.827421176433563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:09.272022+00:00", "epoch": 0, "step": 1401, "train_loss": 3.8617336750030518, "perplexity": 47.547712234891264, "lr": 0.001305, "grad_norm": 0.243481, "tokens_per_sec": 4255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:09.577186+00:00", "epoch": 0, "step": 1402, "train_loss": 3.967297315597534, "perplexity": 52.841523713218514, "lr": 0.001305, "grad_norm": 0.257914, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:09.883606+00:00", "epoch": 0, "step": 1403, "train_loss": 3.7934606075286865, "perplexity": 44.40981961692418, "lr": 0.001305, "grad_norm": 0.259954, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:10.190611+00:00", "epoch": 0, "step": 1404, "train_loss": 3.858492851257324, "perplexity": 47.393867905854194, "lr": 0.001305, "grad_norm": 0.27764, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:10.497534+00:00", "epoch": 0, "step": 1405, "train_loss": 3.776986598968506, "perplexity": 43.684205168198815, "lr": 0.001305, "grad_norm": 0.26989, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:10.803088+00:00", "epoch": 0, "step": 1406, "train_loss": 3.791888952255249, "perplexity": 44.34007750936223, "lr": 0.001305, "grad_norm": 0.25, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:11.109173+00:00", "epoch": 0, "step": 1407, "train_loss": 3.8081765174865723, "perplexity": 45.068182849568856, "lr": 0.001305, "grad_norm": 0.253208, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:11.416225+00:00", "epoch": 0, "step": 1408, "train_loss": 3.9172778129577637, "perplexity": 50.263431912300476, "lr": 0.001305, "grad_norm": 0.280634, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:11.722961+00:00", "epoch": 0, "step": 1409, "train_loss": 3.797305107116699, "perplexity": 44.58088176362741, "lr": 0.001305, "grad_norm": 0.278621, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:12.030390+00:00", "epoch": 0, "step": 1410, "train_loss": 3.8772759437561035, "perplexity": 48.29248428841399, "lr": 0.001305, "grad_norm": 0.266463, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:12.336943+00:00", "epoch": 0, "step": 1411, "train_loss": 3.744220495223999, "perplexity": 42.27603999636587, "lr": 0.001305, "grad_norm": 0.265537, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:12.643410+00:00", "epoch": 0, "step": 1412, "train_loss": 3.8547232151031494, "perplexity": 47.215546582328656, "lr": 0.001305, "grad_norm": 0.282646, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:12.949203+00:00", "epoch": 0, "step": 1413, "train_loss": 3.9489572048187256, "perplexity": 51.881237112570275, "lr": 0.001305, "grad_norm": 0.267952, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:13.255566+00:00", "epoch": 0, "step": 1414, "train_loss": 3.7588512897491455, "perplexity": 42.89911900671713, "lr": 0.001305, "grad_norm": 0.240827, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:13.593958+00:00", "epoch": 0, "step": 1415, "train_loss": 3.8888778686523438, "perplexity": 48.856032866981714, "lr": 0.001305, "grad_norm": 0.253307, "tokens_per_sec": 96836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:13.901393+00:00", "epoch": 0, "step": 1416, "train_loss": 3.9440550804138184, "perplexity": 51.62753119102213, "lr": 0.001305, "grad_norm": 0.256929, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:14.207112+00:00", "epoch": 0, "step": 1417, "train_loss": 3.8972461223602295, "perplexity": 49.26658796393098, "lr": 0.001305, "grad_norm": 0.225611, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:14.512956+00:00", "epoch": 0, "step": 1418, "train_loss": 3.8455252647399902, "perplexity": 46.78325149065134, "lr": 0.001305, "grad_norm": 0.264618, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:14.819408+00:00", "epoch": 0, "step": 1419, "train_loss": 3.876944065093994, "perplexity": 48.27645970259455, "lr": 0.001305, "grad_norm": 0.247754, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:15.126443+00:00", "epoch": 0, "step": 1420, "train_loss": 3.915806770324707, "perplexity": 50.18954661860601, "lr": 0.001305, "grad_norm": 0.271012, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:15.434460+00:00", "epoch": 0, "step": 1421, "train_loss": 3.8600525856018066, "perplexity": 47.46784742854805, "lr": 0.001305, "grad_norm": 0.281237, "tokens_per_sec": 106327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:15.741561+00:00", "epoch": 0, "step": 1422, "train_loss": 3.9074888229370117, "perplexity": 49.77380406978294, "lr": 0.001305, "grad_norm": 0.260808, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:16.047218+00:00", "epoch": 0, "step": 1423, "train_loss": 3.7468104362487793, "perplexity": 42.38567435870035, "lr": 0.001305, "grad_norm": 0.278532, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:16.367817+00:00", "epoch": 0, "step": 1424, "train_loss": 3.891885280609131, "perplexity": 49.00318424588571, "lr": 0.001305, "grad_norm": 0.269006, "tokens_per_sec": 102212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:16.673328+00:00", "epoch": 0, "step": 1425, "train_loss": 3.8534064292907715, "perplexity": 47.15341473660951, "lr": 0.001305, "grad_norm": 0.243225, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:16.980195+00:00", "epoch": 0, "step": 1426, "train_loss": 3.829474687576294, "perplexity": 46.03834736505096, "lr": 0.001305, "grad_norm": 0.233894, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:17.288755+00:00", "epoch": 0, "step": 1427, "train_loss": 3.8547134399414062, "perplexity": 47.21508504497983, "lr": 0.001305, "grad_norm": 0.263592, "tokens_per_sec": 106197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:17.597150+00:00", "epoch": 0, "step": 1428, "train_loss": 3.8301033973693848, "perplexity": 46.06730122573123, "lr": 0.001305, "grad_norm": 0.298794, "tokens_per_sec": 106253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:17.904828+00:00", "epoch": 0, "step": 1429, "train_loss": 3.841357707977295, "perplexity": 46.588685348851186, "lr": 0.001305, "grad_norm": 0.26809, "tokens_per_sec": 106501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:18.211372+00:00", "epoch": 0, "step": 1430, "train_loss": 3.8369829654693604, "perplexity": 46.38531701306325, "lr": 0.001305, "grad_norm": 0.249554, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:18.518596+00:00", "epoch": 0, "step": 1431, "train_loss": 3.9070329666137695, "perplexity": 49.75111953730253, "lr": 0.001305, "grad_norm": 0.269538, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:18.825213+00:00", "epoch": 0, "step": 1432, "train_loss": 3.928954839706421, "perplexity": 50.85379951216538, "lr": 0.001305, "grad_norm": 0.293787, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:19.131535+00:00", "epoch": 0, "step": 1433, "train_loss": 3.910771131515503, "perplexity": 49.93744546773094, "lr": 0.001305, "grad_norm": 0.255196, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:19.438463+00:00", "epoch": 0, "step": 1434, "train_loss": 3.8552820682525635, "perplexity": 47.2419405137164, "lr": 0.001305, "grad_norm": 0.243609, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:19.746287+00:00", "epoch": 0, "step": 1435, "train_loss": 3.902665853500366, "perplexity": 49.53432449954447, "lr": 0.001305, "grad_norm": 0.255093, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:20.053227+00:00", "epoch": 0, "step": 1436, "train_loss": 3.823716402053833, "perplexity": 45.77400721936991, "lr": 0.001305, "grad_norm": 0.227144, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:20.359997+00:00", "epoch": 0, "step": 1437, "train_loss": 3.775568723678589, "perplexity": 43.62231030310036, "lr": 0.001305, "grad_norm": 0.246949, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:20.668004+00:00", "epoch": 0, "step": 1438, "train_loss": 3.838118076324463, "perplexity": 46.43799938443129, "lr": 0.001305, "grad_norm": 0.253512, "tokens_per_sec": 106387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:20.975519+00:00", "epoch": 0, "step": 1439, "train_loss": 3.720888614654541, "perplexity": 41.30107855067892, "lr": 0.001305, "grad_norm": 0.267403, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:21.282884+00:00", "epoch": 0, "step": 1440, "train_loss": 3.8493857383728027, "perplexity": 46.96420605992564, "lr": 0.001305, "grad_norm": 0.290823, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:21.590446+00:00", "epoch": 0, "step": 1441, "train_loss": 3.8066587448120117, "perplexity": 44.99983147719169, "lr": 0.001305, "grad_norm": 0.317222, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:21.897468+00:00", "epoch": 0, "step": 1442, "train_loss": 3.8526222705841064, "perplexity": 47.11645346953848, "lr": 0.001305, "grad_norm": 0.344568, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:22.204038+00:00", "epoch": 0, "step": 1443, "train_loss": 3.8146138191223145, "perplexity": 45.35923613055173, "lr": 0.001305, "grad_norm": 0.305281, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:22.511293+00:00", "epoch": 0, "step": 1444, "train_loss": 3.851464033126831, "perplexity": 47.06191301976924, "lr": 0.001305, "grad_norm": 0.287532, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:22.817864+00:00", "epoch": 0, "step": 1445, "train_loss": 4.041635990142822, "perplexity": 56.9193862304622, "lr": 0.001305, "grad_norm": 0.267922, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:23.124072+00:00", "epoch": 0, "step": 1446, "train_loss": 3.8490428924560547, "perplexity": 46.94810733349368, "lr": 0.001305, "grad_norm": 0.255042, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:23.430874+00:00", "epoch": 0, "step": 1447, "train_loss": 3.901221752166748, "perplexity": 49.46284354077112, "lr": 0.001305, "grad_norm": 0.241521, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:23.738047+00:00", "epoch": 0, "step": 1448, "train_loss": 3.886587142944336, "perplexity": 48.7442451828461, "lr": 0.001305, "grad_norm": 0.236435, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:24.045686+00:00", "epoch": 0, "step": 1449, "train_loss": 3.786688804626465, "perplexity": 44.110101034494896, "lr": 0.001305, "grad_norm": 0.234775, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:24.352349+00:00", "epoch": 0, "step": 1450, "train_loss": 3.817593574523926, "perplexity": 45.49459713060505, "lr": 0.001305, "grad_norm": 0.238083, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:24.659423+00:00", "epoch": 0, "step": 1451, "train_loss": 3.827256917953491, "perplexity": 45.93635805298267, "lr": 0.001305, "grad_norm": 0.264272, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:24.966325+00:00", "epoch": 0, "step": 1452, "train_loss": 3.8193721771240234, "perplexity": 45.57558594145654, "lr": 0.001305, "grad_norm": 0.289245, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:25.272597+00:00", "epoch": 0, "step": 1453, "train_loss": 3.848264694213867, "perplexity": 46.911586610916835, "lr": 0.001305, "grad_norm": 0.283916, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:25.579171+00:00", "epoch": 0, "step": 1454, "train_loss": 3.837066411972046, "perplexity": 46.38918786704632, "lr": 0.001305, "grad_norm": 0.269579, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:25.887024+00:00", "epoch": 0, "step": 1455, "train_loss": 3.7985873222351074, "perplexity": 44.638080707055096, "lr": 0.001305, "grad_norm": 0.263547, "tokens_per_sec": 106440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:26.193205+00:00", "epoch": 0, "step": 1456, "train_loss": 3.979482412338257, "perplexity": 53.48934163817143, "lr": 0.001305, "grad_norm": 0.273006, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:26.500294+00:00", "epoch": 0, "step": 1457, "train_loss": 3.9557580947875977, "perplexity": 52.235278230370625, "lr": 0.001305, "grad_norm": 0.273568, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:26.806204+00:00", "epoch": 0, "step": 1458, "train_loss": 3.8689634799957275, "perplexity": 47.892718583717865, "lr": 0.001305, "grad_norm": 0.280753, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:27.113566+00:00", "epoch": 0, "step": 1459, "train_loss": 3.8792004585266113, "perplexity": 48.38551337694595, "lr": 0.001305, "grad_norm": 0.277006, "tokens_per_sec": 106612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:27.420979+00:00", "epoch": 0, "step": 1460, "train_loss": 3.9116621017456055, "perplexity": 49.98195807176797, "lr": 0.001305, "grad_norm": 0.263083, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:27.726563+00:00", "epoch": 0, "step": 1461, "train_loss": 3.8085074424743652, "perplexity": 45.08309950543916, "lr": 0.001305, "grad_norm": 0.265729, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:28.033037+00:00", "epoch": 0, "step": 1462, "train_loss": 3.8190653324127197, "perplexity": 45.561603459280896, "lr": 0.001305, "grad_norm": 0.248307, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:28.339627+00:00", "epoch": 0, "step": 1463, "train_loss": 3.791200637817383, "perplexity": 44.30956809507616, "lr": 0.001305, "grad_norm": 0.243218, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:28.646667+00:00", "epoch": 0, "step": 1464, "train_loss": 3.8113656044006348, "perplexity": 45.212138623466565, "lr": 0.001305, "grad_norm": 0.241657, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:28.953650+00:00", "epoch": 0, "step": 1465, "train_loss": 3.7506628036499023, "perplexity": 42.54927447041425, "lr": 0.001305, "grad_norm": 0.261123, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:29.259266+00:00", "epoch": 0, "step": 1466, "train_loss": 3.8718950748443604, "perplexity": 48.033326633019726, "lr": 0.001305, "grad_norm": 0.273465, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:29.566226+00:00", "epoch": 0, "step": 1467, "train_loss": 3.8360114097595215, "perplexity": 46.34027297840422, "lr": 0.001305, "grad_norm": 0.293892, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:29.872922+00:00", "epoch": 0, "step": 1468, "train_loss": 4.011974811553955, "perplexity": 55.255882844545205, "lr": 0.001305, "grad_norm": 0.314216, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:30.179663+00:00", "epoch": 0, "step": 1469, "train_loss": 3.894473075866699, "perplexity": 49.13015867473641, "lr": 0.001305, "grad_norm": 0.346641, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:30.486276+00:00", "epoch": 0, "step": 1470, "train_loss": 3.842319965362549, "perplexity": 46.63353723146419, "lr": 0.001305, "grad_norm": 0.362348, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:30.792368+00:00", "epoch": 0, "step": 1471, "train_loss": 3.8154683113098145, "perplexity": 45.39801180785355, "lr": 0.001305, "grad_norm": 0.347867, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:31.098403+00:00", "epoch": 0, "step": 1472, "train_loss": 3.8820104598999023, "perplexity": 48.52166794365135, "lr": 0.001305, "grad_norm": 0.334503, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:31.405649+00:00", "epoch": 0, "step": 1473, "train_loss": 3.779829502105713, "perplexity": 43.808571829541776, "lr": 0.001305, "grad_norm": 0.296792, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:31.712849+00:00", "epoch": 0, "step": 1474, "train_loss": 3.8396098613739014, "perplexity": 46.5073265954304, "lr": 0.001305, "grad_norm": 0.313772, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:32.020486+00:00", "epoch": 0, "step": 1475, "train_loss": 3.999677896499634, "perplexity": 54.580566609898156, "lr": 0.001305, "grad_norm": 0.300319, "tokens_per_sec": 106516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:32.327690+00:00", "epoch": 0, "step": 1476, "train_loss": 3.728928327560425, "perplexity": 41.63446573780532, "lr": 0.001305, "grad_norm": 0.281208, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:32.634685+00:00", "epoch": 0, "step": 1477, "train_loss": 3.9314053058624268, "perplexity": 50.9785678346161, "lr": 0.001305, "grad_norm": 0.249758, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:32.941276+00:00", "epoch": 0, "step": 1478, "train_loss": 3.822603702545166, "perplexity": 45.723102829923825, "lr": 0.001305, "grad_norm": 0.263331, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:33.247960+00:00", "epoch": 0, "step": 1479, "train_loss": 3.967594623565674, "perplexity": 52.85723625488327, "lr": 0.001305, "grad_norm": 0.267552, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:33.555759+00:00", "epoch": 0, "step": 1480, "train_loss": 3.8606975078582764, "perplexity": 47.498470373469395, "lr": 0.001305, "grad_norm": 0.250783, "tokens_per_sec": 106516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:33.863345+00:00", "epoch": 0, "step": 1481, "train_loss": 3.7895684242248535, "perplexity": 44.23730440668276, "lr": 0.001305, "grad_norm": 0.27426, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:34.169818+00:00", "epoch": 0, "step": 1482, "train_loss": 3.828666925430298, "perplexity": 46.001174346285175, "lr": 0.001305, "grad_norm": 0.235242, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:34.475842+00:00", "epoch": 0, "step": 1483, "train_loss": 3.8250136375427246, "perplexity": 45.833425417368716, "lr": 0.001305, "grad_norm": 0.242278, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:34.782530+00:00", "epoch": 0, "step": 1484, "train_loss": 3.864773750305176, "perplexity": 47.69248080270622, "lr": 0.001305, "grad_norm": 0.238467, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:35.089343+00:00", "epoch": 0, "step": 1485, "train_loss": 3.731846570968628, "perplexity": 41.75614269814091, "lr": 0.001305, "grad_norm": 0.253078, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:35.396285+00:00", "epoch": 0, "step": 1486, "train_loss": 3.875624656677246, "perplexity": 48.21280533761893, "lr": 0.001305, "grad_norm": 0.251335, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:35.702829+00:00", "epoch": 0, "step": 1487, "train_loss": 3.7745285034179688, "perplexity": 43.57695708486583, "lr": 0.001305, "grad_norm": 0.23781, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:36.008953+00:00", "epoch": 0, "step": 1488, "train_loss": 3.792290687561035, "perplexity": 44.357894062488626, "lr": 0.001305, "grad_norm": 0.247989, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:36.315558+00:00", "epoch": 0, "step": 1489, "train_loss": 3.9100887775421143, "perplexity": 49.90338207636206, "lr": 0.001305, "grad_norm": 0.227291, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:36.621681+00:00", "epoch": 0, "step": 1490, "train_loss": 3.789904832839966, "perplexity": 44.252188720459834, "lr": 0.001305, "grad_norm": 0.238695, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:36.929160+00:00", "epoch": 0, "step": 1491, "train_loss": 3.8183400630950928, "perplexity": 45.52857100639024, "lr": 0.001305, "grad_norm": 0.230605, "tokens_per_sec": 106511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:37.237844+00:00", "epoch": 0, "step": 1492, "train_loss": 3.8980565071105957, "perplexity": 49.306529037151996, "lr": 0.001305, "grad_norm": 0.251498, "tokens_per_sec": 106155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:37.544180+00:00", "epoch": 0, "step": 1493, "train_loss": 3.880236864089966, "perplexity": 48.43568638749816, "lr": 0.001305, "grad_norm": 0.245611, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:37.849703+00:00", "epoch": 0, "step": 1494, "train_loss": 3.8624014854431152, "perplexity": 47.57947569832864, "lr": 0.001305, "grad_norm": 0.277745, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:38.156258+00:00", "epoch": 0, "step": 1495, "train_loss": 3.816620349884033, "perplexity": 45.45034220618007, "lr": 0.001305, "grad_norm": 0.269761, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:38.463512+00:00", "epoch": 0, "step": 1496, "train_loss": 3.9879584312438965, "perplexity": 53.94464516393581, "lr": 0.001305, "grad_norm": 0.252092, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:38.770385+00:00", "epoch": 0, "step": 1497, "train_loss": 3.916266679763794, "perplexity": 50.21263457361666, "lr": 0.001305, "grad_norm": 0.313376, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:39.076482+00:00", "epoch": 0, "step": 1498, "train_loss": 3.81917667388916, "perplexity": 45.5666766379014, "lr": 0.001305, "grad_norm": 0.324968, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:39.381935+00:00", "epoch": 0, "step": 1499, "train_loss": 3.8435773849487305, "perplexity": 46.69221203624542, "lr": 0.001305, "grad_norm": 0.283675, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:39.689798+00:00", "epoch": 0, "step": 1500, "train_loss": 3.7813520431518555, "perplexity": 43.87532298111439, "lr": 0.001305, "grad_norm": 0.277183, "tokens_per_sec": 106499} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:52:45.399544+00:00", "step": 1500, "epoch": 0, "val_loss": 3.807390022277832, "val_ppl": 45.03275087505657, "eval_train_loss": 3.7813520431518555, "eval_train_ppl": 43.87532298111439} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:52:46.384781+00:00", "step": 1500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p8074_epoch_0000_step_0001500.pt", "val_loss": 3.807390022277832} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:52:48.129264+00:00", "step": 1500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0000_step_0001500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:49.234418+00:00", "epoch": 0, "step": 1501, "train_loss": 3.8576371669769287, "perplexity": 47.35333106394262, "lr": 0.001305, "grad_norm": 0.295735, "tokens_per_sec": 3433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:49.540223+00:00", "epoch": 0, "step": 1502, "train_loss": 3.8262696266174316, "perplexity": 45.89102786539967, "lr": 0.001305, "grad_norm": 0.275887, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:49.846076+00:00", "epoch": 0, "step": 1503, "train_loss": 3.840801477432251, "perplexity": 46.56277850476367, "lr": 0.001305, "grad_norm": 0.287973, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:50.151956+00:00", "epoch": 0, "step": 1504, "train_loss": 3.928265333175659, "perplexity": 50.818747570947586, "lr": 0.001305, "grad_norm": 0.270023, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:50.457772+00:00", "epoch": 0, "step": 1505, "train_loss": 3.7659988403320312, "perplexity": 43.20684104914991, "lr": 0.001305, "grad_norm": 0.28844, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:50.763063+00:00", "epoch": 0, "step": 1506, "train_loss": 3.7150068283081055, "perplexity": 41.05886744603293, "lr": 0.001305, "grad_norm": 0.253901, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:51.068409+00:00", "epoch": 0, "step": 1507, "train_loss": 3.843083143234253, "perplexity": 46.66914049924353, "lr": 0.001305, "grad_norm": 0.250078, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:51.591960+00:00", "epoch": 0, "step": 1508, "train_loss": 3.7746975421905518, "perplexity": 43.58432390282575, "lr": 0.001305, "grad_norm": 0.254208, "tokens_per_sec": 62588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:51.900971+00:00", "epoch": 0, "step": 1509, "train_loss": 3.7459583282470703, "perplexity": 42.34957256991744, "lr": 0.001305, "grad_norm": 0.254927, "tokens_per_sec": 106042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:52.207395+00:00", "epoch": 0, "step": 1510, "train_loss": 3.835536241531372, "perplexity": 46.318258783638285, "lr": 0.001305, "grad_norm": 0.26015, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:52.513681+00:00", "epoch": 0, "step": 1511, "train_loss": 3.847712755203247, "perplexity": 46.88570142039654, "lr": 0.001305, "grad_norm": 0.234121, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:52.819723+00:00", "epoch": 0, "step": 1512, "train_loss": 3.876357316970825, "perplexity": 48.24814188899522, "lr": 0.001305, "grad_norm": 0.233175, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:53.125490+00:00", "epoch": 0, "step": 1513, "train_loss": 3.799467086791992, "perplexity": 44.67736898803303, "lr": 0.001305, "grad_norm": 0.227, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:53.431134+00:00", "epoch": 0, "step": 1514, "train_loss": 4.023834228515625, "perplexity": 55.91508655847049, "lr": 0.001305, "grad_norm": 0.251271, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:53.736572+00:00", "epoch": 0, "step": 1515, "train_loss": 3.8491461277008057, "perplexity": 46.952954283028554, "lr": 0.001305, "grad_norm": 0.264749, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:54.043295+00:00", "epoch": 0, "step": 1516, "train_loss": 3.762178897857666, "perplexity": 43.042108237090474, "lr": 0.001305, "grad_norm": 0.270008, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:54.349515+00:00", "epoch": 0, "step": 1517, "train_loss": 3.8637924194335938, "perplexity": 47.645701655616854, "lr": 0.001305, "grad_norm": 0.274725, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:54.655678+00:00", "epoch": 0, "step": 1518, "train_loss": 3.8329882621765137, "perplexity": 46.200391042558415, "lr": 0.001305, "grad_norm": 0.277455, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:54.961086+00:00", "epoch": 0, "step": 1519, "train_loss": 3.935955286026001, "perplexity": 51.211047795563665, "lr": 0.001305, "grad_norm": 0.262709, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:55.267083+00:00", "epoch": 0, "step": 1520, "train_loss": 3.8465044498443604, "perplexity": 46.82908338893906, "lr": 0.001305, "grad_norm": 0.262224, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:55.573660+00:00", "epoch": 0, "step": 1521, "train_loss": 3.8775579929351807, "perplexity": 48.306107065019305, "lr": 0.001305, "grad_norm": 0.263746, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:55.879798+00:00", "epoch": 0, "step": 1522, "train_loss": 3.8379998207092285, "perplexity": 46.432508154934595, "lr": 0.001305, "grad_norm": 0.23677, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:56.185108+00:00", "epoch": 0, "step": 1523, "train_loss": 3.813504934310913, "perplexity": 45.30896583968193, "lr": 0.001305, "grad_norm": 0.257641, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:56.490696+00:00", "epoch": 0, "step": 1524, "train_loss": 3.807422399520874, "perplexity": 45.03420893498037, "lr": 0.001305, "grad_norm": 0.270602, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:56.795358+00:00", "epoch": 0, "step": 1525, "train_loss": 3.8655219078063965, "perplexity": 47.728175640986116, "lr": 0.001305, "grad_norm": 0.301947, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:57.100623+00:00", "epoch": 0, "step": 1526, "train_loss": 3.87809157371521, "perplexity": 48.331889153108875, "lr": 0.001305, "grad_norm": 0.335189, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:57.407601+00:00", "epoch": 0, "step": 1527, "train_loss": 3.8430542945861816, "perplexity": 46.66779417705336, "lr": 0.001305, "grad_norm": 0.369315, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:57.713873+00:00", "epoch": 0, "step": 1528, "train_loss": 3.750535488128662, "perplexity": 42.54385763218776, "lr": 0.001305, "grad_norm": 0.351948, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:58.019400+00:00", "epoch": 0, "step": 1529, "train_loss": 3.790841579437256, "perplexity": 44.29366122924945, "lr": 0.001305, "grad_norm": 0.312549, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:58.325500+00:00", "epoch": 0, "step": 1530, "train_loss": 3.9231090545654297, "perplexity": 50.55738635454555, "lr": 0.001305, "grad_norm": 0.27757, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:58.631595+00:00", "epoch": 0, "step": 1531, "train_loss": 3.7715489864349365, "perplexity": 43.44731203688426, "lr": 0.001305, "grad_norm": 0.297767, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:58.938640+00:00", "epoch": 0, "step": 1532, "train_loss": 3.7766854763031006, "perplexity": 43.671052844233266, "lr": 0.001305, "grad_norm": 0.262144, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:59.245303+00:00", "epoch": 0, "step": 1533, "train_loss": 3.821756601333618, "perplexity": 45.68438713449226, "lr": 0.001305, "grad_norm": 0.259625, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:59.551439+00:00", "epoch": 0, "step": 1534, "train_loss": 3.8253448009490967, "perplexity": 45.84860628419068, "lr": 0.001305, "grad_norm": 0.249972, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:52:59.858249+00:00", "epoch": 0, "step": 1535, "train_loss": 3.8952038288116455, "perplexity": 49.1660738038186, "lr": 0.001305, "grad_norm": 0.246633, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:00.164322+00:00", "epoch": 0, "step": 1536, "train_loss": 3.8214945793151855, "perplexity": 45.672418387270554, "lr": 0.001305, "grad_norm": 0.265626, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:00.470103+00:00", "epoch": 0, "step": 1537, "train_loss": 3.727280855178833, "perplexity": 41.56593057577458, "lr": 0.001305, "grad_norm": 0.304066, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:00.777514+00:00", "epoch": 0, "step": 1538, "train_loss": 3.728194236755371, "perplexity": 41.603913474773584, "lr": 0.001305, "grad_norm": 0.292985, "tokens_per_sec": 106596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:01.084891+00:00", "epoch": 0, "step": 1539, "train_loss": 3.845979690551758, "perplexity": 46.80451583885552, "lr": 0.001305, "grad_norm": 0.254258, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:01.391219+00:00", "epoch": 0, "step": 1540, "train_loss": 3.7938687801361084, "perplexity": 44.427950188742024, "lr": 0.001305, "grad_norm": 0.238428, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:01.697086+00:00", "epoch": 0, "step": 1541, "train_loss": 3.8792970180511475, "perplexity": 48.390185684686344, "lr": 0.001305, "grad_norm": 0.246741, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:02.003245+00:00", "epoch": 0, "step": 1542, "train_loss": 3.843285083770752, "perplexity": 46.67856584216127, "lr": 0.001305, "grad_norm": 0.263219, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:02.310073+00:00", "epoch": 0, "step": 1543, "train_loss": 3.87172532081604, "perplexity": 48.02517347436549, "lr": 0.001305, "grad_norm": 0.236294, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:02.617772+00:00", "epoch": 0, "step": 1544, "train_loss": 3.830603837966919, "perplexity": 46.09036094301151, "lr": 0.001305, "grad_norm": 0.261753, "tokens_per_sec": 106494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:02.924709+00:00", "epoch": 0, "step": 1545, "train_loss": 3.817647695541382, "perplexity": 45.497059411120475, "lr": 0.001305, "grad_norm": 0.25441, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:03.230698+00:00", "epoch": 0, "step": 1546, "train_loss": 3.8996424674987793, "perplexity": 49.38478928150369, "lr": 0.001305, "grad_norm": 0.259818, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:03.536130+00:00", "epoch": 0, "step": 1547, "train_loss": 3.8287994861602783, "perplexity": 46.00727269972864, "lr": 0.001305, "grad_norm": 0.263742, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:03.843074+00:00", "epoch": 0, "step": 1548, "train_loss": 3.8349645137786865, "perplexity": 46.29178491827851, "lr": 0.001305, "grad_norm": 0.253499, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:04.149803+00:00", "epoch": 0, "step": 1549, "train_loss": 3.8877716064453125, "perplexity": 48.80201516861358, "lr": 0.001305, "grad_norm": 0.254821, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:04.457323+00:00", "epoch": 0, "step": 1550, "train_loss": 3.79980731010437, "perplexity": 44.692571856536965, "lr": 0.001305, "grad_norm": 0.247735, "tokens_per_sec": 106615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:04.764234+00:00", "epoch": 0, "step": 1551, "train_loss": 3.7989702224731445, "perplexity": 44.6551759114534, "lr": 0.001305, "grad_norm": 0.237638, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:05.070821+00:00", "epoch": 0, "step": 1552, "train_loss": 3.8985631465911865, "perplexity": 49.33151600056972, "lr": 0.001305, "grad_norm": 0.27156, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:05.376751+00:00", "epoch": 0, "step": 1553, "train_loss": 3.823812961578369, "perplexity": 45.77842734914251, "lr": 0.001305, "grad_norm": 0.235473, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:05.683587+00:00", "epoch": 0, "step": 1554, "train_loss": 3.78023362159729, "perplexity": 43.826279305043265, "lr": 0.001305, "grad_norm": 0.270924, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:05.991491+00:00", "epoch": 0, "step": 1555, "train_loss": 3.793825387954712, "perplexity": 44.426022404894034, "lr": 0.001305, "grad_norm": 0.293397, "tokens_per_sec": 106424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:06.298736+00:00", "epoch": 0, "step": 1556, "train_loss": 3.810115098953247, "perplexity": 45.15563593365365, "lr": 0.001305, "grad_norm": 0.272527, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:06.605028+00:00", "epoch": 0, "step": 1557, "train_loss": 3.8626868724823, "perplexity": 47.593056201781806, "lr": 0.001305, "grad_norm": 0.259974, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:06.910939+00:00", "epoch": 0, "step": 1558, "train_loss": 3.8933959007263184, "perplexity": 49.077265381953424, "lr": 0.001305, "grad_norm": 0.255722, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:07.217047+00:00", "epoch": 0, "step": 1559, "train_loss": 3.8802297115325928, "perplexity": 48.43533994971133, "lr": 0.001305, "grad_norm": 0.263152, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:07.525459+00:00", "epoch": 0, "step": 1560, "train_loss": 3.826458215713501, "perplexity": 45.89968322899045, "lr": 0.001305, "grad_norm": 0.249568, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:07.832746+00:00", "epoch": 0, "step": 1561, "train_loss": 3.8133785724639893, "perplexity": 45.30324087679228, "lr": 0.001305, "grad_norm": 0.252284, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:08.138689+00:00", "epoch": 0, "step": 1562, "train_loss": 3.775425434112549, "perplexity": 43.616060128990426, "lr": 0.001305, "grad_norm": 0.268219, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:08.444743+00:00", "epoch": 0, "step": 1563, "train_loss": 3.6959617137908936, "perplexity": 40.28429592755217, "lr": 0.001305, "grad_norm": 0.270951, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:08.751250+00:00", "epoch": 0, "step": 1564, "train_loss": 3.7697064876556396, "perplexity": 43.367334119725705, "lr": 0.001305, "grad_norm": 0.271693, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:09.057661+00:00", "epoch": 0, "step": 1565, "train_loss": 3.851649761199951, "perplexity": 47.0706545499404, "lr": 0.001305, "grad_norm": 0.279412, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:09.363690+00:00", "epoch": 0, "step": 1566, "train_loss": 3.8016910552978516, "perplexity": 44.77684061947567, "lr": 0.001305, "grad_norm": 0.257179, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:09.671433+00:00", "epoch": 0, "step": 1567, "train_loss": 3.837033987045288, "perplexity": 46.38768372541335, "lr": 0.001305, "grad_norm": 0.254204, "tokens_per_sec": 106479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:09.978219+00:00", "epoch": 0, "step": 1568, "train_loss": 3.7904000282287598, "perplexity": 44.27410762688103, "lr": 0.001305, "grad_norm": 0.23968, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:10.284423+00:00", "epoch": 0, "step": 1569, "train_loss": 3.78549861907959, "perplexity": 44.05763305927892, "lr": 0.001305, "grad_norm": 0.238397, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:10.591128+00:00", "epoch": 0, "step": 1570, "train_loss": 3.836503028869629, "perplexity": 46.363060343061015, "lr": 0.001305, "grad_norm": 0.240076, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:10.897350+00:00", "epoch": 0, "step": 1571, "train_loss": 3.803448438644409, "perplexity": 44.855599878326764, "lr": 0.001305, "grad_norm": 0.270994, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:11.205205+00:00", "epoch": 0, "step": 1572, "train_loss": 3.921071767807007, "perplexity": 50.45449130973026, "lr": 0.001305, "grad_norm": 0.279544, "tokens_per_sec": 106438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:11.511884+00:00", "epoch": 0, "step": 1573, "train_loss": 3.812389850616455, "perplexity": 45.258470709041944, "lr": 0.001305, "grad_norm": 0.291785, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:11.818818+00:00", "epoch": 0, "step": 1574, "train_loss": 3.803220272064209, "perplexity": 44.845366497001905, "lr": 0.001305, "grad_norm": 0.258723, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:12.124928+00:00", "epoch": 0, "step": 1575, "train_loss": 3.736640214920044, "perplexity": 41.95678730420351, "lr": 0.001305, "grad_norm": 0.210558, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:12.431772+00:00", "epoch": 0, "step": 1576, "train_loss": 3.835750102996826, "perplexity": 46.32816553363731, "lr": 0.001305, "grad_norm": 0.252229, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:12.738896+00:00", "epoch": 0, "step": 1577, "train_loss": 3.8216893672943115, "perplexity": 45.68131569186586, "lr": 0.001305, "grad_norm": 0.274424, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:13.046389+00:00", "epoch": 0, "step": 1578, "train_loss": 3.8646061420440674, "perplexity": 47.684487818794715, "lr": 0.001305, "grad_norm": 0.25754, "tokens_per_sec": 106564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:13.353543+00:00", "epoch": 0, "step": 1579, "train_loss": 3.9022555351257324, "perplexity": 49.51400382528535, "lr": 0.001305, "grad_norm": 0.269643, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:13.659898+00:00", "epoch": 0, "step": 1580, "train_loss": 3.8086347579956055, "perplexity": 45.08883964914876, "lr": 0.001305, "grad_norm": 0.2788, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:13.965453+00:00", "epoch": 0, "step": 1581, "train_loss": 3.9027934074401855, "perplexity": 49.54064320076973, "lr": 0.001305, "grad_norm": 0.27306, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:14.272257+00:00", "epoch": 0, "step": 1582, "train_loss": 3.811837911605835, "perplexity": 45.23349768592343, "lr": 0.001305, "grad_norm": 0.295764, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:14.579539+00:00", "epoch": 0, "step": 1583, "train_loss": 3.7366223335266113, "perplexity": 41.956037065090236, "lr": 0.001305, "grad_norm": 0.314318, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:14.886013+00:00", "epoch": 0, "step": 1584, "train_loss": 3.8548552989959717, "perplexity": 47.22178340740599, "lr": 0.001305, "grad_norm": 0.344277, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:15.193404+00:00", "epoch": 0, "step": 1585, "train_loss": 3.8152902126312256, "perplexity": 45.3899272018902, "lr": 0.001305, "grad_norm": 0.282144, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:15.500117+00:00", "epoch": 0, "step": 1586, "train_loss": 3.8002452850341797, "perplexity": 44.71215036969484, "lr": 0.001305, "grad_norm": 0.315952, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:15.805905+00:00", "epoch": 0, "step": 1587, "train_loss": 3.8644206523895264, "perplexity": 47.67564365989759, "lr": 0.001305, "grad_norm": 0.285857, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:16.112631+00:00", "epoch": 0, "step": 1588, "train_loss": 3.8549180030822754, "perplexity": 47.22474449902348, "lr": 0.001305, "grad_norm": 0.311421, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:16.420634+00:00", "epoch": 0, "step": 1589, "train_loss": 3.924220323562622, "perplexity": 50.61360043928023, "lr": 0.001305, "grad_norm": 0.304849, "tokens_per_sec": 106388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:16.728150+00:00", "epoch": 0, "step": 1590, "train_loss": 3.786112070083618, "perplexity": 44.08466855014091, "lr": 0.001305, "grad_norm": 0.281701, "tokens_per_sec": 106619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:17.034705+00:00", "epoch": 0, "step": 1591, "train_loss": 3.9106686115264893, "perplexity": 49.932326143791215, "lr": 0.001305, "grad_norm": 0.275145, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:17.340939+00:00", "epoch": 0, "step": 1592, "train_loss": 3.849252939224243, "perplexity": 46.95796966745107, "lr": 0.001305, "grad_norm": 0.248201, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:17.647352+00:00", "epoch": 0, "step": 1593, "train_loss": 3.7719674110412598, "perplexity": 43.46549526520889, "lr": 0.001305, "grad_norm": 0.233651, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:17.954329+00:00", "epoch": 0, "step": 1594, "train_loss": 3.7408766746520996, "perplexity": 42.13491258799031, "lr": 0.001305, "grad_norm": 0.236402, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:18.261995+00:00", "epoch": 0, "step": 1595, "train_loss": 3.7973852157592773, "perplexity": 44.58445322060081, "lr": 0.001305, "grad_norm": 0.265752, "tokens_per_sec": 106504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:18.569064+00:00", "epoch": 0, "step": 1596, "train_loss": 3.8179657459259033, "perplexity": 45.5115320697561, "lr": 0.001305, "grad_norm": 0.270399, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:18.875592+00:00", "epoch": 0, "step": 1597, "train_loss": 3.8363425731658936, "perplexity": 46.355621722386864, "lr": 0.001305, "grad_norm": 0.292374, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:19.182228+00:00", "epoch": 0, "step": 1598, "train_loss": 3.9170265197753906, "perplexity": 50.25080264142954, "lr": 0.001305, "grad_norm": 0.282445, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:19.488823+00:00", "epoch": 0, "step": 1599, "train_loss": 3.7799501419067383, "perplexity": 43.813857205737456, "lr": 0.001305, "grad_norm": 0.255957, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:19.796972+00:00", "epoch": 0, "step": 1600, "train_loss": 3.730985164642334, "perplexity": 41.7201891801762, "lr": 0.001305, "grad_norm": 0.271816, "tokens_per_sec": 106394} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:53:25.579280+00:00", "step": 1600, "epoch": 0, "val_loss": 3.782537269592285, "val_ppl": 43.92735600335969, "eval_train_loss": 3.730985164642334, "eval_train_ppl": 41.7201891801762} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:53:26.529090+00:00", "step": 1600, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p7825_epoch_0000_step_0001600.pt", "val_loss": 3.782537269592285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:27.608336+00:00", "epoch": 0, "step": 1601, "train_loss": 3.814389705657959, "perplexity": 45.34907165404256, "lr": 0.001305, "grad_norm": 0.276423, "tokens_per_sec": 4195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:27.913680+00:00", "epoch": 0, "step": 1602, "train_loss": 3.9568357467651367, "perplexity": 52.291600023448034, "lr": 0.001305, "grad_norm": 0.290194, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:28.219768+00:00", "epoch": 0, "step": 1603, "train_loss": 3.851391077041626, "perplexity": 47.05847969207566, "lr": 0.001305, "grad_norm": 0.303658, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:28.524931+00:00", "epoch": 0, "step": 1604, "train_loss": 3.809131383895874, "perplexity": 45.11123749594665, "lr": 0.001305, "grad_norm": 0.270996, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:28.830990+00:00", "epoch": 0, "step": 1605, "train_loss": 3.908996105194092, "perplexity": 49.848883810490754, "lr": 0.001305, "grad_norm": 0.25285, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:29.136927+00:00", "epoch": 0, "step": 1606, "train_loss": 3.7982358932495117, "perplexity": 44.62239634776394, "lr": 0.001305, "grad_norm": 0.227088, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:29.443217+00:00", "epoch": 0, "step": 1607, "train_loss": 3.8295555114746094, "perplexity": 46.04206851413386, "lr": 0.001305, "grad_norm": 0.233717, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:29.749277+00:00", "epoch": 0, "step": 1608, "train_loss": 3.76956844329834, "perplexity": 43.361347917149594, "lr": 0.001305, "grad_norm": 0.237888, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:30.056002+00:00", "epoch": 0, "step": 1609, "train_loss": 3.8585774898529053, "perplexity": 47.397879426035225, "lr": 0.001305, "grad_norm": 0.248939, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:30.364200+00:00", "epoch": 0, "step": 1610, "train_loss": 3.786742687225342, "perplexity": 44.112477865409694, "lr": 0.001305, "grad_norm": 0.25056, "tokens_per_sec": 106383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:30.671065+00:00", "epoch": 0, "step": 1611, "train_loss": 3.8045053482055664, "perplexity": 44.903033252687564, "lr": 0.001305, "grad_norm": 0.24136, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:30.977072+00:00", "epoch": 0, "step": 1612, "train_loss": 3.782228708267212, "perplexity": 43.91380381113306, "lr": 0.001305, "grad_norm": 0.232132, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:31.282872+00:00", "epoch": 0, "step": 1613, "train_loss": 3.8511204719543457, "perplexity": 47.04574715089421, "lr": 0.001305, "grad_norm": 0.250634, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:31.590841+00:00", "epoch": 0, "step": 1614, "train_loss": 3.8871214389801025, "perplexity": 48.7702959986178, "lr": 0.001305, "grad_norm": 0.286869, "tokens_per_sec": 106400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:31.897808+00:00", "epoch": 0, "step": 1615, "train_loss": 3.828233003616333, "perplexity": 45.98121776337978, "lr": 0.001305, "grad_norm": 0.320783, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:32.204259+00:00", "epoch": 0, "step": 1616, "train_loss": 3.8645544052124023, "perplexity": 47.68202083829283, "lr": 0.001305, "grad_norm": 0.299732, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:32.510562+00:00", "epoch": 0, "step": 1617, "train_loss": 3.727177143096924, "perplexity": 41.56161991011599, "lr": 0.001305, "grad_norm": 0.258694, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:32.816640+00:00", "epoch": 0, "step": 1618, "train_loss": 3.8557732105255127, "perplexity": 47.26514872655952, "lr": 0.001305, "grad_norm": 0.26432, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:33.122843+00:00", "epoch": 0, "step": 1619, "train_loss": 3.892834424972534, "perplexity": 49.04971742185897, "lr": 0.001305, "grad_norm": 0.259751, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:33.429906+00:00", "epoch": 0, "step": 1620, "train_loss": 3.743558406829834, "perplexity": 42.24805878497283, "lr": 0.001305, "grad_norm": 0.2619, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:33.736529+00:00", "epoch": 0, "step": 1621, "train_loss": 3.7795143127441406, "perplexity": 43.794766009593516, "lr": 0.001305, "grad_norm": 0.258828, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:34.042375+00:00", "epoch": 0, "step": 1622, "train_loss": 3.764552354812622, "perplexity": 43.14438815872301, "lr": 0.001305, "grad_norm": 0.254019, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:34.348519+00:00", "epoch": 0, "step": 1623, "train_loss": 3.7157113552093506, "perplexity": 41.087804715028604, "lr": 0.001305, "grad_norm": 0.249339, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:34.655493+00:00", "epoch": 0, "step": 1624, "train_loss": 3.993173122406006, "perplexity": 54.22668456337253, "lr": 0.001305, "grad_norm": 0.240808, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:34.962747+00:00", "epoch": 0, "step": 1625, "train_loss": 3.8965864181518555, "perplexity": 49.23409730680891, "lr": 0.001305, "grad_norm": 0.238176, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:35.269667+00:00", "epoch": 0, "step": 1626, "train_loss": 3.74467134475708, "perplexity": 42.295104426530536, "lr": 0.001305, "grad_norm": 0.261218, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:35.575703+00:00", "epoch": 0, "step": 1627, "train_loss": 3.8561251163482666, "perplexity": 47.2817845345569, "lr": 0.001305, "grad_norm": 0.262644, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:35.882148+00:00", "epoch": 0, "step": 1628, "train_loss": 3.8875300884246826, "perplexity": 48.790230025726785, "lr": 0.001305, "grad_norm": 0.232222, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:36.188115+00:00", "epoch": 0, "step": 1629, "train_loss": 3.7845985889434814, "perplexity": 44.01799770098399, "lr": 0.001305, "grad_norm": 0.24747, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:36.495486+00:00", "epoch": 0, "step": 1630, "train_loss": 3.965886354446411, "perplexity": 52.76701895011658, "lr": 0.001305, "grad_norm": 0.241095, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:36.802004+00:00", "epoch": 0, "step": 1631, "train_loss": 3.8465805053710938, "perplexity": 46.832645134986166, "lr": 0.001305, "grad_norm": 0.260311, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:37.107965+00:00", "epoch": 0, "step": 1632, "train_loss": 3.7567975521087646, "perplexity": 42.81110588014472, "lr": 0.001305, "grad_norm": 0.250879, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:37.414752+00:00", "epoch": 0, "step": 1633, "train_loss": 3.7135303020477295, "perplexity": 40.99828768483766, "lr": 0.001305, "grad_norm": 0.255911, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:37.721340+00:00", "epoch": 0, "step": 1634, "train_loss": 3.9140522480010986, "perplexity": 50.10156514395236, "lr": 0.001305, "grad_norm": 0.24654, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:38.029234+00:00", "epoch": 0, "step": 1635, "train_loss": 3.8759613037109375, "perplexity": 48.229038767836535, "lr": 0.001305, "grad_norm": 0.246679, "tokens_per_sec": 106424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:38.335156+00:00", "epoch": 0, "step": 1636, "train_loss": 3.7704644203186035, "perplexity": 43.400216098349404, "lr": 0.001305, "grad_norm": 0.229938, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:38.641515+00:00", "epoch": 0, "step": 1637, "train_loss": 3.8033556938171387, "perplexity": 44.851439946373, "lr": 0.001305, "grad_norm": 0.26118, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:38.947460+00:00", "epoch": 0, "step": 1638, "train_loss": 3.779026746749878, "perplexity": 43.77341837557365, "lr": 0.001305, "grad_norm": 0.279461, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:39.254656+00:00", "epoch": 0, "step": 1639, "train_loss": 3.756624221801758, "perplexity": 42.80368606107755, "lr": 0.001305, "grad_norm": 0.263876, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:39.561853+00:00", "epoch": 0, "step": 1640, "train_loss": 3.7774243354797363, "perplexity": 43.7033315256121, "lr": 0.001305, "grad_norm": 0.25848, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:39.869253+00:00", "epoch": 0, "step": 1641, "train_loss": 3.69407320022583, "perplexity": 40.208290279671886, "lr": 0.001305, "grad_norm": 0.257541, "tokens_per_sec": 106534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:40.174736+00:00", "epoch": 0, "step": 1642, "train_loss": 3.7106316089630127, "perplexity": 40.879618307983925, "lr": 0.001305, "grad_norm": 0.28133, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:40.480722+00:00", "epoch": 0, "step": 1643, "train_loss": 3.7629098892211914, "perplexity": 43.07358314902256, "lr": 0.001305, "grad_norm": 0.281123, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:40.786991+00:00", "epoch": 0, "step": 1644, "train_loss": 3.738879442214966, "perplexity": 42.050843354679216, "lr": 0.001305, "grad_norm": 0.240152, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:41.094551+00:00", "epoch": 0, "step": 1645, "train_loss": 3.767984390258789, "perplexity": 43.29271561513687, "lr": 0.001305, "grad_norm": 0.252329, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:41.401493+00:00", "epoch": 0, "step": 1646, "train_loss": 3.8446805477142334, "perplexity": 46.74374956791914, "lr": 0.001305, "grad_norm": 0.283315, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:41.707620+00:00", "epoch": 0, "step": 1647, "train_loss": 3.7707157135009766, "perplexity": 43.41112364720737, "lr": 0.001305, "grad_norm": 0.280251, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:42.014676+00:00", "epoch": 0, "step": 1648, "train_loss": 3.778837203979492, "perplexity": 43.76512222684783, "lr": 0.001305, "grad_norm": 0.285175, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:42.321373+00:00", "epoch": 0, "step": 1649, "train_loss": 3.7415645122528076, "perplexity": 42.163904534903416, "lr": 0.001305, "grad_norm": 0.287816, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:42.627893+00:00", "epoch": 0, "step": 1650, "train_loss": 3.846085548400879, "perplexity": 46.809470726483625, "lr": 0.001305, "grad_norm": 0.314635, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:42.935422+00:00", "epoch": 0, "step": 1651, "train_loss": 3.9848875999450684, "perplexity": 53.779244348140736, "lr": 0.001305, "grad_norm": 0.282971, "tokens_per_sec": 106497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:43.242522+00:00", "epoch": 0, "step": 1652, "train_loss": 3.7857918739318848, "perplexity": 44.07055506858254, "lr": 0.001305, "grad_norm": 0.25153, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:43.548189+00:00", "epoch": 0, "step": 1653, "train_loss": 3.761155605316162, "perplexity": 42.998086096366386, "lr": 0.001305, "grad_norm": 0.225589, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:43.854214+00:00", "epoch": 0, "step": 1654, "train_loss": 3.8021295070648193, "perplexity": 44.79647740894245, "lr": 0.001305, "grad_norm": 0.244981, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:44.160131+00:00", "epoch": 0, "step": 1655, "train_loss": 3.815415859222412, "perplexity": 45.39563064981922, "lr": 0.001305, "grad_norm": 0.246114, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:44.466593+00:00", "epoch": 0, "step": 1656, "train_loss": 3.8512704372406006, "perplexity": 47.05280290887893, "lr": 0.001305, "grad_norm": 0.254369, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:44.774541+00:00", "epoch": 0, "step": 1657, "train_loss": 3.8423736095428467, "perplexity": 46.636038916443184, "lr": 0.001305, "grad_norm": 0.248598, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:45.080960+00:00", "epoch": 0, "step": 1658, "train_loss": 3.839322328567505, "perplexity": 46.493956135611455, "lr": 0.001305, "grad_norm": 0.270169, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:45.387888+00:00", "epoch": 0, "step": 1659, "train_loss": 3.843982219696045, "perplexity": 46.711118492843745, "lr": 0.001305, "grad_norm": 0.276405, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:45.694305+00:00", "epoch": 0, "step": 1660, "train_loss": 3.830400228500366, "perplexity": 46.080977464521624, "lr": 0.001305, "grad_norm": 0.271498, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:46.000572+00:00", "epoch": 0, "step": 1661, "train_loss": 3.7597742080688477, "perplexity": 42.93872966543718, "lr": 0.001305, "grad_norm": 0.269161, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:46.308016+00:00", "epoch": 0, "step": 1662, "train_loss": 3.7851626873016357, "perplexity": 44.042835185949976, "lr": 0.001305, "grad_norm": 0.261914, "tokens_per_sec": 106582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:46.615507+00:00", "epoch": 0, "step": 1663, "train_loss": 3.7840027809143066, "perplexity": 43.99177923588715, "lr": 0.001305, "grad_norm": 0.295429, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:46.922339+00:00", "epoch": 0, "step": 1664, "train_loss": 3.8188412189483643, "perplexity": 45.55139363461036, "lr": 0.001305, "grad_norm": 0.266542, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:47.228675+00:00", "epoch": 0, "step": 1665, "train_loss": 3.9097208976745605, "perplexity": 49.88502700321127, "lr": 0.001305, "grad_norm": 0.260455, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:47.534357+00:00", "epoch": 0, "step": 1666, "train_loss": 3.747849941253662, "perplexity": 42.42975738762658, "lr": 0.001305, "grad_norm": 0.272467, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:47.841746+00:00", "epoch": 0, "step": 1667, "train_loss": 3.731727361679077, "perplexity": 41.751165274718936, "lr": 0.001305, "grad_norm": 0.264491, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:48.148901+00:00", "epoch": 0, "step": 1668, "train_loss": 3.8118460178375244, "perplexity": 45.23386436062197, "lr": 0.001305, "grad_norm": 0.270438, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:48.456291+00:00", "epoch": 0, "step": 1669, "train_loss": 3.7657527923583984, "perplexity": 43.19621140121799, "lr": 0.001305, "grad_norm": 0.257065, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:48.762910+00:00", "epoch": 0, "step": 1670, "train_loss": 3.8135645389556885, "perplexity": 45.31166654498244, "lr": 0.001305, "grad_norm": 0.244926, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:49.068798+00:00", "epoch": 0, "step": 1671, "train_loss": 3.9081649780273438, "perplexity": 49.807470261270645, "lr": 0.001305, "grad_norm": 0.271771, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:49.375037+00:00", "epoch": 0, "step": 1672, "train_loss": 3.8660829067230225, "perplexity": 47.75495860771848, "lr": 0.001305, "grad_norm": 0.286884, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:49.682296+00:00", "epoch": 0, "step": 1673, "train_loss": 3.68611478805542, "perplexity": 39.88956608444904, "lr": 0.001305, "grad_norm": 0.269561, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:49.988915+00:00", "epoch": 0, "step": 1674, "train_loss": 3.8498919010162354, "perplexity": 46.987983603756206, "lr": 0.001305, "grad_norm": 0.264825, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:50.295994+00:00", "epoch": 0, "step": 1675, "train_loss": 3.771646499633789, "perplexity": 43.45154892983566, "lr": 0.001305, "grad_norm": 0.288499, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:50.601872+00:00", "epoch": 0, "step": 1676, "train_loss": 3.800616502761841, "perplexity": 44.728751393660474, "lr": 0.001305, "grad_norm": 0.280597, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:50.908182+00:00", "epoch": 0, "step": 1677, "train_loss": 3.6442677974700928, "perplexity": 38.254752368010536, "lr": 0.001305, "grad_norm": 0.265822, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:51.214384+00:00", "epoch": 0, "step": 1678, "train_loss": 3.8850910663604736, "perplexity": 48.67137458262118, "lr": 0.001305, "grad_norm": 0.278334, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:51.521177+00:00", "epoch": 0, "step": 1679, "train_loss": 3.819808006286621, "perplexity": 45.59545344001545, "lr": 0.001305, "grad_norm": 0.271336, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:51.829449+00:00", "epoch": 0, "step": 1680, "train_loss": 3.8393588066101074, "perplexity": 46.49565217505804, "lr": 0.001305, "grad_norm": 0.247912, "tokens_per_sec": 106357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:52.137547+00:00", "epoch": 0, "step": 1681, "train_loss": 3.8864450454711914, "perplexity": 48.73731924086636, "lr": 0.001305, "grad_norm": 0.246087, "tokens_per_sec": 106293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:52.444050+00:00", "epoch": 0, "step": 1682, "train_loss": 3.8443264961242676, "perplexity": 46.72720279844152, "lr": 0.001305, "grad_norm": 0.2523, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:52.750287+00:00", "epoch": 0, "step": 1683, "train_loss": 3.8254146575927734, "perplexity": 45.851809225815025, "lr": 0.001305, "grad_norm": 0.244367, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:53.057263+00:00", "epoch": 0, "step": 1684, "train_loss": 3.815187931060791, "perplexity": 45.385284886269794, "lr": 0.001305, "grad_norm": 0.248342, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:53.363889+00:00", "epoch": 0, "step": 1685, "train_loss": 3.810696601867676, "perplexity": 45.1819017036227, "lr": 0.001305, "grad_norm": 0.266722, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:53.671368+00:00", "epoch": 0, "step": 1686, "train_loss": 3.834291696548462, "perplexity": 46.26064948317079, "lr": 0.001305, "grad_norm": 0.246034, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:53.978025+00:00", "epoch": 0, "step": 1687, "train_loss": 3.834698438644409, "perplexity": 46.27946946388119, "lr": 0.001305, "grad_norm": 0.279616, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:54.284921+00:00", "epoch": 0, "step": 1688, "train_loss": 3.895693778991699, "perplexity": 49.19016863268219, "lr": 0.001305, "grad_norm": 0.287095, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:54.592032+00:00", "epoch": 0, "step": 1689, "train_loss": 3.6932473182678223, "perplexity": 40.17509668704893, "lr": 0.001305, "grad_norm": 0.268385, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:54.899131+00:00", "epoch": 0, "step": 1690, "train_loss": 3.7196388244628906, "perplexity": 41.2494931100052, "lr": 0.001305, "grad_norm": 0.239398, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:55.209444+00:00", "epoch": 0, "step": 1691, "train_loss": 3.886544942855835, "perplexity": 48.742188214787895, "lr": 0.001305, "grad_norm": 0.274765, "tokens_per_sec": 105539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:55.516462+00:00", "epoch": 0, "step": 1692, "train_loss": 3.836721897125244, "perplexity": 46.37320885575641, "lr": 0.001305, "grad_norm": 0.286111, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:55.823394+00:00", "epoch": 0, "step": 1693, "train_loss": 3.749598264694214, "perplexity": 42.504003210989104, "lr": 0.001305, "grad_norm": 0.277374, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:56.129724+00:00", "epoch": 0, "step": 1694, "train_loss": 3.8798604011535645, "perplexity": 48.41745557860134, "lr": 0.001305, "grad_norm": 0.271864, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:56.436248+00:00", "epoch": 0, "step": 1695, "train_loss": 3.7788736820220947, "perplexity": 43.76671872195926, "lr": 0.001305, "grad_norm": 0.26282, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:56.742572+00:00", "epoch": 0, "step": 1696, "train_loss": 3.746131658554077, "perplexity": 42.356913670531824, "lr": 0.001305, "grad_norm": 0.25205, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:57.049695+00:00", "epoch": 0, "step": 1697, "train_loss": 3.741804599761963, "perplexity": 42.17402877702286, "lr": 0.001305, "grad_norm": 0.263679, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:57.356379+00:00", "epoch": 0, "step": 1698, "train_loss": 3.7703123092651367, "perplexity": 43.3936149478247, "lr": 0.001305, "grad_norm": 0.262429, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:57.662585+00:00", "epoch": 0, "step": 1699, "train_loss": 3.8300538063049316, "perplexity": 46.06501675587208, "lr": 0.001305, "grad_norm": 0.24576, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:53:57.969259+00:00", "epoch": 0, "step": 1700, "train_loss": 3.7919647693634033, "perplexity": 44.34343937325612, "lr": 0.001305, "grad_norm": 0.251999, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:54:03.649280+00:00", "step": 1700, "epoch": 0, "val_loss": 3.7612058520317078, "val_ppl": 43.00024666324772, "eval_train_loss": 3.7919647693634033, "eval_train_ppl": 44.34343937325612} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:54:04.580726+00:00", "step": 1700, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p7612_epoch_0000_step_0001700.pt", "val_loss": 3.7612058520317078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:05.817059+00:00", "epoch": 0, "step": 1701, "train_loss": 3.856116533279419, "perplexity": 47.28137871348659, "lr": 0.001305, "grad_norm": 0.258276, "tokens_per_sec": 4175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:06.123005+00:00", "epoch": 0, "step": 1702, "train_loss": 3.8469550609588623, "perplexity": 46.85018984944176, "lr": 0.001305, "grad_norm": 0.307557, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:06.429098+00:00", "epoch": 0, "step": 1703, "train_loss": 3.772850751876831, "perplexity": 43.50390707497734, "lr": 0.001305, "grad_norm": 0.308604, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:06.734356+00:00", "epoch": 0, "step": 1704, "train_loss": 3.8073971271514893, "perplexity": 45.033070828198596, "lr": 0.001305, "grad_norm": 0.297187, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:07.040410+00:00", "epoch": 0, "step": 1705, "train_loss": 3.8846375942230225, "perplexity": 48.64930847391739, "lr": 0.001305, "grad_norm": 0.331798, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:07.346060+00:00", "epoch": 0, "step": 1706, "train_loss": 3.73854660987854, "perplexity": 42.03684980311921, "lr": 0.001305, "grad_norm": 0.341475, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:07.651788+00:00", "epoch": 0, "step": 1707, "train_loss": 3.81610369682312, "perplexity": 45.42686621275383, "lr": 0.001305, "grad_norm": 0.339517, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:07.958710+00:00", "epoch": 0, "step": 1708, "train_loss": 3.8056719303131104, "perplexity": 44.955446894318655, "lr": 0.001305, "grad_norm": 0.289636, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:08.265146+00:00", "epoch": 0, "step": 1709, "train_loss": 3.824516773223877, "perplexity": 45.81065808027537, "lr": 0.001305, "grad_norm": 0.31148, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:08.571360+00:00", "epoch": 0, "step": 1710, "train_loss": 3.786203622817993, "perplexity": 44.088704806853045, "lr": 0.001305, "grad_norm": 0.296175, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:08.877586+00:00", "epoch": 0, "step": 1711, "train_loss": 3.80086612701416, "perplexity": 44.73991816847522, "lr": 0.001305, "grad_norm": 0.321961, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:09.183416+00:00", "epoch": 0, "step": 1712, "train_loss": 3.753617286682129, "perplexity": 42.67517146856381, "lr": 0.001305, "grad_norm": 0.269511, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:09.490020+00:00", "epoch": 0, "step": 1713, "train_loss": 3.8384780883789062, "perplexity": 46.45472063373898, "lr": 0.001305, "grad_norm": 0.286841, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:09.796478+00:00", "epoch": 0, "step": 1714, "train_loss": 3.8021016120910645, "perplexity": 44.795227829809384, "lr": 0.001305, "grad_norm": 0.26638, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:10.103172+00:00", "epoch": 0, "step": 1715, "train_loss": 3.768454074859619, "perplexity": 43.31305431300211, "lr": 0.001305, "grad_norm": 0.262309, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:10.408862+00:00", "epoch": 0, "step": 1716, "train_loss": 3.8690128326416016, "perplexity": 47.895082274424794, "lr": 0.001305, "grad_norm": 0.325788, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:10.715168+00:00", "epoch": 0, "step": 1717, "train_loss": 3.793234348297119, "perplexity": 44.399772621919155, "lr": 0.001305, "grad_norm": 0.281561, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:11.021569+00:00", "epoch": 0, "step": 1718, "train_loss": 3.8093483448028564, "perplexity": 45.12102593276454, "lr": 0.001305, "grad_norm": 0.244879, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:11.328409+00:00", "epoch": 0, "step": 1719, "train_loss": 3.8346428871154785, "perplexity": 46.27689864000163, "lr": 0.001305, "grad_norm": 0.226719, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:11.635074+00:00", "epoch": 0, "step": 1720, "train_loss": 3.817359685897827, "perplexity": 45.48395770605596, "lr": 0.001305, "grad_norm": 0.234963, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:11.941358+00:00", "epoch": 0, "step": 1721, "train_loss": 3.7887299060821533, "perplexity": 44.2002261719124, "lr": 0.001305, "grad_norm": 0.224567, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:12.246764+00:00", "epoch": 0, "step": 1722, "train_loss": 3.9168660640716553, "perplexity": 50.24274026037325, "lr": 0.001305, "grad_norm": 0.24813, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:12.551749+00:00", "epoch": 0, "step": 1723, "train_loss": 3.8018569946289062, "perplexity": 44.78427147497335, "lr": 0.001305, "grad_norm": 0.234432, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:12.857715+00:00", "epoch": 0, "step": 1724, "train_loss": 3.7823662757873535, "perplexity": 43.91984533977287, "lr": 0.001305, "grad_norm": 0.236398, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:13.165288+00:00", "epoch": 0, "step": 1725, "train_loss": 3.876047372817993, "perplexity": 48.233189976780295, "lr": 0.001305, "grad_norm": 0.271114, "tokens_per_sec": 106539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:13.472543+00:00", "epoch": 0, "step": 1726, "train_loss": 3.772461175918579, "perplexity": 43.48696229954378, "lr": 0.001305, "grad_norm": 0.248761, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:13.779418+00:00", "epoch": 0, "step": 1727, "train_loss": 3.8061769008636475, "perplexity": 44.97815380376429, "lr": 0.001305, "grad_norm": 0.205514, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:14.085143+00:00", "epoch": 0, "step": 1728, "train_loss": 3.8315999507904053, "perplexity": 46.13629501654766, "lr": 0.001305, "grad_norm": 0.223191, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:14.391376+00:00", "epoch": 0, "step": 1729, "train_loss": 3.7732934951782227, "perplexity": 43.52317240290189, "lr": 0.001305, "grad_norm": 0.23076, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:14.698369+00:00", "epoch": 0, "step": 1730, "train_loss": 3.9581236839294434, "perplexity": 52.35899170731072, "lr": 0.001305, "grad_norm": 0.224327, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:15.005402+00:00", "epoch": 0, "step": 1731, "train_loss": 3.8674120903015137, "perplexity": 47.81847591822293, "lr": 0.001305, "grad_norm": 0.229912, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:15.310907+00:00", "epoch": 0, "step": 1732, "train_loss": 3.8233482837677, "perplexity": 45.757160071344764, "lr": 0.001305, "grad_norm": 0.25517, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:15.617232+00:00", "epoch": 0, "step": 1733, "train_loss": 3.7822210788726807, "perplexity": 43.913468776676474, "lr": 0.001305, "grad_norm": 0.290521, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:15.922735+00:00", "epoch": 0, "step": 1734, "train_loss": 3.8856289386749268, "perplexity": 48.697560609253614, "lr": 0.001305, "grad_norm": 0.2683, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:16.229352+00:00", "epoch": 0, "step": 1735, "train_loss": 3.8296968936920166, "perplexity": 46.048578504061965, "lr": 0.001305, "grad_norm": 0.257364, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:16.535865+00:00", "epoch": 0, "step": 1736, "train_loss": 3.8281474113464355, "perplexity": 45.977282295004, "lr": 0.001305, "grad_norm": 0.236181, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:16.842317+00:00", "epoch": 0, "step": 1737, "train_loss": 3.8956737518310547, "perplexity": 49.189183503137556, "lr": 0.001305, "grad_norm": 0.254485, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:17.148366+00:00", "epoch": 0, "step": 1738, "train_loss": 3.7952258586883545, "perplexity": 44.488283336214195, "lr": 0.001305, "grad_norm": 0.25046, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:17.453843+00:00", "epoch": 0, "step": 1739, "train_loss": 3.8499221801757812, "perplexity": 46.98940638194864, "lr": 0.001305, "grad_norm": 0.263576, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:17.759394+00:00", "epoch": 0, "step": 1740, "train_loss": 3.832782745361328, "perplexity": 46.19089706095087, "lr": 0.001305, "grad_norm": 0.257574, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:18.066436+00:00", "epoch": 0, "step": 1741, "train_loss": 3.804959297180176, "perplexity": 44.923421565866285, "lr": 0.001305, "grad_norm": 0.22968, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:18.373068+00:00", "epoch": 0, "step": 1742, "train_loss": 3.792942523956299, "perplexity": 44.386817577931716, "lr": 0.001305, "grad_norm": 0.238716, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:18.679737+00:00", "epoch": 0, "step": 1743, "train_loss": 3.8831911087036133, "perplexity": 48.578988824119115, "lr": 0.001305, "grad_norm": 0.278245, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:18.985286+00:00", "epoch": 0, "step": 1744, "train_loss": 3.857424259185791, "perplexity": 47.34325024400342, "lr": 0.001305, "grad_norm": 0.270646, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:19.290662+00:00", "epoch": 0, "step": 1745, "train_loss": 3.7664730548858643, "perplexity": 43.22733522093377, "lr": 0.001305, "grad_norm": 0.238813, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:19.596841+00:00", "epoch": 0, "step": 1746, "train_loss": 3.7353622913360596, "perplexity": 41.90320398118754, "lr": 0.001305, "grad_norm": 0.219074, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:19.903640+00:00", "epoch": 0, "step": 1747, "train_loss": 3.834594249725342, "perplexity": 46.27464790716349, "lr": 0.001305, "grad_norm": 0.254075, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:20.211167+00:00", "epoch": 0, "step": 1748, "train_loss": 3.7073299884796143, "perplexity": 40.744871886000695, "lr": 0.001305, "grad_norm": 0.25306, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:20.517861+00:00", "epoch": 0, "step": 1749, "train_loss": 3.9256298542022705, "perplexity": 50.684992162473, "lr": 0.001305, "grad_norm": 0.261228, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:20.823641+00:00", "epoch": 0, "step": 1750, "train_loss": 3.7766571044921875, "perplexity": 43.669813834956145, "lr": 0.001305, "grad_norm": 0.257327, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:54:21.800168+00:00", "step": 1750, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0000_step_0001750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:22.811118+00:00", "epoch": 0, "step": 1751, "train_loss": 3.7745416164398193, "perplexity": 43.57752851420284, "lr": 0.001305, "grad_norm": 0.249954, "tokens_per_sec": 16486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:23.117742+00:00", "epoch": 0, "step": 1752, "train_loss": 3.855437994003296, "perplexity": 47.249307323079776, "lr": 0.001305, "grad_norm": 0.219851, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:23.425412+00:00", "epoch": 0, "step": 1753, "train_loss": 3.801673650741577, "perplexity": 44.776061305215144, "lr": 0.001305, "grad_norm": 0.231352, "tokens_per_sec": 106505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:23.731467+00:00", "epoch": 0, "step": 1754, "train_loss": 3.867469072341919, "perplexity": 47.82120079018347, "lr": 0.001305, "grad_norm": 0.240226, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:24.037935+00:00", "epoch": 0, "step": 1755, "train_loss": 3.7627673149108887, "perplexity": 43.067442400379726, "lr": 0.001305, "grad_norm": 0.242267, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:24.343438+00:00", "epoch": 0, "step": 1756, "train_loss": 3.716343641281128, "perplexity": 41.11379217655902, "lr": 0.001305, "grad_norm": 0.248497, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:24.650323+00:00", "epoch": 0, "step": 1757, "train_loss": 3.710096597671509, "perplexity": 40.857753100184866, "lr": 0.001305, "grad_norm": 0.229951, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:24.956484+00:00", "epoch": 0, "step": 1758, "train_loss": 3.83134126663208, "perplexity": 46.12436183143258, "lr": 0.001305, "grad_norm": 0.227942, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:25.263217+00:00", "epoch": 0, "step": 1759, "train_loss": 3.8373842239379883, "perplexity": 46.403933249046126, "lr": 0.001305, "grad_norm": 0.24725, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:25.569297+00:00", "epoch": 0, "step": 1760, "train_loss": 3.826038122177124, "perplexity": 45.880405118332554, "lr": 0.001305, "grad_norm": 0.282156, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:25.876027+00:00", "epoch": 0, "step": 1761, "train_loss": 3.7407925128936768, "perplexity": 42.13136658887676, "lr": 0.001305, "grad_norm": 0.377936, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:26.181571+00:00", "epoch": 0, "step": 1762, "train_loss": 3.8354384899139404, "perplexity": 46.313731320212604, "lr": 0.001305, "grad_norm": 0.376973, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:26.488571+00:00", "epoch": 0, "step": 1763, "train_loss": 3.828468084335327, "perplexity": 45.99202833174013, "lr": 0.001305, "grad_norm": 0.2842, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:27.180246+00:00", "epoch": 0, "step": 1764, "train_loss": 3.790491819381714, "perplexity": 44.27817178479013, "lr": 0.001305, "grad_norm": 0.319595, "tokens_per_sec": 47375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:27.487778+00:00", "epoch": 0, "step": 1765, "train_loss": 3.7704269886016846, "perplexity": 43.39859158415056, "lr": 0.001305, "grad_norm": 0.300382, "tokens_per_sec": 106552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:27.794815+00:00", "epoch": 0, "step": 1766, "train_loss": 3.8570947647094727, "perplexity": 47.32765347422291, "lr": 0.001305, "grad_norm": 0.284449, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:28.101290+00:00", "epoch": 0, "step": 1767, "train_loss": 3.817082405090332, "perplexity": 45.47134762588255, "lr": 0.001305, "grad_norm": 0.283989, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:28.407773+00:00", "epoch": 0, "step": 1768, "train_loss": 3.759060859680176, "perplexity": 42.90811031424956, "lr": 0.001305, "grad_norm": 0.252724, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:28.713810+00:00", "epoch": 0, "step": 1769, "train_loss": 3.617291212081909, "perplexity": 37.23656512864338, "lr": 0.001305, "grad_norm": 0.246804, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:29.020317+00:00", "epoch": 0, "step": 1770, "train_loss": 3.9069912433624268, "perplexity": 49.749043802141, "lr": 0.001305, "grad_norm": 0.237833, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:29.327100+00:00", "epoch": 0, "step": 1771, "train_loss": 3.7831737995147705, "perplexity": 43.9553259807909, "lr": 0.001305, "grad_norm": 0.248476, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:29.634118+00:00", "epoch": 0, "step": 1772, "train_loss": 3.7531230449676514, "perplexity": 42.654084830029, "lr": 0.001305, "grad_norm": 0.23929, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:29.940124+00:00", "epoch": 0, "step": 1773, "train_loss": 3.789062976837158, "perplexity": 44.214950426588054, "lr": 0.001305, "grad_norm": 0.226467, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:30.245086+00:00", "epoch": 0, "step": 1774, "train_loss": 3.878148078918457, "perplexity": 48.33462023348819, "lr": 0.001305, "grad_norm": 0.229864, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:30.551273+00:00", "epoch": 0, "step": 1775, "train_loss": 3.800407648086548, "perplexity": 44.71941056028421, "lr": 0.001305, "grad_norm": 0.237272, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:30.858033+00:00", "epoch": 0, "step": 1776, "train_loss": 3.66892409324646, "perplexity": 39.20969718007378, "lr": 0.001305, "grad_norm": 0.214119, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:31.165300+00:00", "epoch": 0, "step": 1777, "train_loss": 3.7563178539276123, "perplexity": 42.79057439537256, "lr": 0.001305, "grad_norm": 0.220972, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:31.472204+00:00", "epoch": 0, "step": 1778, "train_loss": 3.7295055389404297, "perplexity": 41.65850456230372, "lr": 0.001305, "grad_norm": 0.231, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:31.778662+00:00", "epoch": 0, "step": 1779, "train_loss": 3.7506253719329834, "perplexity": 42.5476818078254, "lr": 0.001305, "grad_norm": 0.245271, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:32.084388+00:00", "epoch": 0, "step": 1780, "train_loss": 3.7749905586242676, "perplexity": 43.597096697209615, "lr": 0.001305, "grad_norm": 0.281986, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:32.390860+00:00", "epoch": 0, "step": 1781, "train_loss": 3.852292776107788, "perplexity": 47.100931415732475, "lr": 0.001305, "grad_norm": 0.311608, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:32.698163+00:00", "epoch": 0, "step": 1782, "train_loss": 3.8170318603515625, "perplexity": 45.46904934657875, "lr": 0.001305, "grad_norm": 0.293891, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:33.007410+00:00", "epoch": 0, "step": 1783, "train_loss": 3.719327926635742, "perplexity": 41.236670725555335, "lr": 0.001305, "grad_norm": 0.273503, "tokens_per_sec": 105962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:33.313672+00:00", "epoch": 0, "step": 1784, "train_loss": 3.7903122901916504, "perplexity": 44.27022327398832, "lr": 0.001305, "grad_norm": 0.256833, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:33.619952+00:00", "epoch": 0, "step": 1785, "train_loss": 3.85514497756958, "perplexity": 47.23546452773473, "lr": 0.001305, "grad_norm": 0.260931, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:33.925193+00:00", "epoch": 0, "step": 1786, "train_loss": 3.8627705574035645, "perplexity": 47.59703918959848, "lr": 0.001305, "grad_norm": 0.269129, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:34.232876+00:00", "epoch": 0, "step": 1787, "train_loss": 3.6677045822143555, "perplexity": 39.16190966641975, "lr": 0.001305, "grad_norm": 0.26143, "tokens_per_sec": 106500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:34.540882+00:00", "epoch": 0, "step": 1788, "train_loss": 3.9340670108795166, "perplexity": 51.11443848796881, "lr": 0.001305, "grad_norm": 0.282524, "tokens_per_sec": 106387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:34.848392+00:00", "epoch": 0, "step": 1789, "train_loss": 3.698645830154419, "perplexity": 40.392568909080936, "lr": 0.001305, "grad_norm": 0.298756, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:35.155286+00:00", "epoch": 0, "step": 1790, "train_loss": 3.8340163230895996, "perplexity": 46.24791228193746, "lr": 0.001305, "grad_norm": 0.299313, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:35.461070+00:00", "epoch": 0, "step": 1791, "train_loss": 3.7699081897735596, "perplexity": 43.37608228509828, "lr": 0.001305, "grad_norm": 0.29807, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:35.768019+00:00", "epoch": 0, "step": 1792, "train_loss": 3.770253896713257, "perplexity": 43.39108029006894, "lr": 0.001305, "grad_norm": 0.303658, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:36.075622+00:00", "epoch": 0, "step": 1793, "train_loss": 3.770602226257324, "perplexity": 43.406197317983846, "lr": 0.001305, "grad_norm": 0.30079, "tokens_per_sec": 106528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:36.383029+00:00", "epoch": 0, "step": 1794, "train_loss": 3.882672071456909, "perplexity": 48.55378106196182, "lr": 0.001305, "grad_norm": 0.278278, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:36.689906+00:00", "epoch": 0, "step": 1795, "train_loss": 3.805432081222534, "perplexity": 44.94466566425041, "lr": 0.001305, "grad_norm": 0.26927, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:36.996390+00:00", "epoch": 0, "step": 1796, "train_loss": 3.738579750061035, "perplexity": 42.038242935077406, "lr": 0.001305, "grad_norm": 0.262581, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:37.302713+00:00", "epoch": 0, "step": 1797, "train_loss": 3.7529609203338623, "perplexity": 42.64717011268448, "lr": 0.001305, "grad_norm": 0.264655, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:37.609581+00:00", "epoch": 0, "step": 1798, "train_loss": 3.7387077808380127, "perplexity": 42.043625468540775, "lr": 0.001305, "grad_norm": 0.262924, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:37.916187+00:00", "epoch": 0, "step": 1799, "train_loss": 3.8235697746276855, "perplexity": 45.767295986544305, "lr": 0.001305, "grad_norm": 0.278236, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:38.223386+00:00", "epoch": 0, "step": 1800, "train_loss": 3.833423614501953, "perplexity": 46.22050886908854, "lr": 0.001305, "grad_norm": 0.290082, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:54:43.927659+00:00", "step": 1800, "epoch": 0, "val_loss": 3.7423824667930603, "val_ppl": 42.19840680077139, "eval_train_loss": 3.833423614501953, "eval_train_ppl": 46.22050886908854} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:54:44.889983+00:00", "step": 1800, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p7424_epoch_0000_step_0001800.pt", "val_loss": 3.7423824667930603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:45.894160+00:00", "epoch": 0, "step": 1801, "train_loss": 3.8149116039276123, "perplexity": 45.37274543318402, "lr": 0.001305, "grad_norm": 0.240454, "tokens_per_sec": 4272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:46.200317+00:00", "epoch": 0, "step": 1802, "train_loss": 3.7221784591674805, "perplexity": 41.354384891280986, "lr": 0.001305, "grad_norm": 0.243814, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:46.506452+00:00", "epoch": 0, "step": 1803, "train_loss": 3.7255117893218994, "perplexity": 41.492462711087185, "lr": 0.001305, "grad_norm": 0.232615, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:46.810864+00:00", "epoch": 0, "step": 1804, "train_loss": 3.8652114868164062, "perplexity": 47.71336211278721, "lr": 0.001305, "grad_norm": 0.231623, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:47.116821+00:00", "epoch": 0, "step": 1805, "train_loss": 3.761772632598877, "perplexity": 43.02462527544914, "lr": 0.001305, "grad_norm": 0.245739, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:47.422635+00:00", "epoch": 0, "step": 1806, "train_loss": 3.781285285949707, "perplexity": 43.87239408507235, "lr": 0.001305, "grad_norm": 0.261429, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:47.729234+00:00", "epoch": 0, "step": 1807, "train_loss": 3.8049192428588867, "perplexity": 44.92162222474142, "lr": 0.001305, "grad_norm": 0.278982, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:48.034942+00:00", "epoch": 0, "step": 1808, "train_loss": 3.8384127616882324, "perplexity": 46.451685999696174, "lr": 0.001305, "grad_norm": 0.29788, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:48.340569+00:00", "epoch": 0, "step": 1809, "train_loss": 3.844555139541626, "perplexity": 46.737887887264016, "lr": 0.001305, "grad_norm": 0.258451, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:48.646475+00:00", "epoch": 0, "step": 1810, "train_loss": 3.806260347366333, "perplexity": 44.981907229999436, "lr": 0.001305, "grad_norm": 0.250469, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:48.953454+00:00", "epoch": 0, "step": 1811, "train_loss": 3.6930336952209473, "perplexity": 40.16651527711225, "lr": 0.001305, "grad_norm": 0.276041, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:49.259910+00:00", "epoch": 0, "step": 1812, "train_loss": 3.821035146713257, "perplexity": 45.65143980874498, "lr": 0.001305, "grad_norm": 0.279704, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:49.567183+00:00", "epoch": 0, "step": 1813, "train_loss": 3.7186405658721924, "perplexity": 41.20833599528568, "lr": 0.001305, "grad_norm": 0.281156, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:49.873608+00:00", "epoch": 0, "step": 1814, "train_loss": 3.8683583736419678, "perplexity": 47.863747161683804, "lr": 0.001305, "grad_norm": 0.251092, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:50.184685+00:00", "epoch": 0, "step": 1815, "train_loss": 3.8303146362304688, "perplexity": 46.07703345785199, "lr": 0.001305, "grad_norm": 0.245646, "tokens_per_sec": 105336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:50.491065+00:00", "epoch": 0, "step": 1816, "train_loss": 3.7235467433929443, "perplexity": 41.41100817331911, "lr": 0.001305, "grad_norm": 0.249301, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:50.797271+00:00", "epoch": 0, "step": 1817, "train_loss": 3.734513282775879, "perplexity": 41.86764290027565, "lr": 0.001305, "grad_norm": 0.247266, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:51.104198+00:00", "epoch": 0, "step": 1818, "train_loss": 3.859588384628296, "perplexity": 47.44581792101626, "lr": 0.001305, "grad_norm": 0.256881, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:51.410187+00:00", "epoch": 0, "step": 1819, "train_loss": 3.886453628540039, "perplexity": 48.73773755842808, "lr": 0.001305, "grad_norm": 0.297144, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:51.716198+00:00", "epoch": 0, "step": 1820, "train_loss": 3.729581117630005, "perplexity": 41.661653176470764, "lr": 0.001305, "grad_norm": 0.344333, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:52.021591+00:00", "epoch": 0, "step": 1821, "train_loss": 3.8054914474487305, "perplexity": 44.947333938640455, "lr": 0.001305, "grad_norm": 0.316655, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:52.327669+00:00", "epoch": 0, "step": 1822, "train_loss": 3.7382116317749023, "perplexity": 42.02277073711033, "lr": 0.001305, "grad_norm": 0.278212, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:52.636044+00:00", "epoch": 0, "step": 1823, "train_loss": 3.861435651779175, "perplexity": 47.5335440237359, "lr": 0.001305, "grad_norm": 0.260243, "tokens_per_sec": 106260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:52.943238+00:00", "epoch": 0, "step": 1824, "train_loss": 3.7388899326324463, "perplexity": 42.051284487895245, "lr": 0.001305, "grad_norm": 0.290531, "tokens_per_sec": 106670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:53.249434+00:00", "epoch": 0, "step": 1825, "train_loss": 3.681755542755127, "perplexity": 39.71605614193306, "lr": 0.001305, "grad_norm": 0.316507, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:53.555038+00:00", "epoch": 0, "step": 1826, "train_loss": 3.7620849609375, "perplexity": 43.03806518390419, "lr": 0.001305, "grad_norm": 0.323946, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:53.861034+00:00", "epoch": 0, "step": 1827, "train_loss": 3.741241931915283, "perplexity": 42.15030548185868, "lr": 0.001305, "grad_norm": 0.249177, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:54.168830+00:00", "epoch": 0, "step": 1828, "train_loss": 3.7039942741394043, "perplexity": 40.60918506461395, "lr": 0.001305, "grad_norm": 0.236578, "tokens_per_sec": 106459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:54.475677+00:00", "epoch": 0, "step": 1829, "train_loss": 3.8062689304351807, "perplexity": 44.98229331446298, "lr": 0.001305, "grad_norm": 0.257767, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:54.782787+00:00", "epoch": 0, "step": 1830, "train_loss": 3.7139720916748047, "perplexity": 41.01640430463736, "lr": 0.001305, "grad_norm": 0.264538, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:55.089087+00:00", "epoch": 0, "step": 1831, "train_loss": 3.7845957279205322, "perplexity": 44.01787176466254, "lr": 0.001305, "grad_norm": 0.26125, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:55.395452+00:00", "epoch": 0, "step": 1832, "train_loss": 3.717818021774292, "perplexity": 41.17445425825018, "lr": 0.001305, "grad_norm": 0.247854, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:55.701464+00:00", "epoch": 0, "step": 1833, "train_loss": 3.81624698638916, "perplexity": 45.43337587507222, "lr": 0.001305, "grad_norm": 0.228627, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:56.008269+00:00", "epoch": 0, "step": 1834, "train_loss": 3.703683376312256, "perplexity": 40.59656171960204, "lr": 0.001305, "grad_norm": 0.244874, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:56.316145+00:00", "epoch": 0, "step": 1835, "train_loss": 3.7262978553771973, "perplexity": 41.52509135002773, "lr": 0.001305, "grad_norm": 0.251128, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:56.621951+00:00", "epoch": 0, "step": 1836, "train_loss": 3.7071621417999268, "perplexity": 40.73803356845084, "lr": 0.001305, "grad_norm": 0.241189, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:56.927258+00:00", "epoch": 0, "step": 1837, "train_loss": 3.7578125, "perplexity": 42.854578979512624, "lr": 0.001305, "grad_norm": 0.214772, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:57.233299+00:00", "epoch": 0, "step": 1838, "train_loss": 3.6470229625701904, "perplexity": 38.36029585474889, "lr": 0.001305, "grad_norm": 0.232847, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:57.539305+00:00", "epoch": 0, "step": 1839, "train_loss": 3.7137367725372314, "perplexity": 41.006753495304636, "lr": 0.001305, "grad_norm": 0.232816, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:57.846883+00:00", "epoch": 0, "step": 1840, "train_loss": 3.823136568069458, "perplexity": 45.74747358767719, "lr": 0.001305, "grad_norm": 0.216264, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:58.153852+00:00", "epoch": 0, "step": 1841, "train_loss": 3.7319369316101074, "perplexity": 41.75991598045636, "lr": 0.001305, "grad_norm": 0.223067, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:58.459266+00:00", "epoch": 0, "step": 1842, "train_loss": 3.824462652206421, "perplexity": 45.80817882794017, "lr": 0.001305, "grad_norm": 0.233033, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:58.765927+00:00", "epoch": 0, "step": 1843, "train_loss": 3.9261300563812256, "perplexity": 50.71035124779912, "lr": 0.001305, "grad_norm": 0.293717, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:59.072782+00:00", "epoch": 0, "step": 1844, "train_loss": 3.7320070266723633, "perplexity": 41.76284324695908, "lr": 0.001305, "grad_norm": 0.283186, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:59.379913+00:00", "epoch": 0, "step": 1845, "train_loss": 3.8105714321136475, "perplexity": 45.17624665002815, "lr": 0.001305, "grad_norm": 0.277843, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:59.687300+00:00", "epoch": 0, "step": 1846, "train_loss": 3.6508491039276123, "perplexity": 38.50734891269825, "lr": 0.001305, "grad_norm": 0.286794, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:54:59.994633+00:00", "epoch": 0, "step": 1847, "train_loss": 3.67130446434021, "perplexity": 39.30314198236623, "lr": 0.001305, "grad_norm": 0.269566, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:00.300199+00:00", "epoch": 0, "step": 1848, "train_loss": 3.771953821182251, "perplexity": 43.46490457927016, "lr": 0.001305, "grad_norm": 0.296727, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:00.606957+00:00", "epoch": 0, "step": 1849, "train_loss": 3.847275733947754, "perplexity": 46.86521584893824, "lr": 0.001305, "grad_norm": 0.274519, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:00.913626+00:00", "epoch": 0, "step": 1850, "train_loss": 3.6995608806610107, "perplexity": 40.42954706558395, "lr": 0.001305, "grad_norm": 0.303746, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:01.220554+00:00", "epoch": 0, "step": 1851, "train_loss": 3.8905718326568604, "perplexity": 48.9388633641953, "lr": 0.001305, "grad_norm": 0.300995, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:01.527100+00:00", "epoch": 0, "step": 1852, "train_loss": 3.7824630737304688, "perplexity": 43.92409689623134, "lr": 0.001305, "grad_norm": 0.27012, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:01.832665+00:00", "epoch": 0, "step": 1853, "train_loss": 3.734086751937866, "perplexity": 41.849788867382266, "lr": 0.001305, "grad_norm": 0.287067, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:02.139815+00:00", "epoch": 0, "step": 1854, "train_loss": 3.797433376312256, "perplexity": 44.58660048422844, "lr": 0.001305, "grad_norm": 0.271084, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:02.446591+00:00", "epoch": 0, "step": 1855, "train_loss": 3.75689697265625, "perplexity": 42.81536239531881, "lr": 0.001305, "grad_norm": 0.241257, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:02.753840+00:00", "epoch": 0, "step": 1856, "train_loss": 3.820192813873291, "perplexity": 45.61300229266609, "lr": 0.001305, "grad_norm": 0.277406, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:03.061524+00:00", "epoch": 0, "step": 1857, "train_loss": 3.8212554454803467, "perplexity": 45.66149787249957, "lr": 0.001305, "grad_norm": 0.262594, "tokens_per_sec": 106500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:03.368325+00:00", "epoch": 0, "step": 1858, "train_loss": 3.744493007659912, "perplexity": 42.287562312922006, "lr": 0.001305, "grad_norm": 0.237008, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:03.674804+00:00", "epoch": 0, "step": 1859, "train_loss": 3.730517625808716, "perplexity": 41.70068793073952, "lr": 0.001305, "grad_norm": 0.258634, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:03.980904+00:00", "epoch": 0, "step": 1860, "train_loss": 3.755084753036499, "perplexity": 42.737841818929695, "lr": 0.001305, "grad_norm": 0.246279, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:04.287779+00:00", "epoch": 0, "step": 1861, "train_loss": 3.8791518211364746, "perplexity": 48.38316008908422, "lr": 0.001305, "grad_norm": 0.23165, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:04.593557+00:00", "epoch": 0, "step": 1862, "train_loss": 3.8120784759521484, "perplexity": 45.244380561689184, "lr": 0.001305, "grad_norm": 0.245648, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:04.901989+00:00", "epoch": 0, "step": 1863, "train_loss": 3.7562620639801025, "perplexity": 42.78818717806511, "lr": 0.001305, "grad_norm": 0.241598, "tokens_per_sec": 106241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:05.209757+00:00", "epoch": 0, "step": 1864, "train_loss": 3.8034579753875732, "perplexity": 44.856027656702075, "lr": 0.001305, "grad_norm": 0.221123, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:05.516950+00:00", "epoch": 0, "step": 1865, "train_loss": 3.7349979877471924, "perplexity": 41.88794127389048, "lr": 0.001305, "grad_norm": 0.237801, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:05.823854+00:00", "epoch": 0, "step": 1866, "train_loss": 3.6657094955444336, "perplexity": 39.08385615013069, "lr": 0.001305, "grad_norm": 0.238689, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:06.130210+00:00", "epoch": 0, "step": 1867, "train_loss": 3.6877076625823975, "perplexity": 39.95315578992291, "lr": 0.001305, "grad_norm": 0.220874, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:06.437911+00:00", "epoch": 0, "step": 1868, "train_loss": 3.7694251537323, "perplexity": 43.3551351335478, "lr": 0.001305, "grad_norm": 0.236509, "tokens_per_sec": 106493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:06.746198+00:00", "epoch": 0, "step": 1869, "train_loss": 3.7421834468841553, "perplexity": 42.19000931335527, "lr": 0.001305, "grad_norm": 0.225733, "tokens_per_sec": 106293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:07.053613+00:00", "epoch": 0, "step": 1870, "train_loss": 3.773144483566284, "perplexity": 43.516687428005746, "lr": 0.001305, "grad_norm": 0.231691, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:07.360744+00:00", "epoch": 0, "step": 1871, "train_loss": 3.7525219917297363, "perplexity": 42.62845515740206, "lr": 0.001305, "grad_norm": 0.263674, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:07.667400+00:00", "epoch": 0, "step": 1872, "train_loss": 3.760653018951416, "perplexity": 42.97648127418346, "lr": 0.001305, "grad_norm": 0.296655, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:07.974035+00:00", "epoch": 0, "step": 1873, "train_loss": 3.729917287826538, "perplexity": 41.67566093697088, "lr": 0.001305, "grad_norm": 0.308633, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:08.280880+00:00", "epoch": 0, "step": 1874, "train_loss": 3.776411771774292, "perplexity": 43.65910151493304, "lr": 0.001305, "grad_norm": 0.256982, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:08.589293+00:00", "epoch": 0, "step": 1875, "train_loss": 3.6807408332824707, "perplexity": 39.67577632316225, "lr": 0.001305, "grad_norm": 0.276896, "tokens_per_sec": 106247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:08.897266+00:00", "epoch": 0, "step": 1876, "train_loss": 3.8070785999298096, "perplexity": 45.01872885353992, "lr": 0.001305, "grad_norm": 0.310439, "tokens_per_sec": 106399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:09.203911+00:00", "epoch": 0, "step": 1877, "train_loss": 3.770815849304199, "perplexity": 43.4154708725954, "lr": 0.001305, "grad_norm": 0.302456, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:09.510548+00:00", "epoch": 0, "step": 1878, "train_loss": 3.814756155014038, "perplexity": 45.365692837373935, "lr": 0.001305, "grad_norm": 0.299124, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:09.816807+00:00", "epoch": 0, "step": 1879, "train_loss": 3.7994351387023926, "perplexity": 44.675941654245946, "lr": 0.001305, "grad_norm": 0.275754, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:10.123995+00:00", "epoch": 0, "step": 1880, "train_loss": 3.769233465194702, "perplexity": 43.34682524757709, "lr": 0.001305, "grad_norm": 0.243432, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:10.430885+00:00", "epoch": 0, "step": 1881, "train_loss": 3.675163984298706, "perplexity": 39.4551263479631, "lr": 0.001305, "grad_norm": 0.277932, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:10.737066+00:00", "epoch": 0, "step": 1882, "train_loss": 3.78475022315979, "perplexity": 44.02467284164592, "lr": 0.001305, "grad_norm": 0.259006, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:11.043819+00:00", "epoch": 0, "step": 1883, "train_loss": 3.793423891067505, "perplexity": 44.40818907544018, "lr": 0.001305, "grad_norm": 0.24883, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:11.351032+00:00", "epoch": 0, "step": 1884, "train_loss": 3.7980597019195557, "perplexity": 44.61453496098003, "lr": 0.001305, "grad_norm": 0.266861, "tokens_per_sec": 106663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:11.659022+00:00", "epoch": 0, "step": 1885, "train_loss": 3.7205114364624023, "perplexity": 41.28550362198417, "lr": 0.001305, "grad_norm": 0.300608, "tokens_per_sec": 106393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:11.967063+00:00", "epoch": 0, "step": 1886, "train_loss": 3.7688944339752197, "perplexity": 43.332131811460236, "lr": 0.001305, "grad_norm": 0.316292, "tokens_per_sec": 106376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:12.274950+00:00", "epoch": 0, "step": 1887, "train_loss": 3.8157401084899902, "perplexity": 45.41035253646016, "lr": 0.001305, "grad_norm": 0.27219, "tokens_per_sec": 106429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:12.581475+00:00", "epoch": 0, "step": 1888, "train_loss": 3.7729523181915283, "perplexity": 43.50832583088846, "lr": 0.001305, "grad_norm": 0.249277, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:12.889023+00:00", "epoch": 0, "step": 1889, "train_loss": 3.7539849281311035, "perplexity": 42.69086351478431, "lr": 0.001305, "grad_norm": 0.254689, "tokens_per_sec": 106546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:13.195541+00:00", "epoch": 0, "step": 1890, "train_loss": 3.758814573287964, "perplexity": 42.897543931795155, "lr": 0.001305, "grad_norm": 0.245082, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:13.502558+00:00", "epoch": 0, "step": 1891, "train_loss": 3.779874801635742, "perplexity": 43.81055638220622, "lr": 0.001305, "grad_norm": 0.243925, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:13.810641+00:00", "epoch": 0, "step": 1892, "train_loss": 3.728423595428467, "perplexity": 41.61345678754269, "lr": 0.001305, "grad_norm": 0.254969, "tokens_per_sec": 106357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:14.118529+00:00", "epoch": 0, "step": 1893, "train_loss": 3.747728109359741, "perplexity": 42.42458840480532, "lr": 0.001305, "grad_norm": 0.270013, "tokens_per_sec": 106426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:14.424812+00:00", "epoch": 0, "step": 1894, "train_loss": 3.838536500930786, "perplexity": 46.457434251771964, "lr": 0.001305, "grad_norm": 0.261823, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:14.730970+00:00", "epoch": 0, "step": 1895, "train_loss": 3.764039993286133, "perplexity": 43.12228829618967, "lr": 0.001305, "grad_norm": 0.237773, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:15.037309+00:00", "epoch": 0, "step": 1896, "train_loss": 3.703319549560547, "perplexity": 40.58179429097674, "lr": 0.001305, "grad_norm": 0.256436, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:15.344339+00:00", "epoch": 0, "step": 1897, "train_loss": 3.817675828933716, "perplexity": 45.498339415748276, "lr": 0.001305, "grad_norm": 0.238903, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:15.651190+00:00", "epoch": 0, "step": 1898, "train_loss": 3.741377115249634, "perplexity": 42.156003885853565, "lr": 0.001305, "grad_norm": 0.245575, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:15.957562+00:00", "epoch": 0, "step": 1899, "train_loss": 3.841500759124756, "perplexity": 46.5953503904586, "lr": 0.001305, "grad_norm": 0.261796, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:16.265293+00:00", "epoch": 0, "step": 1900, "train_loss": 3.681968927383423, "perplexity": 39.724531842070185, "lr": 0.001305, "grad_norm": 0.257459, "tokens_per_sec": 106544} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:55:21.931777+00:00", "step": 1900, "epoch": 0, "val_loss": 3.729093611240387, "val_ppl": 41.64134780424669, "eval_train_loss": 3.681968927383423, "eval_train_ppl": 39.724531842070185} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:55:22.877809+00:00", "step": 1900, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p7291_epoch_0000_step_0001900.pt", "val_loss": 3.729093611240387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:23.882315+00:00", "epoch": 0, "step": 1901, "train_loss": 3.7094621658325195, "perplexity": 40.8318398617095, "lr": 0.001305, "grad_norm": 0.256515, "tokens_per_sec": 4302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:24.188051+00:00", "epoch": 0, "step": 1902, "train_loss": 3.6750242710113525, "perplexity": 39.449614327618285, "lr": 0.001305, "grad_norm": 0.243601, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:24.494271+00:00", "epoch": 0, "step": 1903, "train_loss": 3.7342114448547363, "perplexity": 41.85500756498709, "lr": 0.001305, "grad_norm": 0.214585, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:24.799766+00:00", "epoch": 0, "step": 1904, "train_loss": 3.823742389678955, "perplexity": 45.775196792566895, "lr": 0.001305, "grad_norm": 0.247188, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:25.106154+00:00", "epoch": 0, "step": 1905, "train_loss": 3.8019230365753174, "perplexity": 44.78722921309648, "lr": 0.001305, "grad_norm": 0.25144, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:25.412330+00:00", "epoch": 0, "step": 1906, "train_loss": 3.6979706287384033, "perplexity": 40.365304994709604, "lr": 0.001305, "grad_norm": 0.274512, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:25.718940+00:00", "epoch": 0, "step": 1907, "train_loss": 3.7282822132110596, "perplexity": 41.60757380063275, "lr": 0.001305, "grad_norm": 0.268165, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:26.024593+00:00", "epoch": 0, "step": 1908, "train_loss": 3.6281490325927734, "perplexity": 37.64307598651721, "lr": 0.001305, "grad_norm": 0.291586, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:26.329359+00:00", "epoch": 0, "step": 1909, "train_loss": 3.7457292079925537, "perplexity": 42.33987053658017, "lr": 0.001305, "grad_norm": 0.279161, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:26.634274+00:00", "epoch": 0, "step": 1910, "train_loss": 3.767192840576172, "perplexity": 43.25846083879927, "lr": 0.001305, "grad_norm": 0.242947, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:26.940073+00:00", "epoch": 0, "step": 1911, "train_loss": 3.7262840270996094, "perplexity": 41.5245171335079, "lr": 0.001305, "grad_norm": 0.262731, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:27.247154+00:00", "epoch": 0, "step": 1912, "train_loss": 3.741291046142578, "perplexity": 42.15237571238114, "lr": 0.001305, "grad_norm": 0.261681, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:27.552380+00:00", "epoch": 0, "step": 1913, "train_loss": 3.8260715007781982, "perplexity": 45.881936567630795, "lr": 0.001305, "grad_norm": 0.258119, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:27.858497+00:00", "epoch": 0, "step": 1914, "train_loss": 3.7138757705688477, "perplexity": 41.01245374947633, "lr": 0.001305, "grad_norm": 0.259848, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:28.178986+00:00", "epoch": 0, "step": 1915, "train_loss": 3.8076913356781006, "perplexity": 45.04632189080797, "lr": 0.001305, "grad_norm": 0.237133, "tokens_per_sec": 102244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:28.484761+00:00", "epoch": 0, "step": 1916, "train_loss": 3.7365288734436035, "perplexity": 41.95211603361627, "lr": 0.001305, "grad_norm": 0.241442, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:28.791296+00:00", "epoch": 0, "step": 1917, "train_loss": 3.866520643234253, "perplexity": 47.77586727260252, "lr": 0.001305, "grad_norm": 0.237476, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:29.098289+00:00", "epoch": 0, "step": 1918, "train_loss": 3.804161548614502, "perplexity": 44.88759826164157, "lr": 0.001305, "grad_norm": 0.236759, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:29.405471+00:00", "epoch": 0, "step": 1919, "train_loss": 3.7214114665985107, "perplexity": 41.32267854619494, "lr": 0.001305, "grad_norm": 0.220549, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:29.712210+00:00", "epoch": 0, "step": 1920, "train_loss": 3.687220573425293, "perplexity": 39.933699779736, "lr": 0.001305, "grad_norm": 0.271937, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:30.017585+00:00", "epoch": 0, "step": 1921, "train_loss": 3.6797971725463867, "perplexity": 39.6383535108688, "lr": 0.001305, "grad_norm": 0.27471, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:30.323689+00:00", "epoch": 0, "step": 1922, "train_loss": 3.7574238777160645, "perplexity": 42.83792797083858, "lr": 0.001305, "grad_norm": 0.31538, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:30.630732+00:00", "epoch": 0, "step": 1923, "train_loss": 3.6969876289367676, "perplexity": 40.325645403785394, "lr": 0.001305, "grad_norm": 0.325464, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:30.937844+00:00", "epoch": 0, "step": 1924, "train_loss": 3.7412936687469482, "perplexity": 42.15248626153086, "lr": 0.001305, "grad_norm": 0.292057, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:31.244321+00:00", "epoch": 0, "step": 1925, "train_loss": 3.6829416751861572, "perplexity": 39.76319259366627, "lr": 0.001305, "grad_norm": 0.271704, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:31.550368+00:00", "epoch": 0, "step": 1926, "train_loss": 3.704115629196167, "perplexity": 40.6141134936114, "lr": 0.001305, "grad_norm": 0.263619, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:31.856375+00:00", "epoch": 0, "step": 1927, "train_loss": 3.833601474761963, "perplexity": 46.2287303919332, "lr": 0.001305, "grad_norm": 0.284097, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:32.163078+00:00", "epoch": 0, "step": 1928, "train_loss": 3.7769482135772705, "perplexity": 43.68252836507521, "lr": 0.001305, "grad_norm": 0.282712, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:32.470270+00:00", "epoch": 0, "step": 1929, "train_loss": 3.7728514671325684, "perplexity": 43.503938191407606, "lr": 0.001305, "grad_norm": 0.249379, "tokens_per_sec": 106670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:32.777687+00:00", "epoch": 0, "step": 1930, "train_loss": 3.6756298542022705, "perplexity": 39.47351158610199, "lr": 0.001305, "grad_norm": 0.253168, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:33.084030+00:00", "epoch": 0, "step": 1931, "train_loss": 3.7961156368255615, "perplexity": 44.52788565411418, "lr": 0.001305, "grad_norm": 0.25907, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:33.395341+00:00", "epoch": 0, "step": 1932, "train_loss": 3.677591562271118, "perplexity": 39.55102309490504, "lr": 0.001305, "grad_norm": 0.278214, "tokens_per_sec": 105258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:33.702004+00:00", "epoch": 0, "step": 1933, "train_loss": 3.689751386642456, "perplexity": 40.03489251088778, "lr": 0.001305, "grad_norm": 0.268795, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:34.007862+00:00", "epoch": 0, "step": 1934, "train_loss": 3.7540245056152344, "perplexity": 42.69255314519304, "lr": 0.001305, "grad_norm": 0.274294, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:34.315912+00:00", "epoch": 0, "step": 1935, "train_loss": 3.762080192565918, "perplexity": 43.0378599629065, "lr": 0.001305, "grad_norm": 0.251957, "tokens_per_sec": 106372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:34.622888+00:00", "epoch": 0, "step": 1936, "train_loss": 3.614819049835205, "perplexity": 37.1446239916648, "lr": 0.001305, "grad_norm": 0.256075, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:34.929110+00:00", "epoch": 0, "step": 1937, "train_loss": 3.7280616760253906, "perplexity": 41.59839879515638, "lr": 0.001305, "grad_norm": 0.266388, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:35.234160+00:00", "epoch": 0, "step": 1938, "train_loss": 3.795771837234497, "perplexity": 44.512579616490825, "lr": 0.001305, "grad_norm": 0.278365, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:35.540262+00:00", "epoch": 0, "step": 1939, "train_loss": 3.7445199489593506, "perplexity": 42.288701610147804, "lr": 0.001305, "grad_norm": 0.257606, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:55:41.196282+00:00", "step": 1939, "epoch": 1, "val_loss": 3.722841036319733, "val_ppl": 41.38179444132346, "eval_train_loss": 3.7445199489593506, "eval_train_ppl": 42.288701610147804} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:55:42.150828+00:00", "step": 1939, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p7228_epoch_0001_step_0001939.pt", "val_loss": 3.722841036319733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:43.236531+00:00", "epoch": 1, "step": 1940, "train_loss": 3.8277101516723633, "perplexity": 45.95718267822915, "lr": 0.001305, "grad_norm": 0.247078, "tokens_per_sec": 4258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:43.542016+00:00", "epoch": 1, "step": 1941, "train_loss": 3.745192766189575, "perplexity": 42.31716375107065, "lr": 0.001305, "grad_norm": 0.254774, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:43.847587+00:00", "epoch": 1, "step": 1942, "train_loss": 3.7206671237945557, "perplexity": 41.29193175227593, "lr": 0.001305, "grad_norm": 0.23991, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:44.153301+00:00", "epoch": 1, "step": 1943, "train_loss": 3.5914714336395264, "perplexity": 36.287431209444236, "lr": 0.001305, "grad_norm": 0.244246, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:44.458778+00:00", "epoch": 1, "step": 1944, "train_loss": 3.6668295860290527, "perplexity": 39.12765813201873, "lr": 0.001305, "grad_norm": 0.224399, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:44.763988+00:00", "epoch": 1, "step": 1945, "train_loss": 3.732232093811035, "perplexity": 41.772243748423776, "lr": 0.001305, "grad_norm": 0.243157, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:45.069497+00:00", "epoch": 1, "step": 1946, "train_loss": 3.6270768642425537, "perplexity": 37.60273790031342, "lr": 0.001305, "grad_norm": 0.252021, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:45.376711+00:00", "epoch": 1, "step": 1947, "train_loss": 3.689119338989258, "perplexity": 40.00959654599996, "lr": 0.001305, "grad_norm": 0.225027, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:45.682191+00:00", "epoch": 1, "step": 1948, "train_loss": 3.6977603435516357, "perplexity": 40.35681766142137, "lr": 0.001305, "grad_norm": 0.232267, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:45.987813+00:00", "epoch": 1, "step": 1949, "train_loss": 3.751028060913086, "perplexity": 42.56481874061374, "lr": 0.001305, "grad_norm": 0.265687, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:46.293390+00:00", "epoch": 1, "step": 1950, "train_loss": 3.7918918132781982, "perplexity": 44.34020436752303, "lr": 0.001305, "grad_norm": 0.272402, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:46.599600+00:00", "epoch": 1, "step": 1951, "train_loss": 3.7440223693847656, "perplexity": 42.26766485015595, "lr": 0.001305, "grad_norm": 0.248972, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:46.906557+00:00", "epoch": 1, "step": 1952, "train_loss": 3.807488203048706, "perplexity": 45.037172442304964, "lr": 0.001305, "grad_norm": 0.253936, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:47.212569+00:00", "epoch": 1, "step": 1953, "train_loss": 3.6968629360198975, "perplexity": 40.32061739492038, "lr": 0.001305, "grad_norm": 0.269084, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:47.526900+00:00", "epoch": 1, "step": 1954, "train_loss": 3.6907246112823486, "perplexity": 40.07387442073555, "lr": 0.001305, "grad_norm": 0.257597, "tokens_per_sec": 104246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:47.832038+00:00", "epoch": 1, "step": 1955, "train_loss": 3.727412700653076, "perplexity": 41.571411216894624, "lr": 0.001305, "grad_norm": 0.243811, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:48.137475+00:00", "epoch": 1, "step": 1956, "train_loss": 3.774869441986084, "perplexity": 43.59181668317831, "lr": 0.001305, "grad_norm": 0.256465, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:48.443891+00:00", "epoch": 1, "step": 1957, "train_loss": 3.6906330585479736, "perplexity": 40.07020571589788, "lr": 0.001305, "grad_norm": 0.232516, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:48.748857+00:00", "epoch": 1, "step": 1958, "train_loss": 3.727325677871704, "perplexity": 41.567793714469815, "lr": 0.001305, "grad_norm": 0.225884, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:49.054306+00:00", "epoch": 1, "step": 1959, "train_loss": 3.600820302963257, "perplexity": 36.6282684006299, "lr": 0.001305, "grad_norm": 0.238739, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:49.360348+00:00", "epoch": 1, "step": 1960, "train_loss": 3.7537145614624023, "perplexity": 42.679322888402496, "lr": 0.001305, "grad_norm": 0.252992, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:49.666451+00:00", "epoch": 1, "step": 1961, "train_loss": 3.7132742404937744, "perplexity": 40.987790943546884, "lr": 0.001305, "grad_norm": 0.234575, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:49.972720+00:00", "epoch": 1, "step": 1962, "train_loss": 3.714987277984619, "perplexity": 41.05806473973898, "lr": 0.001305, "grad_norm": 0.263541, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:50.278745+00:00", "epoch": 1, "step": 1963, "train_loss": 3.700751543045044, "perplexity": 40.47771367587357, "lr": 0.001305, "grad_norm": 0.280086, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:50.584951+00:00", "epoch": 1, "step": 1964, "train_loss": 3.7767245769500732, "perplexity": 43.67276044403736, "lr": 0.001305, "grad_norm": 0.26813, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:50.891956+00:00", "epoch": 1, "step": 1965, "train_loss": 3.732985019683838, "perplexity": 41.80370699476672, "lr": 0.001305, "grad_norm": 0.240603, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:51.198539+00:00", "epoch": 1, "step": 1966, "train_loss": 3.6257524490356445, "perplexity": 37.552969226885615, "lr": 0.001305, "grad_norm": 0.248746, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:51.504454+00:00", "epoch": 1, "step": 1967, "train_loss": 3.6762759685516357, "perplexity": 39.49902422951256, "lr": 0.001305, "grad_norm": 0.244742, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:51.810042+00:00", "epoch": 1, "step": 1968, "train_loss": 3.7002992630004883, "perplexity": 40.45941055310932, "lr": 0.001305, "grad_norm": 0.280191, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:52.115817+00:00", "epoch": 1, "step": 1969, "train_loss": 3.771878480911255, "perplexity": 43.46163004493404, "lr": 0.001305, "grad_norm": 0.304062, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:52.421276+00:00", "epoch": 1, "step": 1970, "train_loss": 3.7633090019226074, "perplexity": 43.09077779422524, "lr": 0.001305, "grad_norm": 0.302217, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:52.727733+00:00", "epoch": 1, "step": 1971, "train_loss": 3.7846968173980713, "perplexity": 44.02232173324032, "lr": 0.001305, "grad_norm": 0.375396, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:53.032694+00:00", "epoch": 1, "step": 1972, "train_loss": 3.850484848022461, "perplexity": 47.01585324975766, "lr": 0.001305, "grad_norm": 0.355262, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:53.337948+00:00", "epoch": 1, "step": 1973, "train_loss": 3.668292999267578, "perplexity": 39.1849599828388, "lr": 0.001305, "grad_norm": 0.307168, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:53.643569+00:00", "epoch": 1, "step": 1974, "train_loss": 3.792058229446411, "perplexity": 44.34758390845208, "lr": 0.001305, "grad_norm": 0.292918, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:53.949888+00:00", "epoch": 1, "step": 1975, "train_loss": 3.71298885345459, "perplexity": 40.97609522822896, "lr": 0.001305, "grad_norm": 0.292578, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:54.256434+00:00", "epoch": 1, "step": 1976, "train_loss": 3.715541124343872, "perplexity": 41.08081089777002, "lr": 0.001305, "grad_norm": 0.306907, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:54.561960+00:00", "epoch": 1, "step": 1977, "train_loss": 3.6028621196746826, "perplexity": 36.70313301506201, "lr": 0.001305, "grad_norm": 0.28807, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:54.866756+00:00", "epoch": 1, "step": 1978, "train_loss": 3.7913763523101807, "perplexity": 44.31735461244202, "lr": 0.001305, "grad_norm": 0.263069, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:55.172173+00:00", "epoch": 1, "step": 1979, "train_loss": 3.7306160926818848, "perplexity": 41.70479426925489, "lr": 0.001305, "grad_norm": 0.28182, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:55.478708+00:00", "epoch": 1, "step": 1980, "train_loss": 3.6716156005859375, "perplexity": 39.315372516990564, "lr": 0.001305, "grad_norm": 0.299807, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:55.784673+00:00", "epoch": 1, "step": 1981, "train_loss": 3.8192992210388184, "perplexity": 45.57226104641246, "lr": 0.001305, "grad_norm": 0.267908, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:56.091245+00:00", "epoch": 1, "step": 1982, "train_loss": 3.781700372695923, "perplexity": 43.89060871444591, "lr": 0.001305, "grad_norm": 0.261729, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:56.397453+00:00", "epoch": 1, "step": 1983, "train_loss": 3.743610382080078, "perplexity": 42.25025469546651, "lr": 0.001305, "grad_norm": 0.259911, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:56.703797+00:00", "epoch": 1, "step": 1984, "train_loss": 3.788390636444092, "perplexity": 44.18523292069809, "lr": 0.001305, "grad_norm": 0.263927, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:57.009736+00:00", "epoch": 1, "step": 1985, "train_loss": 3.730895757675171, "perplexity": 41.716459271334536, "lr": 0.001305, "grad_norm": 0.27485, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:57.316506+00:00", "epoch": 1, "step": 1986, "train_loss": 3.7733399868011475, "perplexity": 43.52519591285951, "lr": 0.001305, "grad_norm": 0.2618, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:57.622528+00:00", "epoch": 1, "step": 1987, "train_loss": 3.8459737300872803, "perplexity": 46.80423686303289, "lr": 0.001305, "grad_norm": 0.228226, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:57.928770+00:00", "epoch": 1, "step": 1988, "train_loss": 3.659318447113037, "perplexity": 38.83486583468255, "lr": 0.001305, "grad_norm": 0.240956, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:58.235257+00:00", "epoch": 1, "step": 1989, "train_loss": 3.7218217849731445, "perplexity": 41.33963747953458, "lr": 0.001305, "grad_norm": 0.234495, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:58.541020+00:00", "epoch": 1, "step": 1990, "train_loss": 3.7742111682891846, "perplexity": 43.56313077947951, "lr": 0.001305, "grad_norm": 0.233975, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:58.847655+00:00", "epoch": 1, "step": 1991, "train_loss": 3.741842031478882, "perplexity": 42.17560745287547, "lr": 0.001305, "grad_norm": 0.21843, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:59.154264+00:00", "epoch": 1, "step": 1992, "train_loss": 3.800086498260498, "perplexity": 44.70505123523251, "lr": 0.001305, "grad_norm": 0.229332, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:59.461232+00:00", "epoch": 1, "step": 1993, "train_loss": 3.6349194049835205, "perplexity": 37.8987983197768, "lr": 0.001305, "grad_norm": 0.215875, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:55:59.767583+00:00", "epoch": 1, "step": 1994, "train_loss": 3.72037935256958, "perplexity": 41.280050832069435, "lr": 0.001305, "grad_norm": 0.24415, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:00.073236+00:00", "epoch": 1, "step": 1995, "train_loss": 3.8220767974853516, "perplexity": 45.699017441606046, "lr": 0.001305, "grad_norm": 0.250499, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:00.378614+00:00", "epoch": 1, "step": 1996, "train_loss": 3.7482709884643555, "perplexity": 42.447626080142626, "lr": 0.001305, "grad_norm": 0.256568, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:00.684867+00:00", "epoch": 1, "step": 1997, "train_loss": 3.6935253143310547, "perplexity": 40.18626675831389, "lr": 0.001305, "grad_norm": 0.25554, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:00.991342+00:00", "epoch": 1, "step": 1998, "train_loss": 3.6543517112731934, "perplexity": 38.642461520932756, "lr": 0.001305, "grad_norm": 0.255223, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:01.297003+00:00", "epoch": 1, "step": 1999, "train_loss": 3.7209601402282715, "perplexity": 41.30403273966672, "lr": 0.001305, "grad_norm": 0.274458, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:01.602830+00:00", "epoch": 1, "step": 2000, "train_loss": 3.7044265270233154, "perplexity": 40.62674229628006, "lr": 0.001305, "grad_norm": 0.227985, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:56:07.348062+00:00", "step": 2000, "epoch": 1, "val_loss": 3.713191533088684, "val_ppl": 40.98440108990249, "eval_train_loss": 3.7044265270233154, "eval_train_ppl": 40.62674229628006} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:56:08.267358+00:00", "step": 2000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p7132_epoch_0001_step_0002000.pt", "val_loss": 3.713191533088684} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:56:09.982614+00:00", "step": 2000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0001_step_0002000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:11.012480+00:00", "epoch": 1, "step": 2001, "train_loss": 3.6656835079193115, "perplexity": 39.082840466726395, "lr": 0.001305, "grad_norm": 0.216482, "tokens_per_sec": 3482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:11.316056+00:00", "epoch": 1, "step": 2002, "train_loss": 3.78790545463562, "perplexity": 44.163800249273464, "lr": 0.001305, "grad_norm": 0.234575, "tokens_per_sec": 107940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:11.621345+00:00", "epoch": 1, "step": 2003, "train_loss": 3.6637635231018066, "perplexity": 39.007873996676814, "lr": 0.001305, "grad_norm": 0.238677, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:11.926456+00:00", "epoch": 1, "step": 2004, "train_loss": 3.709749221801758, "perplexity": 40.84356256753263, "lr": 0.001305, "grad_norm": 0.243703, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:12.232279+00:00", "epoch": 1, "step": 2005, "train_loss": 3.7481489181518555, "perplexity": 42.44244480140873, "lr": 0.001305, "grad_norm": 0.264731, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:12.538075+00:00", "epoch": 1, "step": 2006, "train_loss": 3.775686264038086, "perplexity": 43.62743798648445, "lr": 0.001305, "grad_norm": 0.287127, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:12.843656+00:00", "epoch": 1, "step": 2007, "train_loss": 3.8260788917541504, "perplexity": 45.882275681173795, "lr": 0.001305, "grad_norm": 0.274306, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:13.148508+00:00", "epoch": 1, "step": 2008, "train_loss": 3.6410417556762695, "perplexity": 38.13153978938271, "lr": 0.001305, "grad_norm": 0.263705, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:13.453682+00:00", "epoch": 1, "step": 2009, "train_loss": 3.795912504196167, "perplexity": 44.518841506231745, "lr": 0.001305, "grad_norm": 0.273725, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:13.759968+00:00", "epoch": 1, "step": 2010, "train_loss": 3.7153689861297607, "perplexity": 41.073739928957266, "lr": 0.001305, "grad_norm": 0.257559, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:14.066326+00:00", "epoch": 1, "step": 2011, "train_loss": 3.705462694168091, "perplexity": 40.66886020867694, "lr": 0.001305, "grad_norm": 0.228009, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:14.371535+00:00", "epoch": 1, "step": 2012, "train_loss": 3.722365617752075, "perplexity": 41.36212544375685, "lr": 0.001305, "grad_norm": 0.212024, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:14.677017+00:00", "epoch": 1, "step": 2013, "train_loss": 3.7904317378997803, "perplexity": 44.27551156652772, "lr": 0.001305, "grad_norm": 0.240814, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:14.982856+00:00", "epoch": 1, "step": 2014, "train_loss": 3.688375949859619, "perplexity": 39.979864899307124, "lr": 0.001305, "grad_norm": 0.221225, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:15.648240+00:00", "epoch": 1, "step": 2015, "train_loss": 3.6253163814544678, "perplexity": 37.536597164351136, "lr": 0.001305, "grad_norm": 0.227458, "tokens_per_sec": 49247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:15.954530+00:00", "epoch": 1, "step": 2016, "train_loss": 3.7682600021362305, "perplexity": 43.30464924621679, "lr": 0.001305, "grad_norm": 0.254635, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:16.261055+00:00", "epoch": 1, "step": 2017, "train_loss": 3.8119397163391113, "perplexity": 45.23810290450305, "lr": 0.001305, "grad_norm": 0.226392, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:16.567245+00:00", "epoch": 1, "step": 2018, "train_loss": 3.7633919715881348, "perplexity": 43.09435316996795, "lr": 0.001305, "grad_norm": 0.224039, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:16.873455+00:00", "epoch": 1, "step": 2019, "train_loss": 3.7827885150909424, "perplexity": 43.93839394038139, "lr": 0.001305, "grad_norm": 0.2454, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:17.179010+00:00", "epoch": 1, "step": 2020, "train_loss": 3.8704276084899902, "perplexity": 47.96289103587179, "lr": 0.001305, "grad_norm": 0.250189, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:17.484658+00:00", "epoch": 1, "step": 2021, "train_loss": 3.74900484085083, "perplexity": 42.47878780449613, "lr": 0.001305, "grad_norm": 0.262761, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:17.791064+00:00", "epoch": 1, "step": 2022, "train_loss": 3.7126550674438477, "perplexity": 40.962420263249896, "lr": 0.001305, "grad_norm": 0.286739, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:18.096954+00:00", "epoch": 1, "step": 2023, "train_loss": 3.786205768585205, "perplexity": 44.08879941105173, "lr": 0.001305, "grad_norm": 0.274769, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:18.402595+00:00", "epoch": 1, "step": 2024, "train_loss": 3.636291742324829, "perplexity": 37.95084395982123, "lr": 0.001305, "grad_norm": 0.29384, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:18.708019+00:00", "epoch": 1, "step": 2025, "train_loss": 3.6263012886047363, "perplexity": 37.573585439313284, "lr": 0.001305, "grad_norm": 0.286395, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:19.013127+00:00", "epoch": 1, "step": 2026, "train_loss": 3.7477307319641113, "perplexity": 42.42469966786217, "lr": 0.001305, "grad_norm": 0.286952, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:19.320026+00:00", "epoch": 1, "step": 2027, "train_loss": 3.7210171222686768, "perplexity": 41.30638639478659, "lr": 0.001305, "grad_norm": 0.298484, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:19.626910+00:00", "epoch": 1, "step": 2028, "train_loss": 3.7624642848968506, "perplexity": 43.05439364988593, "lr": 0.001305, "grad_norm": 0.284061, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:19.933401+00:00", "epoch": 1, "step": 2029, "train_loss": 3.7445714473724365, "perplexity": 42.2908794672498, "lr": 0.001305, "grad_norm": 0.29362, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:20.239241+00:00", "epoch": 1, "step": 2030, "train_loss": 3.6939213275909424, "perplexity": 40.2021842043674, "lr": 0.001305, "grad_norm": 0.316108, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:20.546487+00:00", "epoch": 1, "step": 2031, "train_loss": 3.659034490585327, "perplexity": 38.82383998653091, "lr": 0.001305, "grad_norm": 0.294212, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:20.852329+00:00", "epoch": 1, "step": 2032, "train_loss": 3.7140562534332275, "perplexity": 41.01985646261563, "lr": 0.001305, "grad_norm": 0.291373, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:21.160183+00:00", "epoch": 1, "step": 2033, "train_loss": 3.6760692596435547, "perplexity": 39.49086027315413, "lr": 0.001305, "grad_norm": 0.284905, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:21.466991+00:00", "epoch": 1, "step": 2034, "train_loss": 3.7984585762023926, "perplexity": 44.63233410119035, "lr": 0.001305, "grad_norm": 0.296598, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:21.774373+00:00", "epoch": 1, "step": 2035, "train_loss": 3.7668235301971436, "perplexity": 43.2424879898817, "lr": 0.001305, "grad_norm": 0.269963, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:22.080086+00:00", "epoch": 1, "step": 2036, "train_loss": 3.7416908740997314, "perplexity": 42.169232780390345, "lr": 0.001305, "grad_norm": 0.278167, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:22.386231+00:00", "epoch": 1, "step": 2037, "train_loss": 3.837064027786255, "perplexity": 46.3890772667356, "lr": 0.001305, "grad_norm": 0.275016, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:22.692686+00:00", "epoch": 1, "step": 2038, "train_loss": 3.682157278060913, "perplexity": 39.73201468923317, "lr": 0.001305, "grad_norm": 0.24019, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:22.999362+00:00", "epoch": 1, "step": 2039, "train_loss": 3.8432719707489014, "perplexity": 46.67795374912063, "lr": 0.001305, "grad_norm": 0.242921, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:23.305357+00:00", "epoch": 1, "step": 2040, "train_loss": 3.744515895843506, "perplexity": 42.288530209488606, "lr": 0.001305, "grad_norm": 0.26669, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:23.611078+00:00", "epoch": 1, "step": 2041, "train_loss": 3.6919071674346924, "perplexity": 40.12129205896406, "lr": 0.001305, "grad_norm": 0.26857, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:23.916574+00:00", "epoch": 1, "step": 2042, "train_loss": 3.774129629135132, "perplexity": 43.55957882346158, "lr": 0.001305, "grad_norm": 0.274311, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:24.223639+00:00", "epoch": 1, "step": 2043, "train_loss": 3.661459445953369, "perplexity": 38.91810030805984, "lr": 0.001305, "grad_norm": 0.260239, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:24.530067+00:00", "epoch": 1, "step": 2044, "train_loss": 3.685136079788208, "perplexity": 39.85054493462242, "lr": 0.001305, "grad_norm": 0.273673, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:24.836728+00:00", "epoch": 1, "step": 2045, "train_loss": 3.8229098320007324, "perplexity": 45.73710216119446, "lr": 0.001305, "grad_norm": 0.238271, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:25.142903+00:00", "epoch": 1, "step": 2046, "train_loss": 3.7893316745758057, "perplexity": 44.226832480051044, "lr": 0.001305, "grad_norm": 0.253958, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:25.448183+00:00", "epoch": 1, "step": 2047, "train_loss": 3.7529053688049316, "perplexity": 42.64480106298293, "lr": 0.001305, "grad_norm": 0.235572, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:25.754780+00:00", "epoch": 1, "step": 2048, "train_loss": 3.6321892738342285, "perplexity": 37.795470743061465, "lr": 0.001305, "grad_norm": 0.238638, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:26.061213+00:00", "epoch": 1, "step": 2049, "train_loss": 3.692378044128418, "perplexity": 40.14018868895896, "lr": 0.001305, "grad_norm": 0.234155, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:26.367912+00:00", "epoch": 1, "step": 2050, "train_loss": 3.793919086456299, "perplexity": 44.43018525164805, "lr": 0.001305, "grad_norm": 0.236228, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:26.674449+00:00", "epoch": 1, "step": 2051, "train_loss": 3.675055503845215, "perplexity": 39.450846470110065, "lr": 0.001305, "grad_norm": 0.229224, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:26.980788+00:00", "epoch": 1, "step": 2052, "train_loss": 3.7157487869262695, "perplexity": 41.089342730888625, "lr": 0.001305, "grad_norm": 0.210479, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:27.287454+00:00", "epoch": 1, "step": 2053, "train_loss": 3.7089874744415283, "perplexity": 40.81246193847935, "lr": 0.001305, "grad_norm": 0.22422, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:27.593490+00:00", "epoch": 1, "step": 2054, "train_loss": 3.704195499420166, "perplexity": 40.61735748150095, "lr": 0.001305, "grad_norm": 0.247878, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:27.900013+00:00", "epoch": 1, "step": 2055, "train_loss": 3.7282183170318604, "perplexity": 41.60491532057541, "lr": 0.001305, "grad_norm": 0.24334, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:28.207072+00:00", "epoch": 1, "step": 2056, "train_loss": 3.718191623687744, "perplexity": 41.18983998703631, "lr": 0.001305, "grad_norm": 0.23208, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:28.513643+00:00", "epoch": 1, "step": 2057, "train_loss": 3.7674331665039062, "perplexity": 43.26885821786256, "lr": 0.001305, "grad_norm": 0.235761, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:28.819645+00:00", "epoch": 1, "step": 2058, "train_loss": 3.7369186878204346, "perplexity": 41.968472759420976, "lr": 0.001305, "grad_norm": 0.24014, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:29.125898+00:00", "epoch": 1, "step": 2059, "train_loss": 3.6661529541015625, "perplexity": 39.10119206418168, "lr": 0.001305, "grad_norm": 0.221244, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:29.432352+00:00", "epoch": 1, "step": 2060, "train_loss": 3.7188398838043213, "perplexity": 41.21655037421205, "lr": 0.001305, "grad_norm": 0.229564, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:29.739353+00:00", "epoch": 1, "step": 2061, "train_loss": 3.7892308235168457, "perplexity": 44.22237238206757, "lr": 0.001305, "grad_norm": 0.223727, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:30.045423+00:00", "epoch": 1, "step": 2062, "train_loss": 3.6837968826293945, "perplexity": 39.797212917082675, "lr": 0.001305, "grad_norm": 0.231789, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:30.351820+00:00", "epoch": 1, "step": 2063, "train_loss": 3.732693672180176, "perplexity": 41.7915293631374, "lr": 0.001305, "grad_norm": 0.228687, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:30.657354+00:00", "epoch": 1, "step": 2064, "train_loss": 3.6355714797973633, "perplexity": 37.92351923070178, "lr": 0.001305, "grad_norm": 0.237278, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:30.964146+00:00", "epoch": 1, "step": 2065, "train_loss": 3.6356606483459473, "perplexity": 37.92690096663878, "lr": 0.001305, "grad_norm": 0.232207, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:31.271058+00:00", "epoch": 1, "step": 2066, "train_loss": 3.669893741607666, "perplexity": 39.2477352374775, "lr": 0.001305, "grad_norm": 0.261677, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:31.578320+00:00", "epoch": 1, "step": 2067, "train_loss": 3.7495219707489014, "perplexity": 42.500760536592345, "lr": 0.001305, "grad_norm": 0.280248, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:31.884894+00:00", "epoch": 1, "step": 2068, "train_loss": 3.6834468841552734, "perplexity": 39.7832863905611, "lr": 0.001305, "grad_norm": 0.304173, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:32.190840+00:00", "epoch": 1, "step": 2069, "train_loss": 3.649858236312866, "perplexity": 38.46921212510509, "lr": 0.001305, "grad_norm": 0.332023, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:32.497489+00:00", "epoch": 1, "step": 2070, "train_loss": 3.7636547088623047, "perplexity": 43.10567715040268, "lr": 0.001305, "grad_norm": 0.286333, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:32.804959+00:00", "epoch": 1, "step": 2071, "train_loss": 3.723877191543579, "perplexity": 41.424694625592785, "lr": 0.001305, "grad_norm": 0.270234, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:33.111978+00:00", "epoch": 1, "step": 2072, "train_loss": 3.643155813217163, "perplexity": 38.212237328185836, "lr": 0.001305, "grad_norm": 0.280503, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:33.418819+00:00", "epoch": 1, "step": 2073, "train_loss": 3.8271501064300537, "perplexity": 45.93145178262563, "lr": 0.001305, "grad_norm": 0.325112, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:33.725615+00:00", "epoch": 1, "step": 2074, "train_loss": 3.7598772048950195, "perplexity": 42.94315244607484, "lr": 0.001305, "grad_norm": 0.334743, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:34.031692+00:00", "epoch": 1, "step": 2075, "train_loss": 3.7318766117095947, "perplexity": 41.757397102449005, "lr": 0.001305, "grad_norm": 0.271484, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:34.337860+00:00", "epoch": 1, "step": 2076, "train_loss": 3.6837246417999268, "perplexity": 39.79433803725414, "lr": 0.001305, "grad_norm": 0.28567, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:34.645124+00:00", "epoch": 1, "step": 2077, "train_loss": 3.659188747406006, "perplexity": 38.829829290587405, "lr": 0.001305, "grad_norm": 0.281611, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:34.955385+00:00", "epoch": 1, "step": 2078, "train_loss": 3.5662763118743896, "perplexity": 35.384586364033545, "lr": 0.001305, "grad_norm": 0.272772, "tokens_per_sec": 105615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:35.262147+00:00", "epoch": 1, "step": 2079, "train_loss": 3.73146390914917, "perplexity": 41.740167273389986, "lr": 0.001305, "grad_norm": 0.264195, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:35.568528+00:00", "epoch": 1, "step": 2080, "train_loss": 3.7682106494903564, "perplexity": 43.302512099935186, "lr": 0.001305, "grad_norm": 0.292831, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:35.875157+00:00", "epoch": 1, "step": 2081, "train_loss": 3.7576074600219727, "perplexity": 42.84579297835182, "lr": 0.001305, "grad_norm": 0.25474, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:36.182678+00:00", "epoch": 1, "step": 2082, "train_loss": 3.7214198112487793, "perplexity": 41.32302337093429, "lr": 0.001305, "grad_norm": 0.255793, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:36.489911+00:00", "epoch": 1, "step": 2083, "train_loss": 3.640869617462158, "perplexity": 38.124976459138225, "lr": 0.001305, "grad_norm": 0.260674, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:36.798171+00:00", "epoch": 1, "step": 2084, "train_loss": 3.8425798416137695, "perplexity": 46.64565775515098, "lr": 0.001305, "grad_norm": 0.247956, "tokens_per_sec": 106299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:37.104901+00:00", "epoch": 1, "step": 2085, "train_loss": 3.667917251586914, "perplexity": 39.17023909085204, "lr": 0.001305, "grad_norm": 0.269098, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:37.410960+00:00", "epoch": 1, "step": 2086, "train_loss": 3.755850315093994, "perplexity": 42.77057281625173, "lr": 0.001305, "grad_norm": 0.262631, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:37.717641+00:00", "epoch": 1, "step": 2087, "train_loss": 3.695991277694702, "perplexity": 40.28548690620688, "lr": 0.001305, "grad_norm": 0.229597, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:38.023978+00:00", "epoch": 1, "step": 2088, "train_loss": 3.820956230163574, "perplexity": 45.64783729677801, "lr": 0.001305, "grad_norm": 0.217414, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:38.332243+00:00", "epoch": 1, "step": 2089, "train_loss": 3.7508065700531006, "perplexity": 42.555392066305394, "lr": 0.001305, "grad_norm": 0.257382, "tokens_per_sec": 106299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:38.639796+00:00", "epoch": 1, "step": 2090, "train_loss": 3.7664568424224854, "perplexity": 43.22663440502552, "lr": 0.001305, "grad_norm": 0.273112, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:38.946825+00:00", "epoch": 1, "step": 2091, "train_loss": 3.685075044631958, "perplexity": 39.8481127246116, "lr": 0.001305, "grad_norm": 0.274612, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:39.253075+00:00", "epoch": 1, "step": 2092, "train_loss": 3.7112903594970703, "perplexity": 40.90655665022589, "lr": 0.001305, "grad_norm": 0.25698, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:39.559651+00:00", "epoch": 1, "step": 2093, "train_loss": 3.894219160079956, "perplexity": 49.11768533549959, "lr": 0.001305, "grad_norm": 0.228464, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:39.865909+00:00", "epoch": 1, "step": 2094, "train_loss": 3.6257362365722656, "perplexity": 37.55236040568251, "lr": 0.001305, "grad_norm": 0.246688, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:40.173701+00:00", "epoch": 1, "step": 2095, "train_loss": 3.6278865337371826, "perplexity": 37.633196018946656, "lr": 0.001305, "grad_norm": 0.246598, "tokens_per_sec": 106461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:40.482002+00:00", "epoch": 1, "step": 2096, "train_loss": 3.6175808906555176, "perplexity": 37.247353326195324, "lr": 0.001305, "grad_norm": 0.224236, "tokens_per_sec": 106286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:40.789314+00:00", "epoch": 1, "step": 2097, "train_loss": 3.680955171585083, "perplexity": 39.68428127314989, "lr": 0.001305, "grad_norm": 0.226447, "tokens_per_sec": 106627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:41.095549+00:00", "epoch": 1, "step": 2098, "train_loss": 3.661893606185913, "perplexity": 38.93500066800639, "lr": 0.001305, "grad_norm": 0.23322, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:41.402320+00:00", "epoch": 1, "step": 2099, "train_loss": 3.626962900161743, "perplexity": 37.598452783032016, "lr": 0.001305, "grad_norm": 0.248768, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:41.709456+00:00", "epoch": 1, "step": 2100, "train_loss": 3.7406327724456787, "perplexity": 42.12463704300773, "lr": 0.001305, "grad_norm": 0.232263, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:56:47.417831+00:00", "step": 2100, "epoch": 1, "val_loss": 3.703567099571228, "val_ppl": 40.591841558136174, "eval_train_loss": 3.7406327724456787, "eval_train_ppl": 42.12463704300773} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:56:48.378381+00:00", "step": 2100, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p7036_epoch_0001_step_0002100.pt", "val_loss": 3.703567099571228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:49.366999+00:00", "epoch": 1, "step": 2101, "train_loss": 3.6872196197509766, "perplexity": 39.933661696010326, "lr": 0.001305, "grad_norm": 0.2354, "tokens_per_sec": 4279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:49.673315+00:00", "epoch": 1, "step": 2102, "train_loss": 3.7410085201263428, "perplexity": 42.140468251759174, "lr": 0.001305, "grad_norm": 0.239999, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:49.979341+00:00", "epoch": 1, "step": 2103, "train_loss": 3.6867733001708984, "perplexity": 39.91584249771553, "lr": 0.001305, "grad_norm": 0.256048, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:50.285118+00:00", "epoch": 1, "step": 2104, "train_loss": 3.668320417404175, "perplexity": 39.18603437615301, "lr": 0.001305, "grad_norm": 0.269393, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:50.591644+00:00", "epoch": 1, "step": 2105, "train_loss": 3.7034518718719482, "perplexity": 40.58716452309106, "lr": 0.001305, "grad_norm": 0.294648, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:50.898039+00:00", "epoch": 1, "step": 2106, "train_loss": 3.767969846725464, "perplexity": 43.292085990663075, "lr": 0.001305, "grad_norm": 0.236172, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:51.203603+00:00", "epoch": 1, "step": 2107, "train_loss": 3.672908067703247, "perplexity": 39.36621919492502, "lr": 0.001305, "grad_norm": 0.248937, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:51.509132+00:00", "epoch": 1, "step": 2108, "train_loss": 3.6531405448913574, "perplexity": 38.59568740196527, "lr": 0.001305, "grad_norm": 0.30439, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:51.815299+00:00", "epoch": 1, "step": 2109, "train_loss": 3.6900386810302734, "perplexity": 40.04639596318255, "lr": 0.001305, "grad_norm": 0.315627, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:52.121707+00:00", "epoch": 1, "step": 2110, "train_loss": 3.7432761192321777, "perplexity": 42.23613436508988, "lr": 0.001305, "grad_norm": 0.299581, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:52.429894+00:00", "epoch": 1, "step": 2111, "train_loss": 3.7863922119140625, "perplexity": 44.09702023991479, "lr": 0.001305, "grad_norm": 0.301922, "tokens_per_sec": 106270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:52.736474+00:00", "epoch": 1, "step": 2112, "train_loss": 3.7438557147979736, "perplexity": 42.26062133686889, "lr": 0.001305, "grad_norm": 0.322338, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:53.043024+00:00", "epoch": 1, "step": 2113, "train_loss": 3.731492042541504, "perplexity": 41.74134158241054, "lr": 0.001305, "grad_norm": 0.267209, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:53.349697+00:00", "epoch": 1, "step": 2114, "train_loss": 3.6516714096069336, "perplexity": 38.5390267470513, "lr": 0.001305, "grad_norm": 0.241112, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:53.657073+00:00", "epoch": 1, "step": 2115, "train_loss": 3.7753500938415527, "perplexity": 43.61277420698321, "lr": 0.001305, "grad_norm": 0.278118, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:53.963710+00:00", "epoch": 1, "step": 2116, "train_loss": 3.729046583175659, "perplexity": 41.63938953829389, "lr": 0.001305, "grad_norm": 0.288627, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:54.270214+00:00", "epoch": 1, "step": 2117, "train_loss": 3.615152359008789, "perplexity": 37.1570066991109, "lr": 0.001305, "grad_norm": 0.31753, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:54.576414+00:00", "epoch": 1, "step": 2118, "train_loss": 3.7830066680908203, "perplexity": 43.94798027843549, "lr": 0.001305, "grad_norm": 0.283685, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:54.882854+00:00", "epoch": 1, "step": 2119, "train_loss": 3.7388205528259277, "perplexity": 42.048367079119416, "lr": 0.001305, "grad_norm": 0.292114, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:55.190252+00:00", "epoch": 1, "step": 2120, "train_loss": 3.656933546066284, "perplexity": 38.74235887632021, "lr": 0.001305, "grad_norm": 0.254563, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:55.497333+00:00", "epoch": 1, "step": 2121, "train_loss": 3.6645023822784424, "perplexity": 39.036705972413586, "lr": 0.001305, "grad_norm": 0.235712, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:55.803555+00:00", "epoch": 1, "step": 2122, "train_loss": 3.7400166988372803, "perplexity": 42.09869315835406, "lr": 0.001305, "grad_norm": 0.258382, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:56.109858+00:00", "epoch": 1, "step": 2123, "train_loss": 3.727614164352417, "perplexity": 41.57978719088422, "lr": 0.001305, "grad_norm": 0.260677, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:56.416722+00:00", "epoch": 1, "step": 2124, "train_loss": 3.7810564041137695, "perplexity": 43.86235364004876, "lr": 0.001305, "grad_norm": 0.231419, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:56.723543+00:00", "epoch": 1, "step": 2125, "train_loss": 3.7541801929473877, "perplexity": 42.699200352324574, "lr": 0.001305, "grad_norm": 0.229672, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:57.030543+00:00", "epoch": 1, "step": 2126, "train_loss": 3.818112850189209, "perplexity": 45.51822750260442, "lr": 0.001305, "grad_norm": 0.258994, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:57.336402+00:00", "epoch": 1, "step": 2127, "train_loss": 3.6832470893859863, "perplexity": 39.77533869201596, "lr": 0.001305, "grad_norm": 0.264896, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:57.642659+00:00", "epoch": 1, "step": 2128, "train_loss": 3.704061269760132, "perplexity": 40.61190579331203, "lr": 0.001305, "grad_norm": 0.247578, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:57.949958+00:00", "epoch": 1, "step": 2129, "train_loss": 3.7068498134613037, "perplexity": 40.72531191287783, "lr": 0.001305, "grad_norm": 0.229657, "tokens_per_sec": 106632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:58.257883+00:00", "epoch": 1, "step": 2130, "train_loss": 3.796860933303833, "perplexity": 44.561084500431946, "lr": 0.001305, "grad_norm": 0.232855, "tokens_per_sec": 106472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:58.566218+00:00", "epoch": 1, "step": 2131, "train_loss": 3.7666189670562744, "perplexity": 43.23364307542208, "lr": 0.001305, "grad_norm": 0.227602, "tokens_per_sec": 106217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:58.872774+00:00", "epoch": 1, "step": 2132, "train_loss": 3.755232334136963, "perplexity": 42.74414958209865, "lr": 0.001305, "grad_norm": 0.257051, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:59.179268+00:00", "epoch": 1, "step": 2133, "train_loss": 3.6857895851135254, "perplexity": 39.876595989278336, "lr": 0.001305, "grad_norm": 0.246097, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:59.485852+00:00", "epoch": 1, "step": 2134, "train_loss": 3.634282112121582, "perplexity": 37.87465338064743, "lr": 0.001305, "grad_norm": 0.202321, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:56:59.793872+00:00", "epoch": 1, "step": 2135, "train_loss": 3.7229785919189453, "perplexity": 41.38748713037597, "lr": 0.001305, "grad_norm": 0.22944, "tokens_per_sec": 106382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:00.100626+00:00", "epoch": 1, "step": 2136, "train_loss": 3.7920119762420654, "perplexity": 44.34553273802829, "lr": 0.001305, "grad_norm": 0.217367, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:00.407557+00:00", "epoch": 1, "step": 2137, "train_loss": 3.7317962646484375, "perplexity": 41.75404215309221, "lr": 0.001305, "grad_norm": 0.267098, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:00.714378+00:00", "epoch": 1, "step": 2138, "train_loss": 3.6749820709228516, "perplexity": 39.44794958552871, "lr": 0.001305, "grad_norm": 0.295413, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:01.021375+00:00", "epoch": 1, "step": 2139, "train_loss": 3.6670985221862793, "perplexity": 39.13818238915124, "lr": 0.001305, "grad_norm": 0.295347, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:01.329088+00:00", "epoch": 1, "step": 2140, "train_loss": 3.6793625354766846, "perplexity": 39.621128956537085, "lr": 0.001305, "grad_norm": 0.293561, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:01.635943+00:00", "epoch": 1, "step": 2141, "train_loss": 3.6897268295288086, "perplexity": 40.03390938155398, "lr": 0.001305, "grad_norm": 0.262402, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:01.942051+00:00", "epoch": 1, "step": 2142, "train_loss": 3.6785037517547607, "perplexity": 39.58711758224258, "lr": 0.001305, "grad_norm": 0.256243, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:02.248649+00:00", "epoch": 1, "step": 2143, "train_loss": 3.715440034866333, "perplexity": 41.07665826995651, "lr": 0.001305, "grad_norm": 0.260116, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:02.555425+00:00", "epoch": 1, "step": 2144, "train_loss": 3.650334119796753, "perplexity": 38.48752334445142, "lr": 0.001305, "grad_norm": 0.267041, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:02.862626+00:00", "epoch": 1, "step": 2145, "train_loss": 3.794102668762207, "perplexity": 44.43834259625763, "lr": 0.001305, "grad_norm": 0.251972, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:03.169889+00:00", "epoch": 1, "step": 2146, "train_loss": 3.619384288787842, "perplexity": 37.31458573880011, "lr": 0.001305, "grad_norm": 0.260922, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:03.476178+00:00", "epoch": 1, "step": 2147, "train_loss": 3.701371192932129, "perplexity": 40.50280345922018, "lr": 0.001305, "grad_norm": 0.290004, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:03.783129+00:00", "epoch": 1, "step": 2148, "train_loss": 3.6604950428009033, "perplexity": 38.880585661963856, "lr": 0.001305, "grad_norm": 0.262269, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:04.090142+00:00", "epoch": 1, "step": 2149, "train_loss": 3.6989476680755615, "perplexity": 40.40476275830053, "lr": 0.001305, "grad_norm": 0.239856, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:04.397064+00:00", "epoch": 1, "step": 2150, "train_loss": 3.6947214603424072, "perplexity": 40.23436416103153, "lr": 0.001305, "grad_norm": 0.247946, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:04.705268+00:00", "epoch": 1, "step": 2151, "train_loss": 3.7441999912261963, "perplexity": 42.27517317742133, "lr": 0.001305, "grad_norm": 0.257593, "tokens_per_sec": 106257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:05.013331+00:00", "epoch": 1, "step": 2152, "train_loss": 3.7109415531158447, "perplexity": 40.89229067040936, "lr": 0.001305, "grad_norm": 0.274454, "tokens_per_sec": 106369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:05.321015+00:00", "epoch": 1, "step": 2153, "train_loss": 3.6881802082061768, "perplexity": 39.972039940307525, "lr": 0.001305, "grad_norm": 0.259587, "tokens_per_sec": 106500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:05.628151+00:00", "epoch": 1, "step": 2154, "train_loss": 3.765275239944458, "perplexity": 43.17558787099052, "lr": 0.001305, "grad_norm": 0.24185, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:05.934922+00:00", "epoch": 1, "step": 2155, "train_loss": 3.742900848388672, "perplexity": 42.2202873489678, "lr": 0.001305, "grad_norm": 0.257976, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:06.242245+00:00", "epoch": 1, "step": 2156, "train_loss": 3.746983051300049, "perplexity": 42.39299139555005, "lr": 0.001305, "grad_norm": 0.259838, "tokens_per_sec": 106624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:06.549538+00:00", "epoch": 1, "step": 2157, "train_loss": 3.5708861351013184, "perplexity": 35.54807960006261, "lr": 0.001305, "grad_norm": 0.251337, "tokens_per_sec": 106633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:06.858995+00:00", "epoch": 1, "step": 2158, "train_loss": 3.629045248031616, "perplexity": 37.67682741439813, "lr": 0.001305, "grad_norm": 0.278293, "tokens_per_sec": 105889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:07.166797+00:00", "epoch": 1, "step": 2159, "train_loss": 3.6125388145446777, "perplexity": 37.060022002063945, "lr": 0.001305, "grad_norm": 0.258031, "tokens_per_sec": 106458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:07.474325+00:00", "epoch": 1, "step": 2160, "train_loss": 3.7705235481262207, "perplexity": 43.402782333844605, "lr": 0.001305, "grad_norm": 0.270002, "tokens_per_sec": 106611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:07.795279+00:00", "epoch": 1, "step": 2161, "train_loss": 3.7265923023223877, "perplexity": 41.53732008659321, "lr": 0.001305, "grad_norm": 0.259385, "tokens_per_sec": 102050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:08.102900+00:00", "epoch": 1, "step": 2162, "train_loss": 3.8070619106292725, "perplexity": 45.01797752871384, "lr": 0.001305, "grad_norm": 0.265844, "tokens_per_sec": 106512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:08.411419+00:00", "epoch": 1, "step": 2163, "train_loss": 3.627896547317505, "perplexity": 37.63357286386455, "lr": 0.001305, "grad_norm": 0.256837, "tokens_per_sec": 106211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:08.718713+00:00", "epoch": 1, "step": 2164, "train_loss": 3.749049425125122, "perplexity": 42.48068173264259, "lr": 0.001305, "grad_norm": 0.231157, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:09.035712+00:00", "epoch": 1, "step": 2165, "train_loss": 3.5867178440093994, "perplexity": 36.11534499042323, "lr": 0.001305, "grad_norm": 0.247679, "tokens_per_sec": 103369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:09.342303+00:00", "epoch": 1, "step": 2166, "train_loss": 3.6955068111419678, "perplexity": 40.26597466213665, "lr": 0.001305, "grad_norm": 0.24468, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:09.662206+00:00", "epoch": 1, "step": 2167, "train_loss": 3.7248411178588867, "perplexity": 41.46464422999018, "lr": 0.001305, "grad_norm": 0.248529, "tokens_per_sec": 102431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:09.968709+00:00", "epoch": 1, "step": 2168, "train_loss": 3.706555128097534, "perplexity": 40.71331252763059, "lr": 0.001305, "grad_norm": 0.24133, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:10.276482+00:00", "epoch": 1, "step": 2169, "train_loss": 3.531566619873047, "perplexity": 34.17746879763269, "lr": 0.001305, "grad_norm": 0.226785, "tokens_per_sec": 106468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:10.596776+00:00", "epoch": 1, "step": 2170, "train_loss": 3.7438390254974365, "perplexity": 42.25991604254396, "lr": 0.001305, "grad_norm": 0.231916, "tokens_per_sec": 102365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:10.904949+00:00", "epoch": 1, "step": 2171, "train_loss": 3.676872968673706, "perplexity": 39.522612192106926, "lr": 0.001305, "grad_norm": 0.241442, "tokens_per_sec": 106266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:11.212430+00:00", "epoch": 1, "step": 2172, "train_loss": 3.6730499267578125, "perplexity": 39.371804045683284, "lr": 0.001305, "grad_norm": 0.237279, "tokens_per_sec": 106570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:11.532914+00:00", "epoch": 1, "step": 2173, "train_loss": 3.657418966293335, "perplexity": 38.761169766186285, "lr": 0.001305, "grad_norm": 0.238538, "tokens_per_sec": 102245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:11.839409+00:00", "epoch": 1, "step": 2174, "train_loss": 3.676213264465332, "perplexity": 39.49654755693791, "lr": 0.001305, "grad_norm": 0.234054, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:12.146793+00:00", "epoch": 1, "step": 2175, "train_loss": 3.7495474815368652, "perplexity": 42.50184477831257, "lr": 0.001305, "grad_norm": 0.225448, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:12.454889+00:00", "epoch": 1, "step": 2176, "train_loss": 3.7034530639648438, "perplexity": 40.58721290679038, "lr": 0.001305, "grad_norm": 0.222677, "tokens_per_sec": 106356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:12.762925+00:00", "epoch": 1, "step": 2177, "train_loss": 3.8253414630889893, "perplexity": 45.84845324821219, "lr": 0.001305, "grad_norm": 0.222782, "tokens_per_sec": 106377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:13.080501+00:00", "epoch": 1, "step": 2178, "train_loss": 3.6708009243011475, "perplexity": 39.28335625858727, "lr": 0.001305, "grad_norm": 0.252706, "tokens_per_sec": 103182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:13.387449+00:00", "epoch": 1, "step": 2179, "train_loss": 3.7792911529541016, "perplexity": 43.784993869220955, "lr": 0.001305, "grad_norm": 0.244471, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:13.708332+00:00", "epoch": 1, "step": 2180, "train_loss": 3.7602131366729736, "perplexity": 42.95758083896916, "lr": 0.001305, "grad_norm": 0.24881, "tokens_per_sec": 102175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:14.015243+00:00", "epoch": 1, "step": 2181, "train_loss": 3.606184959411621, "perplexity": 36.82529449301983, "lr": 0.001305, "grad_norm": 0.253239, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:14.322179+00:00", "epoch": 1, "step": 2182, "train_loss": 3.621232748031616, "perplexity": 37.38362401730114, "lr": 0.001305, "grad_norm": 0.266382, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:14.629665+00:00", "epoch": 1, "step": 2183, "train_loss": 3.679746150970459, "perplexity": 39.63633115119792, "lr": 0.001305, "grad_norm": 0.296539, "tokens_per_sec": 106568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:14.946448+00:00", "epoch": 1, "step": 2184, "train_loss": 3.7040112018585205, "perplexity": 40.60987249131053, "lr": 0.001305, "grad_norm": 0.307925, "tokens_per_sec": 103439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:15.253687+00:00", "epoch": 1, "step": 2185, "train_loss": 3.7176735401153564, "perplexity": 41.168505734529816, "lr": 0.001305, "grad_norm": 0.306765, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:15.574902+00:00", "epoch": 1, "step": 2186, "train_loss": 3.8015336990356445, "perplexity": 44.7697952575328, "lr": 0.001305, "grad_norm": 0.283726, "tokens_per_sec": 102015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:15.882092+00:00", "epoch": 1, "step": 2187, "train_loss": 3.6672005653381348, "perplexity": 39.14217637641618, "lr": 0.001305, "grad_norm": 0.280818, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:16.189517+00:00", "epoch": 1, "step": 2188, "train_loss": 3.681548833847046, "perplexity": 39.70784732778141, "lr": 0.001305, "grad_norm": 0.248388, "tokens_per_sec": 106589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:16.512956+00:00", "epoch": 1, "step": 2189, "train_loss": 3.713862180709839, "perplexity": 41.01189639979943, "lr": 0.001305, "grad_norm": 0.250151, "tokens_per_sec": 101311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:16.819631+00:00", "epoch": 1, "step": 2190, "train_loss": 3.793966293334961, "perplexity": 44.432282711519036, "lr": 0.001305, "grad_norm": 0.22879, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:17.127386+00:00", "epoch": 1, "step": 2191, "train_loss": 3.774407386779785, "perplexity": 43.57167950992896, "lr": 0.001305, "grad_norm": 0.261817, "tokens_per_sec": 106422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:17.433568+00:00", "epoch": 1, "step": 2192, "train_loss": 3.659797430038452, "perplexity": 38.85347152787711, "lr": 0.001305, "grad_norm": 0.257518, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:17.740267+00:00", "epoch": 1, "step": 2193, "train_loss": 3.689342737197876, "perplexity": 40.01853561664488, "lr": 0.001305, "grad_norm": 0.246057, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:18.059166+00:00", "epoch": 1, "step": 2194, "train_loss": 3.7743382453918457, "perplexity": 43.568667007678286, "lr": 0.001305, "grad_norm": 0.241579, "tokens_per_sec": 102753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:18.366982+00:00", "epoch": 1, "step": 2195, "train_loss": 3.6898305416107178, "perplexity": 40.03806159695664, "lr": 0.001305, "grad_norm": 0.259289, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:18.675313+00:00", "epoch": 1, "step": 2196, "train_loss": 3.6586310863494873, "perplexity": 38.8081814436024, "lr": 0.001305, "grad_norm": 0.238634, "tokens_per_sec": 106280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:18.983630+00:00", "epoch": 1, "step": 2197, "train_loss": 3.675957441329956, "perplexity": 39.48644471863035, "lr": 0.001305, "grad_norm": 0.231954, "tokens_per_sec": 106277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:19.290915+00:00", "epoch": 1, "step": 2198, "train_loss": 3.7210283279418945, "perplexity": 41.306849263247706, "lr": 0.001305, "grad_norm": 0.245504, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:19.598427+00:00", "epoch": 1, "step": 2199, "train_loss": 3.6882410049438477, "perplexity": 39.974470183808634, "lr": 0.001305, "grad_norm": 0.225468, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:19.919792+00:00", "epoch": 1, "step": 2200, "train_loss": 3.6871228218078613, "perplexity": 39.929796386777106, "lr": 0.001305, "grad_norm": 0.230738, "tokens_per_sec": 102016} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:57:25.601208+00:00", "step": 2200, "epoch": 1, "val_loss": 3.692614722251892, "val_ppl": 40.149690117839576, "eval_train_loss": 3.6871228218078613, "eval_train_ppl": 39.929796386777106} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:57:26.552517+00:00", "step": 2200, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p6926_epoch_0001_step_0002200.pt", "val_loss": 3.692614722251892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:27.538529+00:00", "epoch": 1, "step": 2201, "train_loss": 3.6211934089660645, "perplexity": 37.38215340939172, "lr": 0.001305, "grad_norm": 0.235017, "tokens_per_sec": 4301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:27.844749+00:00", "epoch": 1, "step": 2202, "train_loss": 3.7522575855255127, "perplexity": 42.61718541934196, "lr": 0.001305, "grad_norm": 0.224356, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:28.150606+00:00", "epoch": 1, "step": 2203, "train_loss": 3.737445831298828, "perplexity": 41.99060199826377, "lr": 0.001305, "grad_norm": 0.23863, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:28.456163+00:00", "epoch": 1, "step": 2204, "train_loss": 3.766681671142578, "perplexity": 43.23635408650355, "lr": 0.001305, "grad_norm": 0.230657, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:28.762708+00:00", "epoch": 1, "step": 2205, "train_loss": 3.7045464515686035, "perplexity": 40.63161473203296, "lr": 0.001305, "grad_norm": 0.216082, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:29.068822+00:00", "epoch": 1, "step": 2206, "train_loss": 3.686812162399292, "perplexity": 39.9173937464455, "lr": 0.001305, "grad_norm": 0.249569, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:29.374905+00:00", "epoch": 1, "step": 2207, "train_loss": 3.8128933906555176, "perplexity": 45.28126589981466, "lr": 0.001305, "grad_norm": 0.265884, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:29.681595+00:00", "epoch": 1, "step": 2208, "train_loss": 3.6271512508392334, "perplexity": 37.60553514404909, "lr": 0.001305, "grad_norm": 0.227154, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:29.987919+00:00", "epoch": 1, "step": 2209, "train_loss": 3.704709529876709, "perplexity": 40.63824140733786, "lr": 0.001305, "grad_norm": 0.236088, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:30.293876+00:00", "epoch": 1, "step": 2210, "train_loss": 3.683074712753296, "perplexity": 39.76848294397046, "lr": 0.001305, "grad_norm": 0.258338, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:30.600951+00:00", "epoch": 1, "step": 2211, "train_loss": 3.7798283100128174, "perplexity": 43.80851960568567, "lr": 0.001305, "grad_norm": 0.253131, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:30.907245+00:00", "epoch": 1, "step": 2212, "train_loss": 3.7617321014404297, "perplexity": 43.02288147288437, "lr": 0.001305, "grad_norm": 0.223869, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:31.213968+00:00", "epoch": 1, "step": 2213, "train_loss": 3.789384365081787, "perplexity": 44.2291628756267, "lr": 0.001305, "grad_norm": 0.228666, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:31.519839+00:00", "epoch": 1, "step": 2214, "train_loss": 3.6535706520080566, "perplexity": 38.61229125225187, "lr": 0.001305, "grad_norm": 0.210568, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:31.825836+00:00", "epoch": 1, "step": 2215, "train_loss": 3.6421332359313965, "perplexity": 38.17318233403387, "lr": 0.001305, "grad_norm": 0.230517, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:32.132376+00:00", "epoch": 1, "step": 2216, "train_loss": 3.778397560119629, "perplexity": 43.7458853885734, "lr": 0.001305, "grad_norm": 0.229708, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:32.438921+00:00", "epoch": 1, "step": 2217, "train_loss": 3.777224063873291, "perplexity": 43.69457986558421, "lr": 0.001305, "grad_norm": 0.272474, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:32.745289+00:00", "epoch": 1, "step": 2218, "train_loss": 3.690636157989502, "perplexity": 40.07032991134999, "lr": 0.001305, "grad_norm": 0.292495, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:33.051020+00:00", "epoch": 1, "step": 2219, "train_loss": 3.7078239917755127, "perplexity": 40.76500495949623, "lr": 0.001305, "grad_norm": 0.310652, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:33.356950+00:00", "epoch": 1, "step": 2220, "train_loss": 3.6791763305664062, "perplexity": 39.61375199460919, "lr": 0.001305, "grad_norm": 0.332844, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:33.663269+00:00", "epoch": 1, "step": 2221, "train_loss": 3.7513833045959473, "perplexity": 42.57994230970062, "lr": 0.001305, "grad_norm": 0.297345, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:33.970217+00:00", "epoch": 1, "step": 2222, "train_loss": 3.8396947383880615, "perplexity": 46.511274165975024, "lr": 0.001305, "grad_norm": 0.317333, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:34.277377+00:00", "epoch": 1, "step": 2223, "train_loss": 3.7428579330444336, "perplexity": 42.21847548968094, "lr": 0.001305, "grad_norm": 0.301288, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:34.583750+00:00", "epoch": 1, "step": 2224, "train_loss": 3.668818473815918, "perplexity": 39.205556092879384, "lr": 0.001305, "grad_norm": 0.256598, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:34.890844+00:00", "epoch": 1, "step": 2225, "train_loss": 3.619105339050293, "perplexity": 37.30417829654566, "lr": 0.001305, "grad_norm": 0.2736, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:35.197233+00:00", "epoch": 1, "step": 2226, "train_loss": 3.713315963745117, "perplexity": 40.98950112312728, "lr": 0.001305, "grad_norm": 0.285892, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:35.504094+00:00", "epoch": 1, "step": 2227, "train_loss": 3.644437074661255, "perplexity": 38.26122857316144, "lr": 0.001305, "grad_norm": 0.278817, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:35.810432+00:00", "epoch": 1, "step": 2228, "train_loss": 3.6275057792663574, "perplexity": 37.61886973888167, "lr": 0.001305, "grad_norm": 0.243851, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:36.118049+00:00", "epoch": 1, "step": 2229, "train_loss": 3.7353434562683105, "perplexity": 41.90241473893439, "lr": 0.001305, "grad_norm": 0.25986, "tokens_per_sec": 106521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:36.424679+00:00", "epoch": 1, "step": 2230, "train_loss": 3.6882410049438477, "perplexity": 39.974470183808634, "lr": 0.001305, "grad_norm": 0.280775, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:36.731136+00:00", "epoch": 1, "step": 2231, "train_loss": 3.6215648651123047, "perplexity": 37.39604181934325, "lr": 0.001305, "grad_norm": 0.260969, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:37.036805+00:00", "epoch": 1, "step": 2232, "train_loss": 3.5950419902801514, "perplexity": 36.41722912542768, "lr": 0.001305, "grad_norm": 0.234379, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:37.343473+00:00", "epoch": 1, "step": 2233, "train_loss": 3.701256275177002, "perplexity": 40.498149235401904, "lr": 0.001305, "grad_norm": 0.268882, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:37.650219+00:00", "epoch": 1, "step": 2234, "train_loss": 3.703756809234619, "perplexity": 40.59954295322603, "lr": 0.001305, "grad_norm": 0.281581, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:37.957541+00:00", "epoch": 1, "step": 2235, "train_loss": 3.6488780975341797, "perplexity": 38.43152543062815, "lr": 0.001305, "grad_norm": 0.261832, "tokens_per_sec": 106626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:38.263837+00:00", "epoch": 1, "step": 2236, "train_loss": 3.68424391746521, "perplexity": 39.81500763475798, "lr": 0.001305, "grad_norm": 0.302164, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:38.569908+00:00", "epoch": 1, "step": 2237, "train_loss": 3.7712607383728027, "perplexity": 43.434790238163515, "lr": 0.001305, "grad_norm": 0.265588, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:38.875353+00:00", "epoch": 1, "step": 2238, "train_loss": 3.6718249320983887, "perplexity": 39.3236033248358, "lr": 0.001305, "grad_norm": 0.234961, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:39.182075+00:00", "epoch": 1, "step": 2239, "train_loss": 3.6665916442871094, "perplexity": 39.118349136427874, "lr": 0.001305, "grad_norm": 0.266387, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:39.489688+00:00", "epoch": 1, "step": 2240, "train_loss": 3.6036217212677, "perplexity": 36.73102336480498, "lr": 0.001305, "grad_norm": 0.271544, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:39.797850+00:00", "epoch": 1, "step": 2241, "train_loss": 3.5635902881622314, "perplexity": 35.28967005680702, "lr": 0.001305, "grad_norm": 0.250345, "tokens_per_sec": 106266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:40.105024+00:00", "epoch": 1, "step": 2242, "train_loss": 3.71209454536438, "perplexity": 40.93946635595081, "lr": 0.001305, "grad_norm": 0.259671, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:40.453512+00:00", "epoch": 1, "step": 2243, "train_loss": 3.708867073059082, "perplexity": 40.807548357447835, "lr": 0.001305, "grad_norm": 0.248815, "tokens_per_sec": 94029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:40.759967+00:00", "epoch": 1, "step": 2244, "train_loss": 3.651623249053955, "perplexity": 38.53717073090565, "lr": 0.001305, "grad_norm": 0.267619, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:41.066850+00:00", "epoch": 1, "step": 2245, "train_loss": 3.7504210472106934, "perplexity": 42.53898915264835, "lr": 0.001305, "grad_norm": 0.257589, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:41.374133+00:00", "epoch": 1, "step": 2246, "train_loss": 3.625704288482666, "perplexity": 37.55116069867187, "lr": 0.001305, "grad_norm": 0.238686, "tokens_per_sec": 106638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:41.683130+00:00", "epoch": 1, "step": 2247, "train_loss": 3.639695644378662, "perplexity": 38.080245024853646, "lr": 0.001305, "grad_norm": 0.247629, "tokens_per_sec": 106046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:41.990214+00:00", "epoch": 1, "step": 2248, "train_loss": 3.7137794494628906, "perplexity": 41.00850357481882, "lr": 0.001305, "grad_norm": 0.224336, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:42.297147+00:00", "epoch": 1, "step": 2249, "train_loss": 3.747389316558838, "perplexity": 42.4102176941566, "lr": 0.001305, "grad_norm": 0.246351, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:42.603728+00:00", "epoch": 1, "step": 2250, "train_loss": 3.8189144134521484, "perplexity": 45.55472786828643, "lr": 0.001305, "grad_norm": 0.254051, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:57:43.663139+00:00", "step": 2250, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0001_step_0002250.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:44.638919+00:00", "epoch": 1, "step": 2251, "train_loss": 3.7482166290283203, "perplexity": 42.44531871384201, "lr": 0.001305, "grad_norm": 0.26492, "tokens_per_sec": 16099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:44.943875+00:00", "epoch": 1, "step": 2252, "train_loss": 3.669386386871338, "perplexity": 39.22782776361722, "lr": 0.001305, "grad_norm": 0.241496, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:45.250419+00:00", "epoch": 1, "step": 2253, "train_loss": 3.640646457672119, "perplexity": 38.11646944664323, "lr": 0.001305, "grad_norm": 0.243727, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:45.556188+00:00", "epoch": 1, "step": 2254, "train_loss": 3.6768500804901123, "perplexity": 39.52170760165522, "lr": 0.001305, "grad_norm": 0.251068, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:45.862893+00:00", "epoch": 1, "step": 2255, "train_loss": 3.6790003776550293, "perplexity": 39.60678245278874, "lr": 0.001305, "grad_norm": 0.282828, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:46.168479+00:00", "epoch": 1, "step": 2256, "train_loss": 3.6830036640167236, "perplexity": 39.76565754387364, "lr": 0.001305, "grad_norm": 0.262331, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:46.474781+00:00", "epoch": 1, "step": 2257, "train_loss": 3.7159430980682373, "perplexity": 41.097327623749074, "lr": 0.001305, "grad_norm": 0.256224, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:46.780218+00:00", "epoch": 1, "step": 2258, "train_loss": 3.700867176055908, "perplexity": 40.482394506402855, "lr": 0.001305, "grad_norm": 0.241591, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:47.086816+00:00", "epoch": 1, "step": 2259, "train_loss": 3.664071559906006, "perplexity": 39.01989170837508, "lr": 0.001305, "grad_norm": 0.237664, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:47.393471+00:00", "epoch": 1, "step": 2260, "train_loss": 3.7712597846984863, "perplexity": 43.43474881553938, "lr": 0.001305, "grad_norm": 0.228551, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:47.699823+00:00", "epoch": 1, "step": 2261, "train_loss": 3.7889769077301025, "perplexity": 44.2111450490514, "lr": 0.001305, "grad_norm": 0.228751, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:48.005730+00:00", "epoch": 1, "step": 2262, "train_loss": 3.688370704650879, "perplexity": 39.97965519712029, "lr": 0.001305, "grad_norm": 0.227025, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:48.311412+00:00", "epoch": 1, "step": 2263, "train_loss": 3.6776182651519775, "perplexity": 39.55207923526354, "lr": 0.001305, "grad_norm": 0.226742, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:48.618589+00:00", "epoch": 1, "step": 2264, "train_loss": 3.6663882732391357, "perplexity": 39.110394405677376, "lr": 0.001305, "grad_norm": 0.218962, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:48.934636+00:00", "epoch": 1, "step": 2265, "train_loss": 3.8621609210968018, "perplexity": 47.568031149489734, "lr": 0.001305, "grad_norm": 0.229424, "tokens_per_sec": 103684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:49.241414+00:00", "epoch": 1, "step": 2266, "train_loss": 3.804960250854492, "perplexity": 44.92346440820007, "lr": 0.001305, "grad_norm": 0.2365, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:49.548039+00:00", "epoch": 1, "step": 2267, "train_loss": 3.67162823677063, "perplexity": 39.315869316437755, "lr": 0.001305, "grad_norm": 0.254977, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:49.854438+00:00", "epoch": 1, "step": 2268, "train_loss": 3.654276132583618, "perplexity": 38.63954108469181, "lr": 0.001305, "grad_norm": 0.252863, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:50.160498+00:00", "epoch": 1, "step": 2269, "train_loss": 3.6178672313690186, "perplexity": 37.25802028704248, "lr": 0.001305, "grad_norm": 0.256334, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:50.466631+00:00", "epoch": 1, "step": 2270, "train_loss": 3.709394931793213, "perplexity": 40.82909466446966, "lr": 0.001305, "grad_norm": 0.243968, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:50.773260+00:00", "epoch": 1, "step": 2271, "train_loss": 3.7020103931427, "perplexity": 40.52870113573842, "lr": 0.001305, "grad_norm": 0.214191, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:51.080091+00:00", "epoch": 1, "step": 2272, "train_loss": 3.702167510986328, "perplexity": 40.53506941813821, "lr": 0.001305, "grad_norm": 0.240895, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:51.387184+00:00", "epoch": 1, "step": 2273, "train_loss": 3.6618692874908447, "perplexity": 38.93405383111062, "lr": 0.001305, "grad_norm": 0.213289, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:51.693083+00:00", "epoch": 1, "step": 2274, "train_loss": 3.637920379638672, "perplexity": 38.01270247925938, "lr": 0.001305, "grad_norm": 0.223045, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:51.999565+00:00", "epoch": 1, "step": 2275, "train_loss": 3.605451822280884, "perplexity": 36.798306396473315, "lr": 0.001305, "grad_norm": 0.227571, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:52.307781+00:00", "epoch": 1, "step": 2276, "train_loss": 3.7004554271698, "perplexity": 40.4657293567217, "lr": 0.001305, "grad_norm": 0.257765, "tokens_per_sec": 106315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:52.615519+00:00", "epoch": 1, "step": 2277, "train_loss": 3.7011172771453857, "perplexity": 40.49252046357724, "lr": 0.001305, "grad_norm": 0.258696, "tokens_per_sec": 106482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:52.922717+00:00", "epoch": 1, "step": 2278, "train_loss": 3.6168954372406006, "perplexity": 37.22183074892944, "lr": 0.001305, "grad_norm": 0.272991, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:53.229044+00:00", "epoch": 1, "step": 2279, "train_loss": 3.6574392318725586, "perplexity": 38.76195529170251, "lr": 0.001305, "grad_norm": 0.298815, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:53.535639+00:00", "epoch": 1, "step": 2280, "train_loss": 3.8277881145477295, "perplexity": 45.96076577200681, "lr": 0.001305, "grad_norm": 0.311084, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:53.842196+00:00", "epoch": 1, "step": 2281, "train_loss": 3.715960741043091, "perplexity": 41.0980527092632, "lr": 0.001305, "grad_norm": 0.328726, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:54.148576+00:00", "epoch": 1, "step": 2282, "train_loss": 3.718611240386963, "perplexity": 41.2071275585562, "lr": 0.001305, "grad_norm": 0.309011, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:54.455526+00:00", "epoch": 1, "step": 2283, "train_loss": 3.823917865753174, "perplexity": 45.783229949189426, "lr": 0.001305, "grad_norm": 0.307596, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:54.762560+00:00", "epoch": 1, "step": 2284, "train_loss": 3.6870594024658203, "perplexity": 39.927264145659805, "lr": 0.001305, "grad_norm": 0.332825, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:55.069043+00:00", "epoch": 1, "step": 2285, "train_loss": 3.602175235748291, "perplexity": 36.677930879407576, "lr": 0.001305, "grad_norm": 0.293602, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:55.374686+00:00", "epoch": 1, "step": 2286, "train_loss": 3.6635055541992188, "perplexity": 38.997812476065114, "lr": 0.001305, "grad_norm": 0.288388, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:55.680898+00:00", "epoch": 1, "step": 2287, "train_loss": 3.772679328918457, "perplexity": 43.49645014568842, "lr": 0.001305, "grad_norm": 0.2683, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:55.987928+00:00", "epoch": 1, "step": 2288, "train_loss": 3.671884298324585, "perplexity": 39.32593788806205, "lr": 0.001305, "grad_norm": 0.246246, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:56.295009+00:00", "epoch": 1, "step": 2289, "train_loss": 3.621546506881714, "perplexity": 37.39535530048601, "lr": 0.001305, "grad_norm": 0.252059, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:56.600903+00:00", "epoch": 1, "step": 2290, "train_loss": 3.762208938598633, "perplexity": 43.04340127333648, "lr": 0.001305, "grad_norm": 0.241787, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:56.906935+00:00", "epoch": 1, "step": 2291, "train_loss": 3.6104466915130615, "perplexity": 36.98256892541486, "lr": 0.001305, "grad_norm": 0.250296, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:57.213231+00:00", "epoch": 1, "step": 2292, "train_loss": 3.683588981628418, "perplexity": 39.788939896695545, "lr": 0.001305, "grad_norm": 0.232447, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:57.519892+00:00", "epoch": 1, "step": 2293, "train_loss": 3.713632106781006, "perplexity": 41.002461717044724, "lr": 0.001305, "grad_norm": 0.245973, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:57.826350+00:00", "epoch": 1, "step": 2294, "train_loss": 3.639047622680664, "perplexity": 38.055576193644676, "lr": 0.001305, "grad_norm": 0.220843, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:58.133645+00:00", "epoch": 1, "step": 2295, "train_loss": 3.675489902496338, "perplexity": 39.4679875873719, "lr": 0.001305, "grad_norm": 0.272984, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:58.439941+00:00", "epoch": 1, "step": 2296, "train_loss": 3.639279842376709, "perplexity": 38.06441447415286, "lr": 0.001305, "grad_norm": 0.281669, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:58.746022+00:00", "epoch": 1, "step": 2297, "train_loss": 3.639155387878418, "perplexity": 38.059677481322886, "lr": 0.001305, "grad_norm": 0.296993, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:59.052992+00:00", "epoch": 1, "step": 2298, "train_loss": 3.698190450668335, "perplexity": 40.374179149286974, "lr": 0.001305, "grad_norm": 0.298243, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:59.359526+00:00", "epoch": 1, "step": 2299, "train_loss": 3.6540534496307373, "perplexity": 38.630937675536934, "lr": 0.001305, "grad_norm": 0.285848, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:57:59.667771+00:00", "epoch": 1, "step": 2300, "train_loss": 3.7777178287506104, "perplexity": 43.71616004177914, "lr": 0.001305, "grad_norm": 0.264453, "tokens_per_sec": 106372} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:58:05.372527+00:00", "step": 2300, "epoch": 1, "val_loss": 3.683774995803833, "val_ppl": 39.79634189195774, "eval_train_loss": 3.7777178287506104, "eval_train_ppl": 43.71616004177914} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:58:06.284316+00:00", "step": 2300, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p6838_epoch_0001_step_0002300.pt", "val_loss": 3.683774995803833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:07.499345+00:00", "epoch": 1, "step": 2301, "train_loss": 3.750389575958252, "perplexity": 42.53765041844805, "lr": 0.001305, "grad_norm": 0.246677, "tokens_per_sec": 4184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:07.804590+00:00", "epoch": 1, "step": 2302, "train_loss": 3.680936574935913, "perplexity": 39.68354328535557, "lr": 0.001305, "grad_norm": 0.25016, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:08.110012+00:00", "epoch": 1, "step": 2303, "train_loss": 3.7519924640655518, "perplexity": 42.60588818655976, "lr": 0.001305, "grad_norm": 0.26873, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:08.415338+00:00", "epoch": 1, "step": 2304, "train_loss": 3.6343209743499756, "perplexity": 37.87612530267834, "lr": 0.001305, "grad_norm": 0.24768, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:08.721513+00:00", "epoch": 1, "step": 2305, "train_loss": 3.6968657970428467, "perplexity": 40.3207327532971, "lr": 0.001305, "grad_norm": 0.227394, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:09.027849+00:00", "epoch": 1, "step": 2306, "train_loss": 3.66692852973938, "perplexity": 39.13152975922415, "lr": 0.001305, "grad_norm": 0.244991, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:09.333166+00:00", "epoch": 1, "step": 2307, "train_loss": 3.640181541442871, "perplexity": 38.0987526001397, "lr": 0.001305, "grad_norm": 0.225095, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:09.637680+00:00", "epoch": 1, "step": 2308, "train_loss": 3.7048258781433105, "perplexity": 40.6429698713523, "lr": 0.001305, "grad_norm": 0.214316, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:09.942673+00:00", "epoch": 1, "step": 2309, "train_loss": 3.6708173751831055, "perplexity": 39.28400250975968, "lr": 0.001305, "grad_norm": 0.226824, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:10.247960+00:00", "epoch": 1, "step": 2310, "train_loss": 3.6129472255706787, "perplexity": 37.07516081489204, "lr": 0.001305, "grad_norm": 0.242172, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:10.553680+00:00", "epoch": 1, "step": 2311, "train_loss": 3.617914915084839, "perplexity": 37.25979693025201, "lr": 0.001305, "grad_norm": 0.261418, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:10.859876+00:00", "epoch": 1, "step": 2312, "train_loss": 3.7465837001800537, "perplexity": 42.37606508695139, "lr": 0.001305, "grad_norm": 0.247555, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:11.165215+00:00", "epoch": 1, "step": 2313, "train_loss": 3.806021213531494, "perplexity": 44.971151820067625, "lr": 0.001305, "grad_norm": 0.234737, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:12.276133+00:00", "epoch": 1, "step": 2314, "train_loss": 3.697842836380005, "perplexity": 40.36014694677343, "lr": 0.001305, "grad_norm": 0.263157, "tokens_per_sec": 29496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:12.583357+00:00", "epoch": 1, "step": 2315, "train_loss": 3.714285373687744, "perplexity": 41.02925601934187, "lr": 0.001305, "grad_norm": 0.244645, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:12.890165+00:00", "epoch": 1, "step": 2316, "train_loss": 3.7277891635894775, "perplexity": 41.587064258641846, "lr": 0.001305, "grad_norm": 0.233479, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:13.196160+00:00", "epoch": 1, "step": 2317, "train_loss": 3.6974432468414307, "perplexity": 40.34402267603768, "lr": 0.001305, "grad_norm": 0.251011, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:13.500870+00:00", "epoch": 1, "step": 2318, "train_loss": 3.538513422012329, "perplexity": 34.41571949308618, "lr": 0.001305, "grad_norm": 0.228725, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:13.806883+00:00", "epoch": 1, "step": 2319, "train_loss": 3.6530098915100098, "perplexity": 38.590645074306515, "lr": 0.001305, "grad_norm": 0.240243, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:14.113742+00:00", "epoch": 1, "step": 2320, "train_loss": 3.6828272342681885, "perplexity": 39.758642317778325, "lr": 0.001305, "grad_norm": 0.251104, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:14.419574+00:00", "epoch": 1, "step": 2321, "train_loss": 3.6914637088775635, "perplexity": 40.10350387313057, "lr": 0.001305, "grad_norm": 0.273847, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:14.725884+00:00", "epoch": 1, "step": 2322, "train_loss": 3.6737146377563477, "perplexity": 39.39798361682513, "lr": 0.001305, "grad_norm": 0.266382, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:15.031608+00:00", "epoch": 1, "step": 2323, "train_loss": 3.693570852279663, "perplexity": 40.18809680013215, "lr": 0.001305, "grad_norm": 0.2307, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:15.336522+00:00", "epoch": 1, "step": 2324, "train_loss": 3.682704448699951, "perplexity": 39.75376082998325, "lr": 0.001305, "grad_norm": 0.254215, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:15.642761+00:00", "epoch": 1, "step": 2325, "train_loss": 3.746831178665161, "perplexity": 42.386553549124756, "lr": 0.001305, "grad_norm": 0.278165, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:15.949303+00:00", "epoch": 1, "step": 2326, "train_loss": 3.6580517292022705, "perplexity": 38.78570415812928, "lr": 0.001305, "grad_norm": 0.278341, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:16.255322+00:00", "epoch": 1, "step": 2327, "train_loss": 3.6023824214935303, "perplexity": 36.685530811122234, "lr": 0.001305, "grad_norm": 0.296053, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:16.560992+00:00", "epoch": 1, "step": 2328, "train_loss": 3.7565295696258545, "perplexity": 42.799634790788986, "lr": 0.001305, "grad_norm": 0.280909, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:16.866138+00:00", "epoch": 1, "step": 2329, "train_loss": 3.6938960552215576, "perplexity": 40.20116821275642, "lr": 0.001305, "grad_norm": 0.240256, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:17.171907+00:00", "epoch": 1, "step": 2330, "train_loss": 3.786677837371826, "perplexity": 44.1096172704375, "lr": 0.001305, "grad_norm": 0.229243, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:17.478758+00:00", "epoch": 1, "step": 2331, "train_loss": 3.821291208267212, "perplexity": 45.66313088411628, "lr": 0.001305, "grad_norm": 0.242023, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:17.784437+00:00", "epoch": 1, "step": 2332, "train_loss": 3.751955509185791, "perplexity": 42.60431372017701, "lr": 0.001305, "grad_norm": 0.247937, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:18.091438+00:00", "epoch": 1, "step": 2333, "train_loss": 3.6693837642669678, "perplexity": 39.2277248846796, "lr": 0.001305, "grad_norm": 0.230909, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:18.396974+00:00", "epoch": 1, "step": 2334, "train_loss": 3.570476770401001, "perplexity": 35.53353044926782, "lr": 0.001305, "grad_norm": 0.229344, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:18.702996+00:00", "epoch": 1, "step": 2335, "train_loss": 3.745417833328247, "perplexity": 42.326689025905786, "lr": 0.001305, "grad_norm": 0.234309, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:19.008827+00:00", "epoch": 1, "step": 2336, "train_loss": 3.7089030742645264, "perplexity": 40.809017504825306, "lr": 0.001305, "grad_norm": 0.259937, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:19.315447+00:00", "epoch": 1, "step": 2337, "train_loss": 3.6559062004089355, "perplexity": 38.70257752027377, "lr": 0.001305, "grad_norm": 0.21159, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:19.621949+00:00", "epoch": 1, "step": 2338, "train_loss": 3.5931456089019775, "perplexity": 36.34823361185026, "lr": 0.001305, "grad_norm": 0.216921, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:19.927443+00:00", "epoch": 1, "step": 2339, "train_loss": 3.6932942867279053, "perplexity": 40.17698369378856, "lr": 0.001305, "grad_norm": 0.213518, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:20.233041+00:00", "epoch": 1, "step": 2340, "train_loss": 3.696523427963257, "perplexity": 40.306930543995634, "lr": 0.001305, "grad_norm": 0.219882, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:20.538313+00:00", "epoch": 1, "step": 2341, "train_loss": 3.663083076477051, "perplexity": 38.9813402489002, "lr": 0.001305, "grad_norm": 0.213078, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:20.843980+00:00", "epoch": 1, "step": 2342, "train_loss": 3.6675796508789062, "perplexity": 39.15701742235004, "lr": 0.001305, "grad_norm": 0.205646, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:21.151197+00:00", "epoch": 1, "step": 2343, "train_loss": 3.7165215015411377, "perplexity": 41.12110533666653, "lr": 0.001305, "grad_norm": 0.21317, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:21.457225+00:00", "epoch": 1, "step": 2344, "train_loss": 3.650678873062134, "perplexity": 38.50079433127742, "lr": 0.001305, "grad_norm": 0.225265, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:21.762917+00:00", "epoch": 1, "step": 2345, "train_loss": 3.7201757431030273, "perplexity": 41.27164667855189, "lr": 0.001305, "grad_norm": 0.268135, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:22.069738+00:00", "epoch": 1, "step": 2346, "train_loss": 3.638603925704956, "perplexity": 38.03869479496844, "lr": 0.001305, "grad_norm": 0.259301, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:22.376902+00:00", "epoch": 1, "step": 2347, "train_loss": 3.7271037101745605, "perplexity": 41.558568030963414, "lr": 0.001305, "grad_norm": 0.285119, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:22.683423+00:00", "epoch": 1, "step": 2348, "train_loss": 3.695368528366089, "perplexity": 40.26040695635468, "lr": 0.001305, "grad_norm": 0.271772, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:22.991446+00:00", "epoch": 1, "step": 2349, "train_loss": 3.5692155361175537, "perplexity": 35.48874259238063, "lr": 0.001305, "grad_norm": 0.249537, "tokens_per_sec": 106382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:23.296948+00:00", "epoch": 1, "step": 2350, "train_loss": 3.7047412395477295, "perplexity": 40.639530053034896, "lr": 0.001305, "grad_norm": 0.224767, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:23.602600+00:00", "epoch": 1, "step": 2351, "train_loss": 3.610074281692505, "perplexity": 36.96879881777843, "lr": 0.001305, "grad_norm": 0.233533, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:23.908825+00:00", "epoch": 1, "step": 2352, "train_loss": 3.8030245304107666, "perplexity": 44.836589249879005, "lr": 0.001305, "grad_norm": 0.231864, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:24.215486+00:00", "epoch": 1, "step": 2353, "train_loss": 3.7465734481811523, "perplexity": 42.3756306498056, "lr": 0.001305, "grad_norm": 0.225237, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:24.522083+00:00", "epoch": 1, "step": 2354, "train_loss": 3.812258243560791, "perplexity": 45.25251476689867, "lr": 0.001305, "grad_norm": 0.2347, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:24.828711+00:00", "epoch": 1, "step": 2355, "train_loss": 3.6182878017425537, "perplexity": 37.27369320210263, "lr": 0.001305, "grad_norm": 0.256535, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:25.135379+00:00", "epoch": 1, "step": 2356, "train_loss": 3.7847225666046143, "perplexity": 44.02345528768913, "lr": 0.001305, "grad_norm": 0.246928, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:25.441511+00:00", "epoch": 1, "step": 2357, "train_loss": 3.5868070125579834, "perplexity": 36.118565486899, "lr": 0.001305, "grad_norm": 0.263848, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:25.747223+00:00", "epoch": 1, "step": 2358, "train_loss": 3.724008560180664, "perplexity": 41.43013688873093, "lr": 0.001305, "grad_norm": 0.260875, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:26.054447+00:00", "epoch": 1, "step": 2359, "train_loss": 3.7010929584503174, "perplexity": 40.49153575029306, "lr": 0.001305, "grad_norm": 0.271671, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:26.361291+00:00", "epoch": 1, "step": 2360, "train_loss": 3.6923136711120605, "perplexity": 40.13760482710228, "lr": 0.001305, "grad_norm": 0.27205, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:26.667414+00:00", "epoch": 1, "step": 2361, "train_loss": 3.726943016052246, "perplexity": 41.55189034989545, "lr": 0.001305, "grad_norm": 0.22852, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:26.973255+00:00", "epoch": 1, "step": 2362, "train_loss": 3.751906633377075, "perplexity": 42.60223145077587, "lr": 0.001305, "grad_norm": 0.245038, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:27.279067+00:00", "epoch": 1, "step": 2363, "train_loss": 3.634474039077759, "perplexity": 37.8819232452062, "lr": 0.001305, "grad_norm": 0.237676, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:27.585425+00:00", "epoch": 1, "step": 2364, "train_loss": 3.6627330780029297, "perplexity": 38.96769922660164, "lr": 0.001305, "grad_norm": 0.232648, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:27.893658+00:00", "epoch": 1, "step": 2365, "train_loss": 3.6129846572875977, "perplexity": 37.07654862779034, "lr": 0.001305, "grad_norm": 0.265889, "tokens_per_sec": 106308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:28.201089+00:00", "epoch": 1, "step": 2366, "train_loss": 3.676252603530884, "perplexity": 39.4981013447734, "lr": 0.001305, "grad_norm": 0.246869, "tokens_per_sec": 106587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:28.507758+00:00", "epoch": 1, "step": 2367, "train_loss": 3.6935558319091797, "perplexity": 40.18749316456262, "lr": 0.001305, "grad_norm": 0.268677, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:28.814082+00:00", "epoch": 1, "step": 2368, "train_loss": 3.6239709854125977, "perplexity": 37.4861295321898, "lr": 0.001305, "grad_norm": 0.279606, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:29.120419+00:00", "epoch": 1, "step": 2369, "train_loss": 3.653031349182129, "perplexity": 38.59147314859963, "lr": 0.001305, "grad_norm": 0.247826, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:29.427879+00:00", "epoch": 1, "step": 2370, "train_loss": 3.732623815536499, "perplexity": 41.7886100491299, "lr": 0.001305, "grad_norm": 0.235317, "tokens_per_sec": 106631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:29.734881+00:00", "epoch": 1, "step": 2371, "train_loss": 3.645779848098755, "perplexity": 38.3126392432921, "lr": 0.001305, "grad_norm": 0.22594, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:30.040560+00:00", "epoch": 1, "step": 2372, "train_loss": 3.685436248779297, "perplexity": 39.86250862796484, "lr": 0.001305, "grad_norm": 0.223819, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:30.346602+00:00", "epoch": 1, "step": 2373, "train_loss": 3.6417007446289062, "perplexity": 38.15667633429398, "lr": 0.001305, "grad_norm": 0.208458, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:30.652092+00:00", "epoch": 1, "step": 2374, "train_loss": 3.737215757369995, "perplexity": 41.980942166768266, "lr": 0.001305, "grad_norm": 0.248302, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:30.958411+00:00", "epoch": 1, "step": 2375, "train_loss": 3.6973891258239746, "perplexity": 40.34183927556665, "lr": 0.001305, "grad_norm": 0.260178, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:31.264489+00:00", "epoch": 1, "step": 2376, "train_loss": 3.623047113418579, "perplexity": 37.451513139970714, "lr": 0.001305, "grad_norm": 0.254952, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:31.571876+00:00", "epoch": 1, "step": 2377, "train_loss": 3.769761323928833, "perplexity": 43.369712287911504, "lr": 0.001305, "grad_norm": 0.249466, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:31.878681+00:00", "epoch": 1, "step": 2378, "train_loss": 3.650421142578125, "perplexity": 38.490872781517524, "lr": 0.001305, "grad_norm": 0.24864, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:32.185158+00:00", "epoch": 1, "step": 2379, "train_loss": 3.5275509357452393, "perplexity": 34.04049807828778, "lr": 0.001305, "grad_norm": 0.221403, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:32.491629+00:00", "epoch": 1, "step": 2380, "train_loss": 3.7287068367004395, "perplexity": 41.62524510536463, "lr": 0.001305, "grad_norm": 0.244875, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:32.798239+00:00", "epoch": 1, "step": 2381, "train_loss": 3.7077372074127197, "perplexity": 40.76146734802347, "lr": 0.001305, "grad_norm": 0.260047, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:33.105399+00:00", "epoch": 1, "step": 2382, "train_loss": 3.718067169189453, "perplexity": 41.18471404514591, "lr": 0.001305, "grad_norm": 0.233885, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:33.411780+00:00", "epoch": 1, "step": 2383, "train_loss": 3.681360960006714, "perplexity": 39.70038796274434, "lr": 0.001305, "grad_norm": 0.270489, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:33.718595+00:00", "epoch": 1, "step": 2384, "train_loss": 3.6587960720062256, "perplexity": 38.81458476511827, "lr": 0.001305, "grad_norm": 0.305517, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:34.025469+00:00", "epoch": 1, "step": 2385, "train_loss": 3.7925264835357666, "perplexity": 44.36835470859678, "lr": 0.001305, "grad_norm": 0.295291, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:34.331199+00:00", "epoch": 1, "step": 2386, "train_loss": 3.7628278732299805, "perplexity": 43.0700505712715, "lr": 0.001305, "grad_norm": 0.291555, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:34.638250+00:00", "epoch": 1, "step": 2387, "train_loss": 3.819277763366699, "perplexity": 45.57128318226857, "lr": 0.001305, "grad_norm": 0.314716, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:34.945019+00:00", "epoch": 1, "step": 2388, "train_loss": 3.543504476547241, "perplexity": 34.587919598525446, "lr": 0.001305, "grad_norm": 0.297939, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:35.253505+00:00", "epoch": 1, "step": 2389, "train_loss": 3.7319533824920654, "perplexity": 41.76060297355554, "lr": 0.001305, "grad_norm": 0.267462, "tokens_per_sec": 106222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:35.559839+00:00", "epoch": 1, "step": 2390, "train_loss": 3.6753761768341064, "perplexity": 39.46349931956704, "lr": 0.001305, "grad_norm": 0.293573, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:35.866251+00:00", "epoch": 1, "step": 2391, "train_loss": 3.5969061851501465, "perplexity": 36.4851812554698, "lr": 0.001305, "grad_norm": 0.252455, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:36.172597+00:00", "epoch": 1, "step": 2392, "train_loss": 3.6620609760284424, "perplexity": 38.941517758304045, "lr": 0.001305, "grad_norm": 0.225954, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:36.480554+00:00", "epoch": 1, "step": 2393, "train_loss": 3.693976640701294, "perplexity": 40.204407973719896, "lr": 0.001305, "grad_norm": 0.229944, "tokens_per_sec": 106404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:36.786963+00:00", "epoch": 1, "step": 2394, "train_loss": 3.6878180503845215, "perplexity": 39.957566374411364, "lr": 0.001305, "grad_norm": 0.243744, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:37.094415+00:00", "epoch": 1, "step": 2395, "train_loss": 3.7118046283721924, "perplexity": 40.927599029356074, "lr": 0.001305, "grad_norm": 0.239456, "tokens_per_sec": 106579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:37.400827+00:00", "epoch": 1, "step": 2396, "train_loss": 3.738440990447998, "perplexity": 42.03241012944326, "lr": 0.001305, "grad_norm": 0.229242, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:37.707485+00:00", "epoch": 1, "step": 2397, "train_loss": 3.688877820968628, "perplexity": 39.99993467424101, "lr": 0.001305, "grad_norm": 0.225361, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:38.013647+00:00", "epoch": 1, "step": 2398, "train_loss": 3.6386537551879883, "perplexity": 38.040590290690695, "lr": 0.001305, "grad_norm": 0.219491, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:38.321182+00:00", "epoch": 1, "step": 2399, "train_loss": 3.5964126586914062, "perplexity": 36.46717929575539, "lr": 0.001305, "grad_norm": 0.223924, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:38.629560+00:00", "epoch": 1, "step": 2400, "train_loss": 3.67429518699646, "perplexity": 39.42086272685821, "lr": 0.001305, "grad_norm": 0.214526, "tokens_per_sec": 106322} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:58:44.290719+00:00", "step": 2400, "epoch": 1, "val_loss": 3.669832146167755, "val_ppl": 39.24531783041144, "eval_train_loss": 3.67429518699646, "eval_train_ppl": 39.42086272685821} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:58:45.268822+00:00", "step": 2400, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p6698_epoch_0001_step_0002400.pt", "val_loss": 3.669832146167755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:46.267503+00:00", "epoch": 1, "step": 2401, "train_loss": 3.715322256088257, "perplexity": 41.07182059623126, "lr": 0.001305, "grad_norm": 0.218406, "tokens_per_sec": 4290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:46.572877+00:00", "epoch": 1, "step": 2402, "train_loss": 3.5284438133239746, "perplexity": 34.07090564888973, "lr": 0.001305, "grad_norm": 0.246677, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:46.879329+00:00", "epoch": 1, "step": 2403, "train_loss": 3.6426799297332764, "perplexity": 38.1940570817416, "lr": 0.001305, "grad_norm": 0.260785, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:47.184601+00:00", "epoch": 1, "step": 2404, "train_loss": 3.728440523147583, "perplexity": 41.614161214412796, "lr": 0.001305, "grad_norm": 0.257074, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:47.490310+00:00", "epoch": 1, "step": 2405, "train_loss": 3.708272933959961, "perplexity": 40.78331019856135, "lr": 0.001305, "grad_norm": 0.275022, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:47.795173+00:00", "epoch": 1, "step": 2406, "train_loss": 3.761942148208618, "perplexity": 43.03191923923949, "lr": 0.001305, "grad_norm": 0.263751, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:48.101079+00:00", "epoch": 1, "step": 2407, "train_loss": 3.7867894172668457, "perplexity": 44.11453929149607, "lr": 0.001305, "grad_norm": 0.233881, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:48.406764+00:00", "epoch": 1, "step": 2408, "train_loss": 3.591245412826538, "perplexity": 36.27923042155031, "lr": 0.001305, "grad_norm": 0.227091, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:48.713859+00:00", "epoch": 1, "step": 2409, "train_loss": 3.6731820106506348, "perplexity": 39.37700477028749, "lr": 0.001305, "grad_norm": 0.262382, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:49.020243+00:00", "epoch": 1, "step": 2410, "train_loss": 3.679157257080078, "perplexity": 39.612996429457766, "lr": 0.001305, "grad_norm": 0.250058, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:49.325991+00:00", "epoch": 1, "step": 2411, "train_loss": 3.696150302886963, "perplexity": 40.291893822924536, "lr": 0.001305, "grad_norm": 0.244735, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:49.631010+00:00", "epoch": 1, "step": 2412, "train_loss": 3.690460443496704, "perplexity": 40.06328959221454, "lr": 0.001305, "grad_norm": 0.24469, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:49.937410+00:00", "epoch": 1, "step": 2413, "train_loss": 3.6860499382019043, "perplexity": 39.8869793358077, "lr": 0.001305, "grad_norm": 0.247409, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:50.244535+00:00", "epoch": 1, "step": 2414, "train_loss": 3.7347302436828613, "perplexity": 41.87672752752135, "lr": 0.001305, "grad_norm": 0.235856, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:50.558908+00:00", "epoch": 1, "step": 2415, "train_loss": 3.7327678203582764, "perplexity": 41.79462824378648, "lr": 0.001305, "grad_norm": 0.260805, "tokens_per_sec": 104233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:50.865319+00:00", "epoch": 1, "step": 2416, "train_loss": 3.614311695098877, "perplexity": 37.12578327062218, "lr": 0.001305, "grad_norm": 0.262966, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:51.172388+00:00", "epoch": 1, "step": 2417, "train_loss": 3.6968190670013428, "perplexity": 40.31884860780551, "lr": 0.001305, "grad_norm": 0.250833, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:51.478542+00:00", "epoch": 1, "step": 2418, "train_loss": 3.6632583141326904, "perplexity": 38.98817184613817, "lr": 0.001305, "grad_norm": 0.230993, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:51.786075+00:00", "epoch": 1, "step": 2419, "train_loss": 3.6960489749908447, "perplexity": 40.28781133693119, "lr": 0.001305, "grad_norm": 0.221508, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:52.092674+00:00", "epoch": 1, "step": 2420, "train_loss": 3.6491997241973877, "perplexity": 38.443888021877314, "lr": 0.001305, "grad_norm": 0.236374, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:52.400761+00:00", "epoch": 1, "step": 2421, "train_loss": 3.8488872051239014, "perplexity": 46.940798676860666, "lr": 0.001305, "grad_norm": 0.230389, "tokens_per_sec": 106306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:52.707517+00:00", "epoch": 1, "step": 2422, "train_loss": 3.5530824661254883, "perplexity": 34.92079392060563, "lr": 0.001305, "grad_norm": 0.240997, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:53.014297+00:00", "epoch": 1, "step": 2423, "train_loss": 3.7422544956207275, "perplexity": 42.19300696670143, "lr": 0.001305, "grad_norm": 0.223284, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:53.319892+00:00", "epoch": 1, "step": 2424, "train_loss": 3.760211706161499, "perplexity": 42.9575193877008, "lr": 0.001305, "grad_norm": 0.238232, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:53.626490+00:00", "epoch": 1, "step": 2425, "train_loss": 3.650195837020874, "perplexity": 38.48220155085132, "lr": 0.001305, "grad_norm": 0.279035, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:53.933971+00:00", "epoch": 1, "step": 2426, "train_loss": 3.6913137435913086, "perplexity": 40.09749019062549, "lr": 0.001305, "grad_norm": 0.33004, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:54.241371+00:00", "epoch": 1, "step": 2427, "train_loss": 3.6906750202178955, "perplexity": 40.071887163921765, "lr": 0.001305, "grad_norm": 0.318122, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:54.548144+00:00", "epoch": 1, "step": 2428, "train_loss": 3.718461036682129, "perplexity": 41.20093856014822, "lr": 0.001305, "grad_norm": 0.24934, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:54.854930+00:00", "epoch": 1, "step": 2429, "train_loss": 3.7172601222991943, "perplexity": 41.151489458452595, "lr": 0.001305, "grad_norm": 0.277528, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:55.161653+00:00", "epoch": 1, "step": 2430, "train_loss": 3.654460906982422, "perplexity": 38.64668134231387, "lr": 0.001305, "grad_norm": 0.282984, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:55.468379+00:00", "epoch": 1, "step": 2431, "train_loss": 3.69327712059021, "perplexity": 40.176294016073854, "lr": 0.001305, "grad_norm": 0.213716, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:55.774895+00:00", "epoch": 1, "step": 2432, "train_loss": 3.6860458850860596, "perplexity": 39.88681766958738, "lr": 0.001305, "grad_norm": 0.225968, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:56.083147+00:00", "epoch": 1, "step": 2433, "train_loss": 3.7610650062561035, "perplexity": 42.994190686644636, "lr": 0.001305, "grad_norm": 0.249724, "tokens_per_sec": 106304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:56.389491+00:00", "epoch": 1, "step": 2434, "train_loss": 3.778200626373291, "perplexity": 43.73727119571743, "lr": 0.001305, "grad_norm": 0.245508, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:56.695653+00:00", "epoch": 1, "step": 2435, "train_loss": 3.717365264892578, "perplexity": 41.15581646024797, "lr": 0.001305, "grad_norm": 0.218677, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:57.001861+00:00", "epoch": 1, "step": 2436, "train_loss": 3.7741456031799316, "perplexity": 43.56027465168275, "lr": 0.001305, "grad_norm": 0.25199, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:57.309067+00:00", "epoch": 1, "step": 2437, "train_loss": 3.696051597595215, "perplexity": 40.28791699605981, "lr": 0.001305, "grad_norm": 0.232736, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:57.616166+00:00", "epoch": 1, "step": 2438, "train_loss": 3.6876473426818848, "perplexity": 39.95074589222362, "lr": 0.001305, "grad_norm": 0.232445, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:57.922506+00:00", "epoch": 1, "step": 2439, "train_loss": 3.754465103149414, "perplexity": 42.71136752331659, "lr": 0.001305, "grad_norm": 0.226571, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:58.228161+00:00", "epoch": 1, "step": 2440, "train_loss": 3.6381635665893555, "perplexity": 38.021947796587256, "lr": 0.001305, "grad_norm": 0.219104, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:58.534510+00:00", "epoch": 1, "step": 2441, "train_loss": 3.739489793777466, "perplexity": 42.076516986798914, "lr": 0.001305, "grad_norm": 0.278796, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:58.840909+00:00", "epoch": 1, "step": 2442, "train_loss": 3.6548125743865967, "perplexity": 38.660274510418475, "lr": 0.001305, "grad_norm": 0.269814, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:59.147315+00:00", "epoch": 1, "step": 2443, "train_loss": 3.6636035442352295, "perplexity": 39.00163406034952, "lr": 0.001305, "grad_norm": 0.235186, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:59.454827+00:00", "epoch": 1, "step": 2444, "train_loss": 3.736572027206421, "perplexity": 41.95392646434444, "lr": 0.001305, "grad_norm": 0.24309, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:58:59.761620+00:00", "epoch": 1, "step": 2445, "train_loss": 3.649914264678955, "perplexity": 38.47136755258716, "lr": 0.001305, "grad_norm": 0.282443, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:00.068262+00:00", "epoch": 1, "step": 2446, "train_loss": 3.7208714485168457, "perplexity": 41.30036957676275, "lr": 0.001305, "grad_norm": 0.26601, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:00.375379+00:00", "epoch": 1, "step": 2447, "train_loss": 3.6526689529418945, "perplexity": 38.57749027764844, "lr": 0.001305, "grad_norm": 0.240884, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:00.682194+00:00", "epoch": 1, "step": 2448, "train_loss": 3.619084596633911, "perplexity": 37.30340452577161, "lr": 0.001305, "grad_norm": 0.233576, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:00.989394+00:00", "epoch": 1, "step": 2449, "train_loss": 3.734757900238037, "perplexity": 41.87788570956239, "lr": 0.001305, "grad_norm": 0.257622, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:01.295723+00:00", "epoch": 1, "step": 2450, "train_loss": 3.730485439300537, "perplexity": 41.6993457528065, "lr": 0.001305, "grad_norm": 0.226724, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:01.601867+00:00", "epoch": 1, "step": 2451, "train_loss": 3.7438015937805176, "perplexity": 42.25833421093516, "lr": 0.001305, "grad_norm": 0.22087, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:01.908191+00:00", "epoch": 1, "step": 2452, "train_loss": 3.6770122051239014, "perplexity": 39.52811556345706, "lr": 0.001305, "grad_norm": 0.255433, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:02.214534+00:00", "epoch": 1, "step": 2453, "train_loss": 3.6651790142059326, "perplexity": 39.06312839213688, "lr": 0.001305, "grad_norm": 0.275219, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:02.521922+00:00", "epoch": 1, "step": 2454, "train_loss": 3.652048349380493, "perplexity": 38.553556377292274, "lr": 0.001305, "grad_norm": 0.267688, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:02.829403+00:00", "epoch": 1, "step": 2455, "train_loss": 3.613114833831787, "perplexity": 37.08137543892315, "lr": 0.001305, "grad_norm": 0.257817, "tokens_per_sec": 106571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:03.135615+00:00", "epoch": 1, "step": 2456, "train_loss": 3.6089677810668945, "perplexity": 36.927915441670386, "lr": 0.001305, "grad_norm": 0.222556, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:03.441897+00:00", "epoch": 1, "step": 2457, "train_loss": 3.763640880584717, "perplexity": 43.10508107725477, "lr": 0.001305, "grad_norm": 0.263783, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:03.748654+00:00", "epoch": 1, "step": 2458, "train_loss": 3.611178398132324, "perplexity": 37.009639218443446, "lr": 0.001305, "grad_norm": 0.261994, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:04.056642+00:00", "epoch": 1, "step": 2459, "train_loss": 3.6434903144836426, "perplexity": 38.225021508010094, "lr": 0.001305, "grad_norm": 0.243095, "tokens_per_sec": 106393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:04.363495+00:00", "epoch": 1, "step": 2460, "train_loss": 3.6531615257263184, "perplexity": 38.596497180207734, "lr": 0.001305, "grad_norm": 0.231023, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:04.670204+00:00", "epoch": 1, "step": 2461, "train_loss": 3.7143685817718506, "perplexity": 41.03267012716627, "lr": 0.001305, "grad_norm": 0.263824, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:04.976291+00:00", "epoch": 1, "step": 2462, "train_loss": 3.6937320232391357, "perplexity": 40.19457447624543, "lr": 0.001305, "grad_norm": 0.238124, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:05.282959+00:00", "epoch": 1, "step": 2463, "train_loss": 3.6392476558685303, "perplexity": 38.063189333281684, "lr": 0.001305, "grad_norm": 0.241586, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:05.589786+00:00", "epoch": 1, "step": 2464, "train_loss": 3.689171552658081, "perplexity": 40.01168564836314, "lr": 0.001305, "grad_norm": 0.238335, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:05.897482+00:00", "epoch": 1, "step": 2465, "train_loss": 3.743344783782959, "perplexity": 42.23903458985298, "lr": 0.001305, "grad_norm": 0.247269, "tokens_per_sec": 106494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:06.204835+00:00", "epoch": 1, "step": 2466, "train_loss": 3.684394598007202, "perplexity": 39.82100743370294, "lr": 0.001305, "grad_norm": 0.24997, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:06.511179+00:00", "epoch": 1, "step": 2467, "train_loss": 3.615133047103882, "perplexity": 37.156289133459694, "lr": 0.001305, "grad_norm": 0.267267, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:06.816891+00:00", "epoch": 1, "step": 2468, "train_loss": 3.724846124649048, "perplexity": 41.464851835282666, "lr": 0.001305, "grad_norm": 0.23684, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:07.124052+00:00", "epoch": 1, "step": 2469, "train_loss": 3.6897637844085693, "perplexity": 40.035388857198285, "lr": 0.001305, "grad_norm": 0.238849, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:07.431650+00:00", "epoch": 1, "step": 2470, "train_loss": 3.7181198596954346, "perplexity": 41.186884145739, "lr": 0.001305, "grad_norm": 0.264477, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:07.738886+00:00", "epoch": 1, "step": 2471, "train_loss": 3.7049484252929688, "perplexity": 40.64795085666022, "lr": 0.001305, "grad_norm": 0.267707, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:08.045786+00:00", "epoch": 1, "step": 2472, "train_loss": 3.7224934101104736, "perplexity": 41.36741154507017, "lr": 0.001305, "grad_norm": 0.251749, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:08.352908+00:00", "epoch": 1, "step": 2473, "train_loss": 3.734241247177124, "perplexity": 41.85625496000362, "lr": 0.001305, "grad_norm": 0.238727, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:08.659464+00:00", "epoch": 1, "step": 2474, "train_loss": 3.715136766433716, "perplexity": 41.06420290494081, "lr": 0.001305, "grad_norm": 0.207276, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:08.966532+00:00", "epoch": 1, "step": 2475, "train_loss": 3.5602524280548096, "perplexity": 35.172074442955214, "lr": 0.001305, "grad_norm": 0.218589, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:09.272712+00:00", "epoch": 1, "step": 2476, "train_loss": 3.5945756435394287, "perplexity": 36.400250028699816, "lr": 0.001305, "grad_norm": 0.223671, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:09.579992+00:00", "epoch": 1, "step": 2477, "train_loss": 3.757603645324707, "perplexity": 42.84562953493424, "lr": 0.001305, "grad_norm": 0.207356, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:09.887244+00:00", "epoch": 1, "step": 2478, "train_loss": 3.768231153488159, "perplexity": 43.303399983650685, "lr": 0.001305, "grad_norm": 0.205449, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:10.193929+00:00", "epoch": 1, "step": 2479, "train_loss": 3.6499578952789307, "perplexity": 38.4730461180535, "lr": 0.001305, "grad_norm": 0.224651, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:10.501047+00:00", "epoch": 1, "step": 2480, "train_loss": 3.700180768966675, "perplexity": 40.45461663837793, "lr": 0.001305, "grad_norm": 0.235236, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:10.809729+00:00", "epoch": 1, "step": 2481, "train_loss": 3.718668222427368, "perplexity": 41.2094756916638, "lr": 0.001305, "grad_norm": 0.227893, "tokens_per_sec": 106099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:11.116894+00:00", "epoch": 1, "step": 2482, "train_loss": 3.6669604778289795, "perplexity": 39.13277995681367, "lr": 0.001305, "grad_norm": 0.221252, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:11.423497+00:00", "epoch": 1, "step": 2483, "train_loss": 3.642174005508423, "perplexity": 38.17473867025675, "lr": 0.001305, "grad_norm": 0.219363, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:11.729933+00:00", "epoch": 1, "step": 2484, "train_loss": 3.6438798904418945, "perplexity": 38.239915958464806, "lr": 0.001305, "grad_norm": 0.223346, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:12.037224+00:00", "epoch": 1, "step": 2485, "train_loss": 3.578376531600952, "perplexity": 35.81534853660536, "lr": 0.001305, "grad_norm": 0.255126, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:12.344788+00:00", "epoch": 1, "step": 2486, "train_loss": 3.7011048793792725, "perplexity": 40.492018449891134, "lr": 0.001305, "grad_norm": 0.24479, "tokens_per_sec": 106541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:12.651700+00:00", "epoch": 1, "step": 2487, "train_loss": 3.7363102436065674, "perplexity": 41.94294505188617, "lr": 0.001305, "grad_norm": 0.260174, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:12.958673+00:00", "epoch": 1, "step": 2488, "train_loss": 3.6238138675689697, "perplexity": 37.480240255019126, "lr": 0.001305, "grad_norm": 0.267207, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:13.266066+00:00", "epoch": 1, "step": 2489, "train_loss": 3.7274739742279053, "perplexity": 41.57395852391108, "lr": 0.001305, "grad_norm": 0.233979, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:13.573753+00:00", "epoch": 1, "step": 2490, "train_loss": 3.75014591217041, "perplexity": 42.52728679609222, "lr": 0.001305, "grad_norm": 0.217103, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:13.882459+00:00", "epoch": 1, "step": 2491, "train_loss": 3.684635639190674, "perplexity": 39.830607093371974, "lr": 0.001305, "grad_norm": 0.245561, "tokens_per_sec": 106091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:14.189970+00:00", "epoch": 1, "step": 2492, "train_loss": 3.5795187950134277, "perplexity": 35.85628247306199, "lr": 0.001305, "grad_norm": 0.249855, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:14.495982+00:00", "epoch": 1, "step": 2493, "train_loss": 3.6137301921844482, "perplexity": 37.10420079519421, "lr": 0.001305, "grad_norm": 0.262236, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:14.802588+00:00", "epoch": 1, "step": 2494, "train_loss": 3.7113022804260254, "perplexity": 40.90704429728811, "lr": 0.001305, "grad_norm": 0.265025, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:15.109381+00:00", "epoch": 1, "step": 2495, "train_loss": 3.6603715419769287, "perplexity": 38.875784174097944, "lr": 0.001305, "grad_norm": 0.268662, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:15.416926+00:00", "epoch": 1, "step": 2496, "train_loss": 3.695925712585449, "perplexity": 40.28284567044397, "lr": 0.001305, "grad_norm": 0.260071, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:15.724692+00:00", "epoch": 1, "step": 2497, "train_loss": 3.6739745140075684, "perplexity": 39.408223547612906, "lr": 0.001305, "grad_norm": 0.273284, "tokens_per_sec": 106463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:16.031347+00:00", "epoch": 1, "step": 2498, "train_loss": 3.6863977909088135, "perplexity": 39.9008565430122, "lr": 0.001305, "grad_norm": 0.253919, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:16.337941+00:00", "epoch": 1, "step": 2499, "train_loss": 3.6698074340820312, "perplexity": 39.244348008736154, "lr": 0.001305, "grad_norm": 0.246498, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:16.644788+00:00", "epoch": 1, "step": 2500, "train_loss": 3.7505249977111816, "perplexity": 42.54341133170091, "lr": 0.001305, "grad_norm": 0.275897, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T12:59:22.305070+00:00", "step": 2500, "epoch": 1, "val_loss": 3.661393630504608, "val_ppl": 38.91553898011153, "eval_train_loss": 3.7505249977111816, "eval_train_ppl": 42.54341133170091} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T12:59:23.341885+00:00", "step": 2500, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p6614_epoch_0001_step_0002500.pt", "val_loss": 3.661393630504608} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T12:59:25.146959+00:00", "step": 2500, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0001_step_0002500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:26.176681+00:00", "epoch": 1, "step": 2501, "train_loss": 3.71016001701355, "perplexity": 40.8603443541707, "lr": 0.001305, "grad_norm": 0.255139, "tokens_per_sec": 3438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:26.482078+00:00", "epoch": 1, "step": 2502, "train_loss": 3.7090089321136475, "perplexity": 40.81333768830174, "lr": 0.001305, "grad_norm": 0.263429, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:26.787972+00:00", "epoch": 1, "step": 2503, "train_loss": 3.7910354137420654, "perplexity": 44.30224769242997, "lr": 0.001305, "grad_norm": 0.258242, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:27.093329+00:00", "epoch": 1, "step": 2504, "train_loss": 3.7148513793945312, "perplexity": 41.05248538575087, "lr": 0.001305, "grad_norm": 0.261981, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:27.399459+00:00", "epoch": 1, "step": 2505, "train_loss": 3.781338691711426, "perplexity": 43.874737186263886, "lr": 0.001305, "grad_norm": 0.286034, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:27.705918+00:00", "epoch": 1, "step": 2506, "train_loss": 3.5709760189056396, "perplexity": 35.55127494029589, "lr": 0.001305, "grad_norm": 0.278984, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:28.011939+00:00", "epoch": 1, "step": 2507, "train_loss": 3.6565933227539062, "perplexity": 38.72918006465058, "lr": 0.0013028489010989011, "grad_norm": 0.259148, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:28.318348+00:00", "epoch": 1, "step": 2508, "train_loss": 3.7172024250030518, "perplexity": 41.14911519727349, "lr": 0.001300697802197802, "grad_norm": 0.231038, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:28.624345+00:00", "epoch": 1, "step": 2509, "train_loss": 3.7018954753875732, "perplexity": 40.52404393598829, "lr": 0.0012985467032967032, "grad_norm": 0.245461, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:28.930891+00:00", "epoch": 1, "step": 2510, "train_loss": 3.6079423427581787, "perplexity": 36.89006755117039, "lr": 0.0012963956043956044, "grad_norm": 0.253121, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:29.236465+00:00", "epoch": 1, "step": 2511, "train_loss": 3.7286648750305176, "perplexity": 41.62349847721507, "lr": 0.0012942445054945054, "grad_norm": 0.261741, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:29.543212+00:00", "epoch": 1, "step": 2512, "train_loss": 3.6890616416931152, "perplexity": 40.00728816705375, "lr": 0.0012920934065934065, "grad_norm": 0.234288, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:29.849125+00:00", "epoch": 1, "step": 2513, "train_loss": 3.7868800163269043, "perplexity": 44.11853620834652, "lr": 0.0012899423076923077, "grad_norm": 0.255887, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:30.155773+00:00", "epoch": 1, "step": 2514, "train_loss": 3.708073377609253, "perplexity": 40.77517244200584, "lr": 0.0012877912087912087, "grad_norm": 0.233169, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:30.884975+00:00", "epoch": 1, "step": 2515, "train_loss": 3.640347480773926, "perplexity": 38.10507520623019, "lr": 0.0012856401098901098, "grad_norm": 0.277715, "tokens_per_sec": 44936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:31.190556+00:00", "epoch": 1, "step": 2516, "train_loss": 3.712087869644165, "perplexity": 40.93919305643992, "lr": 0.001283489010989011, "grad_norm": 0.270179, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:31.497018+00:00", "epoch": 1, "step": 2517, "train_loss": 3.694768190383911, "perplexity": 40.23624435846917, "lr": 0.001281337912087912, "grad_norm": 0.220177, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:31.803142+00:00", "epoch": 1, "step": 2518, "train_loss": 3.5810811519622803, "perplexity": 35.91234656980518, "lr": 0.001279186813186813, "grad_norm": 0.224832, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:32.108796+00:00", "epoch": 1, "step": 2519, "train_loss": 3.613502025604248, "perplexity": 37.09573582233639, "lr": 0.0012770357142857143, "grad_norm": 0.257778, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:32.415162+00:00", "epoch": 1, "step": 2520, "train_loss": 3.7121992111206055, "perplexity": 40.943751540408556, "lr": 0.0012748846153846154, "grad_norm": 0.242284, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:32.722071+00:00", "epoch": 1, "step": 2521, "train_loss": 3.817544460296631, "perplexity": 45.492362753491186, "lr": 0.0012727335164835164, "grad_norm": 0.24212, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:33.028218+00:00", "epoch": 1, "step": 2522, "train_loss": 3.6486761569976807, "perplexity": 38.42376533132989, "lr": 0.0012705824175824176, "grad_norm": 0.220572, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:33.334943+00:00", "epoch": 1, "step": 2523, "train_loss": 3.6362228393554688, "perplexity": 37.94822912406867, "lr": 0.0012684313186813187, "grad_norm": 0.258867, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:33.640367+00:00", "epoch": 1, "step": 2524, "train_loss": 3.7895655632019043, "perplexity": 44.237177842920694, "lr": 0.0012662802197802197, "grad_norm": 0.286002, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:33.947054+00:00", "epoch": 1, "step": 2525, "train_loss": 3.7159152030944824, "perplexity": 41.09618123086298, "lr": 0.0012641291208791209, "grad_norm": 0.287993, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:34.253475+00:00", "epoch": 1, "step": 2526, "train_loss": 3.645376205444336, "perplexity": 38.297177748559456, "lr": 0.001261978021978022, "grad_norm": 0.278439, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:34.559878+00:00", "epoch": 1, "step": 2527, "train_loss": 3.6733593940734863, "perplexity": 39.38399021770826, "lr": 0.001259826923076923, "grad_norm": 0.293928, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:34.865672+00:00", "epoch": 1, "step": 2528, "train_loss": 3.7089781761169434, "perplexity": 40.81208245272543, "lr": 0.0012576758241758241, "grad_norm": 0.28315, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:35.172539+00:00", "epoch": 1, "step": 2529, "train_loss": 3.639592170715332, "perplexity": 38.07630492625215, "lr": 0.0012555247252747253, "grad_norm": 0.261496, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:35.478867+00:00", "epoch": 1, "step": 2530, "train_loss": 3.6969809532165527, "perplexity": 40.32537620195775, "lr": 0.0012533736263736263, "grad_norm": 0.257043, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:35.785752+00:00", "epoch": 1, "step": 2531, "train_loss": 3.6723239421844482, "perplexity": 39.34323109633578, "lr": 0.0012512225274725274, "grad_norm": 0.242215, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:36.093172+00:00", "epoch": 1, "step": 2532, "train_loss": 3.645141124725342, "perplexity": 38.28817587860336, "lr": 0.0012490714285714286, "grad_norm": 0.219131, "tokens_per_sec": 106591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:36.401571+00:00", "epoch": 1, "step": 2533, "train_loss": 3.5776634216308594, "perplexity": 35.78981735883221, "lr": 0.0012469203296703295, "grad_norm": 0.265476, "tokens_per_sec": 106251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:36.708051+00:00", "epoch": 1, "step": 2534, "train_loss": 3.65928316116333, "perplexity": 38.833495533736155, "lr": 0.0012447692307692307, "grad_norm": 0.253925, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:37.013453+00:00", "epoch": 1, "step": 2535, "train_loss": 3.646141767501831, "perplexity": 38.3265078403231, "lr": 0.0012426181318681319, "grad_norm": 0.236825, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:37.319470+00:00", "epoch": 1, "step": 2536, "train_loss": 3.528968572616577, "perplexity": 34.08878936514912, "lr": 0.0012404670329670328, "grad_norm": 0.232433, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:37.626233+00:00", "epoch": 1, "step": 2537, "train_loss": 3.680103063583374, "perplexity": 39.65048038258483, "lr": 0.001238315934065934, "grad_norm": 0.238239, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:37.933271+00:00", "epoch": 1, "step": 2538, "train_loss": 3.675421714782715, "perplexity": 39.465296447289404, "lr": 0.0012361648351648352, "grad_norm": 0.211568, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:38.240134+00:00", "epoch": 1, "step": 2539, "train_loss": 3.6859073638916016, "perplexity": 39.88129288261954, "lr": 0.0012340137362637361, "grad_norm": 0.231545, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:38.546356+00:00", "epoch": 1, "step": 2540, "train_loss": 3.7211458683013916, "perplexity": 41.311704770513245, "lr": 0.0012318626373626373, "grad_norm": 0.257517, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:38.852997+00:00", "epoch": 1, "step": 2541, "train_loss": 3.6630537509918213, "perplexity": 38.98019711894401, "lr": 0.0012297115384615385, "grad_norm": 0.237059, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:39.159668+00:00", "epoch": 1, "step": 2542, "train_loss": 3.64034366607666, "perplexity": 38.104929847181246, "lr": 0.0012275604395604394, "grad_norm": 0.221474, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:39.467004+00:00", "epoch": 1, "step": 2543, "train_loss": 3.7563581466674805, "perplexity": 42.7922985795913, "lr": 0.0012254093406593406, "grad_norm": 0.231001, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:39.773768+00:00", "epoch": 1, "step": 2544, "train_loss": 3.7174277305603027, "perplexity": 41.15838736609975, "lr": 0.0012232582417582418, "grad_norm": 0.243489, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:40.080502+00:00", "epoch": 1, "step": 2545, "train_loss": 3.7272839546203613, "perplexity": 41.56605940714562, "lr": 0.0012211071428571427, "grad_norm": 0.237549, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:40.387700+00:00", "epoch": 1, "step": 2546, "train_loss": 3.692744493484497, "perplexity": 40.1549007307014, "lr": 0.0012189560439560439, "grad_norm": 0.214196, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:40.694734+00:00", "epoch": 1, "step": 2547, "train_loss": 3.6752283573150635, "perplexity": 39.457666275207394, "lr": 0.001216804945054945, "grad_norm": 0.229006, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:41.001527+00:00", "epoch": 1, "step": 2548, "train_loss": 3.600076198577881, "perplexity": 36.601023283347004, "lr": 0.0012146538461538462, "grad_norm": 0.214348, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:41.309076+00:00", "epoch": 1, "step": 2549, "train_loss": 3.668308973312378, "perplexity": 39.185585930144484, "lr": 0.0012125027472527472, "grad_norm": 0.223088, "tokens_per_sec": 106545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:41.616643+00:00", "epoch": 1, "step": 2550, "train_loss": 3.7453184127807617, "perplexity": 42.32248109249058, "lr": 0.0012103516483516483, "grad_norm": 0.240629, "tokens_per_sec": 106598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:41.925511+00:00", "epoch": 1, "step": 2551, "train_loss": 3.7222886085510254, "perplexity": 41.358940302166886, "lr": 0.0012082005494505495, "grad_norm": 0.21151, "tokens_per_sec": 106035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:42.232773+00:00", "epoch": 1, "step": 2552, "train_loss": 3.7659718990325928, "perplexity": 43.20567701638776, "lr": 0.0012060494505494504, "grad_norm": 0.212686, "tokens_per_sec": 106645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:42.540541+00:00", "epoch": 1, "step": 2553, "train_loss": 3.6588566303253174, "perplexity": 38.816935382301885, "lr": 0.0012038983516483516, "grad_norm": 0.205384, "tokens_per_sec": 106471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:42.848156+00:00", "epoch": 1, "step": 2554, "train_loss": 3.634237289428711, "perplexity": 37.87295577473727, "lr": 0.0012017472527472528, "grad_norm": 0.208437, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:43.157524+00:00", "epoch": 1, "step": 2555, "train_loss": 3.6121604442596436, "perplexity": 37.04600224347294, "lr": 0.0011995961538461537, "grad_norm": 0.239219, "tokens_per_sec": 105919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:43.465320+00:00", "epoch": 1, "step": 2556, "train_loss": 3.6705212593078613, "perplexity": 39.272371615104674, "lr": 0.001197445054945055, "grad_norm": 0.241597, "tokens_per_sec": 106461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:43.773848+00:00", "epoch": 1, "step": 2557, "train_loss": 3.5997562408447266, "perplexity": 36.58931437618372, "lr": 0.001195293956043956, "grad_norm": 0.282542, "tokens_per_sec": 106206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:44.081194+00:00", "epoch": 1, "step": 2558, "train_loss": 3.721709728240967, "perplexity": 41.33500535438446, "lr": 0.001193142857142857, "grad_norm": 0.318281, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:44.387773+00:00", "epoch": 1, "step": 2559, "train_loss": 3.723759412765503, "perplexity": 41.4198159629847, "lr": 0.0011909917582417582, "grad_norm": 0.30625, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:44.694520+00:00", "epoch": 1, "step": 2560, "train_loss": 3.63922119140625, "perplexity": 38.062182024772305, "lr": 0.0011888406593406594, "grad_norm": 0.25122, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:45.002085+00:00", "epoch": 1, "step": 2561, "train_loss": 3.551464319229126, "perplexity": 34.86433263994738, "lr": 0.0011866895604395603, "grad_norm": 0.296367, "tokens_per_sec": 106487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:45.309255+00:00", "epoch": 1, "step": 2562, "train_loss": 3.6763510704040527, "perplexity": 39.50199079079658, "lr": 0.0011845384615384615, "grad_norm": 0.259249, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:45.615433+00:00", "epoch": 1, "step": 2563, "train_loss": 3.6821703910827637, "perplexity": 39.73253569942595, "lr": 0.0011823873626373627, "grad_norm": 0.23141, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:45.922329+00:00", "epoch": 1, "step": 2564, "train_loss": 3.535466432571411, "perplexity": 34.311014757309124, "lr": 0.0011802362637362636, "grad_norm": 0.230114, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:46.229536+00:00", "epoch": 1, "step": 2565, "train_loss": 3.643165349960327, "perplexity": 38.212601750216656, "lr": 0.0011780851648351648, "grad_norm": 0.230512, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:46.537358+00:00", "epoch": 1, "step": 2566, "train_loss": 3.6970932483673096, "perplexity": 40.329904800422746, "lr": 0.001175934065934066, "grad_norm": 0.224566, "tokens_per_sec": 106450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:46.846022+00:00", "epoch": 1, "step": 2567, "train_loss": 3.6809096336364746, "perplexity": 39.682474173534835, "lr": 0.0011737829670329669, "grad_norm": 0.209602, "tokens_per_sec": 106161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:47.154838+00:00", "epoch": 1, "step": 2568, "train_loss": 3.7307369709014893, "perplexity": 41.709835775233124, "lr": 0.001171631868131868, "grad_norm": 0.231286, "tokens_per_sec": 106108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:47.461483+00:00", "epoch": 1, "step": 2569, "train_loss": 3.5920207500457764, "perplexity": 36.30736996659009, "lr": 0.0011694807692307692, "grad_norm": 0.24136, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:47.768513+00:00", "epoch": 1, "step": 2570, "train_loss": 3.548121690750122, "perplexity": 34.7479886844007, "lr": 0.0011673296703296702, "grad_norm": 0.239863, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:48.075331+00:00", "epoch": 1, "step": 2571, "train_loss": 3.7927184104919434, "perplexity": 44.376871009094245, "lr": 0.0011651785714285713, "grad_norm": 0.208796, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:48.382201+00:00", "epoch": 1, "step": 2572, "train_loss": 3.668159246444702, "perplexity": 39.17971923431708, "lr": 0.0011630274725274725, "grad_norm": 0.22053, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:48.689713+00:00", "epoch": 1, "step": 2573, "train_loss": 3.6758692264556885, "perplexity": 39.48296158050878, "lr": 0.0011608763736263737, "grad_norm": 0.211687, "tokens_per_sec": 106558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:48.997664+00:00", "epoch": 1, "step": 2574, "train_loss": 3.707296371459961, "perplexity": 40.74350218786444, "lr": 0.0011587252747252746, "grad_norm": 0.228257, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:49.304935+00:00", "epoch": 1, "step": 2575, "train_loss": 3.760777235031128, "perplexity": 42.98181997577659, "lr": 0.0011565741758241758, "grad_norm": 0.21212, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:49.614609+00:00", "epoch": 1, "step": 2576, "train_loss": 3.6501269340515137, "perplexity": 38.47955010424426, "lr": 0.001154423076923077, "grad_norm": 0.206412, "tokens_per_sec": 105820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:49.922490+00:00", "epoch": 1, "step": 2577, "train_loss": 3.6592838764190674, "perplexity": 38.83352330962657, "lr": 0.001152271978021978, "grad_norm": 0.195937, "tokens_per_sec": 106426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:50.230511+00:00", "epoch": 1, "step": 2578, "train_loss": 3.698782444000244, "perplexity": 40.39808747020971, "lr": 0.001150120879120879, "grad_norm": 0.215742, "tokens_per_sec": 106382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:50.538855+00:00", "epoch": 1, "step": 2579, "train_loss": 3.6882734298706055, "perplexity": 39.97576637409085, "lr": 0.0011479697802197803, "grad_norm": 0.221357, "tokens_per_sec": 106271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:50.847238+00:00", "epoch": 1, "step": 2580, "train_loss": 3.749058485031128, "perplexity": 42.481066605369605, "lr": 0.0011458186813186812, "grad_norm": 0.239967, "tokens_per_sec": 106323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:51.154735+00:00", "epoch": 1, "step": 2581, "train_loss": 3.5666847229003906, "perplexity": 35.39904077072372, "lr": 0.0011436675824175824, "grad_norm": 0.226985, "tokens_per_sec": 106499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:51.460976+00:00", "epoch": 1, "step": 2582, "train_loss": 3.6281752586364746, "perplexity": 37.644063228418744, "lr": 0.0011415164835164836, "grad_norm": 0.221337, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:51.768039+00:00", "epoch": 1, "step": 2583, "train_loss": 3.662869453430176, "perplexity": 38.973013825614586, "lr": 0.0011393653846153845, "grad_norm": 0.200572, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:52.075851+00:00", "epoch": 1, "step": 2584, "train_loss": 3.587371826171875, "perplexity": 36.13897150665664, "lr": 0.0011372142857142857, "grad_norm": 0.221409, "tokens_per_sec": 106454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:52.384570+00:00", "epoch": 1, "step": 2585, "train_loss": 3.518190622329712, "perplexity": 33.72335494257855, "lr": 0.0011350631868131868, "grad_norm": 0.212905, "tokens_per_sec": 106142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:52.691255+00:00", "epoch": 1, "step": 2586, "train_loss": 3.677150249481201, "perplexity": 39.533572573411334, "lr": 0.0011329120879120878, "grad_norm": 0.206743, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:52.998230+00:00", "epoch": 1, "step": 2587, "train_loss": 3.7839291095733643, "perplexity": 43.98853842189943, "lr": 0.001130760989010989, "grad_norm": 0.211551, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:53.304539+00:00", "epoch": 1, "step": 2588, "train_loss": 3.6175057888031006, "perplexity": 37.24455608600317, "lr": 0.0011286098901098901, "grad_norm": 0.230698, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:53.611927+00:00", "epoch": 1, "step": 2589, "train_loss": 3.708771228790283, "perplexity": 40.803637375239646, "lr": 0.001126458791208791, "grad_norm": 0.201497, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:53.919348+00:00", "epoch": 1, "step": 2590, "train_loss": 3.6591591835021973, "perplexity": 38.828681346218296, "lr": 0.0011243076923076922, "grad_norm": 0.187694, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:54.226711+00:00", "epoch": 1, "step": 2591, "train_loss": 3.7221312522888184, "perplexity": 41.35243272592946, "lr": 0.0011221565934065934, "grad_norm": 0.226473, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:54.534365+00:00", "epoch": 1, "step": 2592, "train_loss": 3.6449992656707764, "perplexity": 38.28274473940943, "lr": 0.0011200054945054946, "grad_norm": 0.208744, "tokens_per_sec": 106509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:54.840981+00:00", "epoch": 1, "step": 2593, "train_loss": 3.693272113800049, "perplexity": 40.17609286230383, "lr": 0.0011178543956043955, "grad_norm": 0.198205, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:55.146898+00:00", "epoch": 1, "step": 2594, "train_loss": 3.632826805114746, "perplexity": 37.81957422046606, "lr": 0.0011157032967032967, "grad_norm": 0.215448, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:55.453865+00:00", "epoch": 1, "step": 2595, "train_loss": 3.6042957305908203, "perplexity": 36.75578876211739, "lr": 0.0011135521978021977, "grad_norm": 0.206638, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:55.761376+00:00", "epoch": 1, "step": 2596, "train_loss": 3.5804789066314697, "perplexity": 35.890725038151814, "lr": 0.0011114010989010988, "grad_norm": 0.218666, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:56.070027+00:00", "epoch": 1, "step": 2597, "train_loss": 3.5981194972991943, "perplexity": 36.52947603541063, "lr": 0.00110925, "grad_norm": 0.218934, "tokens_per_sec": 106165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:56.377133+00:00", "epoch": 1, "step": 2598, "train_loss": 3.690880060195923, "perplexity": 40.08010434518191, "lr": 0.0011070989010989012, "grad_norm": 0.204731, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:56.684356+00:00", "epoch": 1, "step": 2599, "train_loss": 3.724196434020996, "perplexity": 41.43792125887055, "lr": 0.0011049478021978021, "grad_norm": 0.195888, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T12:59:56.991858+00:00", "epoch": 1, "step": 2600, "train_loss": 3.6217055320739746, "perplexity": 37.40130257692318, "lr": 0.0011027967032967033, "grad_norm": 0.206634, "tokens_per_sec": 106619} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T13:00:02.717737+00:00", "step": 2600, "epoch": 1, "val_loss": 3.6382956504821777, "val_ppl": 38.02697021514791, "eval_train_loss": 3.6217055320739746, "eval_train_ppl": 37.40130257692318} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T13:00:03.746670+00:00", "step": 2600, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p6383_epoch_0001_step_0002600.pt", "val_loss": 3.6382956504821777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:04.813573+00:00", "epoch": 1, "step": 2601, "train_loss": 3.573699712753296, "perplexity": 35.648237717643916, "lr": 0.0011006456043956044, "grad_norm": 0.220821, "tokens_per_sec": 4189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:05.119982+00:00", "epoch": 1, "step": 2602, "train_loss": 3.5956568717956543, "perplexity": 36.43962829217382, "lr": 0.0010984945054945054, "grad_norm": 0.196626, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:05.425616+00:00", "epoch": 1, "step": 2603, "train_loss": 3.725555896759033, "perplexity": 41.494292877639424, "lr": 0.0010963434065934066, "grad_norm": 0.223035, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:05.731226+00:00", "epoch": 1, "step": 2604, "train_loss": 3.745948553085327, "perplexity": 42.34915859801914, "lr": 0.0010941923076923077, "grad_norm": 0.218851, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:06.037291+00:00", "epoch": 1, "step": 2605, "train_loss": 3.683168888092041, "perplexity": 39.772228330681855, "lr": 0.0010920412087912087, "grad_norm": 0.207078, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:06.344673+00:00", "epoch": 1, "step": 2606, "train_loss": 3.647871971130371, "perplexity": 38.39287790356268, "lr": 0.0010898901098901099, "grad_norm": 0.229093, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:06.651651+00:00", "epoch": 1, "step": 2607, "train_loss": 3.63849139213562, "perplexity": 38.03441440571854, "lr": 0.001087739010989011, "grad_norm": 0.244439, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:06.958785+00:00", "epoch": 1, "step": 2608, "train_loss": 3.5905561447143555, "perplexity": 36.25423292086593, "lr": 0.001085587912087912, "grad_norm": 0.220022, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:07.265045+00:00", "epoch": 1, "step": 2609, "train_loss": 3.723903179168701, "perplexity": 41.42577116901585, "lr": 0.0010834368131868131, "grad_norm": 0.209749, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:07.571339+00:00", "epoch": 1, "step": 2610, "train_loss": 3.5967118740081787, "perplexity": 36.47809246697274, "lr": 0.0010812857142857143, "grad_norm": 0.222317, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:07.878633+00:00", "epoch": 1, "step": 2611, "train_loss": 3.6328985691070557, "perplexity": 37.822288401488635, "lr": 0.0010791346153846153, "grad_norm": 0.21227, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:08.186023+00:00", "epoch": 1, "step": 2612, "train_loss": 3.6864101886749268, "perplexity": 39.90135122756583, "lr": 0.0010769835164835164, "grad_norm": 0.232186, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:08.494104+00:00", "epoch": 1, "step": 2613, "train_loss": 3.6156156063079834, "perplexity": 37.17422356963691, "lr": 0.0010748324175824176, "grad_norm": 0.247589, "tokens_per_sec": 106361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:08.800742+00:00", "epoch": 1, "step": 2614, "train_loss": 3.687331199645996, "perplexity": 39.93811773838786, "lr": 0.0010726813186813186, "grad_norm": 0.23102, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:09.120196+00:00", "epoch": 1, "step": 2615, "train_loss": 3.5380825996398926, "perplexity": 34.400895624621434, "lr": 0.0010705302197802197, "grad_norm": 0.224304, "tokens_per_sec": 102574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:09.428248+00:00", "epoch": 1, "step": 2616, "train_loss": 3.7043330669403076, "perplexity": 40.62294549500015, "lr": 0.001068379120879121, "grad_norm": 0.237218, "tokens_per_sec": 106372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:09.734034+00:00", "epoch": 1, "step": 2617, "train_loss": 3.6369690895080566, "perplexity": 37.976558564953045, "lr": 0.0010662280219780218, "grad_norm": 0.280215, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:10.040800+00:00", "epoch": 1, "step": 2618, "train_loss": 3.6923418045043945, "perplexity": 40.13873404997059, "lr": 0.001064076923076923, "grad_norm": 0.285876, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:10.348131+00:00", "epoch": 1, "step": 2619, "train_loss": 3.608185291290283, "perplexity": 36.89903102771883, "lr": 0.0010619258241758242, "grad_norm": 0.238061, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:10.655074+00:00", "epoch": 1, "step": 2620, "train_loss": 3.6822166442871094, "perplexity": 39.73437349902056, "lr": 0.0010597747252747251, "grad_norm": 0.251007, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:10.960453+00:00", "epoch": 1, "step": 2621, "train_loss": 3.5536582469940186, "perplexity": 34.940906435303944, "lr": 0.0010576236263736263, "grad_norm": 0.204769, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:11.266512+00:00", "epoch": 1, "step": 2622, "train_loss": 3.604619026184082, "perplexity": 36.76767366771633, "lr": 0.0010554725274725275, "grad_norm": 0.239026, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:11.572751+00:00", "epoch": 1, "step": 2623, "train_loss": 3.685724973678589, "perplexity": 39.87401958842446, "lr": 0.0010533214285714286, "grad_norm": 0.206396, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:11.879468+00:00", "epoch": 1, "step": 2624, "train_loss": 3.691304922103882, "perplexity": 40.09713647268009, "lr": 0.0010511703296703296, "grad_norm": 0.212966, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:12.185660+00:00", "epoch": 1, "step": 2625, "train_loss": 3.655771017074585, "perplexity": 38.69734593041647, "lr": 0.0010490192307692308, "grad_norm": 0.222646, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:12.492631+00:00", "epoch": 1, "step": 2626, "train_loss": 3.648710012435913, "perplexity": 38.425066206764456, "lr": 0.0010468681318681317, "grad_norm": 0.226309, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:12.799441+00:00", "epoch": 1, "step": 2627, "train_loss": 3.6402177810668945, "perplexity": 38.100133309627736, "lr": 0.0010447170329670329, "grad_norm": 0.213459, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:13.106417+00:00", "epoch": 1, "step": 2628, "train_loss": 3.712275505065918, "perplexity": 40.9468754199145, "lr": 0.001042565934065934, "grad_norm": 0.198948, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:13.413043+00:00", "epoch": 1, "step": 2629, "train_loss": 3.630772113800049, "perplexity": 37.741946447655465, "lr": 0.0010404148351648352, "grad_norm": 0.208027, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:13.720142+00:00", "epoch": 1, "step": 2630, "train_loss": 3.7362399101257324, "perplexity": 41.93999516230342, "lr": 0.0010382637362637362, "grad_norm": 0.220128, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:14.027104+00:00", "epoch": 1, "step": 2631, "train_loss": 3.615745782852173, "perplexity": 37.17906309658374, "lr": 0.0010361126373626373, "grad_norm": 0.210575, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:14.333275+00:00", "epoch": 1, "step": 2632, "train_loss": 3.7069270610809326, "perplexity": 40.72845796779281, "lr": 0.0010339615384615383, "grad_norm": 0.209395, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:14.639939+00:00", "epoch": 1, "step": 2633, "train_loss": 3.5732550621032715, "perplexity": 35.632390229129676, "lr": 0.0010318104395604395, "grad_norm": 0.2276, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:14.946812+00:00", "epoch": 1, "step": 2634, "train_loss": 3.6105234622955322, "perplexity": 36.98540821515489, "lr": 0.0010296593406593406, "grad_norm": 0.198377, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:15.254011+00:00", "epoch": 1, "step": 2635, "train_loss": 3.5704853534698486, "perplexity": 35.533835437314934, "lr": 0.0010275082417582418, "grad_norm": 0.186822, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:15.560370+00:00", "epoch": 1, "step": 2636, "train_loss": 3.632554054260254, "perplexity": 37.80926030590934, "lr": 0.001025357142857143, "grad_norm": 0.219389, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:15.866902+00:00", "epoch": 1, "step": 2637, "train_loss": 3.6373677253723145, "perplexity": 37.99170040103706, "lr": 0.001023206043956044, "grad_norm": 0.206989, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:16.173638+00:00", "epoch": 1, "step": 2638, "train_loss": 3.6679916381835938, "perplexity": 39.17315294000346, "lr": 0.001021054945054945, "grad_norm": 0.194915, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:16.481997+00:00", "epoch": 1, "step": 2639, "train_loss": 3.6596620082855225, "perplexity": 38.84821027890722, "lr": 0.001018903846153846, "grad_norm": 0.213841, "tokens_per_sec": 106265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:16.790138+00:00", "epoch": 1, "step": 2640, "train_loss": 3.6199140548706055, "perplexity": 37.33435897785015, "lr": 0.0010167527472527472, "grad_norm": 0.216358, "tokens_per_sec": 106406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:17.096807+00:00", "epoch": 1, "step": 2641, "train_loss": 3.583540439605713, "perplexity": 36.000774049707154, "lr": 0.0010146016483516484, "grad_norm": 0.186966, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:17.402966+00:00", "epoch": 1, "step": 2642, "train_loss": 3.591608762741089, "perplexity": 36.29241487196336, "lr": 0.0010124505494505495, "grad_norm": 0.210939, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:17.709911+00:00", "epoch": 1, "step": 2643, "train_loss": 3.648068428039551, "perplexity": 38.40042119063129, "lr": 0.0010102994505494505, "grad_norm": 0.196628, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:18.018519+00:00", "epoch": 1, "step": 2644, "train_loss": 3.6198747158050537, "perplexity": 37.33289030794323, "lr": 0.0010081483516483517, "grad_norm": 0.198964, "tokens_per_sec": 106180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:18.326346+00:00", "epoch": 1, "step": 2645, "train_loss": 3.611081123352051, "perplexity": 37.00603928901449, "lr": 0.0010059972527472526, "grad_norm": 0.217368, "tokens_per_sec": 106448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:18.633193+00:00", "epoch": 1, "step": 2646, "train_loss": 3.711198329925537, "perplexity": 40.90279221056701, "lr": 0.0010038461538461538, "grad_norm": 0.207891, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:18.940531+00:00", "epoch": 1, "step": 2647, "train_loss": 3.681112766265869, "perplexity": 39.69053579761631, "lr": 0.001001695054945055, "grad_norm": 0.204111, "tokens_per_sec": 106619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:19.247279+00:00", "epoch": 1, "step": 2648, "train_loss": 3.62746000289917, "perplexity": 37.61714772310144, "lr": 0.0009995439560439561, "grad_norm": 0.196053, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:19.555267+00:00", "epoch": 1, "step": 2649, "train_loss": 3.5857937335968018, "perplexity": 36.081985840206066, "lr": 0.000997392857142857, "grad_norm": 0.197158, "tokens_per_sec": 106393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:19.863230+00:00", "epoch": 1, "step": 2650, "train_loss": 3.5362305641174316, "perplexity": 34.33724290566273, "lr": 0.0009952417582417582, "grad_norm": 0.197951, "tokens_per_sec": 106456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:20.170330+00:00", "epoch": 1, "step": 2651, "train_loss": 3.656968116760254, "perplexity": 38.74369824970399, "lr": 0.0009930906593406592, "grad_norm": 0.19336, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:20.476819+00:00", "epoch": 1, "step": 2652, "train_loss": 3.618161678314209, "perplexity": 37.26899241257499, "lr": 0.0009909395604395604, "grad_norm": 0.203213, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:20.783472+00:00", "epoch": 1, "step": 2653, "train_loss": 3.816650867462158, "perplexity": 45.45172926171383, "lr": 0.0009887884615384615, "grad_norm": 0.204591, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:21.090747+00:00", "epoch": 1, "step": 2654, "train_loss": 3.5791492462158203, "perplexity": 35.84303427506599, "lr": 0.0009866373626373627, "grad_norm": 0.192854, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:21.398792+00:00", "epoch": 1, "step": 2655, "train_loss": 3.6352016925811768, "perplexity": 37.90949819065792, "lr": 0.0009844862637362639, "grad_norm": 0.195257, "tokens_per_sec": 106375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:21.705626+00:00", "epoch": 1, "step": 2656, "train_loss": 3.746347188949585, "perplexity": 42.366043856768705, "lr": 0.0009823351648351648, "grad_norm": 0.205607, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:22.011821+00:00", "epoch": 1, "step": 2657, "train_loss": 3.5825064182281494, "perplexity": 35.96356771911661, "lr": 0.0009801840659340658, "grad_norm": 0.195437, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:22.318479+00:00", "epoch": 1, "step": 2658, "train_loss": 3.6474926471710205, "perplexity": 38.37831732686782, "lr": 0.000978032967032967, "grad_norm": 0.205884, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:22.626496+00:00", "epoch": 1, "step": 2659, "train_loss": 3.6428840160369873, "perplexity": 38.201852761143776, "lr": 0.0009758818681318681, "grad_norm": 0.208035, "tokens_per_sec": 106384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:22.932077+00:00", "epoch": 1, "step": 2660, "train_loss": 3.715008497238159, "perplexity": 41.05893597046796, "lr": 0.0009737307692307693, "grad_norm": 0.217398, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:23.238134+00:00", "epoch": 1, "step": 2661, "train_loss": 3.6913070678710938, "perplexity": 40.097222511893136, "lr": 0.0009715796703296702, "grad_norm": 0.206138, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:23.544826+00:00", "epoch": 1, "step": 2662, "train_loss": 3.6353330612182617, "perplexity": 37.914478636897854, "lr": 0.0009694285714285714, "grad_norm": 0.202148, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:23.851480+00:00", "epoch": 1, "step": 2663, "train_loss": 3.5769095420837402, "perplexity": 35.762846315266685, "lr": 0.0009672774725274726, "grad_norm": 0.190388, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:24.157696+00:00", "epoch": 1, "step": 2664, "train_loss": 3.7240355014801025, "perplexity": 41.431253085490454, "lr": 0.0009651263736263735, "grad_norm": 0.197862, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:24.464556+00:00", "epoch": 1, "step": 2665, "train_loss": 3.5946896076202393, "perplexity": 36.40439858612439, "lr": 0.0009629752747252747, "grad_norm": 0.199981, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:24.769926+00:00", "epoch": 1, "step": 2666, "train_loss": 3.6137328147888184, "perplexity": 37.10429810496097, "lr": 0.0009608241758241758, "grad_norm": 0.204251, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:25.075933+00:00", "epoch": 1, "step": 2667, "train_loss": 3.7792422771453857, "perplexity": 43.7828538945329, "lr": 0.0009586730769230768, "grad_norm": 0.205292, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:25.382249+00:00", "epoch": 1, "step": 2668, "train_loss": 3.5742030143737793, "perplexity": 35.666184049284645, "lr": 0.000956521978021978, "grad_norm": 0.208326, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:25.688893+00:00", "epoch": 1, "step": 2669, "train_loss": 3.6263206005096436, "perplexity": 37.57431106382888, "lr": 0.0009543708791208791, "grad_norm": 0.221031, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:25.995643+00:00", "epoch": 1, "step": 2670, "train_loss": 3.6617329120635986, "perplexity": 38.928744544921294, "lr": 0.0009522197802197802, "grad_norm": 0.210007, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:26.302183+00:00", "epoch": 1, "step": 2671, "train_loss": 3.5219943523406982, "perplexity": 33.851873750313075, "lr": 0.0009500686813186813, "grad_norm": 0.21122, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:26.608682+00:00", "epoch": 1, "step": 2672, "train_loss": 3.625683546066284, "perplexity": 37.550381804939136, "lr": 0.0009479175824175824, "grad_norm": 0.20821, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:26.915635+00:00", "epoch": 1, "step": 2673, "train_loss": 3.753473997116089, "perplexity": 42.66905699984364, "lr": 0.0009457664835164834, "grad_norm": 0.183354, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:27.221408+00:00", "epoch": 1, "step": 2674, "train_loss": 3.628319025039673, "perplexity": 37.64947556903791, "lr": 0.0009436153846153845, "grad_norm": 0.201711, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:27.529596+00:00", "epoch": 1, "step": 2675, "train_loss": 3.556408405303955, "perplexity": 35.03713171622274, "lr": 0.0009414642857142857, "grad_norm": 0.193064, "tokens_per_sec": 106325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:27.836004+00:00", "epoch": 1, "step": 2676, "train_loss": 3.6243910789489746, "perplexity": 37.50188052112274, "lr": 0.0009393131868131868, "grad_norm": 0.206868, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:28.142603+00:00", "epoch": 1, "step": 2677, "train_loss": 3.5589218139648438, "perplexity": 35.12530510799668, "lr": 0.0009371620879120878, "grad_norm": 0.204777, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:28.448801+00:00", "epoch": 1, "step": 2678, "train_loss": 3.6851603984832764, "perplexity": 39.85151405965687, "lr": 0.000935010989010989, "grad_norm": 0.21816, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:28.755173+00:00", "epoch": 1, "step": 2679, "train_loss": 3.646562337875366, "perplexity": 38.34263022410264, "lr": 0.0009328598901098902, "grad_norm": 0.217265, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:29.061359+00:00", "epoch": 1, "step": 2680, "train_loss": 3.61146879196167, "perplexity": 37.020388149934945, "lr": 0.0009307087912087911, "grad_norm": 0.223296, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:29.369082+00:00", "epoch": 1, "step": 2681, "train_loss": 3.636132001876831, "perplexity": 37.944782159175446, "lr": 0.0009285576923076923, "grad_norm": 0.21471, "tokens_per_sec": 106433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:29.677019+00:00", "epoch": 1, "step": 2682, "train_loss": 3.689875602722168, "perplexity": 40.03986579716184, "lr": 0.0009264065934065933, "grad_norm": 0.190854, "tokens_per_sec": 106412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:29.984937+00:00", "epoch": 1, "step": 2683, "train_loss": 3.6784253120422363, "perplexity": 39.58401250190215, "lr": 0.0009242554945054944, "grad_norm": 0.206544, "tokens_per_sec": 106417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:30.291812+00:00", "epoch": 1, "step": 2684, "train_loss": 3.674781322479248, "perplexity": 39.44003126586783, "lr": 0.0009221043956043956, "grad_norm": 0.179357, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:30.598136+00:00", "epoch": 1, "step": 2685, "train_loss": 3.690945625305176, "perplexity": 40.0827322877519, "lr": 0.0009199532967032966, "grad_norm": 0.193174, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:30.904153+00:00", "epoch": 1, "step": 2686, "train_loss": 3.7112808227539062, "perplexity": 40.906166536761596, "lr": 0.0009178021978021977, "grad_norm": 0.198464, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:31.211272+00:00", "epoch": 1, "step": 2687, "train_loss": 3.5914361476898193, "perplexity": 36.28615079556203, "lr": 0.0009156510989010989, "grad_norm": 0.193928, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:31.518468+00:00", "epoch": 1, "step": 2688, "train_loss": 3.6000006198883057, "perplexity": 36.59825713050256, "lr": 0.0009135, "grad_norm": 0.195691, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:31.824760+00:00", "epoch": 1, "step": 2689, "train_loss": 3.58917498588562, "perplexity": 36.20419463035603, "lr": 0.0009113489010989011, "grad_norm": 0.214714, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:32.131189+00:00", "epoch": 1, "step": 2690, "train_loss": 3.6298999786376953, "perplexity": 37.70904471852572, "lr": 0.0009091978021978022, "grad_norm": 0.214237, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:32.437597+00:00", "epoch": 1, "step": 2691, "train_loss": 3.604217529296875, "perplexity": 36.75291452426224, "lr": 0.0009070467032967032, "grad_norm": 0.210406, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:32.744691+00:00", "epoch": 1, "step": 2692, "train_loss": 3.5169425010681152, "perplexity": 33.68129036257139, "lr": 0.0009048956043956043, "grad_norm": 0.220538, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:33.052731+00:00", "epoch": 1, "step": 2693, "train_loss": 3.62980055809021, "perplexity": 37.70529585101497, "lr": 0.0009027445054945054, "grad_norm": 0.208839, "tokens_per_sec": 106376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:33.359626+00:00", "epoch": 1, "step": 2694, "train_loss": 3.66988205909729, "perplexity": 39.24727672808163, "lr": 0.0009005934065934066, "grad_norm": 0.236379, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:33.666462+00:00", "epoch": 1, "step": 2695, "train_loss": 3.739898204803467, "perplexity": 42.0937050099222, "lr": 0.0008984423076923077, "grad_norm": 0.217683, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:33.972656+00:00", "epoch": 1, "step": 2696, "train_loss": 3.64797043800354, "perplexity": 38.3966585163313, "lr": 0.0008962912087912087, "grad_norm": 0.206241, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:34.279436+00:00", "epoch": 1, "step": 2697, "train_loss": 3.6609690189361572, "perplexity": 38.89901849970984, "lr": 0.0008941401098901099, "grad_norm": 0.239359, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:34.585627+00:00", "epoch": 1, "step": 2698, "train_loss": 3.620213270187378, "perplexity": 37.34553166133386, "lr": 0.0008919890109890108, "grad_norm": 0.199215, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:34.892908+00:00", "epoch": 1, "step": 2699, "train_loss": 3.599102020263672, "perplexity": 36.565384722161724, "lr": 0.000889837912087912, "grad_norm": 0.220878, "tokens_per_sec": 106637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:35.199615+00:00", "epoch": 1, "step": 2700, "train_loss": 3.6080434322357178, "perplexity": 36.89379693732319, "lr": 0.0008876868131868132, "grad_norm": 0.219367, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T13:00:40.887731+00:00", "step": 2700, "epoch": 1, "val_loss": 3.612773048877716, "val_ppl": 37.06870374834181, "eval_train_loss": 3.6080434322357178, "eval_train_ppl": 36.89379693732319} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T13:00:41.816490+00:00", "step": 2700, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p6128_epoch_0001_step_0002700.pt", "val_loss": 3.612773048877716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:42.868851+00:00", "epoch": 1, "step": 2701, "train_loss": 3.537814140319824, "perplexity": 34.391661623104646, "lr": 0.0008855357142857142, "grad_norm": 0.200237, "tokens_per_sec": 4273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:43.173327+00:00", "epoch": 1, "step": 2702, "train_loss": 3.741192102432251, "perplexity": 42.14820520625513, "lr": 0.0008833846153846153, "grad_norm": 0.216485, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:43.478518+00:00", "epoch": 1, "step": 2703, "train_loss": 3.6278235912323, "perplexity": 37.63082736586774, "lr": 0.0008812335164835165, "grad_norm": 0.20724, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:43.784300+00:00", "epoch": 1, "step": 2704, "train_loss": 3.702599048614502, "perplexity": 40.55256560071182, "lr": 0.0008790824175824175, "grad_norm": 0.197591, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:44.090536+00:00", "epoch": 1, "step": 2705, "train_loss": 3.5830183029174805, "perplexity": 35.98198163130353, "lr": 0.0008769313186813186, "grad_norm": 0.214449, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:44.397217+00:00", "epoch": 1, "step": 2706, "train_loss": 3.7254409790039062, "perplexity": 41.48952472062954, "lr": 0.0008747802197802198, "grad_norm": 0.193988, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:44.702237+00:00", "epoch": 1, "step": 2707, "train_loss": 3.638594627380371, "perplexity": 38.03834110048184, "lr": 0.0008726291208791208, "grad_norm": 0.200163, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:45.007748+00:00", "epoch": 1, "step": 2708, "train_loss": 3.635991334915161, "perplexity": 37.939444957350666, "lr": 0.0008704780219780219, "grad_norm": 0.208798, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:45.314413+00:00", "epoch": 1, "step": 2709, "train_loss": 3.5640110969543457, "perplexity": 35.304523385224144, "lr": 0.000868326923076923, "grad_norm": 0.197573, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:45.620931+00:00", "epoch": 1, "step": 2710, "train_loss": 3.6462814807891846, "perplexity": 38.33186293680661, "lr": 0.0008661758241758241, "grad_norm": 0.19864, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:45.927533+00:00", "epoch": 1, "step": 2711, "train_loss": 3.6875720024108887, "perplexity": 39.9477361055821, "lr": 0.0008640247252747252, "grad_norm": 0.201304, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:46.233735+00:00", "epoch": 1, "step": 2712, "train_loss": 3.7559080123901367, "perplexity": 42.773040633850215, "lr": 0.0008618736263736263, "grad_norm": 0.215354, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:46.539131+00:00", "epoch": 1, "step": 2713, "train_loss": 3.6175479888916016, "perplexity": 37.24612784273008, "lr": 0.0008597225274725275, "grad_norm": 0.205475, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:46.845405+00:00", "epoch": 1, "step": 2714, "train_loss": 3.6478493213653564, "perplexity": 38.39200832374785, "lr": 0.0008575714285714286, "grad_norm": 0.19088, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:47.152790+00:00", "epoch": 1, "step": 2715, "train_loss": 3.5364298820495605, "perplexity": 34.344087616027785, "lr": 0.0008554203296703296, "grad_norm": 0.209785, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:47.458972+00:00", "epoch": 1, "step": 2716, "train_loss": 3.6462888717651367, "perplexity": 38.33214624773075, "lr": 0.0008532692307692307, "grad_norm": 0.193594, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:47.764423+00:00", "epoch": 1, "step": 2717, "train_loss": 3.5894622802734375, "perplexity": 36.21459738654396, "lr": 0.0008511181318681317, "grad_norm": 0.191872, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:48.069509+00:00", "epoch": 1, "step": 2718, "train_loss": 3.6351478099823, "perplexity": 37.90745558340427, "lr": 0.0008489670329670329, "grad_norm": 0.188118, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:48.374871+00:00", "epoch": 1, "step": 2719, "train_loss": 3.5925559997558594, "perplexity": 36.32680867765673, "lr": 0.0008468159340659341, "grad_norm": 0.189801, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:48.681259+00:00", "epoch": 1, "step": 2720, "train_loss": 3.5876989364624023, "perplexity": 36.15079486979199, "lr": 0.0008446648351648351, "grad_norm": 0.1982, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:48.987840+00:00", "epoch": 1, "step": 2721, "train_loss": 3.6370413303375244, "perplexity": 37.979302122141334, "lr": 0.0008425137362637362, "grad_norm": 0.209567, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:49.293492+00:00", "epoch": 1, "step": 2722, "train_loss": 3.698338031768799, "perplexity": 40.38013805477625, "lr": 0.0008403626373626374, "grad_norm": 0.210675, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:49.599756+00:00", "epoch": 1, "step": 2723, "train_loss": 3.5819551944732666, "perplexity": 35.9437492089981, "lr": 0.0008382115384615384, "grad_norm": 0.206997, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:49.905653+00:00", "epoch": 1, "step": 2724, "train_loss": 3.6229758262634277, "perplexity": 37.448843423302236, "lr": 0.0008360604395604395, "grad_norm": 0.189092, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:50.211560+00:00", "epoch": 1, "step": 2725, "train_loss": 3.672811269760132, "perplexity": 39.3624088103004, "lr": 0.0008339093406593407, "grad_norm": 0.203178, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:50.519567+00:00", "epoch": 1, "step": 2726, "train_loss": 3.6503889560699463, "perplexity": 38.48963391466344, "lr": 0.0008317582417582417, "grad_norm": 0.187412, "tokens_per_sec": 106388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:50.827009+00:00", "epoch": 1, "step": 2727, "train_loss": 3.6122281551361084, "perplexity": 37.04851074567985, "lr": 0.0008296071428571428, "grad_norm": 0.199985, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:51.132557+00:00", "epoch": 1, "step": 2728, "train_loss": 3.5565619468688965, "perplexity": 35.042511785279046, "lr": 0.000827456043956044, "grad_norm": 0.201828, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:51.438850+00:00", "epoch": 1, "step": 2729, "train_loss": 3.6896963119506836, "perplexity": 40.032687662238835, "lr": 0.000825304945054945, "grad_norm": 0.19245, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:51.744779+00:00", "epoch": 1, "step": 2730, "train_loss": 3.651754856109619, "perplexity": 38.542242828233746, "lr": 0.0008231538461538461, "grad_norm": 0.207755, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:52.052046+00:00", "epoch": 1, "step": 2731, "train_loss": 3.5638387203216553, "perplexity": 35.29843823484819, "lr": 0.0008210027472527472, "grad_norm": 0.196577, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:52.359002+00:00", "epoch": 1, "step": 2732, "train_loss": 3.5682170391082764, "perplexity": 35.45332487422456, "lr": 0.0008188516483516484, "grad_norm": 0.198765, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:52.665206+00:00", "epoch": 1, "step": 2733, "train_loss": 3.5119705200195312, "perplexity": 33.51424324686878, "lr": 0.0008167005494505494, "grad_norm": 0.193789, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:52.970529+00:00", "epoch": 1, "step": 2734, "train_loss": 3.6009910106658936, "perplexity": 36.634521661904884, "lr": 0.0008145494505494505, "grad_norm": 0.20231, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:53.276382+00:00", "epoch": 1, "step": 2735, "train_loss": 3.4703550338745117, "perplexity": 32.14815410258018, "lr": 0.0008123983516483516, "grad_norm": 0.191631, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:53.582335+00:00", "epoch": 1, "step": 2736, "train_loss": 3.702143430709839, "perplexity": 40.53409333421134, "lr": 0.0008102472527472528, "grad_norm": 0.180367, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:53.889548+00:00", "epoch": 1, "step": 2737, "train_loss": 3.62837290763855, "perplexity": 37.651504275283415, "lr": 0.0008080961538461538, "grad_norm": 0.198281, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:54.195704+00:00", "epoch": 1, "step": 2738, "train_loss": 3.621910810470581, "perplexity": 37.408981044431954, "lr": 0.0008059450549450549, "grad_norm": 0.19721, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:54.501577+00:00", "epoch": 1, "step": 2739, "train_loss": 3.646390438079834, "perplexity": 38.33603970027803, "lr": 0.000803793956043956, "grad_norm": 0.194215, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:54.807347+00:00", "epoch": 1, "step": 2740, "train_loss": 3.5750226974487305, "perplexity": 35.69543100167633, "lr": 0.0008016428571428571, "grad_norm": 0.18632, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:55.113382+00:00", "epoch": 1, "step": 2741, "train_loss": 3.6102402210235596, "perplexity": 36.974933904536165, "lr": 0.0007994917582417582, "grad_norm": 0.205696, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:55.418616+00:00", "epoch": 1, "step": 2742, "train_loss": 3.5640058517456055, "perplexity": 35.30433820611517, "lr": 0.0007973406593406593, "grad_norm": 0.196594, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:55.725776+00:00", "epoch": 1, "step": 2743, "train_loss": 3.6876866817474365, "perplexity": 39.95231754814865, "lr": 0.0007951895604395604, "grad_norm": 0.192099, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:56.032046+00:00", "epoch": 1, "step": 2744, "train_loss": 3.5848419666290283, "perplexity": 36.04766053538849, "lr": 0.0007930384615384615, "grad_norm": 0.206839, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:56.337761+00:00", "epoch": 1, "step": 2745, "train_loss": 3.6311073303222656, "perplexity": 37.75460029245578, "lr": 0.0007908873626373626, "grad_norm": 0.194037, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:56.644045+00:00", "epoch": 1, "step": 2746, "train_loss": 3.506770372390747, "perplexity": 33.34041658817307, "lr": 0.0007887362637362637, "grad_norm": 0.195678, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:56.949898+00:00", "epoch": 1, "step": 2747, "train_loss": 3.5920865535736084, "perplexity": 36.309759198229266, "lr": 0.0007865851648351649, "grad_norm": 0.193719, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:57.257469+00:00", "epoch": 1, "step": 2748, "train_loss": 3.6632723808288574, "perplexity": 38.98872028476299, "lr": 0.0007844340659340659, "grad_norm": 0.187454, "tokens_per_sec": 106539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:57.565585+00:00", "epoch": 1, "step": 2749, "train_loss": 3.7459418773651123, "perplexity": 42.348875887828655, "lr": 0.0007822829670329671, "grad_norm": 0.18871, "tokens_per_sec": 106350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:57.871460+00:00", "epoch": 1, "step": 2750, "train_loss": 3.6091580390930176, "perplexity": 36.93494194237404, "lr": 0.000780131868131868, "grad_norm": 0.183084, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T13:00:58.860561+00:00", "step": 2750, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0001_step_0002750.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:00:59.917606+00:00", "epoch": 1, "step": 2751, "train_loss": 3.6561779975891113, "perplexity": 38.71309820139028, "lr": 0.0007779807692307692, "grad_norm": 0.18862, "tokens_per_sec": 16013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:00.224293+00:00", "epoch": 1, "step": 2752, "train_loss": 3.4391326904296875, "perplexity": 31.159921147449452, "lr": 0.0007758296703296703, "grad_norm": 0.197278, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:00.529821+00:00", "epoch": 1, "step": 2753, "train_loss": 3.5269510746002197, "perplexity": 34.02008462936305, "lr": 0.0007736785714285714, "grad_norm": 0.198631, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:00.835049+00:00", "epoch": 1, "step": 2754, "train_loss": 3.5746426582336426, "perplexity": 35.68186791551179, "lr": 0.0007715274725274725, "grad_norm": 0.186956, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:01.141256+00:00", "epoch": 1, "step": 2755, "train_loss": 3.6420388221740723, "perplexity": 38.1695784305924, "lr": 0.0007693763736263735, "grad_norm": 0.19112, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:01.447642+00:00", "epoch": 1, "step": 2756, "train_loss": 3.5769526958465576, "perplexity": 35.76438964995436, "lr": 0.0007672252747252747, "grad_norm": 0.195745, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:01.754165+00:00", "epoch": 1, "step": 2757, "train_loss": 3.627258062362671, "perplexity": 37.60955206307031, "lr": 0.0007650741758241758, "grad_norm": 0.196619, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:02.059612+00:00", "epoch": 1, "step": 2758, "train_loss": 3.5300965309143066, "perplexity": 34.12726179154038, "lr": 0.0007629230769230768, "grad_norm": 0.197286, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:02.364965+00:00", "epoch": 1, "step": 2759, "train_loss": 3.639820098876953, "perplexity": 38.084984577566225, "lr": 0.000760771978021978, "grad_norm": 0.184423, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:02.670328+00:00", "epoch": 1, "step": 2760, "train_loss": 3.6252474784851074, "perplexity": 37.534010870449514, "lr": 0.0007586208791208791, "grad_norm": 0.203619, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:02.977094+00:00", "epoch": 1, "step": 2761, "train_loss": 3.651721954345703, "perplexity": 38.54097474132068, "lr": 0.0007564697802197801, "grad_norm": 0.187441, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:03.283010+00:00", "epoch": 1, "step": 2762, "train_loss": 3.619237184524536, "perplexity": 37.309097007872204, "lr": 0.0007543186813186813, "grad_norm": 0.190372, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:03.589285+00:00", "epoch": 1, "step": 2763, "train_loss": 3.534846305847168, "perplexity": 34.28974417603181, "lr": 0.0007521675824175824, "grad_norm": 0.179922, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:03.894790+00:00", "epoch": 1, "step": 2764, "train_loss": 3.5997815132141113, "perplexity": 36.59023908653693, "lr": 0.0007500164835164835, "grad_norm": 0.194488, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:04.212245+00:00", "epoch": 1, "step": 2765, "train_loss": 3.6117682456970215, "perplexity": 37.03147570347242, "lr": 0.0007478653846153846, "grad_norm": 0.180961, "tokens_per_sec": 103214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:04.517933+00:00", "epoch": 1, "step": 2766, "train_loss": 3.5033886432647705, "perplexity": 33.22785875764004, "lr": 0.0007457142857142857, "grad_norm": 0.195832, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:04.825922+00:00", "epoch": 1, "step": 2767, "train_loss": 3.718630313873291, "perplexity": 41.20791352963589, "lr": 0.0007435631868131867, "grad_norm": 0.175172, "tokens_per_sec": 106393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:05.134132+00:00", "epoch": 1, "step": 2768, "train_loss": 3.57682204246521, "perplexity": 35.759717216756, "lr": 0.0007414120879120879, "grad_norm": 0.205395, "tokens_per_sec": 106319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:05.440816+00:00", "epoch": 1, "step": 2769, "train_loss": 3.481668710708618, "perplexity": 32.51393318043626, "lr": 0.0007392609890109889, "grad_norm": 0.204054, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:05.747733+00:00", "epoch": 1, "step": 2770, "train_loss": 3.65254807472229, "perplexity": 38.572827381133855, "lr": 0.0007371098901098901, "grad_norm": 0.193547, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:06.054421+00:00", "epoch": 1, "step": 2771, "train_loss": 3.6317288875579834, "perplexity": 37.77807423189191, "lr": 0.0007349587912087912, "grad_norm": 0.18743, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:06.360982+00:00", "epoch": 1, "step": 2772, "train_loss": 3.619988441467285, "perplexity": 37.33713625704862, "lr": 0.0007328076923076923, "grad_norm": 0.189905, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:06.668265+00:00", "epoch": 1, "step": 2773, "train_loss": 3.6340441703796387, "perplexity": 37.865642491722355, "lr": 0.0007306565934065934, "grad_norm": 0.175167, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:06.975318+00:00", "epoch": 1, "step": 2774, "train_loss": 3.5801784992218018, "perplexity": 35.87994481772212, "lr": 0.0007285054945054944, "grad_norm": 0.197352, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:07.282172+00:00", "epoch": 1, "step": 2775, "train_loss": 3.5993621349334717, "perplexity": 36.5748971522425, "lr": 0.0007263543956043955, "grad_norm": 0.185649, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:07.588422+00:00", "epoch": 1, "step": 2776, "train_loss": 3.572897434234619, "perplexity": 35.619649371735676, "lr": 0.0007242032967032967, "grad_norm": 0.197409, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:07.895411+00:00", "epoch": 1, "step": 2777, "train_loss": 3.5886642932891846, "perplexity": 36.1857101365472, "lr": 0.0007220521978021977, "grad_norm": 0.177065, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:08.201939+00:00", "epoch": 1, "step": 2778, "train_loss": 3.6486010551452637, "perplexity": 38.42087974373452, "lr": 0.0007199010989010989, "grad_norm": 0.215309, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:08.509396+00:00", "epoch": 1, "step": 2779, "train_loss": 3.7102296352386475, "perplexity": 40.86318907784267, "lr": 0.00071775, "grad_norm": 0.186933, "tokens_per_sec": 106577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:08.816706+00:00", "epoch": 1, "step": 2780, "train_loss": 3.730412721633911, "perplexity": 41.69631358393099, "lr": 0.000715598901098901, "grad_norm": 0.186706, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:09.123225+00:00", "epoch": 1, "step": 2781, "train_loss": 3.632056474685669, "perplexity": 37.79045186998579, "lr": 0.0007134478021978022, "grad_norm": 0.182098, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:09.429551+00:00", "epoch": 1, "step": 2782, "train_loss": 3.744070053100586, "perplexity": 42.269680377528594, "lr": 0.0007112967032967033, "grad_norm": 0.194707, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:09.736624+00:00", "epoch": 1, "step": 2783, "train_loss": 3.631270170211792, "perplexity": 37.76074874798983, "lr": 0.0007091456043956044, "grad_norm": 0.197503, "tokens_per_sec": 106711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:10.044840+00:00", "epoch": 1, "step": 2784, "train_loss": 3.6130874156951904, "perplexity": 37.08035875064409, "lr": 0.0007069945054945055, "grad_norm": 0.185987, "tokens_per_sec": 106316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:10.351947+00:00", "epoch": 1, "step": 2785, "train_loss": 3.564234495162964, "perplexity": 35.31241123353731, "lr": 0.0007048434065934066, "grad_norm": 0.192521, "tokens_per_sec": 106697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:10.659132+00:00", "epoch": 1, "step": 2786, "train_loss": 3.5923261642456055, "perplexity": 36.318460446445144, "lr": 0.0007026923076923077, "grad_norm": 0.189299, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:10.965911+00:00", "epoch": 1, "step": 2787, "train_loss": 3.4628138542175293, "perplexity": 31.906630926289367, "lr": 0.0007005412087912088, "grad_norm": 0.20696, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:11.272058+00:00", "epoch": 1, "step": 2788, "train_loss": 3.6051642894744873, "perplexity": 36.78772719717082, "lr": 0.0006983901098901098, "grad_norm": 0.189641, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:11.578644+00:00", "epoch": 1, "step": 2789, "train_loss": 3.464447498321533, "perplexity": 31.958797605072498, "lr": 0.0006962390109890109, "grad_norm": 0.186682, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:11.885746+00:00", "epoch": 1, "step": 2790, "train_loss": 3.6102845668792725, "perplexity": 36.976573625977245, "lr": 0.0006940879120879121, "grad_norm": 0.183512, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:12.192890+00:00", "epoch": 1, "step": 2791, "train_loss": 3.5880978107452393, "perplexity": 36.165217368360324, "lr": 0.0006919368131868131, "grad_norm": 0.192381, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:12.499529+00:00", "epoch": 1, "step": 2792, "train_loss": 3.553226947784424, "perplexity": 34.92583969934509, "lr": 0.0006897857142857142, "grad_norm": 0.185417, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:12.805780+00:00", "epoch": 1, "step": 2793, "train_loss": 3.6120247840881348, "perplexity": 37.04097691733089, "lr": 0.0006876346153846153, "grad_norm": 0.204421, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:13.111889+00:00", "epoch": 1, "step": 2794, "train_loss": 3.687250852584839, "perplexity": 39.93490895690924, "lr": 0.0006854835164835164, "grad_norm": 0.191684, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:13.418608+00:00", "epoch": 1, "step": 2795, "train_loss": 3.4731578826904297, "perplexity": 32.23838691363223, "lr": 0.0006833324175824176, "grad_norm": 0.179996, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:13.725403+00:00", "epoch": 1, "step": 2796, "train_loss": 3.5944995880126953, "perplexity": 36.39748169378558, "lr": 0.0006811813186813186, "grad_norm": 0.195106, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:14.032080+00:00", "epoch": 1, "step": 2797, "train_loss": 3.708526849746704, "perplexity": 40.79366703968349, "lr": 0.0006790302197802198, "grad_norm": 0.198823, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:14.338436+00:00", "epoch": 1, "step": 2798, "train_loss": 3.6165518760681152, "perplexity": 37.20904496958972, "lr": 0.0006768791208791209, "grad_norm": 0.183598, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:14.645293+00:00", "epoch": 1, "step": 2799, "train_loss": 3.557039737701416, "perplexity": 35.05925877661939, "lr": 0.0006747280219780219, "grad_norm": 0.19475, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:14.952553+00:00", "epoch": 1, "step": 2800, "train_loss": 3.5484442710876465, "perplexity": 34.759199510417616, "lr": 0.0006725769230769231, "grad_norm": 0.184256, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T13:01:20.668071+00:00", "step": 2800, "epoch": 1, "val_loss": 3.5885300993919373, "val_ppl": 36.18085456088092, "eval_train_loss": 3.5484442710876465, "eval_train_ppl": 34.759199510417616} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T13:01:21.593071+00:00", "step": 2800, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p5885_epoch_0001_step_0002800.pt", "val_loss": 3.5885300993919373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:22.689453+00:00", "epoch": 1, "step": 2801, "train_loss": 3.629166841506958, "perplexity": 37.68140894932003, "lr": 0.0006704258241758242, "grad_norm": 0.202758, "tokens_per_sec": 4235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:22.994831+00:00", "epoch": 1, "step": 2802, "train_loss": 3.5222058296203613, "perplexity": 33.859033409511234, "lr": 0.0006682747252747253, "grad_norm": 0.185578, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:23.301336+00:00", "epoch": 1, "step": 2803, "train_loss": 3.4972901344299316, "perplexity": 33.025835015367676, "lr": 0.0006661236263736264, "grad_norm": 0.196467, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:23.607978+00:00", "epoch": 1, "step": 2804, "train_loss": 3.5575778484344482, "perplexity": 35.07812961690473, "lr": 0.0006639725274725275, "grad_norm": 0.203603, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:23.914068+00:00", "epoch": 1, "step": 2805, "train_loss": 3.5601584911346436, "perplexity": 35.16877064178308, "lr": 0.0006618214285714285, "grad_norm": 0.183774, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:24.219950+00:00", "epoch": 1, "step": 2806, "train_loss": 3.562070846557617, "perplexity": 35.236090179964954, "lr": 0.0006596703296703296, "grad_norm": 0.192531, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:24.526890+00:00", "epoch": 1, "step": 2807, "train_loss": 3.6456265449523926, "perplexity": 38.306766245336696, "lr": 0.0006575192307692307, "grad_norm": 0.191862, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:24.833099+00:00", "epoch": 1, "step": 2808, "train_loss": 3.590602159500122, "perplexity": 36.25590119000915, "lr": 0.0006553681318681318, "grad_norm": 0.191737, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:25.140801+00:00", "epoch": 1, "step": 2809, "train_loss": 3.6131582260131836, "perplexity": 37.08298451560305, "lr": 0.0006532170329670328, "grad_norm": 0.197527, "tokens_per_sec": 106494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:25.446929+00:00", "epoch": 1, "step": 2810, "train_loss": 3.6052560806274414, "perplexity": 36.79110414004917, "lr": 0.000651065934065934, "grad_norm": 0.174138, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:25.752990+00:00", "epoch": 1, "step": 2811, "train_loss": 3.4789345264434814, "perplexity": 32.42515551848653, "lr": 0.0006489148351648351, "grad_norm": 0.18831, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:26.059425+00:00", "epoch": 1, "step": 2812, "train_loss": 3.55485463142395, "perplexity": 34.98273420778958, "lr": 0.0006467637362637362, "grad_norm": 0.192071, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:26.367702+00:00", "epoch": 1, "step": 2813, "train_loss": 3.6247854232788086, "perplexity": 37.5166720913586, "lr": 0.0006446126373626373, "grad_norm": 0.197369, "tokens_per_sec": 106294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:26.673538+00:00", "epoch": 1, "step": 2814, "train_loss": 3.5724403858184814, "perplexity": 35.60337318719246, "lr": 0.0006424615384615385, "grad_norm": 0.197194, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:26.979385+00:00", "epoch": 1, "step": 2815, "train_loss": 3.604336738586426, "perplexity": 36.75729607424714, "lr": 0.0006403104395604395, "grad_norm": 0.196885, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:27.285511+00:00", "epoch": 1, "step": 2816, "train_loss": 3.5515966415405273, "perplexity": 34.8689462742644, "lr": 0.0006381593406593406, "grad_norm": 0.185148, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:27.591487+00:00", "epoch": 1, "step": 2817, "train_loss": 3.6770505905151367, "perplexity": 39.529632894759246, "lr": 0.0006360082417582418, "grad_norm": 0.197396, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:27.897873+00:00", "epoch": 1, "step": 2818, "train_loss": 3.615161418914795, "perplexity": 37.15734333962402, "lr": 0.0006338571428571428, "grad_norm": 0.210274, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:28.204288+00:00", "epoch": 1, "step": 2819, "train_loss": 3.5422980785369873, "perplexity": 34.54621796058718, "lr": 0.000631706043956044, "grad_norm": 0.195421, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:28.510580+00:00", "epoch": 1, "step": 2820, "train_loss": 3.5340182781219482, "perplexity": 34.261363068956705, "lr": 0.000629554945054945, "grad_norm": 0.192878, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:28.816348+00:00", "epoch": 1, "step": 2821, "train_loss": 3.4861021041870117, "perplexity": 32.65840024279672, "lr": 0.0006274038461538462, "grad_norm": 0.198728, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:29.123525+00:00", "epoch": 1, "step": 2822, "train_loss": 3.6970906257629395, "perplexity": 40.32979903117686, "lr": 0.0006252527472527472, "grad_norm": 0.188638, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:29.431132+00:00", "epoch": 1, "step": 2823, "train_loss": 3.55033802986145, "perplexity": 34.82508741767082, "lr": 0.0006231016483516482, "grad_norm": 0.188941, "tokens_per_sec": 106525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:29.737058+00:00", "epoch": 1, "step": 2824, "train_loss": 3.6696999073028564, "perplexity": 39.240128417257566, "lr": 0.0006209505494505494, "grad_norm": 0.191065, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:30.042814+00:00", "epoch": 1, "step": 2825, "train_loss": 3.647692918777466, "perplexity": 38.386004183835446, "lr": 0.0006187994505494505, "grad_norm": 0.18818, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:30.348652+00:00", "epoch": 1, "step": 2826, "train_loss": 3.6805965900421143, "perplexity": 39.67005377335123, "lr": 0.0006166483516483516, "grad_norm": 0.203113, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:30.655687+00:00", "epoch": 1, "step": 2827, "train_loss": 3.559541702270508, "perplexity": 35.14708562391211, "lr": 0.0006144972527472527, "grad_norm": 0.199293, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:30.962591+00:00", "epoch": 1, "step": 2828, "train_loss": 3.6427390575408936, "perplexity": 38.19631547936724, "lr": 0.0006123461538461537, "grad_norm": 0.189507, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:31.268919+00:00", "epoch": 1, "step": 2829, "train_loss": 3.563697576522827, "perplexity": 35.293456430766675, "lr": 0.0006101950549450549, "grad_norm": 0.199377, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:31.575094+00:00", "epoch": 1, "step": 2830, "train_loss": 3.5620954036712646, "perplexity": 35.23695548726067, "lr": 0.000608043956043956, "grad_norm": 0.17371, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:31.880537+00:00", "epoch": 1, "step": 2831, "train_loss": 3.575465202331543, "perplexity": 35.71122989947558, "lr": 0.0006058928571428571, "grad_norm": 0.192468, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:32.187273+00:00", "epoch": 1, "step": 2832, "train_loss": 3.608642339706421, "perplexity": 36.91589952597362, "lr": 0.0006037417582417582, "grad_norm": 0.172038, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:32.494602+00:00", "epoch": 1, "step": 2833, "train_loss": 3.5955162048339844, "perplexity": 36.43450280087968, "lr": 0.0006015906593406594, "grad_norm": 0.193965, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:32.801408+00:00", "epoch": 1, "step": 2834, "train_loss": 3.719068765640259, "perplexity": 41.225985173618625, "lr": 0.0005994395604395604, "grad_norm": 0.181923, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:33.110290+00:00", "epoch": 1, "step": 2835, "train_loss": 3.534951686859131, "perplexity": 34.293357854376225, "lr": 0.0005972884615384615, "grad_norm": 0.19292, "tokens_per_sec": 106088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:33.418955+00:00", "epoch": 1, "step": 2836, "train_loss": 3.7281746864318848, "perplexity": 41.60310011275762, "lr": 0.0005951373626373627, "grad_norm": 0.190209, "tokens_per_sec": 106159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:33.725470+00:00", "epoch": 1, "step": 2837, "train_loss": 3.540336847305298, "perplexity": 34.4785312353419, "lr": 0.0005929862637362637, "grad_norm": 0.185928, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:34.032083+00:00", "epoch": 1, "step": 2838, "train_loss": 3.5931613445281982, "perplexity": 36.348805578568275, "lr": 0.0005908351648351649, "grad_norm": 0.180751, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:34.339162+00:00", "epoch": 1, "step": 2839, "train_loss": 3.633596181869507, "perplexity": 37.84868291808801, "lr": 0.0005886840659340658, "grad_norm": 0.18443, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:34.646377+00:00", "epoch": 1, "step": 2840, "train_loss": 3.471968650817871, "perplexity": 32.200070784291576, "lr": 0.0005865329670329669, "grad_norm": 0.184617, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:34.952851+00:00", "epoch": 1, "step": 2841, "train_loss": 3.6209590435028076, "perplexity": 37.37339335025814, "lr": 0.0005843818681318681, "grad_norm": 0.183393, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:35.258198+00:00", "epoch": 1, "step": 2842, "train_loss": 3.540588140487671, "perplexity": 34.48719654390045, "lr": 0.0005822307692307691, "grad_norm": 0.199056, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:35.564451+00:00", "epoch": 1, "step": 2843, "train_loss": 3.6707611083984375, "perplexity": 39.281792187434014, "lr": 0.0005800796703296703, "grad_norm": 0.181667, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:35.871357+00:00", "epoch": 1, "step": 2844, "train_loss": 3.6016957759857178, "perplexity": 36.66034950249466, "lr": 0.0005779285714285714, "grad_norm": 0.203429, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:36.178713+00:00", "epoch": 1, "step": 2845, "train_loss": 3.6547815799713135, "perplexity": 38.65907627638472, "lr": 0.0005757774725274725, "grad_norm": 0.187002, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:36.485623+00:00", "epoch": 1, "step": 2846, "train_loss": 3.6242125034332275, "perplexity": 37.49518420138437, "lr": 0.0005736263736263736, "grad_norm": 0.192661, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:36.793923+00:00", "epoch": 1, "step": 2847, "train_loss": 3.6332640647888184, "perplexity": 37.83611481116636, "lr": 0.0005714752747252746, "grad_norm": 0.189063, "tokens_per_sec": 106286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:37.100210+00:00", "epoch": 1, "step": 2848, "train_loss": 3.5832390785217285, "perplexity": 35.989926452019255, "lr": 0.0005693241758241758, "grad_norm": 0.187381, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:37.406603+00:00", "epoch": 1, "step": 2849, "train_loss": 3.409268379211426, "perplexity": 30.243109675576793, "lr": 0.0005671730769230769, "grad_norm": 0.189158, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:37.712664+00:00", "epoch": 1, "step": 2850, "train_loss": 3.557016372680664, "perplexity": 35.0584396258803, "lr": 0.000565021978021978, "grad_norm": 0.191333, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:38.019908+00:00", "epoch": 1, "step": 2851, "train_loss": 3.632323980331421, "perplexity": 37.80056238146568, "lr": 0.0005628708791208791, "grad_norm": 0.179486, "tokens_per_sec": 106589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:38.325631+00:00", "epoch": 1, "step": 2852, "train_loss": 3.5926926136016846, "perplexity": 36.33177176170202, "lr": 0.0005607197802197803, "grad_norm": 0.188036, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:38.631453+00:00", "epoch": 1, "step": 2853, "train_loss": 3.626330852508545, "perplexity": 37.57469627759923, "lr": 0.0005585686813186813, "grad_norm": 0.197231, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:38.937109+00:00", "epoch": 1, "step": 2854, "train_loss": 3.5087623596191406, "perplexity": 33.40689646375483, "lr": 0.0005564175824175824, "grad_norm": 0.186649, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:39.245014+00:00", "epoch": 1, "step": 2855, "train_loss": 3.6025924682617188, "perplexity": 36.693237297641346, "lr": 0.0005542664835164836, "grad_norm": 0.189472, "tokens_per_sec": 106422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:39.552894+00:00", "epoch": 1, "step": 2856, "train_loss": 3.681795597076416, "perplexity": 39.717646973465726, "lr": 0.0005521153846153845, "grad_norm": 0.191058, "tokens_per_sec": 106431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:39.860089+00:00", "epoch": 1, "step": 2857, "train_loss": 3.6821277141571045, "perplexity": 39.73084007313598, "lr": 0.0005499642857142856, "grad_norm": 0.189949, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:40.167196+00:00", "epoch": 1, "step": 2858, "train_loss": 3.669557809829712, "perplexity": 39.234552890307135, "lr": 0.0005478131868131867, "grad_norm": 0.180341, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:40.474362+00:00", "epoch": 1, "step": 2859, "train_loss": 3.600759506225586, "perplexity": 36.62604158909674, "lr": 0.0005456620879120878, "grad_norm": 0.189983, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:40.781135+00:00", "epoch": 1, "step": 2860, "train_loss": 3.538311719894409, "perplexity": 34.40877846960778, "lr": 0.000543510989010989, "grad_norm": 0.189166, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:41.089484+00:00", "epoch": 1, "step": 2861, "train_loss": 3.6513357162475586, "perplexity": 38.52609162293465, "lr": 0.00054135989010989, "grad_norm": 0.183566, "tokens_per_sec": 106211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:41.396590+00:00", "epoch": 1, "step": 2862, "train_loss": 3.570117950439453, "perplexity": 35.52078259646716, "lr": 0.0005392087912087912, "grad_norm": 0.195172, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:41.704156+00:00", "epoch": 1, "step": 2863, "train_loss": 3.5625197887420654, "perplexity": 35.25191269869407, "lr": 0.0005370576923076923, "grad_norm": 0.182712, "tokens_per_sec": 106539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:42.010716+00:00", "epoch": 1, "step": 2864, "train_loss": 3.605538845062256, "perplexity": 36.80150882678589, "lr": 0.0005349065934065933, "grad_norm": 0.186848, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:42.317381+00:00", "epoch": 1, "step": 2865, "train_loss": 3.53039288520813, "perplexity": 34.13737705088497, "lr": 0.0005327554945054945, "grad_norm": 0.182009, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:42.624373+00:00", "epoch": 1, "step": 2866, "train_loss": 3.558835506439209, "perplexity": 35.12227366064589, "lr": 0.0005306043956043955, "grad_norm": 0.180162, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:42.932023+00:00", "epoch": 1, "step": 2867, "train_loss": 3.5254478454589844, "perplexity": 33.96898306506405, "lr": 0.0005284532967032967, "grad_norm": 0.182671, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:43.238726+00:00", "epoch": 1, "step": 2868, "train_loss": 3.453862428665161, "perplexity": 31.622295596792448, "lr": 0.0005263021978021978, "grad_norm": 0.180628, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:43.545063+00:00", "epoch": 1, "step": 2869, "train_loss": 3.5217366218566895, "perplexity": 33.84315021471562, "lr": 0.0005241510989010989, "grad_norm": 0.183755, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:43.851340+00:00", "epoch": 1, "step": 2870, "train_loss": 3.4659764766693115, "perplexity": 32.00769928991357, "lr": 0.000522, "grad_norm": 0.173487, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:44.157816+00:00", "epoch": 1, "step": 2871, "train_loss": 3.6280457973480225, "perplexity": 37.6391900949385, "lr": 0.0005198489010989011, "grad_norm": 0.197433, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:44.464947+00:00", "epoch": 1, "step": 2872, "train_loss": 3.6019508838653564, "perplexity": 36.669703039552836, "lr": 0.0005176978021978022, "grad_norm": 0.177563, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:44.773442+00:00", "epoch": 1, "step": 2873, "train_loss": 3.553925037384033, "perplexity": 34.95022957696629, "lr": 0.0005155467032967032, "grad_norm": 0.187139, "tokens_per_sec": 106219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:45.080721+00:00", "epoch": 1, "step": 2874, "train_loss": 3.5624003410339355, "perplexity": 35.247702189987834, "lr": 0.0005133956043956044, "grad_norm": 0.187389, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:45.387587+00:00", "epoch": 1, "step": 2875, "train_loss": 3.7137317657470703, "perplexity": 41.00654818360867, "lr": 0.0005112445054945054, "grad_norm": 0.18814, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:45.694060+00:00", "epoch": 1, "step": 2876, "train_loss": 3.5516417026519775, "perplexity": 34.87051754313992, "lr": 0.0005090934065934065, "grad_norm": 0.169745, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:46.001079+00:00", "epoch": 1, "step": 2877, "train_loss": 3.6397483348846436, "perplexity": 38.082251545093726, "lr": 0.0005069423076923076, "grad_norm": 0.187783, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:46.307211+00:00", "epoch": 1, "step": 2878, "train_loss": 3.6025240421295166, "perplexity": 36.690726607234495, "lr": 0.0005047912087912087, "grad_norm": 0.177898, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:46.614486+00:00", "epoch": 1, "step": 2879, "train_loss": 3.586318254470825, "perplexity": 36.10091655930011, "lr": 0.0005026401098901099, "grad_norm": 0.183798, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:46.922043+00:00", "epoch": 1, "step": 2880, "train_loss": 3.611882209777832, "perplexity": 37.035696202050076, "lr": 0.0005004890109890109, "grad_norm": 0.176873, "tokens_per_sec": 106607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:47.228672+00:00", "epoch": 1, "step": 2881, "train_loss": 3.5544240474700928, "perplexity": 34.967674446256154, "lr": 0.0004983379120879121, "grad_norm": 0.175751, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:47.535198+00:00", "epoch": 1, "step": 2882, "train_loss": 3.5294477939605713, "perplexity": 34.10512935552718, "lr": 0.0004961868131868132, "grad_norm": 0.175576, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:47.842607+00:00", "epoch": 1, "step": 2883, "train_loss": 3.704317331314087, "perplexity": 40.622306272543156, "lr": 0.0004940357142857142, "grad_norm": 0.174424, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:48.149272+00:00", "epoch": 1, "step": 2884, "train_loss": 3.6209728717803955, "perplexity": 37.3739101634891, "lr": 0.0004918846153846154, "grad_norm": 0.177674, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:48.457583+00:00", "epoch": 1, "step": 2885, "train_loss": 3.509244203567505, "perplexity": 33.42299725337203, "lr": 0.0004897335164835164, "grad_norm": 0.174437, "tokens_per_sec": 106282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:48.765719+00:00", "epoch": 1, "step": 2886, "train_loss": 3.6264848709106445, "perplexity": 37.5804839179694, "lr": 0.0004875824175824176, "grad_norm": 0.177214, "tokens_per_sec": 106343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:49.072705+00:00", "epoch": 1, "step": 2887, "train_loss": 3.5228326320648193, "perplexity": 33.880262987099805, "lr": 0.0004854313186813187, "grad_norm": 0.175055, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:49.379654+00:00", "epoch": 1, "step": 2888, "train_loss": 3.496028423309326, "perplexity": 32.98419222816967, "lr": 0.00048328021978021984, "grad_norm": 0.174574, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:49.686133+00:00", "epoch": 1, "step": 2889, "train_loss": 3.5866785049438477, "perplexity": 36.11392427444423, "lr": 0.0004811291208791209, "grad_norm": 0.171095, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:49.993767+00:00", "epoch": 1, "step": 2890, "train_loss": 3.5661823749542236, "perplexity": 35.381262601083634, "lr": 0.00047897802197802185, "grad_norm": 0.174314, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:50.301552+00:00", "epoch": 1, "step": 2891, "train_loss": 3.6617302894592285, "perplexity": 38.928642450359604, "lr": 0.000476826923076923, "grad_norm": 0.171525, "tokens_per_sec": 106407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:50.609674+00:00", "epoch": 1, "step": 2892, "train_loss": 3.6257524490356445, "perplexity": 37.552969226885615, "lr": 0.0004746758241758241, "grad_norm": 0.175214, "tokens_per_sec": 106347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:50.916906+00:00", "epoch": 1, "step": 2893, "train_loss": 3.6040847301483154, "perplexity": 36.74803409257209, "lr": 0.00047252472527472525, "grad_norm": 0.173609, "tokens_per_sec": 106656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:51.223765+00:00", "epoch": 1, "step": 2894, "train_loss": 3.600536823272705, "perplexity": 36.61788650203647, "lr": 0.0004703736263736263, "grad_norm": 0.177301, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:51.531247+00:00", "epoch": 1, "step": 2895, "train_loss": 3.7067739963531494, "perplexity": 40.72222435454626, "lr": 0.00046822252747252737, "grad_norm": 0.173774, "tokens_per_sec": 106568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:51.839529+00:00", "epoch": 1, "step": 2896, "train_loss": 3.623729705810547, "perplexity": 37.47708598482948, "lr": 0.00046607142857142854, "grad_norm": 0.171836, "tokens_per_sec": 106292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:52.148096+00:00", "epoch": 1, "step": 2897, "train_loss": 3.509491205215454, "perplexity": 33.43125380842239, "lr": 0.0004639203296703296, "grad_norm": 0.17522, "tokens_per_sec": 106194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:52.455312+00:00", "epoch": 1, "step": 2898, "train_loss": 3.5307984352111816, "perplexity": 34.15122427193261, "lr": 0.00046176923076923077, "grad_norm": 0.170883, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:52.762738+00:00", "epoch": 1, "step": 2899, "train_loss": 3.6166229248046875, "perplexity": 37.21168871914029, "lr": 0.00045961813186813183, "grad_norm": 0.175566, "tokens_per_sec": 106589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:01:53.070754+00:00", "epoch": 1, "step": 2900, "train_loss": 3.688359260559082, "perplexity": 39.97919766889421, "lr": 0.000457467032967033, "grad_norm": 0.181241, "tokens_per_sec": 106441} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T13:01:58.748450+00:00", "step": 2900, "epoch": 1, "val_loss": 3.563421976566315, "val_ppl": 35.28373089594923, "eval_train_loss": 3.688359260559082, "eval_train_ppl": 39.97919766889421} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T13:01:59.713014+00:00", "step": 2900, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p5634_epoch_0001_step_0002900.pt", "val_loss": 3.563421976566315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:00.807708+00:00", "epoch": 1, "step": 2901, "train_loss": 3.505878448486328, "perplexity": 33.31069273131454, "lr": 0.00045531593406593406, "grad_norm": 0.173109, "tokens_per_sec": 4235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:01.112153+00:00", "epoch": 1, "step": 2902, "train_loss": 3.5350661277770996, "perplexity": 34.297282642303195, "lr": 0.0004531648351648351, "grad_norm": 0.181637, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:01.418448+00:00", "epoch": 1, "step": 2903, "train_loss": 3.550668716430664, "perplexity": 34.836605510685914, "lr": 0.0004510137362637363, "grad_norm": 0.193401, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:01.725433+00:00", "epoch": 1, "step": 2904, "train_loss": 3.572601318359375, "perplexity": 35.6091033895795, "lr": 0.00044886263736263734, "grad_norm": 0.176612, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:02.031339+00:00", "epoch": 1, "step": 2905, "train_loss": 3.642087459564209, "perplexity": 38.17143494441751, "lr": 0.0004467115384615385, "grad_norm": 0.187329, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:02.336344+00:00", "epoch": 1, "step": 2906, "train_loss": 3.5657927989959717, "perplexity": 35.36748159635001, "lr": 0.0004445604395604396, "grad_norm": 0.200012, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:02.642360+00:00", "epoch": 1, "step": 2907, "train_loss": 3.570834159851074, "perplexity": 35.546232027744125, "lr": 0.0004424093406593405, "grad_norm": 0.182194, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:02.948577+00:00", "epoch": 1, "step": 2908, "train_loss": 3.538947582244873, "perplexity": 34.430664673931304, "lr": 0.0004402582417582417, "grad_norm": 0.185185, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:03.256025+00:00", "epoch": 1, "step": 2909, "train_loss": 3.5528082847595215, "perplexity": 34.91122060210058, "lr": 0.00043810714285714275, "grad_norm": 0.190618, "tokens_per_sec": 106581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:03.560639+00:00", "epoch": 1, "step": 2910, "train_loss": 3.563049077987671, "perplexity": 35.27057609570447, "lr": 0.0004359560439560439, "grad_norm": 0.176912, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:03.865993+00:00", "epoch": 1, "step": 2911, "train_loss": 3.5578792095184326, "perplexity": 35.088702393101855, "lr": 0.000433804945054945, "grad_norm": 0.195888, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:04.171037+00:00", "epoch": 1, "step": 2912, "train_loss": 3.5265471935272217, "perplexity": 34.00634733538271, "lr": 0.00043165384615384604, "grad_norm": 0.187883, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:04.477642+00:00", "epoch": 1, "step": 2913, "train_loss": 3.5107853412628174, "perplexity": 33.474546406298, "lr": 0.0004295027472527472, "grad_norm": 0.180547, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:04.784918+00:00", "epoch": 1, "step": 2914, "train_loss": 3.5441761016845703, "perplexity": 34.61115751748748, "lr": 0.00042735164835164827, "grad_norm": 0.186233, "tokens_per_sec": 106641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:05.106287+00:00", "epoch": 1, "step": 2915, "train_loss": 3.6096997261047363, "perplexity": 36.9549545404957, "lr": 0.00042520054945054944, "grad_norm": 0.186482, "tokens_per_sec": 101964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:05.414330+00:00", "epoch": 1, "step": 2916, "train_loss": 3.6398253440856934, "perplexity": 38.08518434178411, "lr": 0.0004230494505494505, "grad_norm": 0.180968, "tokens_per_sec": 106376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:05.720249+00:00", "epoch": 1, "step": 2917, "train_loss": 3.592832565307617, "perplexity": 36.33685681096198, "lr": 0.00042089835164835167, "grad_norm": 0.190468, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:06.026040+00:00", "epoch": 1, "step": 2918, "train_loss": 3.591369390487671, "perplexity": 36.28372851451143, "lr": 0.0004187472527472527, "grad_norm": 0.179872, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:06.331744+00:00", "epoch": 1, "step": 2919, "train_loss": 3.615997076034546, "perplexity": 37.1884071156619, "lr": 0.0004165961538461539, "grad_norm": 0.172743, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:06.638770+00:00", "epoch": 1, "step": 2920, "train_loss": 3.646390438079834, "perplexity": 38.33603970027803, "lr": 0.00041444505494505496, "grad_norm": 0.183719, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:06.946460+00:00", "epoch": 1, "step": 2921, "train_loss": 3.6627988815307617, "perplexity": 38.9702635230512, "lr": 0.000412293956043956, "grad_norm": 0.166921, "tokens_per_sec": 106436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:07.253706+00:00", "epoch": 1, "step": 2922, "train_loss": 3.6314163208007812, "perplexity": 37.76626790696437, "lr": 0.0004101428571428572, "grad_norm": 0.183253, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:07.559785+00:00", "epoch": 1, "step": 2923, "train_loss": 3.6013715267181396, "perplexity": 36.648464338001396, "lr": 0.00040799175824175824, "grad_norm": 0.167604, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:07.865604+00:00", "epoch": 1, "step": 2924, "train_loss": 3.5698812007904053, "perplexity": 35.51237405905196, "lr": 0.0004058406593406594, "grad_norm": 0.182608, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:08.172559+00:00", "epoch": 1, "step": 2925, "train_loss": 3.558243751525879, "perplexity": 35.101496030878444, "lr": 0.00040368956043956036, "grad_norm": 0.174641, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:08.479333+00:00", "epoch": 1, "step": 2926, "train_loss": 3.6280367374420166, "perplexity": 37.63884908895884, "lr": 0.0004015384615384614, "grad_norm": 0.171406, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:08.786528+00:00", "epoch": 1, "step": 2927, "train_loss": 3.4128453731536865, "perplexity": 30.35148280514365, "lr": 0.0003993873626373626, "grad_norm": 0.183071, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:09.093072+00:00", "epoch": 1, "step": 2928, "train_loss": 3.5787644386291504, "perplexity": 35.82924425696963, "lr": 0.00039723626373626365, "grad_norm": 0.160548, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:09.399617+00:00", "epoch": 1, "step": 2929, "train_loss": 3.4695346355438232, "perplexity": 32.121790626374306, "lr": 0.0003950851648351648, "grad_norm": 0.184366, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:09.705645+00:00", "epoch": 1, "step": 2930, "train_loss": 3.6239442825317383, "perplexity": 37.48512855790353, "lr": 0.0003929340659340659, "grad_norm": 0.167036, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:10.012352+00:00", "epoch": 1, "step": 2931, "train_loss": 3.487384796142578, "perplexity": 32.70031778797158, "lr": 0.00039078296703296694, "grad_norm": 0.163445, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:10.319216+00:00", "epoch": 1, "step": 2932, "train_loss": 3.65048885345459, "perplexity": 38.49347912048721, "lr": 0.0003886318681318681, "grad_norm": 0.178813, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:10.627989+00:00", "epoch": 1, "step": 2933, "train_loss": 3.5561130046844482, "perplexity": 35.0267832543543, "lr": 0.00038648076923076917, "grad_norm": 0.162026, "tokens_per_sec": 106119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:10.934701+00:00", "epoch": 1, "step": 2934, "train_loss": 3.658578634262085, "perplexity": 38.80614592686141, "lr": 0.00038432967032967034, "grad_norm": 0.166111, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:11.241300+00:00", "epoch": 1, "step": 2935, "train_loss": 3.512390375137329, "perplexity": 33.528317327745796, "lr": 0.0003821785714285714, "grad_norm": 0.169513, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:11.547494+00:00", "epoch": 1, "step": 2936, "train_loss": 3.550964832305908, "perplexity": 34.84692271008512, "lr": 0.00038002747252747257, "grad_norm": 0.163262, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:11.854743+00:00", "epoch": 1, "step": 2937, "train_loss": 3.5778002738952637, "perplexity": 35.79471561154122, "lr": 0.0003778763736263736, "grad_norm": 0.162091, "tokens_per_sec": 106651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:12.161092+00:00", "epoch": 1, "step": 2938, "train_loss": 3.606459617614746, "perplexity": 36.83541025135924, "lr": 0.0003757252747252747, "grad_norm": 0.171548, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:12.468271+00:00", "epoch": 1, "step": 2939, "train_loss": 3.5198631286621094, "perplexity": 33.77980466036614, "lr": 0.00037357417582417585, "grad_norm": 0.165619, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:12.774602+00:00", "epoch": 1, "step": 2940, "train_loss": 3.5105795860290527, "perplexity": 33.4676595517045, "lr": 0.0003714230769230769, "grad_norm": 0.163363, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:13.079837+00:00", "epoch": 1, "step": 2941, "train_loss": 3.4106922149658203, "perplexity": 30.286201567064822, "lr": 0.0003692719780219781, "grad_norm": 0.168994, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:13.385768+00:00", "epoch": 1, "step": 2942, "train_loss": 3.63295316696167, "perplexity": 37.8243534736657, "lr": 0.00036712087912087903, "grad_norm": 0.167648, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:13.692684+00:00", "epoch": 1, "step": 2943, "train_loss": 3.66819167137146, "perplexity": 39.180989654440175, "lr": 0.0003649697802197801, "grad_norm": 0.175713, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:13.998444+00:00", "epoch": 1, "step": 2944, "train_loss": 3.5553653240203857, "perplexity": 35.000604193798964, "lr": 0.00036281868131868126, "grad_norm": 0.162529, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:14.304070+00:00", "epoch": 1, "step": 2945, "train_loss": 3.6171302795410156, "perplexity": 37.230573035777375, "lr": 0.0003606675824175823, "grad_norm": 0.172572, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:14.611076+00:00", "epoch": 1, "step": 2946, "train_loss": 3.56196928024292, "perplexity": 35.23251156187764, "lr": 0.0003585164835164835, "grad_norm": 0.167032, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:14.916674+00:00", "epoch": 1, "step": 2947, "train_loss": 3.5125632286071777, "perplexity": 33.534113314647875, "lr": 0.00035636538461538455, "grad_norm": 0.161257, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:15.223680+00:00", "epoch": 1, "step": 2948, "train_loss": 3.5482382774353027, "perplexity": 34.75204007338253, "lr": 0.0003542142857142857, "grad_norm": 0.169129, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:15.531018+00:00", "epoch": 1, "step": 2949, "train_loss": 3.525020122528076, "perplexity": 33.95445685888469, "lr": 0.0003520631868131868, "grad_norm": 0.164703, "tokens_per_sec": 106619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:15.837891+00:00", "epoch": 1, "step": 2950, "train_loss": 3.544085741043091, "perplexity": 34.60803017238843, "lr": 0.00034991208791208784, "grad_norm": 0.173425, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:16.144319+00:00", "epoch": 1, "step": 2951, "train_loss": 3.614433526992798, "perplexity": 37.13030665065172, "lr": 0.000347760989010989, "grad_norm": 0.170879, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:16.450328+00:00", "epoch": 1, "step": 2952, "train_loss": 3.6174046993255615, "perplexity": 37.240791243583445, "lr": 0.00034560989010989007, "grad_norm": 0.17079, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:16.756439+00:00", "epoch": 1, "step": 2953, "train_loss": 3.4739458560943604, "perplexity": 32.26379991617594, "lr": 0.00034345879120879124, "grad_norm": 0.179483, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:17.063476+00:00", "epoch": 1, "step": 2954, "train_loss": 3.504817008972168, "perplexity": 33.275354203977976, "lr": 0.0003413076923076923, "grad_norm": 0.174714, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:17.370899+00:00", "epoch": 1, "step": 2955, "train_loss": 3.6055212020874023, "perplexity": 36.80085954441874, "lr": 0.00033915659340659347, "grad_norm": 0.177509, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:17.677099+00:00", "epoch": 1, "step": 2956, "train_loss": 3.492765188217163, "perplexity": 32.87673248316839, "lr": 0.0003370054945054945, "grad_norm": 0.166946, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:17.982903+00:00", "epoch": 1, "step": 2957, "train_loss": 3.5691075325012207, "perplexity": 35.4849098868183, "lr": 0.0003348543956043956, "grad_norm": 0.168559, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:18.288796+00:00", "epoch": 1, "step": 2958, "train_loss": 3.4632835388183594, "perplexity": 31.921620499410082, "lr": 0.00033270329670329675, "grad_norm": 0.174668, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:18.595866+00:00", "epoch": 1, "step": 2959, "train_loss": 3.6652369499206543, "perplexity": 39.06539160795943, "lr": 0.0003305521978021977, "grad_norm": 0.162526, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:18.901835+00:00", "epoch": 1, "step": 2960, "train_loss": 3.5644071102142334, "perplexity": 35.31850721332663, "lr": 0.00032840109890109877, "grad_norm": 0.177501, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:19.208669+00:00", "epoch": 1, "step": 2961, "train_loss": 3.5016987323760986, "perplexity": 33.17175405664279, "lr": 0.00032624999999999993, "grad_norm": 0.167798, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:19.515043+00:00", "epoch": 1, "step": 2962, "train_loss": 3.6093742847442627, "perplexity": 36.942929826589555, "lr": 0.0003240989010989011, "grad_norm": 0.1723, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:19.821433+00:00", "epoch": 1, "step": 2963, "train_loss": 3.5875377655029297, "perplexity": 36.14496888099983, "lr": 0.00032194780219780205, "grad_norm": 0.173936, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:20.127053+00:00", "epoch": 1, "step": 2964, "train_loss": 3.568315029144287, "perplexity": 35.4567991170235, "lr": 0.0003197967032967032, "grad_norm": 0.162322, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:20.433589+00:00", "epoch": 1, "step": 2965, "train_loss": 3.524362087249756, "perplexity": 33.93212097812007, "lr": 0.0003176456043956044, "grad_norm": 0.173688, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:20.740901+00:00", "epoch": 1, "step": 2966, "train_loss": 3.6953885555267334, "perplexity": 40.26121326606643, "lr": 0.00031549450549450556, "grad_norm": 0.177081, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:21.047919+00:00", "epoch": 1, "step": 2967, "train_loss": 3.605652332305908, "perplexity": 36.805685565583694, "lr": 0.0003133434065934065, "grad_norm": 0.190029, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:21.353599+00:00", "epoch": 1, "step": 2968, "train_loss": 3.5597004890441895, "perplexity": 35.15266695935205, "lr": 0.0003111923076923077, "grad_norm": 0.164972, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:21.660047+00:00", "epoch": 1, "step": 2969, "train_loss": 3.519484281539917, "perplexity": 33.76700970240187, "lr": 0.00030904120879120885, "grad_norm": 0.178898, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:21.966837+00:00", "epoch": 1, "step": 2970, "train_loss": 3.5259437561035156, "perplexity": 33.98583282299175, "lr": 0.0003068901098901098, "grad_norm": 0.177183, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:22.274407+00:00", "epoch": 1, "step": 2971, "train_loss": 3.6294562816619873, "perplexity": 37.69231704071173, "lr": 0.00030473901098901097, "grad_norm": 0.165341, "tokens_per_sec": 106481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:22.580668+00:00", "epoch": 1, "step": 2972, "train_loss": 3.710522413253784, "perplexity": 40.87515467277913, "lr": 0.00030258791208791214, "grad_norm": 0.174492, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:22.886502+00:00", "epoch": 1, "step": 2973, "train_loss": 3.553786277770996, "perplexity": 34.94538023308895, "lr": 0.0003004368131868133, "grad_norm": 0.168691, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:23.192498+00:00", "epoch": 1, "step": 2974, "train_loss": 3.580143451690674, "perplexity": 35.87868733627519, "lr": 0.00029828571428571426, "grad_norm": 0.17859, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:23.498789+00:00", "epoch": 1, "step": 2975, "train_loss": 3.5226728916168213, "perplexity": 33.87485137095046, "lr": 0.0002961346153846154, "grad_norm": 0.166474, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:23.805376+00:00", "epoch": 1, "step": 2976, "train_loss": 3.522541046142578, "perplexity": 33.87038541952075, "lr": 0.0002939835164835164, "grad_norm": 0.178693, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:24.112830+00:00", "epoch": 1, "step": 2977, "train_loss": 3.6787197589874268, "perplexity": 39.59566960957737, "lr": 0.00029183241758241754, "grad_norm": 0.179358, "tokens_per_sec": 106579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:24.419751+00:00", "epoch": 1, "step": 2978, "train_loss": 3.570249080657959, "perplexity": 35.525440749856195, "lr": 0.0002896813186813187, "grad_norm": 0.172832, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:24.726323+00:00", "epoch": 1, "step": 2979, "train_loss": 3.577692747116089, "perplexity": 35.790866927982016, "lr": 0.00028753021978021966, "grad_norm": 0.183587, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:25.032562+00:00", "epoch": 1, "step": 2980, "train_loss": 3.6528966426849365, "perplexity": 38.58627497655192, "lr": 0.00028537912087912083, "grad_norm": 0.166037, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:25.338751+00:00", "epoch": 1, "step": 2981, "train_loss": 3.5315489768981934, "perplexity": 34.176865810729396, "lr": 0.000283228021978022, "grad_norm": 0.175017, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:25.646268+00:00", "epoch": 1, "step": 2982, "train_loss": 3.5799834728240967, "perplexity": 35.87294796364226, "lr": 0.00028107692307692295, "grad_norm": 0.163966, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:25.953749+00:00", "epoch": 1, "step": 2983, "train_loss": 3.5075459480285645, "perplexity": 33.36628463304957, "lr": 0.0002789258241758241, "grad_norm": 0.161765, "tokens_per_sec": 106567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:26.260149+00:00", "epoch": 1, "step": 2984, "train_loss": 3.5771138668060303, "perplexity": 35.770154295483444, "lr": 0.0002767747252747253, "grad_norm": 0.160951, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:26.566602+00:00", "epoch": 1, "step": 2985, "train_loss": 3.5746395587921143, "perplexity": 35.68175732181995, "lr": 0.00027462362637362646, "grad_norm": 0.167947, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:26.872441+00:00", "epoch": 1, "step": 2986, "train_loss": 3.7410049438476562, "perplexity": 42.14031754597021, "lr": 0.0002724725274725274, "grad_norm": 0.163647, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:27.179631+00:00", "epoch": 1, "step": 2987, "train_loss": 3.6225502490997314, "perplexity": 37.43290944154432, "lr": 0.0002703214285714286, "grad_norm": 0.162507, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:27.486780+00:00", "epoch": 1, "step": 2988, "train_loss": 3.5741066932678223, "perplexity": 35.662748808437506, "lr": 0.00026817032967032975, "grad_norm": 0.161966, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:27.793470+00:00", "epoch": 1, "step": 2989, "train_loss": 3.5593574047088623, "perplexity": 35.140608698591784, "lr": 0.0002660192307692307, "grad_norm": 0.160584, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:28.099890+00:00", "epoch": 1, "step": 2990, "train_loss": 3.62711501121521, "perplexity": 37.60417235828785, "lr": 0.00026386813186813187, "grad_norm": 0.154794, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:28.407038+00:00", "epoch": 1, "step": 2991, "train_loss": 3.6309914588928223, "perplexity": 37.75022586639222, "lr": 0.00026171703296703304, "grad_norm": 0.160365, "tokens_per_sec": 106620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:28.714245+00:00", "epoch": 1, "step": 2992, "train_loss": 3.629103660583496, "perplexity": 37.67902827831256, "lr": 0.0002595659340659342, "grad_norm": 0.159723, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:29.022563+00:00", "epoch": 1, "step": 2993, "train_loss": 3.4885213375091553, "perplexity": 32.73750417977122, "lr": 0.00025741483516483516, "grad_norm": 0.155314, "tokens_per_sec": 106280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:29.329143+00:00", "epoch": 1, "step": 2994, "train_loss": 3.5608444213867188, "perplexity": 35.19290224084532, "lr": 0.0002552637362637361, "grad_norm": 0.173754, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:29.636008+00:00", "epoch": 1, "step": 2995, "train_loss": 3.5472586154937744, "perplexity": 34.71801149330613, "lr": 0.0002531126373626373, "grad_norm": 0.158695, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:29.941956+00:00", "epoch": 1, "step": 2996, "train_loss": 3.5705373287200928, "perplexity": 35.53568236530078, "lr": 0.00025096153846153844, "grad_norm": 0.155671, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:30.248049+00:00", "epoch": 1, "step": 2997, "train_loss": 3.682298183441162, "perplexity": 39.73761353831573, "lr": 0.0002488104395604394, "grad_norm": 0.167008, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:30.554310+00:00", "epoch": 1, "step": 2998, "train_loss": 3.565702438354492, "perplexity": 35.36428591240965, "lr": 0.00024665934065934056, "grad_norm": 0.164777, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:30.862191+00:00", "epoch": 1, "step": 2999, "train_loss": 3.5098278522491455, "perplexity": 33.44251023546466, "lr": 0.00024450824175824173, "grad_norm": 0.164995, "tokens_per_sec": 106431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:31.169414+00:00", "epoch": 1, "step": 3000, "train_loss": 3.6338419914245605, "perplexity": 37.85798762954211, "lr": 0.0002423571428571429, "grad_norm": 0.157939, "tokens_per_sec": 106718} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-08T13:02:36.894749+00:00", "step": 3000, "epoch": 1, "val_loss": 3.543744647502899, "val_ppl": 34.59622760986122, "eval_train_loss": 3.6338419914245605, "eval_train_ppl": 37.85798762954211} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-08T13:02:37.806055+00:00", "step": 3000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p5437_epoch_0001_step_0003000.pt", "val_loss": 3.543744647502899} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T13:02:39.709565+00:00", "step": 3000, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0001_step_0003000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:40.743631+00:00", "epoch": 1, "step": 3001, "train_loss": 3.620394706726074, "perplexity": 37.352308120064286, "lr": 0.00024020604395604385, "grad_norm": 0.162423, "tokens_per_sec": 3422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:41.050109+00:00", "epoch": 1, "step": 3002, "train_loss": 3.5203912258148193, "perplexity": 33.797648390223145, "lr": 0.00023805494505494502, "grad_norm": 0.162056, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:41.354954+00:00", "epoch": 1, "step": 3003, "train_loss": 3.591874361038208, "perplexity": 36.30205535574306, "lr": 0.0002359038461538462, "grad_norm": 0.156518, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:41.659329+00:00", "epoch": 1, "step": 3004, "train_loss": 3.653062105178833, "perplexity": 38.592660086073224, "lr": 0.00023375274725274714, "grad_norm": 0.157852, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:41.964169+00:00", "epoch": 1, "step": 3005, "train_loss": 3.698915481567383, "perplexity": 40.40346229100241, "lr": 0.0002316016483516483, "grad_norm": 0.161799, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:42.268410+00:00", "epoch": 1, "step": 3006, "train_loss": 3.56524395942688, "perplexity": 35.34807584879899, "lr": 0.00022945054945054948, "grad_norm": 0.157909, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:42.574037+00:00", "epoch": 1, "step": 3007, "train_loss": 3.667965888977051, "perplexity": 39.17214427538368, "lr": 0.00022729945054945065, "grad_norm": 0.159098, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:42.880136+00:00", "epoch": 1, "step": 3008, "train_loss": 3.5328011512756348, "perplexity": 34.21968801123475, "lr": 0.0002251483516483516, "grad_norm": 0.155, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:43.186069+00:00", "epoch": 1, "step": 3009, "train_loss": 3.5924999713897705, "perplexity": 36.32477340293839, "lr": 0.00022299725274725277, "grad_norm": 0.160866, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:43.491995+00:00", "epoch": 1, "step": 3010, "train_loss": 3.604189872741699, "perplexity": 36.751898079309576, "lr": 0.00022084615384615372, "grad_norm": 0.152648, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:43.796996+00:00", "epoch": 1, "step": 3011, "train_loss": 3.5342416763305664, "perplexity": 34.2690178510916, "lr": 0.0002186950549450549, "grad_norm": 0.159938, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:44.101939+00:00", "epoch": 1, "step": 3012, "train_loss": 3.5118091106414795, "perplexity": 33.50883417026002, "lr": 0.00021654395604395606, "grad_norm": 0.159656, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:44.408704+00:00", "epoch": 1, "step": 3013, "train_loss": 3.5758445262908936, "perplexity": 35.724778594104116, "lr": 0.000214392857142857, "grad_norm": 0.149438, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:44.714240+00:00", "epoch": 1, "step": 3014, "train_loss": 3.589972734451294, "perplexity": 36.23308799798274, "lr": 0.00021224175824175818, "grad_norm": 0.163745, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:45.138121+00:00", "epoch": 1, "step": 3015, "train_loss": 3.577904462814331, "perplexity": 35.7984452185576, "lr": 0.00021009065934065934, "grad_norm": 0.159735, "tokens_per_sec": 77306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:45.443893+00:00", "epoch": 1, "step": 3016, "train_loss": 3.5654714107513428, "perplexity": 35.35611672988788, "lr": 0.0002079395604395603, "grad_norm": 0.152663, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:45.749960+00:00", "epoch": 1, "step": 3017, "train_loss": 3.5529611110687256, "perplexity": 34.91655636280693, "lr": 0.00020578846153846146, "grad_norm": 0.170292, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:46.055883+00:00", "epoch": 1, "step": 3018, "train_loss": 3.559225559234619, "perplexity": 35.13597587378793, "lr": 0.00020363736263736263, "grad_norm": 0.155974, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:46.361894+00:00", "epoch": 1, "step": 3019, "train_loss": 3.585345506668091, "perplexity": 36.06581654653825, "lr": 0.0002014862637362638, "grad_norm": 0.149297, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:46.667160+00:00", "epoch": 1, "step": 3020, "train_loss": 3.5414798259735107, "perplexity": 34.51796199101881, "lr": 0.00019933516483516475, "grad_norm": 0.157191, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:46.972924+00:00", "epoch": 1, "step": 3021, "train_loss": 3.4548768997192383, "perplexity": 31.654391777909556, "lr": 0.00019718406593406592, "grad_norm": 0.158878, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:47.279540+00:00", "epoch": 1, "step": 3022, "train_loss": 3.572800636291504, "perplexity": 35.61620162981186, "lr": 0.0001950329670329671, "grad_norm": 0.155934, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:47.585793+00:00", "epoch": 1, "step": 3023, "train_loss": 3.4686484336853027, "perplexity": 32.093336845581106, "lr": 0.00019288186813186804, "grad_norm": 0.159736, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:47.892404+00:00", "epoch": 1, "step": 3024, "train_loss": 3.6050145626068115, "perplexity": 36.78221949834421, "lr": 0.0001907307692307692, "grad_norm": 0.157549, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:48.198655+00:00", "epoch": 1, "step": 3025, "train_loss": 3.55098819732666, "perplexity": 34.84773691866935, "lr": 0.00018857967032967038, "grad_norm": 0.160017, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:48.503873+00:00", "epoch": 1, "step": 3026, "train_loss": 3.555368661880493, "perplexity": 35.000721021114416, "lr": 0.00018642857142857155, "grad_norm": 0.162603, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:48.810360+00:00", "epoch": 1, "step": 3027, "train_loss": 3.5708532333374023, "perplexity": 35.54691002478059, "lr": 0.0001842774725274725, "grad_norm": 0.160785, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:49.117421+00:00", "epoch": 1, "step": 3028, "train_loss": 3.543130874633789, "perplexity": 34.57499989914359, "lr": 0.00018212637362637345, "grad_norm": 0.16083, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:49.423985+00:00", "epoch": 1, "step": 3029, "train_loss": 3.591435670852661, "perplexity": 36.28613349298112, "lr": 0.00017997527472527462, "grad_norm": 0.164882, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:49.729810+00:00", "epoch": 1, "step": 3030, "train_loss": 3.644801616668701, "perplexity": 38.27517894082589, "lr": 0.00017782417582417579, "grad_norm": 0.160163, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:50.034864+00:00", "epoch": 1, "step": 3031, "train_loss": 3.5410945415496826, "perplexity": 34.50466531958606, "lr": 0.00017567307692307695, "grad_norm": 0.158838, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:50.341139+00:00", "epoch": 1, "step": 3032, "train_loss": 3.5405900478363037, "perplexity": 34.48726232307036, "lr": 0.0001735219780219779, "grad_norm": 0.161425, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:50.647041+00:00", "epoch": 1, "step": 3033, "train_loss": 3.537008047103882, "perplexity": 34.363949908598364, "lr": 0.00017137087912087907, "grad_norm": 0.153266, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:50.954106+00:00", "epoch": 1, "step": 3034, "train_loss": 3.5267539024353027, "perplexity": 34.01337747687964, "lr": 0.00016921978021978024, "grad_norm": 0.149547, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:51.259806+00:00", "epoch": 1, "step": 3035, "train_loss": 3.5316965579986572, "perplexity": 34.18191004240362, "lr": 0.0001670686813186812, "grad_norm": 0.153993, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:51.564897+00:00", "epoch": 1, "step": 3036, "train_loss": 3.5530154705047607, "perplexity": 34.91845445870835, "lr": 0.00016491758241758236, "grad_norm": 0.162317, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:51.870333+00:00", "epoch": 1, "step": 3037, "train_loss": 3.5920145511627197, "perplexity": 36.30714490214715, "lr": 0.00016276648351648353, "grad_norm": 0.157754, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:52.176303+00:00", "epoch": 1, "step": 3038, "train_loss": 3.5226452350616455, "perplexity": 33.87391452220951, "lr": 0.0001606153846153847, "grad_norm": 0.15016, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:52.483492+00:00", "epoch": 1, "step": 3039, "train_loss": 3.48005747795105, "perplexity": 32.461587847800224, "lr": 0.00015846428571428565, "grad_norm": 0.1537, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:52.790624+00:00", "epoch": 1, "step": 3040, "train_loss": 3.6178958415985107, "perplexity": 37.25908626280214, "lr": 0.00015631318681318682, "grad_norm": 0.155579, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:53.096304+00:00", "epoch": 1, "step": 3041, "train_loss": 3.572038412094116, "perplexity": 35.58906444273476, "lr": 0.000154162087912088, "grad_norm": 0.150938, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:53.402182+00:00", "epoch": 1, "step": 3042, "train_loss": 3.5639803409576416, "perplexity": 35.30343757611693, "lr": 0.00015201098901098894, "grad_norm": 0.147106, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:53.708895+00:00", "epoch": 1, "step": 3043, "train_loss": 3.4003660678863525, "perplexity": 29.97507095009289, "lr": 0.0001498598901098901, "grad_norm": 0.158314, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:54.015130+00:00", "epoch": 1, "step": 3044, "train_loss": 3.6143884658813477, "perplexity": 37.1286335554616, "lr": 0.00014770879120879106, "grad_norm": 0.151301, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:54.322089+00:00", "epoch": 1, "step": 3045, "train_loss": 3.486989974975586, "perplexity": 32.687409558730266, "lr": 0.00014555769230769223, "grad_norm": 0.145682, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:54.628330+00:00", "epoch": 1, "step": 3046, "train_loss": 3.59199595451355, "perplexity": 36.30646971718915, "lr": 0.0001434065934065934, "grad_norm": 0.150385, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:54.934841+00:00", "epoch": 1, "step": 3047, "train_loss": 3.639486074447632, "perplexity": 38.07226538670549, "lr": 0.00014125549450549435, "grad_norm": 0.145734, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:55.240009+00:00", "epoch": 1, "step": 3048, "train_loss": 3.494215488433838, "perplexity": 32.92444820811174, "lr": 0.00013910439560439552, "grad_norm": 0.145485, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:55.544789+00:00", "epoch": 1, "step": 3049, "train_loss": 3.4219398498535156, "perplexity": 30.628772650131243, "lr": 0.00013695329670329669, "grad_norm": 0.1489, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:55.851629+00:00", "epoch": 1, "step": 3050, "train_loss": 3.5033247470855713, "perplexity": 33.225735692251064, "lr": 0.00013480219780219785, "grad_norm": 0.150057, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:56.158869+00:00", "epoch": 1, "step": 3051, "train_loss": 3.650514602661133, "perplexity": 38.49447030979276, "lr": 0.0001326510989010988, "grad_norm": 0.15273, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-08T13:02:56.464967+00:00", "epoch": 1, "step": 3052, "train_loss": 3.4881935119628906, "perplexity": 32.72677374853099, "lr": 0.00013049999999999997, "grad_norm": 0.147662, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-08T13:02:57.478028+00:00", "step": 3052, "epoch": 1, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/last_epoch_0001_step_0003052.pt", "category": "last"} +{"stage": "pretraining", "type": "training_complete", "timestamp": "2026-04-08T13:02:57.480524+00:00", "epochs": 2, "total_steps": 3052, "total_time_hours": 0.35, "best_val_loss": 3.543744647502899, "best_checkpoint_path": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints/best_loss_3p5437_epoch_0001_step_0003000.pt", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c/run_20260408_124138/checkpoints", "stopped_early": false, "stop_reason": null, "run_dir": "runs/big_run/exp_c/run_20260408_124138"}