diff --git "a/runs/big_run/exp_c_8b/run_20260409_205717/metrics.jsonl" "b/runs/big_run/exp_c_8b/run_20260409_205717/metrics.jsonl" new file mode 100644--- /dev/null +++ "b/runs/big_run/exp_c_8b/run_20260409_205717/metrics.jsonl" @@ -0,0 +1,24614 @@ +{"stage": "pretraining", "type": "config", "timestamp": "2026-04-09T20:57:18.390598+00:00", "device": "auto", "train_bin": "data/filter_c/train.bin", "val_bin": "data/filter_c/val.bin", "num_workers": 0, "pin_memory": true, "batch_size": 32, "gradient_accumulation_steps": 1, "train_samples_per_epoch": null, "val_sequence_stride": null, "learning_rate": 0.0026291804804649314, "min_lr": 0.0002629180480464931, "weight_decay": 0.13113982116099054, "beta1": 0.9, "beta2": 0.9606057712410854, "grad_clip": 0.9147944011374611, "warmup_steps": 1500, "max_steps": 122070, "lr_schedule": "wsd", "lr_decay_ratio": 0.25826046407623926, "z_loss_coeff": 0.00013887413811122497, "save_every": 2500, "eval_every": 500, "checkpoint_dir": "checkpoints", "keep_last_checkpoints": 10, "keep_best_checkpoints": 10, "early_stopping_patience": 5, "early_stopping_min_delta": 0.001, "runs_dir": "runs/big_run/exp_c_8b", "log_every": 10, "hf_upload": {"repo_id": "ParrotLabs/Preprocessed", "repo_type": "dataset", "path_in_repo": "", "private": null}, "compile": true} +{"stage": "pretraining", "type": "checkpoint_policy", "timestamp": "2026-04-09T20:57:18.390706+00:00", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints", "keep_last": 10, "keep_best": 10} +{"stage": "pretraining", "type": "model_architecture", "timestamp": "2026-04-09T20:57:20.070349+00:00", "vocab_size": 50258, "context_length": 1024, "n_layers": 14, "n_heads": 6, "d_model": 384, "d_ff": 768, "dropout": 0.01287921140461293, "bias": false, "total_params": 39966592, "total_params_non_embedding": 39966592, "trainable_params": 39966592, "non_trainable_params": 0, "params_size_mb": 159.87} +{"stage": "pretraining", "type": "initial_validation", "timestamp": "2026-04-09T20:57:23.462672+00:00", "val_loss": 10.894190549850464, "val_ppl": 53862.54136245001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:41.836317+00:00", "epoch": 0, "step": 1, "train_loss": 10.907625198364258, "perplexity": 54591.04833395959, "lr": 1.752786986976621e-06, "grad_norm": 7.084432, "tokens_per_sec": 1784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:42.181854+00:00", "epoch": 0, "step": 2, "train_loss": 10.894129753112793, "perplexity": 53859.26679519502, "lr": 3.505573973953242e-06, "grad_norm": 6.896891, "tokens_per_sec": 94775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:42.485465+00:00", "epoch": 0, "step": 3, "train_loss": 10.857537269592285, "perplexity": 51924.04567130085, "lr": 5.258360960929863e-06, "grad_norm": 7.034947, "tokens_per_sec": 107926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:42.788178+00:00", "epoch": 0, "step": 4, "train_loss": 10.821130752563477, "perplexity": 50067.66915753506, "lr": 7.011147947906484e-06, "grad_norm": 6.654187, "tokens_per_sec": 108248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:43.092533+00:00", "epoch": 0, "step": 5, "train_loss": 10.756375312805176, "perplexity": 46928.25911779438, "lr": 8.763934934883105e-06, "grad_norm": 6.850497, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:43.395982+00:00", "epoch": 0, "step": 6, "train_loss": 10.691872596740723, "perplexity": 43996.81820402329, "lr": 1.0516721921859726e-05, "grad_norm": 6.494458, "tokens_per_sec": 107985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:43.699328+00:00", "epoch": 0, "step": 7, "train_loss": 10.596056938171387, "perplexity": 39976.89487866625, "lr": 1.2269508908836347e-05, "grad_norm": 6.086611, "tokens_per_sec": 108023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:44.004625+00:00", "epoch": 0, "step": 8, "train_loss": 10.563188552856445, "perplexity": 38684.278371823464, "lr": 1.4022295895812967e-05, "grad_norm": 5.291498, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:44.309070+00:00", "epoch": 0, "step": 9, "train_loss": 10.457194328308105, "perplexity": 34793.794409945796, "lr": 1.5775082882789588e-05, "grad_norm": 4.258384, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:44.613651+00:00", "epoch": 0, "step": 10, "train_loss": 10.417092323303223, "perplexity": 33426.1004136716, "lr": 1.752786986976621e-05, "grad_norm": 3.318158, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:44.916284+00:00", "epoch": 0, "step": 11, "train_loss": 10.370224952697754, "perplexity": 31895.65117565925, "lr": 1.928065685674283e-05, "grad_norm": 2.708556, "tokens_per_sec": 108215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:45.218323+00:00", "epoch": 0, "step": 12, "train_loss": 10.327241897583008, "perplexity": 30553.725279751412, "lr": 2.103344384371945e-05, "grad_norm": 2.34704, "tokens_per_sec": 108488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:45.521572+00:00", "epoch": 0, "step": 13, "train_loss": 10.298980712890625, "perplexity": 29702.328217816164, "lr": 2.278623083069607e-05, "grad_norm": 2.076818, "tokens_per_sec": 108057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:45.824699+00:00", "epoch": 0, "step": 14, "train_loss": 10.256986618041992, "perplexity": 28480.833123843386, "lr": 2.4539017817672695e-05, "grad_norm": 1.896952, "tokens_per_sec": 108100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:46.126717+00:00", "epoch": 0, "step": 15, "train_loss": 10.227010726928711, "perplexity": 27639.763652672475, "lr": 2.629180480464931e-05, "grad_norm": 1.80941, "tokens_per_sec": 108497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:46.429100+00:00", "epoch": 0, "step": 16, "train_loss": 10.218506813049316, "perplexity": 27405.71406217378, "lr": 2.8044591791625935e-05, "grad_norm": 1.719092, "tokens_per_sec": 108366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:46.732020+00:00", "epoch": 0, "step": 17, "train_loss": 10.213545799255371, "perplexity": 27270.090629697745, "lr": 2.979737877860256e-05, "grad_norm": 1.72923, "tokens_per_sec": 108174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:47.035317+00:00", "epoch": 0, "step": 18, "train_loss": 10.220091819763184, "perplexity": 27449.186746107433, "lr": 3.1550165765579175e-05, "grad_norm": 1.699368, "tokens_per_sec": 108042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:47.339827+00:00", "epoch": 0, "step": 19, "train_loss": 10.178314208984375, "perplexity": 26326.049650857072, "lr": 3.33029527525558e-05, "grad_norm": 1.703214, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:47.643476+00:00", "epoch": 0, "step": 20, "train_loss": 10.14278793334961, "perplexity": 25407.201421394206, "lr": 3.505573973953242e-05, "grad_norm": 1.699245, "tokens_per_sec": 107973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:47.946913+00:00", "epoch": 0, "step": 21, "train_loss": 10.139678955078125, "perplexity": 25328.333646860847, "lr": 3.680852672650904e-05, "grad_norm": 1.710419, "tokens_per_sec": 107931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:48.249817+00:00", "epoch": 0, "step": 22, "train_loss": 10.118196487426758, "perplexity": 24790.02137884378, "lr": 3.856131371348566e-05, "grad_norm": 1.711318, "tokens_per_sec": 108179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:48.552990+00:00", "epoch": 0, "step": 23, "train_loss": 10.114384651184082, "perplexity": 24695.705748964538, "lr": 4.031410070046228e-05, "grad_norm": 1.700977, "tokens_per_sec": 108087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:48.856542+00:00", "epoch": 0, "step": 24, "train_loss": 10.085963249206543, "perplexity": 24003.699642158914, "lr": 4.20668876874389e-05, "grad_norm": 1.687483, "tokens_per_sec": 107947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:49.158820+00:00", "epoch": 0, "step": 25, "train_loss": 10.047271728515625, "perplexity": 23092.697673774343, "lr": 4.3819674674415526e-05, "grad_norm": 1.71227, "tokens_per_sec": 108403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:49.461253+00:00", "epoch": 0, "step": 26, "train_loss": 10.03132438659668, "perplexity": 22727.351421742304, "lr": 4.557246166139214e-05, "grad_norm": 1.692621, "tokens_per_sec": 108349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:49.763771+00:00", "epoch": 0, "step": 27, "train_loss": 9.997313499450684, "perplexity": 21967.371096860654, "lr": 4.732524864836876e-05, "grad_norm": 1.708807, "tokens_per_sec": 108317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:50.066707+00:00", "epoch": 0, "step": 28, "train_loss": 9.966659545898438, "perplexity": 21304.20065063798, "lr": 4.907803563534539e-05, "grad_norm": 1.681106, "tokens_per_sec": 108169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:50.371572+00:00", "epoch": 0, "step": 29, "train_loss": 9.933277130126953, "perplexity": 20604.754520787177, "lr": 5.0830822622322006e-05, "grad_norm": 1.6983, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:50.675517+00:00", "epoch": 0, "step": 30, "train_loss": 9.873411178588867, "perplexity": 19407.42839245464, "lr": 5.258360960929862e-05, "grad_norm": 1.684886, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:50.979690+00:00", "epoch": 0, "step": 31, "train_loss": 9.867805480957031, "perplexity": 19298.940576114488, "lr": 5.433639659627525e-05, "grad_norm": 1.731875, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:51.281840+00:00", "epoch": 0, "step": 32, "train_loss": 9.78109359741211, "perplexity": 17695.994568922324, "lr": 5.608918358325187e-05, "grad_norm": 1.749119, "tokens_per_sec": 108451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:51.586861+00:00", "epoch": 0, "step": 33, "train_loss": 9.77751636505127, "perplexity": 17632.804973809652, "lr": 5.7841970570228486e-05, "grad_norm": 1.715666, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:51.891332+00:00", "epoch": 0, "step": 34, "train_loss": 9.747912406921387, "perplexity": 17118.455113411663, "lr": 5.959475755720512e-05, "grad_norm": 1.69374, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:52.195845+00:00", "epoch": 0, "step": 35, "train_loss": 9.624772071838379, "perplexity": 15135.10368089517, "lr": 6.134754454418173e-05, "grad_norm": 1.680507, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:52.498476+00:00", "epoch": 0, "step": 36, "train_loss": 9.656709671020508, "perplexity": 15626.28437178318, "lr": 6.310033153115835e-05, "grad_norm": 1.622151, "tokens_per_sec": 108277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:52.800572+00:00", "epoch": 0, "step": 37, "train_loss": 9.613618850708008, "perplexity": 14967.236393341824, "lr": 6.485311851813497e-05, "grad_norm": 1.644404, "tokens_per_sec": 108469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:53.104406+00:00", "epoch": 0, "step": 38, "train_loss": 9.584538459777832, "perplexity": 14538.251069549953, "lr": 6.66059055051116e-05, "grad_norm": 1.621659, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:53.408376+00:00", "epoch": 0, "step": 39, "train_loss": 9.569284439086914, "perplexity": 14318.167136862065, "lr": 6.835869249208821e-05, "grad_norm": 1.644453, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:53.712474+00:00", "epoch": 0, "step": 40, "train_loss": 9.533437728881836, "perplexity": 13813.998331723626, "lr": 7.011147947906484e-05, "grad_norm": 1.630085, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:54.016081+00:00", "epoch": 0, "step": 41, "train_loss": 9.492364883422852, "perplexity": 13258.112171394561, "lr": 7.186426646604145e-05, "grad_norm": 1.612092, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:54.319085+00:00", "epoch": 0, "step": 42, "train_loss": 9.467219352722168, "perplexity": 12928.886528710733, "lr": 7.361705345301808e-05, "grad_norm": 1.612146, "tokens_per_sec": 108143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:54.621913+00:00", "epoch": 0, "step": 43, "train_loss": 9.409667015075684, "perplexity": 12205.805950203505, "lr": 7.53698404399947e-05, "grad_norm": 1.579527, "tokens_per_sec": 108207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:54.926065+00:00", "epoch": 0, "step": 44, "train_loss": 9.376564025878906, "perplexity": 11808.37171033892, "lr": 7.712262742697132e-05, "grad_norm": 1.593173, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:55.228733+00:00", "epoch": 0, "step": 45, "train_loss": 9.330817222595215, "perplexity": 11280.346273601748, "lr": 7.887541441394793e-05, "grad_norm": 1.578951, "tokens_per_sec": 108267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:55.532007+00:00", "epoch": 0, "step": 46, "train_loss": 9.271342277526855, "perplexity": 10629.00939596866, "lr": 8.062820140092456e-05, "grad_norm": 1.632975, "tokens_per_sec": 108044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:55.834762+00:00", "epoch": 0, "step": 47, "train_loss": 9.25438117980957, "perplexity": 10450.249992243806, "lr": 8.238098838790118e-05, "grad_norm": 1.644785, "tokens_per_sec": 108233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:56.137430+00:00", "epoch": 0, "step": 48, "train_loss": 9.146221160888672, "perplexity": 9378.931855995073, "lr": 8.41337753748778e-05, "grad_norm": 1.631396, "tokens_per_sec": 108263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:56.439546+00:00", "epoch": 0, "step": 49, "train_loss": 9.15158748626709, "perplexity": 9429.397542576817, "lr": 8.588656236185441e-05, "grad_norm": 1.610557, "tokens_per_sec": 108468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:56.744544+00:00", "epoch": 0, "step": 50, "train_loss": 9.142610549926758, "perplexity": 9345.129242590361, "lr": 8.763934934883105e-05, "grad_norm": 1.570545, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:57.048649+00:00", "epoch": 0, "step": 51, "train_loss": 9.122298240661621, "perplexity": 9157.222953265133, "lr": 8.939213633580767e-05, "grad_norm": 1.588874, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:57.352208+00:00", "epoch": 0, "step": 52, "train_loss": 9.033761024475098, "perplexity": 8381.322728152367, "lr": 9.114492332278428e-05, "grad_norm": 1.620448, "tokens_per_sec": 107947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:57.655847+00:00", "epoch": 0, "step": 53, "train_loss": 9.047924041748047, "perplexity": 8500.872138531555, "lr": 9.289771030976091e-05, "grad_norm": 1.576735, "tokens_per_sec": 107917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:57.958598+00:00", "epoch": 0, "step": 54, "train_loss": 8.99831771850586, "perplexity": 8089.4637191630145, "lr": 9.465049729673752e-05, "grad_norm": 1.577501, "tokens_per_sec": 108234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:58.262637+00:00", "epoch": 0, "step": 55, "train_loss": 8.920747756958008, "perplexity": 7485.684610187312, "lr": 9.640328428371414e-05, "grad_norm": 1.587688, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:58.565029+00:00", "epoch": 0, "step": 56, "train_loss": 8.963961601257324, "perplexity": 7816.261117652883, "lr": 9.815607127069078e-05, "grad_norm": 1.533211, "tokens_per_sec": 108361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:58.867703+00:00", "epoch": 0, "step": 57, "train_loss": 8.931863784790039, "perplexity": 7569.35989543871, "lr": 9.990885825766739e-05, "grad_norm": 1.577538, "tokens_per_sec": 108262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:59.171264+00:00", "epoch": 0, "step": 58, "train_loss": 8.792830467224121, "perplexity": 6586.849677044339, "lr": 0.00010166164524464401, "grad_norm": 1.557575, "tokens_per_sec": 107944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:59.475512+00:00", "epoch": 0, "step": 59, "train_loss": 8.775775909423828, "perplexity": 6475.46636478744, "lr": 0.00010341443223162064, "grad_norm": 1.537498, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:57:59.778754+00:00", "epoch": 0, "step": 60, "train_loss": 8.80052375793457, "perplexity": 6637.719654335719, "lr": 0.00010516721921859725, "grad_norm": 1.55894, "tokens_per_sec": 108119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:00.081491+00:00", "epoch": 0, "step": 61, "train_loss": 8.684514045715332, "perplexity": 5910.66748540339, "lr": 0.00010692000620557387, "grad_norm": 1.530474, "tokens_per_sec": 108179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:00.384259+00:00", "epoch": 0, "step": 62, "train_loss": 8.67270565032959, "perplexity": 5841.282455914892, "lr": 0.0001086727931925505, "grad_norm": 1.555584, "tokens_per_sec": 108228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:00.687198+00:00", "epoch": 0, "step": 63, "train_loss": 8.615456581115723, "perplexity": 5516.26666203462, "lr": 0.00011042558017952712, "grad_norm": 1.563972, "tokens_per_sec": 108167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:00.990311+00:00", "epoch": 0, "step": 64, "train_loss": 8.586727142333984, "perplexity": 5360.042281289336, "lr": 0.00011217836716650374, "grad_norm": 1.498319, "tokens_per_sec": 108105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:01.292948+00:00", "epoch": 0, "step": 65, "train_loss": 8.618507385253906, "perplexity": 5533.12140838661, "lr": 0.00011393115415348035, "grad_norm": 1.521124, "tokens_per_sec": 108275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:01.595675+00:00", "epoch": 0, "step": 66, "train_loss": 8.47697925567627, "perplexity": 4802.919569690415, "lr": 0.00011568394114045697, "grad_norm": 1.47766, "tokens_per_sec": 108243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:01.898500+00:00", "epoch": 0, "step": 67, "train_loss": 8.394466400146484, "perplexity": 4422.526420591646, "lr": 0.00011743672812743361, "grad_norm": 1.515732, "tokens_per_sec": 108208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:02.201232+00:00", "epoch": 0, "step": 68, "train_loss": 8.358550071716309, "perplexity": 4266.504154462321, "lr": 0.00011918951511441023, "grad_norm": 1.533998, "tokens_per_sec": 108241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:02.504207+00:00", "epoch": 0, "step": 69, "train_loss": 8.43408203125, "perplexity": 4601.244232597948, "lr": 0.00012094230210138684, "grad_norm": 1.492722, "tokens_per_sec": 108154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:02.808816+00:00", "epoch": 0, "step": 70, "train_loss": 8.272125244140625, "perplexity": 3913.25678703826, "lr": 0.00012269508908836345, "grad_norm": 1.53787, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:03.111890+00:00", "epoch": 0, "step": 71, "train_loss": 8.324116706848145, "perplexity": 4122.094582014592, "lr": 0.0001244478760753401, "grad_norm": 1.46202, "tokens_per_sec": 108059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:03.415223+00:00", "epoch": 0, "step": 72, "train_loss": 8.295162200927734, "perplexity": 4004.452719804343, "lr": 0.0001262006630623167, "grad_norm": 1.458989, "tokens_per_sec": 108028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:03.718462+00:00", "epoch": 0, "step": 73, "train_loss": 8.201924324035645, "perplexity": 3647.963421108901, "lr": 0.00012795345004929334, "grad_norm": 1.444868, "tokens_per_sec": 108060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:04.022121+00:00", "epoch": 0, "step": 74, "train_loss": 8.138760566711426, "perplexity": 3424.670585694162, "lr": 0.00012970623703626995, "grad_norm": 1.450614, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:04.326682+00:00", "epoch": 0, "step": 75, "train_loss": 8.149968147277832, "perplexity": 3463.268749185265, "lr": 0.00013145902402324656, "grad_norm": 1.407515, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:04.630631+00:00", "epoch": 0, "step": 76, "train_loss": 8.064889907836914, "perplexity": 3180.806016090205, "lr": 0.0001332118110102232, "grad_norm": 1.393021, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:04.934775+00:00", "epoch": 0, "step": 77, "train_loss": 8.019378662109375, "perplexity": 3039.2883211524377, "lr": 0.0001349645979971998, "grad_norm": 1.390204, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:05.237992+00:00", "epoch": 0, "step": 78, "train_loss": 8.063581466674805, "perplexity": 3176.6468401819293, "lr": 0.00013671738498417641, "grad_norm": 1.35331, "tokens_per_sec": 108068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:05.542097+00:00", "epoch": 0, "step": 79, "train_loss": 8.023767471313477, "perplexity": 3052.6564914268606, "lr": 0.00013847017197115305, "grad_norm": 1.332289, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:05.846214+00:00", "epoch": 0, "step": 80, "train_loss": 8.019837379455566, "perplexity": 3040.682815240288, "lr": 0.0001402229589581297, "grad_norm": 1.319537, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:06.150602+00:00", "epoch": 0, "step": 81, "train_loss": 7.88449764251709, "perplexity": 2655.7905346417265, "lr": 0.0001419757459451063, "grad_norm": 1.295746, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:06.454758+00:00", "epoch": 0, "step": 82, "train_loss": 7.834826469421387, "perplexity": 2527.0969427173254, "lr": 0.0001437285329320829, "grad_norm": 1.269593, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:06.758521+00:00", "epoch": 0, "step": 83, "train_loss": 7.751121520996094, "perplexity": 2324.1775674117116, "lr": 0.00014548131991905952, "grad_norm": 1.30313, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:07.063278+00:00", "epoch": 0, "step": 84, "train_loss": 7.809593200683594, "perplexity": 2464.127825853892, "lr": 0.00014723410690603615, "grad_norm": 1.325459, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:07.367617+00:00", "epoch": 0, "step": 85, "train_loss": 7.736644268035889, "perplexity": 2290.7722528901236, "lr": 0.0001489868938930128, "grad_norm": 1.648035, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:07.671834+00:00", "epoch": 0, "step": 86, "train_loss": 7.681135177612305, "perplexity": 2167.078394974577, "lr": 0.0001507396808799894, "grad_norm": 1.286116, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:07.977129+00:00", "epoch": 0, "step": 87, "train_loss": 7.681574821472168, "perplexity": 2168.031347149207, "lr": 0.000152492467866966, "grad_norm": 1.224935, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:08.281509+00:00", "epoch": 0, "step": 88, "train_loss": 7.624613285064697, "perplexity": 2047.9883241913346, "lr": 0.00015424525485394265, "grad_norm": 1.418285, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:08.586542+00:00", "epoch": 0, "step": 89, "train_loss": 7.657384395599365, "perplexity": 2116.2150020686086, "lr": 0.00015599804184091926, "grad_norm": 1.214827, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:08.892105+00:00", "epoch": 0, "step": 90, "train_loss": 7.59259033203125, "perplexity": 1983.4446454068816, "lr": 0.00015775082882789587, "grad_norm": 1.154242, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:09.196816+00:00", "epoch": 0, "step": 91, "train_loss": 7.618843078613281, "perplexity": 2036.2050374432376, "lr": 0.0001595036158148725, "grad_norm": 1.058164, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:09.500960+00:00", "epoch": 0, "step": 92, "train_loss": 7.579423904418945, "perplexity": 1957.500932854652, "lr": 0.00016125640280184912, "grad_norm": 1.117303, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:09.804912+00:00", "epoch": 0, "step": 93, "train_loss": 7.3864240646362305, "perplexity": 1613.9244920777294, "lr": 0.00016300918978882575, "grad_norm": 0.993616, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:10.108691+00:00", "epoch": 0, "step": 94, "train_loss": 7.412735462188721, "perplexity": 1656.9526846464494, "lr": 0.00016476197677580236, "grad_norm": 1.106894, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:10.412413+00:00", "epoch": 0, "step": 95, "train_loss": 7.48423957824707, "perplexity": 1779.7702790552626, "lr": 0.00016651476376277897, "grad_norm": 1.423123, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:10.716509+00:00", "epoch": 0, "step": 96, "train_loss": 7.4378743171691895, "perplexity": 1699.1345584779829, "lr": 0.0001682675507497556, "grad_norm": 1.212349, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:11.019870+00:00", "epoch": 0, "step": 97, "train_loss": 7.394796848297119, "perplexity": 1627.4942617940435, "lr": 0.00017002033773673225, "grad_norm": 0.973642, "tokens_per_sec": 108017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:11.323374+00:00", "epoch": 0, "step": 98, "train_loss": 7.442110061645508, "perplexity": 1706.346922380457, "lr": 0.00017177312472370883, "grad_norm": 1.064888, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:11.627579+00:00", "epoch": 0, "step": 99, "train_loss": 7.3946027755737305, "perplexity": 1627.1784401975276, "lr": 0.00017352591171068547, "grad_norm": 1.346959, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:11.932396+00:00", "epoch": 0, "step": 100, "train_loss": 7.317108154296875, "perplexity": 1505.8430015651809, "lr": 0.0001752786986976621, "grad_norm": 1.094887, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:12.237405+00:00", "epoch": 0, "step": 101, "train_loss": 7.3666276931762695, "perplexity": 1582.2888123103942, "lr": 0.0001770314856846387, "grad_norm": 0.873449, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:12.542089+00:00", "epoch": 0, "step": 102, "train_loss": 7.309309005737305, "perplexity": 1494.1443872032264, "lr": 0.00017878427267161535, "grad_norm": 1.485402, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:12.845163+00:00", "epoch": 0, "step": 103, "train_loss": 7.262707710266113, "perplexity": 1426.1128143124809, "lr": 0.00018053705965859193, "grad_norm": 0.974768, "tokens_per_sec": 108119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:13.148819+00:00", "epoch": 0, "step": 104, "train_loss": 7.283815383911133, "perplexity": 1456.5346760534824, "lr": 0.00018228984664556857, "grad_norm": 1.255386, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:13.455151+00:00", "epoch": 0, "step": 105, "train_loss": 7.1678786277771, "perplexity": 1297.0900712958655, "lr": 0.0001840426336325452, "grad_norm": 0.809414, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:13.759654+00:00", "epoch": 0, "step": 106, "train_loss": 7.140283584594727, "perplexity": 1261.7861614837725, "lr": 0.00018579542061952182, "grad_norm": 1.122076, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:14.064053+00:00", "epoch": 0, "step": 107, "train_loss": 7.186768531799316, "perplexity": 1321.8248616566934, "lr": 0.00018754820760649845, "grad_norm": 1.86463, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:14.367833+00:00", "epoch": 0, "step": 108, "train_loss": 7.148175239562988, "perplexity": 1271.7831369718042, "lr": 0.00018930099459347504, "grad_norm": 0.815183, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:14.671517+00:00", "epoch": 0, "step": 109, "train_loss": 7.305030822753906, "perplexity": 1487.7658181827037, "lr": 0.00019105378158045167, "grad_norm": 0.884428, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:14.975891+00:00", "epoch": 0, "step": 110, "train_loss": 7.164228916168213, "perplexity": 1292.3646949806043, "lr": 0.00019280656856742828, "grad_norm": 0.93821, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:15.280092+00:00", "epoch": 0, "step": 111, "train_loss": 7.162215232849121, "perplexity": 1289.7649002123144, "lr": 0.00019455935555440492, "grad_norm": 0.716543, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:15.583040+00:00", "epoch": 0, "step": 112, "train_loss": 7.111536979675293, "perplexity": 1226.030482636567, "lr": 0.00019631214254138156, "grad_norm": 0.899815, "tokens_per_sec": 108163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:15.886695+00:00", "epoch": 0, "step": 113, "train_loss": 7.1350626945495605, "perplexity": 1255.2156814756572, "lr": 0.00019806492952835817, "grad_norm": 1.05412, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:16.190548+00:00", "epoch": 0, "step": 114, "train_loss": 7.114076614379883, "perplexity": 1229.148109339984, "lr": 0.00019981771651533478, "grad_norm": 1.31346, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:16.495641+00:00", "epoch": 0, "step": 115, "train_loss": 7.002257347106934, "perplexity": 1099.111436230255, "lr": 0.0002015705035023114, "grad_norm": 0.714031, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:16.801358+00:00", "epoch": 0, "step": 116, "train_loss": 7.078845500946045, "perplexity": 1186.5978013764113, "lr": 0.00020332329048928802, "grad_norm": 1.093038, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:17.104886+00:00", "epoch": 0, "step": 117, "train_loss": 7.004420280456543, "perplexity": 1101.4913138411296, "lr": 0.00020507607747626466, "grad_norm": 0.785827, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:17.408900+00:00", "epoch": 0, "step": 118, "train_loss": 6.968133926391602, "perplexity": 1062.2386843323948, "lr": 0.00020682886446324127, "grad_norm": 0.727059, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:17.713860+00:00", "epoch": 0, "step": 119, "train_loss": 7.032924652099609, "perplexity": 1133.340394207257, "lr": 0.0002085816514502179, "grad_norm": 0.662795, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:18.019685+00:00", "epoch": 0, "step": 120, "train_loss": 7.007027626037598, "perplexity": 1104.3670297142273, "lr": 0.0002103344384371945, "grad_norm": 0.593433, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:18.324126+00:00", "epoch": 0, "step": 121, "train_loss": 6.916301727294922, "perplexity": 1008.5830734661406, "lr": 0.00021208722542417113, "grad_norm": 0.614954, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:18.628136+00:00", "epoch": 0, "step": 122, "train_loss": 6.996830940246582, "perplexity": 1093.1633633197466, "lr": 0.00021384001241114774, "grad_norm": 0.532141, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:18.932999+00:00", "epoch": 0, "step": 123, "train_loss": 6.866720199584961, "perplexity": 959.795460311087, "lr": 0.00021559279939812437, "grad_norm": 0.511661, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:19.238491+00:00", "epoch": 0, "step": 124, "train_loss": 6.924229621887207, "perplexity": 1016.6107931718147, "lr": 0.000217345586385101, "grad_norm": 0.463803, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:19.544664+00:00", "epoch": 0, "step": 125, "train_loss": 6.940345287322998, "perplexity": 1033.126878995436, "lr": 0.0002190983733720776, "grad_norm": 0.463999, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:19.849110+00:00", "epoch": 0, "step": 126, "train_loss": 6.767991542816162, "perplexity": 869.5636575746023, "lr": 0.00022085116035905423, "grad_norm": 0.679888, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:20.153441+00:00", "epoch": 0, "step": 127, "train_loss": 6.8003129959106445, "perplexity": 898.1283581651278, "lr": 0.00022260394734603084, "grad_norm": 2.018801, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:20.457941+00:00", "epoch": 0, "step": 128, "train_loss": 6.811952114105225, "perplexity": 908.6428512950791, "lr": 0.00022435673433300748, "grad_norm": 1.288251, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:20.763006+00:00", "epoch": 0, "step": 129, "train_loss": 6.910219192504883, "perplexity": 1002.4669514522224, "lr": 0.00022610952131998412, "grad_norm": 0.884141, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:21.068901+00:00", "epoch": 0, "step": 130, "train_loss": 6.839709281921387, "perplexity": 934.2175013297599, "lr": 0.0002278623083069607, "grad_norm": 1.317613, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:21.373908+00:00", "epoch": 0, "step": 131, "train_loss": 6.852299690246582, "perplexity": 946.0540381935382, "lr": 0.00022961509529393734, "grad_norm": 0.759548, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:21.678480+00:00", "epoch": 0, "step": 132, "train_loss": 6.909698486328125, "perplexity": 1001.9450965969199, "lr": 0.00023136788228091395, "grad_norm": 0.902156, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:21.984267+00:00", "epoch": 0, "step": 133, "train_loss": 6.705643653869629, "perplexity": 817.003724719355, "lr": 0.00023312066926789058, "grad_norm": 0.974512, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:22.289688+00:00", "epoch": 0, "step": 134, "train_loss": 6.886790752410889, "perplexity": 979.2537014397458, "lr": 0.00023487345625486722, "grad_norm": 0.869525, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:22.594518+00:00", "epoch": 0, "step": 135, "train_loss": 6.751105308532715, "perplexity": 855.0032828608437, "lr": 0.00023662624324184383, "grad_norm": 0.904082, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:22.898492+00:00", "epoch": 0, "step": 136, "train_loss": 6.7570390701293945, "perplexity": 860.091750452306, "lr": 0.00023837903022882047, "grad_norm": 0.890015, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:23.201971+00:00", "epoch": 0, "step": 137, "train_loss": 6.872488498687744, "perplexity": 965.3478461191962, "lr": 0.00024013181721579705, "grad_norm": 1.122429, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:23.507022+00:00", "epoch": 0, "step": 138, "train_loss": 6.798379898071289, "perplexity": 896.3938651889381, "lr": 0.00024188460420277369, "grad_norm": 1.018481, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:23.813072+00:00", "epoch": 0, "step": 139, "train_loss": 6.674650192260742, "perplexity": 792.0703330762406, "lr": 0.0002436373911897503, "grad_norm": 0.890306, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:24.117967+00:00", "epoch": 0, "step": 140, "train_loss": 6.680258750915527, "perplexity": 796.525186975084, "lr": 0.0002453901781767269, "grad_norm": 0.686828, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:24.421340+00:00", "epoch": 0, "step": 141, "train_loss": 6.6623735427856445, "perplexity": 782.4058086100466, "lr": 0.00024714296516370357, "grad_norm": 0.721464, "tokens_per_sec": 107947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:24.725047+00:00", "epoch": 0, "step": 142, "train_loss": 6.579498767852783, "perplexity": 720.1782622707541, "lr": 0.0002488957521506802, "grad_norm": 0.784962, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:25.029403+00:00", "epoch": 0, "step": 143, "train_loss": 6.589237213134766, "perplexity": 727.2259398840708, "lr": 0.0002506485391376568, "grad_norm": 0.891665, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:25.336278+00:00", "epoch": 0, "step": 144, "train_loss": 6.674158573150635, "perplexity": 791.6810318657493, "lr": 0.0002524013261246334, "grad_norm": 1.403392, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:25.641247+00:00", "epoch": 0, "step": 145, "train_loss": 6.588869571685791, "perplexity": 726.9586306257924, "lr": 0.00025415411311161, "grad_norm": 0.934459, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:25.947290+00:00", "epoch": 0, "step": 146, "train_loss": 6.497910976409912, "perplexity": 663.7535868176268, "lr": 0.0002559069000985867, "grad_norm": 1.111638, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:26.252379+00:00", "epoch": 0, "step": 147, "train_loss": 6.515249729156494, "perplexity": 675.3625981975181, "lr": 0.0002576596870855633, "grad_norm": 1.116096, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:26.557303+00:00", "epoch": 0, "step": 148, "train_loss": 6.509013652801514, "perplexity": 671.1640901823232, "lr": 0.0002594124740725399, "grad_norm": 0.818035, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:26.862142+00:00", "epoch": 0, "step": 149, "train_loss": 6.599158763885498, "perplexity": 734.4770606536848, "lr": 0.0002611652610595165, "grad_norm": 0.895285, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:27.168043+00:00", "epoch": 0, "step": 150, "train_loss": 6.542880058288574, "perplexity": 694.2832776591089, "lr": 0.0002629180480464931, "grad_norm": 0.79717, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:27.472394+00:00", "epoch": 0, "step": 151, "train_loss": 6.556065559387207, "perplexity": 703.4983698747376, "lr": 0.0002646708350334697, "grad_norm": 0.784444, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:27.776899+00:00", "epoch": 0, "step": 152, "train_loss": 6.470719814300537, "perplexity": 645.9485226457152, "lr": 0.0002664236220204464, "grad_norm": 0.564142, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:28.081774+00:00", "epoch": 0, "step": 153, "train_loss": 6.5576677322387695, "perplexity": 704.6263992718069, "lr": 0.000268176409007423, "grad_norm": 0.955866, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:28.386306+00:00", "epoch": 0, "step": 154, "train_loss": 6.544604301452637, "perplexity": 695.4814235050857, "lr": 0.0002699291959943996, "grad_norm": 1.57807, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:28.692159+00:00", "epoch": 0, "step": 155, "train_loss": 6.414128303527832, "perplexity": 610.4084379578709, "lr": 0.00027168198298137627, "grad_norm": 1.001148, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:28.996177+00:00", "epoch": 0, "step": 156, "train_loss": 6.484420299530029, "perplexity": 654.8592319796861, "lr": 0.00027343476996835283, "grad_norm": 1.100857, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:29.300446+00:00", "epoch": 0, "step": 157, "train_loss": 6.523062229156494, "perplexity": 680.6595327047804, "lr": 0.0002751875569553295, "grad_norm": 1.270184, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:29.604623+00:00", "epoch": 0, "step": 158, "train_loss": 6.706392288208008, "perplexity": 817.6155907656529, "lr": 0.0002769403439423061, "grad_norm": 1.008957, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:29.910458+00:00", "epoch": 0, "step": 159, "train_loss": 6.542902946472168, "perplexity": 694.299168724092, "lr": 0.0002786931309292827, "grad_norm": 1.0861, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:30.216927+00:00", "epoch": 0, "step": 160, "train_loss": 6.381094455718994, "perplexity": 590.5737108583733, "lr": 0.0002804459179162594, "grad_norm": 0.951699, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:30.522033+00:00", "epoch": 0, "step": 161, "train_loss": 6.460882663726807, "perplexity": 639.6253816013807, "lr": 0.00028219870490323593, "grad_norm": 1.120674, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:30.826697+00:00", "epoch": 0, "step": 162, "train_loss": 6.522120952606201, "perplexity": 680.0191452860242, "lr": 0.0002839514918902126, "grad_norm": 1.058402, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:31.131881+00:00", "epoch": 0, "step": 163, "train_loss": 6.568233489990234, "perplexity": 712.1107805650381, "lr": 0.0002857042788771892, "grad_norm": 1.620279, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:31.437753+00:00", "epoch": 0, "step": 164, "train_loss": 6.412965297698975, "perplexity": 609.6989420404569, "lr": 0.0002874570658641658, "grad_norm": 0.989268, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:31.743729+00:00", "epoch": 0, "step": 165, "train_loss": 6.446639060974121, "perplexity": 630.5793884589667, "lr": 0.0002892098528511425, "grad_norm": 0.996291, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:32.049712+00:00", "epoch": 0, "step": 166, "train_loss": 6.45571231842041, "perplexity": 636.3268321797208, "lr": 0.00029096263983811904, "grad_norm": 0.983576, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:32.354252+00:00", "epoch": 0, "step": 167, "train_loss": 6.393120288848877, "perplexity": 597.7187280386596, "lr": 0.0002927154268250957, "grad_norm": 1.249266, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:32.658565+00:00", "epoch": 0, "step": 168, "train_loss": 6.43411922454834, "perplexity": 622.7338525547169, "lr": 0.0002944682138120723, "grad_norm": 1.099393, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:32.964294+00:00", "epoch": 0, "step": 169, "train_loss": 6.335169315338135, "perplexity": 564.0648997395379, "lr": 0.0002962210007990489, "grad_norm": 1.066089, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:33.271360+00:00", "epoch": 0, "step": 170, "train_loss": 6.287862777709961, "perplexity": 538.0022690457039, "lr": 0.0002979737877860256, "grad_norm": 1.156883, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:33.576349+00:00", "epoch": 0, "step": 171, "train_loss": 6.437729835510254, "perplexity": 624.9863662573631, "lr": 0.0002997265747730022, "grad_norm": 1.251636, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:33.882442+00:00", "epoch": 0, "step": 172, "train_loss": 6.281320095062256, "perplexity": 534.4937809133608, "lr": 0.0003014793617599788, "grad_norm": 1.134171, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:34.187232+00:00", "epoch": 0, "step": 173, "train_loss": 6.262197494506836, "perplexity": 524.3699750880896, "lr": 0.0003032321487469554, "grad_norm": 0.858386, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:34.492465+00:00", "epoch": 0, "step": 174, "train_loss": 6.316552639007568, "perplexity": 553.6610292959059, "lr": 0.000304984935733932, "grad_norm": 1.449097, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:34.797922+00:00", "epoch": 0, "step": 175, "train_loss": 6.336670875549316, "perplexity": 564.9125133616973, "lr": 0.0003067377227209087, "grad_norm": 0.849848, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:35.102791+00:00", "epoch": 0, "step": 176, "train_loss": 6.152130126953125, "perplexity": 469.7168784655594, "lr": 0.0003084905097078853, "grad_norm": 1.110948, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:35.408529+00:00", "epoch": 0, "step": 177, "train_loss": 6.274374961853027, "perplexity": 530.7945112352313, "lr": 0.0003102432966948619, "grad_norm": 1.032792, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:35.716468+00:00", "epoch": 0, "step": 178, "train_loss": 6.192694664001465, "perplexity": 489.1624603246008, "lr": 0.0003119960836818385, "grad_norm": 0.976156, "tokens_per_sec": 106410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:36.021571+00:00", "epoch": 0, "step": 179, "train_loss": 6.382201671600342, "perplexity": 591.2279655839143, "lr": 0.0003137488706688151, "grad_norm": 1.320319, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:36.326830+00:00", "epoch": 0, "step": 180, "train_loss": 6.240601062774658, "perplexity": 513.1668637628408, "lr": 0.00031550165765579174, "grad_norm": 1.327497, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:36.632028+00:00", "epoch": 0, "step": 181, "train_loss": 6.411595821380615, "perplexity": 608.8645452517317, "lr": 0.0003172544446427684, "grad_norm": 1.318792, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:36.937158+00:00", "epoch": 0, "step": 182, "train_loss": 6.188788414001465, "perplexity": 487.2553966227068, "lr": 0.000319007231629745, "grad_norm": 0.934941, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:37.242948+00:00", "epoch": 0, "step": 183, "train_loss": 6.235238075256348, "perplexity": 510.4221228618664, "lr": 0.0003207600186167216, "grad_norm": 1.34252, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:37.549073+00:00", "epoch": 0, "step": 184, "train_loss": 6.326612949371338, "perplexity": 559.2591432536059, "lr": 0.00032251280560369823, "grad_norm": 1.247909, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:37.854483+00:00", "epoch": 0, "step": 185, "train_loss": 6.220242977142334, "perplexity": 502.8253922554687, "lr": 0.00032426559259067484, "grad_norm": 0.719663, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:38.161433+00:00", "epoch": 0, "step": 186, "train_loss": 6.182377338409424, "perplexity": 484.1415576362722, "lr": 0.0003260183795776515, "grad_norm": 0.862208, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:38.466895+00:00", "epoch": 0, "step": 187, "train_loss": 6.2852067947387695, "perplexity": 536.5752401018276, "lr": 0.0003277711665646281, "grad_norm": 1.073832, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:38.773409+00:00", "epoch": 0, "step": 188, "train_loss": 6.4159111976623535, "perplexity": 611.4977023145466, "lr": 0.0003295239535516047, "grad_norm": 1.2788, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:39.078487+00:00", "epoch": 0, "step": 189, "train_loss": 6.185566425323486, "perplexity": 485.6879916879446, "lr": 0.00033127674053858133, "grad_norm": 1.463256, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:39.384094+00:00", "epoch": 0, "step": 190, "train_loss": 6.2865705490112305, "perplexity": 537.3074960731288, "lr": 0.00033302952752555794, "grad_norm": 1.133373, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:39.691624+00:00", "epoch": 0, "step": 191, "train_loss": 6.260982036590576, "perplexity": 523.7330126297684, "lr": 0.0003347823145125346, "grad_norm": 1.046754, "tokens_per_sec": 106497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:39.997999+00:00", "epoch": 0, "step": 192, "train_loss": 6.129030227661133, "perplexity": 458.9908281703076, "lr": 0.0003365351014995112, "grad_norm": 1.037273, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:40.303780+00:00", "epoch": 0, "step": 193, "train_loss": 6.01025390625, "perplexity": 407.58679597169703, "lr": 0.0003382878884864879, "grad_norm": 1.133863, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:40.609711+00:00", "epoch": 0, "step": 194, "train_loss": 6.170242786407471, "perplexity": 478.3022172705986, "lr": 0.0003400406754734645, "grad_norm": 0.939359, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:40.913825+00:00", "epoch": 0, "step": 195, "train_loss": 6.164462566375732, "perplexity": 475.54550010437185, "lr": 0.00034179346246044105, "grad_norm": 1.385986, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:41.218320+00:00", "epoch": 0, "step": 196, "train_loss": 6.125377178192139, "perplexity": 457.3171708069721, "lr": 0.00034354624944741766, "grad_norm": 1.347895, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:41.523832+00:00", "epoch": 0, "step": 197, "train_loss": 6.176604270935059, "perplexity": 481.35462806342855, "lr": 0.0003452990364343943, "grad_norm": 1.082201, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:41.830051+00:00", "epoch": 0, "step": 198, "train_loss": 6.130643844604492, "perplexity": 459.7320614199223, "lr": 0.00034705182342137093, "grad_norm": 1.099318, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:42.135098+00:00", "epoch": 0, "step": 199, "train_loss": 6.105199813842773, "perplexity": 448.1821855137626, "lr": 0.00034880461040834754, "grad_norm": 1.175997, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:42.440535+00:00", "epoch": 0, "step": 200, "train_loss": 6.136470794677734, "perplexity": 462.41871709137394, "lr": 0.0003505573973953242, "grad_norm": 1.308471, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:42.745601+00:00", "epoch": 0, "step": 201, "train_loss": 6.223110675811768, "perplexity": 504.2694134830746, "lr": 0.0003523101843823008, "grad_norm": 0.960268, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:43.051319+00:00", "epoch": 0, "step": 202, "train_loss": 5.962231636047363, "perplexity": 388.4760948696022, "lr": 0.0003540629713692774, "grad_norm": 1.525324, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:43.357599+00:00", "epoch": 0, "step": 203, "train_loss": 6.013891220092773, "perplexity": 409.0720165360139, "lr": 0.000355815758356254, "grad_norm": 0.871423, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:43.662950+00:00", "epoch": 0, "step": 204, "train_loss": 6.05271053314209, "perplexity": 425.26416185437375, "lr": 0.0003575685453432307, "grad_norm": 1.148014, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:43.967616+00:00", "epoch": 0, "step": 205, "train_loss": 6.164854049682617, "perplexity": 475.73170467493924, "lr": 0.00035932133233020726, "grad_norm": 1.155222, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:44.273196+00:00", "epoch": 0, "step": 206, "train_loss": 6.006438732147217, "perplexity": 406.03474394000904, "lr": 0.00036107411931718387, "grad_norm": 1.272578, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:44.578389+00:00", "epoch": 0, "step": 207, "train_loss": 6.022676467895508, "perplexity": 412.6816481082751, "lr": 0.00036282690630416053, "grad_norm": 1.66726, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:44.884754+00:00", "epoch": 0, "step": 208, "train_loss": 5.926850318908691, "perplexity": 374.97161092649543, "lr": 0.00036457969329113714, "grad_norm": 1.180076, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:45.191420+00:00", "epoch": 0, "step": 209, "train_loss": 6.1882195472717285, "perplexity": 486.97829206393857, "lr": 0.00036633248027811375, "grad_norm": 1.769935, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:45.497919+00:00", "epoch": 0, "step": 210, "train_loss": 5.9612507820129395, "perplexity": 388.0952433350882, "lr": 0.0003680852672650904, "grad_norm": 0.972495, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:45.807113+00:00", "epoch": 0, "step": 211, "train_loss": 6.035235404968262, "perplexity": 417.89717313102557, "lr": 0.000369838054252067, "grad_norm": 1.099339, "tokens_per_sec": 105922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:46.112864+00:00", "epoch": 0, "step": 212, "train_loss": 6.036691188812256, "perplexity": 418.50598412520384, "lr": 0.00037159084123904363, "grad_norm": 0.938291, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:46.418706+00:00", "epoch": 0, "step": 213, "train_loss": 5.997960567474365, "perplexity": 402.6068661071163, "lr": 0.00037334362822602024, "grad_norm": 1.434096, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:46.724145+00:00", "epoch": 0, "step": 214, "train_loss": 6.082797050476074, "perplexity": 438.2532985776142, "lr": 0.0003750964152129969, "grad_norm": 1.087498, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:47.028910+00:00", "epoch": 0, "step": 215, "train_loss": 6.00768518447876, "perplexity": 406.54116244097503, "lr": 0.0003768492021999735, "grad_norm": 0.965096, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:47.335615+00:00", "epoch": 0, "step": 216, "train_loss": 6.098649501800537, "perplexity": 445.25604637165225, "lr": 0.0003786019891869501, "grad_norm": 1.174209, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:47.641795+00:00", "epoch": 0, "step": 217, "train_loss": 6.002516746520996, "perplexity": 404.44540023937407, "lr": 0.0003803547761739268, "grad_norm": 1.259356, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:47.947902+00:00", "epoch": 0, "step": 218, "train_loss": 5.762890338897705, "perplexity": 318.2668999884083, "lr": 0.00038210756316090335, "grad_norm": 0.667622, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:48.253272+00:00", "epoch": 0, "step": 219, "train_loss": 5.983381271362305, "perplexity": 396.779722431727, "lr": 0.00038386035014787996, "grad_norm": 0.947472, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:48.559323+00:00", "epoch": 0, "step": 220, "train_loss": 5.931420803070068, "perplexity": 376.6893351607301, "lr": 0.00038561313713485657, "grad_norm": 0.984883, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:48.866042+00:00", "epoch": 0, "step": 221, "train_loss": 5.8755269050598145, "perplexity": 356.2123012959615, "lr": 0.00038736592412183323, "grad_norm": 1.110767, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:49.171856+00:00", "epoch": 0, "step": 222, "train_loss": 5.890069007873535, "perplexity": 361.43022504298636, "lr": 0.00038911871110880984, "grad_norm": 1.667441, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:49.477115+00:00", "epoch": 0, "step": 223, "train_loss": 5.89124059677124, "perplexity": 361.8539208321186, "lr": 0.00039087149809578645, "grad_norm": 0.948357, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:49.781547+00:00", "epoch": 0, "step": 224, "train_loss": 5.982260704040527, "perplexity": 396.3353530602229, "lr": 0.0003926242850827631, "grad_norm": 1.068543, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:50.087190+00:00", "epoch": 0, "step": 225, "train_loss": 5.985705375671387, "perplexity": 397.7029523201021, "lr": 0.0003943770720697397, "grad_norm": 0.8394, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:50.392784+00:00", "epoch": 0, "step": 226, "train_loss": 5.8460235595703125, "perplexity": 345.85636526310276, "lr": 0.00039612985905671633, "grad_norm": 1.411471, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:50.698344+00:00", "epoch": 0, "step": 227, "train_loss": 5.972572326660156, "perplexity": 392.514047605671, "lr": 0.000397882646043693, "grad_norm": 1.057194, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:51.002988+00:00", "epoch": 0, "step": 228, "train_loss": 6.012648582458496, "perplexity": 408.5640039562822, "lr": 0.00039963543303066955, "grad_norm": 0.9796, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:51.307709+00:00", "epoch": 0, "step": 229, "train_loss": 5.911764621734619, "perplexity": 369.3573566978731, "lr": 0.00040138822001764616, "grad_norm": 1.491403, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:51.613489+00:00", "epoch": 0, "step": 230, "train_loss": 5.966434001922607, "perplexity": 390.11204858450145, "lr": 0.0004031410070046228, "grad_norm": 1.225319, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:51.920469+00:00", "epoch": 0, "step": 231, "train_loss": 5.87833833694458, "perplexity": 357.21517701538454, "lr": 0.00040489379399159944, "grad_norm": 0.975337, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:52.226450+00:00", "epoch": 0, "step": 232, "train_loss": 5.913355827331543, "perplexity": 369.94554803368334, "lr": 0.00040664658097857605, "grad_norm": 1.224939, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:52.532415+00:00", "epoch": 0, "step": 233, "train_loss": 5.85775899887085, "perplexity": 349.93905084305953, "lr": 0.00040839936796555266, "grad_norm": 1.16501, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:52.838531+00:00", "epoch": 0, "step": 234, "train_loss": 5.785900592803955, "perplexity": 325.6752088294971, "lr": 0.0004101521549525293, "grad_norm": 1.142045, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:53.143918+00:00", "epoch": 0, "step": 235, "train_loss": 5.87370491027832, "perplexity": 355.56387523572494, "lr": 0.00041190494193950593, "grad_norm": 1.260571, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:53.451637+00:00", "epoch": 0, "step": 236, "train_loss": 5.897266864776611, "perplexity": 364.0411332828408, "lr": 0.00041365772892648254, "grad_norm": 0.753697, "tokens_per_sec": 106488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:53.756782+00:00", "epoch": 0, "step": 237, "train_loss": 5.812920570373535, "perplexity": 334.59490803598885, "lr": 0.0004154105159134591, "grad_norm": 1.084927, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:54.064114+00:00", "epoch": 0, "step": 238, "train_loss": 5.913315773010254, "perplexity": 369.9307304125998, "lr": 0.0004171633029004358, "grad_norm": 1.536799, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:54.370847+00:00", "epoch": 0, "step": 239, "train_loss": 5.821900844573975, "perplexity": 337.61319429112046, "lr": 0.00041891608988741237, "grad_norm": 0.830213, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:54.676324+00:00", "epoch": 0, "step": 240, "train_loss": 5.84952449798584, "perplexity": 347.06930908043273, "lr": 0.000420668876874389, "grad_norm": 1.171973, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:54.980890+00:00", "epoch": 0, "step": 241, "train_loss": 5.863598346710205, "perplexity": 351.9884444211414, "lr": 0.00042242166386136565, "grad_norm": 1.331555, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:55.287453+00:00", "epoch": 0, "step": 242, "train_loss": 6.02234411239624, "perplexity": 412.5445138829988, "lr": 0.00042417445084834226, "grad_norm": 1.250112, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:55.593376+00:00", "epoch": 0, "step": 243, "train_loss": 5.745365619659424, "perplexity": 312.7379500723456, "lr": 0.00042592723783531887, "grad_norm": 0.934292, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:55.899872+00:00", "epoch": 0, "step": 244, "train_loss": 5.800647735595703, "perplexity": 330.51357599728345, "lr": 0.0004276800248222955, "grad_norm": 1.120736, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:56.206930+00:00", "epoch": 0, "step": 245, "train_loss": 5.871583938598633, "perplexity": 354.81053351666594, "lr": 0.00042943281180927214, "grad_norm": 1.062695, "tokens_per_sec": 106716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:56.513542+00:00", "epoch": 0, "step": 246, "train_loss": 5.895016670227051, "perplexity": 363.22289085649163, "lr": 0.00043118559879624875, "grad_norm": 1.11661, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:56.903977+00:00", "epoch": 0, "step": 247, "train_loss": 5.863853931427002, "perplexity": 352.07841878557105, "lr": 0.00043293838578322536, "grad_norm": 1.013939, "tokens_per_sec": 83926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:57.211743+00:00", "epoch": 0, "step": 248, "train_loss": 5.753660202026367, "perplexity": 315.3427687657939, "lr": 0.000434691172770202, "grad_norm": 0.967896, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:57.517664+00:00", "epoch": 0, "step": 249, "train_loss": 5.750884056091309, "perplexity": 314.468545267609, "lr": 0.00043644395975717863, "grad_norm": 1.247413, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:57.823874+00:00", "epoch": 0, "step": 250, "train_loss": 5.8320136070251465, "perplexity": 341.04471817377294, "lr": 0.0004381967467441552, "grad_norm": 1.012595, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:58.130614+00:00", "epoch": 0, "step": 251, "train_loss": 5.8039984703063965, "perplexity": 331.62289679067584, "lr": 0.0004399495337311319, "grad_norm": 0.795305, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:58.437421+00:00", "epoch": 0, "step": 252, "train_loss": 5.839847564697266, "perplexity": 343.7269405510135, "lr": 0.00044170232071810846, "grad_norm": 0.909271, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:58.743612+00:00", "epoch": 0, "step": 253, "train_loss": 5.905977725982666, "perplexity": 367.22609681648396, "lr": 0.0004434551077050851, "grad_norm": 1.024007, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:59.049265+00:00", "epoch": 0, "step": 254, "train_loss": 5.876565456390381, "perplexity": 356.58243822526794, "lr": 0.0004452078946920617, "grad_norm": 0.842925, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:59.355685+00:00", "epoch": 0, "step": 255, "train_loss": 5.720139980316162, "perplexity": 304.9476066317909, "lr": 0.00044696068167903835, "grad_norm": 1.235116, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:59.664408+00:00", "epoch": 0, "step": 256, "train_loss": 5.767520427703857, "perplexity": 319.7439207293505, "lr": 0.00044871346866601496, "grad_norm": 0.804542, "tokens_per_sec": 106141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:58:59.971688+00:00", "epoch": 0, "step": 257, "train_loss": 5.729555606842041, "perplexity": 307.83243934687664, "lr": 0.00045046625565299157, "grad_norm": 1.068997, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:00.278942+00:00", "epoch": 0, "step": 258, "train_loss": 5.759598731994629, "perplexity": 317.221012732562, "lr": 0.00045221904263996823, "grad_norm": 1.08113, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:00.585378+00:00", "epoch": 0, "step": 259, "train_loss": 5.743117332458496, "perplexity": 312.0356151630938, "lr": 0.00045397182962694484, "grad_norm": 1.008028, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:00.891149+00:00", "epoch": 0, "step": 260, "train_loss": 5.6836442947387695, "perplexity": 294.01897166771545, "lr": 0.0004557246166139214, "grad_norm": 1.329615, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:01.196444+00:00", "epoch": 0, "step": 261, "train_loss": 5.83113956451416, "perplexity": 340.74676082460525, "lr": 0.000457477403600898, "grad_norm": 1.260699, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:01.502499+00:00", "epoch": 0, "step": 262, "train_loss": 5.805451393127441, "perplexity": 332.1050694604202, "lr": 0.00045923019058787467, "grad_norm": 1.112032, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:01.808746+00:00", "epoch": 0, "step": 263, "train_loss": 5.74357271194458, "perplexity": 312.1777421395653, "lr": 0.0004609829775748513, "grad_norm": 1.172436, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:02.115529+00:00", "epoch": 0, "step": 264, "train_loss": 5.789430141448975, "perplexity": 326.8267262945526, "lr": 0.0004627357645618279, "grad_norm": 1.091344, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:02.421085+00:00", "epoch": 0, "step": 265, "train_loss": 5.692163467407227, "perplexity": 296.53446982294275, "lr": 0.00046448855154880456, "grad_norm": 1.279556, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:02.726367+00:00", "epoch": 0, "step": 266, "train_loss": 5.762818336486816, "perplexity": 318.24398482928626, "lr": 0.00046624133853578117, "grad_norm": 0.752011, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:03.032069+00:00", "epoch": 0, "step": 267, "train_loss": 5.7405595779418945, "perplexity": 311.23852447477833, "lr": 0.0004679941255227578, "grad_norm": 1.077693, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:03.340190+00:00", "epoch": 0, "step": 268, "train_loss": 5.63274621963501, "perplexity": 279.4284367707348, "lr": 0.00046974691250973444, "grad_norm": 1.384682, "tokens_per_sec": 106347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:03.646717+00:00", "epoch": 0, "step": 269, "train_loss": 5.666731357574463, "perplexity": 289.0880628888822, "lr": 0.00047149969949671105, "grad_norm": 0.804382, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:03.952960+00:00", "epoch": 0, "step": 270, "train_loss": 5.674915313720703, "perplexity": 291.4636545276837, "lr": 0.00047325248648368766, "grad_norm": 1.11849, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:04.257950+00:00", "epoch": 0, "step": 271, "train_loss": 5.5997090339660645, "perplexity": 270.34773397311875, "lr": 0.0004750052734706642, "grad_norm": 1.368282, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:04.563587+00:00", "epoch": 0, "step": 272, "train_loss": 5.566479206085205, "perplexity": 261.51174751859196, "lr": 0.00047675806045764093, "grad_norm": 0.954414, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:04.870554+00:00", "epoch": 0, "step": 273, "train_loss": 5.606919288635254, "perplexity": 272.30405429277556, "lr": 0.0004785108474446175, "grad_norm": 1.35085, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:05.176361+00:00", "epoch": 0, "step": 274, "train_loss": 5.5981125831604, "perplexity": 269.91648144372476, "lr": 0.0004802636344315941, "grad_norm": 1.253533, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:05.482040+00:00", "epoch": 0, "step": 275, "train_loss": 5.766750335693359, "perplexity": 319.49778327699846, "lr": 0.00048201642141857076, "grad_norm": 1.244444, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:05.790195+00:00", "epoch": 0, "step": 276, "train_loss": 5.648350715637207, "perplexity": 283.82297479851985, "lr": 0.00048376920840554737, "grad_norm": 0.91033, "tokens_per_sec": 106337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:06.095305+00:00", "epoch": 0, "step": 277, "train_loss": 5.547250747680664, "perplexity": 256.5313162189651, "lr": 0.000485521995392524, "grad_norm": 1.290711, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:06.401011+00:00", "epoch": 0, "step": 278, "train_loss": 5.6827592849731445, "perplexity": 293.7588771165958, "lr": 0.0004872747823795006, "grad_norm": 1.215568, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:06.707058+00:00", "epoch": 0, "step": 279, "train_loss": 5.519473552703857, "perplexity": 249.50365208604353, "lr": 0.0004890275693664772, "grad_norm": 1.049332, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:07.014231+00:00", "epoch": 0, "step": 280, "train_loss": 5.727803707122803, "perplexity": 307.2936198994417, "lr": 0.0004907803563534538, "grad_norm": 1.045655, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:07.319911+00:00", "epoch": 0, "step": 281, "train_loss": 5.6386399269104, "perplexity": 281.080168812877, "lr": 0.0004925331433404304, "grad_norm": 0.892098, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:07.624939+00:00", "epoch": 0, "step": 282, "train_loss": 5.460489273071289, "perplexity": 235.21247934846716, "lr": 0.0004942859303274071, "grad_norm": 0.840288, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:07.930610+00:00", "epoch": 0, "step": 283, "train_loss": 5.639986515045166, "perplexity": 281.45892298882507, "lr": 0.0004960387173143838, "grad_norm": 0.945077, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:08.236997+00:00", "epoch": 0, "step": 284, "train_loss": 5.6223554611206055, "perplexity": 276.53999589618337, "lr": 0.0004977915043013604, "grad_norm": 0.785594, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:08.543445+00:00", "epoch": 0, "step": 285, "train_loss": 5.693501949310303, "perplexity": 296.93164158875396, "lr": 0.000499544291288337, "grad_norm": 0.760417, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:08.848616+00:00", "epoch": 0, "step": 286, "train_loss": 5.547786235809326, "perplexity": 256.66872247985776, "lr": 0.0005012970782753136, "grad_norm": 0.944357, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:09.153774+00:00", "epoch": 0, "step": 287, "train_loss": 5.48287296295166, "perplexity": 240.5367688645247, "lr": 0.0005030498652622902, "grad_norm": 1.109546, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:09.459201+00:00", "epoch": 0, "step": 288, "train_loss": 5.5759501457214355, "perplexity": 264.000275239975, "lr": 0.0005048026522492668, "grad_norm": 0.962084, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:09.766021+00:00", "epoch": 0, "step": 289, "train_loss": 5.5922441482543945, "perplexity": 268.3371328271235, "lr": 0.0005065554392362435, "grad_norm": 1.122486, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:10.071387+00:00", "epoch": 0, "step": 290, "train_loss": 5.505184173583984, "perplexity": 245.9637515331941, "lr": 0.00050830822622322, "grad_norm": 1.295322, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:10.377587+00:00", "epoch": 0, "step": 291, "train_loss": 5.589737415313721, "perplexity": 267.6653256694126, "lr": 0.0005100610132101966, "grad_norm": 0.938109, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:10.682841+00:00", "epoch": 0, "step": 292, "train_loss": 5.556593894958496, "perplexity": 258.9393579145271, "lr": 0.0005118138001971733, "grad_norm": 1.17618, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:10.987114+00:00", "epoch": 0, "step": 293, "train_loss": 5.5735039710998535, "perplexity": 263.3552736814424, "lr": 0.00051356658718415, "grad_norm": 1.222015, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:11.292974+00:00", "epoch": 0, "step": 294, "train_loss": 5.65186882019043, "perplexity": 284.8232522070128, "lr": 0.0005153193741711266, "grad_norm": 0.830881, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:11.599409+00:00", "epoch": 0, "step": 295, "train_loss": 5.567989349365234, "perplexity": 261.90696606996653, "lr": 0.0005170721611581032, "grad_norm": 0.995237, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:11.905536+00:00", "epoch": 0, "step": 296, "train_loss": 5.61421012878418, "perplexity": 274.29663459022805, "lr": 0.0005188249481450798, "grad_norm": 1.14997, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:12.211708+00:00", "epoch": 0, "step": 297, "train_loss": 5.640550136566162, "perplexity": 281.6176040089162, "lr": 0.0005205777351320564, "grad_norm": 1.267639, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:12.517729+00:00", "epoch": 0, "step": 298, "train_loss": 5.489238739013672, "perplexity": 242.07285607614818, "lr": 0.000522330522119033, "grad_norm": 1.260256, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:12.824172+00:00", "epoch": 0, "step": 299, "train_loss": 5.484011173248291, "perplexity": 240.81070616111612, "lr": 0.0005240833091060097, "grad_norm": 1.186518, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:13.130476+00:00", "epoch": 0, "step": 300, "train_loss": 5.504364013671875, "perplexity": 245.76210462701232, "lr": 0.0005258360960929862, "grad_norm": 0.81817, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:13.436419+00:00", "epoch": 0, "step": 301, "train_loss": 5.575558662414551, "perplexity": 263.8969437667987, "lr": 0.0005275888830799628, "grad_norm": 0.793447, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:13.741354+00:00", "epoch": 0, "step": 302, "train_loss": 5.503415584564209, "perplexity": 245.52912719217707, "lr": 0.0005293416700669394, "grad_norm": 1.116902, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:14.046348+00:00", "epoch": 0, "step": 303, "train_loss": 5.58029842376709, "perplexity": 265.15072125775555, "lr": 0.0005310944570539162, "grad_norm": 0.927655, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:14.352397+00:00", "epoch": 0, "step": 304, "train_loss": 5.507343769073486, "perplexity": 246.49550772405246, "lr": 0.0005328472440408928, "grad_norm": 0.877423, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:14.658608+00:00", "epoch": 0, "step": 305, "train_loss": 5.513764381408691, "perplexity": 248.0832515107289, "lr": 0.0005346000310278694, "grad_norm": 0.928125, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:14.965597+00:00", "epoch": 0, "step": 306, "train_loss": 5.404315948486328, "perplexity": 222.36405997512378, "lr": 0.000536352818014846, "grad_norm": 0.778509, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:15.271262+00:00", "epoch": 0, "step": 307, "train_loss": 5.587752342224121, "perplexity": 267.13451745526294, "lr": 0.0005381056050018226, "grad_norm": 0.787903, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:15.577281+00:00", "epoch": 0, "step": 308, "train_loss": 5.487114429473877, "perplexity": 241.5591642121016, "lr": 0.0005398583919887992, "grad_norm": 0.841219, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:15.882937+00:00", "epoch": 0, "step": 309, "train_loss": 5.385982513427734, "perplexity": 218.32450552236696, "lr": 0.0005416111789757759, "grad_norm": 1.127842, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:16.188695+00:00", "epoch": 0, "step": 310, "train_loss": 5.575284481048584, "perplexity": 263.8245980606577, "lr": 0.0005433639659627525, "grad_norm": 1.371189, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:16.495531+00:00", "epoch": 0, "step": 311, "train_loss": 5.51532506942749, "perplexity": 248.47073436029592, "lr": 0.000545116752949729, "grad_norm": 0.983989, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:16.803017+00:00", "epoch": 0, "step": 312, "train_loss": 5.463471412658691, "perplexity": 235.91496272571987, "lr": 0.0005468695399367057, "grad_norm": 1.331901, "tokens_per_sec": 106568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:17.108992+00:00", "epoch": 0, "step": 313, "train_loss": 5.460506916046143, "perplexity": 235.2166292329336, "lr": 0.0005486223269236824, "grad_norm": 0.830106, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:17.414645+00:00", "epoch": 0, "step": 314, "train_loss": 5.358389377593994, "perplexity": 212.38260265178064, "lr": 0.000550375113910659, "grad_norm": 0.790011, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:17.720261+00:00", "epoch": 0, "step": 315, "train_loss": 5.367069721221924, "perplexity": 214.23418116616836, "lr": 0.0005521279008976356, "grad_norm": 0.679354, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:18.026318+00:00", "epoch": 0, "step": 316, "train_loss": 5.383559703826904, "perplexity": 217.7961870802239, "lr": 0.0005538806878846122, "grad_norm": 0.768779, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:18.331073+00:00", "epoch": 0, "step": 317, "train_loss": 5.365716934204102, "perplexity": 213.9445638865284, "lr": 0.0005556334748715888, "grad_norm": 0.89925, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:18.637462+00:00", "epoch": 0, "step": 318, "train_loss": 5.551537036895752, "perplexity": 257.63324353551764, "lr": 0.0005573862618585654, "grad_norm": 1.017837, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:18.942796+00:00", "epoch": 0, "step": 319, "train_loss": 5.504085540771484, "perplexity": 245.69367606912243, "lr": 0.000559139048845542, "grad_norm": 1.00038, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:19.249273+00:00", "epoch": 0, "step": 320, "train_loss": 5.506818771362305, "perplexity": 246.36613211060282, "lr": 0.0005608918358325188, "grad_norm": 1.236331, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:19.556097+00:00", "epoch": 0, "step": 321, "train_loss": 5.511641979217529, "perplexity": 247.55727743583793, "lr": 0.0005626446228194953, "grad_norm": 0.911635, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:19.862950+00:00", "epoch": 0, "step": 322, "train_loss": 5.5089616775512695, "perplexity": 246.89463768654522, "lr": 0.0005643974098064719, "grad_norm": 0.854575, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:20.169929+00:00", "epoch": 0, "step": 323, "train_loss": 5.44044303894043, "perplexity": 230.54430094077256, "lr": 0.0005661501967934486, "grad_norm": 0.825845, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:20.475999+00:00", "epoch": 0, "step": 324, "train_loss": 5.553829669952393, "perplexity": 258.224579625457, "lr": 0.0005679029837804252, "grad_norm": 0.945011, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:20.781494+00:00", "epoch": 0, "step": 325, "train_loss": 5.479264259338379, "perplexity": 239.6703072993953, "lr": 0.0005696557707674018, "grad_norm": 1.19681, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:21.087525+00:00", "epoch": 0, "step": 326, "train_loss": 5.685506343841553, "perplexity": 294.5669594618676, "lr": 0.0005714085577543784, "grad_norm": 0.858567, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:21.394699+00:00", "epoch": 0, "step": 327, "train_loss": 5.431212425231934, "perplexity": 228.4260270804295, "lr": 0.000573161344741355, "grad_norm": 0.909893, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:21.703540+00:00", "epoch": 0, "step": 328, "train_loss": 5.429484844207764, "perplexity": 228.03174328728124, "lr": 0.0005749141317283316, "grad_norm": 0.905283, "tokens_per_sec": 106100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:22.009372+00:00", "epoch": 0, "step": 329, "train_loss": 5.33772611618042, "perplexity": 208.0391152650863, "lr": 0.0005766669187153082, "grad_norm": 1.045031, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:22.314490+00:00", "epoch": 0, "step": 330, "train_loss": 5.4790120124816895, "perplexity": 239.60985884202444, "lr": 0.000578419705702285, "grad_norm": 0.878867, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:22.621019+00:00", "epoch": 0, "step": 331, "train_loss": 5.372130870819092, "perplexity": 215.3212008701912, "lr": 0.0005801724926892616, "grad_norm": 0.838433, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:22.928539+00:00", "epoch": 0, "step": 332, "train_loss": 5.2991180419921875, "perplexity": 200.1601992140602, "lr": 0.0005819252796762381, "grad_norm": 0.879705, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:23.235373+00:00", "epoch": 0, "step": 333, "train_loss": 5.398053169250488, "perplexity": 220.97579469458972, "lr": 0.0005836780666632148, "grad_norm": 0.936421, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:23.541364+00:00", "epoch": 0, "step": 334, "train_loss": 5.320485591888428, "perplexity": 204.48315324889683, "lr": 0.0005854308536501914, "grad_norm": 1.147601, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:23.846832+00:00", "epoch": 0, "step": 335, "train_loss": 5.430795669555664, "perplexity": 228.33084907140136, "lr": 0.000587183640637168, "grad_norm": 1.124303, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:24.152520+00:00", "epoch": 0, "step": 336, "train_loss": 5.369028091430664, "perplexity": 214.6541420895186, "lr": 0.0005889364276241446, "grad_norm": 0.870843, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:24.457865+00:00", "epoch": 0, "step": 337, "train_loss": 5.318353176116943, "perplexity": 204.0475747302126, "lr": 0.0005906892146111212, "grad_norm": 0.761483, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:24.763845+00:00", "epoch": 0, "step": 338, "train_loss": 5.363295555114746, "perplexity": 213.42714967403288, "lr": 0.0005924420015980978, "grad_norm": 0.72463, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:25.069487+00:00", "epoch": 0, "step": 339, "train_loss": 5.366032600402832, "perplexity": 214.01210961414017, "lr": 0.0005941947885850744, "grad_norm": 0.656987, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:25.375914+00:00", "epoch": 0, "step": 340, "train_loss": 5.387282848358154, "perplexity": 218.60858516237533, "lr": 0.0005959475755720512, "grad_norm": 0.722365, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:25.681612+00:00", "epoch": 0, "step": 341, "train_loss": 5.425458908081055, "perplexity": 227.115547564162, "lr": 0.0005977003625590278, "grad_norm": 0.790816, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:25.988118+00:00", "epoch": 0, "step": 342, "train_loss": 5.333930015563965, "perplexity": 207.25087491778592, "lr": 0.0005994531495460044, "grad_norm": 1.06913, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:26.295528+00:00", "epoch": 0, "step": 343, "train_loss": 5.298727035522461, "perplexity": 200.08195058004375, "lr": 0.0006012059365329809, "grad_norm": 1.177721, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:26.601424+00:00", "epoch": 0, "step": 344, "train_loss": 5.339659214019775, "perplexity": 208.44166418717646, "lr": 0.0006029587235199576, "grad_norm": 0.851697, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:26.907467+00:00", "epoch": 0, "step": 345, "train_loss": 5.306052207946777, "perplexity": 201.55296651204617, "lr": 0.0006047115105069342, "grad_norm": 0.840165, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:27.214052+00:00", "epoch": 0, "step": 346, "train_loss": 5.334356784820557, "perplexity": 207.33934209579576, "lr": 0.0006064642974939108, "grad_norm": 0.925917, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:27.520125+00:00", "epoch": 0, "step": 347, "train_loss": 5.35542106628418, "perplexity": 211.75311968322006, "lr": 0.0006082170844808874, "grad_norm": 1.080985, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:27.826203+00:00", "epoch": 0, "step": 348, "train_loss": 5.294137477874756, "perplexity": 199.16576698052435, "lr": 0.000609969871467864, "grad_norm": 1.149949, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:28.132326+00:00", "epoch": 0, "step": 349, "train_loss": 5.449302673339844, "perplexity": 232.59591401106692, "lr": 0.0006117226584548407, "grad_norm": 0.975735, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:28.438580+00:00", "epoch": 0, "step": 350, "train_loss": 5.365603923797607, "perplexity": 213.92038729052558, "lr": 0.0006134754454418174, "grad_norm": 0.874525, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:28.744556+00:00", "epoch": 0, "step": 351, "train_loss": 5.316123008728027, "perplexity": 203.59302153651532, "lr": 0.000615228232428794, "grad_norm": 0.789422, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:29.050568+00:00", "epoch": 0, "step": 352, "train_loss": 5.397727966308594, "perplexity": 220.90394439966443, "lr": 0.0006169810194157706, "grad_norm": 0.75358, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:29.355360+00:00", "epoch": 0, "step": 353, "train_loss": 5.2000603675842285, "perplexity": 181.28318517278777, "lr": 0.0006187338064027471, "grad_norm": 0.662318, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:29.660531+00:00", "epoch": 0, "step": 354, "train_loss": 5.270281791687012, "perplexity": 194.4707549671083, "lr": 0.0006204865933897238, "grad_norm": 0.966669, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:29.966155+00:00", "epoch": 0, "step": 355, "train_loss": 5.361023902893066, "perplexity": 212.94286768369625, "lr": 0.0006222393803767004, "grad_norm": 1.288818, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:30.273667+00:00", "epoch": 0, "step": 356, "train_loss": 5.332644939422607, "perplexity": 206.98471281906382, "lr": 0.000623992167363677, "grad_norm": 0.583763, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:30.580789+00:00", "epoch": 0, "step": 357, "train_loss": 5.256923675537109, "perplexity": 191.89026564415278, "lr": 0.0006257449543506536, "grad_norm": 0.74471, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:30.885448+00:00", "epoch": 0, "step": 358, "train_loss": 5.363386154174805, "perplexity": 213.44648684913605, "lr": 0.0006274977413376303, "grad_norm": 0.73944, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:31.191272+00:00", "epoch": 0, "step": 359, "train_loss": 5.367945671081543, "perplexity": 214.4219217807647, "lr": 0.0006292505283246069, "grad_norm": 1.128855, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:31.496690+00:00", "epoch": 0, "step": 360, "train_loss": 5.289072513580322, "perplexity": 198.15954986010033, "lr": 0.0006310033153115835, "grad_norm": 1.149168, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:31.802672+00:00", "epoch": 0, "step": 361, "train_loss": 5.337021350860596, "perplexity": 207.89254816525164, "lr": 0.0006327561022985602, "grad_norm": 0.858328, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:32.109593+00:00", "epoch": 0, "step": 362, "train_loss": 5.220357418060303, "perplexity": 185.00029470190063, "lr": 0.0006345088892855368, "grad_norm": 0.746101, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:32.415787+00:00", "epoch": 0, "step": 363, "train_loss": 5.232147693634033, "perplexity": 187.19440834286235, "lr": 0.0006362616762725134, "grad_norm": 0.679703, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:32.722160+00:00", "epoch": 0, "step": 364, "train_loss": 5.210297107696533, "perplexity": 183.1484649295268, "lr": 0.00063801446325949, "grad_norm": 0.744632, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:33.026989+00:00", "epoch": 0, "step": 365, "train_loss": 5.383729457855225, "perplexity": 217.83316199856543, "lr": 0.0006397672502464666, "grad_norm": 0.846656, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:33.332456+00:00", "epoch": 0, "step": 366, "train_loss": 5.135343074798584, "perplexity": 169.92260603969612, "lr": 0.0006415200372334432, "grad_norm": 0.695583, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:33.639099+00:00", "epoch": 0, "step": 367, "train_loss": 5.283742427825928, "perplexity": 197.1061523098332, "lr": 0.00064327282422042, "grad_norm": 0.596768, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:33.945296+00:00", "epoch": 0, "step": 368, "train_loss": 5.349009037017822, "perplexity": 210.3996962150772, "lr": 0.0006450256112073965, "grad_norm": 0.729398, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:34.251426+00:00", "epoch": 0, "step": 369, "train_loss": 5.1820197105407715, "perplexity": 178.04204148833495, "lr": 0.0006467783981943731, "grad_norm": 0.825876, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:34.557690+00:00", "epoch": 0, "step": 370, "train_loss": 5.219563961029053, "perplexity": 184.85356313758, "lr": 0.0006485311851813497, "grad_norm": 1.103074, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:34.863862+00:00", "epoch": 0, "step": 371, "train_loss": 5.120078086853027, "perplexity": 167.34843682369967, "lr": 0.0006502839721683264, "grad_norm": 0.865803, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:35.170970+00:00", "epoch": 0, "step": 372, "train_loss": 5.263404369354248, "perplexity": 193.13788605862985, "lr": 0.000652036759155303, "grad_norm": 0.852197, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:35.477540+00:00", "epoch": 0, "step": 373, "train_loss": 5.370265960693359, "perplexity": 214.92002038145316, "lr": 0.0006537895461422796, "grad_norm": 1.165338, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:35.784493+00:00", "epoch": 0, "step": 374, "train_loss": 5.28684139251709, "perplexity": 197.7179247572585, "lr": 0.0006555423331292562, "grad_norm": 0.925048, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:36.090570+00:00", "epoch": 0, "step": 375, "train_loss": 5.280177116394043, "perplexity": 196.404658756284, "lr": 0.0006572951201162328, "grad_norm": 1.126827, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:36.395611+00:00", "epoch": 0, "step": 376, "train_loss": 5.2061381340026855, "perplexity": 182.3883370534013, "lr": 0.0006590479071032094, "grad_norm": 0.890148, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:36.700882+00:00", "epoch": 0, "step": 377, "train_loss": 5.245865821838379, "perplexity": 189.78005983627065, "lr": 0.0006608006940901861, "grad_norm": 0.79561, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:37.006505+00:00", "epoch": 0, "step": 378, "train_loss": 5.265721321105957, "perplexity": 193.58589603037947, "lr": 0.0006625534810771627, "grad_norm": 0.847016, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:37.312449+00:00", "epoch": 0, "step": 379, "train_loss": 5.050705909729004, "perplexity": 156.1326411448315, "lr": 0.0006643062680641393, "grad_norm": 0.694476, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:37.618457+00:00", "epoch": 0, "step": 380, "train_loss": 5.123223304748535, "perplexity": 167.87561272894249, "lr": 0.0006660590550511159, "grad_norm": 0.739835, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:37.922865+00:00", "epoch": 0, "step": 381, "train_loss": 5.31070613861084, "perplexity": 202.4931661583245, "lr": 0.0006678118420380925, "grad_norm": 1.089469, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:38.228304+00:00", "epoch": 0, "step": 382, "train_loss": 5.296958923339844, "perplexity": 199.72849581162384, "lr": 0.0006695646290250692, "grad_norm": 1.067701, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:38.535063+00:00", "epoch": 0, "step": 383, "train_loss": 5.210595607757568, "perplexity": 183.20314291777217, "lr": 0.0006713174160120458, "grad_norm": 0.97797, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:38.840760+00:00", "epoch": 0, "step": 384, "train_loss": 5.149071216583252, "perplexity": 172.27141315813546, "lr": 0.0006730702029990224, "grad_norm": 0.763888, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:39.148156+00:00", "epoch": 0, "step": 385, "train_loss": 5.05088996887207, "perplexity": 156.16138142984028, "lr": 0.000674822989985999, "grad_norm": 0.620955, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:39.453821+00:00", "epoch": 0, "step": 386, "train_loss": 5.095730304718018, "perplexity": 163.32307669468776, "lr": 0.0006765757769729758, "grad_norm": 0.641038, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:39.759328+00:00", "epoch": 0, "step": 387, "train_loss": 5.2150044441223145, "perplexity": 184.01263875250345, "lr": 0.0006783285639599523, "grad_norm": 0.715394, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:40.064597+00:00", "epoch": 0, "step": 388, "train_loss": 5.199681758880615, "perplexity": 181.21456277240137, "lr": 0.000680081350946929, "grad_norm": 0.779303, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:40.370804+00:00", "epoch": 0, "step": 389, "train_loss": 5.151794910430908, "perplexity": 172.74126732526543, "lr": 0.0006818341379339055, "grad_norm": 0.787924, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:40.676711+00:00", "epoch": 0, "step": 390, "train_loss": 5.121131896972656, "perplexity": 167.52488325411667, "lr": 0.0006835869249208821, "grad_norm": 0.680116, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:40.982278+00:00", "epoch": 0, "step": 391, "train_loss": 5.145759582519531, "perplexity": 171.70185687962532, "lr": 0.0006853397119078588, "grad_norm": 0.713684, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:41.287065+00:00", "epoch": 0, "step": 392, "train_loss": 5.217462062835693, "perplexity": 184.4654278205436, "lr": 0.0006870924988948353, "grad_norm": 0.680675, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:41.591314+00:00", "epoch": 0, "step": 393, "train_loss": 5.186108589172363, "perplexity": 178.77152415399635, "lr": 0.000688845285881812, "grad_norm": 0.548858, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:41.896884+00:00", "epoch": 0, "step": 394, "train_loss": 5.211397171020508, "perplexity": 183.35005069686375, "lr": 0.0006905980728687886, "grad_norm": 0.520846, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:42.203785+00:00", "epoch": 0, "step": 395, "train_loss": 5.151406288146973, "perplexity": 172.67414926205979, "lr": 0.0006923508598557653, "grad_norm": 0.519164, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:42.511280+00:00", "epoch": 0, "step": 396, "train_loss": 5.126126289367676, "perplexity": 168.36366110634498, "lr": 0.0006941036468427419, "grad_norm": 0.521091, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:42.815477+00:00", "epoch": 0, "step": 397, "train_loss": 5.069455146789551, "perplexity": 159.0876243223776, "lr": 0.0006958564338297186, "grad_norm": 0.633938, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:43.119714+00:00", "epoch": 0, "step": 398, "train_loss": 5.101625919342041, "perplexity": 164.2888106143307, "lr": 0.0006976092208166951, "grad_norm": 0.669322, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:43.425453+00:00", "epoch": 0, "step": 399, "train_loss": 5.117822647094727, "perplexity": 166.97141783742853, "lr": 0.0006993620078036718, "grad_norm": 0.67479, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:43.732847+00:00", "epoch": 0, "step": 400, "train_loss": 5.06339693069458, "perplexity": 158.12675064521414, "lr": 0.0007011147947906484, "grad_norm": 0.89855, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:44.039738+00:00", "epoch": 0, "step": 401, "train_loss": 5.158649444580078, "perplexity": 173.9293956221317, "lr": 0.0007028675817776249, "grad_norm": 1.055094, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:44.344627+00:00", "epoch": 0, "step": 402, "train_loss": 5.131995677947998, "perplexity": 169.3547585790268, "lr": 0.0007046203687646016, "grad_norm": 1.13094, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:44.650408+00:00", "epoch": 0, "step": 403, "train_loss": 5.097175598144531, "perplexity": 163.5592971266318, "lr": 0.0007063731557515782, "grad_norm": 0.712172, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:44.956019+00:00", "epoch": 0, "step": 404, "train_loss": 5.194971561431885, "perplexity": 180.36301345823085, "lr": 0.0007081259427385549, "grad_norm": 0.647202, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:45.262042+00:00", "epoch": 0, "step": 405, "train_loss": 5.2198381423950195, "perplexity": 184.90425348888223, "lr": 0.0007098787297255315, "grad_norm": 0.714197, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:45.566783+00:00", "epoch": 0, "step": 406, "train_loss": 5.062290668487549, "perplexity": 157.95191772041358, "lr": 0.000711631516712508, "grad_norm": 1.059677, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:45.873280+00:00", "epoch": 0, "step": 407, "train_loss": 5.1718854904174805, "perplexity": 176.24683611619727, "lr": 0.0007133843036994847, "grad_norm": 1.001384, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:46.177878+00:00", "epoch": 0, "step": 408, "train_loss": 5.230887413024902, "perplexity": 186.95863945857278, "lr": 0.0007151370906864614, "grad_norm": 1.041494, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:46.483331+00:00", "epoch": 0, "step": 409, "train_loss": 5.03986930847168, "perplexity": 154.4498284226852, "lr": 0.0007168898776734379, "grad_norm": 1.15012, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:46.788680+00:00", "epoch": 0, "step": 410, "train_loss": 5.190668106079102, "perplexity": 179.58849702628146, "lr": 0.0007186426646604145, "grad_norm": 0.856656, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:47.094149+00:00", "epoch": 0, "step": 411, "train_loss": 5.1985273361206055, "perplexity": 181.00548526182973, "lr": 0.0007203954516473912, "grad_norm": 0.663214, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:47.400131+00:00", "epoch": 0, "step": 412, "train_loss": 5.1288161277771, "perplexity": 168.81714177017406, "lr": 0.0007221482386343677, "grad_norm": 0.666287, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:47.705841+00:00", "epoch": 0, "step": 413, "train_loss": 5.06922721862793, "perplexity": 159.0513679047154, "lr": 0.0007239010256213444, "grad_norm": 0.704974, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:48.011140+00:00", "epoch": 0, "step": 414, "train_loss": 5.152599811553955, "perplexity": 172.8803629369295, "lr": 0.0007256538126083211, "grad_norm": 0.683131, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:48.317258+00:00", "epoch": 0, "step": 415, "train_loss": 5.111547470092773, "perplexity": 165.926923257907, "lr": 0.0007274065995952977, "grad_norm": 0.630107, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:48.623924+00:00", "epoch": 0, "step": 416, "train_loss": 4.931484222412109, "perplexity": 138.58505080822096, "lr": 0.0007291593865822743, "grad_norm": 0.579529, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:48.930529+00:00", "epoch": 0, "step": 417, "train_loss": 5.049273490905762, "perplexity": 155.90915391260526, "lr": 0.000730912173569251, "grad_norm": 0.610951, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:49.237690+00:00", "epoch": 0, "step": 418, "train_loss": 5.048881530761719, "perplexity": 155.84805571299182, "lr": 0.0007326649605562275, "grad_norm": 0.563123, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:49.543799+00:00", "epoch": 0, "step": 419, "train_loss": 5.023307800292969, "perplexity": 151.91297139981012, "lr": 0.0007344177475432042, "grad_norm": 0.576238, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:49.849932+00:00", "epoch": 0, "step": 420, "train_loss": 5.049970626831055, "perplexity": 156.0178816794925, "lr": 0.0007361705345301808, "grad_norm": 0.655041, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:50.155156+00:00", "epoch": 0, "step": 421, "train_loss": 5.139228343963623, "perplexity": 170.58408528173345, "lr": 0.0007379233215171573, "grad_norm": 0.852827, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:50.459854+00:00", "epoch": 0, "step": 422, "train_loss": 5.17752742767334, "perplexity": 177.24402008665544, "lr": 0.000739676108504134, "grad_norm": 0.910916, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:50.765971+00:00", "epoch": 0, "step": 423, "train_loss": 5.11374568939209, "perplexity": 166.2920682101925, "lr": 0.0007414288954911105, "grad_norm": 0.762382, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:51.073610+00:00", "epoch": 0, "step": 424, "train_loss": 5.3125715255737305, "perplexity": 202.87124679426233, "lr": 0.0007431816824780873, "grad_norm": 0.92442, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:51.379922+00:00", "epoch": 0, "step": 425, "train_loss": 4.9825873374938965, "perplexity": 145.85126032157254, "lr": 0.0007449344694650639, "grad_norm": 0.99214, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:51.685446+00:00", "epoch": 0, "step": 426, "train_loss": 5.035129547119141, "perplexity": 153.7195052410625, "lr": 0.0007466872564520405, "grad_norm": 0.848112, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:51.990697+00:00", "epoch": 0, "step": 427, "train_loss": 5.064046382904053, "perplexity": 158.22947976801447, "lr": 0.0007484400434390171, "grad_norm": 0.640648, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:52.296918+00:00", "epoch": 0, "step": 428, "train_loss": 5.052318096160889, "perplexity": 156.38455908523798, "lr": 0.0007501928304259938, "grad_norm": 0.660485, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:52.606789+00:00", "epoch": 0, "step": 429, "train_loss": 5.070928573608398, "perplexity": 159.3222010679964, "lr": 0.0007519456174129703, "grad_norm": 0.776753, "tokens_per_sec": 105747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:52.912486+00:00", "epoch": 0, "step": 430, "train_loss": 5.132652282714844, "perplexity": 169.46599423573258, "lr": 0.000753698404399947, "grad_norm": 0.663062, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:53.217716+00:00", "epoch": 0, "step": 431, "train_loss": 4.990976810455322, "perplexity": 147.08002265359443, "lr": 0.0007554511913869236, "grad_norm": 0.793165, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:53.522679+00:00", "epoch": 0, "step": 432, "train_loss": 5.101473808288574, "perplexity": 164.26382237081756, "lr": 0.0007572039783739001, "grad_norm": 0.667087, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:53.829268+00:00", "epoch": 0, "step": 433, "train_loss": 4.994142055511475, "perplexity": 147.54630452705712, "lr": 0.0007589567653608769, "grad_norm": 0.622575, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:54.134093+00:00", "epoch": 0, "step": 434, "train_loss": 5.0525970458984375, "perplexity": 156.4281886018897, "lr": 0.0007607095523478536, "grad_norm": 0.667553, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:54.440799+00:00", "epoch": 0, "step": 435, "train_loss": 5.193737506866455, "perplexity": 180.14057293813713, "lr": 0.0007624623393348301, "grad_norm": 0.570261, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:54.747536+00:00", "epoch": 0, "step": 436, "train_loss": 4.985287666320801, "perplexity": 146.24563892056565, "lr": 0.0007642151263218067, "grad_norm": 0.648111, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:55.056708+00:00", "epoch": 0, "step": 437, "train_loss": 5.007032871246338, "perplexity": 149.46060871358677, "lr": 0.0007659679133087834, "grad_norm": 0.713311, "tokens_per_sec": 105987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:55.361936+00:00", "epoch": 0, "step": 438, "train_loss": 5.073342800140381, "perplexity": 159.7073056307269, "lr": 0.0007677207002957599, "grad_norm": 0.590367, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:55.668360+00:00", "epoch": 0, "step": 439, "train_loss": 5.119909763336182, "perplexity": 167.3202705168673, "lr": 0.0007694734872827366, "grad_norm": 0.609392, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:55.974256+00:00", "epoch": 0, "step": 440, "train_loss": 5.093424320220947, "perplexity": 162.94689011872177, "lr": 0.0007712262742697131, "grad_norm": 0.671274, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:56.279505+00:00", "epoch": 0, "step": 441, "train_loss": 5.085855960845947, "perplexity": 161.71830454827406, "lr": 0.0007729790612566899, "grad_norm": 0.677504, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:56.585210+00:00", "epoch": 0, "step": 442, "train_loss": 4.982540607452393, "perplexity": 145.8444448453693, "lr": 0.0007747318482436665, "grad_norm": 0.668653, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:56.890386+00:00", "epoch": 0, "step": 443, "train_loss": 5.062612056732178, "perplexity": 158.00268976831816, "lr": 0.000776484635230643, "grad_norm": 0.696771, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:57.195756+00:00", "epoch": 0, "step": 444, "train_loss": 5.043940544128418, "perplexity": 155.07991181001302, "lr": 0.0007782374222176197, "grad_norm": 0.602102, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:57.501871+00:00", "epoch": 0, "step": 445, "train_loss": 5.033796787261963, "perplexity": 153.51477051662803, "lr": 0.0007799902092045964, "grad_norm": 0.602804, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:57.807538+00:00", "epoch": 0, "step": 446, "train_loss": 4.939695835113525, "perplexity": 139.72774282276916, "lr": 0.0007817429961915729, "grad_norm": 0.643028, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:58.113275+00:00", "epoch": 0, "step": 447, "train_loss": 5.0564985275268555, "perplexity": 157.03968239613155, "lr": 0.0007834957831785495, "grad_norm": 0.717037, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:58.418611+00:00", "epoch": 0, "step": 448, "train_loss": 4.97350549697876, "perplexity": 144.5326591622701, "lr": 0.0007852485701655262, "grad_norm": 0.810691, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:58.723310+00:00", "epoch": 0, "step": 449, "train_loss": 4.959803104400635, "perplexity": 142.56572256992817, "lr": 0.0007870013571525027, "grad_norm": 0.720972, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:59.028827+00:00", "epoch": 0, "step": 450, "train_loss": 4.938843250274658, "perplexity": 139.60866383735137, "lr": 0.0007887541441394794, "grad_norm": 0.814996, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:59.333754+00:00", "epoch": 0, "step": 451, "train_loss": 5.011410713195801, "perplexity": 150.1163579721985, "lr": 0.0007905069311264561, "grad_norm": 0.809988, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:59.639867+00:00", "epoch": 0, "step": 452, "train_loss": 4.972068786621094, "perplexity": 144.32515668999903, "lr": 0.0007922597181134327, "grad_norm": 0.837022, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T20:59:59.945590+00:00", "epoch": 0, "step": 453, "train_loss": 5.157489776611328, "perplexity": 173.72781218072595, "lr": 0.0007940125051004093, "grad_norm": 0.785324, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:00.250290+00:00", "epoch": 0, "step": 454, "train_loss": 4.984630107879639, "perplexity": 146.14950547635303, "lr": 0.000795765292087386, "grad_norm": 0.655453, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:00.555253+00:00", "epoch": 0, "step": 455, "train_loss": 4.877976894378662, "perplexity": 131.36463034749443, "lr": 0.0007975180790743625, "grad_norm": 0.639032, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:00.861106+00:00", "epoch": 0, "step": 456, "train_loss": 5.05112886428833, "perplexity": 156.19869212456018, "lr": 0.0007992708660613391, "grad_norm": 0.686255, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:01.167396+00:00", "epoch": 0, "step": 457, "train_loss": 5.051804065704346, "perplexity": 156.30419331593126, "lr": 0.0008010236530483157, "grad_norm": 0.66603, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:01.472443+00:00", "epoch": 0, "step": 458, "train_loss": 4.90056037902832, "perplexity": 134.36505395027868, "lr": 0.0008027764400352923, "grad_norm": 0.687876, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:01.777686+00:00", "epoch": 0, "step": 459, "train_loss": 4.959214210510254, "perplexity": 142.48179120268927, "lr": 0.000804529227022269, "grad_norm": 0.674266, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:02.082068+00:00", "epoch": 0, "step": 460, "train_loss": 4.974020481109619, "perplexity": 144.6071103570761, "lr": 0.0008062820140092455, "grad_norm": 0.757303, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:02.388385+00:00", "epoch": 0, "step": 461, "train_loss": 4.940258502960205, "perplexity": 139.80638525365347, "lr": 0.0008080348009962223, "grad_norm": 0.628211, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:02.694459+00:00", "epoch": 0, "step": 462, "train_loss": 4.987913608551025, "perplexity": 146.6301761851474, "lr": 0.0008097875879831989, "grad_norm": 0.500909, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:03.000641+00:00", "epoch": 0, "step": 463, "train_loss": 5.022953510284424, "perplexity": 151.85915968491412, "lr": 0.0008115403749701754, "grad_norm": 0.502672, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:03.306837+00:00", "epoch": 0, "step": 464, "train_loss": 4.929862976074219, "perplexity": 138.36055233494895, "lr": 0.0008132931619571521, "grad_norm": 0.498058, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:03.612585+00:00", "epoch": 0, "step": 465, "train_loss": 4.969226360321045, "perplexity": 143.91550554626792, "lr": 0.0008150459489441288, "grad_norm": 0.475817, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:03.919218+00:00", "epoch": 0, "step": 466, "train_loss": 4.954796314239502, "perplexity": 141.85370984938302, "lr": 0.0008167987359311053, "grad_norm": 0.486003, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:04.225585+00:00", "epoch": 0, "step": 467, "train_loss": 4.812118053436279, "perplexity": 122.99184513219976, "lr": 0.0008185515229180819, "grad_norm": 0.468859, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:04.533036+00:00", "epoch": 0, "step": 468, "train_loss": 4.944949626922607, "perplexity": 140.46377508062048, "lr": 0.0008203043099050586, "grad_norm": 0.563133, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:04.838427+00:00", "epoch": 0, "step": 469, "train_loss": 4.982804775238037, "perplexity": 145.8829773387103, "lr": 0.0008220570968920351, "grad_norm": 0.600007, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:05.143453+00:00", "epoch": 0, "step": 470, "train_loss": 4.949843883514404, "perplexity": 141.15292590254913, "lr": 0.0008238098838790119, "grad_norm": 0.628047, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:05.448869+00:00", "epoch": 0, "step": 471, "train_loss": 4.9443159103393555, "perplexity": 140.37478905594352, "lr": 0.0008255626708659884, "grad_norm": 0.692038, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:05.754831+00:00", "epoch": 0, "step": 472, "train_loss": 4.990185737609863, "perplexity": 146.9637176504938, "lr": 0.0008273154578529651, "grad_norm": 0.842584, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:06.060566+00:00", "epoch": 0, "step": 473, "train_loss": 4.890616416931152, "perplexity": 133.03555414974974, "lr": 0.0008290682448399417, "grad_norm": 0.974677, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:06.365993+00:00", "epoch": 0, "step": 474, "train_loss": 4.836764812469482, "perplexity": 126.06086078909128, "lr": 0.0008308210318269182, "grad_norm": 0.929624, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:06.672252+00:00", "epoch": 0, "step": 475, "train_loss": 4.81369161605835, "perplexity": 123.18553285243186, "lr": 0.0008325738188138949, "grad_norm": 0.974233, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:06.977873+00:00", "epoch": 0, "step": 476, "train_loss": 4.883096218109131, "perplexity": 132.03885272581553, "lr": 0.0008343266058008716, "grad_norm": 0.815327, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:07.284404+00:00", "epoch": 0, "step": 477, "train_loss": 4.9248127937316895, "perplexity": 137.66356774761618, "lr": 0.0008360793927878481, "grad_norm": 0.703979, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:07.591548+00:00", "epoch": 0, "step": 478, "train_loss": 4.915175437927246, "perplexity": 136.3432274776723, "lr": 0.0008378321797748247, "grad_norm": 0.686903, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:07.897386+00:00", "epoch": 0, "step": 479, "train_loss": 5.010483264923096, "perplexity": 149.97719735740336, "lr": 0.0008395849667618015, "grad_norm": 0.658066, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:08.203309+00:00", "epoch": 0, "step": 480, "train_loss": 4.966926574707031, "perplexity": 143.58491103094423, "lr": 0.000841337753748778, "grad_norm": 0.677328, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:08.508526+00:00", "epoch": 0, "step": 481, "train_loss": 4.849440097808838, "perplexity": 127.6688877408148, "lr": 0.0008430905407357547, "grad_norm": 0.719186, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:08.815202+00:00", "epoch": 0, "step": 482, "train_loss": 4.827392101287842, "perplexity": 124.88484857019593, "lr": 0.0008448433277227313, "grad_norm": 0.571015, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:09.120999+00:00", "epoch": 0, "step": 483, "train_loss": 4.953493118286133, "perplexity": 141.6689670729023, "lr": 0.0008465961147097079, "grad_norm": 0.563341, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:09.427102+00:00", "epoch": 0, "step": 484, "train_loss": 4.844512462615967, "perplexity": 127.04132949543828, "lr": 0.0008483489016966845, "grad_norm": 0.530051, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:09.732798+00:00", "epoch": 0, "step": 485, "train_loss": 4.789851188659668, "perplexity": 120.28346778726062, "lr": 0.0008501016886836612, "grad_norm": 0.520839, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:10.038389+00:00", "epoch": 0, "step": 486, "train_loss": 4.839578151702881, "perplexity": 126.41601210085656, "lr": 0.0008518544756706377, "grad_norm": 0.574773, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:10.343149+00:00", "epoch": 0, "step": 487, "train_loss": 4.690122127532959, "perplexity": 108.86647458953355, "lr": 0.0008536072626576145, "grad_norm": 0.607267, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:10.649097+00:00", "epoch": 0, "step": 488, "train_loss": 4.994503021240234, "perplexity": 147.59957329991886, "lr": 0.000855360049644591, "grad_norm": 0.538047, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:10.955201+00:00", "epoch": 0, "step": 489, "train_loss": 4.8661980628967285, "perplexity": 129.82638565121547, "lr": 0.0008571128366315676, "grad_norm": 0.541917, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:11.260990+00:00", "epoch": 0, "step": 490, "train_loss": 4.945361614227295, "perplexity": 140.52165629506626, "lr": 0.0008588656236185443, "grad_norm": 0.529773, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:11.565476+00:00", "epoch": 0, "step": 491, "train_loss": 4.827022552490234, "perplexity": 124.83870605104879, "lr": 0.0008606184106055208, "grad_norm": 0.533334, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:11.870664+00:00", "epoch": 0, "step": 492, "train_loss": 4.982114791870117, "perplexity": 145.7823552284672, "lr": 0.0008623711975924975, "grad_norm": 0.614895, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:12.177446+00:00", "epoch": 0, "step": 493, "train_loss": 4.819337368011475, "perplexity": 123.88297475456993, "lr": 0.0008641239845794741, "grad_norm": 0.749424, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:12.484426+00:00", "epoch": 0, "step": 494, "train_loss": 4.934336185455322, "perplexity": 138.9808543917854, "lr": 0.0008658767715664507, "grad_norm": 0.750465, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:12.790521+00:00", "epoch": 0, "step": 495, "train_loss": 4.749154567718506, "perplexity": 115.48660713744216, "lr": 0.0008676295585534273, "grad_norm": 0.842868, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:13.096061+00:00", "epoch": 0, "step": 496, "train_loss": 4.798346519470215, "perplexity": 121.30966842230005, "lr": 0.000869382345540404, "grad_norm": 0.90537, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:13.401310+00:00", "epoch": 0, "step": 497, "train_loss": 4.802364826202393, "perplexity": 121.79810857354548, "lr": 0.0008711351325273805, "grad_norm": 0.7615, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:13.706507+00:00", "epoch": 0, "step": 498, "train_loss": 4.822428226470947, "perplexity": 124.26647185739839, "lr": 0.0008728879195143573, "grad_norm": 0.690987, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:14.011595+00:00", "epoch": 0, "step": 499, "train_loss": 4.834755897521973, "perplexity": 125.80786944568572, "lr": 0.0008746407065013339, "grad_norm": 0.549357, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:14.318396+00:00", "epoch": 0, "step": 500, "train_loss": 4.849989414215088, "perplexity": 127.73903762089455, "lr": 0.0008763934934883104, "grad_norm": 0.567948, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:00:17.391464+00:00", "step": 500, "epoch": 0, "val_loss": 4.7747421026229855, "val_ppl": 118.4797550622178, "eval_train_loss": 4.849989414215088, "eval_train_ppl": 127.73903762089455} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:00:18.369517+00:00", "step": 500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_4p7747_epoch_0000_step_0000500.pt", "val_loss": 4.7747421026229855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:19.895710+00:00", "epoch": 0, "step": 501, "train_loss": 4.80290412902832, "perplexity": 121.86381235321639, "lr": 0.0008781462804752871, "grad_norm": 0.623838, "tokens_per_sec": 5875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:20.199973+00:00", "epoch": 0, "step": 502, "train_loss": 4.94523811340332, "perplexity": 140.50430282633343, "lr": 0.0008798990674622638, "grad_norm": 0.614693, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:20.505984+00:00", "epoch": 0, "step": 503, "train_loss": 4.876494407653809, "perplexity": 131.1700283099984, "lr": 0.0008816518544492403, "grad_norm": 0.554097, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:20.810677+00:00", "epoch": 0, "step": 504, "train_loss": 4.955849647521973, "perplexity": 142.003207804973, "lr": 0.0008834046414362169, "grad_norm": 0.515159, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:21.118845+00:00", "epoch": 0, "step": 505, "train_loss": 4.808058738708496, "perplexity": 122.49359448613272, "lr": 0.0008851574284231935, "grad_norm": 0.534184, "tokens_per_sec": 106331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:21.425200+00:00", "epoch": 0, "step": 506, "train_loss": 4.6522135734558105, "perplexity": 104.81674854260349, "lr": 0.0008869102154101701, "grad_norm": 0.541665, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:21.733019+00:00", "epoch": 0, "step": 507, "train_loss": 4.960472583770752, "perplexity": 142.66119933637123, "lr": 0.0008886630023971469, "grad_norm": 0.528085, "tokens_per_sec": 106452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:22.040105+00:00", "epoch": 0, "step": 508, "train_loss": 4.7289652824401855, "perplexity": 113.17839406934044, "lr": 0.0008904157893841234, "grad_norm": 0.48764, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:22.344912+00:00", "epoch": 0, "step": 509, "train_loss": 4.833701133728027, "perplexity": 125.67524181770703, "lr": 0.0008921685763711, "grad_norm": 0.394638, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:22.650183+00:00", "epoch": 0, "step": 510, "train_loss": 4.702431678771973, "perplexity": 110.21485398326077, "lr": 0.0008939213633580767, "grad_norm": 0.438039, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:22.953967+00:00", "epoch": 0, "step": 511, "train_loss": 4.6777238845825195, "perplexity": 107.52505439827893, "lr": 0.0008956741503450532, "grad_norm": 0.441813, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:23.260180+00:00", "epoch": 0, "step": 512, "train_loss": 4.729907035827637, "perplexity": 113.28503041002776, "lr": 0.0008974269373320299, "grad_norm": 0.522623, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:23.566321+00:00", "epoch": 0, "step": 513, "train_loss": 4.759279251098633, "perplexity": 116.6618117174024, "lr": 0.0008991797243190065, "grad_norm": 0.62002, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:23.979457+00:00", "epoch": 0, "step": 514, "train_loss": 4.770457744598389, "perplexity": 117.97323121242573, "lr": 0.0009009325113059831, "grad_norm": 0.545272, "tokens_per_sec": 79315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:24.284651+00:00", "epoch": 0, "step": 515, "train_loss": 4.745258808135986, "perplexity": 115.03757431091418, "lr": 0.0009026852982929597, "grad_norm": 0.553335, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:24.591150+00:00", "epoch": 0, "step": 516, "train_loss": 4.779783725738525, "perplexity": 119.07859362479525, "lr": 0.0009044380852799365, "grad_norm": 0.751088, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:24.896731+00:00", "epoch": 0, "step": 517, "train_loss": 4.776859760284424, "perplexity": 118.73092047056423, "lr": 0.000906190872266913, "grad_norm": 0.890292, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:25.201710+00:00", "epoch": 0, "step": 518, "train_loss": 4.8665971755981445, "perplexity": 129.87821135216763, "lr": 0.0009079436592538897, "grad_norm": 0.992412, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:25.507850+00:00", "epoch": 0, "step": 519, "train_loss": 4.722855567932129, "perplexity": 112.48901449375755, "lr": 0.0009096964462408663, "grad_norm": 0.842137, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:25.813553+00:00", "epoch": 0, "step": 520, "train_loss": 4.823612689971924, "perplexity": 124.41374816219391, "lr": 0.0009114492332278428, "grad_norm": 0.917317, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:26.120348+00:00", "epoch": 0, "step": 521, "train_loss": 4.771324634552002, "perplexity": 118.07554536250608, "lr": 0.0009132020202148195, "grad_norm": 0.688606, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:26.426288+00:00", "epoch": 0, "step": 522, "train_loss": 4.817636966705322, "perplexity": 123.67250297644298, "lr": 0.000914954807201796, "grad_norm": 0.554023, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:26.731960+00:00", "epoch": 0, "step": 523, "train_loss": 4.73953914642334, "perplexity": 114.38147641926784, "lr": 0.0009167075941887727, "grad_norm": 0.471928, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:27.037160+00:00", "epoch": 0, "step": 524, "train_loss": 4.793044090270996, "perplexity": 120.66813484060742, "lr": 0.0009184603811757493, "grad_norm": 0.517993, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:27.342574+00:00", "epoch": 0, "step": 525, "train_loss": 4.785846710205078, "perplexity": 119.8027583696977, "lr": 0.000920213168162726, "grad_norm": 0.515746, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:27.658916+00:00", "epoch": 0, "step": 526, "train_loss": 4.822999954223633, "perplexity": 124.33753876161248, "lr": 0.0009219659551497026, "grad_norm": 0.52207, "tokens_per_sec": 103585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:27.978283+00:00", "epoch": 0, "step": 527, "train_loss": 4.801111221313477, "perplexity": 121.64551753359719, "lr": 0.0009237187421366793, "grad_norm": 0.543202, "tokens_per_sec": 102603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:28.297138+00:00", "epoch": 0, "step": 528, "train_loss": 4.7319769859313965, "perplexity": 113.51976763383506, "lr": 0.0009254715291236558, "grad_norm": 0.534636, "tokens_per_sec": 102767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:28.616598+00:00", "epoch": 0, "step": 529, "train_loss": 4.755581855773926, "perplexity": 116.23126332448668, "lr": 0.0009272243161106325, "grad_norm": 0.611316, "tokens_per_sec": 102574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:28.935320+00:00", "epoch": 0, "step": 530, "train_loss": 4.800248622894287, "perplexity": 121.54063154621826, "lr": 0.0009289771030976091, "grad_norm": 0.642711, "tokens_per_sec": 102864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:29.253443+00:00", "epoch": 0, "step": 531, "train_loss": 4.861891269683838, "perplexity": 129.26845256824626, "lr": 0.0009307298900845856, "grad_norm": 0.750763, "tokens_per_sec": 102950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:29.570701+00:00", "epoch": 0, "step": 532, "train_loss": 4.834955215454102, "perplexity": 125.83294770925998, "lr": 0.0009324826770715623, "grad_norm": 0.705595, "tokens_per_sec": 103286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:29.888405+00:00", "epoch": 0, "step": 533, "train_loss": 4.675136089324951, "perplexity": 107.24716129274776, "lr": 0.000934235464058539, "grad_norm": 0.55639, "tokens_per_sec": 103138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:30.205656+00:00", "epoch": 0, "step": 534, "train_loss": 4.595064163208008, "perplexity": 98.99448714776669, "lr": 0.0009359882510455155, "grad_norm": 0.605311, "tokens_per_sec": 103288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:30.524650+00:00", "epoch": 0, "step": 535, "train_loss": 4.812413215637207, "perplexity": 123.02815303400163, "lr": 0.0009377410380324922, "grad_norm": 0.558975, "tokens_per_sec": 102722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:30.843346+00:00", "epoch": 0, "step": 536, "train_loss": 4.646508693695068, "perplexity": 104.22048402110701, "lr": 0.0009394938250194689, "grad_norm": 0.592212, "tokens_per_sec": 102820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:31.161625+00:00", "epoch": 0, "step": 537, "train_loss": 4.796656608581543, "perplexity": 121.10483901320237, "lr": 0.0009412466120064454, "grad_norm": 0.528573, "tokens_per_sec": 102953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:31.479629+00:00", "epoch": 0, "step": 538, "train_loss": 4.780304908752441, "perplexity": 119.14067154068596, "lr": 0.0009429993989934221, "grad_norm": 0.434294, "tokens_per_sec": 103044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:31.793466+00:00", "epoch": 0, "step": 539, "train_loss": 4.772127628326416, "perplexity": 118.17039736803098, "lr": 0.0009447521859803986, "grad_norm": 0.474619, "tokens_per_sec": 104416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:32.106674+00:00", "epoch": 0, "step": 540, "train_loss": 4.8347249031066895, "perplexity": 125.8039701647625, "lr": 0.0009465049729673753, "grad_norm": 0.503278, "tokens_per_sec": 104694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:32.418793+00:00", "epoch": 0, "step": 541, "train_loss": 4.825031280517578, "perplexity": 124.59036557336232, "lr": 0.0009482577599543519, "grad_norm": 0.557107, "tokens_per_sec": 104905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:32.731879+00:00", "epoch": 0, "step": 542, "train_loss": 4.839413642883301, "perplexity": 126.3952172624439, "lr": 0.0009500105469413284, "grad_norm": 0.590481, "tokens_per_sec": 104662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:33.044730+00:00", "epoch": 0, "step": 543, "train_loss": 4.815932273864746, "perplexity": 123.46185893874497, "lr": 0.0009517633339283051, "grad_norm": 0.726198, "tokens_per_sec": 104740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:33.358768+00:00", "epoch": 0, "step": 544, "train_loss": 4.591517448425293, "perplexity": 98.64400383636274, "lr": 0.0009535161209152819, "grad_norm": 0.773116, "tokens_per_sec": 104344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:33.672019+00:00", "epoch": 0, "step": 545, "train_loss": 4.651241302490234, "perplexity": 104.71488778744099, "lr": 0.0009552689079022584, "grad_norm": 0.663253, "tokens_per_sec": 104607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:33.984404+00:00", "epoch": 0, "step": 546, "train_loss": 4.734766006469727, "perplexity": 113.83681852247376, "lr": 0.000957021694889235, "grad_norm": 0.587932, "tokens_per_sec": 104896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:34.297702+00:00", "epoch": 0, "step": 547, "train_loss": 4.744449615478516, "perplexity": 114.94452440317698, "lr": 0.0009587744818762117, "grad_norm": 0.810285, "tokens_per_sec": 104590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:34.612084+00:00", "epoch": 0, "step": 548, "train_loss": 4.709996223449707, "perplexity": 111.05174051242409, "lr": 0.0009605272688631882, "grad_norm": 0.733178, "tokens_per_sec": 104230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:34.926934+00:00", "epoch": 0, "step": 549, "train_loss": 4.650840759277344, "perplexity": 104.67295334868719, "lr": 0.0009622800558501649, "grad_norm": 0.716296, "tokens_per_sec": 104076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:35.241777+00:00", "epoch": 0, "step": 550, "train_loss": 4.725009918212891, "perplexity": 112.73161646469603, "lr": 0.0009640328428371415, "grad_norm": 0.687163, "tokens_per_sec": 104131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:35.554488+00:00", "epoch": 0, "step": 551, "train_loss": 4.683543682098389, "perplexity": 108.15265291951908, "lr": 0.0009657856298241181, "grad_norm": 0.57112, "tokens_per_sec": 104733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:35.868153+00:00", "epoch": 0, "step": 552, "train_loss": 4.688634872436523, "perplexity": 108.70468271308836, "lr": 0.0009675384168110947, "grad_norm": 0.585812, "tokens_per_sec": 104469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:36.181123+00:00", "epoch": 0, "step": 553, "train_loss": 4.738995552062988, "perplexity": 114.31931619026327, "lr": 0.0009692912037980715, "grad_norm": 0.569481, "tokens_per_sec": 104700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:36.495699+00:00", "epoch": 0, "step": 554, "train_loss": 4.768825054168701, "perplexity": 117.78077460065226, "lr": 0.000971043990785048, "grad_norm": 0.592847, "tokens_per_sec": 104167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:36.813042+00:00", "epoch": 0, "step": 555, "train_loss": 4.663404941558838, "perplexity": 105.99637989117603, "lr": 0.0009727967777720247, "grad_norm": 0.602445, "tokens_per_sec": 103256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:37.128752+00:00", "epoch": 0, "step": 556, "train_loss": 4.7336602210998535, "perplexity": 113.71100900597966, "lr": 0.0009745495647590012, "grad_norm": 0.547858, "tokens_per_sec": 103792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:37.441488+00:00", "epoch": 0, "step": 557, "train_loss": 4.795263290405273, "perplexity": 120.93621893775808, "lr": 0.0009763023517459778, "grad_norm": 0.610031, "tokens_per_sec": 104778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:37.755779+00:00", "epoch": 0, "step": 558, "train_loss": 4.62069845199585, "perplexity": 101.56494560079824, "lr": 0.0009780551387329544, "grad_norm": 0.647386, "tokens_per_sec": 104261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:38.069102+00:00", "epoch": 0, "step": 559, "train_loss": 4.685873508453369, "perplexity": 108.40492357995635, "lr": 0.0009798079257199311, "grad_norm": 0.662976, "tokens_per_sec": 104583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:38.382250+00:00", "epoch": 0, "step": 560, "train_loss": 4.590478420257568, "perplexity": 98.54156316638651, "lr": 0.0009815607127069076, "grad_norm": 0.533591, "tokens_per_sec": 104693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:38.695795+00:00", "epoch": 0, "step": 561, "train_loss": 4.661362648010254, "perplexity": 105.78012507142853, "lr": 0.0009833134996938843, "grad_norm": 0.462852, "tokens_per_sec": 104455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:39.009414+00:00", "epoch": 0, "step": 562, "train_loss": 4.722772121429443, "perplexity": 112.47962807054505, "lr": 0.0009850662866808608, "grad_norm": 0.466959, "tokens_per_sec": 104483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:39.324841+00:00", "epoch": 0, "step": 563, "train_loss": 4.680042743682861, "perplexity": 107.774679159655, "lr": 0.0009868190736678376, "grad_norm": 0.445902, "tokens_per_sec": 103885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:39.638023+00:00", "epoch": 0, "step": 564, "train_loss": 4.670078277587891, "perplexity": 106.7060948012592, "lr": 0.0009885718606548143, "grad_norm": 0.445185, "tokens_per_sec": 104629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:39.951269+00:00", "epoch": 0, "step": 565, "train_loss": 4.7356672286987305, "perplexity": 113.93945703691375, "lr": 0.0009903246476417908, "grad_norm": 0.469018, "tokens_per_sec": 104609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:40.263837+00:00", "epoch": 0, "step": 566, "train_loss": 4.5630388259887695, "perplexity": 95.87438317112611, "lr": 0.0009920774346287675, "grad_norm": 0.522851, "tokens_per_sec": 104838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:40.578042+00:00", "epoch": 0, "step": 567, "train_loss": 4.576258659362793, "perplexity": 97.15024127614687, "lr": 0.0009938302216157442, "grad_norm": 0.617022, "tokens_per_sec": 104285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:40.894688+00:00", "epoch": 0, "step": 568, "train_loss": 4.580948352813721, "perplexity": 97.60691612195339, "lr": 0.0009955830086027207, "grad_norm": 0.533355, "tokens_per_sec": 103485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:41.208906+00:00", "epoch": 0, "step": 569, "train_loss": 4.603454113006592, "perplexity": 99.82853986298215, "lr": 0.0009973357955896972, "grad_norm": 0.546929, "tokens_per_sec": 104284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:41.524723+00:00", "epoch": 0, "step": 570, "train_loss": 4.5784759521484375, "perplexity": 97.36589079596298, "lr": 0.000999088582576674, "grad_norm": 0.582055, "tokens_per_sec": 103822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:41.837889+00:00", "epoch": 0, "step": 571, "train_loss": 4.636992931365967, "perplexity": 103.23345030243577, "lr": 0.0010008413695636504, "grad_norm": 0.618612, "tokens_per_sec": 104568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:42.150060+00:00", "epoch": 0, "step": 572, "train_loss": 4.6823015213012695, "perplexity": 108.01839333721604, "lr": 0.0010025941565506272, "grad_norm": 0.522258, "tokens_per_sec": 104967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:42.463433+00:00", "epoch": 0, "step": 573, "train_loss": 4.654191970825195, "perplexity": 105.02432298685004, "lr": 0.0010043469435376037, "grad_norm": 0.537598, "tokens_per_sec": 104566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:42.776627+00:00", "epoch": 0, "step": 574, "train_loss": 4.672204494476318, "perplexity": 106.93321647152506, "lr": 0.0010060997305245804, "grad_norm": 0.488075, "tokens_per_sec": 104625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:43.090525+00:00", "epoch": 0, "step": 575, "train_loss": 4.694957733154297, "perplexity": 109.39418479696617, "lr": 0.001007852517511557, "grad_norm": 0.505731, "tokens_per_sec": 104391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:43.404386+00:00", "epoch": 0, "step": 576, "train_loss": 4.695950031280518, "perplexity": 109.50279031717515, "lr": 0.0010096053044985336, "grad_norm": 0.61772, "tokens_per_sec": 104404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:43.718866+00:00", "epoch": 0, "step": 577, "train_loss": 4.612132549285889, "perplexity": 100.69866568967028, "lr": 0.0010113580914855103, "grad_norm": 0.613483, "tokens_per_sec": 104197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:44.031660+00:00", "epoch": 0, "step": 578, "train_loss": 4.514529228210449, "perplexity": 91.33455816634408, "lr": 0.001013110878472487, "grad_norm": 0.49507, "tokens_per_sec": 104759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:44.345308+00:00", "epoch": 0, "step": 579, "train_loss": 4.755421161651611, "perplexity": 116.21258714425763, "lr": 0.0010148636654594635, "grad_norm": 0.571335, "tokens_per_sec": 104474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:44.659002+00:00", "epoch": 0, "step": 580, "train_loss": 4.614821434020996, "perplexity": 100.96979715195502, "lr": 0.00101661645244644, "grad_norm": 0.53464, "tokens_per_sec": 104514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:44.972305+00:00", "epoch": 0, "step": 581, "train_loss": 4.575009346008301, "perplexity": 97.02894596603014, "lr": 0.0010183692394334168, "grad_norm": 0.540995, "tokens_per_sec": 104534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:45.285905+00:00", "epoch": 0, "step": 582, "train_loss": 4.6580281257629395, "perplexity": 105.42798632458387, "lr": 0.0010201220264203933, "grad_norm": 0.59158, "tokens_per_sec": 104490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:45.600992+00:00", "epoch": 0, "step": 583, "train_loss": 4.553782939910889, "perplexity": 94.99107500874861, "lr": 0.00102187481340737, "grad_norm": 0.597082, "tokens_per_sec": 103997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:45.915282+00:00", "epoch": 0, "step": 584, "train_loss": 4.57819128036499, "perplexity": 97.33817741897853, "lr": 0.0010236276003943467, "grad_norm": 0.50484, "tokens_per_sec": 104261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:46.229048+00:00", "epoch": 0, "step": 585, "train_loss": 4.56878662109375, "perplexity": 96.42703622817223, "lr": 0.0010253803873813232, "grad_norm": 0.59091, "tokens_per_sec": 104434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:46.543569+00:00", "epoch": 0, "step": 586, "train_loss": 4.666550636291504, "perplexity": 106.33033713346384, "lr": 0.0010271331743683, "grad_norm": 0.616371, "tokens_per_sec": 104184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:46.857490+00:00", "epoch": 0, "step": 587, "train_loss": 4.654463768005371, "perplexity": 105.05287218130727, "lr": 0.0010288859613552764, "grad_norm": 0.667039, "tokens_per_sec": 104383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:47.171596+00:00", "epoch": 0, "step": 588, "train_loss": 4.679917335510254, "perplexity": 107.7611641815503, "lr": 0.0010306387483422531, "grad_norm": 0.638556, "tokens_per_sec": 104322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:47.481459+00:00", "epoch": 0, "step": 589, "train_loss": 4.56715726852417, "perplexity": 96.2700505161838, "lr": 0.0010323915353292296, "grad_norm": 0.582806, "tokens_per_sec": 105750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:47.788180+00:00", "epoch": 0, "step": 590, "train_loss": 4.564202308654785, "perplexity": 95.98599627142359, "lr": 0.0010341443223162064, "grad_norm": 0.575096, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:48.095188+00:00", "epoch": 0, "step": 591, "train_loss": 4.6156182289123535, "perplexity": 101.0502814309801, "lr": 0.0010358971093031829, "grad_norm": 0.525355, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:48.401767+00:00", "epoch": 0, "step": 592, "train_loss": 4.629454612731934, "perplexity": 102.4581694886229, "lr": 0.0010376498962901596, "grad_norm": 0.476498, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:48.707318+00:00", "epoch": 0, "step": 593, "train_loss": 4.658754348754883, "perplexity": 105.50457836032953, "lr": 0.001039402683277136, "grad_norm": 0.459522, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:49.012100+00:00", "epoch": 0, "step": 594, "train_loss": 4.5071187019348145, "perplexity": 90.66022269950213, "lr": 0.0010411554702641128, "grad_norm": 0.50383, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:49.317636+00:00", "epoch": 0, "step": 595, "train_loss": 4.549360275268555, "perplexity": 94.57188898284947, "lr": 0.0010429082572510895, "grad_norm": 0.583254, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:49.624264+00:00", "epoch": 0, "step": 596, "train_loss": 4.5690155029296875, "perplexity": 96.44910915120735, "lr": 0.001044661044238066, "grad_norm": 0.534166, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:49.930686+00:00", "epoch": 0, "step": 597, "train_loss": 4.4954986572265625, "perplexity": 89.61284393769806, "lr": 0.0010464138312250427, "grad_norm": 0.477044, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:50.236714+00:00", "epoch": 0, "step": 598, "train_loss": 4.57972526550293, "perplexity": 97.48760731881403, "lr": 0.0010481666182120195, "grad_norm": 0.435715, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:50.542302+00:00", "epoch": 0, "step": 599, "train_loss": 4.5716753005981445, "perplexity": 96.70598573528225, "lr": 0.001049919405198996, "grad_norm": 0.441077, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:50.851986+00:00", "epoch": 0, "step": 600, "train_loss": 4.583675861358643, "perplexity": 97.8735032137364, "lr": 0.0010516721921859725, "grad_norm": 0.585083, "tokens_per_sec": 105875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:51.157987+00:00", "epoch": 0, "step": 601, "train_loss": 4.53695821762085, "perplexity": 93.40624608970046, "lr": 0.0010534249791729492, "grad_norm": 0.651617, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:51.465871+00:00", "epoch": 0, "step": 602, "train_loss": 4.531757354736328, "perplexity": 92.92171409563618, "lr": 0.0010551777661599257, "grad_norm": 0.617957, "tokens_per_sec": 106430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:51.771950+00:00", "epoch": 0, "step": 603, "train_loss": 4.60443115234375, "perplexity": 99.92612393737618, "lr": 0.0010569305531469024, "grad_norm": 0.692543, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:52.077521+00:00", "epoch": 0, "step": 604, "train_loss": 4.688990116119385, "perplexity": 108.74330622489273, "lr": 0.001058683340133879, "grad_norm": 0.622849, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:52.382603+00:00", "epoch": 0, "step": 605, "train_loss": 4.577705383300781, "perplexity": 97.29089257304943, "lr": 0.0010604361271208556, "grad_norm": 0.601655, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:52.688348+00:00", "epoch": 0, "step": 606, "train_loss": 4.572653770446777, "perplexity": 96.80065593492137, "lr": 0.0010621889141078323, "grad_norm": 0.590213, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:52.995010+00:00", "epoch": 0, "step": 607, "train_loss": 4.571630477905273, "perplexity": 96.70165120972813, "lr": 0.0010639417010948088, "grad_norm": 0.556837, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:53.301008+00:00", "epoch": 0, "step": 608, "train_loss": 4.636773586273193, "perplexity": 103.2108090349181, "lr": 0.0010656944880817856, "grad_norm": 0.571756, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:53.607799+00:00", "epoch": 0, "step": 609, "train_loss": 4.5556793212890625, "perplexity": 95.17138522892293, "lr": 0.0010674472750687623, "grad_norm": 0.627338, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:53.914081+00:00", "epoch": 0, "step": 610, "train_loss": 4.620580196380615, "perplexity": 101.55293568580392, "lr": 0.0010692000620557388, "grad_norm": 0.564693, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:54.219327+00:00", "epoch": 0, "step": 611, "train_loss": 4.477646350860596, "perplexity": 88.02724341784723, "lr": 0.0010709528490427153, "grad_norm": 0.558357, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:54.525469+00:00", "epoch": 0, "step": 612, "train_loss": 4.47296667098999, "perplexity": 87.61626646904936, "lr": 0.001072705636029692, "grad_norm": 0.502838, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:54.832412+00:00", "epoch": 0, "step": 613, "train_loss": 4.513450622558594, "perplexity": 91.23609730546237, "lr": 0.0010744584230166685, "grad_norm": 0.518369, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:55.139451+00:00", "epoch": 0, "step": 614, "train_loss": 4.510520935058594, "perplexity": 90.96919521240387, "lr": 0.0010762112100036452, "grad_norm": 0.448802, "tokens_per_sec": 106722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:55.445870+00:00", "epoch": 0, "step": 615, "train_loss": 4.605975151062012, "perplexity": 100.0805289145255, "lr": 0.001077963996990622, "grad_norm": 0.40241, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:55.751370+00:00", "epoch": 0, "step": 616, "train_loss": 4.562275409698486, "perplexity": 95.80121903609206, "lr": 0.0010797167839775984, "grad_norm": 0.435467, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:56.056818+00:00", "epoch": 0, "step": 617, "train_loss": 4.489563941955566, "perplexity": 89.08259222709103, "lr": 0.0010814695709645751, "grad_norm": 0.454401, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:56.363070+00:00", "epoch": 0, "step": 618, "train_loss": 4.501590251922607, "perplexity": 90.16039509921406, "lr": 0.0010832223579515519, "grad_norm": 0.528121, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:56.669975+00:00", "epoch": 0, "step": 619, "train_loss": 4.4568657875061035, "perplexity": 86.21686321715433, "lr": 0.0010849751449385284, "grad_norm": 0.61045, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:56.976810+00:00", "epoch": 0, "step": 620, "train_loss": 4.4767537117004395, "perplexity": 87.94870191303849, "lr": 0.001086727931925505, "grad_norm": 0.640448, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:57.280811+00:00", "epoch": 0, "step": 621, "train_loss": 4.437175273895264, "perplexity": 84.53581357774391, "lr": 0.0010884807189124816, "grad_norm": 0.638544, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:57.586264+00:00", "epoch": 0, "step": 622, "train_loss": 4.509487628936768, "perplexity": 90.87524473425482, "lr": 0.001090233505899458, "grad_norm": 0.562115, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:57.892022+00:00", "epoch": 0, "step": 623, "train_loss": 4.61605978012085, "perplexity": 101.0949101570737, "lr": 0.0010919862928864348, "grad_norm": 0.483067, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:58.198310+00:00", "epoch": 0, "step": 624, "train_loss": 4.663173675537109, "perplexity": 105.97186936441622, "lr": 0.0010937390798734113, "grad_norm": 0.439231, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:58.504812+00:00", "epoch": 0, "step": 625, "train_loss": 4.577986717224121, "perplexity": 97.31826765215088, "lr": 0.001095491866860388, "grad_norm": 0.479523, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:58.810628+00:00", "epoch": 0, "step": 626, "train_loss": 4.532973766326904, "perplexity": 93.03481391970101, "lr": 0.0010972446538473647, "grad_norm": 0.493187, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:59.115400+00:00", "epoch": 0, "step": 627, "train_loss": 4.432938575744629, "perplexity": 84.17841847487414, "lr": 0.0010989974408343412, "grad_norm": 0.4064, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:59.420501+00:00", "epoch": 0, "step": 628, "train_loss": 4.407933712005615, "perplexity": 82.0996466051884, "lr": 0.001100750227821318, "grad_norm": 0.417889, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:00:59.725664+00:00", "epoch": 0, "step": 629, "train_loss": 4.438573360443115, "perplexity": 84.65408461882959, "lr": 0.0011025030148082947, "grad_norm": 0.445968, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:00.031904+00:00", "epoch": 0, "step": 630, "train_loss": 4.533390045166016, "perplexity": 93.07355040606684, "lr": 0.0011042558017952712, "grad_norm": 0.474949, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:00.338371+00:00", "epoch": 0, "step": 631, "train_loss": 4.534160614013672, "perplexity": 93.14529762408593, "lr": 0.001106008588782248, "grad_norm": 0.525223, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:00.643516+00:00", "epoch": 0, "step": 632, "train_loss": 4.586376667022705, "perplexity": 98.13819780901085, "lr": 0.0011077613757692244, "grad_norm": 0.540007, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:00.948192+00:00", "epoch": 0, "step": 633, "train_loss": 4.634556293487549, "perplexity": 102.98221397738554, "lr": 0.001109514162756201, "grad_norm": 0.523149, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:01.253224+00:00", "epoch": 0, "step": 634, "train_loss": 4.560067176818848, "perplexity": 95.58990103985145, "lr": 0.0011112669497431776, "grad_norm": 0.534544, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:01.559427+00:00", "epoch": 0, "step": 635, "train_loss": 4.565690517425537, "perplexity": 96.12894981892164, "lr": 0.0011130197367301543, "grad_norm": 0.592608, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:01.870402+00:00", "epoch": 0, "step": 636, "train_loss": 4.480339527130127, "perplexity": 88.26463582720632, "lr": 0.0011147725237171308, "grad_norm": 0.596826, "tokens_per_sec": 105371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:02.176515+00:00", "epoch": 0, "step": 637, "train_loss": 4.540511608123779, "perplexity": 93.73874535747933, "lr": 0.0011165253107041076, "grad_norm": 0.573647, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:02.481589+00:00", "epoch": 0, "step": 638, "train_loss": 4.6112871170043945, "perplexity": 100.61356776432812, "lr": 0.001118278097691084, "grad_norm": 0.536494, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:02.787345+00:00", "epoch": 0, "step": 639, "train_loss": 4.412847518920898, "perplexity": 82.5040612103286, "lr": 0.0011200308846780608, "grad_norm": 0.470853, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:03.092417+00:00", "epoch": 0, "step": 640, "train_loss": 4.41187047958374, "perplexity": 82.42349086365927, "lr": 0.0011217836716650375, "grad_norm": 0.45064, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:03.398511+00:00", "epoch": 0, "step": 641, "train_loss": 4.484953880310059, "perplexity": 88.6728611537918, "lr": 0.001123536458652014, "grad_norm": 0.432153, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:03.704707+00:00", "epoch": 0, "step": 642, "train_loss": 4.504384994506836, "perplexity": 90.41272262581512, "lr": 0.0011252892456389905, "grad_norm": 0.467862, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:04.008566+00:00", "epoch": 0, "step": 643, "train_loss": 4.508341312408447, "perplexity": 90.77113262331821, "lr": 0.0011270420326259672, "grad_norm": 0.49574, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:04.313764+00:00", "epoch": 0, "step": 644, "train_loss": 4.496028423309326, "perplexity": 89.66033036023367, "lr": 0.0011287948196129437, "grad_norm": 0.545801, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:04.619834+00:00", "epoch": 0, "step": 645, "train_loss": 4.48207426071167, "perplexity": 88.41788433927763, "lr": 0.0011305476065999204, "grad_norm": 0.504547, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:04.926041+00:00", "epoch": 0, "step": 646, "train_loss": 4.395578861236572, "perplexity": 81.09155793875955, "lr": 0.0011323003935868972, "grad_norm": 0.443981, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:05.231498+00:00", "epoch": 0, "step": 647, "train_loss": 4.430032253265381, "perplexity": 83.93412401618849, "lr": 0.0011340531805738737, "grad_norm": 0.491796, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:05.536166+00:00", "epoch": 0, "step": 648, "train_loss": 4.412810325622559, "perplexity": 82.5009926692307, "lr": 0.0011358059675608504, "grad_norm": 0.495505, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:05.842731+00:00", "epoch": 0, "step": 649, "train_loss": 4.383809566497803, "perplexity": 80.14276177535181, "lr": 0.001137558754547827, "grad_norm": 0.48596, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:06.148421+00:00", "epoch": 0, "step": 650, "train_loss": 4.436734676361084, "perplexity": 84.49857551083433, "lr": 0.0011393115415348036, "grad_norm": 0.535508, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:06.452868+00:00", "epoch": 0, "step": 651, "train_loss": 4.614406585693359, "perplexity": 100.92791868767075, "lr": 0.0011410643285217803, "grad_norm": 0.52168, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:06.757970+00:00", "epoch": 0, "step": 652, "train_loss": 4.447877407073975, "perplexity": 85.44538561202775, "lr": 0.0011428171155087568, "grad_norm": 0.508704, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:07.064710+00:00", "epoch": 0, "step": 653, "train_loss": 4.335502624511719, "perplexity": 76.36333132761771, "lr": 0.0011445699024957333, "grad_norm": 0.505676, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:07.369740+00:00", "epoch": 0, "step": 654, "train_loss": 4.459987163543701, "perplexity": 86.48639891032393, "lr": 0.00114632268948271, "grad_norm": 0.436452, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:07.675088+00:00", "epoch": 0, "step": 655, "train_loss": 4.599192142486572, "perplexity": 99.40397894474567, "lr": 0.0011480754764696865, "grad_norm": 0.595017, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:07.980244+00:00", "epoch": 0, "step": 656, "train_loss": 4.413758754730225, "perplexity": 82.57927612937755, "lr": 0.0011498282634566633, "grad_norm": 0.629703, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:08.287062+00:00", "epoch": 0, "step": 657, "train_loss": 4.407283306121826, "perplexity": 82.04626587343218, "lr": 0.00115158105044364, "grad_norm": 0.500367, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:08.592998+00:00", "epoch": 0, "step": 658, "train_loss": 4.4664483070373535, "perplexity": 87.04700908503867, "lr": 0.0011533338374306165, "grad_norm": 0.442765, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:08.897908+00:00", "epoch": 0, "step": 659, "train_loss": 4.433405876159668, "perplexity": 84.21776427720366, "lr": 0.0011550866244175932, "grad_norm": 0.47462, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:09.203240+00:00", "epoch": 0, "step": 660, "train_loss": 4.4357476234436035, "perplexity": 84.41521209417346, "lr": 0.00115683941140457, "grad_norm": 0.482988, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:09.507482+00:00", "epoch": 0, "step": 661, "train_loss": 4.40923547744751, "perplexity": 82.20659068087953, "lr": 0.0011585921983915464, "grad_norm": 0.487556, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:09.815785+00:00", "epoch": 0, "step": 662, "train_loss": 4.3774261474609375, "perplexity": 79.63280630553432, "lr": 0.0011603449853785231, "grad_norm": 0.567103, "tokens_per_sec": 106284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:10.121283+00:00", "epoch": 0, "step": 663, "train_loss": 4.555737018585205, "perplexity": 95.17687651893556, "lr": 0.0011620977723654996, "grad_norm": 0.567278, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:10.427283+00:00", "epoch": 0, "step": 664, "train_loss": 4.488076686859131, "perplexity": 88.95020216110089, "lr": 0.0011638505593524761, "grad_norm": 0.628777, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:10.732407+00:00", "epoch": 0, "step": 665, "train_loss": 4.569468975067139, "perplexity": 96.49285605314031, "lr": 0.0011656033463394529, "grad_norm": 0.581575, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:11.038749+00:00", "epoch": 0, "step": 666, "train_loss": 4.514797210693359, "perplexity": 91.35903750788788, "lr": 0.0011673561333264296, "grad_norm": 0.540023, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:11.344210+00:00", "epoch": 0, "step": 667, "train_loss": 4.370883941650391, "perplexity": 79.1135325474058, "lr": 0.001169108920313406, "grad_norm": 0.634604, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:11.650113+00:00", "epoch": 0, "step": 668, "train_loss": 4.395516395568848, "perplexity": 81.08649265865078, "lr": 0.0011708617073003828, "grad_norm": 0.653772, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:11.955845+00:00", "epoch": 0, "step": 669, "train_loss": 4.576274394989014, "perplexity": 97.15177000805859, "lr": 0.0011726144942873593, "grad_norm": 0.577228, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:12.263190+00:00", "epoch": 0, "step": 670, "train_loss": 4.495883941650391, "perplexity": 89.64737702274489, "lr": 0.001174367281274336, "grad_norm": 0.509916, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:12.569023+00:00", "epoch": 0, "step": 671, "train_loss": 4.355301856994629, "perplexity": 77.89033353058339, "lr": 0.0011761200682613127, "grad_norm": 0.466457, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:12.874504+00:00", "epoch": 0, "step": 672, "train_loss": 4.526317119598389, "perplexity": 92.41757069425064, "lr": 0.0011778728552482892, "grad_norm": 0.413939, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:13.179925+00:00", "epoch": 0, "step": 673, "train_loss": 4.35439395904541, "perplexity": 77.81964914846509, "lr": 0.001179625642235266, "grad_norm": 0.465127, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:13.485023+00:00", "epoch": 0, "step": 674, "train_loss": 4.456063270568848, "perplexity": 86.14770047999141, "lr": 0.0011813784292222425, "grad_norm": 0.448326, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:13.790592+00:00", "epoch": 0, "step": 675, "train_loss": 4.325713157653809, "perplexity": 75.61942220423248, "lr": 0.001183131216209219, "grad_norm": 0.403884, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:14.095871+00:00", "epoch": 0, "step": 676, "train_loss": 4.37795352935791, "perplexity": 79.67481428213453, "lr": 0.0011848840031961957, "grad_norm": 0.411295, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:14.401758+00:00", "epoch": 0, "step": 677, "train_loss": 4.399261474609375, "perplexity": 81.39073733735975, "lr": 0.0011866367901831724, "grad_norm": 0.396557, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:14.707363+00:00", "epoch": 0, "step": 678, "train_loss": 4.463200092315674, "perplexity": 86.7647204239427, "lr": 0.001188389577170149, "grad_norm": 0.436403, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:15.012111+00:00", "epoch": 0, "step": 679, "train_loss": 4.4481520652771, "perplexity": 85.46885711127763, "lr": 0.0011901423641571256, "grad_norm": 0.489774, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:15.318292+00:00", "epoch": 0, "step": 680, "train_loss": 4.482237815856934, "perplexity": 88.43234672186095, "lr": 0.0011918951511441023, "grad_norm": 0.585093, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:15.624312+00:00", "epoch": 0, "step": 681, "train_loss": 4.416311740875244, "perplexity": 82.79036922141712, "lr": 0.0011936479381310788, "grad_norm": 0.563459, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:15.930397+00:00", "epoch": 0, "step": 682, "train_loss": 4.52533483505249, "perplexity": 92.3268349142612, "lr": 0.0011954007251180556, "grad_norm": 0.565048, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:16.235580+00:00", "epoch": 0, "step": 683, "train_loss": 4.5396623611450195, "perplexity": 93.65917180477388, "lr": 0.001197153512105032, "grad_norm": 0.555788, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:16.539956+00:00", "epoch": 0, "step": 684, "train_loss": 4.48345422744751, "perplexity": 88.53998230470857, "lr": 0.0011989062990920088, "grad_norm": 0.584471, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:16.844953+00:00", "epoch": 0, "step": 685, "train_loss": 4.483532905578613, "perplexity": 88.54694873909366, "lr": 0.0012006590860789853, "grad_norm": 0.574015, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:17.149930+00:00", "epoch": 0, "step": 686, "train_loss": 4.411341667175293, "perplexity": 82.37991582146998, "lr": 0.0012024118730659618, "grad_norm": 0.480873, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:17.454874+00:00", "epoch": 0, "step": 687, "train_loss": 4.422480583190918, "perplexity": 83.30266847645026, "lr": 0.0012041646600529385, "grad_norm": 0.438309, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:17.758950+00:00", "epoch": 0, "step": 688, "train_loss": 4.41759729385376, "perplexity": 82.89686906809246, "lr": 0.0012059174470399152, "grad_norm": 0.425418, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:18.063892+00:00", "epoch": 0, "step": 689, "train_loss": 4.465351104736328, "perplexity": 86.95155328311218, "lr": 0.0012076702340268917, "grad_norm": 0.396972, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:18.369739+00:00", "epoch": 0, "step": 690, "train_loss": 4.358333110809326, "perplexity": 78.12679711072859, "lr": 0.0012094230210138684, "grad_norm": 0.392715, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:18.674679+00:00", "epoch": 0, "step": 691, "train_loss": 4.337828636169434, "perplexity": 76.54116006218652, "lr": 0.0012111758080008451, "grad_norm": 0.432863, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:18.980615+00:00", "epoch": 0, "step": 692, "train_loss": 4.3595075607299805, "perplexity": 78.21860702391585, "lr": 0.0012129285949878217, "grad_norm": 0.4182, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:19.285985+00:00", "epoch": 0, "step": 693, "train_loss": 4.413595199584961, "perplexity": 82.56577096832383, "lr": 0.0012146813819747984, "grad_norm": 0.406911, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:19.590792+00:00", "epoch": 0, "step": 694, "train_loss": 4.32295036315918, "perplexity": 75.41078961806811, "lr": 0.0012164341689617749, "grad_norm": 0.454146, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:19.895647+00:00", "epoch": 0, "step": 695, "train_loss": 4.3053364753723145, "perplexity": 74.09414211306358, "lr": 0.0012181869559487516, "grad_norm": 0.4131, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:20.200452+00:00", "epoch": 0, "step": 696, "train_loss": 4.307303428649902, "perplexity": 74.2400252543704, "lr": 0.001219939742935728, "grad_norm": 0.43286, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:20.506981+00:00", "epoch": 0, "step": 697, "train_loss": 4.352556228637695, "perplexity": 77.67676894077032, "lr": 0.0012216925299227048, "grad_norm": 0.452711, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:20.812718+00:00", "epoch": 0, "step": 698, "train_loss": 4.319825649261475, "perplexity": 75.17552024188397, "lr": 0.0012234453169096813, "grad_norm": 0.454963, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:21.118640+00:00", "epoch": 0, "step": 699, "train_loss": 4.3617262840271, "perplexity": 78.39234513671506, "lr": 0.001225198103896658, "grad_norm": 0.45797, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:21.425271+00:00", "epoch": 0, "step": 700, "train_loss": 4.306266784667969, "perplexity": 74.16310465549824, "lr": 0.0012269508908836347, "grad_norm": 0.478926, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:21.730294+00:00", "epoch": 0, "step": 701, "train_loss": 4.38456916809082, "perplexity": 80.20366147169014, "lr": 0.0012287036778706112, "grad_norm": 0.439866, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:22.037452+00:00", "epoch": 0, "step": 702, "train_loss": 4.407656669616699, "perplexity": 82.07690467334875, "lr": 0.001230456464857588, "grad_norm": 0.456553, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:22.343722+00:00", "epoch": 0, "step": 703, "train_loss": 4.336751937866211, "perplexity": 76.45879267539091, "lr": 0.0012322092518445645, "grad_norm": 0.437747, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:22.649621+00:00", "epoch": 0, "step": 704, "train_loss": 4.395469665527344, "perplexity": 81.08270357201621, "lr": 0.0012339620388315412, "grad_norm": 0.444304, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:22.954210+00:00", "epoch": 0, "step": 705, "train_loss": 4.284361362457275, "perplexity": 72.55619481033192, "lr": 0.0012357148258185177, "grad_norm": 0.528752, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:23.259014+00:00", "epoch": 0, "step": 706, "train_loss": 4.392966270446777, "perplexity": 80.87997539101507, "lr": 0.0012374676128054942, "grad_norm": 0.664467, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:23.564840+00:00", "epoch": 0, "step": 707, "train_loss": 4.373108863830566, "perplexity": 79.2897499630471, "lr": 0.001239220399792471, "grad_norm": 0.621814, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:23.871093+00:00", "epoch": 0, "step": 708, "train_loss": 4.436265468597412, "perplexity": 84.45893742316187, "lr": 0.0012409731867794476, "grad_norm": 0.563864, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:24.176164+00:00", "epoch": 0, "step": 709, "train_loss": 4.478471279144287, "perplexity": 88.09988954047896, "lr": 0.0012427259737664241, "grad_norm": 0.637567, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:24.481209+00:00", "epoch": 0, "step": 710, "train_loss": 4.397420883178711, "perplexity": 81.24106802596103, "lr": 0.0012444787607534008, "grad_norm": 0.626815, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:24.786401+00:00", "epoch": 0, "step": 711, "train_loss": 4.358090877532959, "perplexity": 78.1078744926288, "lr": 0.0012462315477403776, "grad_norm": 0.507204, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:25.092663+00:00", "epoch": 0, "step": 712, "train_loss": 4.342854022979736, "perplexity": 76.92677712430998, "lr": 0.001247984334727354, "grad_norm": 0.480616, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:25.398161+00:00", "epoch": 0, "step": 713, "train_loss": 4.374671936035156, "perplexity": 79.41378247795852, "lr": 0.0012497371217143308, "grad_norm": 0.474332, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:25.703351+00:00", "epoch": 0, "step": 714, "train_loss": 4.299160957336426, "perplexity": 73.63798236318561, "lr": 0.0012514899087013073, "grad_norm": 0.436205, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:26.008305+00:00", "epoch": 0, "step": 715, "train_loss": 4.435781955718994, "perplexity": 84.41811031023312, "lr": 0.001253242695688284, "grad_norm": 0.370625, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:26.313842+00:00", "epoch": 0, "step": 716, "train_loss": 4.363221168518066, "perplexity": 78.50962027222172, "lr": 0.0012549954826752605, "grad_norm": 0.444233, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:26.620842+00:00", "epoch": 0, "step": 717, "train_loss": 4.289638519287109, "perplexity": 72.94009729521049, "lr": 0.0012567482696622372, "grad_norm": 0.415023, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:26.927077+00:00", "epoch": 0, "step": 718, "train_loss": 4.3628644943237305, "perplexity": 78.48162290992863, "lr": 0.0012585010566492137, "grad_norm": 0.393926, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:27.232290+00:00", "epoch": 0, "step": 719, "train_loss": 4.4659318923950195, "perplexity": 87.0020683400039, "lr": 0.0012602538436361904, "grad_norm": 0.395553, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:27.537962+00:00", "epoch": 0, "step": 720, "train_loss": 4.342090606689453, "perplexity": 76.86807238042431, "lr": 0.001262006630623167, "grad_norm": 0.379602, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:27.843085+00:00", "epoch": 0, "step": 721, "train_loss": 4.3205742835998535, "perplexity": 75.23182028918743, "lr": 0.0012637594176101437, "grad_norm": 0.356024, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:28.148045+00:00", "epoch": 0, "step": 722, "train_loss": 4.286542892456055, "perplexity": 72.71465110164235, "lr": 0.0012655122045971204, "grad_norm": 0.397651, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:28.453630+00:00", "epoch": 0, "step": 723, "train_loss": 4.205728054046631, "perplexity": 67.06941005051407, "lr": 0.0012672649915840969, "grad_norm": 0.421779, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:28.759093+00:00", "epoch": 0, "step": 724, "train_loss": 4.437251091003418, "perplexity": 84.5422230816368, "lr": 0.0012690177785710736, "grad_norm": 0.403393, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:29.062827+00:00", "epoch": 0, "step": 725, "train_loss": 4.297603130340576, "perplexity": 73.52335643320585, "lr": 0.00127077056555805, "grad_norm": 0.35865, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:29.366944+00:00", "epoch": 0, "step": 726, "train_loss": 4.29234504699707, "perplexity": 73.13777908465461, "lr": 0.0012725233525450268, "grad_norm": 0.361055, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:29.673071+00:00", "epoch": 0, "step": 727, "train_loss": 4.306610584259033, "perplexity": 74.18860628402442, "lr": 0.0012742761395320033, "grad_norm": 0.371709, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:29.980052+00:00", "epoch": 0, "step": 728, "train_loss": 4.160996437072754, "perplexity": 64.13539765881629, "lr": 0.00127602892651898, "grad_norm": 0.39676, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:30.286141+00:00", "epoch": 0, "step": 729, "train_loss": 4.375532627105713, "perplexity": 79.48216263429013, "lr": 0.0012777817135059565, "grad_norm": 0.383349, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:30.591848+00:00", "epoch": 0, "step": 730, "train_loss": 4.358987331390381, "perplexity": 78.17792599229, "lr": 0.0012795345004929333, "grad_norm": 0.371666, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:30.896602+00:00", "epoch": 0, "step": 731, "train_loss": 4.236160755157471, "perplexity": 69.14188897651235, "lr": 0.00128128728747991, "grad_norm": 0.402605, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:31.200787+00:00", "epoch": 0, "step": 732, "train_loss": 4.314051151275635, "perplexity": 74.74267029999642, "lr": 0.0012830400744668865, "grad_norm": 0.408446, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:31.507925+00:00", "epoch": 0, "step": 733, "train_loss": 4.403502941131592, "perplexity": 81.73668657240762, "lr": 0.0012847928614538632, "grad_norm": 0.447592, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:31.814668+00:00", "epoch": 0, "step": 734, "train_loss": 4.296802997589111, "perplexity": 73.46455151673312, "lr": 0.00128654564844084, "grad_norm": 0.510915, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:32.121201+00:00", "epoch": 0, "step": 735, "train_loss": 4.243221282958984, "perplexity": 69.63179466706313, "lr": 0.0012882984354278164, "grad_norm": 0.563851, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:32.426705+00:00", "epoch": 0, "step": 736, "train_loss": 4.31935453414917, "perplexity": 75.14011225949535, "lr": 0.001290051222414793, "grad_norm": 0.717131, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:32.731600+00:00", "epoch": 0, "step": 737, "train_loss": 4.354048728942871, "perplexity": 77.79278809989808, "lr": 0.0012918040094017696, "grad_norm": 0.67452, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:33.036443+00:00", "epoch": 0, "step": 738, "train_loss": 4.354918003082275, "perplexity": 77.86044075891887, "lr": 0.0012935567963887461, "grad_norm": 0.655713, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:33.342350+00:00", "epoch": 0, "step": 739, "train_loss": 4.439289569854736, "perplexity": 84.71473638805709, "lr": 0.0012953095833757229, "grad_norm": 0.649154, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:33.648506+00:00", "epoch": 0, "step": 740, "train_loss": 4.366182327270508, "perplexity": 78.74244426571907, "lr": 0.0012970623703626994, "grad_norm": 0.623684, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:33.953534+00:00", "epoch": 0, "step": 741, "train_loss": 4.24444055557251, "perplexity": 69.71674658658146, "lr": 0.001298815157349676, "grad_norm": 0.56587, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:34.258106+00:00", "epoch": 0, "step": 742, "train_loss": 4.391821384429932, "perplexity": 80.78743002520345, "lr": 0.0013005679443366528, "grad_norm": 0.49048, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:34.563571+00:00", "epoch": 0, "step": 743, "train_loss": 4.344256401062012, "perplexity": 77.03473323043495, "lr": 0.0013023207313236293, "grad_norm": 0.491241, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:34.868772+00:00", "epoch": 0, "step": 744, "train_loss": 4.428248405456543, "perplexity": 83.78453177753734, "lr": 0.001304073518310606, "grad_norm": 0.479352, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:35.174250+00:00", "epoch": 0, "step": 745, "train_loss": 4.370233058929443, "perplexity": 79.06205567059976, "lr": 0.0013058263052975825, "grad_norm": 0.450157, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:35.478771+00:00", "epoch": 0, "step": 746, "train_loss": 4.391493320465088, "perplexity": 80.76093092753703, "lr": 0.0013075790922845592, "grad_norm": 0.47168, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:35.784165+00:00", "epoch": 0, "step": 747, "train_loss": 4.371370315551758, "perplexity": 79.15202066393078, "lr": 0.0013093318792715357, "grad_norm": 0.431368, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:36.088610+00:00", "epoch": 0, "step": 748, "train_loss": 4.221080303192139, "perplexity": 68.10702079337807, "lr": 0.0013110846662585125, "grad_norm": 0.393001, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:36.393784+00:00", "epoch": 0, "step": 749, "train_loss": 4.297614097595215, "perplexity": 73.52416278699947, "lr": 0.001312837453245489, "grad_norm": 0.38426, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:36.699531+00:00", "epoch": 0, "step": 750, "train_loss": 4.283771514892578, "perplexity": 72.51341033491913, "lr": 0.0013145902402324657, "grad_norm": 0.427597, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:37.004572+00:00", "epoch": 0, "step": 751, "train_loss": 4.260406017303467, "perplexity": 70.83873937012187, "lr": 0.0013163430272194424, "grad_norm": 0.420224, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:37.308800+00:00", "epoch": 0, "step": 752, "train_loss": 4.297423362731934, "perplexity": 73.51014050317413, "lr": 0.001318095814206419, "grad_norm": 0.41894, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:37.613431+00:00", "epoch": 0, "step": 753, "train_loss": 4.237712860107422, "perplexity": 69.24928777018083, "lr": 0.0013198486011933956, "grad_norm": 0.365143, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:37.918466+00:00", "epoch": 0, "step": 754, "train_loss": 4.386749744415283, "perplexity": 80.37874249642466, "lr": 0.0013216013881803721, "grad_norm": 0.333155, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:38.224026+00:00", "epoch": 0, "step": 755, "train_loss": 4.328475475311279, "perplexity": 75.8285958384401, "lr": 0.0013233541751673488, "grad_norm": 0.343409, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:38.529794+00:00", "epoch": 0, "step": 756, "train_loss": 4.26417350769043, "perplexity": 71.10612701363706, "lr": 0.0013251069621543253, "grad_norm": 0.369301, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:38.834987+00:00", "epoch": 0, "step": 757, "train_loss": 4.343482971191406, "perplexity": 76.97517530158994, "lr": 0.001326859749141302, "grad_norm": 0.370904, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:39.139395+00:00", "epoch": 0, "step": 758, "train_loss": 4.275366306304932, "perplexity": 71.90647428092608, "lr": 0.0013286125361282786, "grad_norm": 0.369484, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:39.445186+00:00", "epoch": 0, "step": 759, "train_loss": 4.315769672393799, "perplexity": 74.87122758990456, "lr": 0.0013303653231152553, "grad_norm": 0.389064, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:39.751860+00:00", "epoch": 0, "step": 760, "train_loss": 4.329629898071289, "perplexity": 75.91618464286283, "lr": 0.0013321181101022318, "grad_norm": 0.397071, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:40.056568+00:00", "epoch": 0, "step": 761, "train_loss": 4.264817237854004, "perplexity": 71.15191490835399, "lr": 0.0013338708970892083, "grad_norm": 0.410153, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:40.361583+00:00", "epoch": 0, "step": 762, "train_loss": 4.296492576599121, "perplexity": 73.44175011712173, "lr": 0.001335623684076185, "grad_norm": 0.408176, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:40.667210+00:00", "epoch": 0, "step": 763, "train_loss": 4.22233772277832, "perplexity": 68.19271375999666, "lr": 0.0013373764710631617, "grad_norm": 0.409503, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:40.972598+00:00", "epoch": 0, "step": 764, "train_loss": 4.2994866371154785, "perplexity": 73.6619686707272, "lr": 0.0013391292580501384, "grad_norm": 0.446014, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:41.278522+00:00", "epoch": 0, "step": 765, "train_loss": 4.268945217132568, "perplexity": 71.44623559563456, "lr": 0.0013408820450371152, "grad_norm": 0.508407, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:41.583727+00:00", "epoch": 0, "step": 766, "train_loss": 4.120856761932373, "perplexity": 61.61200647974435, "lr": 0.0013426348320240917, "grad_norm": 0.468902, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:41.887877+00:00", "epoch": 0, "step": 767, "train_loss": 4.282839298248291, "perplexity": 72.44584362516973, "lr": 0.0013443876190110682, "grad_norm": 0.441789, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:42.193203+00:00", "epoch": 0, "step": 768, "train_loss": 4.36619758605957, "perplexity": 78.74364578923326, "lr": 0.0013461404059980449, "grad_norm": 0.394784, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:42.498767+00:00", "epoch": 0, "step": 769, "train_loss": 4.227348327636719, "perplexity": 68.53525796302998, "lr": 0.0013478931929850214, "grad_norm": 0.329934, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:42.803980+00:00", "epoch": 0, "step": 770, "train_loss": 4.2521162033081055, "perplexity": 70.25392674011282, "lr": 0.001349645979971998, "grad_norm": 0.330527, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:43.108829+00:00", "epoch": 0, "step": 771, "train_loss": 4.2659125328063965, "perplexity": 71.22988993663826, "lr": 0.0013513987669589748, "grad_norm": 0.342977, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:43.413967+00:00", "epoch": 0, "step": 772, "train_loss": 4.356150150299072, "perplexity": 77.95643541192601, "lr": 0.0013531515539459515, "grad_norm": 0.384879, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:43.720267+00:00", "epoch": 0, "step": 773, "train_loss": 4.425792217254639, "perplexity": 83.57899372248806, "lr": 0.0013549043409329278, "grad_norm": 0.436984, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:44.025550+00:00", "epoch": 0, "step": 774, "train_loss": 4.343355655670166, "perplexity": 76.96537579085201, "lr": 0.0013566571279199045, "grad_norm": 0.467542, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:44.331243+00:00", "epoch": 0, "step": 775, "train_loss": 4.414602279663086, "perplexity": 82.64896319498693, "lr": 0.0013584099149068813, "grad_norm": 0.497916, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:44.635910+00:00", "epoch": 0, "step": 776, "train_loss": 4.37298583984375, "perplexity": 79.27999602188964, "lr": 0.001360162701893858, "grad_norm": 0.479958, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:44.940669+00:00", "epoch": 0, "step": 777, "train_loss": 4.3208327293396, "perplexity": 75.25126614537561, "lr": 0.0013619154888808345, "grad_norm": 0.426975, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:45.246850+00:00", "epoch": 0, "step": 778, "train_loss": 4.326664924621582, "perplexity": 75.69142853359583, "lr": 0.001363668275867811, "grad_norm": 0.463234, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:45.551835+00:00", "epoch": 0, "step": 779, "train_loss": 4.286052227020264, "perplexity": 72.67898128735929, "lr": 0.0013654210628547877, "grad_norm": 0.409786, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:45.856420+00:00", "epoch": 0, "step": 780, "train_loss": 4.375586986541748, "perplexity": 79.48648335726075, "lr": 0.0013671738498417642, "grad_norm": 0.380992, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:46.161090+00:00", "epoch": 0, "step": 781, "train_loss": 4.307785511016846, "perplexity": 74.27582368967086, "lr": 0.001368926636828741, "grad_norm": 0.486429, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:46.465882+00:00", "epoch": 0, "step": 782, "train_loss": 4.276220798492432, "perplexity": 71.96794406041275, "lr": 0.0013706794238157176, "grad_norm": 0.573558, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:46.771429+00:00", "epoch": 0, "step": 783, "train_loss": 4.241230010986328, "perplexity": 69.4932767850954, "lr": 0.0013724322108026943, "grad_norm": 0.452574, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:47.075816+00:00", "epoch": 0, "step": 784, "train_loss": 4.30191707611084, "perplexity": 73.84121733008976, "lr": 0.0013741849977896706, "grad_norm": 0.469097, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:47.380133+00:00", "epoch": 0, "step": 785, "train_loss": 4.283483505249023, "perplexity": 72.49252878064448, "lr": 0.0013759377847766473, "grad_norm": 0.456878, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:47.685484+00:00", "epoch": 0, "step": 786, "train_loss": 4.287106513977051, "perplexity": 72.75564619566833, "lr": 0.001377690571763624, "grad_norm": 0.486942, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:47.990102+00:00", "epoch": 0, "step": 787, "train_loss": 4.270462512969971, "perplexity": 71.55472295439763, "lr": 0.0013794433587506008, "grad_norm": 0.508256, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:48.295407+00:00", "epoch": 0, "step": 788, "train_loss": 4.323934078216553, "perplexity": 75.485008846605, "lr": 0.0013811961457375773, "grad_norm": 0.448696, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:48.600508+00:00", "epoch": 0, "step": 789, "train_loss": 4.174594402313232, "perplexity": 65.01346500161472, "lr": 0.001382948932724554, "grad_norm": 0.437447, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:48.905476+00:00", "epoch": 0, "step": 790, "train_loss": 4.389307975769043, "perplexity": 80.58463316134359, "lr": 0.0013847017197115305, "grad_norm": 0.379558, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:49.209707+00:00", "epoch": 0, "step": 791, "train_loss": 4.237423896789551, "perplexity": 69.2292801570989, "lr": 0.001386454506698507, "grad_norm": 0.371968, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:49.514765+00:00", "epoch": 0, "step": 792, "train_loss": 4.335084438323975, "perplexity": 76.33140391347347, "lr": 0.0013882072936854837, "grad_norm": 0.405325, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:49.819830+00:00", "epoch": 0, "step": 793, "train_loss": 4.320063591003418, "perplexity": 75.1934097643695, "lr": 0.0013899600806724604, "grad_norm": 0.395565, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:50.126105+00:00", "epoch": 0, "step": 794, "train_loss": 4.217272758483887, "perplexity": 67.84819332802809, "lr": 0.0013917128676594372, "grad_norm": 0.365169, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:50.431809+00:00", "epoch": 0, "step": 795, "train_loss": 4.22783088684082, "perplexity": 68.56833826352532, "lr": 0.0013934656546464134, "grad_norm": 0.427973, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:50.737440+00:00", "epoch": 0, "step": 796, "train_loss": 4.325591087341309, "perplexity": 75.61019188111862, "lr": 0.0013952184416333902, "grad_norm": 0.428499, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:51.041983+00:00", "epoch": 0, "step": 797, "train_loss": 4.2135090827941895, "perplexity": 67.59331467337603, "lr": 0.0013969712286203669, "grad_norm": 0.42548, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:51.346931+00:00", "epoch": 0, "step": 798, "train_loss": 4.20548152923584, "perplexity": 67.05287781477837, "lr": 0.0013987240156073436, "grad_norm": 0.410219, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:51.653227+00:00", "epoch": 0, "step": 799, "train_loss": 4.16830587387085, "perplexity": 64.6059087853299, "lr": 0.00140047680259432, "grad_norm": 0.391979, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:51.959133+00:00", "epoch": 0, "step": 800, "train_loss": 4.337080001831055, "perplexity": 76.48388016498882, "lr": 0.0014022295895812968, "grad_norm": 0.428934, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:52.263932+00:00", "epoch": 0, "step": 801, "train_loss": 4.17711067199707, "perplexity": 65.17726240543516, "lr": 0.0014039823765682733, "grad_norm": 0.432264, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:52.568906+00:00", "epoch": 0, "step": 802, "train_loss": 4.315268516540527, "perplexity": 74.83371483660933, "lr": 0.0014057351635552498, "grad_norm": 0.406124, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:52.873749+00:00", "epoch": 0, "step": 803, "train_loss": 4.223111629486084, "perplexity": 68.24550898525055, "lr": 0.0014074879505422265, "grad_norm": 0.404572, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:53.178668+00:00", "epoch": 0, "step": 804, "train_loss": 4.233608245849609, "perplexity": 68.96562871002119, "lr": 0.0014092407375292033, "grad_norm": 0.431745, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:53.483524+00:00", "epoch": 0, "step": 805, "train_loss": 4.275732040405273, "perplexity": 71.9327777403488, "lr": 0.00141099352451618, "grad_norm": 0.427042, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:53.789337+00:00", "epoch": 0, "step": 806, "train_loss": 4.22309684753418, "perplexity": 68.24450019087504, "lr": 0.0014127463115031565, "grad_norm": 0.420197, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:54.094824+00:00", "epoch": 0, "step": 807, "train_loss": 4.216907024383545, "perplexity": 67.82338346726405, "lr": 0.001414499098490133, "grad_norm": 0.452682, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:54.399599+00:00", "epoch": 0, "step": 808, "train_loss": 4.352148532867432, "perplexity": 77.64510690530294, "lr": 0.0014162518854771097, "grad_norm": 0.404648, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:54.704205+00:00", "epoch": 0, "step": 809, "train_loss": 4.230639457702637, "perplexity": 68.7611879897086, "lr": 0.0014180046724640862, "grad_norm": 0.361051, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:55.009174+00:00", "epoch": 0, "step": 810, "train_loss": 4.138786315917969, "perplexity": 62.72664490098498, "lr": 0.001419757459451063, "grad_norm": 0.356968, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:55.315400+00:00", "epoch": 0, "step": 811, "train_loss": 4.178848743438721, "perplexity": 65.29064364767387, "lr": 0.0014215102464380396, "grad_norm": 0.351463, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:55.621106+00:00", "epoch": 0, "step": 812, "train_loss": 4.3277692794799805, "perplexity": 75.7750649040518, "lr": 0.001423263033425016, "grad_norm": 0.318191, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:55.926057+00:00", "epoch": 0, "step": 813, "train_loss": 4.201010227203369, "perplexity": 66.7537334267651, "lr": 0.0014250158204119926, "grad_norm": 0.351557, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:56.230940+00:00", "epoch": 0, "step": 814, "train_loss": 4.185235500335693, "perplexity": 65.70897358009366, "lr": 0.0014267686073989694, "grad_norm": 0.334153, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:56.536271+00:00", "epoch": 0, "step": 815, "train_loss": 4.1976823806762695, "perplexity": 66.5319564713189, "lr": 0.001428521394385946, "grad_norm": 0.330554, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:56.842610+00:00", "epoch": 0, "step": 816, "train_loss": 4.192288398742676, "perplexity": 66.17405043722071, "lr": 0.0014302741813729228, "grad_norm": 0.332686, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:57.147653+00:00", "epoch": 0, "step": 817, "train_loss": 4.257744789123535, "perplexity": 70.65047194308164, "lr": 0.0014320269683598993, "grad_norm": 0.390389, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:57.451814+00:00", "epoch": 0, "step": 818, "train_loss": 4.3080668449401855, "perplexity": 74.29672293825473, "lr": 0.0014337797553468758, "grad_norm": 0.430615, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:57.757975+00:00", "epoch": 0, "step": 819, "train_loss": 4.293684482574463, "perplexity": 73.23580806510805, "lr": 0.0014355325423338525, "grad_norm": 0.426073, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:58.063220+00:00", "epoch": 0, "step": 820, "train_loss": 4.178183555603027, "perplexity": 65.24722754726666, "lr": 0.001437285329320829, "grad_norm": 0.384166, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:58.368260+00:00", "epoch": 0, "step": 821, "train_loss": 4.303259372711182, "perplexity": 73.94040069693305, "lr": 0.0014390381163078057, "grad_norm": 0.383535, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:58.673874+00:00", "epoch": 0, "step": 822, "train_loss": 4.243847846984863, "perplexity": 69.67543711562527, "lr": 0.0014407909032947825, "grad_norm": 0.465324, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:58.978636+00:00", "epoch": 0, "step": 823, "train_loss": 4.17549991607666, "perplexity": 65.0723622510924, "lr": 0.0014425436902817592, "grad_norm": 0.502044, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:59.282789+00:00", "epoch": 0, "step": 824, "train_loss": 4.344948768615723, "perplexity": 77.08808804867357, "lr": 0.0014442964772687355, "grad_norm": 0.53884, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:59.588667+00:00", "epoch": 0, "step": 825, "train_loss": 4.363834857940674, "perplexity": 78.55781558271926, "lr": 0.0014460492642557122, "grad_norm": 0.502482, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:01:59.893760+00:00", "epoch": 0, "step": 826, "train_loss": 4.275986194610596, "perplexity": 71.95106208173472, "lr": 0.001447802051242689, "grad_norm": 0.452737, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:00.200225+00:00", "epoch": 0, "step": 827, "train_loss": 4.253509044647217, "perplexity": 70.35184749171616, "lr": 0.0014495548382296656, "grad_norm": 0.43835, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:00.507441+00:00", "epoch": 0, "step": 828, "train_loss": 4.243350982666016, "perplexity": 69.64082647613034, "lr": 0.0014513076252166421, "grad_norm": 0.364761, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:00.812151+00:00", "epoch": 0, "step": 829, "train_loss": 4.269178867340088, "perplexity": 71.4629309737713, "lr": 0.0014530604122036186, "grad_norm": 0.336508, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:01.118454+00:00", "epoch": 0, "step": 830, "train_loss": 4.276598930358887, "perplexity": 71.99516257919558, "lr": 0.0014548131991905953, "grad_norm": 0.314319, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:01.423513+00:00", "epoch": 0, "step": 831, "train_loss": 4.133368015289307, "perplexity": 62.38769218458917, "lr": 0.0014565659861775718, "grad_norm": 0.329995, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:01.728744+00:00", "epoch": 0, "step": 832, "train_loss": 4.176961421966553, "perplexity": 65.16753542292638, "lr": 0.0014583187731645486, "grad_norm": 0.324015, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:02.036199+00:00", "epoch": 0, "step": 833, "train_loss": 4.250751495361328, "perplexity": 70.15811603968045, "lr": 0.0014600715601515253, "grad_norm": 0.334339, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:02.342993+00:00", "epoch": 0, "step": 834, "train_loss": 4.177248001098633, "perplexity": 65.18621375494993, "lr": 0.001461824347138502, "grad_norm": 0.396608, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:02.647499+00:00", "epoch": 0, "step": 835, "train_loss": 4.288524150848389, "perplexity": 72.85886042517627, "lr": 0.0014635771341254783, "grad_norm": 0.396747, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:02.953053+00:00", "epoch": 0, "step": 836, "train_loss": 4.1940016746521, "perplexity": 66.28752201997213, "lr": 0.001465329921112455, "grad_norm": 0.409597, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:03.258575+00:00", "epoch": 0, "step": 837, "train_loss": 4.1881256103515625, "perplexity": 65.89915443218517, "lr": 0.0014670827080994317, "grad_norm": 0.427693, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:03.566011+00:00", "epoch": 0, "step": 838, "train_loss": 4.052271842956543, "perplexity": 57.5280032880906, "lr": 0.0014688354950864084, "grad_norm": 0.394202, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:03.871828+00:00", "epoch": 0, "step": 839, "train_loss": 4.172245025634766, "perplexity": 64.86090316593598, "lr": 0.001470588282073385, "grad_norm": 0.396506, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:04.177814+00:00", "epoch": 0, "step": 840, "train_loss": 4.250799179077148, "perplexity": 70.16146151910998, "lr": 0.0014723410690603617, "grad_norm": 0.454628, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:04.481345+00:00", "epoch": 0, "step": 841, "train_loss": 4.161276817321777, "perplexity": 64.15338247875941, "lr": 0.0014740938560473382, "grad_norm": 0.429419, "tokens_per_sec": 107896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:04.785908+00:00", "epoch": 0, "step": 842, "train_loss": 4.255817890167236, "perplexity": 70.51446669869583, "lr": 0.0014758466430343147, "grad_norm": 0.391953, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:05.092673+00:00", "epoch": 0, "step": 843, "train_loss": 4.188600540161133, "perplexity": 65.93045933827861, "lr": 0.0014775994300212914, "grad_norm": 0.394186, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:05.399033+00:00", "epoch": 0, "step": 844, "train_loss": 4.287156105041504, "perplexity": 72.75925431507265, "lr": 0.001479352217008268, "grad_norm": 0.382494, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:05.704361+00:00", "epoch": 0, "step": 845, "train_loss": 4.333242893218994, "perplexity": 76.19096554172725, "lr": 0.0014811050039952448, "grad_norm": 0.382356, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:06.008851+00:00", "epoch": 0, "step": 846, "train_loss": 4.121227741241455, "perplexity": 61.634867499559846, "lr": 0.001482857790982221, "grad_norm": 0.330972, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:06.314152+00:00", "epoch": 0, "step": 847, "train_loss": 4.270071029663086, "perplexity": 71.52671595732583, "lr": 0.0014846105779691978, "grad_norm": 0.334183, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:06.619755+00:00", "epoch": 0, "step": 848, "train_loss": 4.15814733505249, "perplexity": 63.95292942651113, "lr": 0.0014863633649561745, "grad_norm": 0.325229, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:06.926212+00:00", "epoch": 0, "step": 849, "train_loss": 4.170254230499268, "perplexity": 64.73190684066154, "lr": 0.0014881161519431513, "grad_norm": 0.365042, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:07.231635+00:00", "epoch": 0, "step": 850, "train_loss": 4.2617082595825195, "perplexity": 70.93104866301131, "lr": 0.0014898689389301278, "grad_norm": 0.34641, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:07.538004+00:00", "epoch": 0, "step": 851, "train_loss": 4.177072048187256, "perplexity": 65.17474505986284, "lr": 0.0014916217259171045, "grad_norm": 0.354065, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:07.842894+00:00", "epoch": 0, "step": 852, "train_loss": 4.218906402587891, "perplexity": 67.95912371474098, "lr": 0.001493374512904081, "grad_norm": 0.418069, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:08.148005+00:00", "epoch": 0, "step": 853, "train_loss": 4.133655548095703, "perplexity": 62.40563327200961, "lr": 0.0014951272998910575, "grad_norm": 0.402298, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:08.453059+00:00", "epoch": 0, "step": 854, "train_loss": 4.1344475746154785, "perplexity": 62.45507976742742, "lr": 0.0014968800868780342, "grad_norm": 0.40547, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:08.758673+00:00", "epoch": 0, "step": 855, "train_loss": 4.108450412750244, "perplexity": 60.8523484486386, "lr": 0.001498632873865011, "grad_norm": 0.458032, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:09.064612+00:00", "epoch": 0, "step": 856, "train_loss": 4.2017645835876465, "perplexity": 66.80410852975135, "lr": 0.0015003856608519876, "grad_norm": 0.432623, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:09.369779+00:00", "epoch": 0, "step": 857, "train_loss": 4.262917518615723, "perplexity": 71.01687455675162, "lr": 0.0015021384478389643, "grad_norm": 0.392896, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:09.675310+00:00", "epoch": 0, "step": 858, "train_loss": 4.159106731414795, "perplexity": 64.0143150762377, "lr": 0.0015038912348259406, "grad_norm": 0.401105, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:09.980975+00:00", "epoch": 0, "step": 859, "train_loss": 4.183973789215088, "perplexity": 65.62612011696329, "lr": 0.0015056440218129174, "grad_norm": 0.398059, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:10.287317+00:00", "epoch": 0, "step": 860, "train_loss": 4.334047794342041, "perplexity": 76.25231642284211, "lr": 0.001507396808799894, "grad_norm": 0.378537, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:10.592765+00:00", "epoch": 0, "step": 861, "train_loss": 4.181255340576172, "perplexity": 65.44796114777331, "lr": 0.0015091495957868706, "grad_norm": 0.398544, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:10.898627+00:00", "epoch": 0, "step": 862, "train_loss": 4.300116539001465, "perplexity": 73.7083831004521, "lr": 0.0015109023827738473, "grad_norm": 0.414981, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:11.203663+00:00", "epoch": 0, "step": 863, "train_loss": 4.21195125579834, "perplexity": 67.48809795903644, "lr": 0.0015126551697608238, "grad_norm": 0.375768, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:11.508430+00:00", "epoch": 0, "step": 864, "train_loss": 4.262330055236816, "perplexity": 70.97516699569701, "lr": 0.0015144079567478003, "grad_norm": 0.379142, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:11.813491+00:00", "epoch": 0, "step": 865, "train_loss": 4.235958576202393, "perplexity": 69.12791135468505, "lr": 0.001516160743734777, "grad_norm": 0.45713, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:12.119782+00:00", "epoch": 0, "step": 866, "train_loss": 4.302545070648193, "perplexity": 73.88760377489709, "lr": 0.0015179135307217537, "grad_norm": 0.46524, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:12.425281+00:00", "epoch": 0, "step": 867, "train_loss": 4.24959135055542, "perplexity": 70.07676966167759, "lr": 0.0015196663177087304, "grad_norm": 0.408586, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:12.730145+00:00", "epoch": 0, "step": 868, "train_loss": 4.128488540649414, "perplexity": 62.08401451899549, "lr": 0.0015214191046957072, "grad_norm": 0.383887, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:13.035008+00:00", "epoch": 0, "step": 869, "train_loss": 4.153541564941406, "perplexity": 63.659054215973406, "lr": 0.0015231718916826834, "grad_norm": 0.373151, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:13.342071+00:00", "epoch": 0, "step": 870, "train_loss": 4.2858123779296875, "perplexity": 72.66155139014957, "lr": 0.0015249246786696602, "grad_norm": 0.356584, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:13.647769+00:00", "epoch": 0, "step": 871, "train_loss": 4.149952411651611, "perplexity": 63.43098165064424, "lr": 0.0015266774656566369, "grad_norm": 0.387199, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:13.952954+00:00", "epoch": 0, "step": 872, "train_loss": 4.119756698608398, "perplexity": 61.544266636989015, "lr": 0.0015284302526436134, "grad_norm": 0.404173, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:14.257848+00:00", "epoch": 0, "step": 873, "train_loss": 4.2325286865234375, "perplexity": 68.89121639584283, "lr": 0.00153018303963059, "grad_norm": 0.411045, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:14.563214+00:00", "epoch": 0, "step": 874, "train_loss": 4.115071773529053, "perplexity": 61.25661070810421, "lr": 0.0015319358266175668, "grad_norm": 0.400837, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:14.869087+00:00", "epoch": 0, "step": 875, "train_loss": 4.241354942321777, "perplexity": 69.50195921531142, "lr": 0.0015336886136045431, "grad_norm": 0.322457, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:15.174363+00:00", "epoch": 0, "step": 876, "train_loss": 4.290139198303223, "perplexity": 72.97662601517817, "lr": 0.0015354414005915198, "grad_norm": 0.352046, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:15.479955+00:00", "epoch": 0, "step": 877, "train_loss": 4.084899425506592, "perplexity": 59.43595970909166, "lr": 0.0015371941875784965, "grad_norm": 0.36423, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:15.785602+00:00", "epoch": 0, "step": 878, "train_loss": 4.160976886749268, "perplexity": 64.13414380330181, "lr": 0.0015389469745654733, "grad_norm": 0.352657, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:16.089768+00:00", "epoch": 0, "step": 879, "train_loss": 4.210024833679199, "perplexity": 67.35821254158814, "lr": 0.00154069976155245, "grad_norm": 0.359549, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:16.394763+00:00", "epoch": 0, "step": 880, "train_loss": 4.135189533233643, "perplexity": 62.50143604720302, "lr": 0.0015424525485394263, "grad_norm": 0.357406, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:16.700411+00:00", "epoch": 0, "step": 881, "train_loss": 4.117919445037842, "perplexity": 61.43129802113647, "lr": 0.001544205335526403, "grad_norm": 0.354216, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:17.005172+00:00", "epoch": 0, "step": 882, "train_loss": 4.1337738037109375, "perplexity": 62.413013524935856, "lr": 0.0015459581225133797, "grad_norm": 0.378212, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:17.309659+00:00", "epoch": 0, "step": 883, "train_loss": 4.100086688995361, "perplexity": 60.345518663008235, "lr": 0.0015477109095003562, "grad_norm": 0.378557, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:17.615174+00:00", "epoch": 0, "step": 884, "train_loss": 4.136676788330078, "perplexity": 62.59446078510954, "lr": 0.001549463696487333, "grad_norm": 0.408532, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:17.920142+00:00", "epoch": 0, "step": 885, "train_loss": 4.281041145324707, "perplexity": 72.31569197101119, "lr": 0.0015512164834743096, "grad_norm": 0.424438, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:18.224680+00:00", "epoch": 0, "step": 886, "train_loss": 4.213798999786377, "perplexity": 67.61291396480465, "lr": 0.001552969270461286, "grad_norm": 0.463952, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:18.531212+00:00", "epoch": 0, "step": 887, "train_loss": 4.275866985321045, "perplexity": 71.94248535796426, "lr": 0.0015547220574482626, "grad_norm": 0.408095, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:18.836239+00:00", "epoch": 0, "step": 888, "train_loss": 4.2790327072143555, "perplexity": 72.1705961360567, "lr": 0.0015564748444352394, "grad_norm": 0.420688, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:19.142167+00:00", "epoch": 0, "step": 889, "train_loss": 4.225980281829834, "perplexity": 68.44156269507207, "lr": 0.001558227631422216, "grad_norm": 0.406101, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:19.446856+00:00", "epoch": 0, "step": 890, "train_loss": 4.234100818634033, "perplexity": 68.99960766965279, "lr": 0.0015599804184091928, "grad_norm": 0.357076, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:19.752225+00:00", "epoch": 0, "step": 891, "train_loss": 4.201099395751953, "perplexity": 66.75968602567568, "lr": 0.0015617332053961693, "grad_norm": 0.336076, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:20.057775+00:00", "epoch": 0, "step": 892, "train_loss": 4.175055503845215, "perplexity": 65.04344972239554, "lr": 0.0015634859923831458, "grad_norm": 0.30645, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:20.363199+00:00", "epoch": 0, "step": 893, "train_loss": 4.234547138214111, "perplexity": 69.0304104189972, "lr": 0.0015652387793701225, "grad_norm": 0.320467, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:20.669625+00:00", "epoch": 0, "step": 894, "train_loss": 4.239499092102051, "perplexity": 69.37309360365347, "lr": 0.001566991566357099, "grad_norm": 0.319334, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:20.974472+00:00", "epoch": 0, "step": 895, "train_loss": 4.1131439208984375, "perplexity": 61.13863075053431, "lr": 0.0015687443533440757, "grad_norm": 0.309295, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:21.279374+00:00", "epoch": 0, "step": 896, "train_loss": 4.1639862060546875, "perplexity": 64.32743461139569, "lr": 0.0015704971403310525, "grad_norm": 0.281443, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:21.585152+00:00", "epoch": 0, "step": 897, "train_loss": 4.172997951507568, "perplexity": 64.90975700741565, "lr": 0.0015722499273180287, "grad_norm": 0.300652, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:21.892519+00:00", "epoch": 0, "step": 898, "train_loss": 4.2180914878845215, "perplexity": 67.90376538481982, "lr": 0.0015740027143050055, "grad_norm": 0.305714, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:22.197603+00:00", "epoch": 0, "step": 899, "train_loss": 4.202730655670166, "perplexity": 66.86867729801995, "lr": 0.0015757555012919822, "grad_norm": 0.293647, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:22.503501+00:00", "epoch": 0, "step": 900, "train_loss": 4.275238513946533, "perplexity": 71.89728577011684, "lr": 0.001577508288278959, "grad_norm": 0.297313, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:22.808473+00:00", "epoch": 0, "step": 901, "train_loss": 4.168529987335205, "perplexity": 64.6203894619623, "lr": 0.0015792610752659356, "grad_norm": 0.304615, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:23.114855+00:00", "epoch": 0, "step": 902, "train_loss": 4.179673194885254, "perplexity": 65.34449480905792, "lr": 0.0015810138622529121, "grad_norm": 0.331644, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:23.420667+00:00", "epoch": 0, "step": 903, "train_loss": 4.158425331115723, "perplexity": 63.970710560552, "lr": 0.0015827666492398886, "grad_norm": 0.35638, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:23.727335+00:00", "epoch": 0, "step": 904, "train_loss": 4.166400909423828, "perplexity": 64.48295397547946, "lr": 0.0015845194362268653, "grad_norm": 0.38039, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:24.032821+00:00", "epoch": 0, "step": 905, "train_loss": 4.20656681060791, "perplexity": 67.12568855693894, "lr": 0.0015862722232138418, "grad_norm": 0.335509, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:24.338070+00:00", "epoch": 0, "step": 906, "train_loss": 4.1420793533325195, "perplexity": 62.93354657003006, "lr": 0.0015880250102008186, "grad_norm": 0.40507, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:24.644308+00:00", "epoch": 0, "step": 907, "train_loss": 4.2522759437561035, "perplexity": 70.26515003022674, "lr": 0.0015897777971877953, "grad_norm": 0.471956, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:24.949211+00:00", "epoch": 0, "step": 908, "train_loss": 4.168308734893799, "perplexity": 64.606093624582, "lr": 0.001591530584174772, "grad_norm": 0.564899, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:25.255252+00:00", "epoch": 0, "step": 909, "train_loss": 3.9929304122924805, "perplexity": 54.213524795675006, "lr": 0.0015932833711617483, "grad_norm": 0.528847, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:25.561092+00:00", "epoch": 0, "step": 910, "train_loss": 4.205658912658691, "perplexity": 67.06477293872491, "lr": 0.001595036158148725, "grad_norm": 0.490708, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:25.865864+00:00", "epoch": 0, "step": 911, "train_loss": 4.10788106918335, "perplexity": 60.81771241635563, "lr": 0.0015967889451357017, "grad_norm": 0.396526, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:26.170549+00:00", "epoch": 0, "step": 912, "train_loss": 4.216726303100586, "perplexity": 67.81112744588454, "lr": 0.0015985417321226782, "grad_norm": 0.392586, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:26.475908+00:00", "epoch": 0, "step": 913, "train_loss": 4.153731346130371, "perplexity": 63.6711366534443, "lr": 0.001600294519109655, "grad_norm": 0.341566, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:26.781707+00:00", "epoch": 0, "step": 914, "train_loss": 4.193144798278809, "perplexity": 66.2307461369395, "lr": 0.0016020473060966314, "grad_norm": 0.33803, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:27.086804+00:00", "epoch": 0, "step": 915, "train_loss": 4.159348964691162, "perplexity": 64.02982335174737, "lr": 0.001603800093083608, "grad_norm": 0.343046, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:27.391393+00:00", "epoch": 0, "step": 916, "train_loss": 4.254919052124023, "perplexity": 70.45111408956305, "lr": 0.0016055528800705847, "grad_norm": 0.336575, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:27.695995+00:00", "epoch": 0, "step": 917, "train_loss": 4.063541412353516, "perplexity": 58.17998599528793, "lr": 0.0016073056670575614, "grad_norm": 0.367891, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:28.000455+00:00", "epoch": 0, "step": 918, "train_loss": 4.239777088165283, "perplexity": 69.39238173145712, "lr": 0.001609058454044538, "grad_norm": 0.390269, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:28.305855+00:00", "epoch": 0, "step": 919, "train_loss": 4.184878349304199, "perplexity": 65.68550974271737, "lr": 0.0016108112410315148, "grad_norm": 0.352404, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:28.612459+00:00", "epoch": 0, "step": 920, "train_loss": 4.199382781982422, "perplexity": 66.64518373561589, "lr": 0.001612564028018491, "grad_norm": 0.344164, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:28.917474+00:00", "epoch": 0, "step": 921, "train_loss": 4.225059509277344, "perplexity": 68.3785725869246, "lr": 0.0016143168150054678, "grad_norm": 0.392456, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:29.223001+00:00", "epoch": 0, "step": 922, "train_loss": 4.1230244636535645, "perplexity": 61.74570779216465, "lr": 0.0016160696019924445, "grad_norm": 0.397158, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:29.528265+00:00", "epoch": 0, "step": 923, "train_loss": 4.139274597167969, "perplexity": 62.757280624381515, "lr": 0.001617822388979421, "grad_norm": 0.384084, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:29.834016+00:00", "epoch": 0, "step": 924, "train_loss": 4.133732795715332, "perplexity": 62.41045414482938, "lr": 0.0016195751759663978, "grad_norm": 0.36329, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:30.139632+00:00", "epoch": 0, "step": 925, "train_loss": 4.27508020401001, "perplexity": 71.885904616269, "lr": 0.0016213279629533743, "grad_norm": 0.337791, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:30.445622+00:00", "epoch": 0, "step": 926, "train_loss": 4.24271297454834, "perplexity": 69.59640923430403, "lr": 0.0016230807499403508, "grad_norm": 0.320456, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:30.750781+00:00", "epoch": 0, "step": 927, "train_loss": 4.051494121551514, "perplexity": 57.483279921958676, "lr": 0.0016248335369273275, "grad_norm": 0.333372, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:31.055352+00:00", "epoch": 0, "step": 928, "train_loss": 4.163773536682129, "perplexity": 64.3137555908445, "lr": 0.0016265863239143042, "grad_norm": 0.30834, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:31.360431+00:00", "epoch": 0, "step": 929, "train_loss": 4.027438640594482, "perplexity": 56.116991227073534, "lr": 0.001628339110901281, "grad_norm": 0.292086, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:31.666532+00:00", "epoch": 0, "step": 930, "train_loss": 4.158881664276123, "perplexity": 63.99990917871732, "lr": 0.0016300918978882576, "grad_norm": 0.291246, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:31.971398+00:00", "epoch": 0, "step": 931, "train_loss": 4.086012840270996, "perplexity": 59.50217343900016, "lr": 0.001631844684875234, "grad_norm": 0.287202, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:32.275560+00:00", "epoch": 0, "step": 932, "train_loss": 4.199487209320068, "perplexity": 66.65214367811816, "lr": 0.0016335974718622106, "grad_norm": 0.293579, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:32.580789+00:00", "epoch": 0, "step": 933, "train_loss": 4.028711318969727, "perplexity": 56.188455574210636, "lr": 0.0016353502588491874, "grad_norm": 0.322657, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:32.886052+00:00", "epoch": 0, "step": 934, "train_loss": 4.125690460205078, "perplexity": 61.9105412613095, "lr": 0.0016371030458361639, "grad_norm": 0.315191, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:33.191986+00:00", "epoch": 0, "step": 935, "train_loss": 4.212804794311523, "perplexity": 67.54572624030608, "lr": 0.0016388558328231406, "grad_norm": 0.319147, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:33.497568+00:00", "epoch": 0, "step": 936, "train_loss": 4.285381317138672, "perplexity": 72.63023659410618, "lr": 0.0016406086198101173, "grad_norm": 0.357927, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:33.803466+00:00", "epoch": 0, "step": 937, "train_loss": 4.209907531738281, "perplexity": 67.35031175591806, "lr": 0.0016423614067970936, "grad_norm": 0.34508, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:34.108512+00:00", "epoch": 0, "step": 938, "train_loss": 4.179652690887451, "perplexity": 65.3431549994157, "lr": 0.0016441141937840703, "grad_norm": 0.343477, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:34.413859+00:00", "epoch": 0, "step": 939, "train_loss": 4.064262866973877, "perplexity": 58.22197535988606, "lr": 0.001645866980771047, "grad_norm": 0.357692, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:34.720283+00:00", "epoch": 0, "step": 940, "train_loss": 4.084872245788574, "perplexity": 59.43434427842024, "lr": 0.0016476197677580237, "grad_norm": 0.324547, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:35.025487+00:00", "epoch": 0, "step": 941, "train_loss": 4.172986030578613, "perplexity": 64.90898322742598, "lr": 0.0016493725547450004, "grad_norm": 0.378609, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:35.330761+00:00", "epoch": 0, "step": 942, "train_loss": 4.2109198570251465, "perplexity": 67.4185267016185, "lr": 0.0016511253417319767, "grad_norm": 0.376818, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:35.635428+00:00", "epoch": 0, "step": 943, "train_loss": 4.166539192199707, "perplexity": 64.49187147390607, "lr": 0.0016528781287189535, "grad_norm": 0.391421, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:35.939633+00:00", "epoch": 0, "step": 944, "train_loss": 4.169996738433838, "perplexity": 64.7152410340185, "lr": 0.0016546309157059302, "grad_norm": 0.414826, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:36.245092+00:00", "epoch": 0, "step": 945, "train_loss": 4.214256286621094, "perplexity": 67.64383953060573, "lr": 0.0016563837026929067, "grad_norm": 0.407603, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:36.550712+00:00", "epoch": 0, "step": 946, "train_loss": 4.155457973480225, "perplexity": 63.78116794356801, "lr": 0.0016581364896798834, "grad_norm": 0.43298, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:36.856201+00:00", "epoch": 0, "step": 947, "train_loss": 4.172350883483887, "perplexity": 64.86776956506188, "lr": 0.00165988927666686, "grad_norm": 0.351779, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:37.161786+00:00", "epoch": 0, "step": 948, "train_loss": 4.220398426055908, "perplexity": 68.06059600288268, "lr": 0.0016616420636538364, "grad_norm": 0.300588, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:37.466554+00:00", "epoch": 0, "step": 949, "train_loss": 4.037783145904541, "perplexity": 56.700506626449716, "lr": 0.0016633948506408131, "grad_norm": 0.322182, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:37.771504+00:00", "epoch": 0, "step": 950, "train_loss": 4.0450897216796875, "perplexity": 57.116310375563664, "lr": 0.0016651476376277898, "grad_norm": 0.32633, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:38.077355+00:00", "epoch": 0, "step": 951, "train_loss": 4.151505470275879, "perplexity": 63.5295702208409, "lr": 0.0016669004246147665, "grad_norm": 0.358043, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:38.383996+00:00", "epoch": 0, "step": 952, "train_loss": 4.28513765335083, "perplexity": 72.61254139147341, "lr": 0.0016686532116017433, "grad_norm": 0.315986, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:38.691181+00:00", "epoch": 0, "step": 953, "train_loss": 4.0854644775390625, "perplexity": 59.46955360918406, "lr": 0.0016704059985887198, "grad_norm": 0.294024, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:38.997446+00:00", "epoch": 0, "step": 954, "train_loss": 3.997541904449463, "perplexity": 54.46410737580309, "lr": 0.0016721587855756963, "grad_norm": 0.344537, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:39.302415+00:00", "epoch": 0, "step": 955, "train_loss": 4.057794570922852, "perplexity": 57.8465937351172, "lr": 0.001673911572562673, "grad_norm": 0.387722, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:39.607595+00:00", "epoch": 0, "step": 956, "train_loss": 4.109605312347412, "perplexity": 60.922667399196584, "lr": 0.0016756643595496495, "grad_norm": 0.360615, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:39.913337+00:00", "epoch": 0, "step": 957, "train_loss": 4.211330413818359, "perplexity": 67.4462115184502, "lr": 0.0016774171465366262, "grad_norm": 0.388732, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:40.219724+00:00", "epoch": 0, "step": 958, "train_loss": 4.145536422729492, "perplexity": 63.15148871141577, "lr": 0.001679169933523603, "grad_norm": 0.404338, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:40.524724+00:00", "epoch": 0, "step": 959, "train_loss": 4.133652210235596, "perplexity": 62.40542497108347, "lr": 0.0016809227205105792, "grad_norm": 0.407601, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:40.829633+00:00", "epoch": 0, "step": 960, "train_loss": 4.136239528656006, "perplexity": 62.567096734621806, "lr": 0.001682675507497556, "grad_norm": 0.389438, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:41.135276+00:00", "epoch": 0, "step": 961, "train_loss": 4.116021633148193, "perplexity": 61.314823531649665, "lr": 0.0016844282944845326, "grad_norm": 0.367984, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:41.441169+00:00", "epoch": 0, "step": 962, "train_loss": 4.096438884735107, "perplexity": 60.12579102829307, "lr": 0.0016861810814715094, "grad_norm": 0.336209, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:41.747796+00:00", "epoch": 0, "step": 963, "train_loss": 4.10023832321167, "perplexity": 60.354669802233815, "lr": 0.001687933868458486, "grad_norm": 0.327653, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:42.052857+00:00", "epoch": 0, "step": 964, "train_loss": 4.041551113128662, "perplexity": 56.91455528793202, "lr": 0.0016896866554454626, "grad_norm": 0.313332, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:42.357578+00:00", "epoch": 0, "step": 965, "train_loss": 4.047520637512207, "perplexity": 57.255324215766464, "lr": 0.001691439442432439, "grad_norm": 0.319066, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:42.662853+00:00", "epoch": 0, "step": 966, "train_loss": 4.031527519226074, "perplexity": 56.34691654138142, "lr": 0.0016931922294194158, "grad_norm": 0.31817, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:42.968379+00:00", "epoch": 0, "step": 967, "train_loss": 4.161048412322998, "perplexity": 64.13873119878915, "lr": 0.0016949450164063923, "grad_norm": 0.324946, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:43.274853+00:00", "epoch": 0, "step": 968, "train_loss": 4.04713249206543, "perplexity": 57.23310513476487, "lr": 0.001696697803393369, "grad_norm": 0.333023, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:43.580756+00:00", "epoch": 0, "step": 969, "train_loss": 4.125176906585693, "perplexity": 61.878755041429415, "lr": 0.0016984505903803457, "grad_norm": 0.341745, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:43.886191+00:00", "epoch": 0, "step": 970, "train_loss": 4.21212911605835, "perplexity": 67.5001024772187, "lr": 0.0017002033773673225, "grad_norm": 0.342445, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:44.190955+00:00", "epoch": 0, "step": 971, "train_loss": 4.037761688232422, "perplexity": 56.6992899786228, "lr": 0.0017019561643542987, "grad_norm": 0.326533, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:44.496089+00:00", "epoch": 0, "step": 972, "train_loss": 4.1769700050354, "perplexity": 65.16809476276997, "lr": 0.0017037089513412755, "grad_norm": 0.363121, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:44.802780+00:00", "epoch": 0, "step": 973, "train_loss": 4.056094169616699, "perplexity": 57.74831489199015, "lr": 0.0017054617383282522, "grad_norm": 0.361878, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:45.108096+00:00", "epoch": 0, "step": 974, "train_loss": 4.10762882232666, "perplexity": 60.80237327427409, "lr": 0.001707214525315229, "grad_norm": 0.312696, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:45.412988+00:00", "epoch": 0, "step": 975, "train_loss": 4.054993152618408, "perplexity": 57.68476800519218, "lr": 0.0017089673123022054, "grad_norm": 0.328582, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:45.717858+00:00", "epoch": 0, "step": 976, "train_loss": 4.041202545166016, "perplexity": 56.89472015449723, "lr": 0.001710720099289182, "grad_norm": 0.338924, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:46.023240+00:00", "epoch": 0, "step": 977, "train_loss": 4.1388044357299805, "perplexity": 62.72778150629622, "lr": 0.0017124728862761586, "grad_norm": 0.322024, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:46.328164+00:00", "epoch": 0, "step": 978, "train_loss": 4.0744309425354, "perplexity": 58.8170008099357, "lr": 0.0017142256732631351, "grad_norm": 0.273267, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:46.634148+00:00", "epoch": 0, "step": 979, "train_loss": 4.074557781219482, "perplexity": 58.824461554065564, "lr": 0.0017159784602501118, "grad_norm": 0.292124, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:46.939627+00:00", "epoch": 0, "step": 980, "train_loss": 4.090222358703613, "perplexity": 59.75317686599948, "lr": 0.0017177312472370886, "grad_norm": 0.304077, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:47.244926+00:00", "epoch": 0, "step": 981, "train_loss": 4.106418609619141, "perplexity": 60.72883397756147, "lr": 0.0017194840342240653, "grad_norm": 0.305886, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:47.551064+00:00", "epoch": 0, "step": 982, "train_loss": 4.009592056274414, "perplexity": 55.124378331726085, "lr": 0.0017212368212110416, "grad_norm": 0.311435, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:47.856805+00:00", "epoch": 0, "step": 983, "train_loss": 4.138144493103027, "perplexity": 62.68639842611728, "lr": 0.0017229896081980183, "grad_norm": 0.32526, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:48.161931+00:00", "epoch": 0, "step": 984, "train_loss": 4.1504807472229, "perplexity": 63.46450334914516, "lr": 0.001724742395184995, "grad_norm": 0.381642, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:48.467168+00:00", "epoch": 0, "step": 985, "train_loss": 4.0939788818359375, "perplexity": 59.9780631880056, "lr": 0.0017264951821719717, "grad_norm": 0.382479, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:48.772382+00:00", "epoch": 0, "step": 986, "train_loss": 4.116844654083252, "perplexity": 61.36530768695431, "lr": 0.0017282479691589482, "grad_norm": 0.38993, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:49.079293+00:00", "epoch": 0, "step": 987, "train_loss": 4.088383674621582, "perplexity": 59.64341059451767, "lr": 0.001730000756145925, "grad_norm": 0.434482, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:49.385328+00:00", "epoch": 0, "step": 988, "train_loss": 4.166600704193115, "perplexity": 64.49583861949132, "lr": 0.0017317535431329014, "grad_norm": 0.4413, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:49.690937+00:00", "epoch": 0, "step": 989, "train_loss": 4.17551326751709, "perplexity": 65.07323106666057, "lr": 0.001733506330119878, "grad_norm": 0.444571, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:49.996522+00:00", "epoch": 0, "step": 990, "train_loss": 4.126020908355713, "perplexity": 61.93100286573757, "lr": 0.0017352591171068547, "grad_norm": 0.381142, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:50.302123+00:00", "epoch": 0, "step": 991, "train_loss": 4.096060752868652, "perplexity": 60.103059848672, "lr": 0.0017370119040938314, "grad_norm": 0.360177, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:50.608536+00:00", "epoch": 0, "step": 992, "train_loss": 4.113536834716797, "perplexity": 61.16265768333962, "lr": 0.001738764691080808, "grad_norm": 0.345482, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:50.913614+00:00", "epoch": 0, "step": 993, "train_loss": 4.101856231689453, "perplexity": 60.45239716983111, "lr": 0.0017405174780677844, "grad_norm": 0.355843, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:51.218713+00:00", "epoch": 0, "step": 994, "train_loss": 4.032171726226807, "perplexity": 56.38322731405829, "lr": 0.001742270265054761, "grad_norm": 0.357619, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:51.524755+00:00", "epoch": 0, "step": 995, "train_loss": 4.06937313079834, "perplexity": 58.52026653851527, "lr": 0.0017440230520417378, "grad_norm": 0.293471, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:51.831858+00:00", "epoch": 0, "step": 996, "train_loss": 4.214202880859375, "perplexity": 67.64022705629435, "lr": 0.0017457758390287145, "grad_norm": 0.269588, "tokens_per_sec": 106700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:52.137568+00:00", "epoch": 0, "step": 997, "train_loss": 4.076334476470947, "perplexity": 58.92906759458937, "lr": 0.001747528626015691, "grad_norm": 0.261322, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:52.442578+00:00", "epoch": 0, "step": 998, "train_loss": 4.106924057006836, "perplexity": 60.759536966772394, "lr": 0.0017492814130026678, "grad_norm": 0.244259, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:52.747475+00:00", "epoch": 0, "step": 999, "train_loss": 4.055695533752441, "perplexity": 57.72529893037242, "lr": 0.0017510341999896443, "grad_norm": 0.230407, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:53.053099+00:00", "epoch": 0, "step": 1000, "train_loss": 4.082658290863037, "perplexity": 59.302904873306716, "lr": 0.0017527869869766208, "grad_norm": 0.229779, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:02:56.130808+00:00", "step": 1000, "epoch": 0, "val_loss": 4.026695609092712, "val_ppl": 56.07531002194915, "eval_train_loss": 4.082658290863037, "eval_train_ppl": 59.302904873306716} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:02:57.085320+00:00", "step": 1000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_4p0267_epoch_0000_step_0001000.pt", "val_loss": 4.026695609092712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:58.083291+00:00", "epoch": 0, "step": 1001, "train_loss": 4.0540876388549805, "perplexity": 57.6325572961487, "lr": 0.0017545397739635975, "grad_norm": 0.228988, "tokens_per_sec": 6514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:58.387954+00:00", "epoch": 0, "step": 1002, "train_loss": 3.9494359493255615, "perplexity": 51.90608091628854, "lr": 0.0017562925609505742, "grad_norm": 0.23147, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:58.693909+00:00", "epoch": 0, "step": 1003, "train_loss": 4.011699676513672, "perplexity": 55.24068210621665, "lr": 0.001758045347937551, "grad_norm": 0.25587, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:58.997781+00:00", "epoch": 0, "step": 1004, "train_loss": 4.076058387756348, "perplexity": 58.9128001897914, "lr": 0.0017597981349245276, "grad_norm": 0.304096, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:59.301399+00:00", "epoch": 0, "step": 1005, "train_loss": 4.08146333694458, "perplexity": 59.2320829576426, "lr": 0.001761550921911504, "grad_norm": 0.331838, "tokens_per_sec": 107925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:59.605533+00:00", "epoch": 0, "step": 1006, "train_loss": 4.046281814575195, "perplexity": 57.184438923090696, "lr": 0.0017633037088984806, "grad_norm": 0.348593, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:02:59.909780+00:00", "epoch": 0, "step": 1007, "train_loss": 4.1118597984313965, "perplexity": 61.060171647513286, "lr": 0.0017650564958854574, "grad_norm": 0.345254, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:00.213634+00:00", "epoch": 0, "step": 1008, "train_loss": 4.156737327575684, "perplexity": 63.86281886107079, "lr": 0.0017668092828724339, "grad_norm": 0.421911, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:00.518184+00:00", "epoch": 0, "step": 1009, "train_loss": 4.110053062438965, "perplexity": 60.94995163690635, "lr": 0.0017685620698594106, "grad_norm": 0.393086, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:00.821834+00:00", "epoch": 0, "step": 1010, "train_loss": 4.044035911560059, "perplexity": 57.05615233283833, "lr": 0.001770314856846387, "grad_norm": 0.371313, "tokens_per_sec": 107973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:01.125894+00:00", "epoch": 0, "step": 1011, "train_loss": 4.171087265014648, "perplexity": 64.78585321978446, "lr": 0.0017720676438333636, "grad_norm": 0.315292, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:01.430650+00:00", "epoch": 0, "step": 1012, "train_loss": 4.201807975769043, "perplexity": 66.80700736863972, "lr": 0.0017738204308203403, "grad_norm": 0.291228, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:01.735786+00:00", "epoch": 0, "step": 1013, "train_loss": 4.132279872894287, "perplexity": 62.31984241359666, "lr": 0.001775573217807317, "grad_norm": 0.298095, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:02.041379+00:00", "epoch": 0, "step": 1014, "train_loss": 4.04469108581543, "perplexity": 57.09354630340312, "lr": 0.0017773260047942937, "grad_norm": 0.266406, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:02.440736+00:00", "epoch": 0, "step": 1015, "train_loss": 4.052748680114746, "perplexity": 57.555441318912024, "lr": 0.0017790787917812702, "grad_norm": 0.304477, "tokens_per_sec": 82052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:02.746025+00:00", "epoch": 0, "step": 1016, "train_loss": 4.077589988708496, "perplexity": 59.00310022481346, "lr": 0.0017808315787682467, "grad_norm": 0.337491, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:03.050942+00:00", "epoch": 0, "step": 1017, "train_loss": 4.083619117736816, "perplexity": 59.35991208064487, "lr": 0.0017825843657552235, "grad_norm": 0.379825, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:03.355533+00:00", "epoch": 0, "step": 1018, "train_loss": 3.9813144207000732, "perplexity": 53.58742437608607, "lr": 0.0017843371527422, "grad_norm": 0.34473, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:03.660180+00:00", "epoch": 0, "step": 1019, "train_loss": 4.0108537673950195, "perplexity": 55.19397326800377, "lr": 0.0017860899397291767, "grad_norm": 0.318043, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:03.965617+00:00", "epoch": 0, "step": 1020, "train_loss": 4.130575180053711, "perplexity": 62.21369672301252, "lr": 0.0017878427267161534, "grad_norm": 0.343638, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:04.270629+00:00", "epoch": 0, "step": 1021, "train_loss": 4.146631717681885, "perplexity": 63.2206961125772, "lr": 0.0017895955137031301, "grad_norm": 0.362915, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:04.576342+00:00", "epoch": 0, "step": 1022, "train_loss": 4.017765045166016, "perplexity": 55.5767553816451, "lr": 0.0017913483006901064, "grad_norm": 0.332328, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:04.880869+00:00", "epoch": 0, "step": 1023, "train_loss": 4.129775524139404, "perplexity": 62.163967058437244, "lr": 0.0017931010876770831, "grad_norm": 0.363293, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:05.184755+00:00", "epoch": 0, "step": 1024, "train_loss": 4.051927089691162, "perplexity": 57.50817373945341, "lr": 0.0017948538746640598, "grad_norm": 0.383893, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:05.490164+00:00", "epoch": 0, "step": 1025, "train_loss": 4.158767223358154, "perplexity": 63.99258538943957, "lr": 0.0017966066616510365, "grad_norm": 0.337243, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:05.795304+00:00", "epoch": 0, "step": 1026, "train_loss": 4.186832427978516, "perplexity": 65.81398988578842, "lr": 0.001798359448638013, "grad_norm": 0.315001, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:06.101766+00:00", "epoch": 0, "step": 1027, "train_loss": 4.106299877166748, "perplexity": 60.72162392221462, "lr": 0.0018001122356249896, "grad_norm": 0.29494, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:06.406588+00:00", "epoch": 0, "step": 1028, "train_loss": 4.044495105743408, "perplexity": 57.08235820244485, "lr": 0.0018018650226119663, "grad_norm": 0.281379, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:06.712022+00:00", "epoch": 0, "step": 1029, "train_loss": 4.080544948577881, "perplexity": 59.17770987334466, "lr": 0.0018036178095989428, "grad_norm": 0.310454, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:07.017312+00:00", "epoch": 0, "step": 1030, "train_loss": 4.154285907745361, "perplexity": 63.70645601429082, "lr": 0.0018053705965859195, "grad_norm": 0.336458, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:07.322033+00:00", "epoch": 0, "step": 1031, "train_loss": 4.110177993774414, "perplexity": 60.957566671428225, "lr": 0.0018071233835728962, "grad_norm": 0.335859, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:07.627435+00:00", "epoch": 0, "step": 1032, "train_loss": 4.056264877319336, "perplexity": 57.75817381562967, "lr": 0.001808876170559873, "grad_norm": 0.290128, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:07.932298+00:00", "epoch": 0, "step": 1033, "train_loss": 4.056181907653809, "perplexity": 57.75338183806385, "lr": 0.0018106289575468492, "grad_norm": 0.30134, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:08.236957+00:00", "epoch": 0, "step": 1034, "train_loss": 3.993990182876587, "perplexity": 54.27100914924755, "lr": 0.001812381744533826, "grad_norm": 0.317243, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:08.541693+00:00", "epoch": 0, "step": 1035, "train_loss": 4.087003707885742, "perplexity": 59.56116143548927, "lr": 0.0018141345315208026, "grad_norm": 0.329215, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:08.846332+00:00", "epoch": 0, "step": 1036, "train_loss": 4.091047286987305, "perplexity": 59.80248928844621, "lr": 0.0018158873185077794, "grad_norm": 0.281642, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:09.151959+00:00", "epoch": 0, "step": 1037, "train_loss": 4.204650402069092, "perplexity": 66.9971714991407, "lr": 0.0018176401054947559, "grad_norm": 0.285628, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:09.458426+00:00", "epoch": 0, "step": 1038, "train_loss": 4.041499614715576, "perplexity": 56.91162435412307, "lr": 0.0018193928924817326, "grad_norm": 0.296616, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:09.763016+00:00", "epoch": 0, "step": 1039, "train_loss": 4.054538249969482, "perplexity": 57.65853301905871, "lr": 0.001821145679468709, "grad_norm": 0.265238, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:10.067198+00:00", "epoch": 0, "step": 1040, "train_loss": 3.9688100814819336, "perplexity": 52.9215210609449, "lr": 0.0018228984664556856, "grad_norm": 0.255529, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:10.371944+00:00", "epoch": 0, "step": 1041, "train_loss": 3.999671220779419, "perplexity": 54.5802022465225, "lr": 0.0018246512534426623, "grad_norm": 0.290588, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:10.677098+00:00", "epoch": 0, "step": 1042, "train_loss": 4.07534646987915, "perplexity": 58.87087403989913, "lr": 0.001826404040429639, "grad_norm": 0.301403, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:10.981322+00:00", "epoch": 0, "step": 1043, "train_loss": 4.027530193328857, "perplexity": 56.12212912625604, "lr": 0.0018281568274166157, "grad_norm": 0.269176, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:11.286478+00:00", "epoch": 0, "step": 1044, "train_loss": 4.143223762512207, "perplexity": 63.00560952532396, "lr": 0.001829909614403592, "grad_norm": 0.263377, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:11.590990+00:00", "epoch": 0, "step": 1045, "train_loss": 4.160764217376709, "perplexity": 64.1205058854147, "lr": 0.0018316624013905687, "grad_norm": 0.288596, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:11.895807+00:00", "epoch": 0, "step": 1046, "train_loss": 4.172513961791992, "perplexity": 64.87834895378408, "lr": 0.0018334151883775455, "grad_norm": 0.275497, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:12.201453+00:00", "epoch": 0, "step": 1047, "train_loss": 4.0599541664123535, "perplexity": 57.97165396918411, "lr": 0.0018351679753645222, "grad_norm": 0.291333, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:12.506443+00:00", "epoch": 0, "step": 1048, "train_loss": 4.025880813598633, "perplexity": 56.02963872092602, "lr": 0.0018369207623514987, "grad_norm": 0.306982, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:12.811685+00:00", "epoch": 0, "step": 1049, "train_loss": 4.121678352355957, "perplexity": 61.662647114327534, "lr": 0.0018386735493384754, "grad_norm": 0.341972, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:13.117131+00:00", "epoch": 0, "step": 1050, "train_loss": 4.0904693603515625, "perplexity": 59.76793782207078, "lr": 0.001840426336325452, "grad_norm": 0.440612, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:13.422235+00:00", "epoch": 0, "step": 1051, "train_loss": 4.109780311584473, "perplexity": 60.93332975243577, "lr": 0.0018421791233124284, "grad_norm": 0.444994, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:13.726700+00:00", "epoch": 0, "step": 1052, "train_loss": 4.083942413330078, "perplexity": 59.37910598112149, "lr": 0.0018439319102994051, "grad_norm": 0.462711, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:14.032286+00:00", "epoch": 0, "step": 1053, "train_loss": 4.029740333557129, "perplexity": 56.24630407301016, "lr": 0.0018456846972863818, "grad_norm": 0.389594, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:14.338664+00:00", "epoch": 0, "step": 1054, "train_loss": 4.187528610229492, "perplexity": 65.85982437013888, "lr": 0.0018474374842733586, "grad_norm": 0.342099, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:14.644733+00:00", "epoch": 0, "step": 1055, "train_loss": 3.961827039718628, "perplexity": 52.55325517370793, "lr": 0.0018491902712603353, "grad_norm": 0.358065, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:14.950145+00:00", "epoch": 0, "step": 1056, "train_loss": 4.1227545738220215, "perplexity": 61.72904550208259, "lr": 0.0018509430582473116, "grad_norm": 0.317618, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:15.254859+00:00", "epoch": 0, "step": 1057, "train_loss": 4.072327136993408, "perplexity": 58.693391348394776, "lr": 0.0018526958452342883, "grad_norm": 0.307403, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:15.559986+00:00", "epoch": 0, "step": 1058, "train_loss": 3.985180139541626, "perplexity": 53.79497920800808, "lr": 0.001854448632221265, "grad_norm": 0.311302, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:15.865857+00:00", "epoch": 0, "step": 1059, "train_loss": 4.080973148345947, "perplexity": 59.2030551810273, "lr": 0.0018562014192082415, "grad_norm": 0.322311, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:16.172271+00:00", "epoch": 0, "step": 1060, "train_loss": 4.09348726272583, "perplexity": 59.948584072797075, "lr": 0.0018579542061952182, "grad_norm": 0.305172, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:16.477294+00:00", "epoch": 0, "step": 1061, "train_loss": 4.155807018280029, "perplexity": 63.80343431431857, "lr": 0.0018597069931821947, "grad_norm": 0.279982, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:16.782095+00:00", "epoch": 0, "step": 1062, "train_loss": 4.049439907073975, "perplexity": 57.365318137011, "lr": 0.0018614597801691712, "grad_norm": 0.298137, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:17.086831+00:00", "epoch": 0, "step": 1063, "train_loss": 4.105165481567383, "perplexity": 60.65278063439492, "lr": 0.001863212567156148, "grad_norm": 0.29855, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:17.392286+00:00", "epoch": 0, "step": 1064, "train_loss": 4.150347709655762, "perplexity": 63.45606074762396, "lr": 0.0018649653541431247, "grad_norm": 0.279082, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:17.697320+00:00", "epoch": 0, "step": 1065, "train_loss": 3.9990899562835693, "perplexity": 54.54848593145724, "lr": 0.0018667181411301014, "grad_norm": 0.267533, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:18.002671+00:00", "epoch": 0, "step": 1066, "train_loss": 4.121771812438965, "perplexity": 61.668410379758754, "lr": 0.001868470928117078, "grad_norm": 0.271608, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:18.309140+00:00", "epoch": 0, "step": 1067, "train_loss": 4.092818260192871, "perplexity": 59.90849173063962, "lr": 0.0018702237151040544, "grad_norm": 0.272954, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:18.614517+00:00", "epoch": 0, "step": 1068, "train_loss": 4.127701282501221, "perplexity": 62.035157606717355, "lr": 0.001871976502091031, "grad_norm": 0.26574, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:18.919498+00:00", "epoch": 0, "step": 1069, "train_loss": 4.081655025482178, "perplexity": 59.24343815729965, "lr": 0.0018737292890780078, "grad_norm": 0.254532, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:19.225204+00:00", "epoch": 0, "step": 1070, "train_loss": 4.062447547912598, "perplexity": 58.11637977206976, "lr": 0.0018754820760649843, "grad_norm": 0.27368, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:19.530920+00:00", "epoch": 0, "step": 1071, "train_loss": 4.043220520019531, "perplexity": 57.009648191009596, "lr": 0.001877234863051961, "grad_norm": 0.313767, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:19.836063+00:00", "epoch": 0, "step": 1072, "train_loss": 4.03133487701416, "perplexity": 56.33606279222304, "lr": 0.0018789876500389378, "grad_norm": 0.406151, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:20.141605+00:00", "epoch": 0, "step": 1073, "train_loss": 4.120880603790283, "perplexity": 61.61347544195973, "lr": 0.001880740437025914, "grad_norm": 0.427752, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:20.446983+00:00", "epoch": 0, "step": 1074, "train_loss": 4.007477283477783, "perplexity": 55.007925974537976, "lr": 0.0018824932240128908, "grad_norm": 0.322412, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:20.751996+00:00", "epoch": 0, "step": 1075, "train_loss": 4.073915958404541, "perplexity": 58.786718785944, "lr": 0.0018842460109998675, "grad_norm": 0.306395, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:21.058214+00:00", "epoch": 0, "step": 1076, "train_loss": 4.071696758270264, "perplexity": 58.65640394256121, "lr": 0.0018859987979868442, "grad_norm": 0.301802, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:21.363177+00:00", "epoch": 0, "step": 1077, "train_loss": 4.033901214599609, "perplexity": 56.480825823535234, "lr": 0.001887751584973821, "grad_norm": 0.272684, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:21.667364+00:00", "epoch": 0, "step": 1078, "train_loss": 4.17078971862793, "perplexity": 64.76657929083458, "lr": 0.0018895043719607972, "grad_norm": 0.249854, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:21.972315+00:00", "epoch": 0, "step": 1079, "train_loss": 4.081202030181885, "perplexity": 59.21660723584068, "lr": 0.001891257158947774, "grad_norm": 0.251047, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:22.278504+00:00", "epoch": 0, "step": 1080, "train_loss": 3.9766857624053955, "perplexity": 53.33995965638772, "lr": 0.0018930099459347506, "grad_norm": 0.241596, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:22.583943+00:00", "epoch": 0, "step": 1081, "train_loss": 4.0007004737854, "perplexity": 54.63640800376225, "lr": 0.0018947627329217271, "grad_norm": 0.259872, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:22.890846+00:00", "epoch": 0, "step": 1082, "train_loss": 4.059422492980957, "perplexity": 57.94084017315878, "lr": 0.0018965155199087039, "grad_norm": 0.253388, "tokens_per_sec": 106771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:23.195978+00:00", "epoch": 0, "step": 1083, "train_loss": 4.103548526763916, "perplexity": 60.554787076332424, "lr": 0.0018982683068956806, "grad_norm": 0.27311, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:23.500757+00:00", "epoch": 0, "step": 1084, "train_loss": 4.0413408279418945, "perplexity": 56.90258825833213, "lr": 0.0019000210938826569, "grad_norm": 0.267043, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:23.806054+00:00", "epoch": 0, "step": 1085, "train_loss": 4.049970626831055, "perplexity": 57.395771125003904, "lr": 0.0019017738808696336, "grad_norm": 0.262972, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:24.112618+00:00", "epoch": 0, "step": 1086, "train_loss": 4.064040184020996, "perplexity": 58.20901176193007, "lr": 0.0019035266678566103, "grad_norm": 0.278821, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:24.418425+00:00", "epoch": 0, "step": 1087, "train_loss": 4.021273136138916, "perplexity": 55.77206607904726, "lr": 0.001905279454843587, "grad_norm": 0.329512, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:24.728240+00:00", "epoch": 0, "step": 1088, "train_loss": 4.039807319641113, "perplexity": 56.81539454013476, "lr": 0.0019070322418305637, "grad_norm": 0.289774, "tokens_per_sec": 105766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:25.033458+00:00", "epoch": 0, "step": 1089, "train_loss": 4.082388401031494, "perplexity": 59.28690178193341, "lr": 0.0019087850288175402, "grad_norm": 0.279192, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:25.338567+00:00", "epoch": 0, "step": 1090, "train_loss": 4.053747653961182, "perplexity": 57.61296642776538, "lr": 0.0019105378158045167, "grad_norm": 0.274128, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:25.644173+00:00", "epoch": 0, "step": 1091, "train_loss": 3.963317632675171, "perplexity": 52.63164909791382, "lr": 0.0019122906027914935, "grad_norm": 0.272734, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:25.950575+00:00", "epoch": 0, "step": 1092, "train_loss": 4.023057460784912, "perplexity": 55.871670387894994, "lr": 0.00191404338977847, "grad_norm": 0.257907, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:26.255671+00:00", "epoch": 0, "step": 1093, "train_loss": 4.176027774810791, "perplexity": 65.10672033316624, "lr": 0.0019157961767654467, "grad_norm": 0.253655, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:26.560490+00:00", "epoch": 0, "step": 1094, "train_loss": 4.046736240386963, "perplexity": 57.210430913437165, "lr": 0.0019175489637524234, "grad_norm": 0.297172, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:26.867088+00:00", "epoch": 0, "step": 1095, "train_loss": 4.067965030670166, "perplexity": 58.43792213189205, "lr": 0.0019193017507393997, "grad_norm": 0.327365, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:27.172325+00:00", "epoch": 0, "step": 1096, "train_loss": 4.147455215454102, "perplexity": 63.272779657382095, "lr": 0.0019210545377263764, "grad_norm": 0.366057, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:27.478299+00:00", "epoch": 0, "step": 1097, "train_loss": 4.004758834838867, "perplexity": 54.858592822259034, "lr": 0.0019228073247133531, "grad_norm": 0.412201, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:27.783793+00:00", "epoch": 0, "step": 1098, "train_loss": 4.067981719970703, "perplexity": 58.43889742807576, "lr": 0.0019245601117003298, "grad_norm": 0.414019, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:28.088639+00:00", "epoch": 0, "step": 1099, "train_loss": 4.077948570251465, "perplexity": 59.0242614413161, "lr": 0.0019263128986873065, "grad_norm": 0.37142, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:28.395029+00:00", "epoch": 0, "step": 1100, "train_loss": 4.059824466705322, "perplexity": 57.9641355502271, "lr": 0.001928065685674283, "grad_norm": 0.370568, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:28.699942+00:00", "epoch": 0, "step": 1101, "train_loss": 4.029440879821777, "perplexity": 56.22946342877853, "lr": 0.0019298184726612596, "grad_norm": 0.333789, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:29.005226+00:00", "epoch": 0, "step": 1102, "train_loss": 4.032313823699951, "perplexity": 56.39123979745069, "lr": 0.0019315712596482363, "grad_norm": 0.320068, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:29.310879+00:00", "epoch": 0, "step": 1103, "train_loss": 4.081854820251465, "perplexity": 59.25527586887511, "lr": 0.0019333240466352128, "grad_norm": 0.28203, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:29.616712+00:00", "epoch": 0, "step": 1104, "train_loss": 4.026017665863037, "perplexity": 56.037307028559404, "lr": 0.0019350768336221895, "grad_norm": 0.26954, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:29.922649+00:00", "epoch": 0, "step": 1105, "train_loss": 4.079593181610107, "perplexity": 59.12141327872805, "lr": 0.0019368296206091662, "grad_norm": 0.303166, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:30.229474+00:00", "epoch": 0, "step": 1106, "train_loss": 4.068478584289551, "perplexity": 58.46794084576204, "lr": 0.001938582407596143, "grad_norm": 0.291257, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:30.535500+00:00", "epoch": 0, "step": 1107, "train_loss": 4.003531455993652, "perplexity": 54.791301850156145, "lr": 0.0019403351945831192, "grad_norm": 0.329937, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:30.842158+00:00", "epoch": 0, "step": 1108, "train_loss": 4.033063888549805, "perplexity": 56.433552751008236, "lr": 0.001942087981570096, "grad_norm": 0.337713, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:31.147339+00:00", "epoch": 0, "step": 1109, "train_loss": 4.054015159606934, "perplexity": 57.628380283108065, "lr": 0.0019438407685570726, "grad_norm": 0.323453, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:31.452549+00:00", "epoch": 0, "step": 1110, "train_loss": 3.928685188293457, "perplexity": 50.84008856194397, "lr": 0.0019455935555440494, "grad_norm": 0.295479, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:31.758299+00:00", "epoch": 0, "step": 1111, "train_loss": 4.0637054443359375, "perplexity": 58.18953015648046, "lr": 0.0019473463425310259, "grad_norm": 0.287187, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:32.064269+00:00", "epoch": 0, "step": 1112, "train_loss": 3.9820523262023926, "perplexity": 53.626981424275144, "lr": 0.0019490991295180024, "grad_norm": 0.30321, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:32.370574+00:00", "epoch": 0, "step": 1113, "train_loss": 3.988403797149658, "perplexity": 53.96867562046605, "lr": 0.001950851916504979, "grad_norm": 0.283285, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:32.676916+00:00", "epoch": 0, "step": 1114, "train_loss": 4.026247501373291, "perplexity": 56.050187871794456, "lr": 0.0019526047034919556, "grad_norm": 0.251679, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:32.981861+00:00", "epoch": 0, "step": 1115, "train_loss": 4.034475326538086, "perplexity": 56.51326144988479, "lr": 0.0019543574904789325, "grad_norm": 0.251898, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:33.287938+00:00", "epoch": 0, "step": 1116, "train_loss": 4.071524620056152, "perplexity": 58.646307802931005, "lr": 0.001956110277465909, "grad_norm": 0.259666, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:33.593546+00:00", "epoch": 0, "step": 1117, "train_loss": 4.031569957733154, "perplexity": 56.34930787113988, "lr": 0.0019578630644528855, "grad_norm": 0.243413, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:33.899480+00:00", "epoch": 0, "step": 1118, "train_loss": 4.069716453552246, "perplexity": 58.540361326884266, "lr": 0.0019596158514398622, "grad_norm": 0.238044, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:34.205774+00:00", "epoch": 0, "step": 1119, "train_loss": 4.028097152709961, "perplexity": 56.153957115587666, "lr": 0.0019613686384268385, "grad_norm": 0.239453, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:34.511923+00:00", "epoch": 0, "step": 1120, "train_loss": 3.9829294681549072, "perplexity": 53.67404053521214, "lr": 0.0019631214254138152, "grad_norm": 0.246759, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:34.818288+00:00", "epoch": 0, "step": 1121, "train_loss": 4.000133514404297, "perplexity": 54.60544015927922, "lr": 0.001964874212400792, "grad_norm": 0.274668, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:35.124270+00:00", "epoch": 0, "step": 1122, "train_loss": 3.966874122619629, "perplexity": 52.81916628252877, "lr": 0.0019666269993877687, "grad_norm": 0.277361, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:35.430033+00:00", "epoch": 0, "step": 1123, "train_loss": 3.98199725151062, "perplexity": 53.624028016132286, "lr": 0.0019683797863747454, "grad_norm": 0.243822, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:35.736177+00:00", "epoch": 0, "step": 1124, "train_loss": 3.9481327533721924, "perplexity": 51.838481179102324, "lr": 0.0019701325733617217, "grad_norm": 0.234874, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:36.043339+00:00", "epoch": 0, "step": 1125, "train_loss": 3.982499599456787, "perplexity": 53.65097270370899, "lr": 0.0019718853603486984, "grad_norm": 0.225024, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:36.348759+00:00", "epoch": 0, "step": 1126, "train_loss": 4.021067142486572, "perplexity": 55.76057857067441, "lr": 0.001973638147335675, "grad_norm": 0.244603, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:36.654104+00:00", "epoch": 0, "step": 1127, "train_loss": 3.9989993572235107, "perplexity": 54.54354411376963, "lr": 0.001975390934322652, "grad_norm": 0.262798, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:36.960464+00:00", "epoch": 0, "step": 1128, "train_loss": 4.060509204864502, "perplexity": 58.003839397520544, "lr": 0.0019771437213096286, "grad_norm": 0.274722, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:37.266698+00:00", "epoch": 0, "step": 1129, "train_loss": 4.002896785736084, "perplexity": 54.75653847330598, "lr": 0.001978896508296605, "grad_norm": 0.280357, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:37.573285+00:00", "epoch": 0, "step": 1130, "train_loss": 4.0578107833862305, "perplexity": 57.84753157850208, "lr": 0.0019806492952835816, "grad_norm": 0.290059, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:37.879530+00:00", "epoch": 0, "step": 1131, "train_loss": 4.112335205078125, "perplexity": 61.089206960208365, "lr": 0.0019824020822705583, "grad_norm": 0.291234, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:38.187033+00:00", "epoch": 0, "step": 1132, "train_loss": 3.9823760986328125, "perplexity": 53.64434717351068, "lr": 0.001984154869257535, "grad_norm": 0.265804, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:38.494006+00:00", "epoch": 0, "step": 1133, "train_loss": 4.047329902648926, "perplexity": 57.244404670732116, "lr": 0.0019859076562445117, "grad_norm": 0.271307, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:38.799031+00:00", "epoch": 0, "step": 1134, "train_loss": 4.088447093963623, "perplexity": 59.64719326032048, "lr": 0.0019876604432314884, "grad_norm": 0.304981, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:39.104809+00:00", "epoch": 0, "step": 1135, "train_loss": 3.9907121658325195, "perplexity": 54.09339911965049, "lr": 0.0019894132302184647, "grad_norm": 0.294201, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:39.411941+00:00", "epoch": 0, "step": 1136, "train_loss": 4.045525550842285, "perplexity": 57.14120875461104, "lr": 0.0019911660172054414, "grad_norm": 0.264513, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:39.717714+00:00", "epoch": 0, "step": 1137, "train_loss": 3.950005531311035, "perplexity": 51.93565410629362, "lr": 0.001992918804192418, "grad_norm": 0.293753, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:40.024194+00:00", "epoch": 0, "step": 1138, "train_loss": 3.9773173332214355, "perplexity": 53.373658258642166, "lr": 0.0019946715911793944, "grad_norm": 0.30609, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:40.329736+00:00", "epoch": 0, "step": 1139, "train_loss": 4.090953350067139, "perplexity": 59.79687189062885, "lr": 0.001996424378166371, "grad_norm": 0.282312, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:40.635628+00:00", "epoch": 0, "step": 1140, "train_loss": 4.089844226837158, "perplexity": 59.730586557031536, "lr": 0.001998177165153348, "grad_norm": 0.269732, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:40.942272+00:00", "epoch": 0, "step": 1141, "train_loss": 4.0452375411987305, "perplexity": 57.12475390513675, "lr": 0.001999929952140324, "grad_norm": 0.290678, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:41.248699+00:00", "epoch": 0, "step": 1142, "train_loss": 4.0242085456848145, "perplexity": 55.9360204530997, "lr": 0.002001682739127301, "grad_norm": 0.295958, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:41.554991+00:00", "epoch": 0, "step": 1143, "train_loss": 3.967679500579834, "perplexity": 52.86172280967334, "lr": 0.0020034355261142776, "grad_norm": 0.278994, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:41.860390+00:00", "epoch": 0, "step": 1144, "train_loss": 3.98468017578125, "perplexity": 53.76809039019123, "lr": 0.0020051883131012543, "grad_norm": 0.285088, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:42.165809+00:00", "epoch": 0, "step": 1145, "train_loss": 3.944355010986328, "perplexity": 51.643018188405414, "lr": 0.002006941100088231, "grad_norm": 0.269984, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:42.472103+00:00", "epoch": 0, "step": 1146, "train_loss": 3.953784704208374, "perplexity": 52.13229926668508, "lr": 0.0020086938870752073, "grad_norm": 0.264985, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:42.778785+00:00", "epoch": 0, "step": 1147, "train_loss": 4.051050662994385, "perplexity": 57.45779412095138, "lr": 0.002010446674062184, "grad_norm": 0.309761, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:43.085353+00:00", "epoch": 0, "step": 1148, "train_loss": 3.9957525730133057, "perplexity": 54.3667401734098, "lr": 0.0020121994610491608, "grad_norm": 0.33935, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:43.391413+00:00", "epoch": 0, "step": 1149, "train_loss": 4.040157318115234, "perplexity": 56.835283321849, "lr": 0.0020139522480361375, "grad_norm": 0.330105, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:43.696764+00:00", "epoch": 0, "step": 1150, "train_loss": 4.048668384552002, "perplexity": 57.32107657096503, "lr": 0.002015705035023114, "grad_norm": 0.331571, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:44.002884+00:00", "epoch": 0, "step": 1151, "train_loss": 4.057726860046387, "perplexity": 57.842677024158355, "lr": 0.002017457822010091, "grad_norm": 0.306374, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:44.310920+00:00", "epoch": 0, "step": 1152, "train_loss": 4.073215007781982, "perplexity": 58.74552663733051, "lr": 0.002019210608997067, "grad_norm": 0.274126, "tokens_per_sec": 106377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:44.617935+00:00", "epoch": 0, "step": 1153, "train_loss": 3.968142509460449, "perplexity": 52.88620392382812, "lr": 0.002020963395984044, "grad_norm": 0.270515, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:44.924669+00:00", "epoch": 0, "step": 1154, "train_loss": 4.055757522583008, "perplexity": 57.72887736505755, "lr": 0.0020227161829710206, "grad_norm": 0.244598, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:45.229869+00:00", "epoch": 0, "step": 1155, "train_loss": 3.973288059234619, "perplexity": 53.15903384648954, "lr": 0.0020244689699579974, "grad_norm": 0.229676, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:45.535533+00:00", "epoch": 0, "step": 1156, "train_loss": 4.080663681030273, "perplexity": 59.1847366051072, "lr": 0.002026221756944974, "grad_norm": 0.260488, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:45.842012+00:00", "epoch": 0, "step": 1157, "train_loss": 4.047287464141846, "perplexity": 57.24197535520783, "lr": 0.0020279745439319504, "grad_norm": 0.270988, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:46.149315+00:00", "epoch": 0, "step": 1158, "train_loss": 4.051719665527344, "perplexity": 57.496246391655625, "lr": 0.002029727330918927, "grad_norm": 0.270519, "tokens_per_sec": 106632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:46.455855+00:00", "epoch": 0, "step": 1159, "train_loss": 4.111791610717773, "perplexity": 61.05600823596358, "lr": 0.002031480117905904, "grad_norm": 0.319003, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:46.761715+00:00", "epoch": 0, "step": 1160, "train_loss": 4.054335594177246, "perplexity": 57.646849367290386, "lr": 0.00203323290489288, "grad_norm": 0.285235, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:47.067139+00:00", "epoch": 0, "step": 1161, "train_loss": 4.007773399353027, "perplexity": 55.024217106595664, "lr": 0.002034985691879857, "grad_norm": 0.286168, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:47.372905+00:00", "epoch": 0, "step": 1162, "train_loss": 4.0517048835754395, "perplexity": 57.495396491188394, "lr": 0.0020367384788668335, "grad_norm": 0.304388, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:47.679340+00:00", "epoch": 0, "step": 1163, "train_loss": 4.0711894035339355, "perplexity": 58.62665188626652, "lr": 0.00203849126585381, "grad_norm": 0.322617, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:47.985834+00:00", "epoch": 0, "step": 1164, "train_loss": 3.9813666343688965, "perplexity": 53.590222445163626, "lr": 0.0020402440528407865, "grad_norm": 0.321024, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:48.292311+00:00", "epoch": 0, "step": 1165, "train_loss": 3.929396152496338, "perplexity": 50.87624689710069, "lr": 0.0020419968398277632, "grad_norm": 0.313874, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:48.598359+00:00", "epoch": 0, "step": 1166, "train_loss": 4.109826564788818, "perplexity": 60.9361481793684, "lr": 0.00204374962681474, "grad_norm": 0.260681, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:48.904817+00:00", "epoch": 0, "step": 1167, "train_loss": 4.115427017211914, "perplexity": 61.278375597782656, "lr": 0.0020455024138017167, "grad_norm": 0.278418, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:49.210208+00:00", "epoch": 0, "step": 1168, "train_loss": 4.017033100128174, "perplexity": 55.536091135136346, "lr": 0.0020472552007886934, "grad_norm": 0.283694, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:49.517251+00:00", "epoch": 0, "step": 1169, "train_loss": 4.020753860473633, "perplexity": 55.743112520419665, "lr": 0.0020490079877756697, "grad_norm": 0.295796, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:49.823524+00:00", "epoch": 0, "step": 1170, "train_loss": 3.9806249141693115, "perplexity": 53.550488232331574, "lr": 0.0020507607747626464, "grad_norm": 0.276536, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:50.128704+00:00", "epoch": 0, "step": 1171, "train_loss": 3.9544026851654053, "perplexity": 52.16452599160368, "lr": 0.002052513561749623, "grad_norm": 0.256388, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:50.434494+00:00", "epoch": 0, "step": 1172, "train_loss": 4.052488803863525, "perplexity": 57.54048596993889, "lr": 0.0020542663487366, "grad_norm": 0.24781, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:50.740461+00:00", "epoch": 0, "step": 1173, "train_loss": 3.9432570934295654, "perplexity": 51.58634952650404, "lr": 0.0020560191357235766, "grad_norm": 0.266905, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:51.046513+00:00", "epoch": 0, "step": 1174, "train_loss": 4.141757011413574, "perplexity": 62.91326371904482, "lr": 0.002057771922710553, "grad_norm": 0.297864, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:51.352877+00:00", "epoch": 0, "step": 1175, "train_loss": 3.983673334121704, "perplexity": 53.71398168085448, "lr": 0.0020595247096975296, "grad_norm": 0.268103, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:51.658800+00:00", "epoch": 0, "step": 1176, "train_loss": 3.936028480529785, "perplexity": 51.21479629997862, "lr": 0.0020612774966845063, "grad_norm": 0.246768, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:51.965317+00:00", "epoch": 0, "step": 1177, "train_loss": 3.9388928413391113, "perplexity": 51.36170425360743, "lr": 0.002063030283671483, "grad_norm": 0.23619, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:52.272304+00:00", "epoch": 0, "step": 1178, "train_loss": 4.058218479156494, "perplexity": 57.8711205806883, "lr": 0.0020647830706584593, "grad_norm": 0.239422, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:52.577298+00:00", "epoch": 0, "step": 1179, "train_loss": 3.9460415840148926, "perplexity": 51.730191401326614, "lr": 0.002066535857645436, "grad_norm": 0.265487, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:52.883926+00:00", "epoch": 0, "step": 1180, "train_loss": 4.043519973754883, "perplexity": 57.0267224994666, "lr": 0.0020682886446324127, "grad_norm": 0.282492, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:53.191153+00:00", "epoch": 0, "step": 1181, "train_loss": 4.059670448303223, "perplexity": 57.955208694158266, "lr": 0.002070041431619389, "grad_norm": 0.261178, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:53.496984+00:00", "epoch": 0, "step": 1182, "train_loss": 3.889704465866089, "perplexity": 48.89643382298266, "lr": 0.0020717942186063657, "grad_norm": 0.275531, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:53.802861+00:00", "epoch": 0, "step": 1183, "train_loss": 3.9537899494171143, "perplexity": 52.132572712193976, "lr": 0.0020735470055933424, "grad_norm": 0.269218, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:54.109185+00:00", "epoch": 0, "step": 1184, "train_loss": 3.972588539123535, "perplexity": 53.12186103630809, "lr": 0.002075299792580319, "grad_norm": 0.247453, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:54.415683+00:00", "epoch": 0, "step": 1185, "train_loss": 3.9950687885284424, "perplexity": 54.329577746979915, "lr": 0.002077052579567296, "grad_norm": 0.266012, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:54.722419+00:00", "epoch": 0, "step": 1186, "train_loss": 3.9890003204345703, "perplexity": 54.000878796146665, "lr": 0.002078805366554272, "grad_norm": 0.283761, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:55.029256+00:00", "epoch": 0, "step": 1187, "train_loss": 3.9677209854125977, "perplexity": 52.86391581489161, "lr": 0.002080558153541249, "grad_norm": 0.278548, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:55.335877+00:00", "epoch": 0, "step": 1188, "train_loss": 3.9664361476898193, "perplexity": 52.79603787708417, "lr": 0.0020823109405282256, "grad_norm": 0.284096, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:55.642458+00:00", "epoch": 0, "step": 1189, "train_loss": 4.088291168212891, "perplexity": 59.63789345199092, "lr": 0.0020840637275152023, "grad_norm": 0.308512, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:55.948266+00:00", "epoch": 0, "step": 1190, "train_loss": 3.9185967445373535, "perplexity": 50.32976967781845, "lr": 0.002085816514502179, "grad_norm": 0.324514, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:56.254185+00:00", "epoch": 0, "step": 1191, "train_loss": 4.082468032836914, "perplexity": 59.29162309294082, "lr": 0.0020875693014891553, "grad_norm": 0.286605, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:56.560020+00:00", "epoch": 0, "step": 1192, "train_loss": 4.107198715209961, "perplexity": 60.77622736399073, "lr": 0.002089322088476132, "grad_norm": 0.262414, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:56.865476+00:00", "epoch": 0, "step": 1193, "train_loss": 4.014194965362549, "perplexity": 55.378695684540745, "lr": 0.0020910748754631087, "grad_norm": 0.256599, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:57.170890+00:00", "epoch": 0, "step": 1194, "train_loss": 3.9754140377044678, "perplexity": 53.27216902679167, "lr": 0.0020928276624500855, "grad_norm": 0.254553, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:57.477722+00:00", "epoch": 0, "step": 1195, "train_loss": 4.084754467010498, "perplexity": 59.42734458619106, "lr": 0.002094580449437062, "grad_norm": 0.268692, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:57.783485+00:00", "epoch": 0, "step": 1196, "train_loss": 3.9630541801452637, "perplexity": 52.617784983154024, "lr": 0.002096333236424039, "grad_norm": 0.28504, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:58.088717+00:00", "epoch": 0, "step": 1197, "train_loss": 4.068012714385986, "perplexity": 58.440708735601405, "lr": 0.002098086023411015, "grad_norm": 0.300727, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:58.394486+00:00", "epoch": 0, "step": 1198, "train_loss": 3.987344741821289, "perplexity": 53.9115500618881, "lr": 0.002099838810397992, "grad_norm": 0.307145, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:58.701628+00:00", "epoch": 0, "step": 1199, "train_loss": 3.9632728099823, "perplexity": 52.62929005854065, "lr": 0.0021015915973849686, "grad_norm": 0.276487, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:59.008088+00:00", "epoch": 0, "step": 1200, "train_loss": 4.061793804168701, "perplexity": 58.07839896858488, "lr": 0.002103344384371945, "grad_norm": 0.261747, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:59.313078+00:00", "epoch": 0, "step": 1201, "train_loss": 4.071730613708496, "perplexity": 58.65838981443792, "lr": 0.0021050971713589216, "grad_norm": 0.245727, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:59.617778+00:00", "epoch": 0, "step": 1202, "train_loss": 3.9986934661865234, "perplexity": 54.52686228404088, "lr": 0.0021068499583458983, "grad_norm": 0.244348, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:03:59.923649+00:00", "epoch": 0, "step": 1203, "train_loss": 4.044559001922607, "perplexity": 57.08600566356184, "lr": 0.0021086027453328746, "grad_norm": 0.26987, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:00.230032+00:00", "epoch": 0, "step": 1204, "train_loss": 4.081178188323975, "perplexity": 59.21519541873528, "lr": 0.0021103555323198513, "grad_norm": 0.304528, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:00.535634+00:00", "epoch": 0, "step": 1205, "train_loss": 4.022037506103516, "perplexity": 55.81471286810767, "lr": 0.002112108319306828, "grad_norm": 0.30874, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:00.840987+00:00", "epoch": 0, "step": 1206, "train_loss": 3.98068904876709, "perplexity": 53.553922781490726, "lr": 0.002113861106293805, "grad_norm": 0.300666, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:01.146764+00:00", "epoch": 0, "step": 1207, "train_loss": 3.9770994186401367, "perplexity": 53.362028627429474, "lr": 0.0021156138932807815, "grad_norm": 0.271755, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:01.453228+00:00", "epoch": 0, "step": 1208, "train_loss": 4.051353454589844, "perplexity": 57.47519449231498, "lr": 0.002117366680267758, "grad_norm": 0.253682, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:01.759633+00:00", "epoch": 0, "step": 1209, "train_loss": 4.029089450836182, "perplexity": 56.20970623731338, "lr": 0.0021191194672547345, "grad_norm": 0.244418, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:02.066326+00:00", "epoch": 0, "step": 1210, "train_loss": 4.074974536895752, "perplexity": 58.8489820915034, "lr": 0.0021208722542417112, "grad_norm": 0.241584, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:02.371602+00:00", "epoch": 0, "step": 1211, "train_loss": 3.973594903945923, "perplexity": 53.17534791769859, "lr": 0.002122625041228688, "grad_norm": 0.261398, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:02.676545+00:00", "epoch": 0, "step": 1212, "train_loss": 4.091343879699707, "perplexity": 59.82022890154263, "lr": 0.0021243778282156647, "grad_norm": 0.267841, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:02.981594+00:00", "epoch": 0, "step": 1213, "train_loss": 4.005640983581543, "perplexity": 54.9070076123256, "lr": 0.0021261306152026414, "grad_norm": 0.272095, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:03.287763+00:00", "epoch": 0, "step": 1214, "train_loss": 3.9649362564086914, "perplexity": 52.71690891744708, "lr": 0.0021278834021896177, "grad_norm": 0.227285, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:03.594305+00:00", "epoch": 0, "step": 1215, "train_loss": 3.958838939666748, "perplexity": 52.396455172911054, "lr": 0.0021296361891765944, "grad_norm": 0.238909, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:03.899616+00:00", "epoch": 0, "step": 1216, "train_loss": 3.9438490867614746, "perplexity": 51.61689734260084, "lr": 0.002131388976163571, "grad_norm": 0.226142, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:04.204925+00:00", "epoch": 0, "step": 1217, "train_loss": 3.891854763031006, "perplexity": 49.00168881020077, "lr": 0.002133141763150548, "grad_norm": 0.230743, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:04.510052+00:00", "epoch": 0, "step": 1218, "train_loss": 4.010005950927734, "perplexity": 55.14719873948179, "lr": 0.0021348945501375245, "grad_norm": 0.235002, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:04.815359+00:00", "epoch": 0, "step": 1219, "train_loss": 4.008101940155029, "perplexity": 55.04229777696973, "lr": 0.002136647337124501, "grad_norm": 0.250877, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:05.121760+00:00", "epoch": 0, "step": 1220, "train_loss": 3.9123475551605225, "perplexity": 50.016230120216946, "lr": 0.0021384001241114775, "grad_norm": 0.264419, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:05.428129+00:00", "epoch": 0, "step": 1221, "train_loss": 3.9728810787200928, "perplexity": 53.13740355739462, "lr": 0.0021401529110984543, "grad_norm": 0.256939, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:05.733744+00:00", "epoch": 0, "step": 1222, "train_loss": 3.9905638694763184, "perplexity": 54.08537786044297, "lr": 0.0021419056980854305, "grad_norm": 0.243071, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:06.038077+00:00", "epoch": 0, "step": 1223, "train_loss": 4.028526306152344, "perplexity": 56.17806095133934, "lr": 0.0021436584850724073, "grad_norm": 0.277455, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:06.343668+00:00", "epoch": 0, "step": 1224, "train_loss": 4.007555961608887, "perplexity": 55.01225406561036, "lr": 0.002145411272059384, "grad_norm": 0.305612, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:06.650464+00:00", "epoch": 0, "step": 1225, "train_loss": 4.099318027496338, "perplexity": 60.299151208894756, "lr": 0.0021471640590463603, "grad_norm": 0.334556, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:06.956486+00:00", "epoch": 0, "step": 1226, "train_loss": 3.958634376525879, "perplexity": 52.38573788568881, "lr": 0.002148916846033337, "grad_norm": 0.296872, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:07.262592+00:00", "epoch": 0, "step": 1227, "train_loss": 3.9960343837738037, "perplexity": 54.38206346483667, "lr": 0.0021506696330203137, "grad_norm": 0.277216, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:07.568587+00:00", "epoch": 0, "step": 1228, "train_loss": 3.9521396160125732, "perplexity": 52.04660754109036, "lr": 0.0021524224200072904, "grad_norm": 0.284174, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:07.873603+00:00", "epoch": 0, "step": 1229, "train_loss": 4.039280414581299, "perplexity": 56.78546610669014, "lr": 0.002154175206994267, "grad_norm": 0.312019, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:08.179504+00:00", "epoch": 0, "step": 1230, "train_loss": 4.054451942443848, "perplexity": 57.65355686848481, "lr": 0.002155927993981244, "grad_norm": 0.310107, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:08.486211+00:00", "epoch": 0, "step": 1231, "train_loss": 4.09753942489624, "perplexity": 60.19199830123855, "lr": 0.00215768078096822, "grad_norm": 0.286135, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:08.791454+00:00", "epoch": 0, "step": 1232, "train_loss": 4.0857696533203125, "perplexity": 59.48770504621595, "lr": 0.002159433567955197, "grad_norm": 0.304084, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:09.096976+00:00", "epoch": 0, "step": 1233, "train_loss": 3.8242244720458984, "perplexity": 45.79726952779469, "lr": 0.0021611863549421736, "grad_norm": 0.269696, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:09.402269+00:00", "epoch": 0, "step": 1234, "train_loss": 4.064326286315918, "perplexity": 58.22566787634295, "lr": 0.0021629391419291503, "grad_norm": 0.260983, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:09.707639+00:00", "epoch": 0, "step": 1235, "train_loss": 3.875112771987915, "perplexity": 48.18813225616198, "lr": 0.002164691928916127, "grad_norm": 0.264025, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:10.013549+00:00", "epoch": 0, "step": 1236, "train_loss": 4.023643493652344, "perplexity": 55.90442261910524, "lr": 0.0021664447159031037, "grad_norm": 0.270481, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:10.319601+00:00", "epoch": 0, "step": 1237, "train_loss": 3.993318557739258, "perplexity": 54.23457161282708, "lr": 0.00216819750289008, "grad_norm": 0.249723, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:10.626292+00:00", "epoch": 0, "step": 1238, "train_loss": 4.017872333526611, "perplexity": 55.58271844049491, "lr": 0.0021699502898770567, "grad_norm": 0.226573, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:10.930901+00:00", "epoch": 0, "step": 1239, "train_loss": 4.003091335296631, "perplexity": 54.7671923701249, "lr": 0.0021717030768640335, "grad_norm": 0.244952, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:11.236181+00:00", "epoch": 0, "step": 1240, "train_loss": 3.9627254009246826, "perplexity": 52.60048819238707, "lr": 0.00217345586385101, "grad_norm": 0.257564, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:11.540865+00:00", "epoch": 0, "step": 1241, "train_loss": 3.9975855350494385, "perplexity": 54.46648372932552, "lr": 0.0021752086508379865, "grad_norm": 0.247448, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:11.846066+00:00", "epoch": 0, "step": 1242, "train_loss": 3.9785687923431396, "perplexity": 53.44049502314493, "lr": 0.002176961437824963, "grad_norm": 0.274862, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:12.152421+00:00", "epoch": 0, "step": 1243, "train_loss": 3.9714815616607666, "perplexity": 53.06308886909446, "lr": 0.00217871422481194, "grad_norm": 0.264777, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:12.458065+00:00", "epoch": 0, "step": 1244, "train_loss": 3.9900529384613037, "perplexity": 54.05775102174065, "lr": 0.002180467011798916, "grad_norm": 0.254825, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:12.763668+00:00", "epoch": 0, "step": 1245, "train_loss": 4.040029048919678, "perplexity": 56.82799357331315, "lr": 0.002182219798785893, "grad_norm": 0.256411, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:13.069146+00:00", "epoch": 0, "step": 1246, "train_loss": 3.9810447692871094, "perplexity": 53.572976399432164, "lr": 0.0021839725857728696, "grad_norm": 0.308589, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:13.376195+00:00", "epoch": 0, "step": 1247, "train_loss": 3.9296226501464844, "perplexity": 50.887771552575586, "lr": 0.0021857253727598463, "grad_norm": 0.26729, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:13.681388+00:00", "epoch": 0, "step": 1248, "train_loss": 3.9133410453796387, "perplexity": 50.06594544739491, "lr": 0.0021874781597468226, "grad_norm": 0.241302, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:13.987101+00:00", "epoch": 0, "step": 1249, "train_loss": 3.9565577507019043, "perplexity": 52.27706518490894, "lr": 0.0021892309467337993, "grad_norm": 0.253308, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:14.292003+00:00", "epoch": 0, "step": 1250, "train_loss": 4.021023750305176, "perplexity": 55.758159050028816, "lr": 0.002190983733720776, "grad_norm": 0.262745, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:14.597286+00:00", "epoch": 0, "step": 1251, "train_loss": 3.9929769039154053, "perplexity": 54.21604532901862, "lr": 0.0021927365207077528, "grad_norm": 0.26892, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:14.903519+00:00", "epoch": 0, "step": 1252, "train_loss": 4.062769412994385, "perplexity": 58.13508841606637, "lr": 0.0021944893076947295, "grad_norm": 0.260811, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:15.211363+00:00", "epoch": 0, "step": 1253, "train_loss": 3.9850077629089355, "perplexity": 53.78570700981466, "lr": 0.002196242094681706, "grad_norm": 0.255294, "tokens_per_sec": 106445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:15.516695+00:00", "epoch": 0, "step": 1254, "train_loss": 4.076061725616455, "perplexity": 58.91299683280515, "lr": 0.0021979948816686825, "grad_norm": 0.252399, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:15.822635+00:00", "epoch": 0, "step": 1255, "train_loss": 3.8703863620758057, "perplexity": 47.96091277940091, "lr": 0.002199747668655659, "grad_norm": 0.239389, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:16.127677+00:00", "epoch": 0, "step": 1256, "train_loss": 3.949805736541748, "perplexity": 51.92527867077706, "lr": 0.002201500455642636, "grad_norm": 0.220576, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:16.433366+00:00", "epoch": 0, "step": 1257, "train_loss": 3.8623127937316895, "perplexity": 47.57525598032999, "lr": 0.0022032532426296127, "grad_norm": 0.219754, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:16.739889+00:00", "epoch": 0, "step": 1258, "train_loss": 3.962310552597046, "perplexity": 52.578671493451004, "lr": 0.0022050060296165894, "grad_norm": 0.237379, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:17.045404+00:00", "epoch": 0, "step": 1259, "train_loss": 3.9922282695770264, "perplexity": 54.175472524786365, "lr": 0.0022067588166035657, "grad_norm": 0.248802, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:17.351182+00:00", "epoch": 0, "step": 1260, "train_loss": 3.9068851470947266, "perplexity": 49.74376589426005, "lr": 0.0022085116035905424, "grad_norm": 0.248134, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:17.657118+00:00", "epoch": 0, "step": 1261, "train_loss": 3.8955507278442383, "perplexity": 49.18313242589621, "lr": 0.002210264390577519, "grad_norm": 0.237368, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:17.963006+00:00", "epoch": 0, "step": 1262, "train_loss": 3.982307195663452, "perplexity": 53.64065104603957, "lr": 0.002212017177564496, "grad_norm": 0.261435, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:18.269372+00:00", "epoch": 0, "step": 1263, "train_loss": 3.9418697357177734, "perplexity": 51.51483042942211, "lr": 0.002213769964551472, "grad_norm": 0.235087, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:18.574233+00:00", "epoch": 0, "step": 1264, "train_loss": 3.997739315032959, "perplexity": 54.47486022834826, "lr": 0.002215522751538449, "grad_norm": 0.217895, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:18.879695+00:00", "epoch": 0, "step": 1265, "train_loss": 4.1504716873168945, "perplexity": 63.46392836931474, "lr": 0.0022172755385254255, "grad_norm": 0.220142, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:19.184723+00:00", "epoch": 0, "step": 1266, "train_loss": 4.045160293579102, "perplexity": 57.12034132430857, "lr": 0.002219028325512402, "grad_norm": 0.225456, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:19.490069+00:00", "epoch": 0, "step": 1267, "train_loss": 3.9791905879974365, "perplexity": 53.47373442370026, "lr": 0.0022207811124993785, "grad_norm": 0.239476, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:19.795734+00:00", "epoch": 0, "step": 1268, "train_loss": 3.861697196960449, "perplexity": 47.54597781905295, "lr": 0.0022225338994863553, "grad_norm": 0.229979, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:20.101262+00:00", "epoch": 0, "step": 1269, "train_loss": 3.9394965171813965, "perplexity": 51.3927194342997, "lr": 0.002224286686473332, "grad_norm": 0.266052, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:20.407288+00:00", "epoch": 0, "step": 1270, "train_loss": 3.9565696716308594, "perplexity": 52.277688379803514, "lr": 0.0022260394734603087, "grad_norm": 0.287751, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:20.713382+00:00", "epoch": 0, "step": 1271, "train_loss": 3.9930295944213867, "perplexity": 54.218902075140356, "lr": 0.002227792260447285, "grad_norm": 0.306592, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:21.018697+00:00", "epoch": 0, "step": 1272, "train_loss": 3.9562861919403076, "perplexity": 52.262870817217106, "lr": 0.0022295450474342617, "grad_norm": 0.310718, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:21.324632+00:00", "epoch": 0, "step": 1273, "train_loss": 4.015111446380615, "perplexity": 55.42947247236789, "lr": 0.0022312978344212384, "grad_norm": 0.254695, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:21.630452+00:00", "epoch": 0, "step": 1274, "train_loss": 3.9752485752105713, "perplexity": 53.26335521004738, "lr": 0.002233050621408215, "grad_norm": 0.269917, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:21.936614+00:00", "epoch": 0, "step": 1275, "train_loss": 3.935291290283203, "perplexity": 51.17705516457325, "lr": 0.002234803408395192, "grad_norm": 0.233503, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:22.243325+00:00", "epoch": 0, "step": 1276, "train_loss": 3.9767205715179443, "perplexity": 53.34181640536246, "lr": 0.002236556195382168, "grad_norm": 0.217378, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:22.549534+00:00", "epoch": 0, "step": 1277, "train_loss": 3.9573757648468018, "perplexity": 52.31984605897843, "lr": 0.002238308982369145, "grad_norm": 0.214888, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:22.855390+00:00", "epoch": 0, "step": 1278, "train_loss": 3.9214935302734375, "perplexity": 50.47577560857123, "lr": 0.0022400617693561216, "grad_norm": 0.21204, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:23.160639+00:00", "epoch": 0, "step": 1279, "train_loss": 4.010401725769043, "perplexity": 55.16902893294725, "lr": 0.0022418145563430983, "grad_norm": 0.217852, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:23.466996+00:00", "epoch": 0, "step": 1280, "train_loss": 4.071023941040039, "perplexity": 58.616952176727814, "lr": 0.002243567343330075, "grad_norm": 0.233524, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:23.773193+00:00", "epoch": 0, "step": 1281, "train_loss": 3.9627020359039307, "perplexity": 52.59925919524672, "lr": 0.0022453201303170513, "grad_norm": 0.242534, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:24.078474+00:00", "epoch": 0, "step": 1282, "train_loss": 3.9641520977020264, "perplexity": 52.67558669803749, "lr": 0.002247072917304028, "grad_norm": 0.235418, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:24.384163+00:00", "epoch": 0, "step": 1283, "train_loss": 4.065256595611572, "perplexity": 58.27986096067049, "lr": 0.0022488257042910047, "grad_norm": 0.226676, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:24.690012+00:00", "epoch": 0, "step": 1284, "train_loss": 3.9686648845672607, "perplexity": 52.91383757718962, "lr": 0.002250578491277981, "grad_norm": 0.222538, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:24.995501+00:00", "epoch": 0, "step": 1285, "train_loss": 3.9504475593566895, "perplexity": 51.95861619654806, "lr": 0.0022523312782649577, "grad_norm": 0.226887, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:25.301994+00:00", "epoch": 0, "step": 1286, "train_loss": 4.129310131072998, "perplexity": 62.13504311021247, "lr": 0.0022540840652519344, "grad_norm": 0.277119, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:25.608336+00:00", "epoch": 0, "step": 1287, "train_loss": 3.974841594696045, "perplexity": 53.24168247282861, "lr": 0.002255836852238911, "grad_norm": 0.302848, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:25.914296+00:00", "epoch": 0, "step": 1288, "train_loss": 4.114345550537109, "perplexity": 61.212140898453015, "lr": 0.0022575896392258874, "grad_norm": 0.305403, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:26.219511+00:00", "epoch": 0, "step": 1289, "train_loss": 4.07906436920166, "perplexity": 59.09015740675616, "lr": 0.002259342426212864, "grad_norm": 0.349494, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:26.524506+00:00", "epoch": 0, "step": 1290, "train_loss": 4.044027328491211, "perplexity": 57.05566261805631, "lr": 0.002261095213199841, "grad_norm": 0.290209, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:26.829501+00:00", "epoch": 0, "step": 1291, "train_loss": 4.0529465675354, "perplexity": 57.56683194373276, "lr": 0.0022628480001868176, "grad_norm": 0.283003, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:27.137599+00:00", "epoch": 0, "step": 1292, "train_loss": 3.9636588096618652, "perplexity": 52.649608868913866, "lr": 0.0022646007871737943, "grad_norm": 0.26623, "tokens_per_sec": 106356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:27.443195+00:00", "epoch": 0, "step": 1293, "train_loss": 3.9846465587615967, "perplexity": 53.76628289762129, "lr": 0.0022663535741607706, "grad_norm": 0.266032, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:27.749098+00:00", "epoch": 0, "step": 1294, "train_loss": 3.9053759574890137, "perplexity": 49.6687497408613, "lr": 0.0022681063611477473, "grad_norm": 0.254774, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:28.054940+00:00", "epoch": 0, "step": 1295, "train_loss": 3.9679558277130127, "perplexity": 52.87633195635107, "lr": 0.002269859148134724, "grad_norm": 0.262025, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:28.360561+00:00", "epoch": 0, "step": 1296, "train_loss": 4.104910373687744, "perplexity": 60.63730960560767, "lr": 0.0022716119351217008, "grad_norm": 0.267961, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:28.665734+00:00", "epoch": 0, "step": 1297, "train_loss": 4.083495616912842, "perplexity": 59.352581535265365, "lr": 0.0022733647221086775, "grad_norm": 0.259089, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:28.972216+00:00", "epoch": 0, "step": 1298, "train_loss": 3.8771767616271973, "perplexity": 48.287694774533115, "lr": 0.002275117509095654, "grad_norm": 0.266595, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:29.278153+00:00", "epoch": 0, "step": 1299, "train_loss": 3.9900753498077393, "perplexity": 54.05896254230218, "lr": 0.0022768702960826305, "grad_norm": 0.256011, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:29.585083+00:00", "epoch": 0, "step": 1300, "train_loss": 3.9647059440612793, "perplexity": 52.70476896045065, "lr": 0.002278623083069607, "grad_norm": 0.263581, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:29.890682+00:00", "epoch": 0, "step": 1301, "train_loss": 4.072257995605469, "perplexity": 58.68933334614365, "lr": 0.002280375870056584, "grad_norm": 0.248376, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:30.197583+00:00", "epoch": 0, "step": 1302, "train_loss": 4.028787612915039, "perplexity": 56.19274257670149, "lr": 0.0022821286570435606, "grad_norm": 0.253544, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:30.504298+00:00", "epoch": 0, "step": 1303, "train_loss": 3.9752652645111084, "perplexity": 53.26424414560793, "lr": 0.002283881444030537, "grad_norm": 0.28367, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:30.810907+00:00", "epoch": 0, "step": 1304, "train_loss": 3.975982904434204, "perplexity": 53.30248241269278, "lr": 0.0022856342310175136, "grad_norm": 0.287195, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:31.116545+00:00", "epoch": 0, "step": 1305, "train_loss": 4.005385398864746, "perplexity": 54.892976013542324, "lr": 0.0022873870180044904, "grad_norm": 0.270108, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:31.421857+00:00", "epoch": 0, "step": 1306, "train_loss": 3.981895923614502, "perplexity": 53.61859468147086, "lr": 0.0022891398049914666, "grad_norm": 0.233132, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:31.726503+00:00", "epoch": 0, "step": 1307, "train_loss": 4.008272647857666, "perplexity": 55.051694723213814, "lr": 0.0022908925919784434, "grad_norm": 0.24985, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:32.032027+00:00", "epoch": 0, "step": 1308, "train_loss": 3.8828110694885254, "perplexity": 48.56053041101921, "lr": 0.00229264537896542, "grad_norm": 0.237867, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:32.337223+00:00", "epoch": 0, "step": 1309, "train_loss": 4.085444450378418, "perplexity": 59.46836261480663, "lr": 0.002294398165952397, "grad_norm": 0.219784, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:32.644571+00:00", "epoch": 0, "step": 1310, "train_loss": 3.9686057567596436, "perplexity": 52.910708990475214, "lr": 0.002296150952939373, "grad_norm": 0.201687, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:32.950317+00:00", "epoch": 0, "step": 1311, "train_loss": 3.962294578552246, "perplexity": 52.57783160610527, "lr": 0.00229790373992635, "grad_norm": 0.212408, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:33.255366+00:00", "epoch": 0, "step": 1312, "train_loss": 3.9940507411956787, "perplexity": 54.274295809853356, "lr": 0.0022996565269133265, "grad_norm": 0.219533, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:33.560916+00:00", "epoch": 0, "step": 1313, "train_loss": 4.051023960113525, "perplexity": 57.4562598528053, "lr": 0.0023014093139003032, "grad_norm": 0.270342, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:33.866438+00:00", "epoch": 0, "step": 1314, "train_loss": 3.9414803981781006, "perplexity": 51.49477767598421, "lr": 0.00230316210088728, "grad_norm": 0.280272, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:34.172776+00:00", "epoch": 0, "step": 1315, "train_loss": 3.931980609893799, "perplexity": 51.007904448131804, "lr": 0.0023049148878742567, "grad_norm": 0.283187, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:34.477433+00:00", "epoch": 0, "step": 1316, "train_loss": 4.004512310028076, "perplexity": 54.84507048490758, "lr": 0.002306667674861233, "grad_norm": 0.257526, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:34.784709+00:00", "epoch": 0, "step": 1317, "train_loss": 3.996596574783325, "perplexity": 54.41264516756952, "lr": 0.0023084204618482097, "grad_norm": 0.270168, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:35.090307+00:00", "epoch": 0, "step": 1318, "train_loss": 3.975055456161499, "perplexity": 53.253070034702425, "lr": 0.0023101732488351864, "grad_norm": 0.265737, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:35.395682+00:00", "epoch": 0, "step": 1319, "train_loss": 4.0520172119140625, "perplexity": 57.5133567374539, "lr": 0.002311926035822163, "grad_norm": 0.262724, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:35.702096+00:00", "epoch": 0, "step": 1320, "train_loss": 4.066295623779297, "perplexity": 58.34044684759063, "lr": 0.00231367882280914, "grad_norm": 0.244968, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:36.007696+00:00", "epoch": 0, "step": 1321, "train_loss": 3.9984188079833984, "perplexity": 54.51188809051058, "lr": 0.0023154316097961166, "grad_norm": 0.257916, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:36.312560+00:00", "epoch": 0, "step": 1322, "train_loss": 3.938581705093384, "perplexity": 51.34572625156848, "lr": 0.002317184396783093, "grad_norm": 0.245025, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:36.617056+00:00", "epoch": 0, "step": 1323, "train_loss": 3.9377448558807373, "perplexity": 51.3027755951002, "lr": 0.0023189371837700696, "grad_norm": 0.200313, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:36.924907+00:00", "epoch": 0, "step": 1324, "train_loss": 3.9205868244171143, "perplexity": 50.43002966941399, "lr": 0.0023206899707570463, "grad_norm": 0.19558, "tokens_per_sec": 106442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:37.231618+00:00", "epoch": 0, "step": 1325, "train_loss": 4.100114822387695, "perplexity": 60.34721641104197, "lr": 0.0023224427577440226, "grad_norm": 0.200385, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:37.537375+00:00", "epoch": 0, "step": 1326, "train_loss": 4.1108784675598145, "perplexity": 61.00028080722303, "lr": 0.0023241955447309993, "grad_norm": 0.217725, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:37.843430+00:00", "epoch": 0, "step": 1327, "train_loss": 3.957460403442383, "perplexity": 52.32427452467679, "lr": 0.002325948331717976, "grad_norm": 0.211068, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:38.149238+00:00", "epoch": 0, "step": 1328, "train_loss": 4.000436782836914, "perplexity": 54.622002776861294, "lr": 0.0023277011187049523, "grad_norm": 0.211876, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:38.454976+00:00", "epoch": 0, "step": 1329, "train_loss": 4.015248775482178, "perplexity": 55.43708507472656, "lr": 0.002329453905691929, "grad_norm": 0.21384, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:38.762063+00:00", "epoch": 0, "step": 1330, "train_loss": 3.9732508659362793, "perplexity": 53.15705672345232, "lr": 0.0023312066926789057, "grad_norm": 0.246853, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:39.069625+00:00", "epoch": 0, "step": 1331, "train_loss": 3.9755194187164307, "perplexity": 53.27778319768148, "lr": 0.0023329594796658824, "grad_norm": 0.236222, "tokens_per_sec": 106473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:39.374976+00:00", "epoch": 0, "step": 1332, "train_loss": 3.918407440185547, "perplexity": 50.320242935148386, "lr": 0.002334712266652859, "grad_norm": 0.198101, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:39.680600+00:00", "epoch": 0, "step": 1333, "train_loss": 3.9936301708221436, "perplexity": 54.25147444832191, "lr": 0.0023364650536398354, "grad_norm": 0.216276, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:39.985548+00:00", "epoch": 0, "step": 1334, "train_loss": 3.9997594356536865, "perplexity": 54.585017244575276, "lr": 0.002338217840626812, "grad_norm": 0.232248, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:40.291497+00:00", "epoch": 0, "step": 1335, "train_loss": 3.941030502319336, "perplexity": 51.47161559941162, "lr": 0.002339970627613789, "grad_norm": 0.285298, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:40.597833+00:00", "epoch": 0, "step": 1336, "train_loss": 3.9248781204223633, "perplexity": 50.64690485927994, "lr": 0.0023417234146007656, "grad_norm": 0.358788, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:40.903071+00:00", "epoch": 0, "step": 1337, "train_loss": 3.8877198696136475, "perplexity": 48.79949037228293, "lr": 0.0023434762015877423, "grad_norm": 0.291974, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:41.208494+00:00", "epoch": 0, "step": 1338, "train_loss": 3.9489617347717285, "perplexity": 51.88147213266844, "lr": 0.0023452289885747186, "grad_norm": 0.286727, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:41.514061+00:00", "epoch": 0, "step": 1339, "train_loss": 3.879317283630371, "perplexity": 48.391166349764816, "lr": 0.0023469817755616953, "grad_norm": 0.247925, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:41.819171+00:00", "epoch": 0, "step": 1340, "train_loss": 3.94584321975708, "perplexity": 51.71993099798509, "lr": 0.002348734562548672, "grad_norm": 0.270047, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:42.124454+00:00", "epoch": 0, "step": 1341, "train_loss": 3.8870363235473633, "perplexity": 48.76614507042558, "lr": 0.0023504873495356488, "grad_norm": 0.27588, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:42.429347+00:00", "epoch": 0, "step": 1342, "train_loss": 3.922299385070801, "perplexity": 50.51646814843311, "lr": 0.0023522401365226255, "grad_norm": 0.284621, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:42.735927+00:00", "epoch": 0, "step": 1343, "train_loss": 4.045778751373291, "perplexity": 57.155678770840616, "lr": 0.002353992923509602, "grad_norm": 0.287516, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:43.041264+00:00", "epoch": 0, "step": 1344, "train_loss": 3.9049580097198486, "perplexity": 49.64799513518258, "lr": 0.0023557457104965785, "grad_norm": 0.279267, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:43.346386+00:00", "epoch": 0, "step": 1345, "train_loss": 4.001943111419678, "perplexity": 54.70434346138576, "lr": 0.002357498497483555, "grad_norm": 0.257218, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:43.652147+00:00", "epoch": 0, "step": 1346, "train_loss": 3.9370920658111572, "perplexity": 51.269296581222605, "lr": 0.002359251284470532, "grad_norm": 0.255305, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:43.957810+00:00", "epoch": 0, "step": 1347, "train_loss": 3.966485023498535, "perplexity": 52.798618389194196, "lr": 0.002361004071457508, "grad_norm": 0.250082, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:44.262837+00:00", "epoch": 0, "step": 1348, "train_loss": 4.027536869049072, "perplexity": 56.1225037831385, "lr": 0.002362756858444485, "grad_norm": 0.225932, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:44.567805+00:00", "epoch": 0, "step": 1349, "train_loss": 3.858224868774414, "perplexity": 47.38116888109216, "lr": 0.0023645096454314616, "grad_norm": 0.233539, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:44.872867+00:00", "epoch": 0, "step": 1350, "train_loss": 3.898397445678711, "perplexity": 49.32334240055982, "lr": 0.002366262432418438, "grad_norm": 0.240116, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:45.178984+00:00", "epoch": 0, "step": 1351, "train_loss": 3.977853536605835, "perplexity": 53.40228506904866, "lr": 0.0023680152194054146, "grad_norm": 0.22426, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:45.485476+00:00", "epoch": 0, "step": 1352, "train_loss": 3.9418811798095703, "perplexity": 51.51541997324384, "lr": 0.0023697680063923914, "grad_norm": 0.232064, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:45.792082+00:00", "epoch": 0, "step": 1353, "train_loss": 3.972296714782715, "perplexity": 53.106361045975255, "lr": 0.002371520793379368, "grad_norm": 0.236129, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:46.096750+00:00", "epoch": 0, "step": 1354, "train_loss": 4.032650947570801, "perplexity": 56.41025383535616, "lr": 0.002373273580366345, "grad_norm": 0.305638, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:46.400652+00:00", "epoch": 0, "step": 1355, "train_loss": 4.057402610778809, "perplexity": 57.82392461888965, "lr": 0.002375026367353321, "grad_norm": 0.311192, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:46.705018+00:00", "epoch": 0, "step": 1356, "train_loss": 3.9997823238372803, "perplexity": 54.58626661076924, "lr": 0.002376779154340298, "grad_norm": 0.291738, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:47.010498+00:00", "epoch": 0, "step": 1357, "train_loss": 3.933553695678711, "perplexity": 51.08820740269567, "lr": 0.0023785319413272745, "grad_norm": 0.301114, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:47.315867+00:00", "epoch": 0, "step": 1358, "train_loss": 3.924337387084961, "perplexity": 50.6195257924405, "lr": 0.0023802847283142512, "grad_norm": 0.282808, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:47.621295+00:00", "epoch": 0, "step": 1359, "train_loss": 3.9569242000579834, "perplexity": 52.29622559222871, "lr": 0.002382037515301228, "grad_norm": 0.289031, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:47.927054+00:00", "epoch": 0, "step": 1360, "train_loss": 3.8945133686065674, "perplexity": 49.13213830332163, "lr": 0.0023837903022882047, "grad_norm": 0.228124, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:48.232170+00:00", "epoch": 0, "step": 1361, "train_loss": 4.091189861297607, "perplexity": 59.811016194955364, "lr": 0.002385543089275181, "grad_norm": 0.257225, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:48.537598+00:00", "epoch": 0, "step": 1362, "train_loss": 3.978483200073242, "perplexity": 53.43592112561938, "lr": 0.0023872958762621577, "grad_norm": 0.229585, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:48.843960+00:00", "epoch": 0, "step": 1363, "train_loss": 3.9471871852874756, "perplexity": 51.78948753280375, "lr": 0.0023890486632491344, "grad_norm": 0.23671, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:49.150257+00:00", "epoch": 0, "step": 1364, "train_loss": 4.026566028594971, "perplexity": 56.06804422612837, "lr": 0.002390801450236111, "grad_norm": 0.204764, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:49.455231+00:00", "epoch": 0, "step": 1365, "train_loss": 3.8756461143493652, "perplexity": 48.21383988328724, "lr": 0.002392554237223088, "grad_norm": 0.191188, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:49.760688+00:00", "epoch": 0, "step": 1366, "train_loss": 3.959538221359253, "perplexity": 52.43310786855156, "lr": 0.002394307024210064, "grad_norm": 0.192984, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:50.066641+00:00", "epoch": 0, "step": 1367, "train_loss": 3.984218120574951, "perplexity": 53.74325230281939, "lr": 0.002396059811197041, "grad_norm": 0.192067, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:50.372073+00:00", "epoch": 0, "step": 1368, "train_loss": 3.9087376594543457, "perplexity": 49.836002243503586, "lr": 0.0023978125981840175, "grad_norm": 0.200462, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:50.677273+00:00", "epoch": 0, "step": 1369, "train_loss": 4.05280876159668, "perplexity": 57.558899439003255, "lr": 0.002399565385170994, "grad_norm": 0.207737, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:50.982725+00:00", "epoch": 0, "step": 1370, "train_loss": 3.9119999408721924, "perplexity": 49.998846785501556, "lr": 0.0024013181721579705, "grad_norm": 0.210233, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:51.290259+00:00", "epoch": 0, "step": 1371, "train_loss": 3.8687520027160645, "perplexity": 47.88259143274502, "lr": 0.0024030709591449473, "grad_norm": 0.229529, "tokens_per_sec": 106493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:51.595734+00:00", "epoch": 0, "step": 1372, "train_loss": 3.9835758209228516, "perplexity": 53.70874411404777, "lr": 0.0024048237461319236, "grad_norm": 0.227867, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:51.901427+00:00", "epoch": 0, "step": 1373, "train_loss": 4.001883029937744, "perplexity": 54.70105684209586, "lr": 0.0024065765331189003, "grad_norm": 0.220775, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:52.208639+00:00", "epoch": 0, "step": 1374, "train_loss": 3.8709592819213867, "perplexity": 47.988398410924475, "lr": 0.002408329320105877, "grad_norm": 0.207565, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:52.513869+00:00", "epoch": 0, "step": 1375, "train_loss": 3.9837806224823, "perplexity": 53.71974487504638, "lr": 0.0024100821070928537, "grad_norm": 0.214842, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:52.819362+00:00", "epoch": 0, "step": 1376, "train_loss": 4.008149147033691, "perplexity": 55.044896213373725, "lr": 0.0024118348940798304, "grad_norm": 0.21481, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:53.124937+00:00", "epoch": 0, "step": 1377, "train_loss": 3.9565210342407227, "perplexity": 52.275145791311274, "lr": 0.002413587681066807, "grad_norm": 0.223239, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:53.430783+00:00", "epoch": 0, "step": 1378, "train_loss": 3.9952776432037354, "perplexity": 54.34092591831713, "lr": 0.0024153404680537834, "grad_norm": 0.214865, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:53.736979+00:00", "epoch": 0, "step": 1379, "train_loss": 4.034666538238525, "perplexity": 56.52406847988637, "lr": 0.00241709325504076, "grad_norm": 0.234287, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:54.043273+00:00", "epoch": 0, "step": 1380, "train_loss": 3.9826550483703613, "perplexity": 53.65931333738241, "lr": 0.002418846042027737, "grad_norm": 0.219213, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:54.348407+00:00", "epoch": 0, "step": 1381, "train_loss": 3.9447290897369385, "perplexity": 51.66234035790828, "lr": 0.0024205988290147136, "grad_norm": 0.199106, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:54.653689+00:00", "epoch": 0, "step": 1382, "train_loss": 3.789292812347412, "perplexity": 44.22511376018294, "lr": 0.0024223516160016903, "grad_norm": 0.231359, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:54.958731+00:00", "epoch": 0, "step": 1383, "train_loss": 3.9424965381622314, "perplexity": 51.547130172782666, "lr": 0.002424104402988667, "grad_norm": 0.270956, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:55.265239+00:00", "epoch": 0, "step": 1384, "train_loss": 3.9621217250823975, "perplexity": 52.56874413089847, "lr": 0.0024258571899756433, "grad_norm": 0.313516, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:55.572112+00:00", "epoch": 0, "step": 1385, "train_loss": 4.025820255279541, "perplexity": 56.02624576292272, "lr": 0.00242760997696262, "grad_norm": 0.286903, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:55.877653+00:00", "epoch": 0, "step": 1386, "train_loss": 3.96458101272583, "perplexity": 52.698184894566516, "lr": 0.0024293627639495967, "grad_norm": 0.234416, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:56.182985+00:00", "epoch": 0, "step": 1387, "train_loss": 3.8456504344940186, "perplexity": 46.78910770523589, "lr": 0.002431115550936573, "grad_norm": 0.20789, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:56.488169+00:00", "epoch": 0, "step": 1388, "train_loss": 4.055644989013672, "perplexity": 57.722381293953795, "lr": 0.0024328683379235497, "grad_norm": 0.226123, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:56.794233+00:00", "epoch": 0, "step": 1389, "train_loss": 3.9707579612731934, "perplexity": 53.024706285923735, "lr": 0.0024346211249105265, "grad_norm": 0.243592, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:57.101565+00:00", "epoch": 0, "step": 1390, "train_loss": 4.065396308898926, "perplexity": 58.28800400046399, "lr": 0.002436373911897503, "grad_norm": 0.244252, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:57.407377+00:00", "epoch": 0, "step": 1391, "train_loss": 3.988691806793213, "perplexity": 53.98422135804833, "lr": 0.0024381266988844795, "grad_norm": 0.255066, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:57.713153+00:00", "epoch": 0, "step": 1392, "train_loss": 3.9537887573242188, "perplexity": 52.132510565361464, "lr": 0.002439879485871456, "grad_norm": 0.258468, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:58.018069+00:00", "epoch": 0, "step": 1393, "train_loss": 3.995922803878784, "perplexity": 54.37599585842205, "lr": 0.002441632272858433, "grad_norm": 0.257318, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:58.323909+00:00", "epoch": 0, "step": 1394, "train_loss": 3.965604543685913, "perplexity": 52.75215073148743, "lr": 0.0024433850598454096, "grad_norm": 0.254846, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:58.629967+00:00", "epoch": 0, "step": 1395, "train_loss": 4.022678375244141, "perplexity": 55.85049425954782, "lr": 0.002445137846832386, "grad_norm": 0.244333, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:58.935626+00:00", "epoch": 0, "step": 1396, "train_loss": 3.8856329917907715, "perplexity": 48.69775798650811, "lr": 0.0024468906338193626, "grad_norm": 0.257535, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:59.240142+00:00", "epoch": 0, "step": 1397, "train_loss": 3.890263080596924, "perplexity": 48.92375572169854, "lr": 0.0024486434208063393, "grad_norm": 0.272443, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:59.545538+00:00", "epoch": 0, "step": 1398, "train_loss": 3.9665236473083496, "perplexity": 52.800657712372285, "lr": 0.002450396207793316, "grad_norm": 0.273133, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:04:59.851431+00:00", "epoch": 0, "step": 1399, "train_loss": 3.9865968227386475, "perplexity": 53.87124365966461, "lr": 0.0024521489947802928, "grad_norm": 0.262735, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:00.158179+00:00", "epoch": 0, "step": 1400, "train_loss": 3.902482748031616, "perplexity": 49.52525532417091, "lr": 0.0024539017817672695, "grad_norm": 0.216357, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:00.464472+00:00", "epoch": 0, "step": 1401, "train_loss": 3.9326720237731934, "perplexity": 51.04318421628128, "lr": 0.0024556545687542458, "grad_norm": 0.239406, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:00.769661+00:00", "epoch": 0, "step": 1402, "train_loss": 3.858046531677246, "perplexity": 47.37271981438688, "lr": 0.0024574073557412225, "grad_norm": 0.225508, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:01.074600+00:00", "epoch": 0, "step": 1403, "train_loss": 3.9980056285858154, "perplexity": 54.489369553846295, "lr": 0.002459160142728199, "grad_norm": 0.220266, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:01.380308+00:00", "epoch": 0, "step": 1404, "train_loss": 3.945024251937866, "perplexity": 51.67759137864509, "lr": 0.002460912929715176, "grad_norm": 0.207475, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:01.687012+00:00", "epoch": 0, "step": 1405, "train_loss": 3.982039451599121, "perplexity": 53.62629100260912, "lr": 0.0024626657167021527, "grad_norm": 0.204544, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:01.992488+00:00", "epoch": 0, "step": 1406, "train_loss": 3.944422483444214, "perplexity": 51.64650278733113, "lr": 0.002464418503689129, "grad_norm": 0.211302, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:02.297776+00:00", "epoch": 0, "step": 1407, "train_loss": 3.9199628829956055, "perplexity": 50.398574099252386, "lr": 0.0024661712906761057, "grad_norm": 0.217374, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:02.603822+00:00", "epoch": 0, "step": 1408, "train_loss": 4.112461566925049, "perplexity": 61.096926792962854, "lr": 0.0024679240776630824, "grad_norm": 0.213645, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:02.909707+00:00", "epoch": 0, "step": 1409, "train_loss": 4.007511138916016, "perplexity": 55.00978832350324, "lr": 0.0024696768646500587, "grad_norm": 0.242826, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:03.215725+00:00", "epoch": 0, "step": 1410, "train_loss": 3.9400417804718018, "perplexity": 51.42074963882762, "lr": 0.0024714296516370354, "grad_norm": 0.248389, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:03.521026+00:00", "epoch": 0, "step": 1411, "train_loss": 3.924382209777832, "perplexity": 50.621794746748314, "lr": 0.002473182438624012, "grad_norm": 0.231364, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:03.826631+00:00", "epoch": 0, "step": 1412, "train_loss": 3.986372470855713, "perplexity": 53.85915890038347, "lr": 0.0024749352256109884, "grad_norm": 0.244293, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:04.131381+00:00", "epoch": 0, "step": 1413, "train_loss": 4.020749092102051, "perplexity": 55.74284671717975, "lr": 0.002476688012597965, "grad_norm": 0.231058, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:04.438429+00:00", "epoch": 0, "step": 1414, "train_loss": 3.9839911460876465, "perplexity": 53.73105533993393, "lr": 0.002478440799584942, "grad_norm": 0.242201, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:04.744897+00:00", "epoch": 0, "step": 1415, "train_loss": 3.9121806621551514, "perplexity": 50.007883457774014, "lr": 0.0024801935865719185, "grad_norm": 0.24433, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:05.049938+00:00", "epoch": 0, "step": 1416, "train_loss": 3.92478609085083, "perplexity": 50.64224406079519, "lr": 0.0024819463735588953, "grad_norm": 0.241057, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:05.355107+00:00", "epoch": 0, "step": 1417, "train_loss": 3.9328629970550537, "perplexity": 51.052933031539666, "lr": 0.002483699160545872, "grad_norm": 0.231545, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:05.659951+00:00", "epoch": 0, "step": 1418, "train_loss": 4.0212483406066895, "perplexity": 55.770683198130165, "lr": 0.0024854519475328483, "grad_norm": 0.214511, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:05.965342+00:00", "epoch": 0, "step": 1419, "train_loss": 4.026020526885986, "perplexity": 56.03746735281017, "lr": 0.002487204734519825, "grad_norm": 0.229009, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:06.271739+00:00", "epoch": 0, "step": 1420, "train_loss": 4.068500995635986, "perplexity": 58.46925120572312, "lr": 0.0024889575215068017, "grad_norm": 0.251274, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:06.577308+00:00", "epoch": 0, "step": 1421, "train_loss": 3.8789424896240234, "perplexity": 48.37303302899899, "lr": 0.0024907103084937784, "grad_norm": 0.247999, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:06.882550+00:00", "epoch": 0, "step": 1422, "train_loss": 3.8918092250823975, "perplexity": 48.99945742462075, "lr": 0.002492463095480755, "grad_norm": 0.253939, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:07.188503+00:00", "epoch": 0, "step": 1423, "train_loss": 4.101186275482178, "perplexity": 60.41191027484015, "lr": 0.0024942158824677314, "grad_norm": 0.23111, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:07.494849+00:00", "epoch": 0, "step": 1424, "train_loss": 3.991928815841675, "perplexity": 54.159251905958044, "lr": 0.002495968669454708, "grad_norm": 0.238536, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:07.802359+00:00", "epoch": 0, "step": 1425, "train_loss": 3.879882574081421, "perplexity": 48.41852914725292, "lr": 0.002497721456441685, "grad_norm": 0.204489, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:08.106659+00:00", "epoch": 0, "step": 1426, "train_loss": 3.880911111831665, "perplexity": 48.46835505181152, "lr": 0.0024994742434286616, "grad_norm": 0.203397, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:08.411615+00:00", "epoch": 0, "step": 1427, "train_loss": 4.038509368896484, "perplexity": 56.741698793555706, "lr": 0.0025012270304156383, "grad_norm": 0.204253, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:08.716766+00:00", "epoch": 0, "step": 1428, "train_loss": 3.977628231048584, "perplexity": 53.39025459277017, "lr": 0.0025029798174026146, "grad_norm": 0.208435, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:09.023369+00:00", "epoch": 0, "step": 1429, "train_loss": 3.911820411682129, "perplexity": 49.9898713387355, "lr": 0.0025047326043895913, "grad_norm": 0.211324, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:09.328692+00:00", "epoch": 0, "step": 1430, "train_loss": 3.9850645065307617, "perplexity": 53.78875909222516, "lr": 0.002506485391376568, "grad_norm": 0.191852, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:09.634229+00:00", "epoch": 0, "step": 1431, "train_loss": 3.952042818069458, "perplexity": 52.04156978036063, "lr": 0.0025082381783635443, "grad_norm": 0.205778, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:09.939056+00:00", "epoch": 0, "step": 1432, "train_loss": 3.9511778354644775, "perplexity": 51.99657419077134, "lr": 0.002509990965350521, "grad_norm": 0.205919, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:10.244034+00:00", "epoch": 0, "step": 1433, "train_loss": 3.993635654449463, "perplexity": 54.251771944004986, "lr": 0.0025117437523374977, "grad_norm": 0.204204, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:10.548789+00:00", "epoch": 0, "step": 1434, "train_loss": 3.9620449542999268, "perplexity": 52.564708542187674, "lr": 0.0025134965393244744, "grad_norm": 0.201474, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:10.855856+00:00", "epoch": 0, "step": 1435, "train_loss": 3.888568878173828, "perplexity": 48.84093915003527, "lr": 0.0025152493263114507, "grad_norm": 0.208664, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:11.160820+00:00", "epoch": 0, "step": 1436, "train_loss": 3.9413692951202393, "perplexity": 51.48905676653167, "lr": 0.0025170021132984275, "grad_norm": 0.195619, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:11.465623+00:00", "epoch": 0, "step": 1437, "train_loss": 3.8627195358276367, "perplexity": 47.59461077560082, "lr": 0.002518754900285404, "grad_norm": 0.198235, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:11.770490+00:00", "epoch": 0, "step": 1438, "train_loss": 3.989490032196045, "perplexity": 54.02733013786105, "lr": 0.002520507687272381, "grad_norm": 0.230855, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:12.075084+00:00", "epoch": 0, "step": 1439, "train_loss": 4.016220569610596, "perplexity": 55.49098469392369, "lr": 0.0025222604742593576, "grad_norm": 0.268497, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:12.381464+00:00", "epoch": 0, "step": 1440, "train_loss": 3.866236925125122, "perplexity": 47.76231431657829, "lr": 0.002524013261246334, "grad_norm": 0.290034, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:12.687262+00:00", "epoch": 0, "step": 1441, "train_loss": 4.015480041503906, "perplexity": 55.449907271460084, "lr": 0.0025257660482333106, "grad_norm": 0.248493, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:12.993605+00:00", "epoch": 0, "step": 1442, "train_loss": 3.9551689624786377, "perplexity": 52.204513803346366, "lr": 0.0025275188352202873, "grad_norm": 0.214198, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:13.299333+00:00", "epoch": 0, "step": 1443, "train_loss": 3.893030881881714, "perplexity": 49.059354524346496, "lr": 0.002529271622207264, "grad_norm": 0.200797, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:13.603901+00:00", "epoch": 0, "step": 1444, "train_loss": 3.7924835681915283, "perplexity": 44.36645066623778, "lr": 0.0025310244091942408, "grad_norm": 0.217873, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:13.909129+00:00", "epoch": 0, "step": 1445, "train_loss": 4.02019739151001, "perplexity": 55.71210183740784, "lr": 0.0025327771961812175, "grad_norm": 0.206659, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:14.214017+00:00", "epoch": 0, "step": 1446, "train_loss": 3.9922826290130615, "perplexity": 54.17841755296406, "lr": 0.0025345299831681938, "grad_norm": 0.22072, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:14.519016+00:00", "epoch": 0, "step": 1447, "train_loss": 3.8878440856933594, "perplexity": 48.80555243016365, "lr": 0.0025362827701551705, "grad_norm": 0.247789, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:14.823694+00:00", "epoch": 0, "step": 1448, "train_loss": 3.948158025741577, "perplexity": 51.83979127690159, "lr": 0.002538035557142147, "grad_norm": 0.212647, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:15.128166+00:00", "epoch": 0, "step": 1449, "train_loss": 4.107832908630371, "perplexity": 60.81478347222512, "lr": 0.002539788344129124, "grad_norm": 0.232459, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:15.432845+00:00", "epoch": 0, "step": 1450, "train_loss": 3.9612762928009033, "perplexity": 52.524319599225244, "lr": 0.0025415411311161, "grad_norm": 0.233078, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:15.739049+00:00", "epoch": 0, "step": 1451, "train_loss": 3.958317756652832, "perplexity": 52.369154145519374, "lr": 0.002543293918103077, "grad_norm": 0.227016, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:16.044934+00:00", "epoch": 0, "step": 1452, "train_loss": 3.9366962909698486, "perplexity": 51.24900949832749, "lr": 0.0025450467050900536, "grad_norm": 0.273523, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:16.349425+00:00", "epoch": 0, "step": 1453, "train_loss": 3.968280553817749, "perplexity": 52.89350506978822, "lr": 0.00254679949207703, "grad_norm": 0.317396, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:16.655594+00:00", "epoch": 0, "step": 1454, "train_loss": 4.038663387298584, "perplexity": 56.75043873237474, "lr": 0.0025485522790640066, "grad_norm": 0.321435, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:16.959732+00:00", "epoch": 0, "step": 1455, "train_loss": 3.8080928325653076, "perplexity": 45.06441148004114, "lr": 0.0025503050660509834, "grad_norm": 0.272291, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:17.265502+00:00", "epoch": 0, "step": 1456, "train_loss": 3.863198757171631, "perplexity": 47.61742459494156, "lr": 0.00255205785303796, "grad_norm": 0.281443, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:17.571848+00:00", "epoch": 0, "step": 1457, "train_loss": 3.9073221683502197, "perplexity": 49.765509728196854, "lr": 0.0025538106400249364, "grad_norm": 0.280903, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:17.876993+00:00", "epoch": 0, "step": 1458, "train_loss": 3.787550210952759, "perplexity": 44.148114124587075, "lr": 0.002555563427011913, "grad_norm": 0.263815, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:18.182197+00:00", "epoch": 0, "step": 1459, "train_loss": 3.939927816390991, "perplexity": 51.41488985426927, "lr": 0.00255731621399889, "grad_norm": 0.227801, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:18.487026+00:00", "epoch": 0, "step": 1460, "train_loss": 3.979443073272705, "perplexity": 53.487237458842905, "lr": 0.0025590690009858665, "grad_norm": 0.214565, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:18.792734+00:00", "epoch": 0, "step": 1461, "train_loss": 4.0053300857543945, "perplexity": 54.889939796274646, "lr": 0.0025608217879728432, "grad_norm": 0.228396, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:19.098876+00:00", "epoch": 0, "step": 1462, "train_loss": 3.8584303855895996, "perplexity": 47.390907508711955, "lr": 0.00256257457495982, "grad_norm": 0.214828, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:19.404166+00:00", "epoch": 0, "step": 1463, "train_loss": 3.937636137008667, "perplexity": 51.297198318386535, "lr": 0.0025643273619467962, "grad_norm": 0.184863, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:19.709026+00:00", "epoch": 0, "step": 1464, "train_loss": 3.95924711227417, "perplexity": 52.41784633598506, "lr": 0.002566080148933773, "grad_norm": 0.198242, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:20.013405+00:00", "epoch": 0, "step": 1465, "train_loss": 4.016252517700195, "perplexity": 55.492757553194245, "lr": 0.0025678329359207497, "grad_norm": 0.198506, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:20.318175+00:00", "epoch": 0, "step": 1466, "train_loss": 4.013355731964111, "perplexity": 55.33223953006521, "lr": 0.0025695857229077264, "grad_norm": 0.183914, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:20.624572+00:00", "epoch": 0, "step": 1467, "train_loss": 3.933981418609619, "perplexity": 51.11006367438188, "lr": 0.002571338509894703, "grad_norm": 0.194648, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:20.930654+00:00", "epoch": 0, "step": 1468, "train_loss": 3.9224629402160645, "perplexity": 50.52473105242115, "lr": 0.00257309129688168, "grad_norm": 0.185296, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:21.236552+00:00", "epoch": 0, "step": 1469, "train_loss": 4.057590484619141, "perplexity": 57.83478924222817, "lr": 0.002574844083868656, "grad_norm": 0.191478, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:21.541481+00:00", "epoch": 0, "step": 1470, "train_loss": 3.952038049697876, "perplexity": 52.04132162740985, "lr": 0.002576596870855633, "grad_norm": 0.215055, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:21.846730+00:00", "epoch": 0, "step": 1471, "train_loss": 3.8748645782470703, "perplexity": 48.176173747427896, "lr": 0.0025783496578426096, "grad_norm": 0.216102, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:22.151785+00:00", "epoch": 0, "step": 1472, "train_loss": 3.9208998680114746, "perplexity": 50.44581893840122, "lr": 0.002580102444829586, "grad_norm": 0.207539, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:22.456881+00:00", "epoch": 0, "step": 1473, "train_loss": 3.9295756816864014, "perplexity": 50.88538148843796, "lr": 0.0025818552318165626, "grad_norm": 0.212279, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:22.763550+00:00", "epoch": 0, "step": 1474, "train_loss": 3.94350004196167, "perplexity": 51.5988838769375, "lr": 0.0025836080188035393, "grad_norm": 0.198264, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:23.068399+00:00", "epoch": 0, "step": 1475, "train_loss": 3.8427488803863525, "perplexity": 46.65354334635079, "lr": 0.0025853608057905156, "grad_norm": 0.229468, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:23.374330+00:00", "epoch": 0, "step": 1476, "train_loss": 3.9933249950408936, "perplexity": 54.234920738247354, "lr": 0.0025871135927774923, "grad_norm": 0.271849, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:23.678639+00:00", "epoch": 0, "step": 1477, "train_loss": 3.7959954738616943, "perplexity": 44.522535372858506, "lr": 0.002588866379764469, "grad_norm": 0.238673, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:23.983984+00:00", "epoch": 0, "step": 1478, "train_loss": 3.878563404083252, "perplexity": 48.354698986919125, "lr": 0.0025906191667514457, "grad_norm": 0.282424, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:24.290548+00:00", "epoch": 0, "step": 1479, "train_loss": 3.837470531463623, "perplexity": 46.40793843054074, "lr": 0.0025923719537384224, "grad_norm": 0.269371, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:24.596151+00:00", "epoch": 0, "step": 1480, "train_loss": 3.9783577919006348, "perplexity": 53.42922024458083, "lr": 0.0025941247407253987, "grad_norm": 0.234451, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:24.900694+00:00", "epoch": 0, "step": 1481, "train_loss": 3.972846746444702, "perplexity": 53.13557926073845, "lr": 0.0025958775277123754, "grad_norm": 0.210093, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:25.205025+00:00", "epoch": 0, "step": 1482, "train_loss": 4.033744812011719, "perplexity": 56.47199276698469, "lr": 0.002597630314699352, "grad_norm": 0.227129, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:25.510029+00:00", "epoch": 0, "step": 1483, "train_loss": 3.9483320713043213, "perplexity": 51.848814547754266, "lr": 0.002599383101686329, "grad_norm": 0.242938, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:25.815412+00:00", "epoch": 0, "step": 1484, "train_loss": 3.898644208908081, "perplexity": 49.33551508963817, "lr": 0.0026011358886733056, "grad_norm": 0.228014, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:26.122203+00:00", "epoch": 0, "step": 1485, "train_loss": 3.957991600036621, "perplexity": 52.352076384573195, "lr": 0.0026028886756602823, "grad_norm": 0.223253, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:26.427437+00:00", "epoch": 0, "step": 1486, "train_loss": 3.971245288848877, "perplexity": 53.05055298488243, "lr": 0.0026046414626472586, "grad_norm": 0.221469, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:26.732516+00:00", "epoch": 0, "step": 1487, "train_loss": 3.995821237564087, "perplexity": 54.370473369368895, "lr": 0.0026063942496342353, "grad_norm": 0.227559, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:27.037517+00:00", "epoch": 0, "step": 1488, "train_loss": 4.022830009460449, "perplexity": 55.858963747591176, "lr": 0.002608147036621212, "grad_norm": 0.228021, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:27.344296+00:00", "epoch": 0, "step": 1489, "train_loss": 3.8940939903259277, "perplexity": 49.11153767166651, "lr": 0.0026098998236081888, "grad_norm": 0.255742, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:27.651105+00:00", "epoch": 0, "step": 1490, "train_loss": 3.9420313835144043, "perplexity": 51.523158361332634, "lr": 0.002611652610595165, "grad_norm": 0.213404, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:27.956045+00:00", "epoch": 0, "step": 1491, "train_loss": 3.920586585998535, "perplexity": 50.4300176459594, "lr": 0.0026134053975821418, "grad_norm": 0.231238, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:28.261559+00:00", "epoch": 0, "step": 1492, "train_loss": 3.8604769706726074, "perplexity": 47.487996349487986, "lr": 0.0026151581845691185, "grad_norm": 0.242101, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:28.566196+00:00", "epoch": 0, "step": 1493, "train_loss": 3.8352458477020264, "perplexity": 46.30481019988885, "lr": 0.002616910971556095, "grad_norm": 0.221566, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:28.871396+00:00", "epoch": 0, "step": 1494, "train_loss": 4.019364833831787, "perplexity": 55.66573760238124, "lr": 0.0026186637585430715, "grad_norm": 0.256343, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:29.176647+00:00", "epoch": 0, "step": 1495, "train_loss": 3.913464307785034, "perplexity": 50.07211707661627, "lr": 0.002620416545530048, "grad_norm": 0.27412, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:29.482121+00:00", "epoch": 0, "step": 1496, "train_loss": 3.9776012897491455, "perplexity": 53.388816209310136, "lr": 0.002622169332517025, "grad_norm": 0.269386, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:29.786577+00:00", "epoch": 0, "step": 1497, "train_loss": 3.952155590057373, "perplexity": 52.047438942571304, "lr": 0.002623922119504001, "grad_norm": 0.235827, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:30.092464+00:00", "epoch": 0, "step": 1498, "train_loss": 3.8125686645507812, "perplexity": 45.26656427785089, "lr": 0.002625674906490978, "grad_norm": 0.229357, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:30.397328+00:00", "epoch": 0, "step": 1499, "train_loss": 3.8385093212127686, "perplexity": 46.45617156896894, "lr": 0.0026274276934779546, "grad_norm": 0.20786, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:30.704525+00:00", "epoch": 0, "step": 1500, "train_loss": 3.8510854244232178, "perplexity": 47.0440983425, "lr": 0.0026291804804649314, "grad_norm": 0.219108, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:05:33.777798+00:00", "step": 1500, "epoch": 0, "val_loss": 3.869071125984192, "val_ppl": 47.89787432024226, "eval_train_loss": 3.8510854244232178, "eval_train_ppl": 47.0440983425} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:05:34.725826+00:00", "step": 1500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p8691_epoch_0000_step_0001500.pt", "val_loss": 3.869071125984192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:35.773443+00:00", "epoch": 0, "step": 1501, "train_loss": 3.9834861755371094, "perplexity": 53.703929588767146, "lr": 0.0026291804804649314, "grad_norm": 0.201812, "tokens_per_sec": 6464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:36.079028+00:00", "epoch": 0, "step": 1502, "train_loss": 3.8883216381073, "perplexity": 48.82886520563342, "lr": 0.0026291804804649314, "grad_norm": 0.190493, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:36.385269+00:00", "epoch": 0, "step": 1503, "train_loss": 3.8495028018951416, "perplexity": 46.96970417711906, "lr": 0.0026291804804649314, "grad_norm": 0.181681, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:36.690078+00:00", "epoch": 0, "step": 1504, "train_loss": 3.882824420928955, "perplexity": 48.561178768376465, "lr": 0.0026291804804649314, "grad_norm": 0.200273, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:36.995050+00:00", "epoch": 0, "step": 1505, "train_loss": 3.890984535217285, "perplexity": 48.95906472670027, "lr": 0.0026291804804649314, "grad_norm": 0.201854, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:37.299399+00:00", "epoch": 0, "step": 1506, "train_loss": 3.967069387435913, "perplexity": 52.829481014348545, "lr": 0.0026291804804649314, "grad_norm": 0.188032, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:37.604149+00:00", "epoch": 0, "step": 1507, "train_loss": 3.9176692962646484, "perplexity": 50.28311305900972, "lr": 0.0026291804804649314, "grad_norm": 0.197787, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:37.909068+00:00", "epoch": 0, "step": 1508, "train_loss": 3.882277250289917, "perplexity": 48.53461478533588, "lr": 0.0026291804804649314, "grad_norm": 0.203037, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:38.215907+00:00", "epoch": 0, "step": 1509, "train_loss": 3.952033042907715, "perplexity": 52.041061068085035, "lr": 0.0026291804804649314, "grad_norm": 0.209002, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:38.522100+00:00", "epoch": 0, "step": 1510, "train_loss": 4.050751686096191, "perplexity": 57.44061813562741, "lr": 0.0026291804804649314, "grad_norm": 0.19711, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:38.826168+00:00", "epoch": 0, "step": 1511, "train_loss": 3.948021173477173, "perplexity": 51.83269736949882, "lr": 0.0026291804804649314, "grad_norm": 0.190531, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:39.130936+00:00", "epoch": 0, "step": 1512, "train_loss": 3.901711940765381, "perplexity": 49.48709560628815, "lr": 0.0026291804804649314, "grad_norm": 0.182566, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:39.437490+00:00", "epoch": 0, "step": 1513, "train_loss": 3.9351253509521484, "perplexity": 51.16856358283707, "lr": 0.0026291804804649314, "grad_norm": 0.194244, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:39.743093+00:00", "epoch": 0, "step": 1514, "train_loss": 3.9378230571746826, "perplexity": 51.306787695408396, "lr": 0.0026291804804649314, "grad_norm": 0.231433, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:40.159357+00:00", "epoch": 0, "step": 1515, "train_loss": 3.9186694622039795, "perplexity": 50.333429674302835, "lr": 0.0026291804804649314, "grad_norm": 0.25798, "tokens_per_sec": 78719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:40.464027+00:00", "epoch": 0, "step": 1516, "train_loss": 3.907301664352417, "perplexity": 49.764489346755724, "lr": 0.0026291804804649314, "grad_norm": 0.217637, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:40.769248+00:00", "epoch": 0, "step": 1517, "train_loss": 4.002160549163818, "perplexity": 54.71623954370448, "lr": 0.0026291804804649314, "grad_norm": 0.221095, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:41.074336+00:00", "epoch": 0, "step": 1518, "train_loss": 3.857964515686035, "perplexity": 47.36883465313981, "lr": 0.0026291804804649314, "grad_norm": 0.249575, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:41.378320+00:00", "epoch": 0, "step": 1519, "train_loss": 4.085498332977295, "perplexity": 59.4715670110651, "lr": 0.0026291804804649314, "grad_norm": 0.217309, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:41.682967+00:00", "epoch": 0, "step": 1520, "train_loss": 3.9699795246124268, "perplexity": 52.98344597197841, "lr": 0.0026291804804649314, "grad_norm": 0.210654, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:41.987710+00:00", "epoch": 0, "step": 1521, "train_loss": 3.8682844638824463, "perplexity": 47.86020969436958, "lr": 0.0026291804804649314, "grad_norm": 0.212249, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:42.293455+00:00", "epoch": 0, "step": 1522, "train_loss": 4.053595542907715, "perplexity": 57.60420352523276, "lr": 0.0026291804804649314, "grad_norm": 0.213549, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:42.598980+00:00", "epoch": 0, "step": 1523, "train_loss": 3.913308620452881, "perplexity": 50.0643220890995, "lr": 0.0026291804804649314, "grad_norm": 0.201128, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:42.903243+00:00", "epoch": 0, "step": 1524, "train_loss": 3.891087532043457, "perplexity": 48.96410761467573, "lr": 0.0026291804804649314, "grad_norm": 0.217382, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:43.207262+00:00", "epoch": 0, "step": 1525, "train_loss": 3.923255205154419, "perplexity": 50.56477588631808, "lr": 0.0026291804804649314, "grad_norm": 0.224166, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:43.511766+00:00", "epoch": 0, "step": 1526, "train_loss": 3.8948614597320557, "perplexity": 49.14924374159419, "lr": 0.0026291804804649314, "grad_norm": 0.226118, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:43.818415+00:00", "epoch": 0, "step": 1527, "train_loss": 3.9412832260131836, "perplexity": 51.48462534009983, "lr": 0.0026291804804649314, "grad_norm": 0.228552, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:44.125854+00:00", "epoch": 0, "step": 1528, "train_loss": 3.879746675491333, "perplexity": 48.41194958449445, "lr": 0.0026291804804649314, "grad_norm": 0.227438, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:44.430745+00:00", "epoch": 0, "step": 1529, "train_loss": 3.7905824184417725, "perplexity": 44.282183527262255, "lr": 0.0026291804804649314, "grad_norm": 0.22654, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:44.734033+00:00", "epoch": 0, "step": 1530, "train_loss": 3.8880860805511475, "perplexity": 48.817364552061996, "lr": 0.0026291804804649314, "grad_norm": 0.197726, "tokens_per_sec": 108097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:45.038402+00:00", "epoch": 0, "step": 1531, "train_loss": 3.940885543823242, "perplexity": 51.46415489218233, "lr": 0.0026291804804649314, "grad_norm": 0.193213, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:45.343352+00:00", "epoch": 0, "step": 1532, "train_loss": 3.9222958087921143, "perplexity": 50.51628748778781, "lr": 0.0026291804804649314, "grad_norm": 0.19683, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:45.649018+00:00", "epoch": 0, "step": 1533, "train_loss": 3.9036030769348145, "perplexity": 49.5807709912454, "lr": 0.0026291804804649314, "grad_norm": 0.200831, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:45.954126+00:00", "epoch": 0, "step": 1534, "train_loss": 3.861276865005493, "perplexity": 47.525996924844314, "lr": 0.0026291804804649314, "grad_norm": 0.199044, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:46.258463+00:00", "epoch": 0, "step": 1535, "train_loss": 3.7871670722961426, "perplexity": 44.13120251540173, "lr": 0.0026291804804649314, "grad_norm": 0.182278, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:46.564056+00:00", "epoch": 0, "step": 1536, "train_loss": 3.9613492488861084, "perplexity": 52.528151707747384, "lr": 0.0026291804804649314, "grad_norm": 0.168513, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:46.868187+00:00", "epoch": 0, "step": 1537, "train_loss": 3.914667844772339, "perplexity": 50.1324170008665, "lr": 0.0026291804804649314, "grad_norm": 0.187268, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:47.173629+00:00", "epoch": 0, "step": 1538, "train_loss": 3.97583270072937, "perplexity": 53.29447678361017, "lr": 0.0026291804804649314, "grad_norm": 0.194728, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:47.479453+00:00", "epoch": 0, "step": 1539, "train_loss": 4.007593154907227, "perplexity": 55.014300190839, "lr": 0.0026291804804649314, "grad_norm": 0.222264, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:47.784482+00:00", "epoch": 0, "step": 1540, "train_loss": 3.972905158996582, "perplexity": 53.13868313617045, "lr": 0.0026291804804649314, "grad_norm": 0.234739, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:48.089047+00:00", "epoch": 0, "step": 1541, "train_loss": 3.862281084060669, "perplexity": 47.57374740853242, "lr": 0.0026291804804649314, "grad_norm": 0.240266, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:48.393680+00:00", "epoch": 0, "step": 1542, "train_loss": 3.8103933334350586, "perplexity": 45.16820153662947, "lr": 0.0026291804804649314, "grad_norm": 0.208319, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:48.699565+00:00", "epoch": 0, "step": 1543, "train_loss": 3.928729295730591, "perplexity": 50.842331037408655, "lr": 0.0026291804804649314, "grad_norm": 0.201836, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:49.004474+00:00", "epoch": 0, "step": 1544, "train_loss": 3.986490249633789, "perplexity": 53.86550273988447, "lr": 0.0026291804804649314, "grad_norm": 0.201285, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:49.309825+00:00", "epoch": 0, "step": 1545, "train_loss": 3.870858907699585, "perplexity": 47.98358185451156, "lr": 0.0026291804804649314, "grad_norm": 0.226432, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:49.614034+00:00", "epoch": 0, "step": 1546, "train_loss": 3.9765474796295166, "perplexity": 53.33258416866438, "lr": 0.0026291804804649314, "grad_norm": 0.218561, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:49.918677+00:00", "epoch": 0, "step": 1547, "train_loss": 3.9500601291656494, "perplexity": 51.93848975899538, "lr": 0.0026291804804649314, "grad_norm": 0.210923, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:50.223611+00:00", "epoch": 0, "step": 1548, "train_loss": 3.9013211727142334, "perplexity": 49.46776140822061, "lr": 0.0026291804804649314, "grad_norm": 0.196871, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:50.529760+00:00", "epoch": 0, "step": 1549, "train_loss": 4.013501167297363, "perplexity": 55.34028737796737, "lr": 0.0026291804804649314, "grad_norm": 0.193088, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:50.835332+00:00", "epoch": 0, "step": 1550, "train_loss": 3.9221856594085693, "perplexity": 50.51072345630499, "lr": 0.0026291804804649314, "grad_norm": 0.193114, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:51.139852+00:00", "epoch": 0, "step": 1551, "train_loss": 3.932792901992798, "perplexity": 51.04935459843718, "lr": 0.0026291804804649314, "grad_norm": 0.193952, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:51.445312+00:00", "epoch": 0, "step": 1552, "train_loss": 3.9765443801879883, "perplexity": 53.332418867694365, "lr": 0.0026291804804649314, "grad_norm": 0.188538, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:51.749462+00:00", "epoch": 0, "step": 1553, "train_loss": 3.74534010887146, "perplexity": 42.32339933484003, "lr": 0.0026291804804649314, "grad_norm": 0.185151, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:52.054938+00:00", "epoch": 0, "step": 1554, "train_loss": 3.96695876121521, "perplexity": 52.82363701177814, "lr": 0.0026291804804649314, "grad_norm": 0.180064, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:52.359316+00:00", "epoch": 0, "step": 1555, "train_loss": 3.9386682510375977, "perplexity": 51.350170208228704, "lr": 0.0026291804804649314, "grad_norm": 0.184231, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:52.664824+00:00", "epoch": 0, "step": 1556, "train_loss": 3.9104232788085938, "perplexity": 49.92007761305162, "lr": 0.0026291804804649314, "grad_norm": 0.193236, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:52.969238+00:00", "epoch": 0, "step": 1557, "train_loss": 4.021927833557129, "perplexity": 55.808591862080874, "lr": 0.0026291804804649314, "grad_norm": 0.213772, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:53.273581+00:00", "epoch": 0, "step": 1558, "train_loss": 3.916548013687134, "perplexity": 50.22676307842314, "lr": 0.0026291804804649314, "grad_norm": 0.389817, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:53.578110+00:00", "epoch": 0, "step": 1559, "train_loss": 3.940974235534668, "perplexity": 51.468719538576956, "lr": 0.0026291804804649314, "grad_norm": 0.214179, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:53.884103+00:00", "epoch": 0, "step": 1560, "train_loss": 3.944917917251587, "perplexity": 51.672096550328774, "lr": 0.0026291804804649314, "grad_norm": 0.275979, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:54.189133+00:00", "epoch": 0, "step": 1561, "train_loss": 3.929682493209839, "perplexity": 50.89081692383384, "lr": 0.0026291804804649314, "grad_norm": 0.288385, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:54.493721+00:00", "epoch": 0, "step": 1562, "train_loss": 3.808058261871338, "perplexity": 45.0628535989916, "lr": 0.0026291804804649314, "grad_norm": 0.24046, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:54.798667+00:00", "epoch": 0, "step": 1563, "train_loss": 3.9538488388061523, "perplexity": 52.135642857948596, "lr": 0.0026291804804649314, "grad_norm": 0.255238, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:55.103588+00:00", "epoch": 0, "step": 1564, "train_loss": 3.975111722946167, "perplexity": 53.25606649802686, "lr": 0.0026291804804649314, "grad_norm": 0.278816, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:55.408740+00:00", "epoch": 0, "step": 1565, "train_loss": 3.861473560333252, "perplexity": 47.53534598581467, "lr": 0.0026291804804649314, "grad_norm": 0.286346, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:55.714367+00:00", "epoch": 0, "step": 1566, "train_loss": 3.9242753982543945, "perplexity": 50.61638804448646, "lr": 0.0026291804804649314, "grad_norm": 0.251881, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:56.019332+00:00", "epoch": 0, "step": 1567, "train_loss": 3.9453487396240234, "perplexity": 51.69436284161718, "lr": 0.0026291804804649314, "grad_norm": 0.222846, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:56.324639+00:00", "epoch": 0, "step": 1568, "train_loss": 4.001287937164307, "perplexity": 54.66851432233928, "lr": 0.0026291804804649314, "grad_norm": 0.208871, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:56.631132+00:00", "epoch": 0, "step": 1569, "train_loss": 3.8024215698242188, "perplexity": 44.80956270251634, "lr": 0.0026291804804649314, "grad_norm": 0.19216, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:56.937343+00:00", "epoch": 0, "step": 1570, "train_loss": 3.9633634090423584, "perplexity": 52.63405843875363, "lr": 0.0026291804804649314, "grad_norm": 0.187797, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:57.242926+00:00", "epoch": 0, "step": 1571, "train_loss": 3.914001703262329, "perplexity": 50.09903283742813, "lr": 0.0026291804804649314, "grad_norm": 0.185829, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:57.547355+00:00", "epoch": 0, "step": 1572, "train_loss": 3.8997349739074707, "perplexity": 49.3893579023142, "lr": 0.0026291804804649314, "grad_norm": 0.199005, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:57.851346+00:00", "epoch": 0, "step": 1573, "train_loss": 3.9502146244049072, "perplexity": 51.946514628283474, "lr": 0.0026291804804649314, "grad_norm": 0.187733, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:58.155928+00:00", "epoch": 0, "step": 1574, "train_loss": 3.943899154663086, "perplexity": 51.619481757036034, "lr": 0.0026291804804649314, "grad_norm": 0.177868, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:58.462394+00:00", "epoch": 0, "step": 1575, "train_loss": 3.921719551086426, "perplexity": 50.487185473795435, "lr": 0.0026291804804649314, "grad_norm": 0.174954, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:58.768386+00:00", "epoch": 0, "step": 1576, "train_loss": 3.8630223274230957, "perplexity": 47.609024205755446, "lr": 0.0026291804804649314, "grad_norm": 0.179689, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:59.073963+00:00", "epoch": 0, "step": 1577, "train_loss": 3.985605478286743, "perplexity": 53.81786516375483, "lr": 0.0026291804804649314, "grad_norm": 0.203298, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:59.379175+00:00", "epoch": 0, "step": 1578, "train_loss": 3.8318352699279785, "perplexity": 46.14715304720288, "lr": 0.0026291804804649314, "grad_norm": 0.221649, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:59.685691+00:00", "epoch": 0, "step": 1579, "train_loss": 3.9421193599700928, "perplexity": 51.527691385587914, "lr": 0.0026291804804649314, "grad_norm": 0.187386, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:05:59.991315+00:00", "epoch": 0, "step": 1580, "train_loss": 3.9254491329193115, "perplexity": 50.675833133303314, "lr": 0.0026291804804649314, "grad_norm": 0.195113, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:00.297667+00:00", "epoch": 0, "step": 1581, "train_loss": 4.044712066650391, "perplexity": 57.0947441862417, "lr": 0.0026291804804649314, "grad_norm": 0.203332, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:00.604233+00:00", "epoch": 0, "step": 1582, "train_loss": 3.9293715953826904, "perplexity": 50.87499753866407, "lr": 0.0026291804804649314, "grad_norm": 0.190222, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:00.909551+00:00", "epoch": 0, "step": 1583, "train_loss": 3.911694049835205, "perplexity": 49.98355492535089, "lr": 0.0026291804804649314, "grad_norm": 0.209655, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:01.215110+00:00", "epoch": 0, "step": 1584, "train_loss": 4.041849613189697, "perplexity": 56.931546822020074, "lr": 0.0026291804804649314, "grad_norm": 0.199167, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:01.520726+00:00", "epoch": 0, "step": 1585, "train_loss": 3.859121799468994, "perplexity": 47.42368557021811, "lr": 0.0026291804804649314, "grad_norm": 0.218251, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:01.827185+00:00", "epoch": 0, "step": 1586, "train_loss": 3.7964868545532227, "perplexity": 44.54441826305316, "lr": 0.0026291804804649314, "grad_norm": 0.224776, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:02.133045+00:00", "epoch": 0, "step": 1587, "train_loss": 3.8491148948669434, "perplexity": 46.95148783210892, "lr": 0.0026291804804649314, "grad_norm": 0.217733, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:02.438751+00:00", "epoch": 0, "step": 1588, "train_loss": 3.923161745071411, "perplexity": 50.56005031899587, "lr": 0.0026291804804649314, "grad_norm": 0.221056, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:02.744797+00:00", "epoch": 0, "step": 1589, "train_loss": 3.9461417198181152, "perplexity": 51.73537170495605, "lr": 0.0026291804804649314, "grad_norm": 0.232622, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:03.050219+00:00", "epoch": 0, "step": 1590, "train_loss": 3.9091310501098633, "perplexity": 49.85561111781561, "lr": 0.0026291804804649314, "grad_norm": 0.213896, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:03.356955+00:00", "epoch": 0, "step": 1591, "train_loss": 3.816347122192383, "perplexity": 45.43792561045049, "lr": 0.0026291804804649314, "grad_norm": 0.218686, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:03.663086+00:00", "epoch": 0, "step": 1592, "train_loss": 3.9055655002593994, "perplexity": 49.67816498555638, "lr": 0.0026291804804649314, "grad_norm": 0.201276, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:03.968908+00:00", "epoch": 0, "step": 1593, "train_loss": 3.972102642059326, "perplexity": 53.09605554989799, "lr": 0.0026291804804649314, "grad_norm": 0.182711, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:04.274273+00:00", "epoch": 0, "step": 1594, "train_loss": 3.9661872386932373, "perplexity": 52.78289810364443, "lr": 0.0026291804804649314, "grad_norm": 0.184239, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:04.579882+00:00", "epoch": 0, "step": 1595, "train_loss": 3.90262508392334, "perplexity": 49.53230504725272, "lr": 0.0026291804804649314, "grad_norm": 0.173697, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:04.886076+00:00", "epoch": 0, "step": 1596, "train_loss": 3.800792932510376, "perplexity": 44.736643572208216, "lr": 0.0026291804804649314, "grad_norm": 0.172122, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:05.192779+00:00", "epoch": 0, "step": 1597, "train_loss": 3.8704659938812256, "perplexity": 47.96473214554461, "lr": 0.0026291804804649314, "grad_norm": 0.180228, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:05.498696+00:00", "epoch": 0, "step": 1598, "train_loss": 3.907294988632202, "perplexity": 49.76415713405709, "lr": 0.0026291804804649314, "grad_norm": 0.177184, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:05.804597+00:00", "epoch": 0, "step": 1599, "train_loss": 3.878584861755371, "perplexity": 48.355736577327505, "lr": 0.0026291804804649314, "grad_norm": 0.172804, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:06.110764+00:00", "epoch": 0, "step": 1600, "train_loss": 3.838428020477295, "perplexity": 46.45239480158216, "lr": 0.0026291804804649314, "grad_norm": 0.173524, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:06.416218+00:00", "epoch": 0, "step": 1601, "train_loss": 3.932807207107544, "perplexity": 51.05008487053573, "lr": 0.0026291804804649314, "grad_norm": 0.200432, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:06.722326+00:00", "epoch": 0, "step": 1602, "train_loss": 3.8701491355895996, "perplexity": 47.9495365300142, "lr": 0.0026291804804649314, "grad_norm": 0.197241, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:07.029418+00:00", "epoch": 0, "step": 1603, "train_loss": 3.944715738296509, "perplexity": 51.661650595853196, "lr": 0.0026291804804649314, "grad_norm": 0.198458, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:07.334710+00:00", "epoch": 0, "step": 1604, "train_loss": 3.8366096019744873, "perplexity": 46.368001661653665, "lr": 0.0026291804804649314, "grad_norm": 0.223314, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:07.639740+00:00", "epoch": 0, "step": 1605, "train_loss": 3.95131254196167, "perplexity": 52.00357893892854, "lr": 0.0026291804804649314, "grad_norm": 0.227926, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:07.945160+00:00", "epoch": 0, "step": 1606, "train_loss": 3.926046371459961, "perplexity": 50.706107733609194, "lr": 0.0026291804804649314, "grad_norm": 0.251639, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:08.251602+00:00", "epoch": 0, "step": 1607, "train_loss": 3.9156577587127686, "perplexity": 50.18206835055013, "lr": 0.0026291804804649314, "grad_norm": 0.235305, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:08.558265+00:00", "epoch": 0, "step": 1608, "train_loss": 3.9543535709381104, "perplexity": 52.16196403413219, "lr": 0.0026291804804649314, "grad_norm": 0.220269, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:08.863834+00:00", "epoch": 0, "step": 1609, "train_loss": 3.9305880069732666, "perplexity": 50.936920129383815, "lr": 0.0026291804804649314, "grad_norm": 0.226175, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:09.169731+00:00", "epoch": 0, "step": 1610, "train_loss": 3.9703450202941895, "perplexity": 53.002814732069545, "lr": 0.0026291804804649314, "grad_norm": 0.196183, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:09.474939+00:00", "epoch": 0, "step": 1611, "train_loss": 3.8840129375457764, "perplexity": 48.618928847941696, "lr": 0.0026291804804649314, "grad_norm": 0.19907, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:09.781793+00:00", "epoch": 0, "step": 1612, "train_loss": 3.7454745769500732, "perplexity": 42.32909086368487, "lr": 0.0026291804804649314, "grad_norm": 0.206335, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:10.088215+00:00", "epoch": 0, "step": 1613, "train_loss": 3.887988328933716, "perplexity": 48.812592808944885, "lr": 0.0026291804804649314, "grad_norm": 0.207005, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:10.394370+00:00", "epoch": 0, "step": 1614, "train_loss": 3.858704090118408, "perplexity": 47.40388039000864, "lr": 0.0026291804804649314, "grad_norm": 0.19061, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:10.701900+00:00", "epoch": 0, "step": 1615, "train_loss": 3.926757574081421, "perplexity": 50.74218287720112, "lr": 0.0026291804804649314, "grad_norm": 0.183706, "tokens_per_sec": 106553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:11.007379+00:00", "epoch": 0, "step": 1616, "train_loss": 3.8321073055267334, "perplexity": 46.159708423290105, "lr": 0.0026291804804649314, "grad_norm": 0.200308, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:11.313197+00:00", "epoch": 0, "step": 1617, "train_loss": 3.9104154109954834, "perplexity": 49.91968485275559, "lr": 0.0026291804804649314, "grad_norm": 0.205652, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:11.619611+00:00", "epoch": 0, "step": 1618, "train_loss": 3.95564866065979, "perplexity": 52.22956222102565, "lr": 0.0026291804804649314, "grad_norm": 0.192823, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:11.925987+00:00", "epoch": 0, "step": 1619, "train_loss": 3.9053874015808105, "perplexity": 49.669318157845275, "lr": 0.0026291804804649314, "grad_norm": 0.206074, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:12.231938+00:00", "epoch": 0, "step": 1620, "train_loss": 3.826436758041382, "perplexity": 45.8986983392041, "lr": 0.0026291804804649314, "grad_norm": 0.216354, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:12.538641+00:00", "epoch": 0, "step": 1621, "train_loss": 3.9952845573425293, "perplexity": 54.34130164032002, "lr": 0.0026291804804649314, "grad_norm": 0.198351, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:12.847208+00:00", "epoch": 0, "step": 1622, "train_loss": 3.967468738555908, "perplexity": 52.850582739978826, "lr": 0.0026291804804649314, "grad_norm": 0.230909, "tokens_per_sec": 106190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:13.152942+00:00", "epoch": 0, "step": 1623, "train_loss": 3.887925386428833, "perplexity": 48.809520518773496, "lr": 0.0026291804804649314, "grad_norm": 0.215838, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:13.459265+00:00", "epoch": 0, "step": 1624, "train_loss": 3.7839014530181885, "perplexity": 43.98732186728239, "lr": 0.0026291804804649314, "grad_norm": 0.184155, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:13.765920+00:00", "epoch": 0, "step": 1625, "train_loss": 3.772479295730591, "perplexity": 43.487750282264635, "lr": 0.0026291804804649314, "grad_norm": 0.195518, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:14.072270+00:00", "epoch": 0, "step": 1626, "train_loss": 3.99971866607666, "perplexity": 54.582791881874094, "lr": 0.0026291804804649314, "grad_norm": 0.192349, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:14.378159+00:00", "epoch": 0, "step": 1627, "train_loss": 3.9807896614074707, "perplexity": 53.55931125413421, "lr": 0.0026291804804649314, "grad_norm": 0.201041, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:14.682959+00:00", "epoch": 0, "step": 1628, "train_loss": 3.8186278343200684, "perplexity": 45.54167470438332, "lr": 0.0026291804804649314, "grad_norm": 0.192198, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:14.991645+00:00", "epoch": 0, "step": 1629, "train_loss": 3.923459529876709, "perplexity": 50.5751085756847, "lr": 0.0026291804804649314, "grad_norm": 0.181752, "tokens_per_sec": 106153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:15.297361+00:00", "epoch": 0, "step": 1630, "train_loss": 3.8477089405059814, "perplexity": 46.88552256598067, "lr": 0.0026291804804649314, "grad_norm": 0.196246, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:15.603168+00:00", "epoch": 0, "step": 1631, "train_loss": 3.869169235229492, "perplexity": 47.90257377506952, "lr": 0.0026291804804649314, "grad_norm": 0.206435, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:15.909851+00:00", "epoch": 0, "step": 1632, "train_loss": 3.812819480895996, "perplexity": 45.277919296016044, "lr": 0.0026291804804649314, "grad_norm": 0.203343, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:16.215957+00:00", "epoch": 0, "step": 1633, "train_loss": 3.840620517730713, "perplexity": 46.554353280598676, "lr": 0.0026291804804649314, "grad_norm": 0.192878, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:16.522103+00:00", "epoch": 0, "step": 1634, "train_loss": 3.964479684829712, "perplexity": 52.69284536888792, "lr": 0.0026291804804649314, "grad_norm": 0.187439, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:16.827372+00:00", "epoch": 0, "step": 1635, "train_loss": 4.004323482513428, "perplexity": 54.83471520426892, "lr": 0.0026291804804649314, "grad_norm": 0.189218, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:17.135982+00:00", "epoch": 0, "step": 1636, "train_loss": 3.9512698650360107, "perplexity": 52.00135963341305, "lr": 0.0026291804804649314, "grad_norm": 0.199083, "tokens_per_sec": 106180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:17.442396+00:00", "epoch": 0, "step": 1637, "train_loss": 3.9788095951080322, "perplexity": 53.45336519162762, "lr": 0.0026291804804649314, "grad_norm": 0.207198, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:17.749942+00:00", "epoch": 0, "step": 1638, "train_loss": 3.9268100261688232, "perplexity": 50.74484448041509, "lr": 0.0026291804804649314, "grad_norm": 0.20447, "tokens_per_sec": 106548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:18.055431+00:00", "epoch": 0, "step": 1639, "train_loss": 3.92343807220459, "perplexity": 50.57402336323061, "lr": 0.0026291804804649314, "grad_norm": 0.189467, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:18.361604+00:00", "epoch": 0, "step": 1640, "train_loss": 3.82271409034729, "perplexity": 45.72815038134044, "lr": 0.0026291804804649314, "grad_norm": 0.1857, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:18.667864+00:00", "epoch": 0, "step": 1641, "train_loss": 3.9032833576202393, "perplexity": 49.56492159494205, "lr": 0.0026291804804649314, "grad_norm": 0.184705, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:18.973601+00:00", "epoch": 0, "step": 1642, "train_loss": 3.891936779022217, "perplexity": 49.005707897091995, "lr": 0.0026291804804649314, "grad_norm": 0.223684, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:19.281791+00:00", "epoch": 0, "step": 1643, "train_loss": 3.93723201751709, "perplexity": 51.27647230885706, "lr": 0.0026291804804649314, "grad_norm": 0.223677, "tokens_per_sec": 106329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:19.588320+00:00", "epoch": 0, "step": 1644, "train_loss": 3.901637315750122, "perplexity": 49.48340276881416, "lr": 0.0026291804804649314, "grad_norm": 0.224681, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:19.893433+00:00", "epoch": 0, "step": 1645, "train_loss": 3.8852577209472656, "perplexity": 48.67948656637172, "lr": 0.0026291804804649314, "grad_norm": 0.225746, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:20.199679+00:00", "epoch": 0, "step": 1646, "train_loss": 3.835942268371582, "perplexity": 46.33706905837602, "lr": 0.0026291804804649314, "grad_norm": 0.225042, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:20.504754+00:00", "epoch": 0, "step": 1647, "train_loss": 3.8192591667175293, "perplexity": 45.57043571698304, "lr": 0.0026291804804649314, "grad_norm": 0.199156, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:20.810774+00:00", "epoch": 0, "step": 1648, "train_loss": 3.8834896087646484, "perplexity": 48.59349181971299, "lr": 0.0026291804804649314, "grad_norm": 0.196133, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:21.117056+00:00", "epoch": 0, "step": 1649, "train_loss": 3.86669659614563, "perplexity": 47.78427431514281, "lr": 0.0026291804804649314, "grad_norm": 0.188154, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:21.425220+00:00", "epoch": 0, "step": 1650, "train_loss": 3.9288089275360107, "perplexity": 50.84637986522652, "lr": 0.0026291804804649314, "grad_norm": 0.187775, "tokens_per_sec": 106402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:21.731388+00:00", "epoch": 0, "step": 1651, "train_loss": 3.797809362411499, "perplexity": 44.60336757812261, "lr": 0.0026291804804649314, "grad_norm": 0.175537, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:22.037131+00:00", "epoch": 0, "step": 1652, "train_loss": 3.905362129211426, "perplexity": 49.66806291235128, "lr": 0.0026291804804649314, "grad_norm": 0.184623, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:22.342698+00:00", "epoch": 0, "step": 1653, "train_loss": 3.835003614425659, "perplexity": 46.29359499240564, "lr": 0.0026291804804649314, "grad_norm": 0.177364, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:22.648989+00:00", "epoch": 0, "step": 1654, "train_loss": 3.84844970703125, "perplexity": 46.9202666586589, "lr": 0.0026291804804649314, "grad_norm": 0.162619, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:22.955725+00:00", "epoch": 0, "step": 1655, "train_loss": 3.8688549995422363, "perplexity": 47.88752344167776, "lr": 0.0026291804804649314, "grad_norm": 0.170488, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:23.262251+00:00", "epoch": 0, "step": 1656, "train_loss": 3.823824882507324, "perplexity": 45.77897307377538, "lr": 0.0026291804804649314, "grad_norm": 0.174957, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:23.567496+00:00", "epoch": 0, "step": 1657, "train_loss": 3.971832752227783, "perplexity": 53.081727398008255, "lr": 0.0026291804804649314, "grad_norm": 0.192538, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:23.872550+00:00", "epoch": 0, "step": 1658, "train_loss": 3.814401626586914, "perplexity": 45.34961226032618, "lr": 0.0026291804804649314, "grad_norm": 0.183807, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:24.179915+00:00", "epoch": 0, "step": 1659, "train_loss": 3.9093258380889893, "perplexity": 49.86532333743445, "lr": 0.0026291804804649314, "grad_norm": 0.182737, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:24.486594+00:00", "epoch": 0, "step": 1660, "train_loss": 3.8313779830932617, "perplexity": 46.126055385863765, "lr": 0.0026291804804649314, "grad_norm": 0.177883, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:24.792799+00:00", "epoch": 0, "step": 1661, "train_loss": 3.8497533798217773, "perplexity": 46.981475222926996, "lr": 0.0026291804804649314, "grad_norm": 0.180572, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:25.099177+00:00", "epoch": 0, "step": 1662, "train_loss": 3.882760524749756, "perplexity": 48.558075993724536, "lr": 0.0026291804804649314, "grad_norm": 0.204345, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:25.405865+00:00", "epoch": 0, "step": 1663, "train_loss": 3.7946815490722656, "perplexity": 44.464074524933096, "lr": 0.0026291804804649314, "grad_norm": 0.22453, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:25.714093+00:00", "epoch": 0, "step": 1664, "train_loss": 3.8629751205444336, "perplexity": 47.60677678537381, "lr": 0.0026291804804649314, "grad_norm": 0.210748, "tokens_per_sec": 106312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:26.020741+00:00", "epoch": 0, "step": 1665, "train_loss": 4.008399963378906, "perplexity": 55.05870410461471, "lr": 0.0026291804804649314, "grad_norm": 0.182836, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:26.327898+00:00", "epoch": 0, "step": 1666, "train_loss": 3.854233503341675, "perplexity": 47.19243023447853, "lr": 0.0026291804804649314, "grad_norm": 0.181696, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:26.634348+00:00", "epoch": 0, "step": 1667, "train_loss": 3.8821969032287598, "perplexity": 48.53071532833054, "lr": 0.0026291804804649314, "grad_norm": 0.181425, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:26.940386+00:00", "epoch": 0, "step": 1668, "train_loss": 3.901224136352539, "perplexity": 49.46296146952045, "lr": 0.0026291804804649314, "grad_norm": 0.173439, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:27.246520+00:00", "epoch": 0, "step": 1669, "train_loss": 3.944633722305298, "perplexity": 51.65741368812145, "lr": 0.0026291804804649314, "grad_norm": 0.178452, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:27.552767+00:00", "epoch": 0, "step": 1670, "train_loss": 3.9554283618927, "perplexity": 52.218057380160374, "lr": 0.0026291804804649314, "grad_norm": 0.174413, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:27.860735+00:00", "epoch": 0, "step": 1671, "train_loss": 3.8958308696746826, "perplexity": 49.19691260875369, "lr": 0.0026291804804649314, "grad_norm": 0.184872, "tokens_per_sec": 106343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:28.167607+00:00", "epoch": 0, "step": 1672, "train_loss": 3.879049777984619, "perplexity": 48.37822317082564, "lr": 0.0026291804804649314, "grad_norm": 0.185939, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:28.476603+00:00", "epoch": 0, "step": 1673, "train_loss": 3.9015979766845703, "perplexity": 49.48145617627773, "lr": 0.0026291804804649314, "grad_norm": 0.190963, "tokens_per_sec": 106046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:28.782029+00:00", "epoch": 0, "step": 1674, "train_loss": 3.9055445194244385, "perplexity": 49.677122707109625, "lr": 0.0026291804804649314, "grad_norm": 0.185954, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:29.088097+00:00", "epoch": 0, "step": 1675, "train_loss": 3.916447401046753, "perplexity": 50.221709885383845, "lr": 0.0026291804804649314, "grad_norm": 0.189165, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:29.393862+00:00", "epoch": 0, "step": 1676, "train_loss": 3.9419877529144287, "perplexity": 51.520910424060474, "lr": 0.0026291804804649314, "grad_norm": 0.18962, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:29.699894+00:00", "epoch": 0, "step": 1677, "train_loss": 3.933830976486206, "perplexity": 51.10237514622868, "lr": 0.0026291804804649314, "grad_norm": 0.196461, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:30.007439+00:00", "epoch": 0, "step": 1678, "train_loss": 3.8716917037963867, "perplexity": 48.023559038301364, "lr": 0.0026291804804649314, "grad_norm": 0.21893, "tokens_per_sec": 106547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:30.313308+00:00", "epoch": 0, "step": 1679, "train_loss": 3.8180370330810547, "perplexity": 45.51477657304826, "lr": 0.0026291804804649314, "grad_norm": 0.249907, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:30.618556+00:00", "epoch": 0, "step": 1680, "train_loss": 3.9282174110412598, "perplexity": 50.81631228644896, "lr": 0.0026291804804649314, "grad_norm": 0.285808, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:30.923490+00:00", "epoch": 0, "step": 1681, "train_loss": 3.9061167240142822, "perplexity": 49.70555631887864, "lr": 0.0026291804804649314, "grad_norm": 0.261158, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:31.231106+00:00", "epoch": 0, "step": 1682, "train_loss": 4.015100479125977, "perplexity": 55.42886456656233, "lr": 0.0026291804804649314, "grad_norm": 0.217316, "tokens_per_sec": 106523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:31.537361+00:00", "epoch": 0, "step": 1683, "train_loss": 3.8776683807373047, "perplexity": 48.311439764334445, "lr": 0.0026291804804649314, "grad_norm": 0.224599, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:31.844398+00:00", "epoch": 0, "step": 1684, "train_loss": 3.822071075439453, "perplexity": 45.69875595047886, "lr": 0.0026291804804649314, "grad_norm": 0.20361, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:32.150969+00:00", "epoch": 0, "step": 1685, "train_loss": 3.804870843887329, "perplexity": 44.91944811703775, "lr": 0.0026291804804649314, "grad_norm": 0.182632, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:32.456628+00:00", "epoch": 0, "step": 1686, "train_loss": 3.927328109741211, "perplexity": 50.77114136212678, "lr": 0.0026291804804649314, "grad_norm": 0.179376, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:32.762467+00:00", "epoch": 0, "step": 1687, "train_loss": 3.792985439300537, "perplexity": 44.3887224943625, "lr": 0.0026291804804649314, "grad_norm": 0.177054, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:33.068930+00:00", "epoch": 0, "step": 1688, "train_loss": 3.94075608253479, "perplexity": 51.45749270763769, "lr": 0.0026291804804649314, "grad_norm": 0.17182, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:33.375762+00:00", "epoch": 0, "step": 1689, "train_loss": 3.911144256591797, "perplexity": 49.95608185751697, "lr": 0.0026291804804649314, "grad_norm": 0.176528, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:33.682250+00:00", "epoch": 0, "step": 1690, "train_loss": 3.7921266555786133, "perplexity": 44.350618545914514, "lr": 0.0026291804804649314, "grad_norm": 0.1885, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:33.987152+00:00", "epoch": 0, "step": 1691, "train_loss": 3.870626926422119, "perplexity": 47.97245185292147, "lr": 0.0026291804804649314, "grad_norm": 0.192549, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:34.292101+00:00", "epoch": 0, "step": 1692, "train_loss": 3.924560546875, "perplexity": 50.63082329571556, "lr": 0.0026291804804649314, "grad_norm": 0.18227, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:34.598173+00:00", "epoch": 0, "step": 1693, "train_loss": 3.917280673980713, "perplexity": 50.2635757173384, "lr": 0.0026291804804649314, "grad_norm": 0.188859, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:34.904989+00:00", "epoch": 0, "step": 1694, "train_loss": 3.9785683155059814, "perplexity": 53.440469540737226, "lr": 0.0026291804804649314, "grad_norm": 0.182338, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:35.210545+00:00", "epoch": 0, "step": 1695, "train_loss": 3.8756465911865234, "perplexity": 48.21386287344312, "lr": 0.0026291804804649314, "grad_norm": 0.184833, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:35.516394+00:00", "epoch": 0, "step": 1696, "train_loss": 3.951775074005127, "perplexity": 52.02763782413599, "lr": 0.0026291804804649314, "grad_norm": 0.177749, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:35.821291+00:00", "epoch": 0, "step": 1697, "train_loss": 3.9153425693511963, "perplexity": 50.16625398885456, "lr": 0.0026291804804649314, "grad_norm": 0.170508, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:36.127136+00:00", "epoch": 0, "step": 1698, "train_loss": 3.846066951751709, "perplexity": 46.80860023527283, "lr": 0.0026291804804649314, "grad_norm": 0.170941, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:36.434181+00:00", "epoch": 0, "step": 1699, "train_loss": 3.872544288635254, "perplexity": 48.06452065579163, "lr": 0.0026291804804649314, "grad_norm": 0.168458, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:36.741874+00:00", "epoch": 0, "step": 1700, "train_loss": 3.98098087310791, "perplexity": 53.56955340029136, "lr": 0.0026291804804649314, "grad_norm": 0.190131, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:37.046775+00:00", "epoch": 0, "step": 1701, "train_loss": 3.8477022647857666, "perplexity": 46.88520957239463, "lr": 0.0026291804804649314, "grad_norm": 0.179357, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:37.352156+00:00", "epoch": 0, "step": 1702, "train_loss": 3.8050730228424072, "perplexity": 44.928530802253704, "lr": 0.0026291804804649314, "grad_norm": 0.172393, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:37.657748+00:00", "epoch": 0, "step": 1703, "train_loss": 3.957613229751587, "perplexity": 52.33227166150507, "lr": 0.0026291804804649314, "grad_norm": 0.179733, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:37.964458+00:00", "epoch": 0, "step": 1704, "train_loss": 3.7986366748809814, "perplexity": 44.64028376880775, "lr": 0.0026291804804649314, "grad_norm": 0.165316, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:38.270443+00:00", "epoch": 0, "step": 1705, "train_loss": 3.8803818225860596, "perplexity": 48.44270806066746, "lr": 0.0026291804804649314, "grad_norm": 0.174631, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:38.576853+00:00", "epoch": 0, "step": 1706, "train_loss": 3.8208413124084473, "perplexity": 45.64259185119265, "lr": 0.0026291804804649314, "grad_norm": 0.17836, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:38.883244+00:00", "epoch": 0, "step": 1707, "train_loss": 3.8486199378967285, "perplexity": 46.92825461613986, "lr": 0.0026291804804649314, "grad_norm": 0.179571, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:39.188741+00:00", "epoch": 0, "step": 1708, "train_loss": 3.828901767730713, "perplexity": 46.01197863649298, "lr": 0.0026291804804649314, "grad_norm": 0.223503, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:39.495866+00:00", "epoch": 0, "step": 1709, "train_loss": 4.002985954284668, "perplexity": 54.761421252059044, "lr": 0.0026291804804649314, "grad_norm": 0.233213, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:39.801879+00:00", "epoch": 0, "step": 1710, "train_loss": 3.7861428260803223, "perplexity": 44.08602443891229, "lr": 0.0026291804804649314, "grad_norm": 0.237842, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:40.108129+00:00", "epoch": 0, "step": 1711, "train_loss": 3.8779006004333496, "perplexity": 48.32265993491389, "lr": 0.0026291804804649314, "grad_norm": 0.219445, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:40.413999+00:00", "epoch": 0, "step": 1712, "train_loss": 3.89288592338562, "perplexity": 49.05224346951168, "lr": 0.0026291804804649314, "grad_norm": 0.237074, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:40.719220+00:00", "epoch": 0, "step": 1713, "train_loss": 3.8855504989624023, "perplexity": 48.69374093640776, "lr": 0.0026291804804649314, "grad_norm": 0.248111, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:41.025171+00:00", "epoch": 0, "step": 1714, "train_loss": 3.7898197174072266, "perplexity": 44.24842233655821, "lr": 0.0026291804804649314, "grad_norm": 0.267384, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:41.331631+00:00", "epoch": 0, "step": 1715, "train_loss": 3.8974833488464355, "perplexity": 49.278276689863894, "lr": 0.0026291804804649314, "grad_norm": 0.20799, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:41.638000+00:00", "epoch": 0, "step": 1716, "train_loss": 3.9223010540008545, "perplexity": 50.51655245695537, "lr": 0.0026291804804649314, "grad_norm": 0.21892, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:41.945105+00:00", "epoch": 0, "step": 1717, "train_loss": 3.780057907104492, "perplexity": 43.818579069145294, "lr": 0.0026291804804649314, "grad_norm": 0.209884, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:42.250184+00:00", "epoch": 0, "step": 1718, "train_loss": 3.9457175731658936, "perplexity": 51.713432973194664, "lr": 0.0026291804804649314, "grad_norm": 0.17583, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:42.555210+00:00", "epoch": 0, "step": 1719, "train_loss": 3.779398202896118, "perplexity": 43.789681301166034, "lr": 0.0026291804804649314, "grad_norm": 0.168494, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:42.860353+00:00", "epoch": 0, "step": 1720, "train_loss": 3.860058546066284, "perplexity": 47.46813035980967, "lr": 0.0026291804804649314, "grad_norm": 0.176358, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:43.165727+00:00", "epoch": 0, "step": 1721, "train_loss": 3.846351146697998, "perplexity": 46.82190489337131, "lr": 0.0026291804804649314, "grad_norm": 0.181498, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:43.472694+00:00", "epoch": 0, "step": 1722, "train_loss": 3.844788074493408, "perplexity": 46.74877604299222, "lr": 0.0026291804804649314, "grad_norm": 0.168327, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:43.777526+00:00", "epoch": 0, "step": 1723, "train_loss": 3.8732900619506836, "perplexity": 48.100379262249575, "lr": 0.0026291804804649314, "grad_norm": 0.171218, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:44.082657+00:00", "epoch": 0, "step": 1724, "train_loss": 3.845245122909546, "perplexity": 46.77014738053501, "lr": 0.0026291804804649314, "grad_norm": 0.173889, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:44.387761+00:00", "epoch": 0, "step": 1725, "train_loss": 3.900007963180542, "perplexity": 49.402842507717985, "lr": 0.0026291804804649314, "grad_norm": 0.159325, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:44.693124+00:00", "epoch": 0, "step": 1726, "train_loss": 3.8046979904174805, "perplexity": 44.911684305587244, "lr": 0.0026291804804649314, "grad_norm": 0.166417, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:45.000549+00:00", "epoch": 0, "step": 1727, "train_loss": 3.91933536529541, "perplexity": 50.36695802280177, "lr": 0.0026291804804649314, "grad_norm": 0.166227, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:45.306506+00:00", "epoch": 0, "step": 1728, "train_loss": 3.8687961101531982, "perplexity": 47.88470345771411, "lr": 0.0026291804804649314, "grad_norm": 0.166531, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:45.612609+00:00", "epoch": 0, "step": 1729, "train_loss": 3.831345796585083, "perplexity": 46.12457077309722, "lr": 0.0026291804804649314, "grad_norm": 0.164031, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:45.919024+00:00", "epoch": 0, "step": 1730, "train_loss": 3.749772548675537, "perplexity": 42.51141162345595, "lr": 0.0026291804804649314, "grad_norm": 0.16344, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:46.224411+00:00", "epoch": 0, "step": 1731, "train_loss": 3.7890846729278564, "perplexity": 44.215909728569244, "lr": 0.0026291804804649314, "grad_norm": 0.178868, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:46.530113+00:00", "epoch": 0, "step": 1732, "train_loss": 3.832998037338257, "perplexity": 46.20084266106077, "lr": 0.0026291804804649314, "grad_norm": 0.193504, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:46.837066+00:00", "epoch": 0, "step": 1733, "train_loss": 3.9028966426849365, "perplexity": 49.54575780519485, "lr": 0.0026291804804649314, "grad_norm": 0.190176, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:47.142195+00:00", "epoch": 0, "step": 1734, "train_loss": 3.7862167358398438, "perplexity": 44.08928294679311, "lr": 0.0026291804804649314, "grad_norm": 0.185667, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:47.447369+00:00", "epoch": 0, "step": 1735, "train_loss": 3.865151882171631, "perplexity": 47.710518259541715, "lr": 0.0026291804804649314, "grad_norm": 0.205227, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:47.753739+00:00", "epoch": 0, "step": 1736, "train_loss": 3.960103988647461, "perplexity": 52.462781199111035, "lr": 0.0026291804804649314, "grad_norm": 0.215103, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:48.059348+00:00", "epoch": 0, "step": 1737, "train_loss": 3.873819589614868, "perplexity": 48.125856488579466, "lr": 0.0026291804804649314, "grad_norm": 0.172998, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:48.365487+00:00", "epoch": 0, "step": 1738, "train_loss": 3.8290438652038574, "perplexity": 46.01851728694346, "lr": 0.0026291804804649314, "grad_norm": 0.180372, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:48.671513+00:00", "epoch": 0, "step": 1739, "train_loss": 3.8923065662384033, "perplexity": 49.02383293238914, "lr": 0.0026291804804649314, "grad_norm": 0.196241, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:48.978484+00:00", "epoch": 0, "step": 1740, "train_loss": 3.8579976558685303, "perplexity": 47.370404490977, "lr": 0.0026291804804649314, "grad_norm": 0.226808, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:49.283844+00:00", "epoch": 0, "step": 1741, "train_loss": 3.8656625747680664, "perplexity": 47.73488989066506, "lr": 0.0026291804804649314, "grad_norm": 0.246565, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:49.588820+00:00", "epoch": 0, "step": 1742, "train_loss": 3.851824998855591, "perplexity": 47.078903823863754, "lr": 0.0026291804804649314, "grad_norm": 0.224912, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:49.893547+00:00", "epoch": 0, "step": 1743, "train_loss": 3.9051332473754883, "perplexity": 49.656696095802985, "lr": 0.0026291804804649314, "grad_norm": 0.222283, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:50.200872+00:00", "epoch": 0, "step": 1744, "train_loss": 3.8869948387145996, "perplexity": 48.764122057015285, "lr": 0.0026291804804649314, "grad_norm": 0.231062, "tokens_per_sec": 106623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:50.506132+00:00", "epoch": 0, "step": 1745, "train_loss": 3.819873094558716, "perplexity": 45.598421265879516, "lr": 0.0026291804804649314, "grad_norm": 0.241073, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:50.811048+00:00", "epoch": 0, "step": 1746, "train_loss": 3.9996654987335205, "perplexity": 54.579889936993624, "lr": 0.0026291804804649314, "grad_norm": 0.201621, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:51.116213+00:00", "epoch": 0, "step": 1747, "train_loss": 3.8619778156280518, "perplexity": 47.559321980221185, "lr": 0.0026291804804649314, "grad_norm": 0.19407, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:51.421603+00:00", "epoch": 0, "step": 1748, "train_loss": 3.7159626483917236, "perplexity": 41.0981310976526, "lr": 0.0026291804804649314, "grad_norm": 0.206498, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:51.726847+00:00", "epoch": 0, "step": 1749, "train_loss": 3.873565196990967, "perplexity": 48.1136151827848, "lr": 0.0026291804804649314, "grad_norm": 0.193547, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:52.033080+00:00", "epoch": 0, "step": 1750, "train_loss": 3.800189256668091, "perplexity": 44.70964529114369, "lr": 0.0026291804804649314, "grad_norm": 0.180187, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:52.338418+00:00", "epoch": 0, "step": 1751, "train_loss": 3.8729496002197266, "perplexity": 48.084005711308166, "lr": 0.0026291804804649314, "grad_norm": 0.187547, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:52.643752+00:00", "epoch": 0, "step": 1752, "train_loss": 3.8833398818969727, "perplexity": 48.586216613053935, "lr": 0.0026291804804649314, "grad_norm": 0.183472, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:52.948966+00:00", "epoch": 0, "step": 1753, "train_loss": 3.8394432067871094, "perplexity": 46.49957658193943, "lr": 0.0026291804804649314, "grad_norm": 0.180681, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:53.254418+00:00", "epoch": 0, "step": 1754, "train_loss": 3.835465431213379, "perplexity": 46.3149790891244, "lr": 0.0026291804804649314, "grad_norm": 0.182911, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:53.560283+00:00", "epoch": 0, "step": 1755, "train_loss": 3.8303043842315674, "perplexity": 46.07656107857702, "lr": 0.0026291804804649314, "grad_norm": 0.172916, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:53.866475+00:00", "epoch": 0, "step": 1756, "train_loss": 3.912384510040283, "perplexity": 50.018078498140206, "lr": 0.0026291804804649314, "grad_norm": 0.18279, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:54.172177+00:00", "epoch": 0, "step": 1757, "train_loss": 3.889622449874878, "perplexity": 48.892423697945425, "lr": 0.0026291804804649314, "grad_norm": 0.160099, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:54.477331+00:00", "epoch": 0, "step": 1758, "train_loss": 3.758836507797241, "perplexity": 42.89848487869007, "lr": 0.0026291804804649314, "grad_norm": 0.184342, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:54.783764+00:00", "epoch": 0, "step": 1759, "train_loss": 3.9260523319244385, "perplexity": 50.70640996646386, "lr": 0.0026291804804649314, "grad_norm": 0.177739, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:55.090462+00:00", "epoch": 0, "step": 1760, "train_loss": 3.7369978427886963, "perplexity": 41.97179490403063, "lr": 0.0026291804804649314, "grad_norm": 0.162044, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:55.395906+00:00", "epoch": 0, "step": 1761, "train_loss": 3.892892837524414, "perplexity": 49.05258262470366, "lr": 0.0026291804804649314, "grad_norm": 0.181109, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:55.701479+00:00", "epoch": 0, "step": 1762, "train_loss": 3.866206645965576, "perplexity": 47.760868135737496, "lr": 0.0026291804804649314, "grad_norm": 0.171814, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:56.007185+00:00", "epoch": 0, "step": 1763, "train_loss": 3.8211581707000732, "perplexity": 45.657056376353346, "lr": 0.0026291804804649314, "grad_norm": 0.17337, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:56.312022+00:00", "epoch": 0, "step": 1764, "train_loss": 3.9002580642700195, "perplexity": 49.41519975766887, "lr": 0.0026291804804649314, "grad_norm": 0.170067, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:56.617492+00:00", "epoch": 0, "step": 1765, "train_loss": 3.839970588684082, "perplexity": 46.52410608448489, "lr": 0.0026291804804649314, "grad_norm": 0.142115, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:56.924522+00:00", "epoch": 0, "step": 1766, "train_loss": 3.899454355239868, "perplexity": 49.37550027095187, "lr": 0.0026291804804649314, "grad_norm": 0.171087, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:57.230264+00:00", "epoch": 0, "step": 1767, "train_loss": 3.8503496646881104, "perplexity": 47.00949791952507, "lr": 0.0026291804804649314, "grad_norm": 0.17638, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:57.534990+00:00", "epoch": 0, "step": 1768, "train_loss": 3.852969169616699, "perplexity": 47.13280095696611, "lr": 0.0026291804804649314, "grad_norm": 0.174319, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:57.839812+00:00", "epoch": 0, "step": 1769, "train_loss": 3.822002410888672, "perplexity": 45.69561817365851, "lr": 0.0026291804804649314, "grad_norm": 0.207596, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:58.144721+00:00", "epoch": 0, "step": 1770, "train_loss": 3.764319896697998, "perplexity": 43.1343600611965, "lr": 0.0026291804804649314, "grad_norm": 0.217557, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:58.450662+00:00", "epoch": 0, "step": 1771, "train_loss": 3.8543636798858643, "perplexity": 47.19857398183558, "lr": 0.0026291804804649314, "grad_norm": 0.219844, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:58.756183+00:00", "epoch": 0, "step": 1772, "train_loss": 3.8693137168884277, "perplexity": 47.90949531840182, "lr": 0.0026291804804649314, "grad_norm": 0.211704, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:59.061879+00:00", "epoch": 0, "step": 1773, "train_loss": 3.805109977722168, "perplexity": 44.930191161386325, "lr": 0.0026291804804649314, "grad_norm": 0.189808, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:59.368742+00:00", "epoch": 0, "step": 1774, "train_loss": 3.7938952445983887, "perplexity": 44.429125966112075, "lr": 0.0026291804804649314, "grad_norm": 0.192278, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:59.673550+00:00", "epoch": 0, "step": 1775, "train_loss": 3.791595458984375, "perplexity": 44.32706590448584, "lr": 0.0026291804804649314, "grad_norm": 0.171704, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:06:59.979225+00:00", "epoch": 0, "step": 1776, "train_loss": 3.7827651500701904, "perplexity": 43.93736733088859, "lr": 0.0026291804804649314, "grad_norm": 0.17635, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:00.284815+00:00", "epoch": 0, "step": 1777, "train_loss": 3.8713183403015137, "perplexity": 48.005632141295294, "lr": 0.0026291804804649314, "grad_norm": 0.168691, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:00.591699+00:00", "epoch": 0, "step": 1778, "train_loss": 3.859808921813965, "perplexity": 47.45628264205997, "lr": 0.0026291804804649314, "grad_norm": 0.172722, "tokens_per_sec": 106777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:00.898431+00:00", "epoch": 0, "step": 1779, "train_loss": 3.907008171081543, "perplexity": 49.74988594710856, "lr": 0.0026291804804649314, "grad_norm": 0.17576, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:01.203153+00:00", "epoch": 0, "step": 1780, "train_loss": 3.7767505645751953, "perplexity": 43.6738954101113, "lr": 0.0026291804804649314, "grad_norm": 0.173776, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:01.508324+00:00", "epoch": 0, "step": 1781, "train_loss": 3.833247661590576, "perplexity": 46.21237695142597, "lr": 0.0026291804804649314, "grad_norm": 0.172946, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:01.814125+00:00", "epoch": 0, "step": 1782, "train_loss": 3.6692798137664795, "perplexity": 39.22364735497912, "lr": 0.0026291804804649314, "grad_norm": 0.193604, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:02.119411+00:00", "epoch": 0, "step": 1783, "train_loss": 3.907496929168701, "perplexity": 49.77420754940614, "lr": 0.0026291804804649314, "grad_norm": 0.193493, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:02.424664+00:00", "epoch": 0, "step": 1784, "train_loss": 3.8286707401275635, "perplexity": 46.001349827173875, "lr": 0.0026291804804649314, "grad_norm": 0.194203, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:02.730612+00:00", "epoch": 0, "step": 1785, "train_loss": 3.897036075592041, "perplexity": 49.25624076308557, "lr": 0.0026291804804649314, "grad_norm": 0.190604, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:03.035394+00:00", "epoch": 0, "step": 1786, "train_loss": 3.8185670375823975, "perplexity": 45.538906003298074, "lr": 0.0026291804804649314, "grad_norm": 0.19196, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:03.341045+00:00", "epoch": 0, "step": 1787, "train_loss": 3.8313751220703125, "perplexity": 46.12592341834953, "lr": 0.0026291804804649314, "grad_norm": 0.195926, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:03.647527+00:00", "epoch": 0, "step": 1788, "train_loss": 3.9031472206115723, "perplexity": 49.55817443406087, "lr": 0.0026291804804649314, "grad_norm": 0.199079, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:03.953700+00:00", "epoch": 0, "step": 1789, "train_loss": 3.8718795776367188, "perplexity": 48.03258225635108, "lr": 0.0026291804804649314, "grad_norm": 0.215168, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:04.259709+00:00", "epoch": 0, "step": 1790, "train_loss": 3.895637035369873, "perplexity": 49.187377483546435, "lr": 0.0026291804804649314, "grad_norm": 0.237818, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:04.564604+00:00", "epoch": 0, "step": 1791, "train_loss": 3.8340306282043457, "perplexity": 46.24857386836144, "lr": 0.0026291804804649314, "grad_norm": 0.22988, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:04.869048+00:00", "epoch": 0, "step": 1792, "train_loss": 3.867961883544922, "perplexity": 47.84477342162527, "lr": 0.0026291804804649314, "grad_norm": 0.186801, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:05.174639+00:00", "epoch": 0, "step": 1793, "train_loss": 3.801210880279541, "perplexity": 44.755345060439566, "lr": 0.0026291804804649314, "grad_norm": 0.201819, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:05.479732+00:00", "epoch": 0, "step": 1794, "train_loss": 3.770836114883423, "perplexity": 43.41635072117519, "lr": 0.0026291804804649314, "grad_norm": 0.220565, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:05.785651+00:00", "epoch": 0, "step": 1795, "train_loss": 3.7226474285125732, "perplexity": 41.37378337837253, "lr": 0.0026291804804649314, "grad_norm": 0.187073, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:06.090179+00:00", "epoch": 0, "step": 1796, "train_loss": 3.856161594390869, "perplexity": 47.28350931296555, "lr": 0.0026291804804649314, "grad_norm": 0.200742, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:06.395903+00:00", "epoch": 0, "step": 1797, "train_loss": 3.8258345127105713, "perplexity": 45.87106438448437, "lr": 0.0026291804804649314, "grad_norm": 0.211719, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:06.701611+00:00", "epoch": 0, "step": 1798, "train_loss": 3.8226258754730225, "perplexity": 45.724116656223956, "lr": 0.0026291804804649314, "grad_norm": 0.187731, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:07.007987+00:00", "epoch": 0, "step": 1799, "train_loss": 3.8739500045776367, "perplexity": 48.1321332296432, "lr": 0.0026291804804649314, "grad_norm": 0.187339, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:07.314596+00:00", "epoch": 0, "step": 1800, "train_loss": 3.7632243633270264, "perplexity": 43.08713080565041, "lr": 0.0026291804804649314, "grad_norm": 0.170903, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:07.620113+00:00", "epoch": 0, "step": 1801, "train_loss": 3.7376959323883057, "perplexity": 42.00110520694857, "lr": 0.0026291804804649314, "grad_norm": 0.169849, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:07.925100+00:00", "epoch": 0, "step": 1802, "train_loss": 3.80352783203125, "perplexity": 44.859161257692996, "lr": 0.0026291804804649314, "grad_norm": 0.1569, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:08.231547+00:00", "epoch": 0, "step": 1803, "train_loss": 3.856503963470459, "perplexity": 47.299700496050875, "lr": 0.0026291804804649314, "grad_norm": 0.159892, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:08.537453+00:00", "epoch": 0, "step": 1804, "train_loss": 3.880103588104248, "perplexity": 48.429231503801, "lr": 0.0026291804804649314, "grad_norm": 0.163033, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:08.842929+00:00", "epoch": 0, "step": 1805, "train_loss": 3.836019277572632, "perplexity": 46.34063757644579, "lr": 0.0026291804804649314, "grad_norm": 0.166791, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:09.148099+00:00", "epoch": 0, "step": 1806, "train_loss": 3.8355016708374023, "perplexity": 46.31665755696658, "lr": 0.0026291804804649314, "grad_norm": 0.167467, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:09.453217+00:00", "epoch": 0, "step": 1807, "train_loss": 3.8642008304595947, "perplexity": 47.66516465969676, "lr": 0.0026291804804649314, "grad_norm": 0.167026, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:09.758771+00:00", "epoch": 0, "step": 1808, "train_loss": 3.8780884742736816, "perplexity": 48.33173935147664, "lr": 0.0026291804804649314, "grad_norm": 0.154055, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:10.065308+00:00", "epoch": 0, "step": 1809, "train_loss": 3.850346088409424, "perplexity": 47.00932980076021, "lr": 0.0026291804804649314, "grad_norm": 0.146984, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:10.372273+00:00", "epoch": 0, "step": 1810, "train_loss": 3.871061325073242, "perplexity": 47.993295548206234, "lr": 0.0026291804804649314, "grad_norm": 0.156637, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:10.677634+00:00", "epoch": 0, "step": 1811, "train_loss": 3.821790933609009, "perplexity": 45.685955610376936, "lr": 0.0026291804804649314, "grad_norm": 0.178742, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:10.982440+00:00", "epoch": 0, "step": 1812, "train_loss": 3.811866044998169, "perplexity": 45.234770275561516, "lr": 0.0026291804804649314, "grad_norm": 0.199009, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:11.287811+00:00", "epoch": 0, "step": 1813, "train_loss": 3.8901114463806152, "perplexity": 48.91633776876276, "lr": 0.0026291804804649314, "grad_norm": 0.214521, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:11.593079+00:00", "epoch": 0, "step": 1814, "train_loss": 3.751206159591675, "perplexity": 42.5724001536856, "lr": 0.0026291804804649314, "grad_norm": 0.219703, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:11.898878+00:00", "epoch": 0, "step": 1815, "train_loss": 3.8904404640197754, "perplexity": 48.93243475468274, "lr": 0.0026291804804649314, "grad_norm": 0.19975, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:12.204468+00:00", "epoch": 0, "step": 1816, "train_loss": 3.8074231147766113, "perplexity": 45.0342411459682, "lr": 0.0026291804804649314, "grad_norm": 0.21237, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:12.509646+00:00", "epoch": 0, "step": 1817, "train_loss": 3.8581056594848633, "perplexity": 47.375520942261836, "lr": 0.0026291804804649314, "grad_norm": 0.183981, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:12.814353+00:00", "epoch": 0, "step": 1818, "train_loss": 3.9137229919433594, "perplexity": 50.08507161557256, "lr": 0.0026291804804649314, "grad_norm": 0.199421, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:13.121084+00:00", "epoch": 0, "step": 1819, "train_loss": 3.821465015411377, "perplexity": 45.67106815224585, "lr": 0.0026291804804649314, "grad_norm": 0.204041, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:13.427824+00:00", "epoch": 0, "step": 1820, "train_loss": 3.856607675552368, "perplexity": 47.30460630085422, "lr": 0.0026291804804649314, "grad_norm": 0.204706, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:13.733842+00:00", "epoch": 0, "step": 1821, "train_loss": 3.873950719833374, "perplexity": 48.13216765643996, "lr": 0.0026291804804649314, "grad_norm": 0.206609, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:14.039324+00:00", "epoch": 0, "step": 1822, "train_loss": 3.7917118072509766, "perplexity": 44.332223581804946, "lr": 0.0026291804804649314, "grad_norm": 0.205127, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:14.345886+00:00", "epoch": 0, "step": 1823, "train_loss": 3.9502458572387695, "perplexity": 51.9481370904815, "lr": 0.0026291804804649314, "grad_norm": 0.199067, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:14.651797+00:00", "epoch": 0, "step": 1824, "train_loss": 3.803605794906616, "perplexity": 44.862658743226405, "lr": 0.0026291804804649314, "grad_norm": 0.194945, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:14.958446+00:00", "epoch": 0, "step": 1825, "train_loss": 3.8616201877593994, "perplexity": 47.54231648226809, "lr": 0.0026291804804649314, "grad_norm": 0.230918, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:15.265695+00:00", "epoch": 0, "step": 1826, "train_loss": 3.926105260848999, "perplexity": 50.70909387323923, "lr": 0.0026291804804649314, "grad_norm": 0.206823, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:15.570840+00:00", "epoch": 0, "step": 1827, "train_loss": 3.941542625427246, "perplexity": 51.49798215404624, "lr": 0.0026291804804649314, "grad_norm": 0.167769, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:15.877094+00:00", "epoch": 0, "step": 1828, "train_loss": 3.8415987491607666, "perplexity": 46.59991649423397, "lr": 0.0026291804804649314, "grad_norm": 0.176034, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:16.183038+00:00", "epoch": 0, "step": 1829, "train_loss": 3.938824415206909, "perplexity": 51.358189891080535, "lr": 0.0026291804804649314, "grad_norm": 0.163021, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:16.489604+00:00", "epoch": 0, "step": 1830, "train_loss": 3.798177719116211, "perplexity": 44.61980055403536, "lr": 0.0026291804804649314, "grad_norm": 0.16762, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:16.795522+00:00", "epoch": 0, "step": 1831, "train_loss": 3.7890477180480957, "perplexity": 44.21427576513336, "lr": 0.0026291804804649314, "grad_norm": 0.161474, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:17.101875+00:00", "epoch": 0, "step": 1832, "train_loss": 3.929360866546631, "perplexity": 50.87445171208399, "lr": 0.0026291804804649314, "grad_norm": 0.17426, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:17.408292+00:00", "epoch": 0, "step": 1833, "train_loss": 3.829589605331421, "perplexity": 46.04363829258484, "lr": 0.0026291804804649314, "grad_norm": 0.165591, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:17.714368+00:00", "epoch": 0, "step": 1834, "train_loss": 3.7383289337158203, "perplexity": 42.02770037880316, "lr": 0.0026291804804649314, "grad_norm": 0.175215, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:18.019695+00:00", "epoch": 0, "step": 1835, "train_loss": 3.7756669521331787, "perplexity": 43.62659546568607, "lr": 0.0026291804804649314, "grad_norm": 0.178219, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:18.324736+00:00", "epoch": 0, "step": 1836, "train_loss": 3.845905065536499, "perplexity": 46.801023181468295, "lr": 0.0026291804804649314, "grad_norm": 0.173911, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:18.630267+00:00", "epoch": 0, "step": 1837, "train_loss": 3.974048137664795, "perplexity": 53.19945424087423, "lr": 0.0026291804804649314, "grad_norm": 0.183512, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:18.935532+00:00", "epoch": 0, "step": 1838, "train_loss": 3.852855682373047, "perplexity": 47.127452288808435, "lr": 0.0026291804804649314, "grad_norm": 0.163407, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:19.242846+00:00", "epoch": 0, "step": 1839, "train_loss": 3.940140962600708, "perplexity": 51.42584991117051, "lr": 0.0026291804804649314, "grad_norm": 0.163436, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:19.549003+00:00", "epoch": 0, "step": 1840, "train_loss": 3.8854470252990723, "perplexity": 48.68870267731988, "lr": 0.0026291804804649314, "grad_norm": 0.164966, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:19.852861+00:00", "epoch": 0, "step": 1841, "train_loss": 3.8772900104522705, "perplexity": 48.29316360889551, "lr": 0.0026291804804649314, "grad_norm": 0.185046, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:20.157839+00:00", "epoch": 0, "step": 1842, "train_loss": 3.8662617206573486, "perplexity": 47.76349862326484, "lr": 0.0026291804804649314, "grad_norm": 0.192928, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:20.462583+00:00", "epoch": 0, "step": 1843, "train_loss": 3.8996758460998535, "perplexity": 49.386437704195124, "lr": 0.0026291804804649314, "grad_norm": 0.189586, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:20.768126+00:00", "epoch": 0, "step": 1844, "train_loss": 3.89070200920105, "perplexity": 48.94523447097995, "lr": 0.0026291804804649314, "grad_norm": 0.187927, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:21.074077+00:00", "epoch": 0, "step": 1845, "train_loss": 3.8418514728546143, "perplexity": 46.61169488553998, "lr": 0.0026291804804649314, "grad_norm": 0.158343, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:21.379425+00:00", "epoch": 0, "step": 1846, "train_loss": 3.9438092708587646, "perplexity": 51.61484221015181, "lr": 0.0026291804804649314, "grad_norm": 0.16221, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:21.683304+00:00", "epoch": 0, "step": 1847, "train_loss": 3.8623745441436768, "perplexity": 47.578193862693965, "lr": 0.0026291804804649314, "grad_norm": 0.187155, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:21.988900+00:00", "epoch": 0, "step": 1848, "train_loss": 3.859266519546509, "perplexity": 47.43054922631231, "lr": 0.0026291804804649314, "grad_norm": 0.165674, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:22.294681+00:00", "epoch": 0, "step": 1849, "train_loss": 3.8544509410858154, "perplexity": 47.20269276573961, "lr": 0.0026291804804649314, "grad_norm": 0.17773, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:22.599479+00:00", "epoch": 0, "step": 1850, "train_loss": 3.9154250621795654, "perplexity": 50.17039251573183, "lr": 0.0026291804804649314, "grad_norm": 0.169369, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:22.904379+00:00", "epoch": 0, "step": 1851, "train_loss": 3.744661808013916, "perplexity": 42.29470107090587, "lr": 0.0026291804804649314, "grad_norm": 0.159673, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:23.208626+00:00", "epoch": 0, "step": 1852, "train_loss": 3.8693933486938477, "perplexity": 47.91331058991724, "lr": 0.0026291804804649314, "grad_norm": 0.165076, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:23.513448+00:00", "epoch": 0, "step": 1853, "train_loss": 3.8805782794952393, "perplexity": 48.45222590025741, "lr": 0.0026291804804649314, "grad_norm": 0.16547, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:23.818445+00:00", "epoch": 0, "step": 1854, "train_loss": 3.9038078784942627, "perplexity": 49.590926250334014, "lr": 0.0026291804804649314, "grad_norm": 0.174877, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:24.124623+00:00", "epoch": 0, "step": 1855, "train_loss": 3.8393189907073975, "perplexity": 46.49380094554903, "lr": 0.0026291804804649314, "grad_norm": 0.199072, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:24.430331+00:00", "epoch": 0, "step": 1856, "train_loss": 3.855179786682129, "perplexity": 47.2371087809531, "lr": 0.0026291804804649314, "grad_norm": 0.21723, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:24.735582+00:00", "epoch": 0, "step": 1857, "train_loss": 3.9455525875091553, "perplexity": 51.70490170228141, "lr": 0.0026291804804649314, "grad_norm": 0.204777, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:25.040880+00:00", "epoch": 0, "step": 1858, "train_loss": 3.8024682998657227, "perplexity": 44.81165670416721, "lr": 0.0026291804804649314, "grad_norm": 0.198433, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:25.347513+00:00", "epoch": 0, "step": 1859, "train_loss": 3.8469552993774414, "perplexity": 46.85020101939878, "lr": 0.0026291804804649314, "grad_norm": 0.189037, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:25.652652+00:00", "epoch": 0, "step": 1860, "train_loss": 3.909674882888794, "perplexity": 49.882731607192234, "lr": 0.0026291804804649314, "grad_norm": 0.197709, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:25.958332+00:00", "epoch": 0, "step": 1861, "train_loss": 3.801044464111328, "perplexity": 44.74789766710808, "lr": 0.0026291804804649314, "grad_norm": 0.195585, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:26.263068+00:00", "epoch": 0, "step": 1862, "train_loss": 3.829310894012451, "perplexity": 46.03080719759491, "lr": 0.0026291804804649314, "grad_norm": 0.220113, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:26.568141+00:00", "epoch": 0, "step": 1863, "train_loss": 3.7811591625213623, "perplexity": 43.86686109724768, "lr": 0.0026291804804649314, "grad_norm": 0.21691, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:26.872603+00:00", "epoch": 0, "step": 1864, "train_loss": 3.8144936561584473, "perplexity": 45.35378595776054, "lr": 0.0026291804804649314, "grad_norm": 0.223696, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:27.178200+00:00", "epoch": 0, "step": 1865, "train_loss": 3.8923490047454834, "perplexity": 49.02591347481738, "lr": 0.0026291804804649314, "grad_norm": 0.190404, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:27.483822+00:00", "epoch": 0, "step": 1866, "train_loss": 3.8234074115753174, "perplexity": 45.7598656718899, "lr": 0.0026291804804649314, "grad_norm": 0.196053, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:27.788192+00:00", "epoch": 0, "step": 1867, "train_loss": 3.8360140323638916, "perplexity": 46.34039451076602, "lr": 0.0026291804804649314, "grad_norm": 0.185328, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:28.092329+00:00", "epoch": 0, "step": 1868, "train_loss": 3.8017706871032715, "perplexity": 44.78040642210896, "lr": 0.0026291804804649314, "grad_norm": 0.162682, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:28.396881+00:00", "epoch": 0, "step": 1869, "train_loss": 3.834178924560547, "perplexity": 46.255432871915666, "lr": 0.0026291804804649314, "grad_norm": 0.188758, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:28.701928+00:00", "epoch": 0, "step": 1870, "train_loss": 3.906306743621826, "perplexity": 49.715002246610425, "lr": 0.0026291804804649314, "grad_norm": 0.169913, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:29.008306+00:00", "epoch": 0, "step": 1871, "train_loss": 3.9160714149475098, "perplexity": 50.20283076995161, "lr": 0.0026291804804649314, "grad_norm": 0.172555, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:29.312935+00:00", "epoch": 0, "step": 1872, "train_loss": 3.8657875061035156, "perplexity": 47.740853846741366, "lr": 0.0026291804804649314, "grad_norm": 0.179813, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:29.617523+00:00", "epoch": 0, "step": 1873, "train_loss": 3.8343141078948975, "perplexity": 46.261686258230405, "lr": 0.0026291804804649314, "grad_norm": 0.188076, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:29.921320+00:00", "epoch": 0, "step": 1874, "train_loss": 3.7877612113952637, "perplexity": 44.15743037903546, "lr": 0.0026291804804649314, "grad_norm": 0.18498, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:30.226314+00:00", "epoch": 0, "step": 1875, "train_loss": 3.948058843612671, "perplexity": 51.834649951008736, "lr": 0.0026291804804649314, "grad_norm": 0.210746, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:30.532185+00:00", "epoch": 0, "step": 1876, "train_loss": 3.906054973602295, "perplexity": 49.7024870750624, "lr": 0.0026291804804649314, "grad_norm": 0.200772, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:30.837587+00:00", "epoch": 0, "step": 1877, "train_loss": 3.822554111480713, "perplexity": 45.720835428806275, "lr": 0.0026291804804649314, "grad_norm": 0.167997, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:31.141794+00:00", "epoch": 0, "step": 1878, "train_loss": 3.9351234436035156, "perplexity": 51.168465986640356, "lr": 0.0026291804804649314, "grad_norm": 0.192103, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:31.445741+00:00", "epoch": 0, "step": 1879, "train_loss": 3.8681936264038086, "perplexity": 47.855862391045896, "lr": 0.0026291804804649314, "grad_norm": 0.168262, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:31.750357+00:00", "epoch": 0, "step": 1880, "train_loss": 3.8374876976013184, "perplexity": 46.40873508243969, "lr": 0.0026291804804649314, "grad_norm": 0.167283, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:32.056258+00:00", "epoch": 0, "step": 1881, "train_loss": 3.8448410034179688, "perplexity": 46.751250470916524, "lr": 0.0026291804804649314, "grad_norm": 0.1639, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:32.361039+00:00", "epoch": 0, "step": 1882, "train_loss": 3.8991243839263916, "perplexity": 49.359210460006864, "lr": 0.0026291804804649314, "grad_norm": 0.177772, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:32.666783+00:00", "epoch": 0, "step": 1883, "train_loss": 3.9000937938690186, "perplexity": 49.4070829696814, "lr": 0.0026291804804649314, "grad_norm": 0.166095, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:32.972474+00:00", "epoch": 0, "step": 1884, "train_loss": 3.804119110107422, "perplexity": 44.88569333940625, "lr": 0.0026291804804649314, "grad_norm": 0.159633, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:33.276388+00:00", "epoch": 0, "step": 1885, "train_loss": 3.8412654399871826, "perplexity": 46.584386902799615, "lr": 0.0026291804804649314, "grad_norm": 0.13682, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:33.581974+00:00", "epoch": 0, "step": 1886, "train_loss": 3.833635091781616, "perplexity": 46.23028449019326, "lr": 0.0026291804804649314, "grad_norm": 0.143744, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:33.888491+00:00", "epoch": 0, "step": 1887, "train_loss": 3.81339955329895, "perplexity": 45.304191386583526, "lr": 0.0026291804804649314, "grad_norm": 0.157692, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:34.193987+00:00", "epoch": 0, "step": 1888, "train_loss": 3.8391051292419434, "perplexity": 46.483858776305624, "lr": 0.0026291804804649314, "grad_norm": 0.171521, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:34.499106+00:00", "epoch": 0, "step": 1889, "train_loss": 3.7956674098968506, "perplexity": 44.50793152900762, "lr": 0.0026291804804649314, "grad_norm": 0.185005, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:34.804081+00:00", "epoch": 0, "step": 1890, "train_loss": 3.74772310256958, "perplexity": 42.42437599432525, "lr": 0.0026291804804649314, "grad_norm": 0.168113, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:35.108766+00:00", "epoch": 0, "step": 1891, "train_loss": 3.7984719276428223, "perplexity": 44.63293001111846, "lr": 0.0026291804804649314, "grad_norm": 0.15998, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:35.414221+00:00", "epoch": 0, "step": 1892, "train_loss": 3.745107650756836, "perplexity": 42.31356206064758, "lr": 0.0026291804804649314, "grad_norm": 0.182066, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:35.719629+00:00", "epoch": 0, "step": 1893, "train_loss": 3.812915325164795, "perplexity": 45.28225913305464, "lr": 0.0026291804804649314, "grad_norm": 0.212505, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:36.024446+00:00", "epoch": 0, "step": 1894, "train_loss": 3.7517364025115967, "perplexity": 42.59497985328506, "lr": 0.0026291804804649314, "grad_norm": 0.213974, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:36.329969+00:00", "epoch": 0, "step": 1895, "train_loss": 3.862745523452759, "perplexity": 47.595847662575295, "lr": 0.0026291804804649314, "grad_norm": 0.173262, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:36.635234+00:00", "epoch": 0, "step": 1896, "train_loss": 3.8478922843933105, "perplexity": 46.894119528023516, "lr": 0.0026291804804649314, "grad_norm": 0.173336, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:36.940924+00:00", "epoch": 0, "step": 1897, "train_loss": 3.828413248062134, "perplexity": 45.98950636945795, "lr": 0.0026291804804649314, "grad_norm": 0.182147, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:37.246567+00:00", "epoch": 0, "step": 1898, "train_loss": 3.906433343887329, "perplexity": 49.7212965775179, "lr": 0.0026291804804649314, "grad_norm": 0.17385, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:37.551226+00:00", "epoch": 0, "step": 1899, "train_loss": 3.8581435680389404, "perplexity": 47.377316913800534, "lr": 0.0026291804804649314, "grad_norm": 0.203731, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:37.857824+00:00", "epoch": 0, "step": 1900, "train_loss": 3.8881027698516846, "perplexity": 48.81817928652909, "lr": 0.0026291804804649314, "grad_norm": 0.221585, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:38.161477+00:00", "epoch": 0, "step": 1901, "train_loss": 3.7599539756774902, "perplexity": 42.94644935204129, "lr": 0.0026291804804649314, "grad_norm": 0.191598, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:38.466132+00:00", "epoch": 0, "step": 1902, "train_loss": 3.776716470718384, "perplexity": 43.67240642395757, "lr": 0.0026291804804649314, "grad_norm": 0.202572, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:38.772299+00:00", "epoch": 0, "step": 1903, "train_loss": 3.866856098175049, "perplexity": 47.791896611740114, "lr": 0.0026291804804649314, "grad_norm": 0.184529, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:39.077957+00:00", "epoch": 0, "step": 1904, "train_loss": 3.905003547668457, "perplexity": 49.650256054511985, "lr": 0.0026291804804649314, "grad_norm": 0.176671, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:39.383056+00:00", "epoch": 0, "step": 1905, "train_loss": 3.778331756591797, "perplexity": 43.74300684969674, "lr": 0.0026291804804649314, "grad_norm": 0.164552, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:39.687687+00:00", "epoch": 0, "step": 1906, "train_loss": 3.7571516036987305, "perplexity": 42.826265903806664, "lr": 0.0026291804804649314, "grad_norm": 0.167181, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:39.993188+00:00", "epoch": 0, "step": 1907, "train_loss": 3.7732906341552734, "perplexity": 43.523047882284956, "lr": 0.0026291804804649314, "grad_norm": 0.153879, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:40.298544+00:00", "epoch": 0, "step": 1908, "train_loss": 3.914729356765747, "perplexity": 50.135500840616196, "lr": 0.0026291804804649314, "grad_norm": 0.150511, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:40.603976+00:00", "epoch": 0, "step": 1909, "train_loss": 3.756335496902466, "perplexity": 42.79132935506043, "lr": 0.0026291804804649314, "grad_norm": 0.176051, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:40.909722+00:00", "epoch": 0, "step": 1910, "train_loss": 3.959526777267456, "perplexity": 52.432507822685416, "lr": 0.0026291804804649314, "grad_norm": 0.201132, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:41.214035+00:00", "epoch": 0, "step": 1911, "train_loss": 3.7920498847961426, "perplexity": 44.34721384491814, "lr": 0.0026291804804649314, "grad_norm": 0.192416, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:41.517446+00:00", "epoch": 0, "step": 1912, "train_loss": 3.8676750659942627, "perplexity": 47.83105266867108, "lr": 0.0026291804804649314, "grad_norm": 0.175392, "tokens_per_sec": 107999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:41.823015+00:00", "epoch": 0, "step": 1913, "train_loss": 3.8694334030151367, "perplexity": 47.91522976348897, "lr": 0.0026291804804649314, "grad_norm": 0.179636, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:42.128474+00:00", "epoch": 0, "step": 1914, "train_loss": 3.908334493637085, "perplexity": 49.81591412062441, "lr": 0.0026291804804649314, "grad_norm": 0.166738, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:42.434458+00:00", "epoch": 0, "step": 1915, "train_loss": 3.8551888465881348, "perplexity": 47.237536746657305, "lr": 0.0026291804804649314, "grad_norm": 0.176768, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:42.739933+00:00", "epoch": 0, "step": 1916, "train_loss": 3.784606456756592, "perplexity": 44.01834402772581, "lr": 0.0026291804804649314, "grad_norm": 0.162088, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:43.045581+00:00", "epoch": 0, "step": 1917, "train_loss": 3.694391965866089, "perplexity": 40.22110934409669, "lr": 0.0026291804804649314, "grad_norm": 0.163208, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:43.351544+00:00", "epoch": 0, "step": 1918, "train_loss": 3.8761789798736572, "perplexity": 48.23953822262624, "lr": 0.0026291804804649314, "grad_norm": 0.153406, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:43.656618+00:00", "epoch": 0, "step": 1919, "train_loss": 3.7867393493652344, "perplexity": 44.112330624375325, "lr": 0.0026291804804649314, "grad_norm": 0.145405, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:43.961897+00:00", "epoch": 0, "step": 1920, "train_loss": 3.7529852390289307, "perplexity": 42.648207248820825, "lr": 0.0026291804804649314, "grad_norm": 0.16543, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:44.268206+00:00", "epoch": 0, "step": 1921, "train_loss": 3.7388105392456055, "perplexity": 42.047946026526375, "lr": 0.0026291804804649314, "grad_norm": 0.157892, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:44.573159+00:00", "epoch": 0, "step": 1922, "train_loss": 3.8233141899108887, "perplexity": 45.755600059874745, "lr": 0.0026291804804649314, "grad_norm": 0.173027, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:44.878587+00:00", "epoch": 0, "step": 1923, "train_loss": 3.8664333820343018, "perplexity": 47.77169847498539, "lr": 0.0026291804804649314, "grad_norm": 0.167192, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:45.184250+00:00", "epoch": 0, "step": 1924, "train_loss": 3.8894526958465576, "perplexity": 48.884124716481054, "lr": 0.0026291804804649314, "grad_norm": 0.17392, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:45.489561+00:00", "epoch": 0, "step": 1925, "train_loss": 3.7994987964630127, "perplexity": 44.678785715167564, "lr": 0.0026291804804649314, "grad_norm": 0.166566, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:45.794656+00:00", "epoch": 0, "step": 1926, "train_loss": 3.8263373374938965, "perplexity": 45.89413529232042, "lr": 0.0026291804804649314, "grad_norm": 0.179902, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:46.099242+00:00", "epoch": 0, "step": 1927, "train_loss": 3.7738595008850098, "perplexity": 43.54781373977021, "lr": 0.0026291804804649314, "grad_norm": 0.193437, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:46.403008+00:00", "epoch": 0, "step": 1928, "train_loss": 3.8461132049560547, "perplexity": 46.81076533309562, "lr": 0.0026291804804649314, "grad_norm": 0.174476, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:46.708245+00:00", "epoch": 0, "step": 1929, "train_loss": 3.8558473587036133, "perplexity": 47.26865348115923, "lr": 0.0026291804804649314, "grad_norm": 0.185749, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:47.013246+00:00", "epoch": 0, "step": 1930, "train_loss": 3.8421595096588135, "perplexity": 46.62605521471223, "lr": 0.0026291804804649314, "grad_norm": 0.18404, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:47.318097+00:00", "epoch": 0, "step": 1931, "train_loss": 3.730375051498413, "perplexity": 41.69474290773249, "lr": 0.0026291804804649314, "grad_norm": 0.177864, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:47.622729+00:00", "epoch": 0, "step": 1932, "train_loss": 3.8114092350006104, "perplexity": 45.214111299235086, "lr": 0.0026291804804649314, "grad_norm": 0.182141, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:47.926113+00:00", "epoch": 0, "step": 1933, "train_loss": 3.788905620574951, "perplexity": 44.20799347462959, "lr": 0.0026291804804649314, "grad_norm": 0.171162, "tokens_per_sec": 108008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:48.230245+00:00", "epoch": 0, "step": 1934, "train_loss": 3.886378288269043, "perplexity": 48.73406578239073, "lr": 0.0026291804804649314, "grad_norm": 0.175152, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:48.535171+00:00", "epoch": 0, "step": 1935, "train_loss": 3.8016977310180664, "perplexity": 44.7771395381335, "lr": 0.0026291804804649314, "grad_norm": 0.179643, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:48.841478+00:00", "epoch": 0, "step": 1936, "train_loss": 3.8834574222564697, "perplexity": 48.59192779006156, "lr": 0.0026291804804649314, "grad_norm": 0.171012, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:49.147729+00:00", "epoch": 0, "step": 1937, "train_loss": 3.8158183097839355, "perplexity": 45.41390382364284, "lr": 0.0026291804804649314, "grad_norm": 0.166731, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:49.452010+00:00", "epoch": 0, "step": 1938, "train_loss": 3.712498903274536, "perplexity": 40.956023900370845, "lr": 0.0026291804804649314, "grad_norm": 0.164569, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:49.755880+00:00", "epoch": 0, "step": 1939, "train_loss": 3.8073277473449707, "perplexity": 45.02994655083973, "lr": 0.0026291804804649314, "grad_norm": 0.165882, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:50.060475+00:00", "epoch": 0, "step": 1940, "train_loss": 3.6919779777526855, "perplexity": 40.12413316100154, "lr": 0.0026291804804649314, "grad_norm": 0.171201, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:50.366663+00:00", "epoch": 0, "step": 1941, "train_loss": 3.9356422424316406, "perplexity": 51.19501901407537, "lr": 0.0026291804804649314, "grad_norm": 0.19517, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:50.671579+00:00", "epoch": 0, "step": 1942, "train_loss": 3.7172272205352783, "perplexity": 41.15013552413518, "lr": 0.0026291804804649314, "grad_norm": 0.173546, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:50.976450+00:00", "epoch": 0, "step": 1943, "train_loss": 3.7781589031219482, "perplexity": 43.73544637262733, "lr": 0.0026291804804649314, "grad_norm": 0.170836, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:51.280405+00:00", "epoch": 0, "step": 1944, "train_loss": 3.907219886779785, "perplexity": 49.76041989401094, "lr": 0.0026291804804649314, "grad_norm": 0.207188, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:51.585404+00:00", "epoch": 0, "step": 1945, "train_loss": 3.8139150142669678, "perplexity": 45.32754994862721, "lr": 0.0026291804804649314, "grad_norm": 0.189579, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:51.890595+00:00", "epoch": 0, "step": 1946, "train_loss": 3.8626203536987305, "perplexity": 47.589890474868234, "lr": 0.0026291804804649314, "grad_norm": 0.171504, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:52.196445+00:00", "epoch": 0, "step": 1947, "train_loss": 3.845102071762085, "perplexity": 46.763457335806, "lr": 0.0026291804804649314, "grad_norm": 0.18264, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:52.501569+00:00", "epoch": 0, "step": 1948, "train_loss": 3.8602957725524902, "perplexity": 47.479392393355205, "lr": 0.0026291804804649314, "grad_norm": 0.171665, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:52.805555+00:00", "epoch": 0, "step": 1949, "train_loss": 3.875645875930786, "perplexity": 48.213828388213415, "lr": 0.0026291804804649314, "grad_norm": 0.168678, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:53.110117+00:00", "epoch": 0, "step": 1950, "train_loss": 3.685631036758423, "perplexity": 39.87027412175124, "lr": 0.0026291804804649314, "grad_norm": 0.165199, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:53.414720+00:00", "epoch": 0, "step": 1951, "train_loss": 3.846975326538086, "perplexity": 46.8511393052964, "lr": 0.0026291804804649314, "grad_norm": 0.164721, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:53.718852+00:00", "epoch": 0, "step": 1952, "train_loss": 3.809865951538086, "perplexity": 45.14438692506971, "lr": 0.0026291804804649314, "grad_norm": 0.179108, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:54.023686+00:00", "epoch": 0, "step": 1953, "train_loss": 3.912322998046875, "perplexity": 50.01500188105073, "lr": 0.0026291804804649314, "grad_norm": 0.184743, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:54.328017+00:00", "epoch": 0, "step": 1954, "train_loss": 3.8053712844848633, "perplexity": 44.94193325826384, "lr": 0.0026291804804649314, "grad_norm": 0.153994, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:54.632370+00:00", "epoch": 0, "step": 1955, "train_loss": 3.8037450313568115, "perplexity": 44.86890569546777, "lr": 0.0026291804804649314, "grad_norm": 0.181315, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:54.937475+00:00", "epoch": 0, "step": 1956, "train_loss": 3.7864251136779785, "perplexity": 44.09847113353248, "lr": 0.0026291804804649314, "grad_norm": 0.199299, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:55.242599+00:00", "epoch": 0, "step": 1957, "train_loss": 3.7618911266326904, "perplexity": 43.02972373891413, "lr": 0.0026291804804649314, "grad_norm": 0.165736, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:55.548267+00:00", "epoch": 0, "step": 1958, "train_loss": 3.7989397048950195, "perplexity": 44.65381316442781, "lr": 0.0026291804804649314, "grad_norm": 0.180741, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:55.853100+00:00", "epoch": 0, "step": 1959, "train_loss": 3.771385669708252, "perplexity": 43.44021694348864, "lr": 0.0026291804804649314, "grad_norm": 0.171535, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:56.157991+00:00", "epoch": 0, "step": 1960, "train_loss": 3.861229419708252, "perplexity": 47.52374209328453, "lr": 0.0026291804804649314, "grad_norm": 0.184324, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:56.462731+00:00", "epoch": 0, "step": 1961, "train_loss": 3.7920939922332764, "perplexity": 44.34916993000349, "lr": 0.0026291804804649314, "grad_norm": 0.179232, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:56.767867+00:00", "epoch": 0, "step": 1962, "train_loss": 3.810153007507324, "perplexity": 45.15734775096638, "lr": 0.0026291804804649314, "grad_norm": 0.176238, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:57.072858+00:00", "epoch": 0, "step": 1963, "train_loss": 3.772784471511841, "perplexity": 43.501023715694025, "lr": 0.0026291804804649314, "grad_norm": 0.177321, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:57.377359+00:00", "epoch": 0, "step": 1964, "train_loss": 3.918140411376953, "perplexity": 50.30680777449657, "lr": 0.0026291804804649314, "grad_norm": 0.168225, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:57.682528+00:00", "epoch": 0, "step": 1965, "train_loss": 3.7912583351135254, "perplexity": 44.3121247111027, "lr": 0.0026291804804649314, "grad_norm": 0.19295, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:57.986886+00:00", "epoch": 0, "step": 1966, "train_loss": 3.7548389434814453, "perplexity": 42.72733774010259, "lr": 0.0026291804804649314, "grad_norm": 0.181895, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:58.291400+00:00", "epoch": 0, "step": 1967, "train_loss": 3.6990509033203125, "perplexity": 40.408934169187596, "lr": 0.0026291804804649314, "grad_norm": 0.206573, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:58.596632+00:00", "epoch": 0, "step": 1968, "train_loss": 3.8435537815093994, "perplexity": 46.69110995245793, "lr": 0.0026291804804649314, "grad_norm": 0.203673, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:58.901042+00:00", "epoch": 0, "step": 1969, "train_loss": 3.781294822692871, "perplexity": 43.87281248682182, "lr": 0.0026291804804649314, "grad_norm": 0.198663, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:59.206612+00:00", "epoch": 0, "step": 1970, "train_loss": 3.816728115081787, "perplexity": 45.45524043522047, "lr": 0.0026291804804649314, "grad_norm": 0.186293, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:59.511389+00:00", "epoch": 0, "step": 1971, "train_loss": 3.7393195629119873, "perplexity": 42.06935487451951, "lr": 0.0026291804804649314, "grad_norm": 0.174637, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:07:59.816531+00:00", "epoch": 0, "step": 1972, "train_loss": 3.824592113494873, "perplexity": 45.81410959768714, "lr": 0.0026291804804649314, "grad_norm": 0.186691, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:00.122156+00:00", "epoch": 0, "step": 1973, "train_loss": 3.786360502243042, "perplexity": 44.09562196007961, "lr": 0.0026291804804649314, "grad_norm": 0.1621, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:00.427909+00:00", "epoch": 0, "step": 1974, "train_loss": 3.7188401222229004, "perplexity": 41.2165602010046, "lr": 0.0026291804804649314, "grad_norm": 0.161995, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:00.732215+00:00", "epoch": 0, "step": 1975, "train_loss": 3.7378108501434326, "perplexity": 42.005932157017625, "lr": 0.0026291804804649314, "grad_norm": 0.165078, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:01.035899+00:00", "epoch": 0, "step": 1976, "train_loss": 3.781872510910034, "perplexity": 43.898164615757295, "lr": 0.0026291804804649314, "grad_norm": 0.155487, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:01.340061+00:00", "epoch": 0, "step": 1977, "train_loss": 3.8282220363616943, "perplexity": 45.98071347842129, "lr": 0.0026291804804649314, "grad_norm": 0.149653, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:01.643906+00:00", "epoch": 0, "step": 1978, "train_loss": 3.6842753887176514, "perplexity": 39.8162606826316, "lr": 0.0026291804804649314, "grad_norm": 0.156447, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:01.948562+00:00", "epoch": 0, "step": 1979, "train_loss": 3.7676522731781006, "perplexity": 43.27833975217865, "lr": 0.0026291804804649314, "grad_norm": 0.179987, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:02.253858+00:00", "epoch": 0, "step": 1980, "train_loss": 3.8358139991760254, "perplexity": 46.33112582097874, "lr": 0.0026291804804649314, "grad_norm": 0.177251, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:02.557746+00:00", "epoch": 0, "step": 1981, "train_loss": 3.8115577697753906, "perplexity": 45.22082766586846, "lr": 0.0026291804804649314, "grad_norm": 0.174857, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:02.861357+00:00", "epoch": 0, "step": 1982, "train_loss": 3.7795846462249756, "perplexity": 43.7978463562538, "lr": 0.0026291804804649314, "grad_norm": 0.190024, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:03.165322+00:00", "epoch": 0, "step": 1983, "train_loss": 3.835876941680908, "perplexity": 46.33404210987025, "lr": 0.0026291804804649314, "grad_norm": 0.178525, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:03.469493+00:00", "epoch": 0, "step": 1984, "train_loss": 3.759838581085205, "perplexity": 42.94149384995279, "lr": 0.0026291804804649314, "grad_norm": 0.192359, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:03.774377+00:00", "epoch": 0, "step": 1985, "train_loss": 3.769331932067871, "perplexity": 43.351093684067365, "lr": 0.0026291804804649314, "grad_norm": 0.174399, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:04.078278+00:00", "epoch": 0, "step": 1986, "train_loss": 3.798790216445923, "perplexity": 44.64713843406159, "lr": 0.0026291804804649314, "grad_norm": 0.163357, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:04.382718+00:00", "epoch": 0, "step": 1987, "train_loss": 3.8420114517211914, "perplexity": 46.61915236816091, "lr": 0.0026291804804649314, "grad_norm": 0.177644, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:04.687614+00:00", "epoch": 0, "step": 1988, "train_loss": 3.802852153778076, "perplexity": 44.82886113569508, "lr": 0.0026291804804649314, "grad_norm": 0.170456, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:04.993316+00:00", "epoch": 0, "step": 1989, "train_loss": 3.854793071746826, "perplexity": 47.21884501714972, "lr": 0.0026291804804649314, "grad_norm": 0.183803, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:05.298536+00:00", "epoch": 0, "step": 1990, "train_loss": 3.8780293464660645, "perplexity": 48.32888168617504, "lr": 0.0026291804804649314, "grad_norm": 0.207147, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:05.603051+00:00", "epoch": 0, "step": 1991, "train_loss": 3.752187728881836, "perplexity": 42.61420842978808, "lr": 0.0026291804804649314, "grad_norm": 0.19759, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:05.908006+00:00", "epoch": 0, "step": 1992, "train_loss": 3.711399793624878, "perplexity": 40.9110334685284, "lr": 0.0026291804804649314, "grad_norm": 0.174259, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:06.213527+00:00", "epoch": 0, "step": 1993, "train_loss": 3.7483019828796387, "perplexity": 42.44894173988209, "lr": 0.0026291804804649314, "grad_norm": 0.181626, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:06.518379+00:00", "epoch": 0, "step": 1994, "train_loss": 3.8367373943328857, "perplexity": 46.373927516571655, "lr": 0.0026291804804649314, "grad_norm": 0.188171, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:06.823330+00:00", "epoch": 0, "step": 1995, "train_loss": 3.8549184799194336, "perplexity": 47.22476701754181, "lr": 0.0026291804804649314, "grad_norm": 0.185341, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:07.129213+00:00", "epoch": 0, "step": 1996, "train_loss": 3.8122198581695557, "perplexity": 45.25077776475293, "lr": 0.0026291804804649314, "grad_norm": 0.201383, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:07.433923+00:00", "epoch": 0, "step": 1997, "train_loss": 3.817434787750244, "perplexity": 45.487373763809465, "lr": 0.0026291804804649314, "grad_norm": 0.187246, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:07.738000+00:00", "epoch": 0, "step": 1998, "train_loss": 3.79754900932312, "perplexity": 44.59175646518169, "lr": 0.0026291804804649314, "grad_norm": 0.18801, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:08.043757+00:00", "epoch": 0, "step": 1999, "train_loss": 3.737569570541382, "perplexity": 41.99579820503012, "lr": 0.0026291804804649314, "grad_norm": 0.175437, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:08.349918+00:00", "epoch": 0, "step": 2000, "train_loss": 3.823774814605713, "perplexity": 45.776681074033945, "lr": 0.0026291804804649314, "grad_norm": 0.175056, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:08:11.423648+00:00", "step": 2000, "epoch": 0, "val_loss": 3.754971480369568, "val_ppl": 42.73300106377579, "eval_train_loss": 3.823774814605713, "eval_train_ppl": 45.776681074033945} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:08:12.287341+00:00", "step": 2000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p7550_epoch_0000_step_0002000.pt", "val_loss": 3.754971480369568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:13.352779+00:00", "epoch": 0, "step": 2001, "train_loss": 3.8637874126434326, "perplexity": 47.645463104183776, "lr": 0.0026291804804649314, "grad_norm": 0.192298, "tokens_per_sec": 6550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:13.656213+00:00", "epoch": 0, "step": 2002, "train_loss": 3.8578479290008545, "perplexity": 47.3633123996436, "lr": 0.0026291804804649314, "grad_norm": 0.166324, "tokens_per_sec": 107990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:13.959360+00:00", "epoch": 0, "step": 2003, "train_loss": 3.729264259338379, "perplexity": 41.648454427395926, "lr": 0.0026291804804649314, "grad_norm": 0.181935, "tokens_per_sec": 108092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:14.264164+00:00", "epoch": 0, "step": 2004, "train_loss": 3.8423192501068115, "perplexity": 46.63350387657106, "lr": 0.0026291804804649314, "grad_norm": 0.145831, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:14.568915+00:00", "epoch": 0, "step": 2005, "train_loss": 3.8998208045959473, "perplexity": 49.39359720683498, "lr": 0.0026291804804649314, "grad_norm": 0.159103, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:14.873565+00:00", "epoch": 0, "step": 2006, "train_loss": 3.781935691833496, "perplexity": 43.90093822995483, "lr": 0.0026291804804649314, "grad_norm": 0.148841, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:15.177290+00:00", "epoch": 0, "step": 2007, "train_loss": 3.888788938522339, "perplexity": 48.851688286812234, "lr": 0.0026291804804649314, "grad_norm": 0.162356, "tokens_per_sec": 107888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:15.480630+00:00", "epoch": 0, "step": 2008, "train_loss": 3.771831750869751, "perplexity": 43.45959912861099, "lr": 0.0026291804804649314, "grad_norm": 0.148738, "tokens_per_sec": 108024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:15.785687+00:00", "epoch": 0, "step": 2009, "train_loss": 3.8109278678894043, "perplexity": 45.19235195063078, "lr": 0.0026291804804649314, "grad_norm": 0.157497, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:16.090907+00:00", "epoch": 0, "step": 2010, "train_loss": 3.737699508666992, "perplexity": 42.00125541487453, "lr": 0.0026291804804649314, "grad_norm": 0.159417, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:16.396366+00:00", "epoch": 0, "step": 2011, "train_loss": 3.8640921115875244, "perplexity": 47.659982838443895, "lr": 0.0026291804804649314, "grad_norm": 0.165273, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:16.699511+00:00", "epoch": 0, "step": 2012, "train_loss": 3.8767616748809814, "perplexity": 48.26765535176401, "lr": 0.0026291804804649314, "grad_norm": 0.167515, "tokens_per_sec": 108095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:17.003133+00:00", "epoch": 0, "step": 2013, "train_loss": 3.860081434249878, "perplexity": 47.469216831525834, "lr": 0.0026291804804649314, "grad_norm": 0.173145, "tokens_per_sec": 107924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:17.306919+00:00", "epoch": 0, "step": 2014, "train_loss": 3.770001173019409, "perplexity": 43.38011572153949, "lr": 0.0026291804804649314, "grad_norm": 0.162117, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:17.711896+00:00", "epoch": 0, "step": 2015, "train_loss": 3.812037467956543, "perplexity": 45.24252521837215, "lr": 0.0026291804804649314, "grad_norm": 0.150111, "tokens_per_sec": 80914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:18.015552+00:00", "epoch": 0, "step": 2016, "train_loss": 3.829326868057251, "perplexity": 46.03154250164413, "lr": 0.0026291804804649314, "grad_norm": 0.150407, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:18.319424+00:00", "epoch": 0, "step": 2017, "train_loss": 3.75516676902771, "perplexity": 42.7413471491333, "lr": 0.0026291804804649314, "grad_norm": 0.153096, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:18.623961+00:00", "epoch": 0, "step": 2018, "train_loss": 3.749331474304199, "perplexity": 42.49266506391685, "lr": 0.0026291804804649314, "grad_norm": 0.182771, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:18.929118+00:00", "epoch": 0, "step": 2019, "train_loss": 3.729457139968872, "perplexity": 41.656488382317185, "lr": 0.0026291804804649314, "grad_norm": 0.184204, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:19.234034+00:00", "epoch": 0, "step": 2020, "train_loss": 3.815664052963257, "perplexity": 45.406898959512326, "lr": 0.0026291804804649314, "grad_norm": 0.195529, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:19.538085+00:00", "epoch": 0, "step": 2021, "train_loss": 3.7480709552764893, "perplexity": 42.439135995358136, "lr": 0.0026291804804649314, "grad_norm": 0.20583, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:19.842800+00:00", "epoch": 0, "step": 2022, "train_loss": 3.8232953548431396, "perplexity": 45.7547382581638, "lr": 0.0026291804804649314, "grad_norm": 0.199054, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:20.147821+00:00", "epoch": 0, "step": 2023, "train_loss": 3.762380599975586, "perplexity": 43.05079079709757, "lr": 0.0026291804804649314, "grad_norm": 0.24661, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:20.452419+00:00", "epoch": 0, "step": 2024, "train_loss": 3.818950891494751, "perplexity": 45.55638964589937, "lr": 0.0026291804804649314, "grad_norm": 0.205731, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:20.756692+00:00", "epoch": 0, "step": 2025, "train_loss": 3.883289098739624, "perplexity": 48.583749314219844, "lr": 0.0026291804804649314, "grad_norm": 0.168936, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:21.060569+00:00", "epoch": 0, "step": 2026, "train_loss": 3.734145164489746, "perplexity": 41.85223349174332, "lr": 0.0026291804804649314, "grad_norm": 0.177551, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:21.364352+00:00", "epoch": 0, "step": 2027, "train_loss": 3.770387649536133, "perplexity": 43.39688435769194, "lr": 0.0026291804804649314, "grad_norm": 0.176143, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:21.669582+00:00", "epoch": 0, "step": 2028, "train_loss": 3.9080514907836914, "perplexity": 49.80181806948929, "lr": 0.0026291804804649314, "grad_norm": 0.187699, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:21.974920+00:00", "epoch": 0, "step": 2029, "train_loss": 3.707874298095703, "perplexity": 40.767055748471684, "lr": 0.0026291804804649314, "grad_norm": 0.171237, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:22.279701+00:00", "epoch": 0, "step": 2030, "train_loss": 3.8374719619750977, "perplexity": 46.40800481767666, "lr": 0.0026291804804649314, "grad_norm": 0.163673, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:22.583835+00:00", "epoch": 0, "step": 2031, "train_loss": 3.854384183883667, "perplexity": 47.19954175121433, "lr": 0.0026291804804649314, "grad_norm": 0.16608, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:22.888894+00:00", "epoch": 0, "step": 2032, "train_loss": 3.7420716285705566, "perplexity": 42.18529196141122, "lr": 0.0026291804804649314, "grad_norm": 0.170452, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:23.194849+00:00", "epoch": 0, "step": 2033, "train_loss": 3.720641851425171, "perplexity": 41.29088822051039, "lr": 0.0026291804804649314, "grad_norm": 0.153363, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:23.499815+00:00", "epoch": 0, "step": 2034, "train_loss": 3.7307918071746826, "perplexity": 41.71212304989478, "lr": 0.0026291804804649314, "grad_norm": 0.14558, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:23.804492+00:00", "epoch": 0, "step": 2035, "train_loss": 3.853522777557373, "perplexity": 47.15890127384682, "lr": 0.0026291804804649314, "grad_norm": 0.170741, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:24.109518+00:00", "epoch": 0, "step": 2036, "train_loss": 3.801130771636963, "perplexity": 44.75175991410118, "lr": 0.0026291804804649314, "grad_norm": 0.18909, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:24.414240+00:00", "epoch": 0, "step": 2037, "train_loss": 3.849480152130127, "perplexity": 46.96864033640456, "lr": 0.0026291804804649314, "grad_norm": 0.181503, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:24.719018+00:00", "epoch": 0, "step": 2038, "train_loss": 3.7798972129821777, "perplexity": 43.81153824676525, "lr": 0.0026291804804649314, "grad_norm": 0.161083, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:25.023829+00:00", "epoch": 0, "step": 2039, "train_loss": 3.753981351852417, "perplexity": 42.690710840632015, "lr": 0.0026291804804649314, "grad_norm": 0.145546, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:25.327328+00:00", "epoch": 0, "step": 2040, "train_loss": 3.882002592086792, "perplexity": 48.52128618573797, "lr": 0.0026291804804649314, "grad_norm": 0.162494, "tokens_per_sec": 108022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:25.631217+00:00", "epoch": 0, "step": 2041, "train_loss": 3.8190815448760986, "perplexity": 45.56234213109629, "lr": 0.0026291804804649314, "grad_norm": 0.164592, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:25.936724+00:00", "epoch": 0, "step": 2042, "train_loss": 3.751326560974121, "perplexity": 42.57752623810579, "lr": 0.0026291804804649314, "grad_norm": 0.161476, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:26.242652+00:00", "epoch": 0, "step": 2043, "train_loss": 3.8377184867858887, "perplexity": 46.41944695261065, "lr": 0.0026291804804649314, "grad_norm": 0.162692, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:26.547091+00:00", "epoch": 0, "step": 2044, "train_loss": 3.7305922508239746, "perplexity": 41.70379996132887, "lr": 0.0026291804804649314, "grad_norm": 0.161237, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:26.851774+00:00", "epoch": 0, "step": 2045, "train_loss": 3.8232507705688477, "perplexity": 45.75269836183712, "lr": 0.0026291804804649314, "grad_norm": 0.14607, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:27.156002+00:00", "epoch": 0, "step": 2046, "train_loss": 3.8890421390533447, "perplexity": 48.86405912631246, "lr": 0.0026291804804649314, "grad_norm": 0.177943, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:27.460438+00:00", "epoch": 0, "step": 2047, "train_loss": 3.847412586212158, "perplexity": 46.87162989872757, "lr": 0.0026291804804649314, "grad_norm": 0.186172, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:27.765263+00:00", "epoch": 0, "step": 2048, "train_loss": 3.7717714309692383, "perplexity": 43.45697772897731, "lr": 0.0026291804804649314, "grad_norm": 0.173156, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:28.069760+00:00", "epoch": 0, "step": 2049, "train_loss": 3.821617364883423, "perplexity": 45.67802664541454, "lr": 0.0026291804804649314, "grad_norm": 0.158183, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:28.373415+00:00", "epoch": 0, "step": 2050, "train_loss": 3.8607709407806396, "perplexity": 47.501958453025075, "lr": 0.0026291804804649314, "grad_norm": 0.16595, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:28.678864+00:00", "epoch": 0, "step": 2051, "train_loss": 3.906621217727661, "perplexity": 49.73063878600302, "lr": 0.0026291804804649314, "grad_norm": 0.174501, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:28.984948+00:00", "epoch": 0, "step": 2052, "train_loss": 3.880469560623169, "perplexity": 48.446958515245505, "lr": 0.0026291804804649314, "grad_norm": 0.157602, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:29.290878+00:00", "epoch": 0, "step": 2053, "train_loss": 3.7887966632843018, "perplexity": 44.20317695383784, "lr": 0.0026291804804649314, "grad_norm": 0.167032, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:29.594760+00:00", "epoch": 0, "step": 2054, "train_loss": 3.771287441253662, "perplexity": 43.435950087677995, "lr": 0.0026291804804649314, "grad_norm": 0.167464, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:29.899111+00:00", "epoch": 0, "step": 2055, "train_loss": 3.716836452484131, "perplexity": 41.13405850726881, "lr": 0.0026291804804649314, "grad_norm": 0.17622, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:30.204493+00:00", "epoch": 0, "step": 2056, "train_loss": 3.766265392303467, "perplexity": 43.21835945286947, "lr": 0.0026291804804649314, "grad_norm": 0.173444, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:30.508469+00:00", "epoch": 0, "step": 2057, "train_loss": 3.7271525859832764, "perplexity": 41.56059928922429, "lr": 0.0026291804804649314, "grad_norm": 0.17828, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:30.812560+00:00", "epoch": 0, "step": 2058, "train_loss": 3.8053672313690186, "perplexity": 44.9417511037712, "lr": 0.0026291804804649314, "grad_norm": 0.186202, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:31.116814+00:00", "epoch": 0, "step": 2059, "train_loss": 3.799528121948242, "perplexity": 44.68009596144983, "lr": 0.0026291804804649314, "grad_norm": 0.176689, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:31.421023+00:00", "epoch": 0, "step": 2060, "train_loss": 3.7606241703033447, "perplexity": 42.975241478683145, "lr": 0.0026291804804649314, "grad_norm": 0.167797, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:31.726122+00:00", "epoch": 0, "step": 2061, "train_loss": 3.83848237991333, "perplexity": 46.454919996199514, "lr": 0.0026291804804649314, "grad_norm": 0.167532, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:32.031510+00:00", "epoch": 0, "step": 2062, "train_loss": 3.798171043395996, "perplexity": 44.61950268572506, "lr": 0.0026291804804649314, "grad_norm": 0.176119, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:32.336243+00:00", "epoch": 0, "step": 2063, "train_loss": 3.8568880558013916, "perplexity": 47.3178714377031, "lr": 0.0026291804804649314, "grad_norm": 0.159676, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:32.640534+00:00", "epoch": 0, "step": 2064, "train_loss": 3.7969400882720947, "perplexity": 44.56461187126391, "lr": 0.0026291804804649314, "grad_norm": 0.17516, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:32.944646+00:00", "epoch": 0, "step": 2065, "train_loss": 3.944549798965454, "perplexity": 51.65307860734683, "lr": 0.0026291804804649314, "grad_norm": 0.174025, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:33.248937+00:00", "epoch": 0, "step": 2066, "train_loss": 3.8837153911590576, "perplexity": 48.60446461333394, "lr": 0.0026291804804649314, "grad_norm": 0.149747, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:33.553678+00:00", "epoch": 0, "step": 2067, "train_loss": 3.7137033939361572, "perplexity": 41.00538477008156, "lr": 0.0026291804804649314, "grad_norm": 0.196658, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:33.859571+00:00", "epoch": 0, "step": 2068, "train_loss": 3.8771471977233887, "perplexity": 48.286267222871764, "lr": 0.0026291804804649314, "grad_norm": 0.205179, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:34.163103+00:00", "epoch": 0, "step": 2069, "train_loss": 3.782454490661621, "perplexity": 43.92371989430153, "lr": 0.0026291804804649314, "grad_norm": 0.188564, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:34.467491+00:00", "epoch": 0, "step": 2070, "train_loss": 3.8014323711395264, "perplexity": 44.765259058195355, "lr": 0.0026291804804649314, "grad_norm": 0.18816, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:34.771509+00:00", "epoch": 0, "step": 2071, "train_loss": 3.8705058097839355, "perplexity": 47.96664194267312, "lr": 0.0026291804804649314, "grad_norm": 0.183119, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:35.075800+00:00", "epoch": 0, "step": 2072, "train_loss": 3.732083320617676, "perplexity": 41.76602962058682, "lr": 0.0026291804804649314, "grad_norm": 0.165091, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:35.381752+00:00", "epoch": 0, "step": 2073, "train_loss": 3.7516441345214844, "perplexity": 42.59104988141321, "lr": 0.0026291804804649314, "grad_norm": 0.174932, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:35.685781+00:00", "epoch": 0, "step": 2074, "train_loss": 3.7539167404174805, "perplexity": 42.687952621653345, "lr": 0.0026291804804649314, "grad_norm": 0.187424, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:35.991048+00:00", "epoch": 0, "step": 2075, "train_loss": 3.8796229362487793, "perplexity": 48.4059594971344, "lr": 0.0026291804804649314, "grad_norm": 0.206776, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:36.295755+00:00", "epoch": 0, "step": 2076, "train_loss": 3.865720748901367, "perplexity": 47.73766690728714, "lr": 0.0026291804804649314, "grad_norm": 0.208475, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:36.600440+00:00", "epoch": 0, "step": 2077, "train_loss": 3.916144371032715, "perplexity": 50.2064935055586, "lr": 0.0026291804804649314, "grad_norm": 0.171821, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:36.905623+00:00", "epoch": 0, "step": 2078, "train_loss": 3.775404930114746, "perplexity": 43.615165834557715, "lr": 0.0026291804804649314, "grad_norm": 0.18369, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:37.210687+00:00", "epoch": 0, "step": 2079, "train_loss": 3.7781929969787598, "perplexity": 43.73693750809268, "lr": 0.0026291804804649314, "grad_norm": 0.181127, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:37.515307+00:00", "epoch": 0, "step": 2080, "train_loss": 3.7938270568847656, "perplexity": 44.42609654887986, "lr": 0.0026291804804649314, "grad_norm": 0.173846, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:37.819765+00:00", "epoch": 0, "step": 2081, "train_loss": 3.812340497970581, "perplexity": 45.256237138881, "lr": 0.0026291804804649314, "grad_norm": 0.16723, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:38.125228+00:00", "epoch": 0, "step": 2082, "train_loss": 3.8193066120147705, "perplexity": 45.57259787114266, "lr": 0.0026291804804649314, "grad_norm": 0.173449, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:38.430365+00:00", "epoch": 0, "step": 2083, "train_loss": 3.7050559520721436, "perplexity": 40.65232183489029, "lr": 0.0026291804804649314, "grad_norm": 0.165009, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:38.735937+00:00", "epoch": 0, "step": 2084, "train_loss": 3.749563694000244, "perplexity": 42.50253384350028, "lr": 0.0026291804804649314, "grad_norm": 0.150477, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:39.039253+00:00", "epoch": 0, "step": 2085, "train_loss": 3.7868568897247314, "perplexity": 44.117515908309265, "lr": 0.0026291804804649314, "grad_norm": 0.159518, "tokens_per_sec": 108031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:39.343333+00:00", "epoch": 0, "step": 2086, "train_loss": 3.7011866569519043, "perplexity": 40.49532992427125, "lr": 0.0026291804804649314, "grad_norm": 0.16315, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:39.648084+00:00", "epoch": 0, "step": 2087, "train_loss": 3.751051425933838, "perplexity": 42.56581328010559, "lr": 0.0026291804804649314, "grad_norm": 0.160342, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:39.953606+00:00", "epoch": 0, "step": 2088, "train_loss": 3.7998673915863037, "perplexity": 44.695257133152424, "lr": 0.0026291804804649314, "grad_norm": 0.172547, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:40.259503+00:00", "epoch": 0, "step": 2089, "train_loss": 3.726611375808716, "perplexity": 41.538112355655635, "lr": 0.0026291804804649314, "grad_norm": 0.167546, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:40.563637+00:00", "epoch": 0, "step": 2090, "train_loss": 3.8426389694213867, "perplexity": 46.648415892169396, "lr": 0.0026291804804649314, "grad_norm": 0.157082, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:40.867664+00:00", "epoch": 0, "step": 2091, "train_loss": 3.7720112800598145, "perplexity": 43.46740209565225, "lr": 0.0026291804804649314, "grad_norm": 0.16977, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:41.172977+00:00", "epoch": 0, "step": 2092, "train_loss": 3.9218692779541016, "perplexity": 50.4947453278767, "lr": 0.0026291804804649314, "grad_norm": 0.165372, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:41.477993+00:00", "epoch": 0, "step": 2093, "train_loss": 3.7615344524383545, "perplexity": 43.01437888358669, "lr": 0.0026291804804649314, "grad_norm": 0.172066, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:41.783671+00:00", "epoch": 0, "step": 2094, "train_loss": 3.8837928771972656, "perplexity": 48.60823092665251, "lr": 0.0026291804804649314, "grad_norm": 0.180431, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:42.089109+00:00", "epoch": 0, "step": 2095, "train_loss": 3.7917191982269287, "perplexity": 44.33255124141421, "lr": 0.0026291804804649314, "grad_norm": 0.17424, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:42.393722+00:00", "epoch": 0, "step": 2096, "train_loss": 3.7245750427246094, "perplexity": 41.45361298684375, "lr": 0.0026291804804649314, "grad_norm": 0.159537, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:42.697949+00:00", "epoch": 0, "step": 2097, "train_loss": 3.8293540477752686, "perplexity": 46.032793642992, "lr": 0.0026291804804649314, "grad_norm": 0.16492, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:43.002375+00:00", "epoch": 0, "step": 2098, "train_loss": 3.7141001224517822, "perplexity": 41.02165600293164, "lr": 0.0026291804804649314, "grad_norm": 0.163182, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:43.307495+00:00", "epoch": 0, "step": 2099, "train_loss": 3.7145020961761475, "perplexity": 41.03814894541745, "lr": 0.0026291804804649314, "grad_norm": 0.168612, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:43.613663+00:00", "epoch": 0, "step": 2100, "train_loss": 3.9352080821990967, "perplexity": 51.17279699702223, "lr": 0.0026291804804649314, "grad_norm": 0.161898, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:43.919889+00:00", "epoch": 0, "step": 2101, "train_loss": 3.7762629985809326, "perplexity": 43.65260669412106, "lr": 0.0026291804804649314, "grad_norm": 0.157757, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:44.224403+00:00", "epoch": 0, "step": 2102, "train_loss": 3.7996814250946045, "perplexity": 44.68694608579988, "lr": 0.0026291804804649314, "grad_norm": 0.183225, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:44.528542+00:00", "epoch": 0, "step": 2103, "train_loss": 3.804603338241577, "perplexity": 44.90743351812055, "lr": 0.0026291804804649314, "grad_norm": 0.193137, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:44.832845+00:00", "epoch": 0, "step": 2104, "train_loss": 3.7056291103363037, "perplexity": 40.675628727738335, "lr": 0.0026291804804649314, "grad_norm": 0.170013, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:45.138093+00:00", "epoch": 0, "step": 2105, "train_loss": 3.7733914852142334, "perplexity": 43.52743744909556, "lr": 0.0026291804804649314, "grad_norm": 0.184451, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:45.442582+00:00", "epoch": 0, "step": 2106, "train_loss": 3.8071794509887695, "perplexity": 45.023269268966835, "lr": 0.0026291804804649314, "grad_norm": 0.169685, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:45.746428+00:00", "epoch": 0, "step": 2107, "train_loss": 3.9421463012695312, "perplexity": 51.52907962725134, "lr": 0.0026291804804649314, "grad_norm": 0.155024, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:46.051338+00:00", "epoch": 0, "step": 2108, "train_loss": 3.9119131565093994, "perplexity": 49.994507855721245, "lr": 0.0026291804804649314, "grad_norm": 0.157569, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:46.355964+00:00", "epoch": 0, "step": 2109, "train_loss": 3.7299747467041016, "perplexity": 41.67805564246792, "lr": 0.0026291804804649314, "grad_norm": 0.169101, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:46.661483+00:00", "epoch": 0, "step": 2110, "train_loss": 3.764249324798584, "perplexity": 43.13131609488748, "lr": 0.0026291804804649314, "grad_norm": 0.168692, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:46.966829+00:00", "epoch": 0, "step": 2111, "train_loss": 3.878629684448242, "perplexity": 48.357904060232514, "lr": 0.0026291804804649314, "grad_norm": 0.179521, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:47.271204+00:00", "epoch": 0, "step": 2112, "train_loss": 3.760497570037842, "perplexity": 42.96980114608293, "lr": 0.0026291804804649314, "grad_norm": 0.182601, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:47.575515+00:00", "epoch": 0, "step": 2113, "train_loss": 3.793134927749634, "perplexity": 44.39535859164315, "lr": 0.0026291804804649314, "grad_norm": 0.165623, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:47.879666+00:00", "epoch": 0, "step": 2114, "train_loss": 3.8449857234954834, "perplexity": 46.75801680510898, "lr": 0.0026291804804649314, "grad_norm": 0.166079, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:48.184653+00:00", "epoch": 0, "step": 2115, "train_loss": 3.761122465133667, "perplexity": 42.996661155557746, "lr": 0.0026291804804649314, "grad_norm": 0.176108, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:48.489565+00:00", "epoch": 0, "step": 2116, "train_loss": 3.73642897605896, "perplexity": 41.94792533626759, "lr": 0.0026291804804649314, "grad_norm": 0.182076, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:48.794752+00:00", "epoch": 0, "step": 2117, "train_loss": 3.8085944652557373, "perplexity": 45.087022932862304, "lr": 0.0026291804804649314, "grad_norm": 0.18, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:49.099188+00:00", "epoch": 0, "step": 2118, "train_loss": 3.7741332054138184, "perplexity": 43.559734604933475, "lr": 0.0026291804804649314, "grad_norm": 0.186129, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:49.403283+00:00", "epoch": 0, "step": 2119, "train_loss": 3.8588173389434814, "perplexity": 47.4092491277626, "lr": 0.0026291804804649314, "grad_norm": 0.179639, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:49.708690+00:00", "epoch": 0, "step": 2120, "train_loss": 3.77569842338562, "perplexity": 43.62796847089012, "lr": 0.0026291804804649314, "grad_norm": 0.173308, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:50.013126+00:00", "epoch": 0, "step": 2121, "train_loss": 3.7214603424072266, "perplexity": 41.324698274884724, "lr": 0.0026291804804649314, "grad_norm": 0.164399, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:50.318268+00:00", "epoch": 0, "step": 2122, "train_loss": 3.9162604808807373, "perplexity": 50.21232331233171, "lr": 0.0026291804804649314, "grad_norm": 0.166293, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:50.622053+00:00", "epoch": 0, "step": 2123, "train_loss": 3.8563432693481445, "perplexity": 47.29210032286186, "lr": 0.0026291804804649314, "grad_norm": 0.157188, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:50.926628+00:00", "epoch": 0, "step": 2124, "train_loss": 3.735812187194824, "perplexity": 41.92206030049958, "lr": 0.0026291804804649314, "grad_norm": 0.164722, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:51.231066+00:00", "epoch": 0, "step": 2125, "train_loss": 3.7520673274993896, "perplexity": 42.60907792904717, "lr": 0.0026291804804649314, "grad_norm": 0.152635, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:51.536321+00:00", "epoch": 0, "step": 2126, "train_loss": 3.8132758140563965, "perplexity": 45.29858582707784, "lr": 0.0026291804804649314, "grad_norm": 0.140257, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:51.840529+00:00", "epoch": 0, "step": 2127, "train_loss": 3.834861993789673, "perplexity": 46.28703932826034, "lr": 0.0026291804804649314, "grad_norm": 0.146926, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:52.145296+00:00", "epoch": 0, "step": 2128, "train_loss": 3.741642475128174, "perplexity": 42.167191882281486, "lr": 0.0026291804804649314, "grad_norm": 0.15602, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:52.450381+00:00", "epoch": 0, "step": 2129, "train_loss": 3.6572585105895996, "perplexity": 38.754950814360335, "lr": 0.0026291804804649314, "grad_norm": 0.157642, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:52.755195+00:00", "epoch": 0, "step": 2130, "train_loss": 3.734722852706909, "perplexity": 41.876418018779034, "lr": 0.0026291804804649314, "grad_norm": 0.15094, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:53.060817+00:00", "epoch": 0, "step": 2131, "train_loss": 3.8028132915496826, "perplexity": 44.827119020106466, "lr": 0.0026291804804649314, "grad_norm": 0.151661, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:53.365590+00:00", "epoch": 0, "step": 2132, "train_loss": 3.798818349838257, "perplexity": 44.648394527192735, "lr": 0.0026291804804649314, "grad_norm": 0.168801, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:53.669899+00:00", "epoch": 0, "step": 2133, "train_loss": 3.725090742111206, "perplexity": 41.47499610278928, "lr": 0.0026291804804649314, "grad_norm": 0.19271, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:53.973427+00:00", "epoch": 0, "step": 2134, "train_loss": 3.8047499656677246, "perplexity": 44.91401866228177, "lr": 0.0026291804804649314, "grad_norm": 0.181091, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:54.278540+00:00", "epoch": 0, "step": 2135, "train_loss": 3.801564931869507, "perplexity": 44.77119356694659, "lr": 0.0026291804804649314, "grad_norm": 0.163438, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:54.584104+00:00", "epoch": 0, "step": 2136, "train_loss": 3.8077659606933594, "perplexity": 45.049683598698614, "lr": 0.0026291804804649314, "grad_norm": 0.19358, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:54.889838+00:00", "epoch": 0, "step": 2137, "train_loss": 3.6483066082000732, "perplexity": 38.40956849842498, "lr": 0.0026291804804649314, "grad_norm": 0.177635, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:55.194396+00:00", "epoch": 0, "step": 2138, "train_loss": 3.803485631942749, "perplexity": 44.85726823706094, "lr": 0.0026291804804649314, "grad_norm": 0.187553, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:55.498188+00:00", "epoch": 0, "step": 2139, "train_loss": 3.758568286895752, "perplexity": 42.88698015137652, "lr": 0.0026291804804649314, "grad_norm": 0.181356, "tokens_per_sec": 107862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:55.802748+00:00", "epoch": 0, "step": 2140, "train_loss": 3.9007623195648193, "perplexity": 49.44012391733188, "lr": 0.0026291804804649314, "grad_norm": 0.177723, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:56.106733+00:00", "epoch": 0, "step": 2141, "train_loss": 3.7076048851013184, "perplexity": 40.75607405328235, "lr": 0.0026291804804649314, "grad_norm": 0.180351, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:56.412169+00:00", "epoch": 0, "step": 2142, "train_loss": 3.812821626663208, "perplexity": 45.27801645199493, "lr": 0.0026291804804649314, "grad_norm": 0.174027, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:56.717106+00:00", "epoch": 0, "step": 2143, "train_loss": 3.857022285461426, "perplexity": 47.32422332579607, "lr": 0.0026291804804649314, "grad_norm": 0.165784, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:57.021430+00:00", "epoch": 0, "step": 2144, "train_loss": 3.7653844356536865, "perplexity": 43.180302717345256, "lr": 0.0026291804804649314, "grad_norm": 0.161923, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:57.325559+00:00", "epoch": 0, "step": 2145, "train_loss": 3.783843994140625, "perplexity": 43.984794477752054, "lr": 0.0026291804804649314, "grad_norm": 0.17588, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:57.629776+00:00", "epoch": 0, "step": 2146, "train_loss": 3.8439886569976807, "perplexity": 46.71141918737106, "lr": 0.0026291804804649314, "grad_norm": 0.152233, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:57.935437+00:00", "epoch": 0, "step": 2147, "train_loss": 3.9233639240264893, "perplexity": 50.57027353056235, "lr": 0.0026291804804649314, "grad_norm": 0.167043, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:58.240807+00:00", "epoch": 0, "step": 2148, "train_loss": 3.7935054302215576, "perplexity": 44.411810229241304, "lr": 0.0026291804804649314, "grad_norm": 0.144902, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:58.544977+00:00", "epoch": 0, "step": 2149, "train_loss": 3.82704758644104, "perplexity": 45.92674313206302, "lr": 0.0026291804804649314, "grad_norm": 0.15974, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:58.848889+00:00", "epoch": 0, "step": 2150, "train_loss": 3.7256696224212646, "perplexity": 41.49901211191969, "lr": 0.0026291804804649314, "grad_norm": 0.171915, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:59.153922+00:00", "epoch": 0, "step": 2151, "train_loss": 3.839890718460083, "perplexity": 46.52039034210113, "lr": 0.0026291804804649314, "grad_norm": 0.187918, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:59.458588+00:00", "epoch": 0, "step": 2152, "train_loss": 3.846757173538208, "perplexity": 46.840919703468174, "lr": 0.0026291804804649314, "grad_norm": 0.172451, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:08:59.764443+00:00", "epoch": 0, "step": 2153, "train_loss": 3.8826043605804443, "perplexity": 48.5504935541917, "lr": 0.0026291804804649314, "grad_norm": 0.162005, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:00.069342+00:00", "epoch": 0, "step": 2154, "train_loss": 3.8435118198394775, "perplexity": 46.68915075661967, "lr": 0.0026291804804649314, "grad_norm": 0.172432, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:00.374789+00:00", "epoch": 0, "step": 2155, "train_loss": 3.7824933528900146, "perplexity": 43.925426901104785, "lr": 0.0026291804804649314, "grad_norm": 0.186617, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:00.679018+00:00", "epoch": 0, "step": 2156, "train_loss": 3.8606395721435547, "perplexity": 47.495718595354006, "lr": 0.0026291804804649314, "grad_norm": 0.197667, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:00.984299+00:00", "epoch": 0, "step": 2157, "train_loss": 3.777878522872925, "perplexity": 43.72318553621073, "lr": 0.0026291804804649314, "grad_norm": 0.197067, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:01.290193+00:00", "epoch": 0, "step": 2158, "train_loss": 3.7766995429992676, "perplexity": 43.671667155985574, "lr": 0.0026291804804649314, "grad_norm": 0.17548, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:01.594639+00:00", "epoch": 0, "step": 2159, "train_loss": 3.716163396835327, "perplexity": 41.10638231168694, "lr": 0.0026291804804649314, "grad_norm": 0.163113, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:01.898941+00:00", "epoch": 0, "step": 2160, "train_loss": 3.7436835765838623, "perplexity": 42.25334729507297, "lr": 0.0026291804804649314, "grad_norm": 0.184116, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:02.203316+00:00", "epoch": 0, "step": 2161, "train_loss": 3.9147443771362305, "perplexity": 50.136253900068795, "lr": 0.0026291804804649314, "grad_norm": 0.1805, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:02.507634+00:00", "epoch": 0, "step": 2162, "train_loss": 3.7844080924987793, "perplexity": 44.00961322755059, "lr": 0.0026291804804649314, "grad_norm": 0.167703, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:02.812599+00:00", "epoch": 0, "step": 2163, "train_loss": 3.820463180541992, "perplexity": 45.62533619540838, "lr": 0.0026291804804649314, "grad_norm": 0.165699, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:03.117629+00:00", "epoch": 0, "step": 2164, "train_loss": 3.8285727500915527, "perplexity": 45.996842374094115, "lr": 0.0026291804804649314, "grad_norm": 0.194946, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:03.421844+00:00", "epoch": 0, "step": 2165, "train_loss": 3.83870267868042, "perplexity": 46.46515508514727, "lr": 0.0026291804804649314, "grad_norm": 0.18151, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:03.725042+00:00", "epoch": 0, "step": 2166, "train_loss": 3.679452896118164, "perplexity": 39.62470930892477, "lr": 0.0026291804804649314, "grad_norm": 0.190659, "tokens_per_sec": 108075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:04.029605+00:00", "epoch": 0, "step": 2167, "train_loss": 3.799923896789551, "perplexity": 44.69778271909461, "lr": 0.0026291804804649314, "grad_norm": 0.20422, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:04.334598+00:00", "epoch": 0, "step": 2168, "train_loss": 3.8284800052642822, "perplexity": 45.992576602710315, "lr": 0.0026291804804649314, "grad_norm": 0.187409, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:04.639105+00:00", "epoch": 0, "step": 2169, "train_loss": 3.8261971473693848, "perplexity": 45.887701838743304, "lr": 0.0026291804804649314, "grad_norm": 0.150421, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:04.944202+00:00", "epoch": 0, "step": 2170, "train_loss": 3.666886329650879, "perplexity": 39.12987844004829, "lr": 0.0026291804804649314, "grad_norm": 0.163814, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:05.248526+00:00", "epoch": 0, "step": 2171, "train_loss": 3.8116047382354736, "perplexity": 45.22295166838778, "lr": 0.0026291804804649314, "grad_norm": 0.149958, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:05.552898+00:00", "epoch": 0, "step": 2172, "train_loss": 3.794938087463379, "perplexity": 44.47548273033283, "lr": 0.0026291804804649314, "grad_norm": 0.155433, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:05.857590+00:00", "epoch": 0, "step": 2173, "train_loss": 3.7711968421936035, "perplexity": 43.43201500968717, "lr": 0.0026291804804649314, "grad_norm": 0.148613, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:06.162376+00:00", "epoch": 0, "step": 2174, "train_loss": 3.7368216514587402, "perplexity": 41.964400489100875, "lr": 0.0026291804804649314, "grad_norm": 0.149794, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:06.467171+00:00", "epoch": 0, "step": 2175, "train_loss": 3.7482028007507324, "perplexity": 42.444731772250734, "lr": 0.0026291804804649314, "grad_norm": 0.157835, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:06.771470+00:00", "epoch": 0, "step": 2176, "train_loss": 3.7524449825286865, "perplexity": 42.62517250052741, "lr": 0.0026291804804649314, "grad_norm": 0.15353, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:07.076093+00:00", "epoch": 0, "step": 2177, "train_loss": 3.808317184448242, "perplexity": 45.07452289982564, "lr": 0.0026291804804649314, "grad_norm": 0.137633, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:07.381226+00:00", "epoch": 0, "step": 2178, "train_loss": 3.807265281677246, "perplexity": 45.027133813011524, "lr": 0.0026291804804649314, "grad_norm": 0.144358, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:07.686884+00:00", "epoch": 0, "step": 2179, "train_loss": 3.777146577835083, "perplexity": 43.691194276868885, "lr": 0.0026291804804649314, "grad_norm": 0.152511, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:07.993163+00:00", "epoch": 0, "step": 2180, "train_loss": 3.7600278854370117, "perplexity": 42.94962363108886, "lr": 0.0026291804804649314, "grad_norm": 0.157055, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:08.297687+00:00", "epoch": 0, "step": 2181, "train_loss": 3.6869683265686035, "perplexity": 39.923627899844014, "lr": 0.0026291804804649314, "grad_norm": 0.169623, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:08.604271+00:00", "epoch": 0, "step": 2182, "train_loss": 3.7508108615875244, "perplexity": 42.555574694627246, "lr": 0.0026291804804649314, "grad_norm": 0.180575, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:08.908541+00:00", "epoch": 0, "step": 2183, "train_loss": 3.7970986366271973, "perplexity": 44.57167807732475, "lr": 0.0026291804804649314, "grad_norm": 0.164285, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:09.213347+00:00", "epoch": 0, "step": 2184, "train_loss": 3.725008487701416, "perplexity": 41.47158474176547, "lr": 0.0026291804804649314, "grad_norm": 0.171772, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:09.518753+00:00", "epoch": 0, "step": 2185, "train_loss": 3.757967948913574, "perplexity": 42.86124119506034, "lr": 0.0026291804804649314, "grad_norm": 0.179128, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:09.824060+00:00", "epoch": 0, "step": 2186, "train_loss": 3.8632123470306396, "perplexity": 47.61807171342528, "lr": 0.0026291804804649314, "grad_norm": 0.183589, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:10.128267+00:00", "epoch": 0, "step": 2187, "train_loss": 3.813986301422119, "perplexity": 45.33078133588986, "lr": 0.0026291804804649314, "grad_norm": 0.183415, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:10.433014+00:00", "epoch": 0, "step": 2188, "train_loss": 3.7421491146087646, "perplexity": 42.18856085920129, "lr": 0.0026291804804649314, "grad_norm": 0.174426, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:10.737705+00:00", "epoch": 0, "step": 2189, "train_loss": 3.721773862838745, "perplexity": 41.337656443339384, "lr": 0.0026291804804649314, "grad_norm": 0.178778, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:11.042976+00:00", "epoch": 0, "step": 2190, "train_loss": 3.8084475994110107, "perplexity": 45.08040167538325, "lr": 0.0026291804804649314, "grad_norm": 0.194179, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:11.348126+00:00", "epoch": 0, "step": 2191, "train_loss": 3.7913742065429688, "perplexity": 44.3172595178176, "lr": 0.0026291804804649314, "grad_norm": 0.182262, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:11.652617+00:00", "epoch": 0, "step": 2192, "train_loss": 3.7424674034118652, "perplexity": 42.2019911429828, "lr": 0.0026291804804649314, "grad_norm": 0.189774, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:11.956394+00:00", "epoch": 0, "step": 2193, "train_loss": 3.80304217338562, "perplexity": 44.837380307673946, "lr": 0.0026291804804649314, "grad_norm": 0.163616, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:12.260753+00:00", "epoch": 0, "step": 2194, "train_loss": 3.759406328201294, "perplexity": 42.92293627646797, "lr": 0.0026291804804649314, "grad_norm": 0.176183, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:12.565492+00:00", "epoch": 0, "step": 2195, "train_loss": 3.7802517414093018, "perplexity": 43.82707343618018, "lr": 0.0026291804804649314, "grad_norm": 0.178788, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:12.870230+00:00", "epoch": 0, "step": 2196, "train_loss": 3.705056667327881, "perplexity": 40.65235091170712, "lr": 0.0026291804804649314, "grad_norm": 0.179487, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:13.174406+00:00", "epoch": 0, "step": 2197, "train_loss": 3.739560842514038, "perplexity": 42.07950657637221, "lr": 0.0026291804804649314, "grad_norm": 0.18353, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:13.479050+00:00", "epoch": 0, "step": 2198, "train_loss": 3.6897900104522705, "perplexity": 40.036438840824445, "lr": 0.0026291804804649314, "grad_norm": 0.168514, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:13.783456+00:00", "epoch": 0, "step": 2199, "train_loss": 3.7639355659484863, "perplexity": 43.11778538554726, "lr": 0.0026291804804649314, "grad_norm": 0.164078, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:14.089457+00:00", "epoch": 0, "step": 2200, "train_loss": 3.735713243484497, "perplexity": 41.91791258150763, "lr": 0.0026291804804649314, "grad_norm": 0.157741, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:14.395289+00:00", "epoch": 0, "step": 2201, "train_loss": 3.839545488357544, "perplexity": 46.50433287489318, "lr": 0.0026291804804649314, "grad_norm": 0.156606, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:14.700671+00:00", "epoch": 0, "step": 2202, "train_loss": 3.8869845867156982, "perplexity": 48.76362212985216, "lr": 0.0026291804804649314, "grad_norm": 0.186981, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:15.006320+00:00", "epoch": 0, "step": 2203, "train_loss": 3.7264633178710938, "perplexity": 41.531962763666726, "lr": 0.0026291804804649314, "grad_norm": 0.170001, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:15.310388+00:00", "epoch": 0, "step": 2204, "train_loss": 3.7419888973236084, "perplexity": 42.18180206396815, "lr": 0.0026291804804649314, "grad_norm": 0.18122, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:15.614768+00:00", "epoch": 0, "step": 2205, "train_loss": 3.760202646255493, "perplexity": 42.957130198375914, "lr": 0.0026291804804649314, "grad_norm": 0.161242, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:15.920635+00:00", "epoch": 0, "step": 2206, "train_loss": 3.8058712482452393, "perplexity": 44.96440821407776, "lr": 0.0026291804804649314, "grad_norm": 0.155429, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:16.226220+00:00", "epoch": 0, "step": 2207, "train_loss": 3.7860236167907715, "perplexity": 44.08076928849741, "lr": 0.0026291804804649314, "grad_norm": 0.173514, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:16.529861+00:00", "epoch": 0, "step": 2208, "train_loss": 3.783311367034912, "perplexity": 43.96137322192554, "lr": 0.0026291804804649314, "grad_norm": 0.158854, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:16.834056+00:00", "epoch": 0, "step": 2209, "train_loss": 3.827033042907715, "perplexity": 45.92607519980082, "lr": 0.0026291804804649314, "grad_norm": 0.177757, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:17.139388+00:00", "epoch": 0, "step": 2210, "train_loss": 3.742197275161743, "perplexity": 42.19059273254932, "lr": 0.0026291804804649314, "grad_norm": 0.161347, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:17.444687+00:00", "epoch": 0, "step": 2211, "train_loss": 3.749662399291992, "perplexity": 42.5067292755556, "lr": 0.0026291804804649314, "grad_norm": 0.17257, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:17.750131+00:00", "epoch": 0, "step": 2212, "train_loss": 3.7391512393951416, "perplexity": 42.0622742086936, "lr": 0.0026291804804649314, "grad_norm": 0.161614, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:18.055062+00:00", "epoch": 0, "step": 2213, "train_loss": 3.7028348445892334, "perplexity": 40.5621288598902, "lr": 0.0026291804804649314, "grad_norm": 0.151086, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:18.360908+00:00", "epoch": 0, "step": 2214, "train_loss": 3.81966495513916, "perplexity": 45.58893142458387, "lr": 0.0026291804804649314, "grad_norm": 0.145666, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:18.665847+00:00", "epoch": 0, "step": 2215, "train_loss": 3.670942544937134, "perplexity": 39.28891998644433, "lr": 0.0026291804804649314, "grad_norm": 0.138942, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:18.970731+00:00", "epoch": 0, "step": 2216, "train_loss": 3.8699631690979004, "perplexity": 47.940620352007976, "lr": 0.0026291804804649314, "grad_norm": 0.130271, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:19.275965+00:00", "epoch": 0, "step": 2217, "train_loss": 3.7281806468963623, "perplexity": 41.60334808729702, "lr": 0.0026291804804649314, "grad_norm": 0.131356, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:19.581446+00:00", "epoch": 0, "step": 2218, "train_loss": 3.787163257598877, "perplexity": 44.13103416854526, "lr": 0.0026291804804649314, "grad_norm": 0.131546, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:19.885922+00:00", "epoch": 0, "step": 2219, "train_loss": 3.6969733238220215, "perplexity": 40.32506854492671, "lr": 0.0026291804804649314, "grad_norm": 0.15897, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:20.190762+00:00", "epoch": 0, "step": 2220, "train_loss": 3.7314860820770264, "perplexity": 41.74109278536827, "lr": 0.0026291804804649314, "grad_norm": 0.172373, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:20.495122+00:00", "epoch": 0, "step": 2221, "train_loss": 3.845050811767578, "perplexity": 46.76106030267633, "lr": 0.0026291804804649314, "grad_norm": 0.171292, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:20.800503+00:00", "epoch": 0, "step": 2222, "train_loss": 3.8323020935058594, "perplexity": 46.16870065537185, "lr": 0.0026291804804649314, "grad_norm": 0.168789, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:21.105980+00:00", "epoch": 0, "step": 2223, "train_loss": 3.7227394580841064, "perplexity": 41.377591165141354, "lr": 0.0026291804804649314, "grad_norm": 0.182313, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:21.409918+00:00", "epoch": 0, "step": 2224, "train_loss": 3.712491035461426, "perplexity": 40.955701667296694, "lr": 0.0026291804804649314, "grad_norm": 0.179135, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:21.714262+00:00", "epoch": 0, "step": 2225, "train_loss": 3.794645071029663, "perplexity": 44.46245259211094, "lr": 0.0026291804804649314, "grad_norm": 0.162718, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:22.018913+00:00", "epoch": 0, "step": 2226, "train_loss": 3.7692031860351562, "perplexity": 43.34551276201019, "lr": 0.0026291804804649314, "grad_norm": 0.183045, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:22.325265+00:00", "epoch": 0, "step": 2227, "train_loss": 3.795055389404297, "perplexity": 44.48070009677798, "lr": 0.0026291804804649314, "grad_norm": 0.190409, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:22.630670+00:00", "epoch": 0, "step": 2228, "train_loss": 3.7243669033050537, "perplexity": 41.44498575376298, "lr": 0.0026291804804649314, "grad_norm": 0.182244, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:22.935672+00:00", "epoch": 0, "step": 2229, "train_loss": 3.830852508544922, "perplexity": 46.101823684871825, "lr": 0.0026291804804649314, "grad_norm": 0.171995, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:23.240658+00:00", "epoch": 0, "step": 2230, "train_loss": 3.898594617843628, "perplexity": 49.33306854959329, "lr": 0.0026291804804649314, "grad_norm": 0.199162, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:23.546656+00:00", "epoch": 0, "step": 2231, "train_loss": 3.769322395324707, "perplexity": 43.35068025779239, "lr": 0.0026291804804649314, "grad_norm": 0.16255, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:23.851685+00:00", "epoch": 0, "step": 2232, "train_loss": 3.761265516281128, "perplexity": 43.002812317227914, "lr": 0.0026291804804649314, "grad_norm": 0.159705, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:24.157006+00:00", "epoch": 0, "step": 2233, "train_loss": 3.735945701599121, "perplexity": 41.927657873077465, "lr": 0.0026291804804649314, "grad_norm": 0.145879, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:24.462193+00:00", "epoch": 0, "step": 2234, "train_loss": 3.7649104595184326, "perplexity": 43.15984113387839, "lr": 0.0026291804804649314, "grad_norm": 0.151799, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:24.766388+00:00", "epoch": 0, "step": 2235, "train_loss": 3.793116331100464, "perplexity": 44.39453299441134, "lr": 0.0026291804804649314, "grad_norm": 0.168663, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:25.071101+00:00", "epoch": 0, "step": 2236, "train_loss": 3.7251768112182617, "perplexity": 41.47856597229452, "lr": 0.0026291804804649314, "grad_norm": 0.149813, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:25.376224+00:00", "epoch": 0, "step": 2237, "train_loss": 3.710732936859131, "perplexity": 40.8837607635708, "lr": 0.0026291804804649314, "grad_norm": 0.152824, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:25.682461+00:00", "epoch": 0, "step": 2238, "train_loss": 3.855896472930908, "perplexity": 47.27097510156205, "lr": 0.0026291804804649314, "grad_norm": 0.162516, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:25.988684+00:00", "epoch": 0, "step": 2239, "train_loss": 3.7188355922698975, "perplexity": 41.21637349234683, "lr": 0.0026291804804649314, "grad_norm": 0.154469, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:26.293445+00:00", "epoch": 0, "step": 2240, "train_loss": 3.6897833347320557, "perplexity": 40.03617156965246, "lr": 0.0026291804804649314, "grad_norm": 0.163104, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:26.597799+00:00", "epoch": 0, "step": 2241, "train_loss": 3.7971487045288086, "perplexity": 44.57390974358434, "lr": 0.0026291804804649314, "grad_norm": 0.181014, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:26.903045+00:00", "epoch": 0, "step": 2242, "train_loss": 3.752210855484009, "perplexity": 42.61519396302931, "lr": 0.0026291804804649314, "grad_norm": 0.204503, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:27.209616+00:00", "epoch": 0, "step": 2243, "train_loss": 3.7625677585601807, "perplexity": 43.05884887621466, "lr": 0.0026291804804649314, "grad_norm": 0.207337, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:27.514877+00:00", "epoch": 0, "step": 2244, "train_loss": 3.8024020195007324, "perplexity": 44.808686669633616, "lr": 0.0026291804804649314, "grad_norm": 0.198542, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:27.819742+00:00", "epoch": 0, "step": 2245, "train_loss": 3.783210515975952, "perplexity": 43.95693989443942, "lr": 0.0026291804804649314, "grad_norm": 0.200929, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:28.125393+00:00", "epoch": 0, "step": 2246, "train_loss": 3.720675230026245, "perplexity": 41.29226647559829, "lr": 0.0026291804804649314, "grad_norm": 0.164298, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:28.430204+00:00", "epoch": 0, "step": 2247, "train_loss": 3.7670083045959473, "perplexity": 43.250478832831675, "lr": 0.0026291804804649314, "grad_norm": 0.151318, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:28.736202+00:00", "epoch": 0, "step": 2248, "train_loss": 3.7986977100372314, "perplexity": 44.64300847865332, "lr": 0.0026291804804649314, "grad_norm": 0.155815, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:29.042900+00:00", "epoch": 0, "step": 2249, "train_loss": 3.7145981788635254, "perplexity": 41.042092190488894, "lr": 0.0026291804804649314, "grad_norm": 0.164469, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:29.348423+00:00", "epoch": 0, "step": 2250, "train_loss": 3.7998764514923096, "perplexity": 44.6956620698153, "lr": 0.0026291804804649314, "grad_norm": 0.155381, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:29.654019+00:00", "epoch": 0, "step": 2251, "train_loss": 3.7408270835876465, "perplexity": 42.132823124634214, "lr": 0.0026291804804649314, "grad_norm": 0.156894, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:29.959878+00:00", "epoch": 0, "step": 2252, "train_loss": 3.860280990600586, "perplexity": 47.478690560447646, "lr": 0.0026291804804649314, "grad_norm": 0.146411, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:30.267138+00:00", "epoch": 0, "step": 2253, "train_loss": 3.8207693099975586, "perplexity": 45.63930559285084, "lr": 0.0026291804804649314, "grad_norm": 0.142947, "tokens_per_sec": 106646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:30.573184+00:00", "epoch": 0, "step": 2254, "train_loss": 3.744683265686035, "perplexity": 42.29560862647081, "lr": 0.0026291804804649314, "grad_norm": 0.157601, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:30.878328+00:00", "epoch": 0, "step": 2255, "train_loss": 3.7513535022735596, "perplexity": 42.578673347441764, "lr": 0.0026291804804649314, "grad_norm": 0.173555, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:31.183382+00:00", "epoch": 0, "step": 2256, "train_loss": 3.7943274974823, "perplexity": 44.4483347351644, "lr": 0.0026291804804649314, "grad_norm": 0.16548, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:31.488704+00:00", "epoch": 0, "step": 2257, "train_loss": 3.82348370552063, "perplexity": 45.7633570057611, "lr": 0.0026291804804649314, "grad_norm": 0.189455, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:31.794070+00:00", "epoch": 0, "step": 2258, "train_loss": 3.8759450912475586, "perplexity": 48.228256862650035, "lr": 0.0026291804804649314, "grad_norm": 0.21178, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:32.098686+00:00", "epoch": 0, "step": 2259, "train_loss": 3.768977642059326, "perplexity": 43.33573754513953, "lr": 0.0026291804804649314, "grad_norm": 0.191651, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:32.404077+00:00", "epoch": 0, "step": 2260, "train_loss": 3.593951463699341, "perplexity": 36.37753681575998, "lr": 0.0026291804804649314, "grad_norm": 0.166664, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:32.709278+00:00", "epoch": 0, "step": 2261, "train_loss": 3.7496767044067383, "perplexity": 42.5073373435446, "lr": 0.0026291804804649314, "grad_norm": 0.191094, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:33.014142+00:00", "epoch": 0, "step": 2262, "train_loss": 3.823995590209961, "perplexity": 45.7867885641611, "lr": 0.0026291804804649314, "grad_norm": 0.163261, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:33.319939+00:00", "epoch": 0, "step": 2263, "train_loss": 3.832772731781006, "perplexity": 46.190434527008804, "lr": 0.0026291804804649314, "grad_norm": 0.173928, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:33.624937+00:00", "epoch": 0, "step": 2264, "train_loss": 3.7736942768096924, "perplexity": 43.54061918688601, "lr": 0.0026291804804649314, "grad_norm": 0.15846, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:33.929183+00:00", "epoch": 0, "step": 2265, "train_loss": 3.8028295040130615, "perplexity": 44.82784578402326, "lr": 0.0026291804804649314, "grad_norm": 0.159995, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:34.233610+00:00", "epoch": 0, "step": 2266, "train_loss": 3.7712790966033936, "perplexity": 43.435587631377715, "lr": 0.0026291804804649314, "grad_norm": 0.171184, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:34.538699+00:00", "epoch": 0, "step": 2267, "train_loss": 3.7726714611053467, "perplexity": 43.49610792509398, "lr": 0.0026291804804649314, "grad_norm": 0.156422, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:34.844447+00:00", "epoch": 0, "step": 2268, "train_loss": 3.7416305541992188, "perplexity": 42.16668921317897, "lr": 0.0026291804804649314, "grad_norm": 0.169709, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:35.148908+00:00", "epoch": 0, "step": 2269, "train_loss": 3.7731125354766846, "perplexity": 43.51529717518479, "lr": 0.0026291804804649314, "grad_norm": 0.188958, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:35.454355+00:00", "epoch": 0, "step": 2270, "train_loss": 3.7631759643554688, "perplexity": 43.085045483296184, "lr": 0.0026291804804649314, "grad_norm": 0.177339, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:35.760259+00:00", "epoch": 0, "step": 2271, "train_loss": 3.7161741256713867, "perplexity": 41.10682333768961, "lr": 0.0026291804804649314, "grad_norm": 0.159019, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:36.065006+00:00", "epoch": 0, "step": 2272, "train_loss": 3.7803354263305664, "perplexity": 43.83074125483836, "lr": 0.0026291804804649314, "grad_norm": 0.16927, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:36.370853+00:00", "epoch": 0, "step": 2273, "train_loss": 3.774294853210449, "perplexity": 43.566776509193005, "lr": 0.0026291804804649314, "grad_norm": 0.148162, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:36.675820+00:00", "epoch": 0, "step": 2274, "train_loss": 3.7815499305725098, "perplexity": 43.88400621473251, "lr": 0.0026291804804649314, "grad_norm": 0.156093, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:36.980102+00:00", "epoch": 0, "step": 2275, "train_loss": 3.7944085597991943, "perplexity": 44.45193796620129, "lr": 0.0026291804804649314, "grad_norm": 0.15297, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:37.284784+00:00", "epoch": 0, "step": 2276, "train_loss": 3.7324767112731934, "perplexity": 41.78246321855763, "lr": 0.0026291804804649314, "grad_norm": 0.146079, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:37.589856+00:00", "epoch": 0, "step": 2277, "train_loss": 3.808298349380493, "perplexity": 45.07367392612833, "lr": 0.0026291804804649314, "grad_norm": 0.135416, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:37.894303+00:00", "epoch": 0, "step": 2278, "train_loss": 3.786841630935669, "perplexity": 44.116842733575986, "lr": 0.0026291804804649314, "grad_norm": 0.156753, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:38.198626+00:00", "epoch": 0, "step": 2279, "train_loss": 3.7679483890533447, "perplexity": 43.29115705324299, "lr": 0.0026291804804649314, "grad_norm": 0.165384, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:38.503242+00:00", "epoch": 0, "step": 2280, "train_loss": 3.728783130645752, "perplexity": 41.628420980686805, "lr": 0.0026291804804649314, "grad_norm": 0.154112, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:38.807205+00:00", "epoch": 0, "step": 2281, "train_loss": 3.746150493621826, "perplexity": 42.35771147338376, "lr": 0.0026291804804649314, "grad_norm": 0.147414, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:39.113499+00:00", "epoch": 0, "step": 2282, "train_loss": 3.6529059410095215, "perplexity": 38.58663376592925, "lr": 0.0026291804804649314, "grad_norm": 0.154552, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:39.418596+00:00", "epoch": 0, "step": 2283, "train_loss": 3.7897660732269287, "perplexity": 44.24604872987815, "lr": 0.0026291804804649314, "grad_norm": 0.159697, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:39.723564+00:00", "epoch": 0, "step": 2284, "train_loss": 3.7632644176483154, "perplexity": 43.088856665994975, "lr": 0.0026291804804649314, "grad_norm": 0.178917, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:40.027742+00:00", "epoch": 0, "step": 2285, "train_loss": 3.7379863262176514, "perplexity": 42.013303839844504, "lr": 0.0026291804804649314, "grad_norm": 0.149167, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:40.332394+00:00", "epoch": 0, "step": 2286, "train_loss": 3.7431278228759766, "perplexity": 42.229871364665065, "lr": 0.0026291804804649314, "grad_norm": 0.144174, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:40.637630+00:00", "epoch": 0, "step": 2287, "train_loss": 3.7940728664398193, "perplexity": 44.43701825017959, "lr": 0.0026291804804649314, "grad_norm": 0.177547, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:40.942311+00:00", "epoch": 0, "step": 2288, "train_loss": 3.7782161235809326, "perplexity": 43.737949006542905, "lr": 0.0026291804804649314, "grad_norm": 0.187368, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:41.247762+00:00", "epoch": 0, "step": 2289, "train_loss": 3.7040460109710693, "perplexity": 40.61128610953593, "lr": 0.0026291804804649314, "grad_norm": 0.197577, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:41.552837+00:00", "epoch": 0, "step": 2290, "train_loss": 3.8616578578948975, "perplexity": 47.544107441504494, "lr": 0.0026291804804649314, "grad_norm": 0.205106, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:41.856707+00:00", "epoch": 0, "step": 2291, "train_loss": 3.891644239425659, "perplexity": 48.99137388381021, "lr": 0.0026291804804649314, "grad_norm": 0.214715, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:42.161718+00:00", "epoch": 0, "step": 2292, "train_loss": 3.872359037399292, "perplexity": 48.055617468622906, "lr": 0.0026291804804649314, "grad_norm": 0.180991, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:42.466990+00:00", "epoch": 0, "step": 2293, "train_loss": 3.7547740936279297, "perplexity": 42.72456696835208, "lr": 0.0026291804804649314, "grad_norm": 0.176556, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:42.772133+00:00", "epoch": 0, "step": 2294, "train_loss": 3.7887582778930664, "perplexity": 44.20148023016153, "lr": 0.0026291804804649314, "grad_norm": 0.160326, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:43.076603+00:00", "epoch": 0, "step": 2295, "train_loss": 3.748718023300171, "perplexity": 42.46660588970006, "lr": 0.0026291804804649314, "grad_norm": 0.171374, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:43.382007+00:00", "epoch": 0, "step": 2296, "train_loss": 3.787627696990967, "perplexity": 44.1515351195829, "lr": 0.0026291804804649314, "grad_norm": 0.154516, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:43.687980+00:00", "epoch": 0, "step": 2297, "train_loss": 3.7126007080078125, "perplexity": 40.96019362970557, "lr": 0.0026291804804649314, "grad_norm": 0.157028, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:43.993899+00:00", "epoch": 0, "step": 2298, "train_loss": 3.8104147911071777, "perplexity": 45.16917075148676, "lr": 0.0026291804804649314, "grad_norm": 0.178922, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:44.300224+00:00", "epoch": 0, "step": 2299, "train_loss": 3.83971905708313, "perplexity": 46.51240527322218, "lr": 0.0026291804804649314, "grad_norm": 0.178784, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:44.606295+00:00", "epoch": 0, "step": 2300, "train_loss": 3.7310004234313965, "perplexity": 41.72082578459944, "lr": 0.0026291804804649314, "grad_norm": 0.151374, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:44.911130+00:00", "epoch": 0, "step": 2301, "train_loss": 3.7654755115509033, "perplexity": 43.18423558124917, "lr": 0.0026291804804649314, "grad_norm": 0.147289, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:45.215402+00:00", "epoch": 0, "step": 2302, "train_loss": 3.772186279296875, "perplexity": 43.47500952348361, "lr": 0.0026291804804649314, "grad_norm": 0.145054, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:45.520890+00:00", "epoch": 0, "step": 2303, "train_loss": 3.8134586811065674, "perplexity": 45.30687020329158, "lr": 0.0026291804804649314, "grad_norm": 0.156917, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:45.826752+00:00", "epoch": 0, "step": 2304, "train_loss": 3.73856258392334, "perplexity": 42.03752130700451, "lr": 0.0026291804804649314, "grad_norm": 0.160759, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:46.132828+00:00", "epoch": 0, "step": 2305, "train_loss": 3.7742083072662354, "perplexity": 43.5630061445409, "lr": 0.0026291804804649314, "grad_norm": 0.140835, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:46.438738+00:00", "epoch": 0, "step": 2306, "train_loss": 3.7802674770355225, "perplexity": 43.82776308805215, "lr": 0.0026291804804649314, "grad_norm": 0.138438, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:46.743781+00:00", "epoch": 0, "step": 2307, "train_loss": 3.627666711807251, "perplexity": 37.6249243263513, "lr": 0.0026291804804649314, "grad_norm": 0.14165, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:47.048094+00:00", "epoch": 0, "step": 2308, "train_loss": 3.7561206817626953, "perplexity": 42.78213811690796, "lr": 0.0026291804804649314, "grad_norm": 0.151574, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:47.353205+00:00", "epoch": 0, "step": 2309, "train_loss": 3.7708053588867188, "perplexity": 43.41501542856974, "lr": 0.0026291804804649314, "grad_norm": 0.164402, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:47.659721+00:00", "epoch": 0, "step": 2310, "train_loss": 3.651435136795044, "perplexity": 38.529922098467104, "lr": 0.0026291804804649314, "grad_norm": 0.189332, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:47.964842+00:00", "epoch": 0, "step": 2311, "train_loss": 3.724489212036133, "perplexity": 41.45005514738934, "lr": 0.0026291804804649314, "grad_norm": 0.196997, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:48.269999+00:00", "epoch": 0, "step": 2312, "train_loss": 3.7812082767486572, "perplexity": 43.86901563714317, "lr": 0.0026291804804649314, "grad_norm": 0.184971, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:48.574684+00:00", "epoch": 0, "step": 2313, "train_loss": 3.824650764465332, "perplexity": 45.81679671847611, "lr": 0.0026291804804649314, "grad_norm": 0.193973, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:48.880484+00:00", "epoch": 0, "step": 2314, "train_loss": 3.7236976623535156, "perplexity": 41.41725835125215, "lr": 0.0026291804804649314, "grad_norm": 0.175268, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:49.186051+00:00", "epoch": 0, "step": 2315, "train_loss": 3.670917510986328, "perplexity": 39.287936441865234, "lr": 0.0026291804804649314, "grad_norm": 0.16281, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:49.491229+00:00", "epoch": 0, "step": 2316, "train_loss": 3.7943756580352783, "perplexity": 44.45047544309264, "lr": 0.0026291804804649314, "grad_norm": 0.160025, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:49.796715+00:00", "epoch": 0, "step": 2317, "train_loss": 3.7074697017669678, "perplexity": 40.75056488367914, "lr": 0.0026291804804649314, "grad_norm": 0.180911, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:50.101554+00:00", "epoch": 0, "step": 2318, "train_loss": 3.7846364974975586, "perplexity": 44.01966639125882, "lr": 0.0026291804804649314, "grad_norm": 0.150068, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:50.406715+00:00", "epoch": 0, "step": 2319, "train_loss": 3.8159008026123047, "perplexity": 45.41765029954311, "lr": 0.0026291804804649314, "grad_norm": 0.153893, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:50.712327+00:00", "epoch": 0, "step": 2320, "train_loss": 3.7655115127563477, "perplexity": 43.18579029377188, "lr": 0.0026291804804649314, "grad_norm": 0.142872, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:51.017084+00:00", "epoch": 0, "step": 2321, "train_loss": 3.769988536834717, "perplexity": 43.37956756584856, "lr": 0.0026291804804649314, "grad_norm": 0.151916, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:51.322191+00:00", "epoch": 0, "step": 2322, "train_loss": 3.6978421211242676, "perplexity": 40.36011807895709, "lr": 0.0026291804804649314, "grad_norm": 0.153277, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:51.626960+00:00", "epoch": 0, "step": 2323, "train_loss": 3.7330989837646484, "perplexity": 41.808471387288506, "lr": 0.0026291804804649314, "grad_norm": 0.158691, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:51.932306+00:00", "epoch": 0, "step": 2324, "train_loss": 3.6933858394622803, "perplexity": 40.1806621748896, "lr": 0.0026291804804649314, "grad_norm": 0.166299, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:52.237780+00:00", "epoch": 0, "step": 2325, "train_loss": 3.7433230876922607, "perplexity": 42.23811817786882, "lr": 0.0026291804804649314, "grad_norm": 0.175073, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:52.542884+00:00", "epoch": 0, "step": 2326, "train_loss": 3.785257577896118, "perplexity": 44.04701463505553, "lr": 0.0026291804804649314, "grad_norm": 0.157531, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:52.848112+00:00", "epoch": 0, "step": 2327, "train_loss": 3.8050475120544434, "perplexity": 44.92738465465051, "lr": 0.0026291804804649314, "grad_norm": 0.146843, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:53.152988+00:00", "epoch": 0, "step": 2328, "train_loss": 3.7825169563293457, "perplexity": 43.92646370448975, "lr": 0.0026291804804649314, "grad_norm": 0.144273, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:53.457939+00:00", "epoch": 0, "step": 2329, "train_loss": 3.8670413494110107, "perplexity": 47.80075093976868, "lr": 0.0026291804804649314, "grad_norm": 0.143977, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:53.764038+00:00", "epoch": 0, "step": 2330, "train_loss": 3.654639720916748, "perplexity": 38.65359252534284, "lr": 0.0026291804804649314, "grad_norm": 0.166268, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:54.070106+00:00", "epoch": 0, "step": 2331, "train_loss": 3.745457649230957, "perplexity": 42.3283743347889, "lr": 0.0026291804804649314, "grad_norm": 0.15261, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:54.375521+00:00", "epoch": 0, "step": 2332, "train_loss": 3.7219719886779785, "perplexity": 41.345847312598785, "lr": 0.0026291804804649314, "grad_norm": 0.152883, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:54.679588+00:00", "epoch": 0, "step": 2333, "train_loss": 3.7985053062438965, "perplexity": 44.634419820747816, "lr": 0.0026291804804649314, "grad_norm": 0.144497, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:54.984264+00:00", "epoch": 0, "step": 2334, "train_loss": 3.7708187103271484, "perplexity": 43.415595085431605, "lr": 0.0026291804804649314, "grad_norm": 0.170668, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:55.288546+00:00", "epoch": 0, "step": 2335, "train_loss": 3.7736566066741943, "perplexity": 43.5389790367541, "lr": 0.0026291804804649314, "grad_norm": 0.177913, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:55.593830+00:00", "epoch": 0, "step": 2336, "train_loss": 3.874594211578369, "perplexity": 48.1631502764566, "lr": 0.0026291804804649314, "grad_norm": 0.176551, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:55.899218+00:00", "epoch": 0, "step": 2337, "train_loss": 3.8600893020629883, "perplexity": 47.46959031192159, "lr": 0.0026291804804649314, "grad_norm": 0.16772, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:56.204784+00:00", "epoch": 0, "step": 2338, "train_loss": 3.745961904525757, "perplexity": 42.34972402406203, "lr": 0.0026291804804649314, "grad_norm": 0.178851, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:56.510295+00:00", "epoch": 0, "step": 2339, "train_loss": 3.7889113426208496, "perplexity": 44.208246435521055, "lr": 0.0026291804804649314, "grad_norm": 0.152472, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:56.815077+00:00", "epoch": 0, "step": 2340, "train_loss": 3.7557365894317627, "perplexity": 42.76570898111086, "lr": 0.0026291804804649314, "grad_norm": 0.158784, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:57.121391+00:00", "epoch": 0, "step": 2341, "train_loss": 3.735353469848633, "perplexity": 41.9028343342309, "lr": 0.0026291804804649314, "grad_norm": 0.165105, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:57.426747+00:00", "epoch": 0, "step": 2342, "train_loss": 3.771012306213379, "perplexity": 43.42400097968541, "lr": 0.0026291804804649314, "grad_norm": 0.192258, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:57.733200+00:00", "epoch": 0, "step": 2343, "train_loss": 3.7536890506744385, "perplexity": 42.67823411913359, "lr": 0.0026291804804649314, "grad_norm": 0.181935, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:58.038442+00:00", "epoch": 0, "step": 2344, "train_loss": 3.769542694091797, "perplexity": 43.360231411220724, "lr": 0.0026291804804649314, "grad_norm": 0.177206, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:58.343309+00:00", "epoch": 0, "step": 2345, "train_loss": 3.8156352043151855, "perplexity": 45.40558905075887, "lr": 0.0026291804804649314, "grad_norm": 0.186637, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:58.649279+00:00", "epoch": 0, "step": 2346, "train_loss": 3.7840113639831543, "perplexity": 43.99215682197749, "lr": 0.0026291804804649314, "grad_norm": 0.170268, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:58.954891+00:00", "epoch": 0, "step": 2347, "train_loss": 3.774885654449463, "perplexity": 43.59252341963886, "lr": 0.0026291804804649314, "grad_norm": 0.192505, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:59.259576+00:00", "epoch": 0, "step": 2348, "train_loss": 3.800170421600342, "perplexity": 44.708803189876136, "lr": 0.0026291804804649314, "grad_norm": 0.177365, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:59.564871+00:00", "epoch": 0, "step": 2349, "train_loss": 3.8444440364837646, "perplexity": 46.73269545345474, "lr": 0.0026291804804649314, "grad_norm": 0.15879, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:09:59.869470+00:00", "epoch": 0, "step": 2350, "train_loss": 3.8121261596679688, "perplexity": 45.24653803331207, "lr": 0.0026291804804649314, "grad_norm": 0.159819, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:00.174129+00:00", "epoch": 0, "step": 2351, "train_loss": 3.788790464401245, "perplexity": 44.20290294436245, "lr": 0.0026291804804649314, "grad_norm": 0.150986, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:00.479254+00:00", "epoch": 0, "step": 2352, "train_loss": 3.817697048187256, "perplexity": 45.499304866791036, "lr": 0.0026291804804649314, "grad_norm": 0.152056, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:00.785215+00:00", "epoch": 0, "step": 2353, "train_loss": 3.774359941482544, "perplexity": 43.56961228768369, "lr": 0.0026291804804649314, "grad_norm": 0.167125, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:01.090903+00:00", "epoch": 0, "step": 2354, "train_loss": 3.8416311740875244, "perplexity": 46.601427517610496, "lr": 0.0026291804804649314, "grad_norm": 0.15645, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:01.395235+00:00", "epoch": 0, "step": 2355, "train_loss": 3.762662410736084, "perplexity": 43.062924682841626, "lr": 0.0026291804804649314, "grad_norm": 0.168022, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:01.699323+00:00", "epoch": 0, "step": 2356, "train_loss": 3.765904664993286, "perplexity": 43.2027722218526, "lr": 0.0026291804804649314, "grad_norm": 0.166466, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:02.004492+00:00", "epoch": 0, "step": 2357, "train_loss": 3.7853922843933105, "perplexity": 44.05294845376179, "lr": 0.0026291804804649314, "grad_norm": 0.171033, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:02.310772+00:00", "epoch": 0, "step": 2358, "train_loss": 3.7712512016296387, "perplexity": 43.43437601369981, "lr": 0.0026291804804649314, "grad_norm": 0.18809, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:02.616850+00:00", "epoch": 0, "step": 2359, "train_loss": 3.7370243072509766, "perplexity": 41.97290567971168, "lr": 0.0026291804804649314, "grad_norm": 0.183692, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:02.922795+00:00", "epoch": 0, "step": 2360, "train_loss": 3.752941846847534, "perplexity": 42.64635669022583, "lr": 0.0026291804804649314, "grad_norm": 0.142696, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:03.227456+00:00", "epoch": 0, "step": 2361, "train_loss": 3.7890713214874268, "perplexity": 44.21531938642542, "lr": 0.0026291804804649314, "grad_norm": 0.166729, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:03.532552+00:00", "epoch": 0, "step": 2362, "train_loss": 3.7351207733154297, "perplexity": 41.89308482433251, "lr": 0.0026291804804649314, "grad_norm": 0.157013, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:03.838073+00:00", "epoch": 0, "step": 2363, "train_loss": 3.7936205863952637, "perplexity": 44.416924817857236, "lr": 0.0026291804804649314, "grad_norm": 0.147903, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:04.143542+00:00", "epoch": 0, "step": 2364, "train_loss": 3.8642966747283936, "perplexity": 47.66973331148679, "lr": 0.0026291804804649314, "grad_norm": 0.151302, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:04.448951+00:00", "epoch": 0, "step": 2365, "train_loss": 3.695817470550537, "perplexity": 40.27848560923175, "lr": 0.0026291804804649314, "grad_norm": 0.158266, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:04.754242+00:00", "epoch": 0, "step": 2366, "train_loss": 3.70672345161438, "perplexity": 40.72016611237124, "lr": 0.0026291804804649314, "grad_norm": 0.162607, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:05.059399+00:00", "epoch": 0, "step": 2367, "train_loss": 3.784001350402832, "perplexity": 43.99171630518718, "lr": 0.0026291804804649314, "grad_norm": 0.158449, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:05.364638+00:00", "epoch": 0, "step": 2368, "train_loss": 3.7165119647979736, "perplexity": 41.12071317711628, "lr": 0.0026291804804649314, "grad_norm": 0.172631, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:05.670144+00:00", "epoch": 0, "step": 2369, "train_loss": 3.681016206741333, "perplexity": 39.68670348337731, "lr": 0.0026291804804649314, "grad_norm": 0.151866, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:05.976186+00:00", "epoch": 0, "step": 2370, "train_loss": 3.81276273727417, "perplexity": 45.27535013577886, "lr": 0.0026291804804649314, "grad_norm": 0.152496, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:06.280686+00:00", "epoch": 0, "step": 2371, "train_loss": 3.733877420425415, "perplexity": 41.84102930464114, "lr": 0.0026291804804649314, "grad_norm": 0.150626, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:06.585570+00:00", "epoch": 0, "step": 2372, "train_loss": 3.74402117729187, "perplexity": 42.26761446320301, "lr": 0.0026291804804649314, "grad_norm": 0.157466, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:06.890800+00:00", "epoch": 0, "step": 2373, "train_loss": 3.7670211791992188, "perplexity": 43.25103566917247, "lr": 0.0026291804804649314, "grad_norm": 0.145917, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:07.196030+00:00", "epoch": 0, "step": 2374, "train_loss": 3.810241937637329, "perplexity": 45.16136377834282, "lr": 0.0026291804804649314, "grad_norm": 0.148375, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:07.500980+00:00", "epoch": 0, "step": 2375, "train_loss": 3.7376532554626465, "perplexity": 41.99931276715223, "lr": 0.0026291804804649314, "grad_norm": 0.143181, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:07.805687+00:00", "epoch": 0, "step": 2376, "train_loss": 3.815714120864868, "perplexity": 45.40917244457574, "lr": 0.0026291804804649314, "grad_norm": 0.157804, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:08.110784+00:00", "epoch": 0, "step": 2377, "train_loss": 3.8055245876312256, "perplexity": 44.94882352617231, "lr": 0.0026291804804649314, "grad_norm": 0.150128, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:08.415855+00:00", "epoch": 0, "step": 2378, "train_loss": 3.6747121810913086, "perplexity": 39.437304421635716, "lr": 0.0026291804804649314, "grad_norm": 0.144754, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:08.721183+00:00", "epoch": 0, "step": 2379, "train_loss": 3.8291146755218506, "perplexity": 46.02177598815959, "lr": 0.0026291804804649314, "grad_norm": 0.145618, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:09.026665+00:00", "epoch": 0, "step": 2380, "train_loss": 3.8129138946533203, "perplexity": 45.28219435630968, "lr": 0.0026291804804649314, "grad_norm": 0.14926, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:09.331473+00:00", "epoch": 0, "step": 2381, "train_loss": 3.7266273498535156, "perplexity": 41.53877589262297, "lr": 0.0026291804804649314, "grad_norm": 0.17291, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:09.635862+00:00", "epoch": 0, "step": 2382, "train_loss": 3.783082962036133, "perplexity": 43.95133337114815, "lr": 0.0026291804804649314, "grad_norm": 0.187817, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:09.940556+00:00", "epoch": 0, "step": 2383, "train_loss": 3.7143301963806152, "perplexity": 41.03109510229917, "lr": 0.0026291804804649314, "grad_norm": 0.190379, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:10.246275+00:00", "epoch": 0, "step": 2384, "train_loss": 3.7670998573303223, "perplexity": 43.25443871369804, "lr": 0.0026291804804649314, "grad_norm": 0.170142, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:10.551566+00:00", "epoch": 0, "step": 2385, "train_loss": 3.819011926651001, "perplexity": 45.55917027211671, "lr": 0.0026291804804649314, "grad_norm": 0.159262, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:10.858059+00:00", "epoch": 0, "step": 2386, "train_loss": 3.771456480026245, "perplexity": 43.44329306797349, "lr": 0.0026291804804649314, "grad_norm": 0.174318, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:11.162755+00:00", "epoch": 0, "step": 2387, "train_loss": 3.7266409397125244, "perplexity": 41.539340402566545, "lr": 0.0026291804804649314, "grad_norm": 0.187689, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:11.467435+00:00", "epoch": 0, "step": 2388, "train_loss": 3.6104800701141357, "perplexity": 36.983803372431666, "lr": 0.0026291804804649314, "grad_norm": 0.200992, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:11.772652+00:00", "epoch": 0, "step": 2389, "train_loss": 3.77180814743042, "perplexity": 43.45857334470566, "lr": 0.0026291804804649314, "grad_norm": 0.175679, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:12.078387+00:00", "epoch": 0, "step": 2390, "train_loss": 3.724642276763916, "perplexity": 41.456400174384605, "lr": 0.0026291804804649314, "grad_norm": 0.212895, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:12.384267+00:00", "epoch": 0, "step": 2391, "train_loss": 3.770120143890381, "perplexity": 43.38527699870446, "lr": 0.0026291804804649314, "grad_norm": 0.200776, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:12.689918+00:00", "epoch": 0, "step": 2392, "train_loss": 3.7027249336242676, "perplexity": 40.55767088216095, "lr": 0.0026291804804649314, "grad_norm": 0.195439, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:12.994492+00:00", "epoch": 0, "step": 2393, "train_loss": 3.655738115310669, "perplexity": 38.6960727404217, "lr": 0.0026291804804649314, "grad_norm": 0.21018, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:13.298908+00:00", "epoch": 0, "step": 2394, "train_loss": 3.6675267219543457, "perplexity": 39.15494493837653, "lr": 0.0026291804804649314, "grad_norm": 0.16971, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:13.604052+00:00", "epoch": 0, "step": 2395, "train_loss": 3.6504135131835938, "perplexity": 38.490579120583455, "lr": 0.0026291804804649314, "grad_norm": 0.176033, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:13.909390+00:00", "epoch": 0, "step": 2396, "train_loss": 3.751650094985962, "perplexity": 42.59130374460966, "lr": 0.0026291804804649314, "grad_norm": 0.167617, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:14.214858+00:00", "epoch": 0, "step": 2397, "train_loss": 3.7149715423583984, "perplexity": 41.057418670462056, "lr": 0.0026291804804649314, "grad_norm": 0.149665, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:14.519238+00:00", "epoch": 0, "step": 2398, "train_loss": 3.778745412826538, "perplexity": 43.76110516018786, "lr": 0.0026291804804649314, "grad_norm": 0.13449, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:14.823469+00:00", "epoch": 0, "step": 2399, "train_loss": 3.7323529720306396, "perplexity": 41.777293408067756, "lr": 0.0026291804804649314, "grad_norm": 0.142767, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:15.129698+00:00", "epoch": 0, "step": 2400, "train_loss": 3.7047274112701416, "perplexity": 40.638968082217836, "lr": 0.0026291804804649314, "grad_norm": 0.146493, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:15.435283+00:00", "epoch": 0, "step": 2401, "train_loss": 3.724527359008789, "perplexity": 41.45163637166892, "lr": 0.0026291804804649314, "grad_norm": 0.138806, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:15.741268+00:00", "epoch": 0, "step": 2402, "train_loss": 3.6781725883483887, "perplexity": 39.57400994803978, "lr": 0.0026291804804649314, "grad_norm": 0.135534, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:16.046455+00:00", "epoch": 0, "step": 2403, "train_loss": 3.697082996368408, "perplexity": 40.329491340402434, "lr": 0.0026291804804649314, "grad_norm": 0.137958, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:16.351470+00:00", "epoch": 0, "step": 2404, "train_loss": 3.7022950649261475, "perplexity": 40.54024015571031, "lr": 0.0026291804804649314, "grad_norm": 0.170134, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:16.655538+00:00", "epoch": 0, "step": 2405, "train_loss": 3.707979202270508, "perplexity": 40.77133260714044, "lr": 0.0026291804804649314, "grad_norm": 0.167565, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:16.959814+00:00", "epoch": 0, "step": 2406, "train_loss": 3.733006715774536, "perplexity": 41.8046139816242, "lr": 0.0026291804804649314, "grad_norm": 0.137646, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:17.265157+00:00", "epoch": 0, "step": 2407, "train_loss": 3.706003189086914, "perplexity": 40.690847462438256, "lr": 0.0026291804804649314, "grad_norm": 0.163199, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:17.571923+00:00", "epoch": 0, "step": 2408, "train_loss": 3.76975154876709, "perplexity": 43.369288344031204, "lr": 0.0026291804804649314, "grad_norm": 0.150575, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:17.876190+00:00", "epoch": 0, "step": 2409, "train_loss": 3.6940224170684814, "perplexity": 40.20624842758618, "lr": 0.0026291804804649314, "grad_norm": 0.158339, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:18.180428+00:00", "epoch": 0, "step": 2410, "train_loss": 3.7442543506622314, "perplexity": 42.27747129445515, "lr": 0.0026291804804649314, "grad_norm": 0.171402, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:18.484688+00:00", "epoch": 0, "step": 2411, "train_loss": 3.707308053970337, "perplexity": 40.74397817703188, "lr": 0.0026291804804649314, "grad_norm": 0.207326, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:18.789786+00:00", "epoch": 0, "step": 2412, "train_loss": 3.7655858993530273, "perplexity": 43.189002857221105, "lr": 0.0026291804804649314, "grad_norm": 0.222374, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:19.094633+00:00", "epoch": 0, "step": 2413, "train_loss": 3.8571200370788574, "perplexity": 47.32884957127767, "lr": 0.0026291804804649314, "grad_norm": 0.172115, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:19.398856+00:00", "epoch": 0, "step": 2414, "train_loss": 3.8258581161499023, "perplexity": 45.87214711214762, "lr": 0.0026291804804649314, "grad_norm": 0.156201, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:19.704284+00:00", "epoch": 0, "step": 2415, "train_loss": 3.740811347961426, "perplexity": 42.132160143494126, "lr": 0.0026291804804649314, "grad_norm": 0.178635, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:20.009869+00:00", "epoch": 0, "step": 2416, "train_loss": 3.745236873626709, "perplexity": 42.31903029387439, "lr": 0.0026291804804649314, "grad_norm": 0.181969, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:20.315543+00:00", "epoch": 0, "step": 2417, "train_loss": 3.78222918510437, "perplexity": 43.91382475087147, "lr": 0.0026291804804649314, "grad_norm": 0.184809, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:20.620942+00:00", "epoch": 0, "step": 2418, "train_loss": 3.7578089237213135, "perplexity": 42.85442571986925, "lr": 0.0026291804804649314, "grad_norm": 0.188812, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:20.927039+00:00", "epoch": 0, "step": 2419, "train_loss": 3.8035032749176025, "perplexity": 44.85805965969795, "lr": 0.0026291804804649314, "grad_norm": 0.165007, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:21.231899+00:00", "epoch": 0, "step": 2420, "train_loss": 3.7664363384246826, "perplexity": 43.22574809529514, "lr": 0.0026291804804649314, "grad_norm": 0.140168, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:21.536012+00:00", "epoch": 0, "step": 2421, "train_loss": 3.7871978282928467, "perplexity": 44.1325598353936, "lr": 0.0026291804804649314, "grad_norm": 0.171205, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:21.840733+00:00", "epoch": 0, "step": 2422, "train_loss": 3.7038774490356445, "perplexity": 40.6044411694636, "lr": 0.0026291804804649314, "grad_norm": 0.16729, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:22.147407+00:00", "epoch": 0, "step": 2423, "train_loss": 3.749438762664795, "perplexity": 42.49722427685977, "lr": 0.0026291804804649314, "grad_norm": 0.179286, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:22.452429+00:00", "epoch": 0, "step": 2424, "train_loss": 3.6325039863586426, "perplexity": 37.80736732297358, "lr": 0.0026291804804649314, "grad_norm": 0.162756, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:22.757434+00:00", "epoch": 0, "step": 2425, "train_loss": 3.8473293781280518, "perplexity": 46.86772996246012, "lr": 0.0026291804804649314, "grad_norm": 0.139815, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:23.061555+00:00", "epoch": 0, "step": 2426, "train_loss": 3.819964647293091, "perplexity": 45.602596117136194, "lr": 0.0026291804804649314, "grad_norm": 0.163561, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:23.366000+00:00", "epoch": 0, "step": 2427, "train_loss": 3.671949863433838, "perplexity": 39.32851638200485, "lr": 0.0026291804804649314, "grad_norm": 0.151158, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:23.671444+00:00", "epoch": 0, "step": 2428, "train_loss": 3.7467710971832275, "perplexity": 42.384006978675096, "lr": 0.0026291804804649314, "grad_norm": 0.163858, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:23.978818+00:00", "epoch": 0, "step": 2429, "train_loss": 3.78757905960083, "perplexity": 44.149387756365655, "lr": 0.0026291804804649314, "grad_norm": 0.148459, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:24.287065+00:00", "epoch": 0, "step": 2430, "train_loss": 3.7495675086975098, "perplexity": 42.502695978109166, "lr": 0.0026291804804649314, "grad_norm": 0.152775, "tokens_per_sec": 106365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:24.591736+00:00", "epoch": 0, "step": 2431, "train_loss": 3.7719554901123047, "perplexity": 43.46497711921623, "lr": 0.0026291804804649314, "grad_norm": 0.182701, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:24.896434+00:00", "epoch": 0, "step": 2432, "train_loss": 3.753159523010254, "perplexity": 42.655640795931724, "lr": 0.0026291804804649314, "grad_norm": 0.16919, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:25.201906+00:00", "epoch": 0, "step": 2433, "train_loss": 3.705810308456421, "perplexity": 40.68299974298527, "lr": 0.0026291804804649314, "grad_norm": 0.142527, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:25.507596+00:00", "epoch": 0, "step": 2434, "train_loss": 3.7320830821990967, "perplexity": 41.76601966279057, "lr": 0.0026291804804649314, "grad_norm": 0.152877, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:25.813288+00:00", "epoch": 0, "step": 2435, "train_loss": 3.814129114151001, "perplexity": 45.33725561076822, "lr": 0.0026291804804649314, "grad_norm": 0.158426, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:26.118117+00:00", "epoch": 0, "step": 2436, "train_loss": 3.853834629058838, "perplexity": 47.17361014138851, "lr": 0.0026291804804649314, "grad_norm": 0.155528, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:26.423146+00:00", "epoch": 0, "step": 2437, "train_loss": 3.710265874862671, "perplexity": 40.8646699712846, "lr": 0.0026291804804649314, "grad_norm": 0.14964, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:26.727876+00:00", "epoch": 0, "step": 2438, "train_loss": 3.840317964553833, "perplexity": 46.540270243657254, "lr": 0.0026291804804649314, "grad_norm": 0.14869, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:27.033600+00:00", "epoch": 0, "step": 2439, "train_loss": 3.8483622074127197, "perplexity": 46.916161332834776, "lr": 0.0026291804804649314, "grad_norm": 0.145358, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:27.339831+00:00", "epoch": 0, "step": 2440, "train_loss": 3.746281147003174, "perplexity": 42.3632460131592, "lr": 0.0026291804804649314, "grad_norm": 0.162301, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:27.644821+00:00", "epoch": 0, "step": 2441, "train_loss": 3.737218141555786, "perplexity": 41.98104225725339, "lr": 0.0026291804804649314, "grad_norm": 0.192146, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:27.949141+00:00", "epoch": 0, "step": 2442, "train_loss": 3.8072941303253174, "perplexity": 45.028432803685526, "lr": 0.0026291804804649314, "grad_norm": 0.210428, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:28.253967+00:00", "epoch": 0, "step": 2443, "train_loss": 3.6655259132385254, "perplexity": 39.07668170426568, "lr": 0.0026291804804649314, "grad_norm": 0.208995, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:28.559740+00:00", "epoch": 0, "step": 2444, "train_loss": 3.785635232925415, "perplexity": 44.06365235311918, "lr": 0.0026291804804649314, "grad_norm": 0.211812, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:28.865603+00:00", "epoch": 0, "step": 2445, "train_loss": 3.628251552581787, "perplexity": 37.646935352081464, "lr": 0.0026291804804649314, "grad_norm": 0.17301, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:29.170755+00:00", "epoch": 0, "step": 2446, "train_loss": 3.7050790786743164, "perplexity": 40.65326199583609, "lr": 0.0026291804804649314, "grad_norm": 0.192765, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:29.475532+00:00", "epoch": 0, "step": 2447, "train_loss": 3.6992225646972656, "perplexity": 40.41587141788008, "lr": 0.0026291804804649314, "grad_norm": 0.168271, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:29.780562+00:00", "epoch": 0, "step": 2448, "train_loss": 3.8171303272247314, "perplexity": 45.47352676212881, "lr": 0.0026291804804649314, "grad_norm": 0.16952, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:30.087049+00:00", "epoch": 0, "step": 2449, "train_loss": 3.741257667541504, "perplexity": 42.15096874852928, "lr": 0.0026291804804649314, "grad_norm": 0.161757, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:30.391847+00:00", "epoch": 0, "step": 2450, "train_loss": 3.8031551837921143, "perplexity": 44.84244768457642, "lr": 0.0026291804804649314, "grad_norm": 0.173942, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:30.697815+00:00", "epoch": 0, "step": 2451, "train_loss": 3.7468326091766357, "perplexity": 42.38661418361935, "lr": 0.0026291804804649314, "grad_norm": 0.169423, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:31.002901+00:00", "epoch": 0, "step": 2452, "train_loss": 3.8267571926116943, "perplexity": 45.913408225536514, "lr": 0.0026291804804649314, "grad_norm": 0.15496, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:31.308227+00:00", "epoch": 0, "step": 2453, "train_loss": 3.754288911819458, "perplexity": 42.703842813582185, "lr": 0.0026291804804649314, "grad_norm": 0.141714, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:31.613533+00:00", "epoch": 0, "step": 2454, "train_loss": 3.72084903717041, "perplexity": 41.29944399024411, "lr": 0.0026291804804649314, "grad_norm": 0.157738, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:31.920091+00:00", "epoch": 0, "step": 2455, "train_loss": 3.654634952545166, "perplexity": 38.65340821109014, "lr": 0.0026291804804649314, "grad_norm": 0.144297, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:32.227399+00:00", "epoch": 0, "step": 2456, "train_loss": 3.7412259578704834, "perplexity": 42.149632176368314, "lr": 0.0026291804804649314, "grad_norm": 0.18771, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:32.532991+00:00", "epoch": 0, "step": 2457, "train_loss": 3.8225605487823486, "perplexity": 45.72112974856228, "lr": 0.0026291804804649314, "grad_norm": 0.14987, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:32.838357+00:00", "epoch": 0, "step": 2458, "train_loss": 3.679161548614502, "perplexity": 39.61316643036036, "lr": 0.0026291804804649314, "grad_norm": 0.155974, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:33.143572+00:00", "epoch": 0, "step": 2459, "train_loss": 3.7993364334106445, "perplexity": 44.6715321200166, "lr": 0.0026291804804649314, "grad_norm": 0.162206, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:33.449979+00:00", "epoch": 0, "step": 2460, "train_loss": 3.6768903732299805, "perplexity": 39.52330007162104, "lr": 0.0026291804804649314, "grad_norm": 0.169843, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:33.756852+00:00", "epoch": 0, "step": 2461, "train_loss": 3.784276008605957, "perplexity": 44.00380065039625, "lr": 0.0026291804804649314, "grad_norm": 0.169098, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:34.061863+00:00", "epoch": 0, "step": 2462, "train_loss": 3.8267223834991455, "perplexity": 45.91181004835782, "lr": 0.0026291804804649314, "grad_norm": 0.171711, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:34.366219+00:00", "epoch": 0, "step": 2463, "train_loss": 3.812185764312744, "perplexity": 45.24923501751446, "lr": 0.0026291804804649314, "grad_norm": 0.182518, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:34.670691+00:00", "epoch": 0, "step": 2464, "train_loss": 3.693331718444824, "perplexity": 40.17848761541585, "lr": 0.0026291804804649314, "grad_norm": 0.205045, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:34.976716+00:00", "epoch": 0, "step": 2465, "train_loss": 3.763261079788208, "perplexity": 43.08871284165927, "lr": 0.0026291804804649314, "grad_norm": 0.174022, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:35.281914+00:00", "epoch": 0, "step": 2466, "train_loss": 3.7206993103027344, "perplexity": 41.29326081676385, "lr": 0.0026291804804649314, "grad_norm": 0.16866, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:35.587776+00:00", "epoch": 0, "step": 2467, "train_loss": 3.7168025970458984, "perplexity": 41.13266591926524, "lr": 0.0026291804804649314, "grad_norm": 0.182923, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:35.893161+00:00", "epoch": 0, "step": 2468, "train_loss": 3.8208749294281006, "perplexity": 45.64412624489067, "lr": 0.0026291804804649314, "grad_norm": 0.172477, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:36.197835+00:00", "epoch": 0, "step": 2469, "train_loss": 3.7607860565185547, "perplexity": 42.982199141033476, "lr": 0.0026291804804649314, "grad_norm": 0.18033, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:36.502961+00:00", "epoch": 0, "step": 2470, "train_loss": 3.8077664375305176, "perplexity": 45.049705080066836, "lr": 0.0026291804804649314, "grad_norm": 0.164213, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:36.809068+00:00", "epoch": 0, "step": 2471, "train_loss": 3.7311270236968994, "perplexity": 41.72610798657781, "lr": 0.0026291804804649314, "grad_norm": 0.175766, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:37.114911+00:00", "epoch": 0, "step": 2472, "train_loss": 3.7700467109680176, "perplexity": 43.3820912079993, "lr": 0.0026291804804649314, "grad_norm": 0.185943, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:37.419849+00:00", "epoch": 0, "step": 2473, "train_loss": 3.749445915222168, "perplexity": 42.497528241781666, "lr": 0.0026291804804649314, "grad_norm": 0.173132, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:37.725078+00:00", "epoch": 0, "step": 2474, "train_loss": 3.9004905223846436, "perplexity": 49.42668805706079, "lr": 0.0026291804804649314, "grad_norm": 0.180908, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:38.030222+00:00", "epoch": 0, "step": 2475, "train_loss": 3.773282289505005, "perplexity": 43.522684699187074, "lr": 0.0026291804804649314, "grad_norm": 0.197656, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:38.336966+00:00", "epoch": 0, "step": 2476, "train_loss": 3.744272470474243, "perplexity": 42.27823736122781, "lr": 0.0026291804804649314, "grad_norm": 0.151835, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:38.641652+00:00", "epoch": 0, "step": 2477, "train_loss": 3.7406156063079834, "perplexity": 42.123913931894414, "lr": 0.0026291804804649314, "grad_norm": 0.174456, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:38.945898+00:00", "epoch": 0, "step": 2478, "train_loss": 3.7606844902038574, "perplexity": 42.97783381915772, "lr": 0.0026291804804649314, "grad_norm": 0.174688, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:39.251355+00:00", "epoch": 0, "step": 2479, "train_loss": 3.7215499877929688, "perplexity": 41.328403009455954, "lr": 0.0026291804804649314, "grad_norm": 0.154112, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:39.557920+00:00", "epoch": 0, "step": 2480, "train_loss": 3.709658145904541, "perplexity": 40.8398428728161, "lr": 0.0026291804804649314, "grad_norm": 0.163148, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:39.862907+00:00", "epoch": 0, "step": 2481, "train_loss": 3.700601816177368, "perplexity": 40.47165352828898, "lr": 0.0026291804804649314, "grad_norm": 0.155188, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:40.167638+00:00", "epoch": 0, "step": 2482, "train_loss": 3.7909512519836426, "perplexity": 44.29851929425856, "lr": 0.0026291804804649314, "grad_norm": 0.166758, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:40.472537+00:00", "epoch": 0, "step": 2483, "train_loss": 3.6830546855926514, "perplexity": 39.76768650214922, "lr": 0.0026291804804649314, "grad_norm": 0.144973, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:40.778556+00:00", "epoch": 0, "step": 2484, "train_loss": 3.793901205062866, "perplexity": 44.42939078512838, "lr": 0.0026291804804649314, "grad_norm": 0.173197, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:41.085355+00:00", "epoch": 0, "step": 2485, "train_loss": 3.692584276199341, "perplexity": 40.14846773687277, "lr": 0.0026291804804649314, "grad_norm": 0.165163, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:41.391056+00:00", "epoch": 0, "step": 2486, "train_loss": 3.70536208152771, "perplexity": 40.66476861310663, "lr": 0.0026291804804649314, "grad_norm": 0.158112, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:41.696260+00:00", "epoch": 0, "step": 2487, "train_loss": 3.8132336139678955, "perplexity": 45.29667426308133, "lr": 0.0026291804804649314, "grad_norm": 0.158196, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:42.001926+00:00", "epoch": 0, "step": 2488, "train_loss": 3.6954755783081055, "perplexity": 40.26471706127904, "lr": 0.0026291804804649314, "grad_norm": 0.149155, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:42.307104+00:00", "epoch": 0, "step": 2489, "train_loss": 3.685762405395508, "perplexity": 39.875512169372875, "lr": 0.0026291804804649314, "grad_norm": 0.155678, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:42.613259+00:00", "epoch": 0, "step": 2490, "train_loss": 3.725440740585327, "perplexity": 41.48951482875719, "lr": 0.0026291804804649314, "grad_norm": 0.149177, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:42.917771+00:00", "epoch": 0, "step": 2491, "train_loss": 3.7568414211273193, "perplexity": 42.812984002538315, "lr": 0.0026291804804649314, "grad_norm": 0.156307, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:43.222091+00:00", "epoch": 0, "step": 2492, "train_loss": 3.5781474113464355, "perplexity": 35.80714345484416, "lr": 0.0026291804804649314, "grad_norm": 0.157523, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:43.526508+00:00", "epoch": 0, "step": 2493, "train_loss": 3.678821563720703, "perplexity": 39.59970084135663, "lr": 0.0026291804804649314, "grad_norm": 0.156763, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:43.831655+00:00", "epoch": 0, "step": 2494, "train_loss": 3.8131022453308105, "perplexity": 45.29072409156043, "lr": 0.0026291804804649314, "grad_norm": 0.15753, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:44.137879+00:00", "epoch": 0, "step": 2495, "train_loss": 3.738248586654663, "perplexity": 42.024323712244964, "lr": 0.0026291804804649314, "grad_norm": 0.153278, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:44.442378+00:00", "epoch": 0, "step": 2496, "train_loss": 3.7637810707092285, "perplexity": 43.11112440753579, "lr": 0.0026291804804649314, "grad_norm": 0.138781, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:44.747703+00:00", "epoch": 0, "step": 2497, "train_loss": 3.6737165451049805, "perplexity": 39.39805876258698, "lr": 0.0026291804804649314, "grad_norm": 0.144488, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:45.052828+00:00", "epoch": 0, "step": 2498, "train_loss": 3.855297565460205, "perplexity": 47.242672637550854, "lr": 0.0026291804804649314, "grad_norm": 0.14588, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:45.358243+00:00", "epoch": 0, "step": 2499, "train_loss": 3.777082681655884, "perplexity": 43.688402665677536, "lr": 0.0026291804804649314, "grad_norm": 0.155321, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:45.665261+00:00", "epoch": 0, "step": 2500, "train_loss": 3.668839693069458, "perplexity": 39.20638801434065, "lr": 0.0026291804804649314, "grad_norm": 0.172606, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:10:48.737908+00:00", "step": 2500, "epoch": 0, "val_loss": 3.688628840446472, "val_ppl": 39.9899767093407, "eval_train_loss": 3.668839693069458, "eval_train_ppl": 39.20638801434065} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:10:49.634526+00:00", "step": 2500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p6886_epoch_0000_step_0002500.pt", "val_loss": 3.688628840446472} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T21:10:51.310508+00:00", "step": 2500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0002500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:52.566102+00:00", "epoch": 0, "step": 2501, "train_loss": 3.867047071456909, "perplexity": 47.80102445864208, "lr": 0.0026291804804649314, "grad_norm": 0.18061, "tokens_per_sec": 4748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:52.870275+00:00", "epoch": 0, "step": 2502, "train_loss": 3.6008317470550537, "perplexity": 36.6286875802944, "lr": 0.0026291804804649314, "grad_norm": 0.165777, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:53.174473+00:00", "epoch": 0, "step": 2503, "train_loss": 3.7868261337280273, "perplexity": 44.116159051001254, "lr": 0.0026291804804649314, "grad_norm": 0.158545, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:53.478182+00:00", "epoch": 0, "step": 2504, "train_loss": 3.8132848739624023, "perplexity": 45.29899622986673, "lr": 0.0026291804804649314, "grad_norm": 0.175904, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:53.781641+00:00", "epoch": 0, "step": 2505, "train_loss": 3.7034685611724854, "perplexity": 40.58784190013019, "lr": 0.0026291804804649314, "grad_norm": 0.180413, "tokens_per_sec": 107981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:54.086872+00:00", "epoch": 0, "step": 2506, "train_loss": 3.8202192783355713, "perplexity": 45.614209432217834, "lr": 0.0026291804804649314, "grad_norm": 0.180233, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:54.390964+00:00", "epoch": 0, "step": 2507, "train_loss": 3.557276964187622, "perplexity": 35.067576747970506, "lr": 0.0026291804804649314, "grad_norm": 0.163654, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:54.694965+00:00", "epoch": 0, "step": 2508, "train_loss": 3.7712364196777344, "perplexity": 43.4337339735879, "lr": 0.0026291804804649314, "grad_norm": 0.153469, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:54.999154+00:00", "epoch": 0, "step": 2509, "train_loss": 3.812718629837036, "perplexity": 45.273353200159214, "lr": 0.0026291804804649314, "grad_norm": 0.182226, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:55.303036+00:00", "epoch": 0, "step": 2510, "train_loss": 3.765723466873169, "perplexity": 43.19494466993241, "lr": 0.0026291804804649314, "grad_norm": 0.157375, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:55.607747+00:00", "epoch": 0, "step": 2511, "train_loss": 3.7235262393951416, "perplexity": 41.41015909080334, "lr": 0.0026291804804649314, "grad_norm": 0.178311, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:55.912186+00:00", "epoch": 0, "step": 2512, "train_loss": 3.794948101043701, "perplexity": 44.47592809138134, "lr": 0.0026291804804649314, "grad_norm": 0.161642, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:56.216691+00:00", "epoch": 0, "step": 2513, "train_loss": 3.8307888507843018, "perplexity": 46.098889039423014, "lr": 0.0026291804804649314, "grad_norm": 0.163928, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:56.521347+00:00", "epoch": 0, "step": 2514, "train_loss": 3.814140796661377, "perplexity": 45.33778526682116, "lr": 0.0026291804804649314, "grad_norm": 0.152517, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:57.124173+00:00", "epoch": 0, "step": 2515, "train_loss": 3.667747974395752, "perplexity": 39.163609023976974, "lr": 0.0026291804804649314, "grad_norm": 0.15305, "tokens_per_sec": 54357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:57.428386+00:00", "epoch": 0, "step": 2516, "train_loss": 3.8124899864196777, "perplexity": 45.263002929273995, "lr": 0.0026291804804649314, "grad_norm": 0.157593, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:57.731688+00:00", "epoch": 0, "step": 2517, "train_loss": 3.7380788326263428, "perplexity": 42.017190519468606, "lr": 0.0026291804804649314, "grad_norm": 0.168715, "tokens_per_sec": 108038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:58.036340+00:00", "epoch": 0, "step": 2518, "train_loss": 3.76662278175354, "perplexity": 43.23380799899667, "lr": 0.0026291804804649314, "grad_norm": 0.171683, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:58.340754+00:00", "epoch": 0, "step": 2519, "train_loss": 3.6499886512756348, "perplexity": 38.47422941312973, "lr": 0.0026291804804649314, "grad_norm": 0.168023, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:58.645239+00:00", "epoch": 0, "step": 2520, "train_loss": 3.6438331604003906, "perplexity": 38.2381290473565, "lr": 0.0026291804804649314, "grad_norm": 0.172082, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:58.950704+00:00", "epoch": 0, "step": 2521, "train_loss": 3.6454708576202393, "perplexity": 38.30080283132259, "lr": 0.0026291804804649314, "grad_norm": 0.172781, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:59.254769+00:00", "epoch": 0, "step": 2522, "train_loss": 3.7679827213287354, "perplexity": 43.292643362682966, "lr": 0.0026291804804649314, "grad_norm": 0.158992, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:59.559335+00:00", "epoch": 0, "step": 2523, "train_loss": 3.774813413619995, "perplexity": 43.589374373334664, "lr": 0.0026291804804649314, "grad_norm": 0.166867, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:10:59.863809+00:00", "epoch": 0, "step": 2524, "train_loss": 3.7279725074768066, "perplexity": 41.59468969168254, "lr": 0.0026291804804649314, "grad_norm": 0.169502, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:00.168811+00:00", "epoch": 0, "step": 2525, "train_loss": 3.7657930850982666, "perplexity": 43.197951929992364, "lr": 0.0026291804804649314, "grad_norm": 0.184965, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:00.473622+00:00", "epoch": 0, "step": 2526, "train_loss": 3.7872064113616943, "perplexity": 44.132938629818696, "lr": 0.0026291804804649314, "grad_norm": 0.169147, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:00.779007+00:00", "epoch": 0, "step": 2527, "train_loss": 3.8124325275421143, "perplexity": 45.26040224264751, "lr": 0.0026291804804649314, "grad_norm": 0.171687, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:01.083175+00:00", "epoch": 0, "step": 2528, "train_loss": 3.6863653659820557, "perplexity": 39.8995627816364, "lr": 0.0026291804804649314, "grad_norm": 0.165567, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:01.387370+00:00", "epoch": 0, "step": 2529, "train_loss": 3.595226764678955, "perplexity": 36.42395871875457, "lr": 0.0026291804804649314, "grad_norm": 0.157357, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:01.691727+00:00", "epoch": 0, "step": 2530, "train_loss": 3.734672784805298, "perplexity": 41.874321406888754, "lr": 0.0026291804804649314, "grad_norm": 0.172708, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:02.006691+00:00", "epoch": 0, "step": 2531, "train_loss": 3.6623754501342773, "perplexity": 38.953765783023314, "lr": 0.0026291804804649314, "grad_norm": 0.154669, "tokens_per_sec": 103983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:02.311270+00:00", "epoch": 0, "step": 2532, "train_loss": 3.7730002403259277, "perplexity": 43.51041089268633, "lr": 0.0026291804804649314, "grad_norm": 0.143312, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:02.615500+00:00", "epoch": 0, "step": 2533, "train_loss": 3.757436513900757, "perplexity": 42.83846928222831, "lr": 0.0026291804804649314, "grad_norm": 0.153022, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:02.920965+00:00", "epoch": 0, "step": 2534, "train_loss": 3.7683849334716797, "perplexity": 43.31005969183834, "lr": 0.0026291804804649314, "grad_norm": 0.164463, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:03.225134+00:00", "epoch": 0, "step": 2535, "train_loss": 3.806805372238159, "perplexity": 45.00643017042114, "lr": 0.0026291804804649314, "grad_norm": 0.15895, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:03.530036+00:00", "epoch": 0, "step": 2536, "train_loss": 3.5721302032470703, "perplexity": 35.592331353926994, "lr": 0.0026291804804649314, "grad_norm": 0.173256, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:03.835727+00:00", "epoch": 0, "step": 2537, "train_loss": 3.70763897895813, "perplexity": 40.757463608722844, "lr": 0.0026291804804649314, "grad_norm": 0.174443, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:04.139733+00:00", "epoch": 0, "step": 2538, "train_loss": 3.662015438079834, "perplexity": 38.93974448184562, "lr": 0.0026291804804649314, "grad_norm": 0.171076, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:04.443487+00:00", "epoch": 0, "step": 2539, "train_loss": 3.750967264175415, "perplexity": 42.56223101715812, "lr": 0.0026291804804649314, "grad_norm": 0.159241, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:04.748201+00:00", "epoch": 0, "step": 2540, "train_loss": 3.6655616760253906, "perplexity": 39.078079220294256, "lr": 0.0026291804804649314, "grad_norm": 0.151159, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:05.052754+00:00", "epoch": 0, "step": 2541, "train_loss": 3.849590539932251, "perplexity": 46.97382538755798, "lr": 0.0026291804804649314, "grad_norm": 0.163938, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:05.357559+00:00", "epoch": 0, "step": 2542, "train_loss": 3.6232292652130127, "perplexity": 37.45833562163819, "lr": 0.0026291804804649314, "grad_norm": 0.169662, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:05.663305+00:00", "epoch": 0, "step": 2543, "train_loss": 3.7206709384918213, "perplexity": 41.292089268795515, "lr": 0.0026291804804649314, "grad_norm": 0.161515, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:05.968027+00:00", "epoch": 0, "step": 2544, "train_loss": 3.761298656463623, "perplexity": 43.004237461890554, "lr": 0.0026291804804649314, "grad_norm": 0.154155, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:06.274611+00:00", "epoch": 0, "step": 2545, "train_loss": 3.76310133934021, "perplexity": 43.081830381084586, "lr": 0.0026291804804649314, "grad_norm": 0.157073, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:06.578999+00:00", "epoch": 0, "step": 2546, "train_loss": 3.836455821990967, "perplexity": 46.36087173935598, "lr": 0.0026291804804649314, "grad_norm": 0.169924, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:06.883343+00:00", "epoch": 0, "step": 2547, "train_loss": 3.756199598312378, "perplexity": 42.78551446885946, "lr": 0.0026291804804649314, "grad_norm": 0.162238, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:07.188061+00:00", "epoch": 0, "step": 2548, "train_loss": 3.7294812202453613, "perplexity": 41.65749149415256, "lr": 0.0026291804804649314, "grad_norm": 0.168493, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:07.492489+00:00", "epoch": 0, "step": 2549, "train_loss": 3.713001251220703, "perplexity": 40.97660324342295, "lr": 0.0026291804804649314, "grad_norm": 0.159559, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:07.797305+00:00", "epoch": 0, "step": 2550, "train_loss": 3.6988067626953125, "perplexity": 40.39906991092603, "lr": 0.0026291804804649314, "grad_norm": 0.170659, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:08.101231+00:00", "epoch": 0, "step": 2551, "train_loss": 3.821699380874634, "perplexity": 45.68177312768005, "lr": 0.0026291804804649314, "grad_norm": 0.164714, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:08.406417+00:00", "epoch": 0, "step": 2552, "train_loss": 3.6558456420898438, "perplexity": 38.70023382820037, "lr": 0.0026291804804649314, "grad_norm": 0.147983, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:08.712070+00:00", "epoch": 0, "step": 2553, "train_loss": 3.715358018875122, "perplexity": 41.07328946526268, "lr": 0.0026291804804649314, "grad_norm": 0.148677, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:09.016679+00:00", "epoch": 0, "step": 2554, "train_loss": 3.660055160522461, "perplexity": 38.86348654243125, "lr": 0.0026291804804649314, "grad_norm": 0.143785, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:09.320166+00:00", "epoch": 0, "step": 2555, "train_loss": 3.631730079650879, "perplexity": 37.778119266892645, "lr": 0.0026291804804649314, "grad_norm": 0.146242, "tokens_per_sec": 107972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:09.624681+00:00", "epoch": 0, "step": 2556, "train_loss": 3.6923651695251465, "perplexity": 40.139671903281055, "lr": 0.0026291804804649314, "grad_norm": 0.155812, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:09.929264+00:00", "epoch": 0, "step": 2557, "train_loss": 3.7736330032348633, "perplexity": 43.53795137923205, "lr": 0.0026291804804649314, "grad_norm": 0.13048, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:10.233981+00:00", "epoch": 0, "step": 2558, "train_loss": 3.8763842582702637, "perplexity": 48.24944177414342, "lr": 0.0026291804804649314, "grad_norm": 0.18632, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:10.539558+00:00", "epoch": 0, "step": 2559, "train_loss": 3.7810168266296387, "perplexity": 43.86061771279567, "lr": 0.0026291804804649314, "grad_norm": 0.169974, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:10.843970+00:00", "epoch": 0, "step": 2560, "train_loss": 3.6520779132843018, "perplexity": 38.55469618777303, "lr": 0.0026291804804649314, "grad_norm": 0.160474, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:11.148815+00:00", "epoch": 0, "step": 2561, "train_loss": 3.8059980869293213, "perplexity": 44.97011180215631, "lr": 0.0026291804804649314, "grad_norm": 0.180209, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:11.456160+00:00", "epoch": 0, "step": 2562, "train_loss": 3.7324774265289307, "perplexity": 41.782493103714856, "lr": 0.0026291804804649314, "grad_norm": 0.176117, "tokens_per_sec": 106617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:11.761665+00:00", "epoch": 0, "step": 2563, "train_loss": 3.768723964691162, "perplexity": 43.32474564354891, "lr": 0.0026291804804649314, "grad_norm": 0.170961, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:12.067978+00:00", "epoch": 0, "step": 2564, "train_loss": 3.7833030223846436, "perplexity": 43.96100638117126, "lr": 0.0026291804804649314, "grad_norm": 0.184758, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:12.372017+00:00", "epoch": 0, "step": 2565, "train_loss": 3.7265048027038574, "perplexity": 41.533685745934896, "lr": 0.0026291804804649314, "grad_norm": 0.189943, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:12.675496+00:00", "epoch": 0, "step": 2566, "train_loss": 3.694256544113159, "perplexity": 40.21566289975638, "lr": 0.0026291804804649314, "grad_norm": 0.150911, "tokens_per_sec": 107975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:12.979355+00:00", "epoch": 0, "step": 2567, "train_loss": 3.5964198112487793, "perplexity": 36.46744013028036, "lr": 0.0026291804804649314, "grad_norm": 0.177527, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:13.283904+00:00", "epoch": 0, "step": 2568, "train_loss": 3.7419891357421875, "perplexity": 42.18181212089466, "lr": 0.0026291804804649314, "grad_norm": 0.182011, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:13.589779+00:00", "epoch": 0, "step": 2569, "train_loss": 3.654726982116699, "perplexity": 38.656965631377524, "lr": 0.0026291804804649314, "grad_norm": 0.154126, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:13.894908+00:00", "epoch": 0, "step": 2570, "train_loss": 3.7339391708374023, "perplexity": 41.84361308521261, "lr": 0.0026291804804649314, "grad_norm": 0.16587, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:14.198883+00:00", "epoch": 0, "step": 2571, "train_loss": 3.7204341888427734, "perplexity": 41.28231453828034, "lr": 0.0026291804804649314, "grad_norm": 0.172996, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:14.503046+00:00", "epoch": 0, "step": 2572, "train_loss": 3.7233476638793945, "perplexity": 41.402764910515955, "lr": 0.0026291804804649314, "grad_norm": 0.1431, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:14.808636+00:00", "epoch": 0, "step": 2573, "train_loss": 3.6645138263702393, "perplexity": 39.03715271461646, "lr": 0.0026291804804649314, "grad_norm": 0.154558, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:15.114471+00:00", "epoch": 0, "step": 2574, "train_loss": 3.6984431743621826, "perplexity": 40.384383950420855, "lr": 0.0026291804804649314, "grad_norm": 0.161743, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:15.419170+00:00", "epoch": 0, "step": 2575, "train_loss": 3.7598421573638916, "perplexity": 42.94164742097662, "lr": 0.0026291804804649314, "grad_norm": 0.186125, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:15.723533+00:00", "epoch": 0, "step": 2576, "train_loss": 3.673018455505371, "perplexity": 39.37056498519659, "lr": 0.0026291804804649314, "grad_norm": 0.164959, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:16.027893+00:00", "epoch": 0, "step": 2577, "train_loss": 3.8531060218811035, "perplexity": 47.13925162888925, "lr": 0.0026291804804649314, "grad_norm": 0.153486, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:16.332199+00:00", "epoch": 0, "step": 2578, "train_loss": 3.6485722064971924, "perplexity": 38.41977136928383, "lr": 0.0026291804804649314, "grad_norm": 0.19405, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:16.637278+00:00", "epoch": 0, "step": 2579, "train_loss": 3.6653189659118652, "perplexity": 39.06859572616686, "lr": 0.0026291804804649314, "grad_norm": 0.182606, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:16.942607+00:00", "epoch": 0, "step": 2580, "train_loss": 3.7463560104370117, "perplexity": 42.36641758994035, "lr": 0.0026291804804649314, "grad_norm": 0.182222, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:17.246993+00:00", "epoch": 0, "step": 2581, "train_loss": 3.7342376708984375, "perplexity": 41.85610527063878, "lr": 0.0026291804804649314, "grad_norm": 0.212173, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:17.551119+00:00", "epoch": 0, "step": 2582, "train_loss": 3.8320369720458984, "perplexity": 46.15646196449108, "lr": 0.0026291804804649314, "grad_norm": 0.206035, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:17.854753+00:00", "epoch": 0, "step": 2583, "train_loss": 3.776909828186035, "perplexity": 43.6808516263151, "lr": 0.0026291804804649314, "grad_norm": 0.19282, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:18.160157+00:00", "epoch": 0, "step": 2584, "train_loss": 3.846229314804077, "perplexity": 46.81620083949606, "lr": 0.0026291804804649314, "grad_norm": 0.15741, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:18.465901+00:00", "epoch": 0, "step": 2585, "train_loss": 3.72859263420105, "perplexity": 41.620491669768164, "lr": 0.0026291804804649314, "grad_norm": 0.164204, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:18.771152+00:00", "epoch": 0, "step": 2586, "train_loss": 3.718634605407715, "perplexity": 41.20809037519481, "lr": 0.0026291804804649314, "grad_norm": 0.156844, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:19.075236+00:00", "epoch": 0, "step": 2587, "train_loss": 3.731764078140259, "perplexity": 41.75269825790072, "lr": 0.0026291804804649314, "grad_norm": 0.158571, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:19.379706+00:00", "epoch": 0, "step": 2588, "train_loss": 3.7572855949401855, "perplexity": 42.83200463280313, "lr": 0.0026291804804649314, "grad_norm": 0.150902, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:19.684716+00:00", "epoch": 0, "step": 2589, "train_loss": 3.8204739093780518, "perplexity": 45.625825704786514, "lr": 0.0026291804804649314, "grad_norm": 0.169252, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:19.989759+00:00", "epoch": 0, "step": 2590, "train_loss": 3.87660551071167, "perplexity": 48.26011826198839, "lr": 0.0026291804804649314, "grad_norm": 0.148452, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:20.294784+00:00", "epoch": 0, "step": 2591, "train_loss": 3.7409439086914062, "perplexity": 42.137745583595226, "lr": 0.0026291804804649314, "grad_norm": 0.150912, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:20.599333+00:00", "epoch": 0, "step": 2592, "train_loss": 3.695725440979004, "perplexity": 40.27477896802205, "lr": 0.0026291804804649314, "grad_norm": 0.137497, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:20.903324+00:00", "epoch": 0, "step": 2593, "train_loss": 3.7099409103393555, "perplexity": 40.85139256074748, "lr": 0.0026291804804649314, "grad_norm": 0.164637, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:21.207604+00:00", "epoch": 0, "step": 2594, "train_loss": 3.6964519023895264, "perplexity": 40.30404767076397, "lr": 0.0026291804804649314, "grad_norm": 0.152278, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:21.513046+00:00", "epoch": 0, "step": 2595, "train_loss": 3.798698663711548, "perplexity": 44.643051053564214, "lr": 0.0026291804804649314, "grad_norm": 0.169207, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:21.818400+00:00", "epoch": 0, "step": 2596, "train_loss": 3.7946856021881104, "perplexity": 44.4642547433433, "lr": 0.0026291804804649314, "grad_norm": 0.149946, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:22.122995+00:00", "epoch": 0, "step": 2597, "train_loss": 3.67423677444458, "perplexity": 39.418560120920226, "lr": 0.0026291804804649314, "grad_norm": 0.158195, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:22.426326+00:00", "epoch": 0, "step": 2598, "train_loss": 3.718888998031616, "perplexity": 41.21857474294769, "lr": 0.0026291804804649314, "grad_norm": 0.156455, "tokens_per_sec": 108027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:22.731534+00:00", "epoch": 0, "step": 2599, "train_loss": 3.6726372241973877, "perplexity": 39.355558553853704, "lr": 0.0026291804804649314, "grad_norm": 0.147362, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:23.037482+00:00", "epoch": 0, "step": 2600, "train_loss": 3.730088472366333, "perplexity": 41.682795776478706, "lr": 0.0026291804804649314, "grad_norm": 0.154779, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:23.343443+00:00", "epoch": 0, "step": 2601, "train_loss": 3.726764678955078, "perplexity": 41.54448076710999, "lr": 0.0026291804804649314, "grad_norm": 0.184871, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:23.648723+00:00", "epoch": 0, "step": 2602, "train_loss": 3.646129846572876, "perplexity": 38.32605095546929, "lr": 0.0026291804804649314, "grad_norm": 0.18032, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:23.952852+00:00", "epoch": 0, "step": 2603, "train_loss": 3.7315773963928223, "perplexity": 41.7449045187268, "lr": 0.0026291804804649314, "grad_norm": 0.164196, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:24.257804+00:00", "epoch": 0, "step": 2604, "train_loss": 3.7467119693756104, "perplexity": 42.38150097935227, "lr": 0.0026291804804649314, "grad_norm": 0.142972, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:24.563263+00:00", "epoch": 0, "step": 2605, "train_loss": 3.659574270248413, "perplexity": 38.84480196271381, "lr": 0.0026291804804649314, "grad_norm": 0.162797, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:24.868720+00:00", "epoch": 0, "step": 2606, "train_loss": 3.6983795166015625, "perplexity": 40.38181325279784, "lr": 0.0026291804804649314, "grad_norm": 0.155055, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:25.174693+00:00", "epoch": 0, "step": 2607, "train_loss": 3.7360100746154785, "perplexity": 41.93035696975712, "lr": 0.0026291804804649314, "grad_norm": 0.145348, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:25.478850+00:00", "epoch": 0, "step": 2608, "train_loss": 3.723874092102051, "perplexity": 41.42456623237294, "lr": 0.0026291804804649314, "grad_norm": 0.15083, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:25.783316+00:00", "epoch": 0, "step": 2609, "train_loss": 3.712322473526001, "perplexity": 40.948798676764035, "lr": 0.0026291804804649314, "grad_norm": 0.16414, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:26.088477+00:00", "epoch": 0, "step": 2610, "train_loss": 3.687699556350708, "perplexity": 39.95283192169803, "lr": 0.0026291804804649314, "grad_norm": 0.16024, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:26.393748+00:00", "epoch": 0, "step": 2611, "train_loss": 3.7322704792022705, "perplexity": 41.77384722311764, "lr": 0.0026291804804649314, "grad_norm": 0.157247, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:26.699336+00:00", "epoch": 0, "step": 2612, "train_loss": 3.7321698665618896, "perplexity": 41.769644457478876, "lr": 0.0026291804804649314, "grad_norm": 0.160365, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:27.004322+00:00", "epoch": 0, "step": 2613, "train_loss": 3.7903261184692383, "perplexity": 44.270835459157354, "lr": 0.0026291804804649314, "grad_norm": 0.159113, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:27.309316+00:00", "epoch": 0, "step": 2614, "train_loss": 3.7596025466918945, "perplexity": 42.93135937659326, "lr": 0.0026291804804649314, "grad_norm": 0.168306, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:27.613706+00:00", "epoch": 0, "step": 2615, "train_loss": 3.7509655952453613, "perplexity": 42.5621599838309, "lr": 0.0026291804804649314, "grad_norm": 0.153715, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:27.918772+00:00", "epoch": 0, "step": 2616, "train_loss": 3.67345929145813, "perplexity": 39.38792477185039, "lr": 0.0026291804804649314, "grad_norm": 0.167542, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:28.223740+00:00", "epoch": 0, "step": 2617, "train_loss": 3.7797324657440186, "perplexity": 43.804321011365715, "lr": 0.0026291804804649314, "grad_norm": 0.166972, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:28.528529+00:00", "epoch": 0, "step": 2618, "train_loss": 3.867264986038208, "perplexity": 47.81144213391309, "lr": 0.0026291804804649314, "grad_norm": 0.155705, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:28.833283+00:00", "epoch": 0, "step": 2619, "train_loss": 3.7290661334991455, "perplexity": 41.640203609786795, "lr": 0.0026291804804649314, "grad_norm": 0.150073, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:29.137886+00:00", "epoch": 0, "step": 2620, "train_loss": 3.5786526203155518, "perplexity": 35.82523811528348, "lr": 0.0026291804804649314, "grad_norm": 0.165087, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:29.442253+00:00", "epoch": 0, "step": 2621, "train_loss": 3.6722264289855957, "perplexity": 39.33939479906623, "lr": 0.0026291804804649314, "grad_norm": 0.177753, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:29.747113+00:00", "epoch": 0, "step": 2622, "train_loss": 3.6491289138793945, "perplexity": 38.441165894320086, "lr": 0.0026291804804649314, "grad_norm": 0.162364, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:30.053463+00:00", "epoch": 0, "step": 2623, "train_loss": 3.7145509719848633, "perplexity": 41.040154767153034, "lr": 0.0026291804804649314, "grad_norm": 0.14781, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:30.357458+00:00", "epoch": 0, "step": 2624, "train_loss": 3.718911647796631, "perplexity": 41.21950834455274, "lr": 0.0026291804804649314, "grad_norm": 0.150141, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:30.662396+00:00", "epoch": 0, "step": 2625, "train_loss": 3.6325266361236572, "perplexity": 37.80822366065716, "lr": 0.0026291804804649314, "grad_norm": 0.146237, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:30.966979+00:00", "epoch": 0, "step": 2626, "train_loss": 3.7681097984313965, "perplexity": 43.29814521594042, "lr": 0.0026291804804649314, "grad_norm": 0.142105, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:31.271659+00:00", "epoch": 0, "step": 2627, "train_loss": 3.8252112865448, "perplexity": 45.84248524346745, "lr": 0.0026291804804649314, "grad_norm": 0.144945, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:31.576346+00:00", "epoch": 0, "step": 2628, "train_loss": 3.7328598499298096, "perplexity": 41.79847476251016, "lr": 0.0026291804804649314, "grad_norm": 0.163966, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:31.880270+00:00", "epoch": 0, "step": 2629, "train_loss": 3.8106133937835693, "perplexity": 45.178142360551696, "lr": 0.0026291804804649314, "grad_norm": 0.164075, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:32.183835+00:00", "epoch": 0, "step": 2630, "train_loss": 3.778118848800659, "perplexity": 43.73369461408958, "lr": 0.0026291804804649314, "grad_norm": 0.162945, "tokens_per_sec": 107998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:32.488736+00:00", "epoch": 0, "step": 2631, "train_loss": 3.7252790927886963, "perplexity": 41.48280868213335, "lr": 0.0026291804804649314, "grad_norm": 0.156952, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:32.793406+00:00", "epoch": 0, "step": 2632, "train_loss": 3.711625099182129, "perplexity": 40.920251990174826, "lr": 0.0026291804804649314, "grad_norm": 0.168758, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:33.098912+00:00", "epoch": 0, "step": 2633, "train_loss": 3.6691951751708984, "perplexity": 39.220327661042525, "lr": 0.0026291804804649314, "grad_norm": 0.206861, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:33.403735+00:00", "epoch": 0, "step": 2634, "train_loss": 3.7646589279174805, "perplexity": 43.148986435148245, "lr": 0.0026291804804649314, "grad_norm": 0.215643, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:33.708637+00:00", "epoch": 0, "step": 2635, "train_loss": 3.690964460372925, "perplexity": 40.083487255840026, "lr": 0.0026291804804649314, "grad_norm": 0.17469, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:34.013385+00:00", "epoch": 0, "step": 2636, "train_loss": 3.79940128326416, "perplexity": 44.67442915626608, "lr": 0.0026291804804649314, "grad_norm": 0.165821, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:34.318563+00:00", "epoch": 0, "step": 2637, "train_loss": 3.849088430404663, "perplexity": 46.95024530267169, "lr": 0.0026291804804649314, "grad_norm": 0.174709, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:34.623028+00:00", "epoch": 0, "step": 2638, "train_loss": 3.7549803256988525, "perplexity": 42.733379052913236, "lr": 0.0026291804804649314, "grad_norm": 0.178271, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:34.928135+00:00", "epoch": 0, "step": 2639, "train_loss": 3.778108596801758, "perplexity": 43.73324625859872, "lr": 0.0026291804804649314, "grad_norm": 0.16315, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:35.233538+00:00", "epoch": 0, "step": 2640, "train_loss": 3.6657681465148926, "perplexity": 39.086148523447484, "lr": 0.0026291804804649314, "grad_norm": 0.144812, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:35.537150+00:00", "epoch": 0, "step": 2641, "train_loss": 3.7895243167877197, "perplexity": 44.23535325559013, "lr": 0.0026291804804649314, "grad_norm": 0.154851, "tokens_per_sec": 107870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:35.841669+00:00", "epoch": 0, "step": 2642, "train_loss": 3.8759124279022217, "perplexity": 48.22668159216808, "lr": 0.0026291804804649314, "grad_norm": 0.152619, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:36.147200+00:00", "epoch": 0, "step": 2643, "train_loss": 3.7647769451141357, "perplexity": 43.15407905806847, "lr": 0.0026291804804649314, "grad_norm": 0.158632, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:36.452807+00:00", "epoch": 0, "step": 2644, "train_loss": 3.727546215057373, "perplexity": 41.576961969643506, "lr": 0.0026291804804649314, "grad_norm": 0.15242, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:36.757929+00:00", "epoch": 0, "step": 2645, "train_loss": 3.575244188308716, "perplexity": 35.7033380890279, "lr": 0.0026291804804649314, "grad_norm": 0.169822, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:37.062557+00:00", "epoch": 0, "step": 2646, "train_loss": 3.696890115737915, "perplexity": 40.32171331282478, "lr": 0.0026291804804649314, "grad_norm": 0.160684, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:37.367530+00:00", "epoch": 0, "step": 2647, "train_loss": 3.787754774093628, "perplexity": 44.157146125251565, "lr": 0.0026291804804649314, "grad_norm": 0.152633, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:37.672834+00:00", "epoch": 0, "step": 2648, "train_loss": 3.7350873947143555, "perplexity": 41.89168651510332, "lr": 0.0026291804804649314, "grad_norm": 0.173541, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:37.978508+00:00", "epoch": 0, "step": 2649, "train_loss": 3.6738295555114746, "perplexity": 39.402511404815534, "lr": 0.0026291804804649314, "grad_norm": 0.196367, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:38.283774+00:00", "epoch": 0, "step": 2650, "train_loss": 3.8099822998046875, "perplexity": 45.14963970180504, "lr": 0.0026291804804649314, "grad_norm": 0.165338, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:38.588208+00:00", "epoch": 0, "step": 2651, "train_loss": 3.688565492630005, "perplexity": 39.98744351187271, "lr": 0.0026291804804649314, "grad_norm": 0.157945, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:38.893241+00:00", "epoch": 0, "step": 2652, "train_loss": 3.831139326095581, "perplexity": 46.11504839347065, "lr": 0.0026291804804649314, "grad_norm": 0.154317, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:39.198309+00:00", "epoch": 0, "step": 2653, "train_loss": 3.597269058227539, "perplexity": 36.498423147889014, "lr": 0.0026291804804649314, "grad_norm": 0.139389, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:39.504992+00:00", "epoch": 0, "step": 2654, "train_loss": 3.693807601928711, "perplexity": 40.19761244431375, "lr": 0.0026291804804649314, "grad_norm": 0.14335, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:39.810489+00:00", "epoch": 0, "step": 2655, "train_loss": 3.890427350997925, "perplexity": 48.93179310680358, "lr": 0.0026291804804649314, "grad_norm": 0.1664, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:40.115464+00:00", "epoch": 0, "step": 2656, "train_loss": 3.710695266723633, "perplexity": 40.88222069577063, "lr": 0.0026291804804649314, "grad_norm": 0.1685, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:40.421292+00:00", "epoch": 0, "step": 2657, "train_loss": 3.773263454437256, "perplexity": 43.52186495419215, "lr": 0.0026291804804649314, "grad_norm": 0.156849, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:40.725303+00:00", "epoch": 0, "step": 2658, "train_loss": 3.727602481842041, "perplexity": 41.57930143742635, "lr": 0.0026291804804649314, "grad_norm": 0.163325, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:41.030552+00:00", "epoch": 0, "step": 2659, "train_loss": 3.8232908248901367, "perplexity": 45.75453099181928, "lr": 0.0026291804804649314, "grad_norm": 0.162414, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:41.336821+00:00", "epoch": 0, "step": 2660, "train_loss": 3.6855099201202393, "perplexity": 39.86544546060766, "lr": 0.0026291804804649314, "grad_norm": 0.17414, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:41.642097+00:00", "epoch": 0, "step": 2661, "train_loss": 3.711179256439209, "perplexity": 40.90201205915913, "lr": 0.0026291804804649314, "grad_norm": 0.160284, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:41.948042+00:00", "epoch": 0, "step": 2662, "train_loss": 3.692631721496582, "perplexity": 40.15037263804727, "lr": 0.0026291804804649314, "grad_norm": 0.156192, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:42.252539+00:00", "epoch": 0, "step": 2663, "train_loss": 3.756598949432373, "perplexity": 42.80260432418147, "lr": 0.0026291804804649314, "grad_norm": 0.159877, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:42.557367+00:00", "epoch": 0, "step": 2664, "train_loss": 3.7255616188049316, "perplexity": 41.49453031056709, "lr": 0.0026291804804649314, "grad_norm": 0.161374, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:42.862560+00:00", "epoch": 0, "step": 2665, "train_loss": 3.7844367027282715, "perplexity": 44.010872370696994, "lr": 0.0026291804804649314, "grad_norm": 0.145103, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:43.167216+00:00", "epoch": 0, "step": 2666, "train_loss": 3.8564016819000244, "perplexity": 47.29486285580799, "lr": 0.0026291804804649314, "grad_norm": 0.144598, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:43.472300+00:00", "epoch": 0, "step": 2667, "train_loss": 3.6305150985717773, "perplexity": 37.73224743912365, "lr": 0.0026291804804649314, "grad_norm": 0.156902, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:43.777601+00:00", "epoch": 0, "step": 2668, "train_loss": 3.8058969974517822, "perplexity": 44.96556602681826, "lr": 0.0026291804804649314, "grad_norm": 0.163237, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:44.083333+00:00", "epoch": 0, "step": 2669, "train_loss": 3.72005558013916, "perplexity": 41.26668765311488, "lr": 0.0026291804804649314, "grad_norm": 0.160337, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:44.388709+00:00", "epoch": 0, "step": 2670, "train_loss": 3.771104097366333, "perplexity": 43.42798710174379, "lr": 0.0026291804804649314, "grad_norm": 0.148469, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:44.694047+00:00", "epoch": 0, "step": 2671, "train_loss": 3.6196353435516357, "perplexity": 37.323954919348324, "lr": 0.0026291804804649314, "grad_norm": 0.136673, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:44.998318+00:00", "epoch": 0, "step": 2672, "train_loss": 3.8030459880828857, "perplexity": 44.83755134903224, "lr": 0.0026291804804649314, "grad_norm": 0.148684, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:45.303452+00:00", "epoch": 0, "step": 2673, "train_loss": 3.70058536529541, "perplexity": 40.470987739370564, "lr": 0.0026291804804649314, "grad_norm": 0.155231, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:45.608621+00:00", "epoch": 0, "step": 2674, "train_loss": 3.5916531085968018, "perplexity": 36.294024325842784, "lr": 0.0026291804804649314, "grad_norm": 0.168712, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:45.914000+00:00", "epoch": 0, "step": 2675, "train_loss": 3.6254215240478516, "perplexity": 37.540544067013954, "lr": 0.0026291804804649314, "grad_norm": 0.157135, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:46.219227+00:00", "epoch": 0, "step": 2676, "train_loss": 3.6752023696899414, "perplexity": 39.45664087749193, "lr": 0.0026291804804649314, "grad_norm": 0.170823, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:46.523836+00:00", "epoch": 0, "step": 2677, "train_loss": 3.704784870147705, "perplexity": 40.6413032187957, "lr": 0.0026291804804649314, "grad_norm": 0.154227, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:46.828754+00:00", "epoch": 0, "step": 2678, "train_loss": 3.781156539916992, "perplexity": 43.866746051976925, "lr": 0.0026291804804649314, "grad_norm": 0.145378, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:47.135083+00:00", "epoch": 0, "step": 2679, "train_loss": 3.7390387058258057, "perplexity": 42.05754105716672, "lr": 0.0026291804804649314, "grad_norm": 0.157103, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:47.440726+00:00", "epoch": 0, "step": 2680, "train_loss": 3.68072772026062, "perplexity": 39.675256057251524, "lr": 0.0026291804804649314, "grad_norm": 0.158826, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:47.744811+00:00", "epoch": 0, "step": 2681, "train_loss": 3.8420932292938232, "perplexity": 46.62296492516844, "lr": 0.0026291804804649314, "grad_norm": 0.168936, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:48.049104+00:00", "epoch": 0, "step": 2682, "train_loss": 3.7873692512512207, "perplexity": 44.140125817834225, "lr": 0.0026291804804649314, "grad_norm": 0.163759, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:48.354470+00:00", "epoch": 0, "step": 2683, "train_loss": 3.6076574325561523, "perplexity": 36.87955869168334, "lr": 0.0026291804804649314, "grad_norm": 0.169689, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:48.660627+00:00", "epoch": 0, "step": 2684, "train_loss": 3.747232675552368, "perplexity": 42.403575035242355, "lr": 0.0026291804804649314, "grad_norm": 0.173065, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:48.965695+00:00", "epoch": 0, "step": 2685, "train_loss": 3.711078643798828, "perplexity": 40.89789700674558, "lr": 0.0026291804804649314, "grad_norm": 0.158028, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:49.270125+00:00", "epoch": 0, "step": 2686, "train_loss": 3.8007102012634277, "perplexity": 44.732942606996055, "lr": 0.0026291804804649314, "grad_norm": 0.154544, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:49.576322+00:00", "epoch": 0, "step": 2687, "train_loss": 3.6877477169036865, "perplexity": 39.954756118511256, "lr": 0.0026291804804649314, "grad_norm": 0.14601, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:49.882039+00:00", "epoch": 0, "step": 2688, "train_loss": 3.6822736263275146, "perplexity": 39.73663770920581, "lr": 0.0026291804804649314, "grad_norm": 0.150625, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:50.188057+00:00", "epoch": 0, "step": 2689, "train_loss": 3.7399919033050537, "perplexity": 42.09764931179258, "lr": 0.0026291804804649314, "grad_norm": 0.164676, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:50.493922+00:00", "epoch": 0, "step": 2690, "train_loss": 3.7364397048950195, "perplexity": 41.94837539109584, "lr": 0.0026291804804649314, "grad_norm": 0.190545, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:50.799285+00:00", "epoch": 0, "step": 2691, "train_loss": 3.74350643157959, "perplexity": 42.245862988609176, "lr": 0.0026291804804649314, "grad_norm": 0.190328, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:51.103643+00:00", "epoch": 0, "step": 2692, "train_loss": 3.7998619079589844, "perplexity": 44.695012041691356, "lr": 0.0026291804804649314, "grad_norm": 0.167749, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:51.409728+00:00", "epoch": 0, "step": 2693, "train_loss": 3.686042070388794, "perplexity": 39.886665513743296, "lr": 0.0026291804804649314, "grad_norm": 0.169063, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:51.715053+00:00", "epoch": 0, "step": 2694, "train_loss": 3.6103174686431885, "perplexity": 36.97779024048738, "lr": 0.0026291804804649314, "grad_norm": 0.158175, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:52.018934+00:00", "epoch": 0, "step": 2695, "train_loss": 3.629486560821533, "perplexity": 37.69345834967191, "lr": 0.0026291804804649314, "grad_norm": 0.185872, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:52.323313+00:00", "epoch": 0, "step": 2696, "train_loss": 3.7590954303741455, "perplexity": 42.9095937030408, "lr": 0.0026291804804649314, "grad_norm": 0.21325, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:52.627397+00:00", "epoch": 0, "step": 2697, "train_loss": 3.7657089233398438, "perplexity": 43.19431646738328, "lr": 0.0026291804804649314, "grad_norm": 0.144315, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:52.933969+00:00", "epoch": 0, "step": 2698, "train_loss": 3.6699061393737793, "perplexity": 39.248221824735744, "lr": 0.0026291804804649314, "grad_norm": 0.16577, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:53.239280+00:00", "epoch": 0, "step": 2699, "train_loss": 3.728308916091919, "perplexity": 41.608684857552944, "lr": 0.0026291804804649314, "grad_norm": 0.168438, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:53.545082+00:00", "epoch": 0, "step": 2700, "train_loss": 3.7071354389190674, "perplexity": 40.73694576011789, "lr": 0.0026291804804649314, "grad_norm": 0.169459, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:53.849569+00:00", "epoch": 0, "step": 2701, "train_loss": 3.774395704269409, "perplexity": 43.57117048630433, "lr": 0.0026291804804649314, "grad_norm": 0.163116, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:54.153693+00:00", "epoch": 0, "step": 2702, "train_loss": 3.7384085655212402, "perplexity": 42.03104725371905, "lr": 0.0026291804804649314, "grad_norm": 0.157405, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:54.458530+00:00", "epoch": 0, "step": 2703, "train_loss": 3.713930368423462, "perplexity": 41.014693002592175, "lr": 0.0026291804804649314, "grad_norm": 0.172572, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:54.763323+00:00", "epoch": 0, "step": 2704, "train_loss": 3.7116153240203857, "perplexity": 40.91985199004809, "lr": 0.0026291804804649314, "grad_norm": 0.143991, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:55.069940+00:00", "epoch": 0, "step": 2705, "train_loss": 3.7191390991210938, "perplexity": 41.228884842627444, "lr": 0.0026291804804649314, "grad_norm": 0.145686, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:55.375512+00:00", "epoch": 0, "step": 2706, "train_loss": 3.672585964202881, "perplexity": 39.35354123984262, "lr": 0.0026291804804649314, "grad_norm": 0.147419, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:55.680100+00:00", "epoch": 0, "step": 2707, "train_loss": 3.79902720451355, "perplexity": 44.65772052698949, "lr": 0.0026291804804649314, "grad_norm": 0.141841, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:55.984406+00:00", "epoch": 0, "step": 2708, "train_loss": 3.7065865993499756, "perplexity": 40.71459384672912, "lr": 0.0026291804804649314, "grad_norm": 0.140908, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:56.289281+00:00", "epoch": 0, "step": 2709, "train_loss": 3.681427478790283, "perplexity": 39.703028872092915, "lr": 0.0026291804804649314, "grad_norm": 0.139247, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:56.594630+00:00", "epoch": 0, "step": 2710, "train_loss": 3.692183256149292, "perplexity": 40.13237062417974, "lr": 0.0026291804804649314, "grad_norm": 0.146825, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:56.899645+00:00", "epoch": 0, "step": 2711, "train_loss": 3.711516857147217, "perplexity": 40.91582293853939, "lr": 0.0026291804804649314, "grad_norm": 0.17124, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:57.204755+00:00", "epoch": 0, "step": 2712, "train_loss": 3.771794080734253, "perplexity": 43.45796203045817, "lr": 0.0026291804804649314, "grad_norm": 0.163873, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:57.510681+00:00", "epoch": 0, "step": 2713, "train_loss": 3.6613070964813232, "perplexity": 38.912171607653534, "lr": 0.0026291804804649314, "grad_norm": 0.164406, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:57.815765+00:00", "epoch": 0, "step": 2714, "train_loss": 3.609314441680908, "perplexity": 36.94071911464799, "lr": 0.0026291804804649314, "grad_norm": 0.165668, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:58.121652+00:00", "epoch": 0, "step": 2715, "train_loss": 3.8284530639648438, "perplexity": 45.99133751962344, "lr": 0.0026291804804649314, "grad_norm": 0.168565, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:58.427061+00:00", "epoch": 0, "step": 2716, "train_loss": 3.790278673171997, "perplexity": 44.26873506603716, "lr": 0.0026291804804649314, "grad_norm": 0.166483, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:58.733007+00:00", "epoch": 0, "step": 2717, "train_loss": 3.639441967010498, "perplexity": 38.070586153687, "lr": 0.0026291804804649314, "grad_norm": 0.176668, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:59.037602+00:00", "epoch": 0, "step": 2718, "train_loss": 3.7455856800079346, "perplexity": 42.333794016378796, "lr": 0.0026291804804649314, "grad_norm": 0.167846, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:59.342231+00:00", "epoch": 0, "step": 2719, "train_loss": 3.6527888774871826, "perplexity": 38.58211694304816, "lr": 0.0026291804804649314, "grad_norm": 0.172404, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:59.648558+00:00", "epoch": 0, "step": 2720, "train_loss": 3.7409820556640625, "perplexity": 42.139353041683435, "lr": 0.0026291804804649314, "grad_norm": 0.173769, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:11:59.954073+00:00", "epoch": 0, "step": 2721, "train_loss": 3.6775972843170166, "perplexity": 39.55124940832201, "lr": 0.0026291804804649314, "grad_norm": 0.15876, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:00.258986+00:00", "epoch": 0, "step": 2722, "train_loss": 3.6984522342681885, "perplexity": 40.38474983080097, "lr": 0.0026291804804649314, "grad_norm": 0.180351, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:00.563273+00:00", "epoch": 0, "step": 2723, "train_loss": 3.71937894821167, "perplexity": 41.238774739156355, "lr": 0.0026291804804649314, "grad_norm": 0.160197, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:00.867340+00:00", "epoch": 0, "step": 2724, "train_loss": 3.726351499557495, "perplexity": 41.5273189892644, "lr": 0.0026291804804649314, "grad_norm": 0.143884, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:01.172832+00:00", "epoch": 0, "step": 2725, "train_loss": 3.842919111251831, "perplexity": 46.6614858954266, "lr": 0.0026291804804649314, "grad_norm": 0.157198, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:01.478675+00:00", "epoch": 0, "step": 2726, "train_loss": 3.670474052429199, "perplexity": 39.27051773278141, "lr": 0.0026291804804649314, "grad_norm": 0.144437, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:01.784561+00:00", "epoch": 0, "step": 2727, "train_loss": 3.669985294342041, "perplexity": 39.251328639446896, "lr": 0.0026291804804649314, "grad_norm": 0.139015, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:02.090039+00:00", "epoch": 0, "step": 2728, "train_loss": 3.7587270736694336, "perplexity": 42.89379057727618, "lr": 0.0026291804804649314, "grad_norm": 0.144961, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:02.394840+00:00", "epoch": 0, "step": 2729, "train_loss": 3.794513463973999, "perplexity": 44.45660140467491, "lr": 0.0026291804804649314, "grad_norm": 0.152237, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:02.700682+00:00", "epoch": 0, "step": 2730, "train_loss": 3.729402780532837, "perplexity": 41.65422402064679, "lr": 0.0026291804804649314, "grad_norm": 0.149224, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:03.006688+00:00", "epoch": 0, "step": 2731, "train_loss": 3.7687900066375732, "perplexity": 43.32760698856233, "lr": 0.0026291804804649314, "grad_norm": 0.155386, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:03.312478+00:00", "epoch": 0, "step": 2732, "train_loss": 3.7183117866516113, "perplexity": 41.19478977767516, "lr": 0.0026291804804649314, "grad_norm": 0.1754, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:03.617295+00:00", "epoch": 0, "step": 2733, "train_loss": 3.7291746139526367, "perplexity": 41.64472100297785, "lr": 0.0026291804804649314, "grad_norm": 0.171802, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:03.922683+00:00", "epoch": 0, "step": 2734, "train_loss": 3.7188353538513184, "perplexity": 41.2163636655988, "lr": 0.0026291804804649314, "grad_norm": 0.148, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:04.228544+00:00", "epoch": 0, "step": 2735, "train_loss": 3.6812379360198975, "perplexity": 39.69550416315741, "lr": 0.0026291804804649314, "grad_norm": 0.153694, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:04.534408+00:00", "epoch": 0, "step": 2736, "train_loss": 3.786034345626831, "perplexity": 44.08124222638151, "lr": 0.0026291804804649314, "grad_norm": 0.154563, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:04.840788+00:00", "epoch": 0, "step": 2737, "train_loss": 3.668544054031372, "perplexity": 39.19479878869943, "lr": 0.0026291804804649314, "grad_norm": 0.191309, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:05.146714+00:00", "epoch": 0, "step": 2738, "train_loss": 3.7353172302246094, "perplexity": 41.90131581878449, "lr": 0.0026291804804649314, "grad_norm": 0.206001, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:05.451481+00:00", "epoch": 0, "step": 2739, "train_loss": 3.777860641479492, "perplexity": 43.7224037117181, "lr": 0.0026291804804649314, "grad_norm": 0.180861, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:05.756762+00:00", "epoch": 0, "step": 2740, "train_loss": 3.639787197113037, "perplexity": 38.083731535008674, "lr": 0.0026291804804649314, "grad_norm": 0.160254, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:06.063289+00:00", "epoch": 0, "step": 2741, "train_loss": 3.611459732055664, "perplexity": 37.02005275021735, "lr": 0.0026291804804649314, "grad_norm": 0.168796, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:06.370183+00:00", "epoch": 0, "step": 2742, "train_loss": 3.749429941177368, "perplexity": 42.49684938978367, "lr": 0.0026291804804649314, "grad_norm": 0.151546, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:06.675207+00:00", "epoch": 0, "step": 2743, "train_loss": 3.739036798477173, "perplexity": 42.05746083884979, "lr": 0.0026291804804649314, "grad_norm": 0.140316, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:06.980912+00:00", "epoch": 0, "step": 2744, "train_loss": 3.7497522830963135, "perplexity": 42.51055011380531, "lr": 0.0026291804804649314, "grad_norm": 0.16435, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:07.286220+00:00", "epoch": 0, "step": 2745, "train_loss": 3.7256202697753906, "perplexity": 41.49696407640921, "lr": 0.0026291804804649314, "grad_norm": 0.166276, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:07.591899+00:00", "epoch": 0, "step": 2746, "train_loss": 3.771270275115967, "perplexity": 43.435204466577595, "lr": 0.0026291804804649314, "grad_norm": 0.168258, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:07.897764+00:00", "epoch": 0, "step": 2747, "train_loss": 3.6394505500793457, "perplexity": 38.07091291755135, "lr": 0.0026291804804649314, "grad_norm": 0.153744, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:08.203943+00:00", "epoch": 0, "step": 2748, "train_loss": 3.860640287399292, "perplexity": 47.495752566951374, "lr": 0.0026291804804649314, "grad_norm": 0.160156, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:08.509249+00:00", "epoch": 0, "step": 2749, "train_loss": 3.7224371433258057, "perplexity": 41.36508399931486, "lr": 0.0026291804804649314, "grad_norm": 0.176245, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:08.814463+00:00", "epoch": 0, "step": 2750, "train_loss": 3.681208848953247, "perplexity": 39.69434955417427, "lr": 0.0026291804804649314, "grad_norm": 0.169543, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:09.119342+00:00", "epoch": 0, "step": 2751, "train_loss": 3.7759292125701904, "perplexity": 43.638038496139735, "lr": 0.0026291804804649314, "grad_norm": 0.161786, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:09.424790+00:00", "epoch": 0, "step": 2752, "train_loss": 3.7997725009918213, "perplexity": 44.69101617484922, "lr": 0.0026291804804649314, "grad_norm": 0.156142, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:09.730739+00:00", "epoch": 0, "step": 2753, "train_loss": 3.7241857051849365, "perplexity": 41.43747668059163, "lr": 0.0026291804804649314, "grad_norm": 0.15511, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:10.036611+00:00", "epoch": 0, "step": 2754, "train_loss": 3.65303373336792, "perplexity": 38.59156515795125, "lr": 0.0026291804804649314, "grad_norm": 0.158613, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:10.342175+00:00", "epoch": 0, "step": 2755, "train_loss": 3.7902281284332275, "perplexity": 44.26649757093487, "lr": 0.0026291804804649314, "grad_norm": 0.150677, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:10.647035+00:00", "epoch": 0, "step": 2756, "train_loss": 3.7612526416778564, "perplexity": 43.00225867664373, "lr": 0.0026291804804649314, "grad_norm": 0.15348, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:10.952653+00:00", "epoch": 0, "step": 2757, "train_loss": 3.636836051940918, "perplexity": 37.9715065920518, "lr": 0.0026291804804649314, "grad_norm": 0.131449, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:11.258888+00:00", "epoch": 0, "step": 2758, "train_loss": 3.663145065307617, "perplexity": 38.98375673149284, "lr": 0.0026291804804649314, "grad_norm": 0.144363, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:11.565458+00:00", "epoch": 0, "step": 2759, "train_loss": 3.701920509338379, "perplexity": 40.52505842560892, "lr": 0.0026291804804649314, "grad_norm": 0.154631, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:11.871553+00:00", "epoch": 0, "step": 2760, "train_loss": 3.7837071418762207, "perplexity": 43.978775470895165, "lr": 0.0026291804804649314, "grad_norm": 0.154861, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:12.176633+00:00", "epoch": 0, "step": 2761, "train_loss": 3.660611629486084, "perplexity": 38.88511888481565, "lr": 0.0026291804804649314, "grad_norm": 0.150365, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:12.483127+00:00", "epoch": 0, "step": 2762, "train_loss": 3.794691324234009, "perplexity": 44.4645091705777, "lr": 0.0026291804804649314, "grad_norm": 0.155929, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:12.788654+00:00", "epoch": 0, "step": 2763, "train_loss": 3.7553720474243164, "perplexity": 42.75012192495001, "lr": 0.0026291804804649314, "grad_norm": 0.154463, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:13.095341+00:00", "epoch": 0, "step": 2764, "train_loss": 3.667145252227783, "perplexity": 39.1400113607723, "lr": 0.0026291804804649314, "grad_norm": 0.170813, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:13.401425+00:00", "epoch": 0, "step": 2765, "train_loss": 3.6891677379608154, "perplexity": 40.011533016186426, "lr": 0.0026291804804649314, "grad_norm": 0.173239, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:13.706508+00:00", "epoch": 0, "step": 2766, "train_loss": 3.692112922668457, "perplexity": 40.12954807412063, "lr": 0.0026291804804649314, "grad_norm": 0.171504, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:14.011668+00:00", "epoch": 0, "step": 2767, "train_loss": 3.683095932006836, "perplexity": 39.76932681044603, "lr": 0.0026291804804649314, "grad_norm": 0.189789, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:14.317538+00:00", "epoch": 0, "step": 2768, "train_loss": 3.6364030838012695, "perplexity": 37.95506969806547, "lr": 0.0026291804804649314, "grad_norm": 0.188503, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:14.624067+00:00", "epoch": 0, "step": 2769, "train_loss": 3.6645736694335938, "perplexity": 39.03948888732071, "lr": 0.0026291804804649314, "grad_norm": 0.167207, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:14.930073+00:00", "epoch": 0, "step": 2770, "train_loss": 3.7276203632354736, "perplexity": 41.580044939921414, "lr": 0.0026291804804649314, "grad_norm": 0.144544, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:15.235825+00:00", "epoch": 0, "step": 2771, "train_loss": 3.7046749591827393, "perplexity": 40.63683653941447, "lr": 0.0026291804804649314, "grad_norm": 0.15327, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:15.541256+00:00", "epoch": 0, "step": 2772, "train_loss": 3.666874885559082, "perplexity": 39.129430636689776, "lr": 0.0026291804804649314, "grad_norm": 0.15491, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:15.846983+00:00", "epoch": 0, "step": 2773, "train_loss": 3.759235382080078, "perplexity": 42.91559939412397, "lr": 0.0026291804804649314, "grad_norm": 0.152292, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:16.151692+00:00", "epoch": 0, "step": 2774, "train_loss": 3.736445665359497, "perplexity": 41.9486254236424, "lr": 0.0026291804804649314, "grad_norm": 0.166421, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:16.458224+00:00", "epoch": 0, "step": 2775, "train_loss": 3.6676547527313232, "perplexity": 39.15995829732478, "lr": 0.0026291804804649314, "grad_norm": 0.15131, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:16.764425+00:00", "epoch": 0, "step": 2776, "train_loss": 3.701176166534424, "perplexity": 40.49490511358256, "lr": 0.0026291804804649314, "grad_norm": 0.152091, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:17.070154+00:00", "epoch": 0, "step": 2777, "train_loss": 3.730806827545166, "perplexity": 41.71274958614203, "lr": 0.0026291804804649314, "grad_norm": 0.14976, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:17.374301+00:00", "epoch": 0, "step": 2778, "train_loss": 3.6754424571990967, "perplexity": 39.46611506139094, "lr": 0.0026291804804649314, "grad_norm": 0.170734, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:17.679967+00:00", "epoch": 0, "step": 2779, "train_loss": 3.7185275554656982, "perplexity": 41.20367928761707, "lr": 0.0026291804804649314, "grad_norm": 0.184948, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:17.986181+00:00", "epoch": 0, "step": 2780, "train_loss": 3.7563228607177734, "perplexity": 42.790788639335766, "lr": 0.0026291804804649314, "grad_norm": 0.218082, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:18.292397+00:00", "epoch": 0, "step": 2781, "train_loss": 3.6945903301239014, "perplexity": 40.22908856597018, "lr": 0.0026291804804649314, "grad_norm": 0.205556, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:18.597848+00:00", "epoch": 0, "step": 2782, "train_loss": 3.7119922637939453, "perplexity": 40.935279217176394, "lr": 0.0026291804804649314, "grad_norm": 0.162695, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:18.903308+00:00", "epoch": 0, "step": 2783, "train_loss": 3.681565284729004, "perplexity": 39.70850056226373, "lr": 0.0026291804804649314, "grad_norm": 0.198479, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:19.208218+00:00", "epoch": 0, "step": 2784, "train_loss": 3.725820541381836, "perplexity": 41.505275572317935, "lr": 0.0026291804804649314, "grad_norm": 0.204337, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:19.514815+00:00", "epoch": 0, "step": 2785, "train_loss": 3.667947769165039, "perplexity": 39.171434489923925, "lr": 0.0026291804804649314, "grad_norm": 0.19286, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:19.821424+00:00", "epoch": 0, "step": 2786, "train_loss": 3.8220434188842773, "perplexity": 45.69749209779044, "lr": 0.0026291804804649314, "grad_norm": 0.177078, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:20.127386+00:00", "epoch": 0, "step": 2787, "train_loss": 3.719088315963745, "perplexity": 41.22679116284346, "lr": 0.0026291804804649314, "grad_norm": 0.163618, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:20.432734+00:00", "epoch": 0, "step": 2788, "train_loss": 3.7578394412994385, "perplexity": 42.85573355311001, "lr": 0.0026291804804649314, "grad_norm": 0.172323, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:20.738006+00:00", "epoch": 0, "step": 2789, "train_loss": 3.6972391605377197, "perplexity": 40.33578985370462, "lr": 0.0026291804804649314, "grad_norm": 0.172948, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:21.044623+00:00", "epoch": 0, "step": 2790, "train_loss": 3.780853271484375, "perplexity": 43.853444669704345, "lr": 0.0026291804804649314, "grad_norm": 0.158172, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:21.350607+00:00", "epoch": 0, "step": 2791, "train_loss": 3.6698246002197266, "perplexity": 39.24502168840006, "lr": 0.0026291804804649314, "grad_norm": 0.170536, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:21.656781+00:00", "epoch": 0, "step": 2792, "train_loss": 3.659682035446167, "perplexity": 38.84898830604604, "lr": 0.0026291804804649314, "grad_norm": 0.172727, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:21.962382+00:00", "epoch": 0, "step": 2793, "train_loss": 3.7447280883789062, "perplexity": 42.297504472034205, "lr": 0.0026291804804649314, "grad_norm": 0.162504, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:22.267098+00:00", "epoch": 0, "step": 2794, "train_loss": 3.695845127105713, "perplexity": 40.27959958879575, "lr": 0.0026291804804649314, "grad_norm": 0.163652, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:22.572193+00:00", "epoch": 0, "step": 2795, "train_loss": 3.687270164489746, "perplexity": 39.9356801835204, "lr": 0.0026291804804649314, "grad_norm": 0.175999, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:22.877996+00:00", "epoch": 0, "step": 2796, "train_loss": 3.7196993827819824, "perplexity": 41.251991185610194, "lr": 0.0026291804804649314, "grad_norm": 0.141319, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:23.186085+00:00", "epoch": 0, "step": 2797, "train_loss": 3.738237142562866, "perplexity": 42.02384278477859, "lr": 0.0026291804804649314, "grad_norm": 0.157051, "tokens_per_sec": 106359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:23.492185+00:00", "epoch": 0, "step": 2798, "train_loss": 3.667069911956787, "perplexity": 39.13706265278921, "lr": 0.0026291804804649314, "grad_norm": 0.132933, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:23.797640+00:00", "epoch": 0, "step": 2799, "train_loss": 3.800271987915039, "perplexity": 44.71334432886013, "lr": 0.0026291804804649314, "grad_norm": 0.13689, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:24.103651+00:00", "epoch": 0, "step": 2800, "train_loss": 3.688040018081665, "perplexity": 39.96643664782361, "lr": 0.0026291804804649314, "grad_norm": 0.137694, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:24.408810+00:00", "epoch": 0, "step": 2801, "train_loss": 3.623574733734131, "perplexity": 37.471278533004444, "lr": 0.0026291804804649314, "grad_norm": 0.144376, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:24.713529+00:00", "epoch": 0, "step": 2802, "train_loss": 3.790844678878784, "perplexity": 44.29379851507526, "lr": 0.0026291804804649314, "grad_norm": 0.134686, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:25.021535+00:00", "epoch": 0, "step": 2803, "train_loss": 3.7244534492492676, "perplexity": 41.448572804408066, "lr": 0.0026291804804649314, "grad_norm": 0.155932, "tokens_per_sec": 106387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:25.327037+00:00", "epoch": 0, "step": 2804, "train_loss": 3.776484251022339, "perplexity": 43.662266008459916, "lr": 0.0026291804804649314, "grad_norm": 0.159591, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:25.632144+00:00", "epoch": 0, "step": 2805, "train_loss": 3.7663111686706543, "perplexity": 43.22033787764323, "lr": 0.0026291804804649314, "grad_norm": 0.16094, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:25.936951+00:00", "epoch": 0, "step": 2806, "train_loss": 3.672485589981079, "perplexity": 39.34959135700205, "lr": 0.0026291804804649314, "grad_norm": 0.173438, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:26.242560+00:00", "epoch": 0, "step": 2807, "train_loss": 3.7569355964660645, "perplexity": 42.81701611966946, "lr": 0.0026291804804649314, "grad_norm": 0.136227, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:26.548840+00:00", "epoch": 0, "step": 2808, "train_loss": 3.708618640899658, "perplexity": 40.79741170927543, "lr": 0.0026291804804649314, "grad_norm": 0.143506, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:26.855428+00:00", "epoch": 0, "step": 2809, "train_loss": 3.7337851524353027, "perplexity": 41.83716889506182, "lr": 0.0026291804804649314, "grad_norm": 0.154274, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:27.160091+00:00", "epoch": 0, "step": 2810, "train_loss": 3.6957123279571533, "perplexity": 40.27425084742804, "lr": 0.0026291804804649314, "grad_norm": 0.146334, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:27.464876+00:00", "epoch": 0, "step": 2811, "train_loss": 3.7100284099578857, "perplexity": 40.85496719840041, "lr": 0.0026291804804649314, "grad_norm": 0.168077, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:27.771064+00:00", "epoch": 0, "step": 2812, "train_loss": 3.668013095855713, "perplexity": 39.17399351369347, "lr": 0.0026291804804649314, "grad_norm": 0.153644, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:28.076861+00:00", "epoch": 0, "step": 2813, "train_loss": 3.724217414855957, "perplexity": 41.438790670178065, "lr": 0.0026291804804649314, "grad_norm": 0.165288, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:28.382617+00:00", "epoch": 0, "step": 2814, "train_loss": 3.6340925693511963, "perplexity": 37.86747519422642, "lr": 0.0026291804804649314, "grad_norm": 0.170882, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:28.688294+00:00", "epoch": 0, "step": 2815, "train_loss": 3.689790725708008, "perplexity": 40.03646747712727, "lr": 0.0026291804804649314, "grad_norm": 0.160293, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:28.993327+00:00", "epoch": 0, "step": 2816, "train_loss": 3.715271234512329, "perplexity": 41.06972510067642, "lr": 0.0026291804804649314, "grad_norm": 0.152211, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:29.298410+00:00", "epoch": 0, "step": 2817, "train_loss": 3.686671257019043, "perplexity": 39.91176956714897, "lr": 0.0026291804804649314, "grad_norm": 0.154589, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:29.604535+00:00", "epoch": 0, "step": 2818, "train_loss": 3.7305057048797607, "perplexity": 41.70019082276432, "lr": 0.0026291804804649314, "grad_norm": 0.155827, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:29.910860+00:00", "epoch": 0, "step": 2819, "train_loss": 3.7622196674346924, "perplexity": 43.04386308140951, "lr": 0.0026291804804649314, "grad_norm": 0.159506, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:30.217422+00:00", "epoch": 0, "step": 2820, "train_loss": 3.649728536605835, "perplexity": 38.46422300311357, "lr": 0.0026291804804649314, "grad_norm": 0.167475, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:30.522200+00:00", "epoch": 0, "step": 2821, "train_loss": 3.702622890472412, "perplexity": 40.55353246074459, "lr": 0.0026291804804649314, "grad_norm": 0.175414, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:30.827191+00:00", "epoch": 0, "step": 2822, "train_loss": 3.695513963699341, "perplexity": 40.26626266786059, "lr": 0.0026291804804649314, "grad_norm": 0.165923, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:31.133220+00:00", "epoch": 0, "step": 2823, "train_loss": 3.7274692058563232, "perplexity": 41.57376028430134, "lr": 0.0026291804804649314, "grad_norm": 0.141767, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:31.439314+00:00", "epoch": 0, "step": 2824, "train_loss": 3.5871410369873047, "perplexity": 36.13063198526409, "lr": 0.0026291804804649314, "grad_norm": 0.160398, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:31.745937+00:00", "epoch": 0, "step": 2825, "train_loss": 3.7490198612213135, "perplexity": 42.479425856418516, "lr": 0.0026291804804649314, "grad_norm": 0.166558, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:32.052526+00:00", "epoch": 0, "step": 2826, "train_loss": 3.697096586227417, "perplexity": 40.33003941622778, "lr": 0.0026291804804649314, "grad_norm": 0.145896, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:32.358281+00:00", "epoch": 0, "step": 2827, "train_loss": 3.7467031478881836, "perplexity": 42.381127113123284, "lr": 0.0026291804804649314, "grad_norm": 0.17362, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:32.662992+00:00", "epoch": 0, "step": 2828, "train_loss": 3.750235080718994, "perplexity": 42.53107906160393, "lr": 0.0026291804804649314, "grad_norm": 0.173887, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:32.967602+00:00", "epoch": 0, "step": 2829, "train_loss": 3.810823917388916, "perplexity": 45.18765442718645, "lr": 0.0026291804804649314, "grad_norm": 0.172611, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:33.273732+00:00", "epoch": 0, "step": 2830, "train_loss": 3.7150285243988037, "perplexity": 41.0597582726087, "lr": 0.0026291804804649314, "grad_norm": 0.190648, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:33.580013+00:00", "epoch": 0, "step": 2831, "train_loss": 3.669851303100586, "perplexity": 39.24606965753036, "lr": 0.0026291804804649314, "grad_norm": 0.1694, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:33.886185+00:00", "epoch": 0, "step": 2832, "train_loss": 3.7935047149658203, "perplexity": 44.41177846345059, "lr": 0.0026291804804649314, "grad_norm": 0.162031, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:34.191547+00:00", "epoch": 0, "step": 2833, "train_loss": 3.765404224395752, "perplexity": 43.181157209672676, "lr": 0.0026291804804649314, "grad_norm": 0.139889, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:34.497346+00:00", "epoch": 0, "step": 2834, "train_loss": 3.693364143371582, "perplexity": 40.17979042105559, "lr": 0.0026291804804649314, "grad_norm": 0.169124, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:34.803510+00:00", "epoch": 0, "step": 2835, "train_loss": 3.635214328765869, "perplexity": 37.90997722510523, "lr": 0.0026291804804649314, "grad_norm": 0.15971, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:35.109949+00:00", "epoch": 0, "step": 2836, "train_loss": 3.6893274784088135, "perplexity": 40.01792498691006, "lr": 0.0026291804804649314, "grad_norm": 0.163256, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:35.416233+00:00", "epoch": 0, "step": 2837, "train_loss": 3.697281837463379, "perplexity": 40.33751129794232, "lr": 0.0026291804804649314, "grad_norm": 0.142663, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:35.721096+00:00", "epoch": 0, "step": 2838, "train_loss": 3.6888082027435303, "perplexity": 39.997150046716506, "lr": 0.0026291804804649314, "grad_norm": 0.143608, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:36.026887+00:00", "epoch": 0, "step": 2839, "train_loss": 3.7303147315979004, "perplexity": 41.69222796083983, "lr": 0.0026291804804649314, "grad_norm": 0.147331, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:36.332282+00:00", "epoch": 0, "step": 2840, "train_loss": 3.7511696815490723, "perplexity": 42.57084722418319, "lr": 0.0026291804804649314, "grad_norm": 0.164983, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:36.638668+00:00", "epoch": 0, "step": 2841, "train_loss": 3.7207767963409424, "perplexity": 41.29646059191658, "lr": 0.0026291804804649314, "grad_norm": 0.165253, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:36.944746+00:00", "epoch": 0, "step": 2842, "train_loss": 3.748981237411499, "perplexity": 42.477785170838175, "lr": 0.0026291804804649314, "grad_norm": 0.159453, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:37.249897+00:00", "epoch": 0, "step": 2843, "train_loss": 3.780226945877075, "perplexity": 43.82598673404113, "lr": 0.0026291804804649314, "grad_norm": 0.154453, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:37.555125+00:00", "epoch": 0, "step": 2844, "train_loss": 3.646977424621582, "perplexity": 38.358549045341015, "lr": 0.0026291804804649314, "grad_norm": 0.191985, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:37.860377+00:00", "epoch": 0, "step": 2845, "train_loss": 3.6993227005004883, "perplexity": 40.41991869626281, "lr": 0.0026291804804649314, "grad_norm": 0.187816, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:38.165621+00:00", "epoch": 0, "step": 2846, "train_loss": 3.6580803394317627, "perplexity": 38.78681384190033, "lr": 0.0026291804804649314, "grad_norm": 0.175279, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:38.470681+00:00", "epoch": 0, "step": 2847, "train_loss": 3.5885086059570312, "perplexity": 36.1800769183957, "lr": 0.0026291804804649314, "grad_norm": 0.188151, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:38.775841+00:00", "epoch": 0, "step": 2848, "train_loss": 3.7203712463378906, "perplexity": 41.27971620776951, "lr": 0.0026291804804649314, "grad_norm": 0.159787, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:39.081696+00:00", "epoch": 0, "step": 2849, "train_loss": 3.7880640029907227, "perplexity": 44.17080290227295, "lr": 0.0026291804804649314, "grad_norm": 0.151407, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:39.387556+00:00", "epoch": 0, "step": 2850, "train_loss": 3.6350176334381104, "perplexity": 37.90252124301207, "lr": 0.0026291804804649314, "grad_norm": 0.152557, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:39.692865+00:00", "epoch": 0, "step": 2851, "train_loss": 3.702157497406006, "perplexity": 40.53466351899698, "lr": 0.0026291804804649314, "grad_norm": 0.144516, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:39.999867+00:00", "epoch": 0, "step": 2852, "train_loss": 3.8300371170043945, "perplexity": 46.064247969378464, "lr": 0.0026291804804649314, "grad_norm": 0.154322, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:40.306163+00:00", "epoch": 0, "step": 2853, "train_loss": 3.7370553016662598, "perplexity": 41.97420662554188, "lr": 0.0026291804804649314, "grad_norm": 0.150004, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:40.611560+00:00", "epoch": 0, "step": 2854, "train_loss": 3.7387402057647705, "perplexity": 42.04498875211929, "lr": 0.0026291804804649314, "grad_norm": 0.159129, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:40.917599+00:00", "epoch": 0, "step": 2855, "train_loss": 3.738543748855591, "perplexity": 42.03672953489926, "lr": 0.0026291804804649314, "grad_norm": 0.160247, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:41.222438+00:00", "epoch": 0, "step": 2856, "train_loss": 3.606123924255371, "perplexity": 36.823046924007556, "lr": 0.0026291804804649314, "grad_norm": 0.140549, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:41.528426+00:00", "epoch": 0, "step": 2857, "train_loss": 3.8368780612945557, "perplexity": 46.38045125488258, "lr": 0.0026291804804649314, "grad_norm": 0.139273, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:41.834597+00:00", "epoch": 0, "step": 2858, "train_loss": 3.613320827484131, "perplexity": 37.08901475368191, "lr": 0.0026291804804649314, "grad_norm": 0.144719, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:42.142039+00:00", "epoch": 0, "step": 2859, "train_loss": 3.7420384883880615, "perplexity": 42.18389395630221, "lr": 0.0026291804804649314, "grad_norm": 0.133067, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:42.448973+00:00", "epoch": 0, "step": 2860, "train_loss": 3.6282854080200195, "perplexity": 37.64820992715144, "lr": 0.0026291804804649314, "grad_norm": 0.124361, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:42.754649+00:00", "epoch": 0, "step": 2861, "train_loss": 3.722317934036255, "perplexity": 41.36015319094402, "lr": 0.0026291804804649314, "grad_norm": 0.136878, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:43.060491+00:00", "epoch": 0, "step": 2862, "train_loss": 3.816429376602173, "perplexity": 45.44166323391956, "lr": 0.0026291804804649314, "grad_norm": 0.144006, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:43.366741+00:00", "epoch": 0, "step": 2863, "train_loss": 3.7453691959381104, "perplexity": 42.32463041628154, "lr": 0.0026291804804649314, "grad_norm": 0.164736, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:43.673443+00:00", "epoch": 0, "step": 2864, "train_loss": 3.7176952362060547, "perplexity": 41.169398939853636, "lr": 0.0026291804804649314, "grad_norm": 0.187459, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:43.978171+00:00", "epoch": 0, "step": 2865, "train_loss": 3.8385472297668457, "perplexity": 46.45793268864162, "lr": 0.0026291804804649314, "grad_norm": 0.166258, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:44.283785+00:00", "epoch": 0, "step": 2866, "train_loss": 3.7024238109588623, "perplexity": 40.545459886798284, "lr": 0.0026291804804649314, "grad_norm": 0.193344, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:44.588727+00:00", "epoch": 0, "step": 2867, "train_loss": 3.710022449493408, "perplexity": 40.85472368454542, "lr": 0.0026291804804649314, "grad_norm": 0.192892, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:44.894804+00:00", "epoch": 0, "step": 2868, "train_loss": 3.6969351768493652, "perplexity": 40.32353029497954, "lr": 0.0026291804804649314, "grad_norm": 0.179992, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:45.201503+00:00", "epoch": 0, "step": 2869, "train_loss": 3.673558235168457, "perplexity": 39.391822152076855, "lr": 0.0026291804804649314, "grad_norm": 0.181941, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:45.507687+00:00", "epoch": 0, "step": 2870, "train_loss": 3.715853214263916, "perplexity": 41.09363380560451, "lr": 0.0026291804804649314, "grad_norm": 0.189102, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:45.813431+00:00", "epoch": 0, "step": 2871, "train_loss": 3.6863315105438232, "perplexity": 39.89821198731914, "lr": 0.0026291804804649314, "grad_norm": 0.189922, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:46.118996+00:00", "epoch": 0, "step": 2872, "train_loss": 3.771392822265625, "perplexity": 43.44052765324381, "lr": 0.0026291804804649314, "grad_norm": 0.203349, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:46.425007+00:00", "epoch": 0, "step": 2873, "train_loss": 3.65246319770813, "perplexity": 38.5695535736555, "lr": 0.0026291804804649314, "grad_norm": 0.179227, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:46.730287+00:00", "epoch": 0, "step": 2874, "train_loss": 3.781449556350708, "perplexity": 43.879601612817105, "lr": 0.0026291804804649314, "grad_norm": 0.165101, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:47.036192+00:00", "epoch": 0, "step": 2875, "train_loss": 3.645291566848755, "perplexity": 38.29393646639082, "lr": 0.0026291804804649314, "grad_norm": 0.14721, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:47.342515+00:00", "epoch": 0, "step": 2876, "train_loss": 3.787022113800049, "perplexity": 44.1248057862956, "lr": 0.0026291804804649314, "grad_norm": 0.144956, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:47.647674+00:00", "epoch": 0, "step": 2877, "train_loss": 3.6342933177948, "perplexity": 37.875077794014366, "lr": 0.0026291804804649314, "grad_norm": 0.142143, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:47.952963+00:00", "epoch": 0, "step": 2878, "train_loss": 3.667970657348633, "perplexity": 39.172331063168585, "lr": 0.0026291804804649314, "grad_norm": 0.158009, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:48.260690+00:00", "epoch": 0, "step": 2879, "train_loss": 3.6845812797546387, "perplexity": 39.82844198288107, "lr": 0.0026291804804649314, "grad_norm": 0.166317, "tokens_per_sec": 106486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:48.567412+00:00", "epoch": 0, "step": 2880, "train_loss": 3.822364568710327, "perplexity": 45.712170196236784, "lr": 0.0026291804804649314, "grad_norm": 0.155422, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:48.873536+00:00", "epoch": 0, "step": 2881, "train_loss": 3.616337299346924, "perplexity": 37.20106163127139, "lr": 0.0026291804804649314, "grad_norm": 0.152302, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:49.178797+00:00", "epoch": 0, "step": 2882, "train_loss": 3.6986019611358643, "perplexity": 40.39079696559298, "lr": 0.0026291804804649314, "grad_norm": 0.153157, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:49.485886+00:00", "epoch": 0, "step": 2883, "train_loss": 3.7317588329315186, "perplexity": 41.75247925685724, "lr": 0.0026291804804649314, "grad_norm": 0.153788, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:49.793578+00:00", "epoch": 0, "step": 2884, "train_loss": 3.654132843017578, "perplexity": 38.63400483827044, "lr": 0.0026291804804649314, "grad_norm": 0.138413, "tokens_per_sec": 106497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:50.100153+00:00", "epoch": 0, "step": 2885, "train_loss": 3.6605043411254883, "perplexity": 38.880947187950184, "lr": 0.0026291804804649314, "grad_norm": 0.152903, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:50.406514+00:00", "epoch": 0, "step": 2886, "train_loss": 3.6653738021850586, "perplexity": 39.07073816109642, "lr": 0.0026291804804649314, "grad_norm": 0.155671, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:50.711791+00:00", "epoch": 0, "step": 2887, "train_loss": 3.719432830810547, "perplexity": 41.240996851379855, "lr": 0.0026291804804649314, "grad_norm": 0.16592, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:51.019378+00:00", "epoch": 0, "step": 2888, "train_loss": 3.7561473846435547, "perplexity": 42.783280538497905, "lr": 0.0026291804804649314, "grad_norm": 0.154116, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:51.325783+00:00", "epoch": 0, "step": 2889, "train_loss": 3.7337441444396973, "perplexity": 41.83545327180101, "lr": 0.0026291804804649314, "grad_norm": 0.15017, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:51.632364+00:00", "epoch": 0, "step": 2890, "train_loss": 3.746488094329834, "perplexity": 42.37201388088237, "lr": 0.0026291804804649314, "grad_norm": 0.17215, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:51.938480+00:00", "epoch": 0, "step": 2891, "train_loss": 3.6783828735351562, "perplexity": 39.58233265115277, "lr": 0.0026291804804649314, "grad_norm": 0.14961, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:52.244374+00:00", "epoch": 0, "step": 2892, "train_loss": 3.6603586673736572, "perplexity": 38.87528366702176, "lr": 0.0026291804804649314, "grad_norm": 0.151995, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:52.550682+00:00", "epoch": 0, "step": 2893, "train_loss": 3.6548118591308594, "perplexity": 38.66024685844521, "lr": 0.0026291804804649314, "grad_norm": 0.151632, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:52.857310+00:00", "epoch": 0, "step": 2894, "train_loss": 3.6275575160980225, "perplexity": 37.62081607036085, "lr": 0.0026291804804649314, "grad_norm": 0.161952, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:53.163370+00:00", "epoch": 0, "step": 2895, "train_loss": 3.7594711780548096, "perplexity": 42.92571991285619, "lr": 0.0026291804804649314, "grad_norm": 0.157273, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:53.468470+00:00", "epoch": 0, "step": 2896, "train_loss": 3.6772284507751465, "perplexity": 39.53666427082665, "lr": 0.0026291804804649314, "grad_norm": 0.166717, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:53.774011+00:00", "epoch": 0, "step": 2897, "train_loss": 3.7260310649871826, "perplexity": 41.51401433239769, "lr": 0.0026291804804649314, "grad_norm": 0.174816, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:54.080065+00:00", "epoch": 0, "step": 2898, "train_loss": 3.7927143573760986, "perplexity": 44.37669114485973, "lr": 0.0026291804804649314, "grad_norm": 0.176393, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:54.386418+00:00", "epoch": 0, "step": 2899, "train_loss": 3.7746939659118652, "perplexity": 43.58416803341583, "lr": 0.0026291804804649314, "grad_norm": 0.150477, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:54.693797+00:00", "epoch": 0, "step": 2900, "train_loss": 3.6995584964752197, "perplexity": 40.429450674147205, "lr": 0.0026291804804649314, "grad_norm": 0.159134, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:54.999994+00:00", "epoch": 0, "step": 2901, "train_loss": 3.706186056137085, "perplexity": 40.698289158082396, "lr": 0.0026291804804649314, "grad_norm": 0.176606, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:55.305917+00:00", "epoch": 0, "step": 2902, "train_loss": 3.794572591781616, "perplexity": 44.45923010376392, "lr": 0.0026291804804649314, "grad_norm": 0.193706, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:55.612125+00:00", "epoch": 0, "step": 2903, "train_loss": 3.58073353767395, "perplexity": 35.899865094505316, "lr": 0.0026291804804649314, "grad_norm": 0.190136, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:55.918346+00:00", "epoch": 0, "step": 2904, "train_loss": 3.627088785171509, "perplexity": 37.60318616255229, "lr": 0.0026291804804649314, "grad_norm": 0.174366, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:56.224725+00:00", "epoch": 0, "step": 2905, "train_loss": 3.8220438957214355, "perplexity": 45.697513888057905, "lr": 0.0026291804804649314, "grad_norm": 0.154987, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:56.530114+00:00", "epoch": 0, "step": 2906, "train_loss": 3.553189992904663, "perplexity": 34.924549042986634, "lr": 0.0026291804804649314, "grad_norm": 0.179915, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:56.836243+00:00", "epoch": 0, "step": 2907, "train_loss": 3.5509448051452637, "perplexity": 34.84622483215432, "lr": 0.0026291804804649314, "grad_norm": 0.181011, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:57.142121+00:00", "epoch": 0, "step": 2908, "train_loss": 3.7089290618896484, "perplexity": 40.810078048054265, "lr": 0.0026291804804649314, "grad_norm": 0.148223, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:57.448488+00:00", "epoch": 0, "step": 2909, "train_loss": 3.7935774326324463, "perplexity": 44.41500810177562, "lr": 0.0026291804804649314, "grad_norm": 0.155585, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:57.754068+00:00", "epoch": 0, "step": 2910, "train_loss": 3.634359121322632, "perplexity": 37.87757018975344, "lr": 0.0026291804804649314, "grad_norm": 0.15214, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:58.058801+00:00", "epoch": 0, "step": 2911, "train_loss": 3.7654807567596436, "perplexity": 43.18446209217313, "lr": 0.0026291804804649314, "grad_norm": 0.142605, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:58.363366+00:00", "epoch": 0, "step": 2912, "train_loss": 3.739558458328247, "perplexity": 42.07940625113013, "lr": 0.0026291804804649314, "grad_norm": 0.143544, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:58.668888+00:00", "epoch": 0, "step": 2913, "train_loss": 3.7012991905212402, "perplexity": 40.49988726471115, "lr": 0.0026291804804649314, "grad_norm": 0.170919, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:58.974587+00:00", "epoch": 0, "step": 2914, "train_loss": 3.671999216079712, "perplexity": 39.330457396243304, "lr": 0.0026291804804649314, "grad_norm": 0.168926, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:59.280894+00:00", "epoch": 0, "step": 2915, "train_loss": 3.6832852363586426, "perplexity": 39.77685602971417, "lr": 0.0026291804804649314, "grad_norm": 0.174122, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:59.585927+00:00", "epoch": 0, "step": 2916, "train_loss": 3.694652557373047, "perplexity": 40.23159198937704, "lr": 0.0026291804804649314, "grad_norm": 0.182761, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:12:59.891988+00:00", "epoch": 0, "step": 2917, "train_loss": 3.651134967803955, "perplexity": 38.51835834625082, "lr": 0.0026291804804649314, "grad_norm": 0.154812, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:00.198026+00:00", "epoch": 0, "step": 2918, "train_loss": 3.75787091255188, "perplexity": 42.85708229794245, "lr": 0.0026291804804649314, "grad_norm": 0.149836, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:00.504546+00:00", "epoch": 0, "step": 2919, "train_loss": 3.6888322830200195, "perplexity": 39.998113200744875, "lr": 0.0026291804804649314, "grad_norm": 0.148594, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:00.811583+00:00", "epoch": 0, "step": 2920, "train_loss": 3.607532262802124, "perplexity": 36.87494277528582, "lr": 0.0026291804804649314, "grad_norm": 0.143115, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:01.116411+00:00", "epoch": 0, "step": 2921, "train_loss": 3.5732407569885254, "perplexity": 35.6318805073446, "lr": 0.0026291804804649314, "grad_norm": 0.158918, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:01.422065+00:00", "epoch": 0, "step": 2922, "train_loss": 3.719792127609253, "perplexity": 41.25581727182938, "lr": 0.0026291804804649314, "grad_norm": 0.152349, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:01.728098+00:00", "epoch": 0, "step": 2923, "train_loss": 3.601968765258789, "perplexity": 36.670358750802436, "lr": 0.0026291804804649314, "grad_norm": 0.15469, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:02.034205+00:00", "epoch": 0, "step": 2924, "train_loss": 3.7676174640655518, "perplexity": 43.27683329779862, "lr": 0.0026291804804649314, "grad_norm": 0.16421, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:02.340432+00:00", "epoch": 0, "step": 2925, "train_loss": 3.674253463745117, "perplexity": 39.419217994606534, "lr": 0.0026291804804649314, "grad_norm": 0.169158, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:02.648050+00:00", "epoch": 0, "step": 2926, "train_loss": 3.706667900085449, "perplexity": 40.71790410771488, "lr": 0.0026291804804649314, "grad_norm": 0.133814, "tokens_per_sec": 106523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:02.953722+00:00", "epoch": 0, "step": 2927, "train_loss": 3.7167184352874756, "perplexity": 41.12920426744425, "lr": 0.0026291804804649314, "grad_norm": 0.150967, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:03.258792+00:00", "epoch": 0, "step": 2928, "train_loss": 3.607391119003296, "perplexity": 36.86973847306708, "lr": 0.0026291804804649314, "grad_norm": 0.152315, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:03.564506+00:00", "epoch": 0, "step": 2929, "train_loss": 3.689391851425171, "perplexity": 40.020501144366456, "lr": 0.0026291804804649314, "grad_norm": 0.161995, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:03.870581+00:00", "epoch": 0, "step": 2930, "train_loss": 3.756817579269409, "perplexity": 42.8119632736251, "lr": 0.0026291804804649314, "grad_norm": 0.151978, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:04.176277+00:00", "epoch": 0, "step": 2931, "train_loss": 3.6194546222686768, "perplexity": 37.31721029579707, "lr": 0.0026291804804649314, "grad_norm": 0.179836, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:04.481952+00:00", "epoch": 0, "step": 2932, "train_loss": 3.779254198074341, "perplexity": 43.7833758299346, "lr": 0.0026291804804649314, "grad_norm": 0.170747, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:04.787435+00:00", "epoch": 0, "step": 2933, "train_loss": 3.7016868591308594, "perplexity": 40.51559084339238, "lr": 0.0026291804804649314, "grad_norm": 0.151521, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:05.094184+00:00", "epoch": 0, "step": 2934, "train_loss": 3.6816837787628174, "perplexity": 39.71320606145332, "lr": 0.0026291804804649314, "grad_norm": 0.169996, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:05.399323+00:00", "epoch": 0, "step": 2935, "train_loss": 3.6722137928009033, "perplexity": 39.33889770234857, "lr": 0.0026291804804649314, "grad_norm": 0.153031, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:05.704430+00:00", "epoch": 0, "step": 2936, "train_loss": 3.7329697608947754, "perplexity": 41.80306912568623, "lr": 0.0026291804804649314, "grad_norm": 0.172884, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:06.010452+00:00", "epoch": 0, "step": 2937, "train_loss": 3.6823136806488037, "perplexity": 39.7382293651357, "lr": 0.0026291804804649314, "grad_norm": 0.168332, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:06.315863+00:00", "epoch": 0, "step": 2938, "train_loss": 3.6730661392211914, "perplexity": 39.37244236478889, "lr": 0.0026291804804649314, "grad_norm": 0.156571, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:06.621416+00:00", "epoch": 0, "step": 2939, "train_loss": 3.6248505115509033, "perplexity": 37.51911406619087, "lr": 0.0026291804804649314, "grad_norm": 0.162029, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:06.927873+00:00", "epoch": 0, "step": 2940, "train_loss": 3.6712000370025635, "perplexity": 39.29903787418214, "lr": 0.0026291804804649314, "grad_norm": 0.187031, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:07.233974+00:00", "epoch": 0, "step": 2941, "train_loss": 3.6777946949005127, "perplexity": 39.55905801427107, "lr": 0.0026291804804649314, "grad_norm": 0.208605, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:07.539480+00:00", "epoch": 0, "step": 2942, "train_loss": 3.6881866455078125, "perplexity": 39.97229725321381, "lr": 0.0026291804804649314, "grad_norm": 0.183797, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:07.845019+00:00", "epoch": 0, "step": 2943, "train_loss": 3.612856864929199, "perplexity": 37.07181083093348, "lr": 0.0026291804804649314, "grad_norm": 0.164471, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:08.150049+00:00", "epoch": 0, "step": 2944, "train_loss": 3.6893150806427, "perplexity": 40.01742885711119, "lr": 0.0026291804804649314, "grad_norm": 0.164493, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:08.454970+00:00", "epoch": 0, "step": 2945, "train_loss": 3.734255313873291, "perplexity": 41.85684374336594, "lr": 0.0026291804804649314, "grad_norm": 0.163869, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:08.760412+00:00", "epoch": 0, "step": 2946, "train_loss": 3.7915894985198975, "perplexity": 44.32680169537153, "lr": 0.0026291804804649314, "grad_norm": 0.164813, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:09.066566+00:00", "epoch": 0, "step": 2947, "train_loss": 3.6616742610931396, "perplexity": 38.92646140322988, "lr": 0.0026291804804649314, "grad_norm": 0.1631, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:09.371951+00:00", "epoch": 0, "step": 2948, "train_loss": 3.6723620891571045, "perplexity": 39.34473195012295, "lr": 0.0026291804804649314, "grad_norm": 0.176274, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:09.677161+00:00", "epoch": 0, "step": 2949, "train_loss": 3.6495368480682373, "perplexity": 38.456850559085346, "lr": 0.0026291804804649314, "grad_norm": 0.145968, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:09.981402+00:00", "epoch": 0, "step": 2950, "train_loss": 3.650115728378296, "perplexity": 38.4791189173961, "lr": 0.0026291804804649314, "grad_norm": 0.161933, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:10.286443+00:00", "epoch": 0, "step": 2951, "train_loss": 3.653371572494507, "perplexity": 38.604605101195325, "lr": 0.0026291804804649314, "grad_norm": 0.163681, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:10.592690+00:00", "epoch": 0, "step": 2952, "train_loss": 3.7171199321746826, "perplexity": 41.14572083038337, "lr": 0.0026291804804649314, "grad_norm": 0.151068, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:10.899475+00:00", "epoch": 0, "step": 2953, "train_loss": 3.8033652305603027, "perplexity": 44.85186768507592, "lr": 0.0026291804804649314, "grad_norm": 0.147957, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:11.204729+00:00", "epoch": 0, "step": 2954, "train_loss": 3.564483404159546, "perplexity": 35.32120190437748, "lr": 0.0026291804804649314, "grad_norm": 0.158725, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:11.509622+00:00", "epoch": 0, "step": 2955, "train_loss": 3.7356276512145996, "perplexity": 41.91432488576211, "lr": 0.0026291804804649314, "grad_norm": 0.177168, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:11.815249+00:00", "epoch": 0, "step": 2956, "train_loss": 3.6753604412078857, "perplexity": 39.46287834157814, "lr": 0.0026291804804649314, "grad_norm": 0.168896, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:12.121625+00:00", "epoch": 0, "step": 2957, "train_loss": 3.7040932178497314, "perplexity": 40.61320328684324, "lr": 0.0026291804804649314, "grad_norm": 0.191913, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:12.427425+00:00", "epoch": 0, "step": 2958, "train_loss": 3.652841091156006, "perplexity": 38.58413150951822, "lr": 0.0026291804804649314, "grad_norm": 0.16365, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:12.733487+00:00", "epoch": 0, "step": 2959, "train_loss": 3.62032151222229, "perplexity": 37.34957423646009, "lr": 0.0026291804804649314, "grad_norm": 0.158267, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:13.038045+00:00", "epoch": 0, "step": 2960, "train_loss": 3.741661310195923, "perplexity": 42.16798611167703, "lr": 0.0026291804804649314, "grad_norm": 0.16857, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:13.342980+00:00", "epoch": 0, "step": 2961, "train_loss": 3.73945951461792, "perplexity": 42.07524296451619, "lr": 0.0026291804804649314, "grad_norm": 0.164635, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:13.648928+00:00", "epoch": 0, "step": 2962, "train_loss": 3.663856029510498, "perplexity": 39.01148264191976, "lr": 0.0026291804804649314, "grad_norm": 0.150297, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:13.955074+00:00", "epoch": 0, "step": 2963, "train_loss": 3.791787624359131, "perplexity": 44.33558485021419, "lr": 0.0026291804804649314, "grad_norm": 0.150387, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:14.261047+00:00", "epoch": 0, "step": 2964, "train_loss": 3.6395792961120605, "perplexity": 38.07581471208785, "lr": 0.0026291804804649314, "grad_norm": 0.154751, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:14.566551+00:00", "epoch": 0, "step": 2965, "train_loss": 3.6787476539611816, "perplexity": 39.59677414514736, "lr": 0.0026291804804649314, "grad_norm": 0.141005, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:14.871014+00:00", "epoch": 0, "step": 2966, "train_loss": 3.6471593379974365, "perplexity": 38.36552761321891, "lr": 0.0026291804804649314, "grad_norm": 0.163028, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:15.175703+00:00", "epoch": 0, "step": 2967, "train_loss": 3.716043710708618, "perplexity": 41.101462742413005, "lr": 0.0026291804804649314, "grad_norm": 0.153083, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:15.481840+00:00", "epoch": 0, "step": 2968, "train_loss": 3.6866321563720703, "perplexity": 39.91020902164644, "lr": 0.0026291804804649314, "grad_norm": 0.150872, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:15.787222+00:00", "epoch": 0, "step": 2969, "train_loss": 3.7224411964416504, "perplexity": 41.365251657132006, "lr": 0.0026291804804649314, "grad_norm": 0.158557, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:16.092666+00:00", "epoch": 0, "step": 2970, "train_loss": 3.6662962436676025, "perplexity": 39.106795258454255, "lr": 0.0026291804804649314, "grad_norm": 0.162652, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:16.397754+00:00", "epoch": 0, "step": 2971, "train_loss": 3.7234818935394287, "perplexity": 41.408322762580326, "lr": 0.0026291804804649314, "grad_norm": 0.156965, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:16.702992+00:00", "epoch": 0, "step": 2972, "train_loss": 3.587153196334839, "perplexity": 36.13107131284599, "lr": 0.0026291804804649314, "grad_norm": 0.155755, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:17.009265+00:00", "epoch": 0, "step": 2973, "train_loss": 3.7818868160247803, "perplexity": 43.89879258853087, "lr": 0.0026291804804649314, "grad_norm": 0.163999, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:17.315099+00:00", "epoch": 0, "step": 2974, "train_loss": 3.6368165016174316, "perplexity": 37.97076424407126, "lr": 0.0026291804804649314, "grad_norm": 0.14832, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:17.621723+00:00", "epoch": 0, "step": 2975, "train_loss": 3.7112696170806885, "perplexity": 40.90570815819502, "lr": 0.0026291804804649314, "grad_norm": 0.174413, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:17.926855+00:00", "epoch": 0, "step": 2976, "train_loss": 3.7311794757843018, "perplexity": 41.728296665440766, "lr": 0.0026291804804649314, "grad_norm": 0.151938, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:18.232118+00:00", "epoch": 0, "step": 2977, "train_loss": 3.578523874282837, "perplexity": 35.820626054903684, "lr": 0.0026291804804649314, "grad_norm": 0.160592, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:18.537669+00:00", "epoch": 0, "step": 2978, "train_loss": 3.722641706466675, "perplexity": 41.37354663636237, "lr": 0.0026291804804649314, "grad_norm": 0.157635, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:18.843732+00:00", "epoch": 0, "step": 2979, "train_loss": 3.722139358520508, "perplexity": 41.35276793968872, "lr": 0.0026291804804649314, "grad_norm": 0.155894, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:19.151093+00:00", "epoch": 0, "step": 2980, "train_loss": 3.7287750244140625, "perplexity": 41.62808353242919, "lr": 0.0026291804804649314, "grad_norm": 0.152981, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:19.456005+00:00", "epoch": 0, "step": 2981, "train_loss": 3.702615737915039, "perplexity": 40.553242400314325, "lr": 0.0026291804804649314, "grad_norm": 0.166378, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:19.761379+00:00", "epoch": 0, "step": 2982, "train_loss": 3.7280752658843994, "perplexity": 41.5989641153723, "lr": 0.0026291804804649314, "grad_norm": 0.178223, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:20.067060+00:00", "epoch": 0, "step": 2983, "train_loss": 3.788271427154541, "perplexity": 44.179965944415464, "lr": 0.0026291804804649314, "grad_norm": 0.165885, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:20.372859+00:00", "epoch": 0, "step": 2984, "train_loss": 3.7790164947509766, "perplexity": 43.77296961283691, "lr": 0.0026291804804649314, "grad_norm": 0.166196, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:20.679063+00:00", "epoch": 0, "step": 2985, "train_loss": 3.6891415119171143, "perplexity": 40.01048368573295, "lr": 0.0026291804804649314, "grad_norm": 0.166595, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:20.985529+00:00", "epoch": 0, "step": 2986, "train_loss": 3.6074187755584717, "perplexity": 36.87075817712416, "lr": 0.0026291804804649314, "grad_norm": 0.170136, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:21.290818+00:00", "epoch": 0, "step": 2987, "train_loss": 3.7418267726898193, "perplexity": 42.17496390908762, "lr": 0.0026291804804649314, "grad_norm": 0.176957, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:21.597446+00:00", "epoch": 0, "step": 2988, "train_loss": 3.7236685752868652, "perplexity": 41.41605366221853, "lr": 0.0026291804804649314, "grad_norm": 0.173016, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:21.903228+00:00", "epoch": 0, "step": 2989, "train_loss": 3.6185832023620605, "perplexity": 37.28470550060566, "lr": 0.0026291804804649314, "grad_norm": 0.186509, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:22.209905+00:00", "epoch": 0, "step": 2990, "train_loss": 3.696963310241699, "perplexity": 40.324664748635556, "lr": 0.0026291804804649314, "grad_norm": 0.164403, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:22.515708+00:00", "epoch": 0, "step": 2991, "train_loss": 3.6705775260925293, "perplexity": 39.27458140735011, "lr": 0.0026291804804649314, "grad_norm": 0.157651, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:22.822257+00:00", "epoch": 0, "step": 2992, "train_loss": 3.730912446975708, "perplexity": 41.71715549567039, "lr": 0.0026291804804649314, "grad_norm": 0.166705, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:23.127688+00:00", "epoch": 0, "step": 2993, "train_loss": 3.6151933670043945, "perplexity": 37.15853046472141, "lr": 0.0026291804804649314, "grad_norm": 0.159405, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:23.433729+00:00", "epoch": 0, "step": 2994, "train_loss": 3.6741573810577393, "perplexity": 39.41543067215825, "lr": 0.0026291804804649314, "grad_norm": 0.154823, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:23.739205+00:00", "epoch": 0, "step": 2995, "train_loss": 3.5452511310577393, "perplexity": 34.64838553547924, "lr": 0.0026291804804649314, "grad_norm": 0.154581, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:24.045029+00:00", "epoch": 0, "step": 2996, "train_loss": 3.579493284225464, "perplexity": 35.85536776271019, "lr": 0.0026291804804649314, "grad_norm": 0.137503, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:24.351318+00:00", "epoch": 0, "step": 2997, "train_loss": 3.6363251209259033, "perplexity": 37.95211072704353, "lr": 0.0026291804804649314, "grad_norm": 0.156218, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:24.656428+00:00", "epoch": 0, "step": 2998, "train_loss": 3.7586655616760254, "perplexity": 42.891152175860434, "lr": 0.0026291804804649314, "grad_norm": 0.184665, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:24.962356+00:00", "epoch": 0, "step": 2999, "train_loss": 3.692194700241089, "perplexity": 40.13282990534121, "lr": 0.0026291804804649314, "grad_norm": 0.146255, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:25.268630+00:00", "epoch": 0, "step": 3000, "train_loss": 3.7677595615386963, "perplexity": 43.282983263392865, "lr": 0.0026291804804649314, "grad_norm": 0.173937, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:13:28.341608+00:00", "step": 3000, "epoch": 0, "val_loss": 3.6426683902740478, "val_ppl": 38.19361634552006, "eval_train_loss": 3.7677595615386963, "eval_train_ppl": 43.282983263392865} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:13:29.177242+00:00", "step": 3000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p6427_epoch_0000_step_0003000.pt", "val_loss": 3.6426683902740478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:30.210251+00:00", "epoch": 0, "step": 3001, "train_loss": 3.7165091037750244, "perplexity": 41.120595529980484, "lr": 0.0026291804804649314, "grad_norm": 0.166692, "tokens_per_sec": 6631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:30.514720+00:00", "epoch": 0, "step": 3002, "train_loss": 3.5991315841674805, "perplexity": 36.56646575365805, "lr": 0.0026291804804649314, "grad_norm": 0.149408, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:30.818168+00:00", "epoch": 0, "step": 3003, "train_loss": 3.662341356277466, "perplexity": 38.952437721549934, "lr": 0.0026291804804649314, "grad_norm": 0.163138, "tokens_per_sec": 107981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:31.122250+00:00", "epoch": 0, "step": 3004, "train_loss": 3.6117563247680664, "perplexity": 37.03103425651269, "lr": 0.0026291804804649314, "grad_norm": 0.137961, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:31.426773+00:00", "epoch": 0, "step": 3005, "train_loss": 3.7049403190612793, "perplexity": 40.64762135628838, "lr": 0.0026291804804649314, "grad_norm": 0.165268, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:31.730626+00:00", "epoch": 0, "step": 3006, "train_loss": 3.757190227508545, "perplexity": 42.82792004930049, "lr": 0.0026291804804649314, "grad_norm": 0.182154, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:32.034070+00:00", "epoch": 0, "step": 3007, "train_loss": 3.719302177429199, "perplexity": 41.235608927673965, "lr": 0.0026291804804649314, "grad_norm": 0.218778, "tokens_per_sec": 107985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:32.337775+00:00", "epoch": 0, "step": 3008, "train_loss": 3.634193181991577, "perplexity": 37.871285332561094, "lr": 0.0026291804804649314, "grad_norm": 0.173269, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:32.642418+00:00", "epoch": 0, "step": 3009, "train_loss": 3.669727087020874, "perplexity": 39.24119496737712, "lr": 0.0026291804804649314, "grad_norm": 0.16795, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:32.947288+00:00", "epoch": 0, "step": 3010, "train_loss": 3.7652807235717773, "perplexity": 43.175824630472846, "lr": 0.0026291804804649314, "grad_norm": 0.170855, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:33.252708+00:00", "epoch": 0, "step": 3011, "train_loss": 3.798640727996826, "perplexity": 44.64046470141587, "lr": 0.0026291804804649314, "grad_norm": 0.162801, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:33.557254+00:00", "epoch": 0, "step": 3012, "train_loss": 3.7379283905029297, "perplexity": 42.01086983956718, "lr": 0.0026291804804649314, "grad_norm": 0.148898, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:33.861553+00:00", "epoch": 0, "step": 3013, "train_loss": 3.7084121704101562, "perplexity": 40.788989117247674, "lr": 0.0026291804804649314, "grad_norm": 0.154601, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:34.167340+00:00", "epoch": 0, "step": 3014, "train_loss": 3.6488475799560547, "perplexity": 38.43035261144425, "lr": 0.0026291804804649314, "grad_norm": 0.160904, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:34.567744+00:00", "epoch": 0, "step": 3015, "train_loss": 3.733799695968628, "perplexity": 41.83777735974648, "lr": 0.0026291804804649314, "grad_norm": 0.139922, "tokens_per_sec": 81837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:34.871655+00:00", "epoch": 0, "step": 3016, "train_loss": 3.6140670776367188, "perplexity": 37.11670276640811, "lr": 0.0026291804804649314, "grad_norm": 0.167473, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:35.176436+00:00", "epoch": 0, "step": 3017, "train_loss": 3.775966167449951, "perplexity": 43.639651164403155, "lr": 0.0026291804804649314, "grad_norm": 0.178025, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:35.481119+00:00", "epoch": 0, "step": 3018, "train_loss": 3.726958990097046, "perplexity": 41.552554106954844, "lr": 0.0026291804804649314, "grad_norm": 0.179996, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:35.786396+00:00", "epoch": 0, "step": 3019, "train_loss": 3.6954116821289062, "perplexity": 40.26214438189536, "lr": 0.0026291804804649314, "grad_norm": 0.151302, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:36.091849+00:00", "epoch": 0, "step": 3020, "train_loss": 3.734922170639038, "perplexity": 41.88476557170436, "lr": 0.0026291804804649314, "grad_norm": 0.163425, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:36.395709+00:00", "epoch": 0, "step": 3021, "train_loss": 3.685027837753296, "perplexity": 39.846231663989144, "lr": 0.0026291804804649314, "grad_norm": 0.153153, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:36.699012+00:00", "epoch": 0, "step": 3022, "train_loss": 3.5981693267822266, "perplexity": 36.5312963256686, "lr": 0.0026291804804649314, "grad_norm": 0.156847, "tokens_per_sec": 108037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:37.004899+00:00", "epoch": 0, "step": 3023, "train_loss": 3.830042600631714, "perplexity": 46.064500569239655, "lr": 0.0026291804804649314, "grad_norm": 0.155445, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:37.310213+00:00", "epoch": 0, "step": 3024, "train_loss": 3.6767890453338623, "perplexity": 39.51929546166991, "lr": 0.0026291804804649314, "grad_norm": 0.161125, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:37.615827+00:00", "epoch": 0, "step": 3025, "train_loss": 3.668459415435791, "perplexity": 39.19148153636164, "lr": 0.0026291804804649314, "grad_norm": 0.164969, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:37.920375+00:00", "epoch": 0, "step": 3026, "train_loss": 3.590991973876953, "perplexity": 36.27003701653319, "lr": 0.0026291804804649314, "grad_norm": 0.163241, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:38.224695+00:00", "epoch": 0, "step": 3027, "train_loss": 3.710700273513794, "perplexity": 40.882425384983385, "lr": 0.0026291804804649314, "grad_norm": 0.191013, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:38.530535+00:00", "epoch": 0, "step": 3028, "train_loss": 3.5804245471954346, "perplexity": 35.88877409160651, "lr": 0.0026291804804649314, "grad_norm": 0.190868, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:38.836655+00:00", "epoch": 0, "step": 3029, "train_loss": 3.724471092224121, "perplexity": 41.44930408698675, "lr": 0.0026291804804649314, "grad_norm": 0.157605, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:39.142890+00:00", "epoch": 0, "step": 3030, "train_loss": 3.669856071472168, "perplexity": 39.2462567978198, "lr": 0.0026291804804649314, "grad_norm": 0.15796, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:39.461593+00:00", "epoch": 0, "step": 3031, "train_loss": 3.609795331954956, "perplexity": 36.95848781924274, "lr": 0.0026291804804649314, "grad_norm": 0.151899, "tokens_per_sec": 102764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:39.766873+00:00", "epoch": 0, "step": 3032, "train_loss": 3.788536548614502, "perplexity": 44.19168055431629, "lr": 0.0026291804804649314, "grad_norm": 0.180072, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:40.071414+00:00", "epoch": 0, "step": 3033, "train_loss": 3.78281831741333, "perplexity": 43.93970342607556, "lr": 0.0026291804804649314, "grad_norm": 0.167781, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:40.375724+00:00", "epoch": 0, "step": 3034, "train_loss": 3.5947861671447754, "perplexity": 36.40791394726096, "lr": 0.0026291804804649314, "grad_norm": 0.183739, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:40.681084+00:00", "epoch": 0, "step": 3035, "train_loss": 3.7644355297088623, "perplexity": 43.13934810550782, "lr": 0.0026291804804649314, "grad_norm": 0.181066, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:40.988010+00:00", "epoch": 0, "step": 3036, "train_loss": 3.598120927810669, "perplexity": 36.52952829128264, "lr": 0.0026291804804649314, "grad_norm": 0.16976, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:41.292881+00:00", "epoch": 0, "step": 3037, "train_loss": 3.799923896789551, "perplexity": 44.69778271909461, "lr": 0.0026291804804649314, "grad_norm": 0.156103, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:41.597505+00:00", "epoch": 0, "step": 3038, "train_loss": 3.6111040115356445, "perplexity": 37.00688629972904, "lr": 0.0026291804804649314, "grad_norm": 0.163852, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:41.903712+00:00", "epoch": 0, "step": 3039, "train_loss": 3.698211431503296, "perplexity": 40.37502624216272, "lr": 0.0026291804804649314, "grad_norm": 0.149472, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:42.208699+00:00", "epoch": 0, "step": 3040, "train_loss": 3.758302927017212, "perplexity": 42.875601177360956, "lr": 0.0026291804804649314, "grad_norm": 0.136324, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:42.513817+00:00", "epoch": 0, "step": 3041, "train_loss": 3.6404929161071777, "perplexity": 38.11061743354947, "lr": 0.0026291804804649314, "grad_norm": 0.149366, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:42.818733+00:00", "epoch": 0, "step": 3042, "train_loss": 3.616698980331421, "perplexity": 37.21451898135364, "lr": 0.0026291804804649314, "grad_norm": 0.137589, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:43.123463+00:00", "epoch": 0, "step": 3043, "train_loss": 3.6911184787750244, "perplexity": 40.08966132594575, "lr": 0.0026291804804649314, "grad_norm": 0.148805, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:43.428499+00:00", "epoch": 0, "step": 3044, "train_loss": 3.6304290294647217, "perplexity": 37.728999998033544, "lr": 0.0026291804804649314, "grad_norm": 0.134004, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:43.733946+00:00", "epoch": 0, "step": 3045, "train_loss": 3.7660746574401855, "perplexity": 43.21011699107539, "lr": 0.0026291804804649314, "grad_norm": 0.147778, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:44.039442+00:00", "epoch": 0, "step": 3046, "train_loss": 3.7014169692993164, "perplexity": 40.50465757286043, "lr": 0.0026291804804649314, "grad_norm": 0.185628, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:44.345108+00:00", "epoch": 0, "step": 3047, "train_loss": 3.7413487434387207, "perplexity": 42.15480786064924, "lr": 0.0026291804804649314, "grad_norm": 0.157481, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:44.650931+00:00", "epoch": 0, "step": 3048, "train_loss": 3.628225088119507, "perplexity": 37.6459390593641, "lr": 0.0026291804804649314, "grad_norm": 0.153122, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:44.955221+00:00", "epoch": 0, "step": 3049, "train_loss": 3.765739917755127, "perplexity": 43.195655270713345, "lr": 0.0026291804804649314, "grad_norm": 0.153468, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:45.260337+00:00", "epoch": 0, "step": 3050, "train_loss": 3.6752841472625732, "perplexity": 39.459867677745244, "lr": 0.0026291804804649314, "grad_norm": 0.146237, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:45.565678+00:00", "epoch": 0, "step": 3051, "train_loss": 3.687413454055786, "perplexity": 39.941402959800676, "lr": 0.0026291804804649314, "grad_norm": 0.149116, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:45.870744+00:00", "epoch": 0, "step": 3052, "train_loss": 3.6272430419921875, "perplexity": 37.608987157907166, "lr": 0.0026291804804649314, "grad_norm": 0.152152, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:46.176211+00:00", "epoch": 0, "step": 3053, "train_loss": 3.6160855293273926, "perplexity": 37.19169669821206, "lr": 0.0026291804804649314, "grad_norm": 0.151168, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:46.479780+00:00", "epoch": 0, "step": 3054, "train_loss": 3.764580488204956, "perplexity": 43.145601973796374, "lr": 0.0026291804804649314, "grad_norm": 0.149463, "tokens_per_sec": 107942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:46.784007+00:00", "epoch": 0, "step": 3055, "train_loss": 3.722942590713501, "perplexity": 41.38599715776953, "lr": 0.0026291804804649314, "grad_norm": 0.161407, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:47.089314+00:00", "epoch": 0, "step": 3056, "train_loss": 3.798285961151123, "perplexity": 44.62463055344447, "lr": 0.0026291804804649314, "grad_norm": 0.146087, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:47.395487+00:00", "epoch": 0, "step": 3057, "train_loss": 3.6364288330078125, "perplexity": 37.956047023577106, "lr": 0.0026291804804649314, "grad_norm": 0.149529, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:47.700575+00:00", "epoch": 0, "step": 3058, "train_loss": 3.721999406814575, "perplexity": 41.34698095422921, "lr": 0.0026291804804649314, "grad_norm": 0.142366, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:48.004942+00:00", "epoch": 0, "step": 3059, "train_loss": 3.710808038711548, "perplexity": 40.8868313250389, "lr": 0.0026291804804649314, "grad_norm": 0.140911, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:48.309861+00:00", "epoch": 0, "step": 3060, "train_loss": 3.655162811279297, "perplexity": 38.6738171362591, "lr": 0.0026291804804649314, "grad_norm": 0.136824, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:48.615676+00:00", "epoch": 0, "step": 3061, "train_loss": 3.6748876571655273, "perplexity": 39.44422533220275, "lr": 0.0026291804804649314, "grad_norm": 0.149752, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:48.922338+00:00", "epoch": 0, "step": 3062, "train_loss": 3.6829686164855957, "perplexity": 39.76426388017543, "lr": 0.0026291804804649314, "grad_norm": 0.153943, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:49.228132+00:00", "epoch": 0, "step": 3063, "train_loss": 3.6927616596221924, "perplexity": 40.155590041172864, "lr": 0.0026291804804649314, "grad_norm": 0.163266, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:49.533486+00:00", "epoch": 0, "step": 3064, "train_loss": 3.6789615154266357, "perplexity": 39.60524327487126, "lr": 0.0026291804804649314, "grad_norm": 0.160562, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:49.839078+00:00", "epoch": 0, "step": 3065, "train_loss": 3.701296091079712, "perplexity": 40.4997617378732, "lr": 0.0026291804804649314, "grad_norm": 0.148624, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:50.145752+00:00", "epoch": 0, "step": 3066, "train_loss": 3.663184404373169, "perplexity": 38.98529034621964, "lr": 0.0026291804804649314, "grad_norm": 0.152603, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:50.452120+00:00", "epoch": 0, "step": 3067, "train_loss": 3.776442527770996, "perplexity": 43.66044431476481, "lr": 0.0026291804804649314, "grad_norm": 0.177035, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:50.758962+00:00", "epoch": 0, "step": 3068, "train_loss": 3.61873459815979, "perplexity": 37.29035067565514, "lr": 0.0026291804804649314, "grad_norm": 0.177208, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:51.063644+00:00", "epoch": 0, "step": 3069, "train_loss": 3.698298931121826, "perplexity": 40.37855919612087, "lr": 0.0026291804804649314, "grad_norm": 0.167564, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:51.368277+00:00", "epoch": 0, "step": 3070, "train_loss": 3.618021249771118, "perplexity": 37.26375914972636, "lr": 0.0026291804804649314, "grad_norm": 0.173263, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:51.672791+00:00", "epoch": 0, "step": 3071, "train_loss": 3.6527960300445557, "perplexity": 38.58239290484009, "lr": 0.0026291804804649314, "grad_norm": 0.170964, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:51.978468+00:00", "epoch": 0, "step": 3072, "train_loss": 3.7646872997283936, "perplexity": 43.150210667399236, "lr": 0.0026291804804649314, "grad_norm": 0.168446, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:52.285279+00:00", "epoch": 0, "step": 3073, "train_loss": 3.7002956867218018, "perplexity": 40.459265859240425, "lr": 0.0026291804804649314, "grad_norm": 0.160211, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:52.590879+00:00", "epoch": 0, "step": 3074, "train_loss": 3.8187079429626465, "perplexity": 45.54532313225797, "lr": 0.0026291804804649314, "grad_norm": 0.156816, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:52.896802+00:00", "epoch": 0, "step": 3075, "train_loss": 3.723438262939453, "perplexity": 41.40651613202668, "lr": 0.0026291804804649314, "grad_norm": 0.157917, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:53.202085+00:00", "epoch": 0, "step": 3076, "train_loss": 3.6659586429595947, "perplexity": 39.093595005019914, "lr": 0.0026291804804649314, "grad_norm": 0.153367, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:53.507599+00:00", "epoch": 0, "step": 3077, "train_loss": 3.686817169189453, "perplexity": 39.91759360496009, "lr": 0.0026291804804649314, "grad_norm": 0.16164, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:53.812508+00:00", "epoch": 0, "step": 3078, "train_loss": 3.7285315990448, "perplexity": 41.61795143407853, "lr": 0.0026291804804649314, "grad_norm": 0.183672, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:54.118287+00:00", "epoch": 0, "step": 3079, "train_loss": 3.745666980743408, "perplexity": 42.33723592488067, "lr": 0.0026291804804649314, "grad_norm": 0.165792, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:54.423008+00:00", "epoch": 0, "step": 3080, "train_loss": 3.666508913040161, "perplexity": 39.11511296049354, "lr": 0.0026291804804649314, "grad_norm": 0.185051, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:54.727216+00:00", "epoch": 0, "step": 3081, "train_loss": 3.638277530670166, "perplexity": 38.02628117983885, "lr": 0.0026291804804649314, "grad_norm": 0.164833, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:55.032511+00:00", "epoch": 0, "step": 3082, "train_loss": 3.702547550201416, "perplexity": 40.550477261710355, "lr": 0.0026291804804649314, "grad_norm": 0.168203, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:55.339055+00:00", "epoch": 0, "step": 3083, "train_loss": 3.55782413482666, "perplexity": 35.08676994684779, "lr": 0.0026291804804649314, "grad_norm": 0.161051, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:55.645748+00:00", "epoch": 0, "step": 3084, "train_loss": 3.704819440841675, "perplexity": 40.64270824113797, "lr": 0.0026291804804649314, "grad_norm": 0.152071, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:55.951101+00:00", "epoch": 0, "step": 3085, "train_loss": 3.6613833904266357, "perplexity": 38.915140483998364, "lr": 0.0026291804804649314, "grad_norm": 0.168674, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:56.256137+00:00", "epoch": 0, "step": 3086, "train_loss": 3.72017240524292, "perplexity": 41.271508919798784, "lr": 0.0026291804804649314, "grad_norm": 0.194122, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:56.561029+00:00", "epoch": 0, "step": 3087, "train_loss": 3.6297452449798584, "perplexity": 37.703210311504094, "lr": 0.0026291804804649314, "grad_norm": 0.16623, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:56.867294+00:00", "epoch": 0, "step": 3088, "train_loss": 3.6679701805114746, "perplexity": 39.172312384350015, "lr": 0.0026291804804649314, "grad_norm": 0.144952, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:57.173176+00:00", "epoch": 0, "step": 3089, "train_loss": 3.6736106872558594, "perplexity": 39.393888389564076, "lr": 0.0026291804804649314, "grad_norm": 0.172872, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:57.478282+00:00", "epoch": 0, "step": 3090, "train_loss": 3.615821361541748, "perplexity": 37.181873147640864, "lr": 0.0026291804804649314, "grad_norm": 0.161782, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:57.783622+00:00", "epoch": 0, "step": 3091, "train_loss": 3.7132632732391357, "perplexity": 40.987341422471545, "lr": 0.0026291804804649314, "grad_norm": 0.137974, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:58.088512+00:00", "epoch": 0, "step": 3092, "train_loss": 3.7229342460632324, "perplexity": 41.385651807538146, "lr": 0.0026291804804649314, "grad_norm": 0.171015, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:58.394018+00:00", "epoch": 0, "step": 3093, "train_loss": 3.6998109817504883, "perplexity": 40.439659803902856, "lr": 0.0026291804804649314, "grad_norm": 0.150863, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:58.699240+00:00", "epoch": 0, "step": 3094, "train_loss": 3.6054320335388184, "perplexity": 36.79757821148454, "lr": 0.0026291804804649314, "grad_norm": 0.154035, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:59.004672+00:00", "epoch": 0, "step": 3095, "train_loss": 3.7075836658477783, "perplexity": 40.755209248989004, "lr": 0.0026291804804649314, "grad_norm": 0.168313, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:59.309726+00:00", "epoch": 0, "step": 3096, "train_loss": 3.7564704418182373, "perplexity": 42.79710421703137, "lr": 0.0026291804804649314, "grad_norm": 0.169268, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:59.613791+00:00", "epoch": 0, "step": 3097, "train_loss": 3.7489657402038574, "perplexity": 42.47712688888201, "lr": 0.0026291804804649314, "grad_norm": 0.150964, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:13:59.919605+00:00", "epoch": 0, "step": 3098, "train_loss": 3.793693780899048, "perplexity": 44.420176011612206, "lr": 0.0026291804804649314, "grad_norm": 0.168535, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:00.224929+00:00", "epoch": 0, "step": 3099, "train_loss": 3.6589393615722656, "perplexity": 38.8201468886129, "lr": 0.0026291804804649314, "grad_norm": 0.175616, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:00.531091+00:00", "epoch": 0, "step": 3100, "train_loss": 3.5984885692596436, "perplexity": 36.542960528966375, "lr": 0.0026291804804649314, "grad_norm": 0.159184, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:00.837430+00:00", "epoch": 0, "step": 3101, "train_loss": 3.6250181198120117, "perplexity": 37.52540310669068, "lr": 0.0026291804804649314, "grad_norm": 0.167864, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:01.143159+00:00", "epoch": 0, "step": 3102, "train_loss": 3.661081075668335, "perplexity": 38.90337764083883, "lr": 0.0026291804804649314, "grad_norm": 0.155029, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:01.448586+00:00", "epoch": 0, "step": 3103, "train_loss": 3.657268524169922, "perplexity": 38.755338892116214, "lr": 0.0026291804804649314, "grad_norm": 0.157433, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:01.753993+00:00", "epoch": 0, "step": 3104, "train_loss": 3.6416430473327637, "perplexity": 38.15447486074984, "lr": 0.0026291804804649314, "grad_norm": 0.14211, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:02.060153+00:00", "epoch": 0, "step": 3105, "train_loss": 3.6869959831237793, "perplexity": 39.92473206513048, "lr": 0.0026291804804649314, "grad_norm": 0.155715, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:02.365535+00:00", "epoch": 0, "step": 3106, "train_loss": 3.6622986793518066, "perplexity": 38.95077538673297, "lr": 0.0026291804804649314, "grad_norm": 0.170324, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:02.670707+00:00", "epoch": 0, "step": 3107, "train_loss": 3.763679265975952, "perplexity": 43.1067357144129, "lr": 0.0026291804804649314, "grad_norm": 0.160129, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:02.977687+00:00", "epoch": 0, "step": 3108, "train_loss": 3.588491439819336, "perplexity": 36.17945585154417, "lr": 0.0026291804804649314, "grad_norm": 0.156455, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:03.284308+00:00", "epoch": 0, "step": 3109, "train_loss": 3.6711959838867188, "perplexity": 39.29887859095185, "lr": 0.0026291804804649314, "grad_norm": 0.156725, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:03.590179+00:00", "epoch": 0, "step": 3110, "train_loss": 3.7876529693603516, "perplexity": 44.152650947587105, "lr": 0.0026291804804649314, "grad_norm": 0.148395, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:03.895038+00:00", "epoch": 0, "step": 3111, "train_loss": 3.662123680114746, "perplexity": 38.94395962715114, "lr": 0.0026291804804649314, "grad_norm": 0.150573, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:04.201582+00:00", "epoch": 0, "step": 3112, "train_loss": 3.633274793624878, "perplexity": 37.83652075081693, "lr": 0.0026291804804649314, "grad_norm": 0.1545, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:04.507565+00:00", "epoch": 0, "step": 3113, "train_loss": 3.6827592849731445, "perplexity": 39.755940837843795, "lr": 0.0026291804804649314, "grad_norm": 0.148449, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:04.812448+00:00", "epoch": 0, "step": 3114, "train_loss": 3.6882448196411133, "perplexity": 39.97462267460159, "lr": 0.0026291804804649314, "grad_norm": 0.165041, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:05.118427+00:00", "epoch": 0, "step": 3115, "train_loss": 3.66841721534729, "perplexity": 39.18982768726886, "lr": 0.0026291804804649314, "grad_norm": 0.156755, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:05.423653+00:00", "epoch": 0, "step": 3116, "train_loss": 3.6347384452819824, "perplexity": 37.891940785031636, "lr": 0.0026291804804649314, "grad_norm": 0.173925, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:05.729838+00:00", "epoch": 0, "step": 3117, "train_loss": 3.7380428314208984, "perplexity": 42.015677877189155, "lr": 0.0026291804804649314, "grad_norm": 0.206344, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:06.035058+00:00", "epoch": 0, "step": 3118, "train_loss": 3.641819953918457, "perplexity": 38.16122523570167, "lr": 0.0026291804804649314, "grad_norm": 0.189813, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:06.340993+00:00", "epoch": 0, "step": 3119, "train_loss": 3.655423164367676, "perplexity": 38.683887294831486, "lr": 0.0026291804804649314, "grad_norm": 0.139247, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:06.646945+00:00", "epoch": 0, "step": 3120, "train_loss": 3.6751017570495605, "perplexity": 39.45267124037387, "lr": 0.0026291804804649314, "grad_norm": 0.154832, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:06.952134+00:00", "epoch": 0, "step": 3121, "train_loss": 3.7447116374969482, "perplexity": 42.2968086465045, "lr": 0.0026291804804649314, "grad_norm": 0.169203, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:07.257642+00:00", "epoch": 0, "step": 3122, "train_loss": 3.7144253253936768, "perplexity": 41.03499853554302, "lr": 0.0026291804804649314, "grad_norm": 0.138496, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:07.563723+00:00", "epoch": 0, "step": 3123, "train_loss": 3.6268575191497803, "perplexity": 37.59449082879054, "lr": 0.0026291804804649314, "grad_norm": 0.165865, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:07.869234+00:00", "epoch": 0, "step": 3124, "train_loss": 3.7759387493133545, "perplexity": 43.638454662889494, "lr": 0.0026291804804649314, "grad_norm": 0.167604, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:08.174329+00:00", "epoch": 0, "step": 3125, "train_loss": 3.7621214389801025, "perplexity": 43.039635156914095, "lr": 0.0026291804804649314, "grad_norm": 0.172705, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:08.479451+00:00", "epoch": 0, "step": 3126, "train_loss": 3.7526187896728516, "perplexity": 42.63258170389686, "lr": 0.0026291804804649314, "grad_norm": 0.164331, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:08.784363+00:00", "epoch": 0, "step": 3127, "train_loss": 3.6922123432159424, "perplexity": 40.133537974096235, "lr": 0.0026291804804649314, "grad_norm": 0.177374, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:09.089379+00:00", "epoch": 0, "step": 3128, "train_loss": 3.744401693344116, "perplexity": 42.28370102940064, "lr": 0.0026291804804649314, "grad_norm": 0.158407, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:09.394540+00:00", "epoch": 0, "step": 3129, "train_loss": 3.670154094696045, "perplexity": 39.25795483685286, "lr": 0.0026291804804649314, "grad_norm": 0.145255, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:09.699567+00:00", "epoch": 0, "step": 3130, "train_loss": 3.5961642265319824, "perplexity": 36.45812080091216, "lr": 0.0026291804804649314, "grad_norm": 0.15331, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:10.004299+00:00", "epoch": 0, "step": 3131, "train_loss": 3.7613511085510254, "perplexity": 43.0064931830707, "lr": 0.0026291804804649314, "grad_norm": 0.142775, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:10.308379+00:00", "epoch": 0, "step": 3132, "train_loss": 3.815329074859619, "perplexity": 45.391691189883915, "lr": 0.0026291804804649314, "grad_norm": 0.167987, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:10.613556+00:00", "epoch": 0, "step": 3133, "train_loss": 3.5935635566711426, "perplexity": 36.363428450104614, "lr": 0.0026291804804649314, "grad_norm": 0.142647, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:10.918944+00:00", "epoch": 0, "step": 3134, "train_loss": 3.663451910018921, "perplexity": 38.99572052649239, "lr": 0.0026291804804649314, "grad_norm": 0.151047, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:11.225246+00:00", "epoch": 0, "step": 3135, "train_loss": 3.6936278343200684, "perplexity": 40.1903868651335, "lr": 0.0026291804804649314, "grad_norm": 0.158415, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:11.530231+00:00", "epoch": 0, "step": 3136, "train_loss": 3.679656505584717, "perplexity": 39.632778096262335, "lr": 0.0026291804804649314, "grad_norm": 0.166856, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:11.834387+00:00", "epoch": 0, "step": 3137, "train_loss": 3.645169734954834, "perplexity": 38.28927132777254, "lr": 0.0026291804804649314, "grad_norm": 0.188662, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:12.139353+00:00", "epoch": 0, "step": 3138, "train_loss": 3.6930670738220215, "perplexity": 40.16785600157785, "lr": 0.0026291804804649314, "grad_norm": 0.189724, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:12.444525+00:00", "epoch": 0, "step": 3139, "train_loss": 3.620102882385254, "perplexity": 37.34140939770261, "lr": 0.0026291804804649314, "grad_norm": 0.174022, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:12.751500+00:00", "epoch": 0, "step": 3140, "train_loss": 3.7435622215270996, "perplexity": 42.24821994883455, "lr": 0.0026291804804649314, "grad_norm": 0.161512, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:13.056066+00:00", "epoch": 0, "step": 3141, "train_loss": 3.670431613922119, "perplexity": 39.268851185999694, "lr": 0.0026291804804649314, "grad_norm": 0.167136, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:13.360716+00:00", "epoch": 0, "step": 3142, "train_loss": 3.7437713146209717, "perplexity": 42.25705468346305, "lr": 0.0026291804804649314, "grad_norm": 0.158411, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:13.665508+00:00", "epoch": 0, "step": 3143, "train_loss": 3.6041879653930664, "perplexity": 36.751827980693875, "lr": 0.0026291804804649314, "grad_norm": 0.146869, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:13.970346+00:00", "epoch": 0, "step": 3144, "train_loss": 3.6736018657684326, "perplexity": 39.393540878405744, "lr": 0.0026291804804649314, "grad_norm": 0.149121, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:14.275652+00:00", "epoch": 0, "step": 3145, "train_loss": 3.6272926330566406, "perplexity": 37.61085227385949, "lr": 0.0026291804804649314, "grad_norm": 0.145768, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:14.580489+00:00", "epoch": 0, "step": 3146, "train_loss": 3.7578654289245605, "perplexity": 42.85684728631949, "lr": 0.0026291804804649314, "grad_norm": 0.179475, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:14.886593+00:00", "epoch": 0, "step": 3147, "train_loss": 3.6833579540252686, "perplexity": 39.77974861504012, "lr": 0.0026291804804649314, "grad_norm": 0.166761, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:15.191209+00:00", "epoch": 0, "step": 3148, "train_loss": 3.7088727951049805, "perplexity": 40.807781860780594, "lr": 0.0026291804804649314, "grad_norm": 0.180143, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:15.496515+00:00", "epoch": 0, "step": 3149, "train_loss": 3.6619157791137695, "perplexity": 38.935863980538336, "lr": 0.0026291804804649314, "grad_norm": 0.162806, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:15.803165+00:00", "epoch": 0, "step": 3150, "train_loss": 3.6221425533294678, "perplexity": 37.41765131324487, "lr": 0.0026291804804649314, "grad_norm": 0.160361, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:16.107964+00:00", "epoch": 0, "step": 3151, "train_loss": 3.7294201850891113, "perplexity": 41.65494900024178, "lr": 0.0026291804804649314, "grad_norm": 0.156337, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:16.412802+00:00", "epoch": 0, "step": 3152, "train_loss": 3.672041893005371, "perplexity": 39.33213593506693, "lr": 0.0026291804804649314, "grad_norm": 0.153042, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:16.716914+00:00", "epoch": 0, "step": 3153, "train_loss": 3.615002155303955, "perplexity": 37.151425998175625, "lr": 0.0026291804804649314, "grad_norm": 0.152816, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:17.022338+00:00", "epoch": 0, "step": 3154, "train_loss": 3.641084909439087, "perplexity": 38.133185344312324, "lr": 0.0026291804804649314, "grad_norm": 0.158037, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:17.329261+00:00", "epoch": 0, "step": 3155, "train_loss": 3.668105125427246, "perplexity": 39.177598845427795, "lr": 0.0026291804804649314, "grad_norm": 0.151292, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:17.634324+00:00", "epoch": 0, "step": 3156, "train_loss": 3.6567013263702393, "perplexity": 38.73336318204701, "lr": 0.0026291804804649314, "grad_norm": 0.147466, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:17.940299+00:00", "epoch": 0, "step": 3157, "train_loss": 3.625612735748291, "perplexity": 37.54772294460125, "lr": 0.0026291804804649314, "grad_norm": 0.138644, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:18.244898+00:00", "epoch": 0, "step": 3158, "train_loss": 3.602811574935913, "perplexity": 36.70127791167499, "lr": 0.0026291804804649314, "grad_norm": 0.15301, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:18.549204+00:00", "epoch": 0, "step": 3159, "train_loss": 3.704477310180664, "perplexity": 40.62880550291428, "lr": 0.0026291804804649314, "grad_norm": 0.152791, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:18.854221+00:00", "epoch": 0, "step": 3160, "train_loss": 3.6741294860839844, "perplexity": 39.414331195089126, "lr": 0.0026291804804649314, "grad_norm": 0.146867, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:19.160137+00:00", "epoch": 0, "step": 3161, "train_loss": 3.6171629428863525, "perplexity": 37.23178913070229, "lr": 0.0026291804804649314, "grad_norm": 0.154264, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:19.466240+00:00", "epoch": 0, "step": 3162, "train_loss": 3.6520347595214844, "perplexity": 38.55303244345692, "lr": 0.0026291804804649314, "grad_norm": 0.14041, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:19.770967+00:00", "epoch": 0, "step": 3163, "train_loss": 3.6322922706604004, "perplexity": 37.79936375707227, "lr": 0.0026291804804649314, "grad_norm": 0.149474, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:20.075451+00:00", "epoch": 0, "step": 3164, "train_loss": 3.7113964557647705, "perplexity": 40.91089691344973, "lr": 0.0026291804804649314, "grad_norm": 0.147562, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:20.380986+00:00", "epoch": 0, "step": 3165, "train_loss": 3.68143367767334, "perplexity": 39.70327498728871, "lr": 0.0026291804804649314, "grad_norm": 0.17399, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:20.686642+00:00", "epoch": 0, "step": 3166, "train_loss": 3.6058058738708496, "perplexity": 36.811337202013625, "lr": 0.0026291804804649314, "grad_norm": 0.151509, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:20.992101+00:00", "epoch": 0, "step": 3167, "train_loss": 3.5926105976104736, "perplexity": 36.32879209762026, "lr": 0.0026291804804649314, "grad_norm": 0.14078, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:21.297793+00:00", "epoch": 0, "step": 3168, "train_loss": 3.565802574157715, "perplexity": 35.367827320892815, "lr": 0.0026291804804649314, "grad_norm": 0.157225, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:21.602597+00:00", "epoch": 0, "step": 3169, "train_loss": 3.6274025440216064, "perplexity": 37.61498634611187, "lr": 0.0026291804804649314, "grad_norm": 0.159388, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:21.907196+00:00", "epoch": 0, "step": 3170, "train_loss": 3.634096384048462, "perplexity": 37.867619647456024, "lr": 0.0026291804804649314, "grad_norm": 0.170849, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:22.212923+00:00", "epoch": 0, "step": 3171, "train_loss": 3.711264133453369, "perplexity": 40.90548384715127, "lr": 0.0026291804804649314, "grad_norm": 0.160722, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:22.519174+00:00", "epoch": 0, "step": 3172, "train_loss": 3.7088892459869385, "perplexity": 40.80845319030492, "lr": 0.0026291804804649314, "grad_norm": 0.151707, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:22.824180+00:00", "epoch": 0, "step": 3173, "train_loss": 3.5801167488098145, "perplexity": 35.877729284753286, "lr": 0.0026291804804649314, "grad_norm": 0.165115, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:23.128831+00:00", "epoch": 0, "step": 3174, "train_loss": 3.638307809829712, "perplexity": 38.02743260110558, "lr": 0.0026291804804649314, "grad_norm": 0.153323, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:23.434416+00:00", "epoch": 0, "step": 3175, "train_loss": 3.7231876850128174, "perplexity": 41.39614187289956, "lr": 0.0026291804804649314, "grad_norm": 0.170071, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:23.739403+00:00", "epoch": 0, "step": 3176, "train_loss": 3.690635919570923, "perplexity": 40.070320357840004, "lr": 0.0026291804804649314, "grad_norm": 0.161013, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:24.045225+00:00", "epoch": 0, "step": 3177, "train_loss": 3.7417232990264893, "perplexity": 42.17060013684278, "lr": 0.0026291804804649314, "grad_norm": 0.151336, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:24.350706+00:00", "epoch": 0, "step": 3178, "train_loss": 3.7906110286712646, "perplexity": 44.28345046881904, "lr": 0.0026291804804649314, "grad_norm": 0.157407, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:24.655893+00:00", "epoch": 0, "step": 3179, "train_loss": 3.6205410957336426, "perplexity": 37.35777648762505, "lr": 0.0026291804804649314, "grad_norm": 0.153562, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:24.961203+00:00", "epoch": 0, "step": 3180, "train_loss": 3.683697462081909, "perplexity": 39.793256453066235, "lr": 0.0026291804804649314, "grad_norm": 0.146047, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:25.266121+00:00", "epoch": 0, "step": 3181, "train_loss": 3.625854730606079, "perplexity": 37.556810399989885, "lr": 0.0026291804804649314, "grad_norm": 0.141649, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:25.571997+00:00", "epoch": 0, "step": 3182, "train_loss": 3.7709481716156006, "perplexity": 43.42121608815356, "lr": 0.0026291804804649314, "grad_norm": 0.143314, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:25.877435+00:00", "epoch": 0, "step": 3183, "train_loss": 3.776163339614868, "perplexity": 43.64825653724158, "lr": 0.0026291804804649314, "grad_norm": 0.152543, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:26.182914+00:00", "epoch": 0, "step": 3184, "train_loss": 3.628054141998291, "perplexity": 37.6395041821267, "lr": 0.0026291804804649314, "grad_norm": 0.175549, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:26.488563+00:00", "epoch": 0, "step": 3185, "train_loss": 3.7144980430603027, "perplexity": 41.037982613382795, "lr": 0.0026291804804649314, "grad_norm": 0.209144, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:26.794593+00:00", "epoch": 0, "step": 3186, "train_loss": 3.700174570083618, "perplexity": 40.45436586571754, "lr": 0.0026291804804649314, "grad_norm": 0.193376, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:27.099861+00:00", "epoch": 0, "step": 3187, "train_loss": 3.792073965072632, "perplexity": 44.34828175094673, "lr": 0.0026291804804649314, "grad_norm": 0.164931, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:27.405528+00:00", "epoch": 0, "step": 3188, "train_loss": 3.633958101272583, "perplexity": 37.86238356993326, "lr": 0.0026291804804649314, "grad_norm": 0.171825, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:27.711106+00:00", "epoch": 0, "step": 3189, "train_loss": 3.6370291709899902, "perplexity": 37.97884032141533, "lr": 0.0026291804804649314, "grad_norm": 0.189674, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:28.016101+00:00", "epoch": 0, "step": 3190, "train_loss": 3.671417236328125, "perplexity": 39.307574525747555, "lr": 0.0026291804804649314, "grad_norm": 0.170625, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:28.320023+00:00", "epoch": 0, "step": 3191, "train_loss": 3.6232759952545166, "perplexity": 37.460086092115915, "lr": 0.0026291804804649314, "grad_norm": 0.162233, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:28.625357+00:00", "epoch": 0, "step": 3192, "train_loss": 3.7015323638916016, "perplexity": 40.509331860995296, "lr": 0.0026291804804649314, "grad_norm": 0.187542, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:28.931719+00:00", "epoch": 0, "step": 3193, "train_loss": 3.652251720428467, "perplexity": 38.56139785179325, "lr": 0.0026291804804649314, "grad_norm": 0.19166, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:29.237229+00:00", "epoch": 0, "step": 3194, "train_loss": 3.593892812728882, "perplexity": 36.37540330048981, "lr": 0.0026291804804649314, "grad_norm": 0.163636, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:29.542719+00:00", "epoch": 0, "step": 3195, "train_loss": 3.720604181289673, "perplexity": 41.289332816452614, "lr": 0.0026291804804649314, "grad_norm": 0.159423, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:29.848023+00:00", "epoch": 0, "step": 3196, "train_loss": 3.6345431804656982, "perplexity": 37.88454254450714, "lr": 0.0026291804804649314, "grad_norm": 0.148198, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:30.152750+00:00", "epoch": 0, "step": 3197, "train_loss": 3.7065539360046387, "perplexity": 40.713263993608905, "lr": 0.0026291804804649314, "grad_norm": 0.167586, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:30.458547+00:00", "epoch": 0, "step": 3198, "train_loss": 3.7449655532836914, "perplexity": 42.30754983756999, "lr": 0.0026291804804649314, "grad_norm": 0.168106, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:30.766008+00:00", "epoch": 0, "step": 3199, "train_loss": 3.6768133640289307, "perplexity": 39.52025653105149, "lr": 0.0026291804804649314, "grad_norm": 0.158039, "tokens_per_sec": 106585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:31.073294+00:00", "epoch": 0, "step": 3200, "train_loss": 3.6480929851531982, "perplexity": 38.4013642057174, "lr": 0.0026291804804649314, "grad_norm": 0.173924, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:31.378894+00:00", "epoch": 0, "step": 3201, "train_loss": 3.6930932998657227, "perplexity": 40.168909459338686, "lr": 0.0026291804804649314, "grad_norm": 0.16211, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:31.683470+00:00", "epoch": 0, "step": 3202, "train_loss": 3.683802366256714, "perplexity": 39.797431150765014, "lr": 0.0026291804804649314, "grad_norm": 0.16552, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:31.988542+00:00", "epoch": 0, "step": 3203, "train_loss": 3.7077343463897705, "perplexity": 40.76135072869677, "lr": 0.0026291804804649314, "grad_norm": 0.167898, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:32.294135+00:00", "epoch": 0, "step": 3204, "train_loss": 3.763171911239624, "perplexity": 43.08487085496956, "lr": 0.0026291804804649314, "grad_norm": 0.16369, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:32.599536+00:00", "epoch": 0, "step": 3205, "train_loss": 3.7234039306640625, "perplexity": 41.40509457651463, "lr": 0.0026291804804649314, "grad_norm": 0.14988, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:32.905371+00:00", "epoch": 0, "step": 3206, "train_loss": 3.684260606765747, "perplexity": 39.815672124931204, "lr": 0.0026291804804649314, "grad_norm": 0.142363, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:33.210404+00:00", "epoch": 0, "step": 3207, "train_loss": 3.7637391090393066, "perplexity": 43.10931543071755, "lr": 0.0026291804804649314, "grad_norm": 0.156797, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:33.515502+00:00", "epoch": 0, "step": 3208, "train_loss": 3.6941702365875244, "perplexity": 40.21219213517845, "lr": 0.0026291804804649314, "grad_norm": 0.145363, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:33.819730+00:00", "epoch": 0, "step": 3209, "train_loss": 3.590013265609741, "perplexity": 36.23455659677523, "lr": 0.0026291804804649314, "grad_norm": 0.154102, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:34.125549+00:00", "epoch": 0, "step": 3210, "train_loss": 3.644343376159668, "perplexity": 38.25764372132551, "lr": 0.0026291804804649314, "grad_norm": 0.153591, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:34.431619+00:00", "epoch": 0, "step": 3211, "train_loss": 3.668577194213867, "perplexity": 39.196097733007655, "lr": 0.0026291804804649314, "grad_norm": 0.138108, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:34.736133+00:00", "epoch": 0, "step": 3212, "train_loss": 3.697204113006592, "perplexity": 40.33437620862668, "lr": 0.0026291804804649314, "grad_norm": 0.133813, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:35.041658+00:00", "epoch": 0, "step": 3213, "train_loss": 3.797922372817993, "perplexity": 44.608408507657, "lr": 0.0026291804804649314, "grad_norm": 0.130183, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:35.347170+00:00", "epoch": 0, "step": 3214, "train_loss": 3.5924839973449707, "perplexity": 36.32419315401518, "lr": 0.0026291804804649314, "grad_norm": 0.151532, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:35.652322+00:00", "epoch": 0, "step": 3215, "train_loss": 3.7188432216644287, "perplexity": 41.21668794952091, "lr": 0.0026291804804649314, "grad_norm": 0.147188, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:35.957823+00:00", "epoch": 0, "step": 3216, "train_loss": 3.7031588554382324, "perplexity": 40.57527355909688, "lr": 0.0026291804804649314, "grad_norm": 0.135256, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:36.264153+00:00", "epoch": 0, "step": 3217, "train_loss": 3.736361026763916, "perplexity": 41.945075101149264, "lr": 0.0026291804804649314, "grad_norm": 0.131355, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:36.568871+00:00", "epoch": 0, "step": 3218, "train_loss": 3.586530923843384, "perplexity": 36.1085949350223, "lr": 0.0026291804804649314, "grad_norm": 0.155287, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:36.873351+00:00", "epoch": 0, "step": 3219, "train_loss": 3.6183722019195557, "perplexity": 37.276839241167664, "lr": 0.0026291804804649314, "grad_norm": 0.1669, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:37.178610+00:00", "epoch": 0, "step": 3220, "train_loss": 3.764115333557129, "perplexity": 43.12553726346338, "lr": 0.0026291804804649314, "grad_norm": 0.166574, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:37.484167+00:00", "epoch": 0, "step": 3221, "train_loss": 3.677229642868042, "perplexity": 39.53671140223133, "lr": 0.0026291804804649314, "grad_norm": 0.166456, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:37.790226+00:00", "epoch": 0, "step": 3222, "train_loss": 3.8554885387420654, "perplexity": 47.251695587332044, "lr": 0.0026291804804649314, "grad_norm": 0.150025, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:38.095193+00:00", "epoch": 0, "step": 3223, "train_loss": 3.697695732116699, "perplexity": 40.35421023375852, "lr": 0.0026291804804649314, "grad_norm": 0.160484, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:38.400550+00:00", "epoch": 0, "step": 3224, "train_loss": 3.7264487743377686, "perplexity": 41.5313587465745, "lr": 0.0026291804804649314, "grad_norm": 0.190855, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:38.706162+00:00", "epoch": 0, "step": 3225, "train_loss": 3.6775121688842773, "perplexity": 39.54788312987639, "lr": 0.0026291804804649314, "grad_norm": 0.185753, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:39.011764+00:00", "epoch": 0, "step": 3226, "train_loss": 3.645711898803711, "perplexity": 38.31003601490901, "lr": 0.0026291804804649314, "grad_norm": 0.184506, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:39.318305+00:00", "epoch": 0, "step": 3227, "train_loss": 3.7045891284942627, "perplexity": 40.6333488014364, "lr": 0.0026291804804649314, "grad_norm": 0.161985, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:39.623892+00:00", "epoch": 0, "step": 3228, "train_loss": 3.6235296726226807, "perplexity": 37.4695900735885, "lr": 0.0026291804804649314, "grad_norm": 0.157549, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:39.928894+00:00", "epoch": 0, "step": 3229, "train_loss": 3.707041025161743, "perplexity": 40.733099813564756, "lr": 0.0026291804804649314, "grad_norm": 0.171325, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:40.235471+00:00", "epoch": 0, "step": 3230, "train_loss": 3.6281368732452393, "perplexity": 37.64261827405678, "lr": 0.0026291804804649314, "grad_norm": 0.136114, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:40.540643+00:00", "epoch": 0, "step": 3231, "train_loss": 3.5572078227996826, "perplexity": 35.06515221086137, "lr": 0.0026291804804649314, "grad_norm": 0.15875, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:40.846531+00:00", "epoch": 0, "step": 3232, "train_loss": 3.7052924633026123, "perplexity": 40.661937702634404, "lr": 0.0026291804804649314, "grad_norm": 0.172171, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:41.153124+00:00", "epoch": 0, "step": 3233, "train_loss": 3.7765724658966064, "perplexity": 43.666117839658, "lr": 0.0026291804804649314, "grad_norm": 0.169456, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:41.458891+00:00", "epoch": 0, "step": 3234, "train_loss": 3.659818172454834, "perplexity": 38.85427745111979, "lr": 0.0026291804804649314, "grad_norm": 0.175341, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:41.764741+00:00", "epoch": 0, "step": 3235, "train_loss": 3.6268129348754883, "perplexity": 37.59281474306337, "lr": 0.0026291804804649314, "grad_norm": 0.163936, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:42.069950+00:00", "epoch": 0, "step": 3236, "train_loss": 3.724539041519165, "perplexity": 41.45212063366962, "lr": 0.0026291804804649314, "grad_norm": 0.147292, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:42.375594+00:00", "epoch": 0, "step": 3237, "train_loss": 3.692232847213745, "perplexity": 40.134360880507074, "lr": 0.0026291804804649314, "grad_norm": 0.135863, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:42.681090+00:00", "epoch": 0, "step": 3238, "train_loss": 3.722104787826538, "perplexity": 41.35133837051421, "lr": 0.0026291804804649314, "grad_norm": 0.144101, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:42.987102+00:00", "epoch": 0, "step": 3239, "train_loss": 3.6654114723205566, "perplexity": 39.07220998881876, "lr": 0.0026291804804649314, "grad_norm": 0.162758, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:43.292553+00:00", "epoch": 0, "step": 3240, "train_loss": 3.6541686058044434, "perplexity": 38.63538652265752, "lr": 0.0026291804804649314, "grad_norm": 0.179277, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:43.597621+00:00", "epoch": 0, "step": 3241, "train_loss": 3.668562173843384, "perplexity": 39.195508997519724, "lr": 0.0026291804804649314, "grad_norm": 0.181881, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:43.903087+00:00", "epoch": 0, "step": 3242, "train_loss": 3.6892497539520264, "perplexity": 40.014814736301545, "lr": 0.0026291804804649314, "grad_norm": 0.165512, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:44.208734+00:00", "epoch": 0, "step": 3243, "train_loss": 3.7430410385131836, "perplexity": 42.226206631210935, "lr": 0.0026291804804649314, "grad_norm": 0.148381, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:44.514152+00:00", "epoch": 0, "step": 3244, "train_loss": 3.7229514122009277, "perplexity": 41.38636224543341, "lr": 0.0026291804804649314, "grad_norm": 0.184222, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:44.820249+00:00", "epoch": 0, "step": 3245, "train_loss": 3.738687753677368, "perplexity": 42.042783462530956, "lr": 0.0026291804804649314, "grad_norm": 0.172434, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:45.125582+00:00", "epoch": 0, "step": 3246, "train_loss": 3.668940544128418, "perplexity": 39.21034221947942, "lr": 0.0026291804804649314, "grad_norm": 0.170917, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:45.429973+00:00", "epoch": 0, "step": 3247, "train_loss": 3.574575424194336, "perplexity": 35.67946896004845, "lr": 0.0026291804804649314, "grad_norm": 0.184503, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:45.734370+00:00", "epoch": 0, "step": 3248, "train_loss": 3.61263370513916, "perplexity": 37.06353881643692, "lr": 0.0026291804804649314, "grad_norm": 0.155001, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:46.039627+00:00", "epoch": 0, "step": 3249, "train_loss": 3.7074975967407227, "perplexity": 40.75170163547182, "lr": 0.0026291804804649314, "grad_norm": 0.190261, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:46.345843+00:00", "epoch": 0, "step": 3250, "train_loss": 3.8321192264556885, "perplexity": 46.16025869317467, "lr": 0.0026291804804649314, "grad_norm": 0.155748, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:46.650508+00:00", "epoch": 0, "step": 3251, "train_loss": 3.70350980758667, "perplexity": 40.58951603759391, "lr": 0.0026291804804649314, "grad_norm": 0.164712, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:46.954848+00:00", "epoch": 0, "step": 3252, "train_loss": 3.6655514240264893, "perplexity": 39.077678593922634, "lr": 0.0026291804804649314, "grad_norm": 0.155003, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:47.259605+00:00", "epoch": 0, "step": 3253, "train_loss": 3.600050210952759, "perplexity": 36.6000721220341, "lr": 0.0026291804804649314, "grad_norm": 0.166709, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:47.565178+00:00", "epoch": 0, "step": 3254, "train_loss": 3.6642415523529053, "perplexity": 39.02652535906362, "lr": 0.0026291804804649314, "grad_norm": 0.1577, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:47.871026+00:00", "epoch": 0, "step": 3255, "train_loss": 3.681347608566284, "perplexity": 39.69985790891792, "lr": 0.0026291804804649314, "grad_norm": 0.167868, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:48.176672+00:00", "epoch": 0, "step": 3256, "train_loss": 3.624297857284546, "perplexity": 37.49838469634719, "lr": 0.0026291804804649314, "grad_norm": 0.15699, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:48.482581+00:00", "epoch": 0, "step": 3257, "train_loss": 3.739806652069092, "perplexity": 42.08985139253587, "lr": 0.0026291804804649314, "grad_norm": 0.152577, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:48.788428+00:00", "epoch": 0, "step": 3258, "train_loss": 3.749678611755371, "perplexity": 42.507418419933686, "lr": 0.0026291804804649314, "grad_norm": 0.181203, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:49.092968+00:00", "epoch": 0, "step": 3259, "train_loss": 3.6362454891204834, "perplexity": 37.94908865227507, "lr": 0.0026291804804649314, "grad_norm": 0.174631, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:49.398365+00:00", "epoch": 0, "step": 3260, "train_loss": 3.750377655029297, "perplexity": 42.53714333316197, "lr": 0.0026291804804649314, "grad_norm": 0.157985, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:49.703789+00:00", "epoch": 0, "step": 3261, "train_loss": 3.623499631881714, "perplexity": 37.46846447624594, "lr": 0.0026291804804649314, "grad_norm": 0.15611, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:50.009562+00:00", "epoch": 0, "step": 3262, "train_loss": 3.622713327407837, "perplexity": 37.43901443486597, "lr": 0.0026291804804649314, "grad_norm": 0.154566, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:50.314899+00:00", "epoch": 0, "step": 3263, "train_loss": 3.6827077865600586, "perplexity": 39.7538935226971, "lr": 0.0026291804804649314, "grad_norm": 0.144341, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:50.620097+00:00", "epoch": 0, "step": 3264, "train_loss": 3.6466023921966553, "perplexity": 38.34416604289059, "lr": 0.0026291804804649314, "grad_norm": 0.142647, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:50.925442+00:00", "epoch": 0, "step": 3265, "train_loss": 3.655895948410034, "perplexity": 38.70218074352543, "lr": 0.0026291804804649314, "grad_norm": 0.139021, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:51.231028+00:00", "epoch": 0, "step": 3266, "train_loss": 3.7065281867980957, "perplexity": 40.712215672862065, "lr": 0.0026291804804649314, "grad_norm": 0.158909, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:51.535985+00:00", "epoch": 0, "step": 3267, "train_loss": 3.6206018924713135, "perplexity": 37.360047787605254, "lr": 0.0026291804804649314, "grad_norm": 0.13856, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:51.841284+00:00", "epoch": 0, "step": 3268, "train_loss": 3.7008707523345947, "perplexity": 40.48253928298639, "lr": 0.0026291804804649314, "grad_norm": 0.153475, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:52.147552+00:00", "epoch": 0, "step": 3269, "train_loss": 3.689561367034912, "perplexity": 40.02728581905798, "lr": 0.0026291804804649314, "grad_norm": 0.155021, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:52.453193+00:00", "epoch": 0, "step": 3270, "train_loss": 3.6294150352478027, "perplexity": 37.69076239985339, "lr": 0.0026291804804649314, "grad_norm": 0.145468, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:52.760979+00:00", "epoch": 0, "step": 3271, "train_loss": 3.7836625576019287, "perplexity": 43.97681475281547, "lr": 0.0026291804804649314, "grad_norm": 0.137842, "tokens_per_sec": 106407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:53.066638+00:00", "epoch": 0, "step": 3272, "train_loss": 3.6324126720428467, "perplexity": 37.803915126714315, "lr": 0.0026291804804649314, "grad_norm": 0.150229, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:53.372591+00:00", "epoch": 0, "step": 3273, "train_loss": 3.643941640853882, "perplexity": 38.242277361937674, "lr": 0.0026291804804649314, "grad_norm": 0.154693, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:53.677989+00:00", "epoch": 0, "step": 3274, "train_loss": 3.597289800643921, "perplexity": 36.499180221230965, "lr": 0.0026291804804649314, "grad_norm": 0.156742, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:53.983497+00:00", "epoch": 0, "step": 3275, "train_loss": 3.793701648712158, "perplexity": 44.420525502630255, "lr": 0.0026291804804649314, "grad_norm": 0.151007, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:54.289423+00:00", "epoch": 0, "step": 3276, "train_loss": 3.6764743328094482, "perplexity": 39.50686020130083, "lr": 0.0026291804804649314, "grad_norm": 0.149698, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:54.596750+00:00", "epoch": 0, "step": 3277, "train_loss": 3.6948087215423584, "perplexity": 40.23787521311458, "lr": 0.0026291804804649314, "grad_norm": 0.138087, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:54.901803+00:00", "epoch": 0, "step": 3278, "train_loss": 3.6022322177886963, "perplexity": 36.68002092229382, "lr": 0.0026291804804649314, "grad_norm": 0.160684, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:55.206607+00:00", "epoch": 0, "step": 3279, "train_loss": 3.598222255706787, "perplexity": 36.5332299390675, "lr": 0.0026291804804649314, "grad_norm": 0.156837, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:55.512523+00:00", "epoch": 0, "step": 3280, "train_loss": 3.6667637825012207, "perplexity": 39.125083478789406, "lr": 0.0026291804804649314, "grad_norm": 0.157495, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:55.819243+00:00", "epoch": 0, "step": 3281, "train_loss": 3.675318479537964, "perplexity": 39.46122244804527, "lr": 0.0026291804804649314, "grad_norm": 0.168182, "tokens_per_sec": 106778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:56.126097+00:00", "epoch": 0, "step": 3282, "train_loss": 3.7620060443878174, "perplexity": 43.034668902308034, "lr": 0.0026291804804649314, "grad_norm": 0.155001, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:56.432276+00:00", "epoch": 0, "step": 3283, "train_loss": 3.6363747119903564, "perplexity": 37.95399285928081, "lr": 0.0026291804804649314, "grad_norm": 0.148826, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:56.738349+00:00", "epoch": 0, "step": 3284, "train_loss": 3.6954689025878906, "perplexity": 40.26444826619062, "lr": 0.0026291804804649314, "grad_norm": 0.154614, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:57.042995+00:00", "epoch": 0, "step": 3285, "train_loss": 3.6671481132507324, "perplexity": 39.140123341403225, "lr": 0.0026291804804649314, "grad_norm": 0.160489, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:57.348464+00:00", "epoch": 0, "step": 3286, "train_loss": 3.6495988368988037, "perplexity": 38.45923452816774, "lr": 0.0026291804804649314, "grad_norm": 0.154405, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:57.654399+00:00", "epoch": 0, "step": 3287, "train_loss": 3.700274705886841, "perplexity": 40.458416998965724, "lr": 0.0026291804804649314, "grad_norm": 0.165785, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:57.960039+00:00", "epoch": 0, "step": 3288, "train_loss": 3.7847602367401123, "perplexity": 44.02511368845081, "lr": 0.0026291804804649314, "grad_norm": 0.156637, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:58.265163+00:00", "epoch": 0, "step": 3289, "train_loss": 3.5981342792510986, "perplexity": 36.530016016359454, "lr": 0.0026291804804649314, "grad_norm": 0.16785, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:58.570401+00:00", "epoch": 0, "step": 3290, "train_loss": 3.7776308059692383, "perplexity": 43.71235590546706, "lr": 0.0026291804804649314, "grad_norm": 0.182576, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:58.874798+00:00", "epoch": 0, "step": 3291, "train_loss": 3.6569571495056152, "perplexity": 38.743273340029695, "lr": 0.0026291804804649314, "grad_norm": 0.186449, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:59.181533+00:00", "epoch": 0, "step": 3292, "train_loss": 3.7265398502349854, "perplexity": 41.535141424587756, "lr": 0.0026291804804649314, "grad_norm": 0.155699, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:59.487915+00:00", "epoch": 0, "step": 3293, "train_loss": 3.6606736183166504, "perplexity": 38.88752940257357, "lr": 0.0026291804804649314, "grad_norm": 0.138146, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:14:59.794417+00:00", "epoch": 0, "step": 3294, "train_loss": 3.679713249206543, "perplexity": 39.635027067441335, "lr": 0.0026291804804649314, "grad_norm": 0.147255, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:00.098720+00:00", "epoch": 0, "step": 3295, "train_loss": 3.7018110752105713, "perplexity": 40.52062384383748, "lr": 0.0026291804804649314, "grad_norm": 0.143912, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:00.404617+00:00", "epoch": 0, "step": 3296, "train_loss": 3.7670764923095703, "perplexity": 43.253428084646615, "lr": 0.0026291804804649314, "grad_norm": 0.155925, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:00.711363+00:00", "epoch": 0, "step": 3297, "train_loss": 3.7038307189941406, "perplexity": 40.60254376657571, "lr": 0.0026291804804649314, "grad_norm": 0.183917, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:01.018148+00:00", "epoch": 0, "step": 3298, "train_loss": 3.791074752807617, "perplexity": 44.30399053573672, "lr": 0.0026291804804649314, "grad_norm": 0.163491, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:01.324012+00:00", "epoch": 0, "step": 3299, "train_loss": 3.623009204864502, "perplexity": 37.45009343416925, "lr": 0.0026291804804649314, "grad_norm": 0.165273, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:01.630409+00:00", "epoch": 0, "step": 3300, "train_loss": 3.7296645641326904, "perplexity": 41.66512984078016, "lr": 0.0026291804804649314, "grad_norm": 0.163966, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:01.935084+00:00", "epoch": 0, "step": 3301, "train_loss": 3.712843179702759, "perplexity": 40.97012652145418, "lr": 0.0026291804804649314, "grad_norm": 0.172838, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:02.240260+00:00", "epoch": 0, "step": 3302, "train_loss": 3.7639753818511963, "perplexity": 43.119502193273135, "lr": 0.0026291804804649314, "grad_norm": 0.160218, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:02.545486+00:00", "epoch": 0, "step": 3303, "train_loss": 3.6527421474456787, "perplexity": 38.58031404124727, "lr": 0.0026291804804649314, "grad_norm": 0.182383, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:02.851905+00:00", "epoch": 0, "step": 3304, "train_loss": 3.6000983715057373, "perplexity": 36.60183484419304, "lr": 0.0026291804804649314, "grad_norm": 0.190648, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:03.158494+00:00", "epoch": 0, "step": 3305, "train_loss": 3.5958518981933594, "perplexity": 36.44673567465443, "lr": 0.0026291804804649314, "grad_norm": 0.142243, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:03.464166+00:00", "epoch": 0, "step": 3306, "train_loss": 3.7842206954956055, "perplexity": 44.00136673062944, "lr": 0.0026291804804649314, "grad_norm": 0.148797, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:03.768499+00:00", "epoch": 0, "step": 3307, "train_loss": 3.6950395107269287, "perplexity": 40.247162751214304, "lr": 0.0026291804804649314, "grad_norm": 0.139514, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:04.074698+00:00", "epoch": 0, "step": 3308, "train_loss": 3.584721803665161, "perplexity": 36.04332920189623, "lr": 0.0026291804804649314, "grad_norm": 0.15847, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:04.379864+00:00", "epoch": 0, "step": 3309, "train_loss": 3.7262699604034424, "perplexity": 41.52393302485015, "lr": 0.0026291804804649314, "grad_norm": 0.17142, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:04.685172+00:00", "epoch": 0, "step": 3310, "train_loss": 3.7238810062408447, "perplexity": 41.42485264856351, "lr": 0.0026291804804649314, "grad_norm": 0.183818, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:04.989970+00:00", "epoch": 0, "step": 3311, "train_loss": 3.6008312702178955, "perplexity": 36.62867011437927, "lr": 0.0026291804804649314, "grad_norm": 0.159507, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:05.295645+00:00", "epoch": 0, "step": 3312, "train_loss": 3.7349042892456055, "perplexity": 41.884016620428504, "lr": 0.0026291804804649314, "grad_norm": 0.156581, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:05.600328+00:00", "epoch": 0, "step": 3313, "train_loss": 3.5809974670410156, "perplexity": 35.909341373656616, "lr": 0.0026291804804649314, "grad_norm": 0.150116, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:05.906517+00:00", "epoch": 0, "step": 3314, "train_loss": 3.5363712310791016, "perplexity": 34.342073361029165, "lr": 0.0026291804804649314, "grad_norm": 0.18681, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:06.213113+00:00", "epoch": 0, "step": 3315, "train_loss": 3.7310900688171387, "perplexity": 41.72456603176583, "lr": 0.0026291804804649314, "grad_norm": 0.174021, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:06.518979+00:00", "epoch": 0, "step": 3316, "train_loss": 3.6362881660461426, "perplexity": 37.950708237269524, "lr": 0.0026291804804649314, "grad_norm": 0.146812, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:06.824726+00:00", "epoch": 0, "step": 3317, "train_loss": 3.6923234462738037, "perplexity": 40.1379971805991, "lr": 0.0026291804804649314, "grad_norm": 0.179353, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:07.129820+00:00", "epoch": 0, "step": 3318, "train_loss": 3.6296274662017822, "perplexity": 37.698769934960254, "lr": 0.0026291804804649314, "grad_norm": 0.163642, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:07.435863+00:00", "epoch": 0, "step": 3319, "train_loss": 3.654006004333496, "perplexity": 38.62910486269573, "lr": 0.0026291804804649314, "grad_norm": 0.190916, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:07.740659+00:00", "epoch": 0, "step": 3320, "train_loss": 3.7123873233795166, "perplexity": 40.951454286466884, "lr": 0.0026291804804649314, "grad_norm": 0.173129, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:08.045932+00:00", "epoch": 0, "step": 3321, "train_loss": 3.6525325775146484, "perplexity": 38.572229614650475, "lr": 0.0026291804804649314, "grad_norm": 0.154151, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:08.350936+00:00", "epoch": 0, "step": 3322, "train_loss": 3.6394906044006348, "perplexity": 38.07243785266903, "lr": 0.0026291804804649314, "grad_norm": 0.183142, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:08.657311+00:00", "epoch": 0, "step": 3323, "train_loss": 3.755744218826294, "perplexity": 42.76603525882174, "lr": 0.0026291804804649314, "grad_norm": 0.166978, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:08.962235+00:00", "epoch": 0, "step": 3324, "train_loss": 3.7439253330230713, "perplexity": 42.26356354873248, "lr": 0.0026291804804649314, "grad_norm": 0.165009, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:09.266851+00:00", "epoch": 0, "step": 3325, "train_loss": 3.625009298324585, "perplexity": 37.525072078279074, "lr": 0.0026291804804649314, "grad_norm": 0.167622, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:09.571959+00:00", "epoch": 0, "step": 3326, "train_loss": 3.680298089981079, "perplexity": 39.65821402704902, "lr": 0.0026291804804649314, "grad_norm": 0.162464, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:09.877603+00:00", "epoch": 0, "step": 3327, "train_loss": 3.650226354598999, "perplexity": 38.483375952363424, "lr": 0.0026291804804649314, "grad_norm": 0.144565, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:10.183833+00:00", "epoch": 0, "step": 3328, "train_loss": 3.659146785736084, "perplexity": 38.82819996029254, "lr": 0.0026291804804649314, "grad_norm": 0.161446, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:10.490208+00:00", "epoch": 0, "step": 3329, "train_loss": 3.719524621963501, "perplexity": 41.24478258377553, "lr": 0.0026291804804649314, "grad_norm": 0.16591, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:10.794707+00:00", "epoch": 0, "step": 3330, "train_loss": 3.6503825187683105, "perplexity": 38.48938614607757, "lr": 0.0026291804804649314, "grad_norm": 0.17463, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:11.100363+00:00", "epoch": 0, "step": 3331, "train_loss": 3.7497637271881104, "perplexity": 42.511036611226906, "lr": 0.0026291804804649314, "grad_norm": 0.178155, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:11.405593+00:00", "epoch": 0, "step": 3332, "train_loss": 3.51706862449646, "perplexity": 33.685538630280405, "lr": 0.0026291804804649314, "grad_norm": 0.185471, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:11.711455+00:00", "epoch": 0, "step": 3333, "train_loss": 3.748619794845581, "perplexity": 42.462434665501654, "lr": 0.0026291804804649314, "grad_norm": 0.177977, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:12.016947+00:00", "epoch": 0, "step": 3334, "train_loss": 3.7214672565460205, "perplexity": 41.32498400057199, "lr": 0.0026291804804649314, "grad_norm": 0.186389, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:12.322383+00:00", "epoch": 0, "step": 3335, "train_loss": 3.602235794067383, "perplexity": 36.68015210050543, "lr": 0.0026291804804649314, "grad_norm": 0.165592, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:12.627935+00:00", "epoch": 0, "step": 3336, "train_loss": 3.7826735973358154, "perplexity": 43.93334492890197, "lr": 0.0026291804804649314, "grad_norm": 0.148868, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:12.933570+00:00", "epoch": 0, "step": 3337, "train_loss": 3.7687323093414307, "perplexity": 43.3251071749077, "lr": 0.0026291804804649314, "grad_norm": 0.150497, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:13.238751+00:00", "epoch": 0, "step": 3338, "train_loss": 3.6122987270355225, "perplexity": 37.05112542171388, "lr": 0.0026291804804649314, "grad_norm": 0.162192, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:13.543862+00:00", "epoch": 0, "step": 3339, "train_loss": 3.7020413875579834, "perplexity": 40.529957318599536, "lr": 0.0026291804804649314, "grad_norm": 0.172691, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:13.849750+00:00", "epoch": 0, "step": 3340, "train_loss": 3.676210403442383, "perplexity": 39.496434556570584, "lr": 0.0026291804804649314, "grad_norm": 0.155022, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:14.154674+00:00", "epoch": 0, "step": 3341, "train_loss": 3.801693916320801, "perplexity": 44.77696872722753, "lr": 0.0026291804804649314, "grad_norm": 0.159689, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:14.460413+00:00", "epoch": 0, "step": 3342, "train_loss": 3.630199670791626, "perplexity": 37.720347516955485, "lr": 0.0026291804804649314, "grad_norm": 0.148264, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:14.767202+00:00", "epoch": 0, "step": 3343, "train_loss": 3.6470279693603516, "perplexity": 38.360487917181565, "lr": 0.0026291804804649314, "grad_norm": 0.179318, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:15.072463+00:00", "epoch": 0, "step": 3344, "train_loss": 3.6129143238067627, "perplexity": 37.07394099677095, "lr": 0.0026291804804649314, "grad_norm": 0.143231, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:15.378093+00:00", "epoch": 0, "step": 3345, "train_loss": 3.6536056995391846, "perplexity": 38.61364454144604, "lr": 0.0026291804804649314, "grad_norm": 0.171144, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:15.683203+00:00", "epoch": 0, "step": 3346, "train_loss": 3.517650842666626, "perplexity": 33.705156673379186, "lr": 0.0026291804804649314, "grad_norm": 0.137768, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:15.988612+00:00", "epoch": 0, "step": 3347, "train_loss": 3.718885660171509, "perplexity": 41.21843716134098, "lr": 0.0026291804804649314, "grad_norm": 0.15482, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:16.293989+00:00", "epoch": 0, "step": 3348, "train_loss": 3.638655662536621, "perplexity": 38.04066284742777, "lr": 0.0026291804804649314, "grad_norm": 0.164586, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:16.598675+00:00", "epoch": 0, "step": 3349, "train_loss": 3.6356632709503174, "perplexity": 37.92700043402544, "lr": 0.0026291804804649314, "grad_norm": 0.145732, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:16.904017+00:00", "epoch": 0, "step": 3350, "train_loss": 3.6259186267852783, "perplexity": 37.55921021334599, "lr": 0.0026291804804649314, "grad_norm": 0.156855, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:17.208970+00:00", "epoch": 0, "step": 3351, "train_loss": 3.756801128387451, "perplexity": 42.811258984864004, "lr": 0.0026291804804649314, "grad_norm": 0.174308, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:17.513978+00:00", "epoch": 0, "step": 3352, "train_loss": 3.6570446491241455, "perplexity": 38.746663509984685, "lr": 0.0026291804804649314, "grad_norm": 0.160347, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:17.818952+00:00", "epoch": 0, "step": 3353, "train_loss": 3.653851270675659, "perplexity": 38.623128102416274, "lr": 0.0026291804804649314, "grad_norm": 0.154476, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:18.124201+00:00", "epoch": 0, "step": 3354, "train_loss": 3.672497034072876, "perplexity": 39.35004167991447, "lr": 0.0026291804804649314, "grad_norm": 0.17138, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:18.428882+00:00", "epoch": 0, "step": 3355, "train_loss": 3.7266719341278076, "perplexity": 41.54062791008624, "lr": 0.0026291804804649314, "grad_norm": 0.163216, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:18.734749+00:00", "epoch": 0, "step": 3356, "train_loss": 3.61326265335083, "perplexity": 37.08685719515129, "lr": 0.0026291804804649314, "grad_norm": 0.163649, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:19.039703+00:00", "epoch": 0, "step": 3357, "train_loss": 3.7330856323242188, "perplexity": 41.80791318769971, "lr": 0.0026291804804649314, "grad_norm": 0.151721, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:19.345117+00:00", "epoch": 0, "step": 3358, "train_loss": 3.663896322250366, "perplexity": 39.01305455310982, "lr": 0.0026291804804649314, "grad_norm": 0.142263, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:19.651520+00:00", "epoch": 0, "step": 3359, "train_loss": 3.5954782962799072, "perplexity": 36.4331216477389, "lr": 0.0026291804804649314, "grad_norm": 0.163792, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:19.956428+00:00", "epoch": 0, "step": 3360, "train_loss": 3.656324863433838, "perplexity": 38.718784250792616, "lr": 0.0026291804804649314, "grad_norm": 0.152966, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:20.262479+00:00", "epoch": 0, "step": 3361, "train_loss": 3.598527193069458, "perplexity": 36.54437198458162, "lr": 0.0026291804804649314, "grad_norm": 0.144869, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:20.567481+00:00", "epoch": 0, "step": 3362, "train_loss": 3.6924376487731934, "perplexity": 40.142581301951694, "lr": 0.0026291804804649314, "grad_norm": 0.161125, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:20.873444+00:00", "epoch": 0, "step": 3363, "train_loss": 3.6284282207489014, "perplexity": 37.65358695469351, "lr": 0.0026291804804649314, "grad_norm": 0.179387, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:21.178506+00:00", "epoch": 0, "step": 3364, "train_loss": 3.6795873641967773, "perplexity": 39.630037925707555, "lr": 0.0026291804804649314, "grad_norm": 0.164613, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:21.484463+00:00", "epoch": 0, "step": 3365, "train_loss": 3.7001326084136963, "perplexity": 40.45266836858534, "lr": 0.0026291804804649314, "grad_norm": 0.144616, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:21.789880+00:00", "epoch": 0, "step": 3366, "train_loss": 3.628600835800171, "perplexity": 37.66008709153076, "lr": 0.0026291804804649314, "grad_norm": 0.148279, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:22.094669+00:00", "epoch": 0, "step": 3367, "train_loss": 3.598414421081543, "perplexity": 36.54025103547399, "lr": 0.0026291804804649314, "grad_norm": 0.16783, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:22.401757+00:00", "epoch": 0, "step": 3368, "train_loss": 3.615901231765747, "perplexity": 37.184842990777305, "lr": 0.0026291804804649314, "grad_norm": 0.144687, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:22.708551+00:00", "epoch": 0, "step": 3369, "train_loss": 3.606813669204712, "perplexity": 36.848454195904175, "lr": 0.0026291804804649314, "grad_norm": 0.149325, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:23.015023+00:00", "epoch": 0, "step": 3370, "train_loss": 3.7893972396850586, "perplexity": 44.22973231221738, "lr": 0.0026291804804649314, "grad_norm": 0.137945, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:23.320981+00:00", "epoch": 0, "step": 3371, "train_loss": 3.6647093296051025, "perplexity": 39.044785350330194, "lr": 0.0026291804804649314, "grad_norm": 0.149456, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:23.626565+00:00", "epoch": 0, "step": 3372, "train_loss": 3.6387314796447754, "perplexity": 38.04354708981321, "lr": 0.0026291804804649314, "grad_norm": 0.139082, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:23.931448+00:00", "epoch": 0, "step": 3373, "train_loss": 3.658439874649048, "perplexity": 38.8007615746431, "lr": 0.0026291804804649314, "grad_norm": 0.148779, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:24.236696+00:00", "epoch": 0, "step": 3374, "train_loss": 3.6847918033599854, "perplexity": 39.83682769274661, "lr": 0.0026291804804649314, "grad_norm": 0.141017, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:24.541501+00:00", "epoch": 0, "step": 3375, "train_loss": 3.6079704761505127, "perplexity": 36.891105408513184, "lr": 0.0026291804804649314, "grad_norm": 0.132788, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:24.846819+00:00", "epoch": 0, "step": 3376, "train_loss": 3.61893367767334, "perplexity": 37.297775159534545, "lr": 0.0026291804804649314, "grad_norm": 0.162074, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:25.151701+00:00", "epoch": 0, "step": 3377, "train_loss": 3.726497173309326, "perplexity": 41.53336887026879, "lr": 0.0026291804804649314, "grad_norm": 0.157099, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:25.456235+00:00", "epoch": 0, "step": 3378, "train_loss": 3.682013750076294, "perplexity": 39.72631244246575, "lr": 0.0026291804804649314, "grad_norm": 0.149142, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:25.762176+00:00", "epoch": 0, "step": 3379, "train_loss": 3.6468677520751953, "perplexity": 38.354342396272685, "lr": 0.0026291804804649314, "grad_norm": 0.163485, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:26.067946+00:00", "epoch": 0, "step": 3380, "train_loss": 3.626718759536743, "perplexity": 37.58927459370056, "lr": 0.0026291804804649314, "grad_norm": 0.156531, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:26.372557+00:00", "epoch": 0, "step": 3381, "train_loss": 3.709240436553955, "perplexity": 40.822787250966016, "lr": 0.0026291804804649314, "grad_norm": 0.145549, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:26.678461+00:00", "epoch": 0, "step": 3382, "train_loss": 3.5723230838775635, "perplexity": 35.59919708735148, "lr": 0.0026291804804649314, "grad_norm": 0.162048, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:26.984056+00:00", "epoch": 0, "step": 3383, "train_loss": 3.6266467571258545, "perplexity": 37.58656817274184, "lr": 0.0026291804804649314, "grad_norm": 0.154345, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:27.289398+00:00", "epoch": 0, "step": 3384, "train_loss": 3.7681849002838135, "perplexity": 43.301397108962426, "lr": 0.0026291804804649314, "grad_norm": 0.143309, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:27.595302+00:00", "epoch": 0, "step": 3385, "train_loss": 3.668787717819214, "perplexity": 39.204350305468104, "lr": 0.0026291804804649314, "grad_norm": 0.16984, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:27.901245+00:00", "epoch": 0, "step": 3386, "train_loss": 3.664747714996338, "perplexity": 39.046284128456975, "lr": 0.0026291804804649314, "grad_norm": 0.162877, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:28.207605+00:00", "epoch": 0, "step": 3387, "train_loss": 3.6404216289520264, "perplexity": 38.107900732885646, "lr": 0.0026291804804649314, "grad_norm": 0.146467, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:28.512556+00:00", "epoch": 0, "step": 3388, "train_loss": 3.568923234939575, "perplexity": 35.4783707070469, "lr": 0.0026291804804649314, "grad_norm": 0.168664, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:28.817523+00:00", "epoch": 0, "step": 3389, "train_loss": 3.6745922565460205, "perplexity": 39.43257520441585, "lr": 0.0026291804804649314, "grad_norm": 0.160129, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:29.124383+00:00", "epoch": 0, "step": 3390, "train_loss": 3.794705629348755, "perplexity": 44.465145245033064, "lr": 0.0026291804804649314, "grad_norm": 0.174143, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:29.430183+00:00", "epoch": 0, "step": 3391, "train_loss": 3.7141828536987305, "perplexity": 41.02504991607404, "lr": 0.0026291804804649314, "grad_norm": 0.166813, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:29.735757+00:00", "epoch": 0, "step": 3392, "train_loss": 3.6565937995910645, "perplexity": 38.729198532167146, "lr": 0.0026291804804649314, "grad_norm": 0.154689, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:30.041353+00:00", "epoch": 0, "step": 3393, "train_loss": 3.6886491775512695, "perplexity": 39.99078999795783, "lr": 0.0026291804804649314, "grad_norm": 0.1511, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:30.346911+00:00", "epoch": 0, "step": 3394, "train_loss": 3.6822288036346436, "perplexity": 39.734856646014364, "lr": 0.0026291804804649314, "grad_norm": 0.165094, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:30.652246+00:00", "epoch": 0, "step": 3395, "train_loss": 3.7002651691436768, "perplexity": 40.45803115927381, "lr": 0.0026291804804649314, "grad_norm": 0.148387, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:30.957970+00:00", "epoch": 0, "step": 3396, "train_loss": 3.7287230491638184, "perplexity": 41.62591995859704, "lr": 0.0026291804804649314, "grad_norm": 0.178463, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:31.263582+00:00", "epoch": 0, "step": 3397, "train_loss": 3.775268316268921, "perplexity": 43.60920780600062, "lr": 0.0026291804804649314, "grad_norm": 0.16785, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:31.569689+00:00", "epoch": 0, "step": 3398, "train_loss": 3.656442880630493, "perplexity": 38.72335400281715, "lr": 0.0026291804804649314, "grad_norm": 0.174913, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:31.875250+00:00", "epoch": 0, "step": 3399, "train_loss": 3.744189500808716, "perplexity": 42.274729695531796, "lr": 0.0026291804804649314, "grad_norm": 0.130295, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:32.181546+00:00", "epoch": 0, "step": 3400, "train_loss": 3.669292688369751, "perplexity": 39.22415234712846, "lr": 0.0026291804804649314, "grad_norm": 0.141955, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:32.486347+00:00", "epoch": 0, "step": 3401, "train_loss": 3.7094993591308594, "perplexity": 40.83335856075378, "lr": 0.0026291804804649314, "grad_norm": 0.146825, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:32.791595+00:00", "epoch": 0, "step": 3402, "train_loss": 3.639484167098999, "perplexity": 38.072192769691405, "lr": 0.0026291804804649314, "grad_norm": 0.147043, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:33.097050+00:00", "epoch": 0, "step": 3403, "train_loss": 3.6430087089538574, "perplexity": 38.206616558594135, "lr": 0.0026291804804649314, "grad_norm": 0.165499, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:33.402843+00:00", "epoch": 0, "step": 3404, "train_loss": 3.551344633102417, "perplexity": 34.8601601127152, "lr": 0.0026291804804649314, "grad_norm": 0.145818, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:33.707331+00:00", "epoch": 0, "step": 3405, "train_loss": 3.591498613357544, "perplexity": 36.28841750499565, "lr": 0.0026291804804649314, "grad_norm": 0.157549, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:34.012451+00:00", "epoch": 0, "step": 3406, "train_loss": 3.666429042816162, "perplexity": 39.11198895241891, "lr": 0.0026291804804649314, "grad_norm": 0.170585, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:34.318463+00:00", "epoch": 0, "step": 3407, "train_loss": 3.6735851764678955, "perplexity": 39.39288343324897, "lr": 0.0026291804804649314, "grad_norm": 0.147092, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:34.623544+00:00", "epoch": 0, "step": 3408, "train_loss": 3.6344711780548096, "perplexity": 37.881814864309476, "lr": 0.0026291804804649314, "grad_norm": 0.161907, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:34.929308+00:00", "epoch": 0, "step": 3409, "train_loss": 3.6363461017608643, "perplexity": 37.95290700236835, "lr": 0.0026291804804649314, "grad_norm": 0.147101, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:35.235002+00:00", "epoch": 0, "step": 3410, "train_loss": 3.6616125106811523, "perplexity": 38.924057752414996, "lr": 0.0026291804804649314, "grad_norm": 0.15134, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:35.539538+00:00", "epoch": 0, "step": 3411, "train_loss": 3.6200647354125977, "perplexity": 37.339984963148474, "lr": 0.0026291804804649314, "grad_norm": 0.149595, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:35.844177+00:00", "epoch": 0, "step": 3412, "train_loss": 3.7060859203338623, "perplexity": 40.69421400624547, "lr": 0.0026291804804649314, "grad_norm": 0.152444, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:36.150236+00:00", "epoch": 0, "step": 3413, "train_loss": 3.628446340560913, "perplexity": 37.65426923679208, "lr": 0.0026291804804649314, "grad_norm": 0.161812, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:36.456826+00:00", "epoch": 0, "step": 3414, "train_loss": 3.761310577392578, "perplexity": 43.004750115405756, "lr": 0.0026291804804649314, "grad_norm": 0.175767, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:36.761858+00:00", "epoch": 0, "step": 3415, "train_loss": 3.573908805847168, "perplexity": 35.65569229728278, "lr": 0.0026291804804649314, "grad_norm": 0.162507, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:37.066750+00:00", "epoch": 0, "step": 3416, "train_loss": 3.6437134742736816, "perplexity": 38.23355274766348, "lr": 0.0026291804804649314, "grad_norm": 0.169626, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:37.372533+00:00", "epoch": 0, "step": 3417, "train_loss": 3.6626856327056885, "perplexity": 38.965850436387576, "lr": 0.0026291804804649314, "grad_norm": 0.166715, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:37.679181+00:00", "epoch": 0, "step": 3418, "train_loss": 3.6069562435150146, "perplexity": 36.853708213381935, "lr": 0.0026291804804649314, "grad_norm": 0.177198, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:37.984928+00:00", "epoch": 0, "step": 3419, "train_loss": 3.7002649307250977, "perplexity": 40.45802151332865, "lr": 0.0026291804804649314, "grad_norm": 0.153221, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:38.290402+00:00", "epoch": 0, "step": 3420, "train_loss": 3.6723175048828125, "perplexity": 39.34297783290505, "lr": 0.0026291804804649314, "grad_norm": 0.152867, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:38.594854+00:00", "epoch": 0, "step": 3421, "train_loss": 3.6808011531829834, "perplexity": 39.67816963422425, "lr": 0.0026291804804649314, "grad_norm": 0.166113, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:38.899942+00:00", "epoch": 0, "step": 3422, "train_loss": 3.7043657302856445, "perplexity": 40.624272397967886, "lr": 0.0026291804804649314, "grad_norm": 0.144557, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:39.205805+00:00", "epoch": 0, "step": 3423, "train_loss": 3.6155166625976562, "perplexity": 37.170545595987576, "lr": 0.0026291804804649314, "grad_norm": 0.142923, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:39.511163+00:00", "epoch": 0, "step": 3424, "train_loss": 3.67962384223938, "perplexity": 39.63148357828648, "lr": 0.0026291804804649314, "grad_norm": 0.149567, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:39.817143+00:00", "epoch": 0, "step": 3425, "train_loss": 3.626976490020752, "perplexity": 37.598963744176224, "lr": 0.0026291804804649314, "grad_norm": 0.166176, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:40.122500+00:00", "epoch": 0, "step": 3426, "train_loss": 3.7397589683532715, "perplexity": 42.08784443987301, "lr": 0.0026291804804649314, "grad_norm": 0.171881, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:40.427361+00:00", "epoch": 0, "step": 3427, "train_loss": 3.7115366458892822, "perplexity": 40.916632619217175, "lr": 0.0026291804804649314, "grad_norm": 0.165708, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:40.732018+00:00", "epoch": 0, "step": 3428, "train_loss": 3.639608860015869, "perplexity": 38.076940398451185, "lr": 0.0026291804804649314, "grad_norm": 0.145769, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:41.038217+00:00", "epoch": 0, "step": 3429, "train_loss": 3.5474178791046143, "perplexity": 34.72354124951051, "lr": 0.0026291804804649314, "grad_norm": 0.161182, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:41.344776+00:00", "epoch": 0, "step": 3430, "train_loss": 3.677570104598999, "perplexity": 39.550174431124695, "lr": 0.0026291804804649314, "grad_norm": 0.165157, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:41.650269+00:00", "epoch": 0, "step": 3431, "train_loss": 3.574789524078369, "perplexity": 35.68710874802485, "lr": 0.0026291804804649314, "grad_norm": 0.159028, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:41.956764+00:00", "epoch": 0, "step": 3432, "train_loss": 3.653336763381958, "perplexity": 38.60326133253929, "lr": 0.0026291804804649314, "grad_norm": 0.154286, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:42.261877+00:00", "epoch": 0, "step": 3433, "train_loss": 3.7040297985076904, "perplexity": 40.6106277058843, "lr": 0.0026291804804649314, "grad_norm": 0.149013, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:42.567574+00:00", "epoch": 0, "step": 3434, "train_loss": 3.588628053665161, "perplexity": 36.18439880377808, "lr": 0.0026291804804649314, "grad_norm": 0.149183, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:42.873757+00:00", "epoch": 0, "step": 3435, "train_loss": 3.7203686237335205, "perplexity": 41.279607947547355, "lr": 0.0026291804804649314, "grad_norm": 0.147543, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:43.179931+00:00", "epoch": 0, "step": 3436, "train_loss": 3.6793723106384277, "perplexity": 39.62151626137406, "lr": 0.0026291804804649314, "grad_norm": 0.17156, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:43.484721+00:00", "epoch": 0, "step": 3437, "train_loss": 3.7362489700317383, "perplexity": 41.94037513643874, "lr": 0.0026291804804649314, "grad_norm": 0.154136, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:43.789229+00:00", "epoch": 0, "step": 3438, "train_loss": 3.6927545070648193, "perplexity": 40.155302827038405, "lr": 0.0026291804804649314, "grad_norm": 0.162546, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:44.093294+00:00", "epoch": 0, "step": 3439, "train_loss": 3.596240758895874, "perplexity": 36.460911133854125, "lr": 0.0026291804804649314, "grad_norm": 0.154934, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:44.399278+00:00", "epoch": 0, "step": 3440, "train_loss": 3.7129294872283936, "perplexity": 40.97366270429659, "lr": 0.0026291804804649314, "grad_norm": 0.157565, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:44.703974+00:00", "epoch": 0, "step": 3441, "train_loss": 3.688141107559204, "perplexity": 39.97047703824048, "lr": 0.0026291804804649314, "grad_norm": 0.139777, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:45.009643+00:00", "epoch": 0, "step": 3442, "train_loss": 3.647172689437866, "perplexity": 38.36603985169494, "lr": 0.0026291804804649314, "grad_norm": 0.15334, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:45.314943+00:00", "epoch": 0, "step": 3443, "train_loss": 3.7798633575439453, "perplexity": 43.810055013046174, "lr": 0.0026291804804649314, "grad_norm": 0.160075, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:45.620352+00:00", "epoch": 0, "step": 3444, "train_loss": 3.544647216796875, "perplexity": 34.627467198415246, "lr": 0.0026291804804649314, "grad_norm": 0.181623, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:45.926559+00:00", "epoch": 0, "step": 3445, "train_loss": 3.7822320461273193, "perplexity": 43.9139503895116, "lr": 0.0026291804804649314, "grad_norm": 0.194196, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:46.232129+00:00", "epoch": 0, "step": 3446, "train_loss": 3.6683528423309326, "perplexity": 39.187305001047434, "lr": 0.0026291804804649314, "grad_norm": 0.151662, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:46.538511+00:00", "epoch": 0, "step": 3447, "train_loss": 3.722856044769287, "perplexity": 41.38241552255852, "lr": 0.0026291804804649314, "grad_norm": 0.145196, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:46.843700+00:00", "epoch": 0, "step": 3448, "train_loss": 3.6514177322387695, "perplexity": 38.529251508105375, "lr": 0.0026291804804649314, "grad_norm": 0.156181, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:47.148016+00:00", "epoch": 0, "step": 3449, "train_loss": 3.672283172607422, "perplexity": 39.34162712214202, "lr": 0.0026291804804649314, "grad_norm": 0.161294, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:47.453774+00:00", "epoch": 0, "step": 3450, "train_loss": 3.6924962997436523, "perplexity": 40.1449357723471, "lr": 0.0026291804804649314, "grad_norm": 0.173474, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:47.758766+00:00", "epoch": 0, "step": 3451, "train_loss": 3.720905065536499, "perplexity": 41.301757995435615, "lr": 0.0026291804804649314, "grad_norm": 0.161457, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:48.064382+00:00", "epoch": 0, "step": 3452, "train_loss": 3.6974315643310547, "perplexity": 40.34355135932725, "lr": 0.0026291804804649314, "grad_norm": 0.157795, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:48.370231+00:00", "epoch": 0, "step": 3453, "train_loss": 3.680438756942749, "perplexity": 39.66379301990225, "lr": 0.0026291804804649314, "grad_norm": 0.156738, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:48.675982+00:00", "epoch": 0, "step": 3454, "train_loss": 3.6552987098693848, "perplexity": 38.6790732106197, "lr": 0.0026291804804649314, "grad_norm": 0.162486, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:48.981215+00:00", "epoch": 0, "step": 3455, "train_loss": 3.668039321899414, "perplexity": 39.17502090603146, "lr": 0.0026291804804649314, "grad_norm": 0.185177, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:49.286819+00:00", "epoch": 0, "step": 3456, "train_loss": 3.6277618408203125, "perplexity": 37.628503718518296, "lr": 0.0026291804804649314, "grad_norm": 0.22207, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:49.592954+00:00", "epoch": 0, "step": 3457, "train_loss": 3.6462326049804688, "perplexity": 38.329989481789674, "lr": 0.0026291804804649314, "grad_norm": 0.206972, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:49.898409+00:00", "epoch": 0, "step": 3458, "train_loss": 3.6385066509246826, "perplexity": 38.03499476925289, "lr": 0.0026291804804649314, "grad_norm": 0.195004, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:50.204204+00:00", "epoch": 0, "step": 3459, "train_loss": 3.5795536041259766, "perplexity": 35.8575306201575, "lr": 0.0026291804804649314, "grad_norm": 0.200029, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:50.509063+00:00", "epoch": 0, "step": 3460, "train_loss": 3.7323293685913086, "perplexity": 41.77630733189482, "lr": 0.0026291804804649314, "grad_norm": 0.177044, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:50.814397+00:00", "epoch": 0, "step": 3461, "train_loss": 3.6310203075408936, "perplexity": 37.75131492508171, "lr": 0.0026291804804649314, "grad_norm": 0.161349, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:51.120184+00:00", "epoch": 0, "step": 3462, "train_loss": 3.6478607654571533, "perplexity": 38.39244768792943, "lr": 0.0026291804804649314, "grad_norm": 0.162524, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:51.425854+00:00", "epoch": 0, "step": 3463, "train_loss": 3.6180362701416016, "perplexity": 37.26431886939798, "lr": 0.0026291804804649314, "grad_norm": 0.161375, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:51.731626+00:00", "epoch": 0, "step": 3464, "train_loss": 3.6360275745391846, "perplexity": 37.94081989348501, "lr": 0.0026291804804649314, "grad_norm": 0.149802, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:52.036785+00:00", "epoch": 0, "step": 3465, "train_loss": 3.731337308883667, "perplexity": 41.73488329161484, "lr": 0.0026291804804649314, "grad_norm": 0.150315, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:52.342565+00:00", "epoch": 0, "step": 3466, "train_loss": 3.6599340438842773, "perplexity": 38.85877981263053, "lr": 0.0026291804804649314, "grad_norm": 0.156144, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:52.647078+00:00", "epoch": 0, "step": 3467, "train_loss": 3.629730701446533, "perplexity": 37.70266197759583, "lr": 0.0026291804804649314, "grad_norm": 0.151339, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:52.952673+00:00", "epoch": 0, "step": 3468, "train_loss": 3.6183178424835205, "perplexity": 37.27481294828391, "lr": 0.0026291804804649314, "grad_norm": 0.154755, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:53.257930+00:00", "epoch": 0, "step": 3469, "train_loss": 3.7474324703216553, "perplexity": 42.41204789412173, "lr": 0.0026291804804649314, "grad_norm": 0.154643, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:53.562975+00:00", "epoch": 0, "step": 3470, "train_loss": 3.5392065048217773, "perplexity": 34.43958070458421, "lr": 0.0026291804804649314, "grad_norm": 0.154218, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:53.866689+00:00", "epoch": 0, "step": 3471, "train_loss": 3.6986382007598877, "perplexity": 40.39226073941217, "lr": 0.0026291804804649314, "grad_norm": 0.142417, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:54.171617+00:00", "epoch": 0, "step": 3472, "train_loss": 3.6218225955963135, "perplexity": 37.40568116142416, "lr": 0.0026291804804649314, "grad_norm": 0.166632, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:54.478047+00:00", "epoch": 0, "step": 3473, "train_loss": 3.685971736907959, "perplexity": 39.88386024437215, "lr": 0.0026291804804649314, "grad_norm": 0.143228, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:54.783657+00:00", "epoch": 0, "step": 3474, "train_loss": 3.6197915077209473, "perplexity": 37.329784038901686, "lr": 0.0026291804804649314, "grad_norm": 0.169013, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:55.089537+00:00", "epoch": 0, "step": 3475, "train_loss": 3.6305506229400635, "perplexity": 37.733587877186906, "lr": 0.0026291804804649314, "grad_norm": 0.151151, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:55.394981+00:00", "epoch": 0, "step": 3476, "train_loss": 3.527157783508301, "perplexity": 34.027117610774525, "lr": 0.0026291804804649314, "grad_norm": 0.164846, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:55.700241+00:00", "epoch": 0, "step": 3477, "train_loss": 3.590054750442505, "perplexity": 36.23605981247602, "lr": 0.0026291804804649314, "grad_norm": 0.173637, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:56.006250+00:00", "epoch": 0, "step": 3478, "train_loss": 3.753608465194702, "perplexity": 42.674795011735725, "lr": 0.0026291804804649314, "grad_norm": 0.166997, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:56.312411+00:00", "epoch": 0, "step": 3479, "train_loss": 3.624803066253662, "perplexity": 37.51733400289992, "lr": 0.0026291804804649314, "grad_norm": 0.170088, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:56.617839+00:00", "epoch": 0, "step": 3480, "train_loss": 3.656777858734131, "perplexity": 38.73632765133028, "lr": 0.0026291804804649314, "grad_norm": 0.156321, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:56.922435+00:00", "epoch": 0, "step": 3481, "train_loss": 3.5858540534973145, "perplexity": 36.08416236764555, "lr": 0.0026291804804649314, "grad_norm": 0.148807, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:57.227422+00:00", "epoch": 0, "step": 3482, "train_loss": 3.734304904937744, "perplexity": 41.85891952027139, "lr": 0.0026291804804649314, "grad_norm": 0.181188, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:57.532452+00:00", "epoch": 0, "step": 3483, "train_loss": 3.5945472717285156, "perplexity": 36.39921730233904, "lr": 0.0026291804804649314, "grad_norm": 0.145501, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:57.837783+00:00", "epoch": 0, "step": 3484, "train_loss": 3.640986680984497, "perplexity": 38.12943976441164, "lr": 0.0026291804804649314, "grad_norm": 0.1485, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:58.143005+00:00", "epoch": 0, "step": 3485, "train_loss": 3.6451456546783447, "perplexity": 38.28834932263351, "lr": 0.0026291804804649314, "grad_norm": 0.152356, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:58.448288+00:00", "epoch": 0, "step": 3486, "train_loss": 3.718383312225342, "perplexity": 41.1977363640256, "lr": 0.0026291804804649314, "grad_norm": 0.145811, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:58.752802+00:00", "epoch": 0, "step": 3487, "train_loss": 3.7002313137054443, "perplexity": 40.45666145808494, "lr": 0.0026291804804649314, "grad_norm": 0.139538, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:59.058195+00:00", "epoch": 0, "step": 3488, "train_loss": 3.6599526405334473, "perplexity": 38.85950246244529, "lr": 0.0026291804804649314, "grad_norm": 0.167844, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:59.363878+00:00", "epoch": 0, "step": 3489, "train_loss": 3.7007246017456055, "perplexity": 40.47662316835875, "lr": 0.0026291804804649314, "grad_norm": 0.172968, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:59.670114+00:00", "epoch": 0, "step": 3490, "train_loss": 3.6222167015075684, "perplexity": 37.420425866781315, "lr": 0.0026291804804649314, "grad_norm": 0.186436, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:15:59.975237+00:00", "epoch": 0, "step": 3491, "train_loss": 3.689728021621704, "perplexity": 40.033957105721385, "lr": 0.0026291804804649314, "grad_norm": 0.175876, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:00.280110+00:00", "epoch": 0, "step": 3492, "train_loss": 3.6475961208343506, "perplexity": 38.38228867741563, "lr": 0.0026291804804649314, "grad_norm": 0.161006, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:00.584674+00:00", "epoch": 0, "step": 3493, "train_loss": 3.6270453929901123, "perplexity": 37.6015545136779, "lr": 0.0026291804804649314, "grad_norm": 0.153349, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:00.889387+00:00", "epoch": 0, "step": 3494, "train_loss": 3.6416680812835693, "perplexity": 38.155430029952306, "lr": 0.0026291804804649314, "grad_norm": 0.17592, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:01.194699+00:00", "epoch": 0, "step": 3495, "train_loss": 3.621386766433716, "perplexity": 37.3893822267607, "lr": 0.0026291804804649314, "grad_norm": 0.150933, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:01.499806+00:00", "epoch": 0, "step": 3496, "train_loss": 3.618457794189453, "perplexity": 37.28002998700256, "lr": 0.0026291804804649314, "grad_norm": 0.171878, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:01.804691+00:00", "epoch": 0, "step": 3497, "train_loss": 3.7353665828704834, "perplexity": 41.903383810615765, "lr": 0.0026291804804649314, "grad_norm": 0.16376, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:02.108932+00:00", "epoch": 0, "step": 3498, "train_loss": 3.716315507888794, "perplexity": 41.11263552238376, "lr": 0.0026291804804649314, "grad_norm": 0.168891, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:02.413871+00:00", "epoch": 0, "step": 3499, "train_loss": 3.6547839641571045, "perplexity": 38.65916844691494, "lr": 0.0026291804804649314, "grad_norm": 0.201334, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:02.719887+00:00", "epoch": 0, "step": 3500, "train_loss": 3.726304769515991, "perplexity": 41.52537846126532, "lr": 0.0026291804804649314, "grad_norm": 0.182176, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:16:05.792423+00:00", "step": 3500, "epoch": 0, "val_loss": 3.6162638783454897, "val_ppl": 37.19833039233839, "eval_train_loss": 3.726304769515991, "eval_train_ppl": 41.52537846126532} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:16:06.699851+00:00", "step": 3500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p6163_epoch_0000_step_0003500.pt", "val_loss": 3.6162638783454897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:07.756126+00:00", "epoch": 0, "step": 3501, "train_loss": 3.6904075145721436, "perplexity": 40.06116914149915, "lr": 0.0026291804804649314, "grad_norm": 0.149278, "tokens_per_sec": 6506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:08.061081+00:00", "epoch": 0, "step": 3502, "train_loss": 3.680232524871826, "perplexity": 39.65561391715274, "lr": 0.0026291804804649314, "grad_norm": 0.14561, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:08.364473+00:00", "epoch": 0, "step": 3503, "train_loss": 3.771984100341797, "perplexity": 43.46622067997568, "lr": 0.0026291804804649314, "grad_norm": 0.13779, "tokens_per_sec": 108011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:08.667349+00:00", "epoch": 0, "step": 3504, "train_loss": 3.6228227615356445, "perplexity": 37.44311176494641, "lr": 0.0026291804804649314, "grad_norm": 0.137503, "tokens_per_sec": 108183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:08.970298+00:00", "epoch": 0, "step": 3505, "train_loss": 3.7896952629089355, "perplexity": 44.24291576402211, "lr": 0.0026291804804649314, "grad_norm": 0.135157, "tokens_per_sec": 108164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:09.275073+00:00", "epoch": 0, "step": 3506, "train_loss": 3.725693941116333, "perplexity": 41.50002132601222, "lr": 0.0026291804804649314, "grad_norm": 0.14443, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:09.579932+00:00", "epoch": 0, "step": 3507, "train_loss": 3.678999900817871, "perplexity": 39.60676356680765, "lr": 0.0026291804804649314, "grad_norm": 0.14967, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:09.883722+00:00", "epoch": 0, "step": 3508, "train_loss": 3.695011615753174, "perplexity": 40.24604007332426, "lr": 0.0026291804804649314, "grad_norm": 0.147303, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:10.186788+00:00", "epoch": 0, "step": 3509, "train_loss": 3.786929130554199, "perplexity": 44.120703109374176, "lr": 0.0026291804804649314, "grad_norm": 0.137392, "tokens_per_sec": 108121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:10.490737+00:00", "epoch": 0, "step": 3510, "train_loss": 3.7634761333465576, "perplexity": 43.09798021913611, "lr": 0.0026291804804649314, "grad_norm": 0.167821, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:10.795101+00:00", "epoch": 0, "step": 3511, "train_loss": 3.6386563777923584, "perplexity": 38.04069005623985, "lr": 0.0026291804804649314, "grad_norm": 0.143836, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:11.099336+00:00", "epoch": 0, "step": 3512, "train_loss": 3.7348759174346924, "perplexity": 41.88282831188598, "lr": 0.0026291804804649314, "grad_norm": 0.150019, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:11.402727+00:00", "epoch": 0, "step": 3513, "train_loss": 3.7142958641052246, "perplexity": 41.02968643562404, "lr": 0.0026291804804649314, "grad_norm": 0.140539, "tokens_per_sec": 108006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:11.707658+00:00", "epoch": 0, "step": 3514, "train_loss": 3.689495086669922, "perplexity": 40.02463288386406, "lr": 0.0026291804804649314, "grad_norm": 0.151043, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:12.094922+00:00", "epoch": 0, "step": 3515, "train_loss": 3.814262866973877, "perplexity": 45.3433200022433, "lr": 0.0026291804804649314, "grad_norm": 0.183792, "tokens_per_sec": 84614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:12.398914+00:00", "epoch": 0, "step": 3516, "train_loss": 3.635277271270752, "perplexity": 37.9123634491285, "lr": 0.0026291804804649314, "grad_norm": 0.162913, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:12.703032+00:00", "epoch": 0, "step": 3517, "train_loss": 3.6185741424560547, "perplexity": 37.28436770620856, "lr": 0.0026291804804649314, "grad_norm": 0.156677, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:13.007883+00:00", "epoch": 0, "step": 3518, "train_loss": 3.6245691776275635, "perplexity": 37.50856015128715, "lr": 0.0026291804804649314, "grad_norm": 0.146687, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:13.312820+00:00", "epoch": 0, "step": 3519, "train_loss": 3.614368200302124, "perplexity": 37.12788112982101, "lr": 0.0026291804804649314, "grad_norm": 0.146143, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:13.617151+00:00", "epoch": 0, "step": 3520, "train_loss": 3.592966318130493, "perplexity": 36.34171729317925, "lr": 0.0026291804804649314, "grad_norm": 0.141353, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:13.921652+00:00", "epoch": 0, "step": 3521, "train_loss": 3.6584157943725586, "perplexity": 38.79982725282579, "lr": 0.0026291804804649314, "grad_norm": 0.15347, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:14.225862+00:00", "epoch": 0, "step": 3522, "train_loss": 3.738199472427368, "perplexity": 42.022259770743105, "lr": 0.0026291804804649314, "grad_norm": 0.179755, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:14.530429+00:00", "epoch": 0, "step": 3523, "train_loss": 3.6684112548828125, "perplexity": 39.1895940983892, "lr": 0.0026291804804649314, "grad_norm": 0.191894, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:14.834583+00:00", "epoch": 0, "step": 3524, "train_loss": 3.685051918029785, "perplexity": 39.84719118381738, "lr": 0.0026291804804649314, "grad_norm": 0.204345, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:15.139592+00:00", "epoch": 0, "step": 3525, "train_loss": 3.721886157989502, "perplexity": 41.342298722349454, "lr": 0.0026291804804649314, "grad_norm": 0.161816, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:15.443512+00:00", "epoch": 0, "step": 3526, "train_loss": 3.73773193359375, "perplexity": 42.00261732458488, "lr": 0.0026291804804649314, "grad_norm": 0.151413, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:15.748357+00:00", "epoch": 0, "step": 3527, "train_loss": 3.6159350872039795, "perplexity": 37.186101921243065, "lr": 0.0026291804804649314, "grad_norm": 0.145693, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:16.053601+00:00", "epoch": 0, "step": 3528, "train_loss": 3.7178292274475098, "perplexity": 41.1749156483146, "lr": 0.0026291804804649314, "grad_norm": 0.152714, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:16.358550+00:00", "epoch": 0, "step": 3529, "train_loss": 3.6111679077148438, "perplexity": 37.00925097391367, "lr": 0.0026291804804649314, "grad_norm": 0.153544, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:16.662423+00:00", "epoch": 0, "step": 3530, "train_loss": 3.6398231983184814, "perplexity": 38.08510261993196, "lr": 0.0026291804804649314, "grad_norm": 0.152723, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:16.966028+00:00", "epoch": 0, "step": 3531, "train_loss": 3.6198174953460693, "perplexity": 37.330754163940746, "lr": 0.0026291804804649314, "grad_norm": 0.14444, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:17.270772+00:00", "epoch": 0, "step": 3532, "train_loss": 3.78016996383667, "perplexity": 43.823489511043356, "lr": 0.0026291804804649314, "grad_norm": 0.144314, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:17.575381+00:00", "epoch": 0, "step": 3533, "train_loss": 3.685394287109375, "perplexity": 39.86083596562963, "lr": 0.0026291804804649314, "grad_norm": 0.143929, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:17.878696+00:00", "epoch": 0, "step": 3534, "train_loss": 3.8001368045806885, "perplexity": 44.707300238423144, "lr": 0.0026291804804649314, "grad_norm": 0.144143, "tokens_per_sec": 108033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:18.183395+00:00", "epoch": 0, "step": 3535, "train_loss": 3.6374053955078125, "perplexity": 37.99313158049516, "lr": 0.0026291804804649314, "grad_norm": 0.16994, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:18.486921+00:00", "epoch": 0, "step": 3536, "train_loss": 3.6031670570373535, "perplexity": 36.7143268782722, "lr": 0.0026291804804649314, "grad_norm": 0.209296, "tokens_per_sec": 107958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:18.792321+00:00", "epoch": 0, "step": 3537, "train_loss": 3.583115816116333, "perplexity": 35.98549052051218, "lr": 0.0026291804804649314, "grad_norm": 0.189791, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:19.095843+00:00", "epoch": 0, "step": 3538, "train_loss": 3.5982487201690674, "perplexity": 36.534196784146665, "lr": 0.0026291804804649314, "grad_norm": 0.148762, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:19.399482+00:00", "epoch": 0, "step": 3539, "train_loss": 3.580554723739624, "perplexity": 35.89344627229052, "lr": 0.0026291804804649314, "grad_norm": 0.147934, "tokens_per_sec": 107918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:19.703949+00:00", "epoch": 0, "step": 3540, "train_loss": 3.7073779106140137, "perplexity": 40.74682451401398, "lr": 0.0026291804804649314, "grad_norm": 0.175437, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:20.009353+00:00", "epoch": 0, "step": 3541, "train_loss": 3.652787685394287, "perplexity": 38.582070949608074, "lr": 0.0026291804804649314, "grad_norm": 0.182255, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:20.314900+00:00", "epoch": 0, "step": 3542, "train_loss": 3.574592113494873, "perplexity": 35.68006443039791, "lr": 0.0026291804804649314, "grad_norm": 0.162837, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:20.619266+00:00", "epoch": 0, "step": 3543, "train_loss": 3.6366658210754395, "perplexity": 37.965043219769456, "lr": 0.0026291804804649314, "grad_norm": 0.157521, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:20.922925+00:00", "epoch": 0, "step": 3544, "train_loss": 3.8003928661346436, "perplexity": 44.718749524993825, "lr": 0.0026291804804649314, "grad_norm": 0.148107, "tokens_per_sec": 107909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:21.226792+00:00", "epoch": 0, "step": 3545, "train_loss": 3.652165412902832, "perplexity": 38.55806985657723, "lr": 0.0026291804804649314, "grad_norm": 0.149329, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:21.531326+00:00", "epoch": 0, "step": 3546, "train_loss": 3.706050395965576, "perplexity": 40.69276839567735, "lr": 0.0026291804804649314, "grad_norm": 0.150739, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:21.835774+00:00", "epoch": 0, "step": 3547, "train_loss": 3.741347312927246, "perplexity": 42.154747557756025, "lr": 0.0026291804804649314, "grad_norm": 0.173096, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:22.140198+00:00", "epoch": 0, "step": 3548, "train_loss": 3.618783950805664, "perplexity": 37.292191098541046, "lr": 0.0026291804804649314, "grad_norm": 0.16683, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:22.443729+00:00", "epoch": 0, "step": 3549, "train_loss": 3.614441156387329, "perplexity": 37.13058993349085, "lr": 0.0026291804804649314, "grad_norm": 0.148508, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:22.747666+00:00", "epoch": 0, "step": 3550, "train_loss": 3.5296764373779297, "perplexity": 34.112928160391206, "lr": 0.0026291804804649314, "grad_norm": 0.160164, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:23.051642+00:00", "epoch": 0, "step": 3551, "train_loss": 3.8077166080474854, "perplexity": 45.047460332479716, "lr": 0.0026291804804649314, "grad_norm": 0.182182, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:23.358340+00:00", "epoch": 0, "step": 3552, "train_loss": 3.673236846923828, "perplexity": 39.37916411768335, "lr": 0.0026291804804649314, "grad_norm": 0.167021, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:23.663398+00:00", "epoch": 0, "step": 3553, "train_loss": 3.6640255451202393, "perplexity": 39.018096257726434, "lr": 0.0026291804804649314, "grad_norm": 0.159714, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:23.967440+00:00", "epoch": 0, "step": 3554, "train_loss": 3.6581971645355225, "perplexity": 38.791345380145486, "lr": 0.0026291804804649314, "grad_norm": 0.165741, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:24.270981+00:00", "epoch": 0, "step": 3555, "train_loss": 3.5510854721069336, "perplexity": 34.8511268894978, "lr": 0.0026291804804649314, "grad_norm": 0.166572, "tokens_per_sec": 107954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:24.575334+00:00", "epoch": 0, "step": 3556, "train_loss": 3.6828691959381104, "perplexity": 39.76031069180744, "lr": 0.0026291804804649314, "grad_norm": 0.157614, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:24.879655+00:00", "epoch": 0, "step": 3557, "train_loss": 3.6583454608917236, "perplexity": 38.79709842188452, "lr": 0.0026291804804649314, "grad_norm": 0.169579, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:25.184488+00:00", "epoch": 0, "step": 3558, "train_loss": 3.6687240600585938, "perplexity": 39.201854723753506, "lr": 0.0026291804804649314, "grad_norm": 0.185922, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:25.489583+00:00", "epoch": 0, "step": 3559, "train_loss": 3.6034114360809326, "perplexity": 36.72330018675998, "lr": 0.0026291804804649314, "grad_norm": 0.173414, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:25.794690+00:00", "epoch": 0, "step": 3560, "train_loss": 3.7536768913269043, "perplexity": 42.67771518280776, "lr": 0.0026291804804649314, "grad_norm": 0.177692, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:26.098066+00:00", "epoch": 0, "step": 3561, "train_loss": 3.6549577713012695, "perplexity": 38.665888270538254, "lr": 0.0026291804804649314, "grad_norm": 0.168656, "tokens_per_sec": 107949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:26.402884+00:00", "epoch": 0, "step": 3562, "train_loss": 3.6044816970825195, "perplexity": 36.76262474281509, "lr": 0.0026291804804649314, "grad_norm": 0.16687, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:26.707865+00:00", "epoch": 0, "step": 3563, "train_loss": 3.6764605045318604, "perplexity": 39.506313893248596, "lr": 0.0026291804804649314, "grad_norm": 0.156194, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:27.012202+00:00", "epoch": 0, "step": 3564, "train_loss": 3.6116480827331543, "perplexity": 37.027026158936096, "lr": 0.0026291804804649314, "grad_norm": 0.159022, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:27.316024+00:00", "epoch": 0, "step": 3565, "train_loss": 3.616088390350342, "perplexity": 37.19180310466206, "lr": 0.0026291804804649314, "grad_norm": 0.137244, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:27.620471+00:00", "epoch": 0, "step": 3566, "train_loss": 3.6765296459198, "perplexity": 39.509045509056314, "lr": 0.0026291804804649314, "grad_norm": 0.144356, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:27.925358+00:00", "epoch": 0, "step": 3567, "train_loss": 3.62607479095459, "perplexity": 37.565076074215675, "lr": 0.0026291804804649314, "grad_norm": 0.14502, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:28.230831+00:00", "epoch": 0, "step": 3568, "train_loss": 3.694507598876953, "perplexity": 40.22576050097891, "lr": 0.0026291804804649314, "grad_norm": 0.145225, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:28.535890+00:00", "epoch": 0, "step": 3569, "train_loss": 3.7127292156219482, "perplexity": 40.96545766469051, "lr": 0.0026291804804649314, "grad_norm": 0.167311, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:28.839919+00:00", "epoch": 0, "step": 3570, "train_loss": 3.6816914081573486, "perplexity": 39.71350905032627, "lr": 0.0026291804804649314, "grad_norm": 0.146937, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:29.143248+00:00", "epoch": 0, "step": 3571, "train_loss": 3.643068790435791, "perplexity": 38.20891213769685, "lr": 0.0026291804804649314, "grad_norm": 0.141891, "tokens_per_sec": 107971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:29.447408+00:00", "epoch": 0, "step": 3572, "train_loss": 3.6423614025115967, "perplexity": 38.18189317222568, "lr": 0.0026291804804649314, "grad_norm": 0.149325, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:29.751684+00:00", "epoch": 0, "step": 3573, "train_loss": 3.706800937652588, "perplexity": 40.72332147896531, "lr": 0.0026291804804649314, "grad_norm": 0.14963, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:30.056864+00:00", "epoch": 0, "step": 3574, "train_loss": 3.722622871398926, "perplexity": 41.37276737014725, "lr": 0.0026291804804649314, "grad_norm": 0.151002, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:30.361285+00:00", "epoch": 0, "step": 3575, "train_loss": 3.6799674034118652, "perplexity": 39.64510175645665, "lr": 0.0026291804804649314, "grad_norm": 0.137869, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:30.665298+00:00", "epoch": 0, "step": 3576, "train_loss": 3.6580512523651123, "perplexity": 38.78568566366874, "lr": 0.0026291804804649314, "grad_norm": 0.143514, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:30.969455+00:00", "epoch": 0, "step": 3577, "train_loss": 3.5650339126586914, "perplexity": 35.34065187942296, "lr": 0.0026291804804649314, "grad_norm": 0.150524, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:31.274415+00:00", "epoch": 0, "step": 3578, "train_loss": 3.729996919631958, "perplexity": 41.678979777234225, "lr": 0.0026291804804649314, "grad_norm": 0.141951, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:31.578731+00:00", "epoch": 0, "step": 3579, "train_loss": 3.5603857040405273, "perplexity": 35.176762348231996, "lr": 0.0026291804804649314, "grad_norm": 0.151652, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:31.883079+00:00", "epoch": 0, "step": 3580, "train_loss": 3.6355226039886475, "perplexity": 37.921665733325995, "lr": 0.0026291804804649314, "grad_norm": 0.160834, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:32.187190+00:00", "epoch": 0, "step": 3581, "train_loss": 3.6684937477111816, "perplexity": 39.19282709219658, "lr": 0.0026291804804649314, "grad_norm": 0.154225, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:32.490542+00:00", "epoch": 0, "step": 3582, "train_loss": 3.6013967990875244, "perplexity": 36.64939054323318, "lr": 0.0026291804804649314, "grad_norm": 0.15406, "tokens_per_sec": 108020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:32.795524+00:00", "epoch": 0, "step": 3583, "train_loss": 3.524120807647705, "perplexity": 33.92393483708787, "lr": 0.0026291804804649314, "grad_norm": 0.170387, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:33.100328+00:00", "epoch": 0, "step": 3584, "train_loss": 3.6672189235687256, "perplexity": 39.1428949641119, "lr": 0.0026291804804649314, "grad_norm": 0.182809, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:33.405479+00:00", "epoch": 0, "step": 3585, "train_loss": 3.657074213027954, "perplexity": 38.74780902955053, "lr": 0.0026291804804649314, "grad_norm": 0.159079, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:33.710505+00:00", "epoch": 0, "step": 3586, "train_loss": 3.6103289127349854, "perplexity": 36.978213420134786, "lr": 0.0026291804804649314, "grad_norm": 0.164438, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:34.015833+00:00", "epoch": 0, "step": 3587, "train_loss": 3.631619930267334, "perplexity": 37.773958259514316, "lr": 0.0026291804804649314, "grad_norm": 0.160084, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:34.320495+00:00", "epoch": 0, "step": 3588, "train_loss": 3.7152907848358154, "perplexity": 41.07052803493642, "lr": 0.0026291804804649314, "grad_norm": 0.159057, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:34.626141+00:00", "epoch": 0, "step": 3589, "train_loss": 3.7307567596435547, "perplexity": 41.710661168581595, "lr": 0.0026291804804649314, "grad_norm": 0.154421, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:34.930182+00:00", "epoch": 0, "step": 3590, "train_loss": 3.696307897567749, "perplexity": 40.29824411144253, "lr": 0.0026291804804649314, "grad_norm": 0.175034, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:35.233814+00:00", "epoch": 0, "step": 3591, "train_loss": 3.740739345550537, "perplexity": 42.1291266355991, "lr": 0.0026291804804649314, "grad_norm": 0.188165, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:35.537822+00:00", "epoch": 0, "step": 3592, "train_loss": 3.738837480545044, "perplexity": 42.04907886809109, "lr": 0.0026291804804649314, "grad_norm": 0.171454, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:35.841810+00:00", "epoch": 0, "step": 3593, "train_loss": 3.57060170173645, "perplexity": 35.537969977992425, "lr": 0.0026291804804649314, "grad_norm": 0.167644, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:36.146172+00:00", "epoch": 0, "step": 3594, "train_loss": 3.7227320671081543, "perplexity": 41.37728534549025, "lr": 0.0026291804804649314, "grad_norm": 0.162343, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:36.450311+00:00", "epoch": 0, "step": 3595, "train_loss": 3.707261562347412, "perplexity": 40.7420839673948, "lr": 0.0026291804804649314, "grad_norm": 0.169243, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:36.753961+00:00", "epoch": 0, "step": 3596, "train_loss": 3.647430658340454, "perplexity": 38.37593837359265, "lr": 0.0026291804804649314, "grad_norm": 0.180197, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:37.057735+00:00", "epoch": 0, "step": 3597, "train_loss": 3.695828676223755, "perplexity": 40.278936959308034, "lr": 0.0026291804804649314, "grad_norm": 0.156681, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:37.361687+00:00", "epoch": 0, "step": 3598, "train_loss": 3.766127347946167, "perplexity": 43.21239381398612, "lr": 0.0026291804804649314, "grad_norm": 0.147096, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:37.667076+00:00", "epoch": 0, "step": 3599, "train_loss": 3.606707811355591, "perplexity": 36.84455370425203, "lr": 0.0026291804804649314, "grad_norm": 0.163847, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:37.973132+00:00", "epoch": 0, "step": 3600, "train_loss": 3.623403549194336, "perplexity": 37.46486457843384, "lr": 0.0026291804804649314, "grad_norm": 0.173184, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:38.276269+00:00", "epoch": 0, "step": 3601, "train_loss": 3.624002695083618, "perplexity": 37.48731822387151, "lr": 0.0026291804804649314, "grad_norm": 0.161615, "tokens_per_sec": 108020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:38.580437+00:00", "epoch": 0, "step": 3602, "train_loss": 3.5719780921936035, "perplexity": 35.58691777865217, "lr": 0.0026291804804649314, "grad_norm": 0.15139, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:38.884292+00:00", "epoch": 0, "step": 3603, "train_loss": 3.6708414554595947, "perplexity": 39.284948490791415, "lr": 0.0026291804804649314, "grad_norm": 0.14601, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:39.188422+00:00", "epoch": 0, "step": 3604, "train_loss": 3.6355559825897217, "perplexity": 37.92293152660366, "lr": 0.0026291804804649314, "grad_norm": 0.154746, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:39.493117+00:00", "epoch": 0, "step": 3605, "train_loss": 3.678680658340454, "perplexity": 39.594121423546056, "lr": 0.0026291804804649314, "grad_norm": 0.154895, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:39.797206+00:00", "epoch": 0, "step": 3606, "train_loss": 3.6848526000976562, "perplexity": 39.83924971553428, "lr": 0.0026291804804649314, "grad_norm": 0.150722, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:40.100926+00:00", "epoch": 0, "step": 3607, "train_loss": 3.670297145843506, "perplexity": 39.26357113403854, "lr": 0.0026291804804649314, "grad_norm": 0.169417, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:40.404376+00:00", "epoch": 0, "step": 3608, "train_loss": 3.649298667907715, "perplexity": 38.447691990983905, "lr": 0.0026291804804649314, "grad_norm": 0.155323, "tokens_per_sec": 107985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:40.708535+00:00", "epoch": 0, "step": 3609, "train_loss": 3.7423276901245117, "perplexity": 42.196095375935435, "lr": 0.0026291804804649314, "grad_norm": 0.171125, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:41.013895+00:00", "epoch": 0, "step": 3610, "train_loss": 3.7207958698272705, "perplexity": 41.2972482669049, "lr": 0.0026291804804649314, "grad_norm": 0.167188, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:41.318377+00:00", "epoch": 0, "step": 3611, "train_loss": 3.6898601055145264, "perplexity": 40.039245295855665, "lr": 0.0026291804804649314, "grad_norm": 0.150781, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:41.622384+00:00", "epoch": 0, "step": 3612, "train_loss": 3.643054962158203, "perplexity": 38.20838377790664, "lr": 0.0026291804804649314, "grad_norm": 0.153155, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:41.926093+00:00", "epoch": 0, "step": 3613, "train_loss": 3.6390867233276367, "perplexity": 38.057064220385975, "lr": 0.0026291804804649314, "grad_norm": 0.155911, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:42.230376+00:00", "epoch": 0, "step": 3614, "train_loss": 3.6696105003356934, "perplexity": 39.23662023321507, "lr": 0.0026291804804649314, "grad_norm": 0.146575, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:42.534462+00:00", "epoch": 0, "step": 3615, "train_loss": 3.4735591411590576, "perplexity": 32.25132543506828, "lr": 0.0026291804804649314, "grad_norm": 0.156808, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:42.838844+00:00", "epoch": 0, "step": 3616, "train_loss": 3.717621326446533, "perplexity": 41.16635623192264, "lr": 0.0026291804804649314, "grad_norm": 0.149163, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:43.143299+00:00", "epoch": 0, "step": 3617, "train_loss": 3.725836992263794, "perplexity": 41.50595837632336, "lr": 0.0026291804804649314, "grad_norm": 0.151424, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:43.447145+00:00", "epoch": 0, "step": 3618, "train_loss": 3.5713951587677, "perplexity": 35.566179019999204, "lr": 0.0026291804804649314, "grad_norm": 0.130176, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:43.751313+00:00", "epoch": 0, "step": 3619, "train_loss": 3.696816921234131, "perplexity": 40.31876209303497, "lr": 0.0026291804804649314, "grad_norm": 0.151982, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:44.055876+00:00", "epoch": 0, "step": 3620, "train_loss": 3.6120095252990723, "perplexity": 37.04041172118955, "lr": 0.0026291804804649314, "grad_norm": 0.167007, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:44.361125+00:00", "epoch": 0, "step": 3621, "train_loss": 3.66975474357605, "perplexity": 39.24228025865854, "lr": 0.0026291804804649314, "grad_norm": 0.168756, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:44.665323+00:00", "epoch": 0, "step": 3622, "train_loss": 3.6058835983276367, "perplexity": 36.81419845439446, "lr": 0.0026291804804649314, "grad_norm": 0.175638, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:44.969748+00:00", "epoch": 0, "step": 3623, "train_loss": 3.7547430992126465, "perplexity": 42.723242765902214, "lr": 0.0026291804804649314, "grad_norm": 0.172346, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:45.274294+00:00", "epoch": 0, "step": 3624, "train_loss": 3.6417198181152344, "perplexity": 38.157404122079065, "lr": 0.0026291804804649314, "grad_norm": 0.156801, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:45.580050+00:00", "epoch": 0, "step": 3625, "train_loss": 3.5718445777893066, "perplexity": 35.58216672969798, "lr": 0.0026291804804649314, "grad_norm": 0.163478, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:45.884820+00:00", "epoch": 0, "step": 3626, "train_loss": 3.5828750133514404, "perplexity": 35.9768261581419, "lr": 0.0026291804804649314, "grad_norm": 0.173884, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:46.190163+00:00", "epoch": 0, "step": 3627, "train_loss": 3.5035171508789062, "perplexity": 33.23212906486945, "lr": 0.0026291804804649314, "grad_norm": 0.16351, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:46.493684+00:00", "epoch": 0, "step": 3628, "train_loss": 3.707115411758423, "perplexity": 40.73612992293046, "lr": 0.0026291804804649314, "grad_norm": 0.167821, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:46.797901+00:00", "epoch": 0, "step": 3629, "train_loss": 3.682224750518799, "perplexity": 39.73469559636368, "lr": 0.0026291804804649314, "grad_norm": 0.16576, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:47.103052+00:00", "epoch": 0, "step": 3630, "train_loss": 3.623589038848877, "perplexity": 37.47181456777755, "lr": 0.0026291804804649314, "grad_norm": 0.145351, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:47.408369+00:00", "epoch": 0, "step": 3631, "train_loss": 3.731985569000244, "perplexity": 41.76194712317632, "lr": 0.0026291804804649314, "grad_norm": 0.145904, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:47.712816+00:00", "epoch": 0, "step": 3632, "train_loss": 3.6280901432037354, "perplexity": 37.64085927404191, "lr": 0.0026291804804649314, "grad_norm": 0.161099, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:48.017044+00:00", "epoch": 0, "step": 3633, "train_loss": 3.7218120098114014, "perplexity": 41.33923337986688, "lr": 0.0026291804804649314, "grad_norm": 0.138863, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:48.321339+00:00", "epoch": 0, "step": 3634, "train_loss": 3.6040773391723633, "perplexity": 36.74776248973953, "lr": 0.0026291804804649314, "grad_norm": 0.162424, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:48.625700+00:00", "epoch": 0, "step": 3635, "train_loss": 3.807971477508545, "perplexity": 45.058943017647486, "lr": 0.0026291804804649314, "grad_norm": 0.150608, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:48.930682+00:00", "epoch": 0, "step": 3636, "train_loss": 3.622089147567749, "perplexity": 37.415653048434656, "lr": 0.0026291804804649314, "grad_norm": 0.143449, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:49.235634+00:00", "epoch": 0, "step": 3637, "train_loss": 3.621849775314331, "perplexity": 37.40669785110699, "lr": 0.0026291804804649314, "grad_norm": 0.163883, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:49.539412+00:00", "epoch": 0, "step": 3638, "train_loss": 3.6642541885375977, "perplexity": 39.02701850856172, "lr": 0.0026291804804649314, "grad_norm": 0.164465, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:49.843054+00:00", "epoch": 0, "step": 3639, "train_loss": 3.651492118835449, "perplexity": 38.53211767459854, "lr": 0.0026291804804649314, "grad_norm": 0.165393, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:50.147468+00:00", "epoch": 0, "step": 3640, "train_loss": 3.6799697875976562, "perplexity": 39.645196277857615, "lr": 0.0026291804804649314, "grad_norm": 0.162577, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:50.451384+00:00", "epoch": 0, "step": 3641, "train_loss": 3.5991158485412598, "perplexity": 36.56589036194783, "lr": 0.0026291804804649314, "grad_norm": 0.161945, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:50.756240+00:00", "epoch": 0, "step": 3642, "train_loss": 3.638429641723633, "perplexity": 38.03206583747257, "lr": 0.0026291804804649314, "grad_norm": 0.179192, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:51.060924+00:00", "epoch": 0, "step": 3643, "train_loss": 3.6749379634857178, "perplexity": 39.44620967594407, "lr": 0.0026291804804649314, "grad_norm": 0.179159, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:51.364902+00:00", "epoch": 0, "step": 3644, "train_loss": 3.6713361740112305, "perplexity": 39.30438829182849, "lr": 0.0026291804804649314, "grad_norm": 0.147735, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:51.668575+00:00", "epoch": 0, "step": 3645, "train_loss": 3.676586627960205, "perplexity": 39.51129687922711, "lr": 0.0026291804804649314, "grad_norm": 0.178998, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:51.973024+00:00", "epoch": 0, "step": 3646, "train_loss": 3.619370222091675, "perplexity": 37.31406084955166, "lr": 0.0026291804804649314, "grad_norm": 0.186778, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:52.278545+00:00", "epoch": 0, "step": 3647, "train_loss": 3.56375789642334, "perplexity": 35.29558539275607, "lr": 0.0026291804804649314, "grad_norm": 0.201215, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:52.583885+00:00", "epoch": 0, "step": 3648, "train_loss": 3.5850255489349365, "perplexity": 36.05427885551682, "lr": 0.0026291804804649314, "grad_norm": 0.203713, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:52.887419+00:00", "epoch": 0, "step": 3649, "train_loss": 3.6309595108032227, "perplexity": 37.749019838059084, "lr": 0.0026291804804649314, "grad_norm": 0.15729, "tokens_per_sec": 107955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:53.191113+00:00", "epoch": 0, "step": 3650, "train_loss": 3.6868791580200195, "perplexity": 39.920068126602246, "lr": 0.0026291804804649314, "grad_norm": 0.175873, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:53.494719+00:00", "epoch": 0, "step": 3651, "train_loss": 3.7791261672973633, "perplexity": 43.777770569139, "lr": 0.0026291804804649314, "grad_norm": 0.150398, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:53.798967+00:00", "epoch": 0, "step": 3652, "train_loss": 3.625053644180298, "perplexity": 37.52673619660918, "lr": 0.0026291804804649314, "grad_norm": 0.157716, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:54.103944+00:00", "epoch": 0, "step": 3653, "train_loss": 3.6790285110473633, "perplexity": 39.60789674161286, "lr": 0.0026291804804649314, "grad_norm": 0.144833, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:54.408010+00:00", "epoch": 0, "step": 3654, "train_loss": 3.6822426319122314, "perplexity": 39.73540611444107, "lr": 0.0026291804804649314, "grad_norm": 0.138434, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:54.712076+00:00", "epoch": 0, "step": 3655, "train_loss": 3.5467143058776855, "perplexity": 34.69911928786982, "lr": 0.0026291804804649314, "grad_norm": 0.14334, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:55.016230+00:00", "epoch": 0, "step": 3656, "train_loss": 3.5927000045776367, "perplexity": 36.33204028994575, "lr": 0.0026291804804649314, "grad_norm": 0.150569, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:55.320623+00:00", "epoch": 0, "step": 3657, "train_loss": 3.5900208950042725, "perplexity": 36.234833045557735, "lr": 0.0026291804804649314, "grad_norm": 0.139567, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:55.625437+00:00", "epoch": 0, "step": 3658, "train_loss": 3.771400213241577, "perplexity": 43.44084872232555, "lr": 0.0026291804804649314, "grad_norm": 0.144584, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:55.930085+00:00", "epoch": 0, "step": 3659, "train_loss": 3.6607491970062256, "perplexity": 38.8904685821549, "lr": 0.0026291804804649314, "grad_norm": 0.141042, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:56.234630+00:00", "epoch": 0, "step": 3660, "train_loss": 3.761531352996826, "perplexity": 43.01424556324107, "lr": 0.0026291804804649314, "grad_norm": 0.152162, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:56.539274+00:00", "epoch": 0, "step": 3661, "train_loss": 3.7828660011291504, "perplexity": 43.94179868436141, "lr": 0.0026291804804649314, "grad_norm": 0.157296, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:56.844997+00:00", "epoch": 0, "step": 3662, "train_loss": 3.7637009620666504, "perplexity": 43.10767097220634, "lr": 0.0026291804804649314, "grad_norm": 0.164427, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:57.150126+00:00", "epoch": 0, "step": 3663, "train_loss": 3.6608998775482178, "perplexity": 38.89632906055814, "lr": 0.0026291804804649314, "grad_norm": 0.154173, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:57.454899+00:00", "epoch": 0, "step": 3664, "train_loss": 3.6087148189544678, "perplexity": 36.918575259578645, "lr": 0.0026291804804649314, "grad_norm": 0.147563, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:57.758202+00:00", "epoch": 0, "step": 3665, "train_loss": 3.716860771179199, "perplexity": 41.135058846057994, "lr": 0.0026291804804649314, "grad_norm": 0.163847, "tokens_per_sec": 108037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:58.061258+00:00", "epoch": 0, "step": 3666, "train_loss": 3.7167747020721436, "perplexity": 41.13151854063208, "lr": 0.0026291804804649314, "grad_norm": 0.173418, "tokens_per_sec": 108125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:58.365897+00:00", "epoch": 0, "step": 3667, "train_loss": 3.5967392921447754, "perplexity": 36.479092642006194, "lr": 0.0026291804804649314, "grad_norm": 0.185878, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:58.671331+00:00", "epoch": 0, "step": 3668, "train_loss": 3.58259916305542, "perplexity": 35.9669033086697, "lr": 0.0026291804804649314, "grad_norm": 0.173859, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:58.977308+00:00", "epoch": 0, "step": 3669, "train_loss": 3.614750623703003, "perplexity": 37.142082415669044, "lr": 0.0026291804804649314, "grad_norm": 0.179212, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:59.281166+00:00", "epoch": 0, "step": 3670, "train_loss": 3.656055212020874, "perplexity": 38.70834508344251, "lr": 0.0026291804804649314, "grad_norm": 0.151217, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:59.584804+00:00", "epoch": 0, "step": 3671, "train_loss": 3.60088849067688, "perplexity": 36.630766083660795, "lr": 0.0026291804804649314, "grad_norm": 0.156855, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:16:59.888967+00:00", "epoch": 0, "step": 3672, "train_loss": 3.661688804626465, "perplexity": 38.927027535635304, "lr": 0.0026291804804649314, "grad_norm": 0.148011, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:00.193145+00:00", "epoch": 0, "step": 3673, "train_loss": 3.5937020778656006, "perplexity": 36.36846590453715, "lr": 0.0026291804804649314, "grad_norm": 0.170154, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:00.498959+00:00", "epoch": 0, "step": 3674, "train_loss": 3.6793689727783203, "perplexity": 39.62138401051626, "lr": 0.0026291804804649314, "grad_norm": 0.168663, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:00.804479+00:00", "epoch": 0, "step": 3675, "train_loss": 3.698258399963379, "perplexity": 40.376922639506205, "lr": 0.0026291804804649314, "grad_norm": 0.156167, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:01.108789+00:00", "epoch": 0, "step": 3676, "train_loss": 3.6950809955596924, "perplexity": 40.248832432663235, "lr": 0.0026291804804649314, "grad_norm": 0.148358, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:01.412851+00:00", "epoch": 0, "step": 3677, "train_loss": 3.7184042930603027, "perplexity": 41.19860073600061, "lr": 0.0026291804804649314, "grad_norm": 0.141149, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:01.717079+00:00", "epoch": 0, "step": 3678, "train_loss": 3.6727161407470703, "perplexity": 39.35866448129854, "lr": 0.0026291804804649314, "grad_norm": 0.144978, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:02.021282+00:00", "epoch": 0, "step": 3679, "train_loss": 3.6853089332580566, "perplexity": 39.85743383495774, "lr": 0.0026291804804649314, "grad_norm": 0.15803, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:02.327303+00:00", "epoch": 0, "step": 3680, "train_loss": 3.6197519302368164, "perplexity": 37.328306649202155, "lr": 0.0026291804804649314, "grad_norm": 0.13962, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:02.630636+00:00", "epoch": 0, "step": 3681, "train_loss": 3.588958740234375, "perplexity": 36.19636647714293, "lr": 0.0026291804804649314, "grad_norm": 0.142744, "tokens_per_sec": 107964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:02.934638+00:00", "epoch": 0, "step": 3682, "train_loss": 3.735974073410034, "perplexity": 41.92884745353387, "lr": 0.0026291804804649314, "grad_norm": 0.161435, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:03.238560+00:00", "epoch": 0, "step": 3683, "train_loss": 3.6454081535339355, "perplexity": 38.298401289770375, "lr": 0.0026291804804649314, "grad_norm": 0.152881, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:03.542895+00:00", "epoch": 0, "step": 3684, "train_loss": 3.6379897594451904, "perplexity": 38.01533988469292, "lr": 0.0026291804804649314, "grad_norm": 0.165437, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:03.847464+00:00", "epoch": 0, "step": 3685, "train_loss": 3.6338422298431396, "perplexity": 37.8579966555908, "lr": 0.0026291804804649314, "grad_norm": 0.182296, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:04.151793+00:00", "epoch": 0, "step": 3686, "train_loss": 3.6584298610687256, "perplexity": 38.800373042045806, "lr": 0.0026291804804649314, "grad_norm": 0.182318, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:04.455634+00:00", "epoch": 0, "step": 3687, "train_loss": 3.63922119140625, "perplexity": 38.062182024772305, "lr": 0.0026291804804649314, "grad_norm": 0.197398, "tokens_per_sec": 107846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:04.759503+00:00", "epoch": 0, "step": 3688, "train_loss": 3.653899908065796, "perplexity": 38.62500667625019, "lr": 0.0026291804804649314, "grad_norm": 0.163444, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:05.064666+00:00", "epoch": 0, "step": 3689, "train_loss": 3.6593308448791504, "perplexity": 38.83534730325077, "lr": 0.0026291804804649314, "grad_norm": 0.148988, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:05.369263+00:00", "epoch": 0, "step": 3690, "train_loss": 3.5824663639068604, "perplexity": 35.96212725166916, "lr": 0.0026291804804649314, "grad_norm": 0.164077, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:05.672387+00:00", "epoch": 0, "step": 3691, "train_loss": 3.7806007862091064, "perplexity": 43.842373718340355, "lr": 0.0026291804804649314, "grad_norm": 0.152466, "tokens_per_sec": 108019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:05.975521+00:00", "epoch": 0, "step": 3692, "train_loss": 3.584165096282959, "perplexity": 36.023269198744394, "lr": 0.0026291804804649314, "grad_norm": 0.159762, "tokens_per_sec": 108097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:06.279583+00:00", "epoch": 0, "step": 3693, "train_loss": 3.56292462348938, "perplexity": 35.266186786992314, "lr": 0.0026291804804649314, "grad_norm": 0.174187, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:06.583267+00:00", "epoch": 0, "step": 3694, "train_loss": 3.660493850708008, "perplexity": 38.880539312721545, "lr": 0.0026291804804649314, "grad_norm": 0.168577, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:06.888170+00:00", "epoch": 0, "step": 3695, "train_loss": 3.6997029781341553, "perplexity": 40.43529241025216, "lr": 0.0026291804804649314, "grad_norm": 0.140885, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:07.192838+00:00", "epoch": 0, "step": 3696, "train_loss": 3.6932780742645264, "perplexity": 40.17633233119186, "lr": 0.0026291804804649314, "grad_norm": 0.140994, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:07.496518+00:00", "epoch": 0, "step": 3697, "train_loss": 3.639065980911255, "perplexity": 38.056274833100574, "lr": 0.0026291804804649314, "grad_norm": 0.186606, "tokens_per_sec": 107903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:07.799586+00:00", "epoch": 0, "step": 3698, "train_loss": 3.716646432876587, "perplexity": 41.12624297219054, "lr": 0.0026291804804649314, "grad_norm": 0.165482, "tokens_per_sec": 108121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:08.103686+00:00", "epoch": 0, "step": 3699, "train_loss": 3.6919338703155518, "perplexity": 40.12236342735008, "lr": 0.0026291804804649314, "grad_norm": 0.138246, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:08.410766+00:00", "epoch": 0, "step": 3700, "train_loss": 3.598985433578491, "perplexity": 36.56112193366163, "lr": 0.0026291804804649314, "grad_norm": 0.161363, "tokens_per_sec": 106781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:08.714209+00:00", "epoch": 0, "step": 3701, "train_loss": 3.573270797729492, "perplexity": 35.63295093151517, "lr": 0.0026291804804649314, "grad_norm": 0.167158, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:09.018055+00:00", "epoch": 0, "step": 3702, "train_loss": 3.6457018852233887, "perplexity": 38.30965239620693, "lr": 0.0026291804804649314, "grad_norm": 0.163783, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:09.322502+00:00", "epoch": 0, "step": 3703, "train_loss": 3.6878814697265625, "perplexity": 39.96010053733701, "lr": 0.0026291804804649314, "grad_norm": 0.152748, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:09.626583+00:00", "epoch": 0, "step": 3704, "train_loss": 3.561807155609131, "perplexity": 35.226799966850834, "lr": 0.0026291804804649314, "grad_norm": 0.155543, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:09.931472+00:00", "epoch": 0, "step": 3705, "train_loss": 3.6104490756988525, "perplexity": 36.98265709883532, "lr": 0.0026291804804649314, "grad_norm": 0.158646, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:10.236785+00:00", "epoch": 0, "step": 3706, "train_loss": 3.6535918712615967, "perplexity": 38.613110584942504, "lr": 0.0026291804804649314, "grad_norm": 0.151203, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:10.541178+00:00", "epoch": 0, "step": 3707, "train_loss": 3.5475502014160156, "perplexity": 34.728136252752705, "lr": 0.0026291804804649314, "grad_norm": 0.142225, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:10.845211+00:00", "epoch": 0, "step": 3708, "train_loss": 3.570919990539551, "perplexity": 35.54928311624845, "lr": 0.0026291804804649314, "grad_norm": 0.159511, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:11.149157+00:00", "epoch": 0, "step": 3709, "train_loss": 3.7012431621551514, "perplexity": 40.49761818576791, "lr": 0.0026291804804649314, "grad_norm": 0.146606, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:11.454081+00:00", "epoch": 0, "step": 3710, "train_loss": 3.7060000896453857, "perplexity": 40.69072134373126, "lr": 0.0026291804804649314, "grad_norm": 0.153198, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:11.758738+00:00", "epoch": 0, "step": 3711, "train_loss": 3.6564834117889404, "perplexity": 38.724923537021155, "lr": 0.0026291804804649314, "grad_norm": 0.142032, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:12.062674+00:00", "epoch": 0, "step": 3712, "train_loss": 3.6077067852020264, "perplexity": 36.88137884039765, "lr": 0.0026291804804649314, "grad_norm": 0.14853, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:12.366223+00:00", "epoch": 0, "step": 3713, "train_loss": 3.698657751083374, "perplexity": 40.393050428895286, "lr": 0.0026291804804649314, "grad_norm": 0.146606, "tokens_per_sec": 107951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:12.670193+00:00", "epoch": 0, "step": 3714, "train_loss": 3.5385665893554688, "perplexity": 34.41754933409732, "lr": 0.0026291804804649314, "grad_norm": 0.142185, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:12.976211+00:00", "epoch": 0, "step": 3715, "train_loss": 3.6924285888671875, "perplexity": 40.14221761558575, "lr": 0.0026291804804649314, "grad_norm": 0.141305, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:13.281817+00:00", "epoch": 0, "step": 3716, "train_loss": 3.6401889324188232, "perplexity": 38.09903418814458, "lr": 0.0026291804804649314, "grad_norm": 0.1469, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:13.587137+00:00", "epoch": 0, "step": 3717, "train_loss": 3.597534418106079, "perplexity": 36.508109650170134, "lr": 0.0026291804804649314, "grad_norm": 0.164454, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:13.890466+00:00", "epoch": 0, "step": 3718, "train_loss": 3.6486964225769043, "perplexity": 38.424544019080535, "lr": 0.0026291804804649314, "grad_norm": 0.142773, "tokens_per_sec": 108029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:14.194507+00:00", "epoch": 0, "step": 3719, "train_loss": 3.6622281074523926, "perplexity": 38.94802665352308, "lr": 0.0026291804804649314, "grad_norm": 0.154809, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:14.499147+00:00", "epoch": 0, "step": 3720, "train_loss": 3.665055513381958, "perplexity": 39.05830436148544, "lr": 0.0026291804804649314, "grad_norm": 0.170043, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:14.803698+00:00", "epoch": 0, "step": 3721, "train_loss": 3.6582865715026855, "perplexity": 38.79481375173408, "lr": 0.0026291804804649314, "grad_norm": 0.167489, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:15.108692+00:00", "epoch": 0, "step": 3722, "train_loss": 3.6619160175323486, "perplexity": 38.93587326357281, "lr": 0.0026291804804649314, "grad_norm": 0.160609, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:15.412958+00:00", "epoch": 0, "step": 3723, "train_loss": 3.6566994190216064, "perplexity": 38.73328930409016, "lr": 0.0026291804804649314, "grad_norm": 0.139911, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:15.717300+00:00", "epoch": 0, "step": 3724, "train_loss": 3.7069315910339355, "perplexity": 40.72864246621116, "lr": 0.0026291804804649314, "grad_norm": 0.141257, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:16.021821+00:00", "epoch": 0, "step": 3725, "train_loss": 3.6502225399017334, "perplexity": 38.483229150214406, "lr": 0.0026291804804649314, "grad_norm": 0.149045, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:16.326883+00:00", "epoch": 0, "step": 3726, "train_loss": 3.7563607692718506, "perplexity": 42.79241080700773, "lr": 0.0026291804804649314, "grad_norm": 0.138659, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:16.631628+00:00", "epoch": 0, "step": 3727, "train_loss": 3.652513265609741, "perplexity": 38.5714847186128, "lr": 0.0026291804804649314, "grad_norm": 0.144847, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:16.936106+00:00", "epoch": 0, "step": 3728, "train_loss": 3.576319456100464, "perplexity": 35.74174938604726, "lr": 0.0026291804804649314, "grad_norm": 0.143967, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:17.241013+00:00", "epoch": 0, "step": 3729, "train_loss": 3.664170026779175, "perplexity": 39.023734064272276, "lr": 0.0026291804804649314, "grad_norm": 0.202632, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:17.545253+00:00", "epoch": 0, "step": 3730, "train_loss": 3.695918083190918, "perplexity": 40.28253833789389, "lr": 0.0026291804804649314, "grad_norm": 0.190245, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:17.850952+00:00", "epoch": 0, "step": 3731, "train_loss": 3.643667221069336, "perplexity": 38.2317843642324, "lr": 0.0026291804804649314, "grad_norm": 0.177279, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:18.156610+00:00", "epoch": 0, "step": 3732, "train_loss": 3.631664991378784, "perplexity": 37.775660434408024, "lr": 0.0026291804804649314, "grad_norm": 0.183867, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:18.460837+00:00", "epoch": 0, "step": 3733, "train_loss": 3.6007721424102783, "perplexity": 36.626504405446916, "lr": 0.0026291804804649314, "grad_norm": 0.210846, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:18.764674+00:00", "epoch": 0, "step": 3734, "train_loss": 3.671144962310791, "perplexity": 39.29687355138452, "lr": 0.0026291804804649314, "grad_norm": 0.172237, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:19.068626+00:00", "epoch": 0, "step": 3735, "train_loss": 3.67745041847229, "perplexity": 39.54544110719863, "lr": 0.0026291804804649314, "grad_norm": 0.157663, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:19.372807+00:00", "epoch": 0, "step": 3736, "train_loss": 3.710047960281372, "perplexity": 40.8557659340329, "lr": 0.0026291804804649314, "grad_norm": 0.16633, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:19.678696+00:00", "epoch": 0, "step": 3737, "train_loss": 3.665837526321411, "perplexity": 39.08886040694348, "lr": 0.0026291804804649314, "grad_norm": 0.159699, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:19.984685+00:00", "epoch": 0, "step": 3738, "train_loss": 3.614124059677124, "perplexity": 37.11881781212409, "lr": 0.0026291804804649314, "grad_norm": 0.197437, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:20.290378+00:00", "epoch": 0, "step": 3739, "train_loss": 3.7174651622772217, "perplexity": 41.15992802403904, "lr": 0.0026291804804649314, "grad_norm": 0.159168, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:20.594746+00:00", "epoch": 0, "step": 3740, "train_loss": 3.6155812740325928, "perplexity": 37.172947315864356, "lr": 0.0026291804804649314, "grad_norm": 0.177072, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:20.899609+00:00", "epoch": 0, "step": 3741, "train_loss": 3.7371392250061035, "perplexity": 41.97772938896816, "lr": 0.0026291804804649314, "grad_norm": 0.182698, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:21.204765+00:00", "epoch": 0, "step": 3742, "train_loss": 3.5762174129486084, "perplexity": 35.73810237136668, "lr": 0.0026291804804649314, "grad_norm": 0.182839, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:21.509261+00:00", "epoch": 0, "step": 3743, "train_loss": 3.6337008476257324, "perplexity": 37.85264458642966, "lr": 0.0026291804804649314, "grad_norm": 0.17254, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:21.813746+00:00", "epoch": 0, "step": 3744, "train_loss": 3.639711618423462, "perplexity": 38.08085332525216, "lr": 0.0026291804804649314, "grad_norm": 0.183013, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:22.117572+00:00", "epoch": 0, "step": 3745, "train_loss": 3.6375908851623535, "perplexity": 38.000179566991044, "lr": 0.0026291804804649314, "grad_norm": 0.149607, "tokens_per_sec": 107851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:22.423350+00:00", "epoch": 0, "step": 3746, "train_loss": 3.6163687705993652, "perplexity": 37.20223241369598, "lr": 0.0026291804804649314, "grad_norm": 0.164962, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:22.728149+00:00", "epoch": 0, "step": 3747, "train_loss": 3.544977903366089, "perplexity": 34.63891993027159, "lr": 0.0026291804804649314, "grad_norm": 0.145335, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:23.033056+00:00", "epoch": 0, "step": 3748, "train_loss": 3.6584672927856445, "perplexity": 38.80182543380845, "lr": 0.0026291804804649314, "grad_norm": 0.160796, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:23.337417+00:00", "epoch": 0, "step": 3749, "train_loss": 3.6323413848876953, "perplexity": 37.801220289206135, "lr": 0.0026291804804649314, "grad_norm": 0.161044, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:23.641858+00:00", "epoch": 0, "step": 3750, "train_loss": 3.7669155597686768, "perplexity": 43.24646776064893, "lr": 0.0026291804804649314, "grad_norm": 0.136159, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:23.946785+00:00", "epoch": 0, "step": 3751, "train_loss": 3.6264448165893555, "perplexity": 37.57897868733805, "lr": 0.0026291804804649314, "grad_norm": 0.160769, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:24.251437+00:00", "epoch": 0, "step": 3752, "train_loss": 3.566631555557251, "perplexity": 35.39715874780777, "lr": 0.0026291804804649314, "grad_norm": 0.142355, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:24.555489+00:00", "epoch": 0, "step": 3753, "train_loss": 3.74204683303833, "perplexity": 42.184245967612945, "lr": 0.0026291804804649314, "grad_norm": 0.161323, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:24.859949+00:00", "epoch": 0, "step": 3754, "train_loss": 3.7082409858703613, "perplexity": 40.7820072705261, "lr": 0.0026291804804649314, "grad_norm": 0.163241, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:25.165558+00:00", "epoch": 0, "step": 3755, "train_loss": 3.58577299118042, "perplexity": 36.081237420393926, "lr": 0.0026291804804649314, "grad_norm": 0.161169, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:25.469889+00:00", "epoch": 0, "step": 3756, "train_loss": 3.5511274337768555, "perplexity": 34.85258933166378, "lr": 0.0026291804804649314, "grad_norm": 0.151619, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:25.774045+00:00", "epoch": 0, "step": 3757, "train_loss": 3.7073845863342285, "perplexity": 40.74709652932203, "lr": 0.0026291804804649314, "grad_norm": 0.171868, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:26.078476+00:00", "epoch": 0, "step": 3758, "train_loss": 3.678999662399292, "perplexity": 39.60675412382049, "lr": 0.0026291804804649314, "grad_norm": 0.159025, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:26.382955+00:00", "epoch": 0, "step": 3759, "train_loss": 3.623544216156006, "perplexity": 37.470135017783115, "lr": 0.0026291804804649314, "grad_norm": 0.150276, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:26.688009+00:00", "epoch": 0, "step": 3760, "train_loss": 3.689563035964966, "perplexity": 40.027352621854, "lr": 0.0026291804804649314, "grad_norm": 0.150166, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:26.991644+00:00", "epoch": 0, "step": 3761, "train_loss": 3.6494786739349365, "perplexity": 38.45461343020674, "lr": 0.0026291804804649314, "grad_norm": 0.145826, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:27.295831+00:00", "epoch": 0, "step": 3762, "train_loss": 3.6655831336975098, "perplexity": 39.07891775390166, "lr": 0.0026291804804649314, "grad_norm": 0.140337, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:27.600750+00:00", "epoch": 0, "step": 3763, "train_loss": 3.786048412322998, "perplexity": 44.081862308183815, "lr": 0.0026291804804649314, "grad_norm": 0.143471, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:27.904394+00:00", "epoch": 0, "step": 3764, "train_loss": 3.6141796112060547, "perplexity": 37.12087987648054, "lr": 0.0026291804804649314, "grad_norm": 0.130751, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:28.209822+00:00", "epoch": 0, "step": 3765, "train_loss": 3.590820074081421, "perplexity": 36.26380274043684, "lr": 0.0026291804804649314, "grad_norm": 0.139768, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:28.513996+00:00", "epoch": 0, "step": 3766, "train_loss": 3.666332244873047, "perplexity": 39.10820317556783, "lr": 0.0026291804804649314, "grad_norm": 0.154602, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:28.818199+00:00", "epoch": 0, "step": 3767, "train_loss": 3.623204231262207, "perplexity": 37.45739790324443, "lr": 0.0026291804804649314, "grad_norm": 0.151344, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:29.122341+00:00", "epoch": 0, "step": 3768, "train_loss": 3.6968069076538086, "perplexity": 40.318358359893665, "lr": 0.0026291804804649314, "grad_norm": 0.153457, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:29.427385+00:00", "epoch": 0, "step": 3769, "train_loss": 3.7279889583587646, "perplexity": 41.59537396664118, "lr": 0.0026291804804649314, "grad_norm": 0.145361, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:29.732658+00:00", "epoch": 0, "step": 3770, "train_loss": 3.6106717586517334, "perplexity": 36.99089342313394, "lr": 0.0026291804804649314, "grad_norm": 0.150437, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:30.036162+00:00", "epoch": 0, "step": 3771, "train_loss": 3.716505289077759, "perplexity": 41.12043866765635, "lr": 0.0026291804804649314, "grad_norm": 0.152155, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:30.340515+00:00", "epoch": 0, "step": 3772, "train_loss": 3.667840003967285, "perplexity": 39.16721339998719, "lr": 0.0026291804804649314, "grad_norm": 0.139843, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:30.644634+00:00", "epoch": 0, "step": 3773, "train_loss": 3.630863666534424, "perplexity": 37.745401984232885, "lr": 0.0026291804804649314, "grad_norm": 0.147065, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:30.948965+00:00", "epoch": 0, "step": 3774, "train_loss": 3.6823434829711914, "perplexity": 39.73941367430585, "lr": 0.0026291804804649314, "grad_norm": 0.155588, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:31.254662+00:00", "epoch": 0, "step": 3775, "train_loss": 3.7412750720977783, "perplexity": 42.151702373821074, "lr": 0.0026291804804649314, "grad_norm": 0.167504, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:31.560601+00:00", "epoch": 0, "step": 3776, "train_loss": 3.8584067821502686, "perplexity": 47.38978893350289, "lr": 0.0026291804804649314, "grad_norm": 0.153692, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:31.864515+00:00", "epoch": 0, "step": 3777, "train_loss": 3.550379514694214, "perplexity": 34.82653216056557, "lr": 0.0026291804804649314, "grad_norm": 0.165846, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:32.168698+00:00", "epoch": 0, "step": 3778, "train_loss": 3.7230725288391113, "perplexity": 41.39137512606072, "lr": 0.0026291804804649314, "grad_norm": 0.160156, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:32.473154+00:00", "epoch": 0, "step": 3779, "train_loss": 3.7325778007507324, "perplexity": 41.78668719943111, "lr": 0.0026291804804649314, "grad_norm": 0.163529, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:32.778171+00:00", "epoch": 0, "step": 3780, "train_loss": 3.6582727432250977, "perplexity": 38.79427728998973, "lr": 0.0026291804804649314, "grad_norm": 0.174964, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:33.082453+00:00", "epoch": 0, "step": 3781, "train_loss": 3.693608045578003, "perplexity": 40.189591555803425, "lr": 0.0026291804804649314, "grad_norm": 0.209479, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:33.385838+00:00", "epoch": 0, "step": 3782, "train_loss": 3.5909175872802734, "perplexity": 36.26733911226328, "lr": 0.0026291804804649314, "grad_norm": 0.194599, "tokens_per_sec": 108008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:33.689700+00:00", "epoch": 0, "step": 3783, "train_loss": 3.719223976135254, "perplexity": 41.23238437578291, "lr": 0.0026291804804649314, "grad_norm": 0.154585, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:33.993532+00:00", "epoch": 0, "step": 3784, "train_loss": 3.7431468963623047, "perplexity": 42.2306768432208, "lr": 0.0026291804804649314, "grad_norm": 0.166739, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:34.299074+00:00", "epoch": 0, "step": 3785, "train_loss": 3.584212303161621, "perplexity": 36.02496978498184, "lr": 0.0026291804804649314, "grad_norm": 0.156836, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:34.603873+00:00", "epoch": 0, "step": 3786, "train_loss": 3.678736925125122, "perplexity": 39.59634932012801, "lr": 0.0026291804804649314, "grad_norm": 0.161507, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:34.908212+00:00", "epoch": 0, "step": 3787, "train_loss": 3.5453343391418457, "perplexity": 34.651268681205885, "lr": 0.0026291804804649314, "grad_norm": 0.154569, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:35.212125+00:00", "epoch": 0, "step": 3788, "train_loss": 3.608931541442871, "perplexity": 36.92657721214743, "lr": 0.0026291804804649314, "grad_norm": 0.148142, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:35.515376+00:00", "epoch": 0, "step": 3789, "train_loss": 3.5288984775543213, "perplexity": 34.08639999307891, "lr": 0.0026291804804649314, "grad_norm": 0.141549, "tokens_per_sec": 108056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:35.820187+00:00", "epoch": 0, "step": 3790, "train_loss": 3.718451499938965, "perplexity": 41.200545639252645, "lr": 0.0026291804804649314, "grad_norm": 0.146562, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:36.124338+00:00", "epoch": 0, "step": 3791, "train_loss": 3.6779911518096924, "perplexity": 39.566830427985806, "lr": 0.0026291804804649314, "grad_norm": 0.166954, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:36.428507+00:00", "epoch": 0, "step": 3792, "train_loss": 3.5967764854431152, "perplexity": 36.480449445013825, "lr": 0.0026291804804649314, "grad_norm": 0.157413, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:36.732752+00:00", "epoch": 0, "step": 3793, "train_loss": 3.7537307739257812, "perplexity": 42.6800148309709, "lr": 0.0026291804804649314, "grad_norm": 0.158826, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:37.037737+00:00", "epoch": 0, "step": 3794, "train_loss": 3.7717936038970947, "perplexity": 43.457941308091996, "lr": 0.0026291804804649314, "grad_norm": 0.14802, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:37.341736+00:00", "epoch": 0, "step": 3795, "train_loss": 3.753603935241699, "perplexity": 42.674601697357765, "lr": 0.0026291804804649314, "grad_norm": 0.171132, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:37.646143+00:00", "epoch": 0, "step": 3796, "train_loss": 3.6058051586151123, "perplexity": 36.81131087250291, "lr": 0.0026291804804649314, "grad_norm": 0.192389, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:37.950129+00:00", "epoch": 0, "step": 3797, "train_loss": 3.6894662380218506, "perplexity": 40.023478243970786, "lr": 0.0026291804804649314, "grad_norm": 0.16738, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:38.254754+00:00", "epoch": 0, "step": 3798, "train_loss": 3.639019727706909, "perplexity": 38.05451464915147, "lr": 0.0026291804804649314, "grad_norm": 0.179041, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:38.557774+00:00", "epoch": 0, "step": 3799, "train_loss": 3.5979597568511963, "perplexity": 36.5236412665803, "lr": 0.0026291804804649314, "grad_norm": 0.158203, "tokens_per_sec": 108134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:38.862996+00:00", "epoch": 0, "step": 3800, "train_loss": 3.705204963684082, "perplexity": 40.6583799542498, "lr": 0.0026291804804649314, "grad_norm": 0.146779, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:39.167860+00:00", "epoch": 0, "step": 3801, "train_loss": 3.6602721214294434, "perplexity": 38.87191931447785, "lr": 0.0026291804804649314, "grad_norm": 0.136606, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:39.472127+00:00", "epoch": 0, "step": 3802, "train_loss": 3.6159889698028564, "perplexity": 37.1881056590395, "lr": 0.0026291804804649314, "grad_norm": 0.157028, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:39.775847+00:00", "epoch": 0, "step": 3803, "train_loss": 3.7599940299987793, "perplexity": 42.948169577372866, "lr": 0.0026291804804649314, "grad_norm": 0.153405, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:40.079186+00:00", "epoch": 0, "step": 3804, "train_loss": 3.6121208667755127, "perplexity": 37.044536084920665, "lr": 0.0026291804804649314, "grad_norm": 0.133267, "tokens_per_sec": 108025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:40.383867+00:00", "epoch": 0, "step": 3805, "train_loss": 3.555241346359253, "perplexity": 34.99626516972937, "lr": 0.0026291804804649314, "grad_norm": 0.150633, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:40.688944+00:00", "epoch": 0, "step": 3806, "train_loss": 3.6220598220825195, "perplexity": 37.41455583234211, "lr": 0.0026291804804649314, "grad_norm": 0.144039, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:40.993199+00:00", "epoch": 0, "step": 3807, "train_loss": 3.668562889099121, "perplexity": 39.195537032342436, "lr": 0.0026291804804649314, "grad_norm": 0.151411, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:41.297699+00:00", "epoch": 0, "step": 3808, "train_loss": 3.569890260696411, "perplexity": 35.51269579928044, "lr": 0.0026291804804649314, "grad_norm": 0.160073, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:41.601788+00:00", "epoch": 0, "step": 3809, "train_loss": 3.6717584133148193, "perplexity": 39.32098765357367, "lr": 0.0026291804804649314, "grad_norm": 0.185064, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:41.905896+00:00", "epoch": 0, "step": 3810, "train_loss": 3.535802125930786, "perplexity": 34.322534670587586, "lr": 0.0026291804804649314, "grad_norm": 0.159317, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:42.210657+00:00", "epoch": 0, "step": 3811, "train_loss": 3.7096896171569824, "perplexity": 40.84112817404573, "lr": 0.0026291804804649314, "grad_norm": 0.14423, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:42.516345+00:00", "epoch": 0, "step": 3812, "train_loss": 3.6472883224487305, "perplexity": 38.37047648890385, "lr": 0.0026291804804649314, "grad_norm": 0.146898, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:42.821298+00:00", "epoch": 0, "step": 3813, "train_loss": 3.692178726196289, "perplexity": 40.132188826838686, "lr": 0.0026291804804649314, "grad_norm": 0.155431, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:43.125656+00:00", "epoch": 0, "step": 3814, "train_loss": 3.6281614303588867, "perplexity": 37.643542679462044, "lr": 0.0026291804804649314, "grad_norm": 0.174174, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:43.429503+00:00", "epoch": 0, "step": 3815, "train_loss": 3.5979385375976562, "perplexity": 36.52286627039851, "lr": 0.0026291804804649314, "grad_norm": 0.19498, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:43.733763+00:00", "epoch": 0, "step": 3816, "train_loss": 3.6476938724517822, "perplexity": 38.3860407915992, "lr": 0.0026291804804649314, "grad_norm": 0.209411, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:44.040291+00:00", "epoch": 0, "step": 3817, "train_loss": 3.733405113220215, "perplexity": 41.82127215111782, "lr": 0.0026291804804649314, "grad_norm": 0.168732, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:44.345670+00:00", "epoch": 0, "step": 3818, "train_loss": 3.655705451965332, "perplexity": 38.69480881787668, "lr": 0.0026291804804649314, "grad_norm": 0.160103, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:44.649695+00:00", "epoch": 0, "step": 3819, "train_loss": 3.5886776447296143, "perplexity": 36.18619327112576, "lr": 0.0026291804804649314, "grad_norm": 0.170994, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:44.954671+00:00", "epoch": 0, "step": 3820, "train_loss": 3.62791109085083, "perplexity": 37.63412019296568, "lr": 0.0026291804804649314, "grad_norm": 0.158145, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:45.258750+00:00", "epoch": 0, "step": 3821, "train_loss": 3.6247501373291016, "perplexity": 37.51534830330971, "lr": 0.0026291804804649314, "grad_norm": 0.157002, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:45.564191+00:00", "epoch": 0, "step": 3822, "train_loss": 3.4962663650512695, "perplexity": 32.9920414781201, "lr": 0.0026291804804649314, "grad_norm": 0.166118, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:45.869839+00:00", "epoch": 0, "step": 3823, "train_loss": 3.6334574222564697, "perplexity": 37.8434314138491, "lr": 0.0026291804804649314, "grad_norm": 0.161814, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:46.174198+00:00", "epoch": 0, "step": 3824, "train_loss": 3.5713682174682617, "perplexity": 35.565220833827794, "lr": 0.0026291804804649314, "grad_norm": 0.152788, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:46.478920+00:00", "epoch": 0, "step": 3825, "train_loss": 3.6002655029296875, "perplexity": 36.607952672196156, "lr": 0.0026291804804649314, "grad_norm": 0.166645, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:46.782612+00:00", "epoch": 0, "step": 3826, "train_loss": 3.702530860900879, "perplexity": 40.5498005082557, "lr": 0.0026291804804649314, "grad_norm": 0.168027, "tokens_per_sec": 107899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:47.087659+00:00", "epoch": 0, "step": 3827, "train_loss": 3.6320972442626953, "perplexity": 37.791992602131444, "lr": 0.0026291804804649314, "grad_norm": 0.158843, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:47.392405+00:00", "epoch": 0, "step": 3828, "train_loss": 3.5467679500579834, "perplexity": 34.70098074360876, "lr": 0.0026291804804649314, "grad_norm": 0.149736, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:47.697976+00:00", "epoch": 0, "step": 3829, "train_loss": 3.5222010612487793, "perplexity": 33.85887195744346, "lr": 0.0026291804804649314, "grad_norm": 0.145374, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:48.002978+00:00", "epoch": 0, "step": 3830, "train_loss": 3.6500184535980225, "perplexity": 38.47537605160447, "lr": 0.0026291804804649314, "grad_norm": 0.135461, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:48.306108+00:00", "epoch": 0, "step": 3831, "train_loss": 3.6196582317352295, "perplexity": 37.32480920665747, "lr": 0.0026291804804649314, "grad_norm": 0.161051, "tokens_per_sec": 108045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:48.611097+00:00", "epoch": 0, "step": 3832, "train_loss": 3.6996891498565674, "perplexity": 40.43473326367039, "lr": 0.0026291804804649314, "grad_norm": 0.157061, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:48.916376+00:00", "epoch": 0, "step": 3833, "train_loss": 3.616145610809326, "perplexity": 37.193931297593664, "lr": 0.0026291804804649314, "grad_norm": 0.143827, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:49.219883+00:00", "epoch": 0, "step": 3834, "train_loss": 3.6514317989349365, "perplexity": 38.529793491191825, "lr": 0.0026291804804649314, "grad_norm": 0.148221, "tokens_per_sec": 107965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:49.524480+00:00", "epoch": 0, "step": 3835, "train_loss": 3.644670009613037, "perplexity": 38.27014198867696, "lr": 0.0026291804804649314, "grad_norm": 0.13244, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:49.827901+00:00", "epoch": 0, "step": 3836, "train_loss": 3.5411975383758545, "perplexity": 34.5082193736271, "lr": 0.0026291804804649314, "grad_norm": 0.141296, "tokens_per_sec": 107994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:50.132056+00:00", "epoch": 0, "step": 3837, "train_loss": 3.713364601135254, "perplexity": 40.991494793968, "lr": 0.0026291804804649314, "grad_norm": 0.157937, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:50.437281+00:00", "epoch": 0, "step": 3838, "train_loss": 3.691366195678711, "perplexity": 40.099593442845084, "lr": 0.0026291804804649314, "grad_norm": 0.165673, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:50.742426+00:00", "epoch": 0, "step": 3839, "train_loss": 3.621098756790161, "perplexity": 37.37861527468028, "lr": 0.0026291804804649314, "grad_norm": 0.169835, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:51.048428+00:00", "epoch": 0, "step": 3840, "train_loss": 3.637575387954712, "perplexity": 37.99959067488098, "lr": 0.0026291804804649314, "grad_norm": 0.157412, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:51.352777+00:00", "epoch": 0, "step": 3841, "train_loss": 3.6708219051361084, "perplexity": 39.28418046484788, "lr": 0.0026291804804649314, "grad_norm": 0.135201, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:51.656930+00:00", "epoch": 0, "step": 3842, "train_loss": 3.5121607780456543, "perplexity": 33.52062020725065, "lr": 0.0026291804804649314, "grad_norm": 0.162085, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:51.961384+00:00", "epoch": 0, "step": 3843, "train_loss": 3.7009775638580322, "perplexity": 40.48686351561465, "lr": 0.0026291804804649314, "grad_norm": 0.1618, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:52.266875+00:00", "epoch": 0, "step": 3844, "train_loss": 3.6721694469451904, "perplexity": 39.33715322394764, "lr": 0.0026291804804649314, "grad_norm": 0.15011, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:52.572294+00:00", "epoch": 0, "step": 3845, "train_loss": 3.8032171726226807, "perplexity": 44.845227501626034, "lr": 0.0026291804804649314, "grad_norm": 0.15092, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:52.876986+00:00", "epoch": 0, "step": 3846, "train_loss": 3.661609172821045, "perplexity": 38.92392782957224, "lr": 0.0026291804804649314, "grad_norm": 0.144445, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:53.181499+00:00", "epoch": 0, "step": 3847, "train_loss": 3.6082875728607178, "perplexity": 36.90280531157641, "lr": 0.0026291804804649314, "grad_norm": 0.160079, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:53.486144+00:00", "epoch": 0, "step": 3848, "train_loss": 3.6897389888763428, "perplexity": 40.03439617073082, "lr": 0.0026291804804649314, "grad_norm": 0.166368, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:53.790833+00:00", "epoch": 0, "step": 3849, "train_loss": 3.6089911460876465, "perplexity": 36.92877827326101, "lr": 0.0026291804804649314, "grad_norm": 0.147516, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:54.095997+00:00", "epoch": 0, "step": 3850, "train_loss": 3.568021774291992, "perplexity": 35.44640276310071, "lr": 0.0026291804804649314, "grad_norm": 0.161648, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:54.400940+00:00", "epoch": 0, "step": 3851, "train_loss": 3.745225667953491, "perplexity": 42.31855608330696, "lr": 0.0026291804804649314, "grad_norm": 0.125974, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:54.704654+00:00", "epoch": 0, "step": 3852, "train_loss": 3.6899256706237793, "perplexity": 40.0418705594098, "lr": 0.0026291804804649314, "grad_norm": 0.147334, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:55.008958+00:00", "epoch": 0, "step": 3853, "train_loss": 3.6334118843078613, "perplexity": 37.84170814085167, "lr": 0.0026291804804649314, "grad_norm": 0.170328, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:55.313128+00:00", "epoch": 0, "step": 3854, "train_loss": 3.564375877380371, "perplexity": 35.31740413348482, "lr": 0.0026291804804649314, "grad_norm": 0.167627, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:55.617908+00:00", "epoch": 0, "step": 3855, "train_loss": 3.6042118072509766, "perplexity": 36.75270422300011, "lr": 0.0026291804804649314, "grad_norm": 0.145796, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:55.924043+00:00", "epoch": 0, "step": 3856, "train_loss": 3.553337574005127, "perplexity": 34.92970362671781, "lr": 0.0026291804804649314, "grad_norm": 0.163735, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:56.228731+00:00", "epoch": 0, "step": 3857, "train_loss": 3.6661906242370605, "perplexity": 39.10266503912828, "lr": 0.0026291804804649314, "grad_norm": 0.155357, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:56.532773+00:00", "epoch": 0, "step": 3858, "train_loss": 3.789886474609375, "perplexity": 44.25137633603215, "lr": 0.0026291804804649314, "grad_norm": 0.149306, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:56.836362+00:00", "epoch": 0, "step": 3859, "train_loss": 3.5697710514068604, "perplexity": 35.50846260836703, "lr": 0.0026291804804649314, "grad_norm": 0.162007, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:57.141933+00:00", "epoch": 0, "step": 3860, "train_loss": 3.701949119567871, "perplexity": 40.52621787341661, "lr": 0.0026291804804649314, "grad_norm": 0.153423, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:57.447585+00:00", "epoch": 0, "step": 3861, "train_loss": 3.7022719383239746, "perplexity": 40.53930260854541, "lr": 0.0026291804804649314, "grad_norm": 0.154128, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:57.751478+00:00", "epoch": 0, "step": 3862, "train_loss": 3.5744059085845947, "perplexity": 35.673421245717925, "lr": 0.0026291804804649314, "grad_norm": 0.169784, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:58.055268+00:00", "epoch": 0, "step": 3863, "train_loss": 3.623769760131836, "perplexity": 37.47858713413605, "lr": 0.0026291804804649314, "grad_norm": 0.159261, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:58.359796+00:00", "epoch": 0, "step": 3864, "train_loss": 3.711005449295044, "perplexity": 40.8949036150196, "lr": 0.0026291804804649314, "grad_norm": 0.146739, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:58.664843+00:00", "epoch": 0, "step": 3865, "train_loss": 3.578120231628418, "perplexity": 35.80617024000795, "lr": 0.0026291804804649314, "grad_norm": 0.136605, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:58.970744+00:00", "epoch": 0, "step": 3866, "train_loss": 3.6640539169311523, "perplexity": 39.0192032874798, "lr": 0.0026291804804649314, "grad_norm": 0.149655, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:59.275503+00:00", "epoch": 0, "step": 3867, "train_loss": 3.675457000732422, "perplexity": 39.46668904232439, "lr": 0.0026291804804649314, "grad_norm": 0.148468, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:59.579820+00:00", "epoch": 0, "step": 3868, "train_loss": 3.613696575164795, "perplexity": 37.10295348351242, "lr": 0.0026291804804649314, "grad_norm": 0.143811, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:17:59.884409+00:00", "epoch": 0, "step": 3869, "train_loss": 3.6023128032684326, "perplexity": 36.682976918480165, "lr": 0.0026291804804649314, "grad_norm": 0.151226, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:00.189794+00:00", "epoch": 0, "step": 3870, "train_loss": 3.5923824310302734, "perplexity": 36.32050402693087, "lr": 0.0026291804804649314, "grad_norm": 0.154834, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:00.495247+00:00", "epoch": 0, "step": 3871, "train_loss": 3.773250102996826, "perplexity": 43.521283878483935, "lr": 0.0026291804804649314, "grad_norm": 0.142176, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:00.801360+00:00", "epoch": 0, "step": 3872, "train_loss": 3.599848747253418, "perplexity": 36.59269927881333, "lr": 0.0026291804804649314, "grad_norm": 0.147288, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:01.105875+00:00", "epoch": 0, "step": 3873, "train_loss": 3.5539214611053467, "perplexity": 34.95010458542867, "lr": 0.0026291804804649314, "grad_norm": 0.153023, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:01.409707+00:00", "epoch": 0, "step": 3874, "train_loss": 3.5405657291412354, "perplexity": 34.48642364805197, "lr": 0.0026291804804649314, "grad_norm": 0.185374, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:01.714720+00:00", "epoch": 0, "step": 3875, "train_loss": 3.7658421993255615, "perplexity": 43.200073616124186, "lr": 0.0026291804804649314, "grad_norm": 0.18914, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:02.018879+00:00", "epoch": 0, "step": 3876, "train_loss": 3.686602830886841, "perplexity": 39.909038652562174, "lr": 0.0026291804804649314, "grad_norm": 0.180882, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:02.324363+00:00", "epoch": 0, "step": 3877, "train_loss": 3.6976053714752197, "perplexity": 40.35056396617742, "lr": 0.0026291804804649314, "grad_norm": 0.168896, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:02.629329+00:00", "epoch": 0, "step": 3878, "train_loss": 3.5635228157043457, "perplexity": 35.28728905635719, "lr": 0.0026291804804649314, "grad_norm": 0.156709, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:02.935101+00:00", "epoch": 0, "step": 3879, "train_loss": 3.583308219909668, "perplexity": 35.99241493151346, "lr": 0.0026291804804649314, "grad_norm": 0.156448, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:03.239908+00:00", "epoch": 0, "step": 3880, "train_loss": 3.5229268074035645, "perplexity": 33.88345382259004, "lr": 0.0026291804804649314, "grad_norm": 0.188102, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:03.544368+00:00", "epoch": 0, "step": 3881, "train_loss": 3.5842432975769043, "perplexity": 36.02608637515986, "lr": 0.0026291804804649314, "grad_norm": 0.162443, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:03.850382+00:00", "epoch": 0, "step": 3882, "train_loss": 3.712678909301758, "perplexity": 40.963396895095784, "lr": 0.0026291804804649314, "grad_norm": 0.160282, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:04.155446+00:00", "epoch": 0, "step": 3883, "train_loss": 3.646468162536621, "perplexity": 38.33901946393786, "lr": 0.0026291804804649314, "grad_norm": 0.151598, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:04.460360+00:00", "epoch": 0, "step": 3884, "train_loss": 3.7015373706817627, "perplexity": 40.50953468322724, "lr": 0.0026291804804649314, "grad_norm": 0.164124, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:04.765428+00:00", "epoch": 0, "step": 3885, "train_loss": 3.6949362754821777, "perplexity": 40.243008039977155, "lr": 0.0026291804804649314, "grad_norm": 0.156111, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:05.071383+00:00", "epoch": 0, "step": 3886, "train_loss": 3.6389079093933105, "perplexity": 38.0502596953939, "lr": 0.0026291804804649314, "grad_norm": 0.164173, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:05.377396+00:00", "epoch": 0, "step": 3887, "train_loss": 3.6782755851745605, "perplexity": 39.578086155377925, "lr": 0.0026291804804649314, "grad_norm": 0.19932, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:05.682461+00:00", "epoch": 0, "step": 3888, "train_loss": 3.6461374759674072, "perplexity": 38.3263433611483, "lr": 0.0026291804804649314, "grad_norm": 0.166079, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:05.986126+00:00", "epoch": 0, "step": 3889, "train_loss": 3.723541021347046, "perplexity": 41.41077121830757, "lr": 0.0026291804804649314, "grad_norm": 0.158036, "tokens_per_sec": 107907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:06.290741+00:00", "epoch": 0, "step": 3890, "train_loss": 3.593855619430542, "perplexity": 36.374050404422114, "lr": 0.0026291804804649314, "grad_norm": 0.147865, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:06.595753+00:00", "epoch": 0, "step": 3891, "train_loss": 3.6242594718933105, "perplexity": 37.49694533380535, "lr": 0.0026291804804649314, "grad_norm": 0.161527, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:06.900472+00:00", "epoch": 0, "step": 3892, "train_loss": 3.7017483711242676, "perplexity": 40.518083114800774, "lr": 0.0026291804804649314, "grad_norm": 0.146272, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:07.205903+00:00", "epoch": 0, "step": 3893, "train_loss": 3.6512062549591064, "perplexity": 38.52110430831317, "lr": 0.0026291804804649314, "grad_norm": 0.160787, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:07.510334+00:00", "epoch": 0, "step": 3894, "train_loss": 3.652615785598755, "perplexity": 38.57543926950919, "lr": 0.0026291804804649314, "grad_norm": 0.16138, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:07.814504+00:00", "epoch": 0, "step": 3895, "train_loss": 3.6547207832336426, "perplexity": 38.65672600211097, "lr": 0.0026291804804649314, "grad_norm": 0.145316, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:08.118608+00:00", "epoch": 0, "step": 3896, "train_loss": 3.5831940174102783, "perplexity": 35.9883047424706, "lr": 0.0026291804804649314, "grad_norm": 0.149502, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:08.423298+00:00", "epoch": 0, "step": 3897, "train_loss": 3.610466718673706, "perplexity": 36.98330958868044, "lr": 0.0026291804804649314, "grad_norm": 0.13393, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:08.728205+00:00", "epoch": 0, "step": 3898, "train_loss": 3.6949257850646973, "perplexity": 40.24258587623648, "lr": 0.0026291804804649314, "grad_norm": 0.152921, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:09.034206+00:00", "epoch": 0, "step": 3899, "train_loss": 3.5913443565368652, "perplexity": 36.28282020080614, "lr": 0.0026291804804649314, "grad_norm": 0.139844, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:09.339543+00:00", "epoch": 0, "step": 3900, "train_loss": 3.642960786819458, "perplexity": 38.204785659851105, "lr": 0.0026291804804649314, "grad_norm": 0.139068, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:09.643589+00:00", "epoch": 0, "step": 3901, "train_loss": 3.623249053955078, "perplexity": 37.45907688231429, "lr": 0.0026291804804649314, "grad_norm": 0.1459, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:09.947443+00:00", "epoch": 0, "step": 3902, "train_loss": 3.6113760471343994, "perplexity": 37.016954859642915, "lr": 0.0026291804804649314, "grad_norm": 0.141813, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:10.252441+00:00", "epoch": 0, "step": 3903, "train_loss": 3.642402410507202, "perplexity": 38.183458967237925, "lr": 0.0026291804804649314, "grad_norm": 0.156277, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:10.557447+00:00", "epoch": 0, "step": 3904, "train_loss": 3.681274890899658, "perplexity": 39.69697113284648, "lr": 0.0026291804804649314, "grad_norm": 0.160432, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:10.862740+00:00", "epoch": 0, "step": 3905, "train_loss": 3.6252458095550537, "perplexity": 37.53394822886301, "lr": 0.0026291804804649314, "grad_norm": 0.155314, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:11.167094+00:00", "epoch": 0, "step": 3906, "train_loss": 3.6807310581207275, "perplexity": 39.67538848792699, "lr": 0.0026291804804649314, "grad_norm": 0.137159, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:11.470576+00:00", "epoch": 0, "step": 3907, "train_loss": 3.6592063903808594, "perplexity": 38.830514370332544, "lr": 0.0026291804804649314, "grad_norm": 0.156894, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:11.775862+00:00", "epoch": 0, "step": 3908, "train_loss": 3.6057374477386475, "perplexity": 36.80881843076357, "lr": 0.0026291804804649314, "grad_norm": 0.161772, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:12.080437+00:00", "epoch": 0, "step": 3909, "train_loss": 3.5715513229370117, "perplexity": 35.57173361650464, "lr": 0.0026291804804649314, "grad_norm": 0.177547, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:12.384880+00:00", "epoch": 0, "step": 3910, "train_loss": 3.644110679626465, "perplexity": 38.248742335963954, "lr": 0.0026291804804649314, "grad_norm": 0.172565, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:12.689367+00:00", "epoch": 0, "step": 3911, "train_loss": 3.6664953231811523, "perplexity": 39.11458139523525, "lr": 0.0026291804804649314, "grad_norm": 0.176454, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:12.993663+00:00", "epoch": 0, "step": 3912, "train_loss": 3.713512659072876, "perplexity": 40.99756435945983, "lr": 0.0026291804804649314, "grad_norm": 0.194832, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:13.298524+00:00", "epoch": 0, "step": 3913, "train_loss": 3.6546385288238525, "perplexity": 38.65354644669727, "lr": 0.0026291804804649314, "grad_norm": 0.183851, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:13.603307+00:00", "epoch": 0, "step": 3914, "train_loss": 3.6352007389068604, "perplexity": 37.909462037360385, "lr": 0.0026291804804649314, "grad_norm": 0.15625, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:13.907986+00:00", "epoch": 0, "step": 3915, "train_loss": 3.653390407562256, "perplexity": 38.605332228395554, "lr": 0.0026291804804649314, "grad_norm": 0.15422, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:14.213560+00:00", "epoch": 0, "step": 3916, "train_loss": 3.7694342136383057, "perplexity": 43.35552792877632, "lr": 0.0026291804804649314, "grad_norm": 0.163368, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:14.518674+00:00", "epoch": 0, "step": 3917, "train_loss": 3.60866379737854, "perplexity": 36.91669166374032, "lr": 0.0026291804804649314, "grad_norm": 0.149591, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:14.823597+00:00", "epoch": 0, "step": 3918, "train_loss": 3.694689989089966, "perplexity": 40.23309795512485, "lr": 0.0026291804804649314, "grad_norm": 0.159751, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:15.129241+00:00", "epoch": 0, "step": 3919, "train_loss": 3.62225604057312, "perplexity": 37.42189798032306, "lr": 0.0026291804804649314, "grad_norm": 0.175647, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:15.434420+00:00", "epoch": 0, "step": 3920, "train_loss": 3.6517045497894287, "perplexity": 38.54030395859428, "lr": 0.0026291804804649314, "grad_norm": 0.163591, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:15.737545+00:00", "epoch": 0, "step": 3921, "train_loss": 3.7577013969421387, "perplexity": 42.849817969230934, "lr": 0.0026291804804649314, "grad_norm": 0.186196, "tokens_per_sec": 108043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:16.041484+00:00", "epoch": 0, "step": 3922, "train_loss": 3.6624863147735596, "perplexity": 38.95808461761413, "lr": 0.0026291804804649314, "grad_norm": 0.196196, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:16.345745+00:00", "epoch": 0, "step": 3923, "train_loss": 3.686866283416748, "perplexity": 39.91955417487101, "lr": 0.0026291804804649314, "grad_norm": 0.136816, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:16.650827+00:00", "epoch": 0, "step": 3924, "train_loss": 3.7111852169036865, "perplexity": 40.902255854875634, "lr": 0.0026291804804649314, "grad_norm": 0.160886, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:16.956323+00:00", "epoch": 0, "step": 3925, "train_loss": 3.653937339782715, "perplexity": 38.62645250362582, "lr": 0.0026291804804649314, "grad_norm": 0.17608, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:17.261024+00:00", "epoch": 0, "step": 3926, "train_loss": 3.533750534057617, "perplexity": 34.25219102029469, "lr": 0.0026291804804649314, "grad_norm": 0.151387, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:17.565904+00:00", "epoch": 0, "step": 3927, "train_loss": 3.77962064743042, "perplexity": 43.79942315990174, "lr": 0.0026291804804649314, "grad_norm": 0.138767, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:17.870667+00:00", "epoch": 0, "step": 3928, "train_loss": 3.64925479888916, "perplexity": 38.44600536546614, "lr": 0.0026291804804649314, "grad_norm": 0.161661, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:18.175092+00:00", "epoch": 0, "step": 3929, "train_loss": 3.6884963512420654, "perplexity": 39.98467882010656, "lr": 0.0026291804804649314, "grad_norm": 0.157701, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:18.480496+00:00", "epoch": 0, "step": 3930, "train_loss": 3.7146947383880615, "perplexity": 41.04605538673587, "lr": 0.0026291804804649314, "grad_norm": 0.153732, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:18.786068+00:00", "epoch": 0, "step": 3931, "train_loss": 3.767801523208618, "perplexity": 43.28479952775628, "lr": 0.0026291804804649314, "grad_norm": 0.159216, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:19.090882+00:00", "epoch": 0, "step": 3932, "train_loss": 3.6256091594696045, "perplexity": 37.54758866372007, "lr": 0.0026291804804649314, "grad_norm": 0.165497, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:19.394653+00:00", "epoch": 0, "step": 3933, "train_loss": 3.6461408138275146, "perplexity": 38.326471289334364, "lr": 0.0026291804804649314, "grad_norm": 0.146097, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:19.699107+00:00", "epoch": 0, "step": 3934, "train_loss": 3.6652042865753174, "perplexity": 39.0641156224217, "lr": 0.0026291804804649314, "grad_norm": 0.179112, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:20.004104+00:00", "epoch": 0, "step": 3935, "train_loss": 3.805384874343872, "perplexity": 44.94254401695045, "lr": 0.0026291804804649314, "grad_norm": 0.173326, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:20.309780+00:00", "epoch": 0, "step": 3936, "train_loss": 3.6440298557281494, "perplexity": 38.24565104842934, "lr": 0.0026291804804649314, "grad_norm": 0.166784, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:20.615105+00:00", "epoch": 0, "step": 3937, "train_loss": 3.649284601211548, "perplexity": 38.447151162786184, "lr": 0.0026291804804649314, "grad_norm": 0.147503, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:20.919523+00:00", "epoch": 0, "step": 3938, "train_loss": 3.612030506134033, "perplexity": 37.04118886810733, "lr": 0.0026291804804649314, "grad_norm": 0.160022, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:21.223951+00:00", "epoch": 0, "step": 3939, "train_loss": 3.690910816192627, "perplexity": 40.08133706769576, "lr": 0.0026291804804649314, "grad_norm": 0.149868, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:21.529770+00:00", "epoch": 0, "step": 3940, "train_loss": 3.7639899253845215, "perplexity": 43.12012930775047, "lr": 0.0026291804804649314, "grad_norm": 0.155437, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:21.835531+00:00", "epoch": 0, "step": 3941, "train_loss": 3.5569679737091064, "perplexity": 35.05674287451883, "lr": 0.0026291804804649314, "grad_norm": 0.176836, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:22.142191+00:00", "epoch": 0, "step": 3942, "train_loss": 3.672898769378662, "perplexity": 39.365853156743036, "lr": 0.0026291804804649314, "grad_norm": 0.150357, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:22.446635+00:00", "epoch": 0, "step": 3943, "train_loss": 3.645050525665283, "perplexity": 38.28470716299096, "lr": 0.0026291804804649314, "grad_norm": 0.174719, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:22.751146+00:00", "epoch": 0, "step": 3944, "train_loss": 3.6199610233306885, "perplexity": 37.33611255638065, "lr": 0.0026291804804649314, "grad_norm": 0.172265, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:23.057166+00:00", "epoch": 0, "step": 3945, "train_loss": 3.803844690322876, "perplexity": 44.87337750704218, "lr": 0.0026291804804649314, "grad_norm": 0.163048, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:23.362838+00:00", "epoch": 0, "step": 3946, "train_loss": 3.7332818508148193, "perplexity": 41.81611747821099, "lr": 0.0026291804804649314, "grad_norm": 0.176706, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:23.667741+00:00", "epoch": 0, "step": 3947, "train_loss": 3.7359697818756104, "perplexity": 41.92866751482777, "lr": 0.0026291804804649314, "grad_norm": 0.154769, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:23.972231+00:00", "epoch": 0, "step": 3948, "train_loss": 3.6793463230133057, "perplexity": 39.62048660564191, "lr": 0.0026291804804649314, "grad_norm": 0.155481, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:24.277485+00:00", "epoch": 0, "step": 3949, "train_loss": 3.654588222503662, "perplexity": 38.65160197792317, "lr": 0.0026291804804649314, "grad_norm": 0.155018, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:24.583796+00:00", "epoch": 0, "step": 3950, "train_loss": 3.542677402496338, "perplexity": 34.55932465444893, "lr": 0.0026291804804649314, "grad_norm": 0.144676, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:24.890572+00:00", "epoch": 0, "step": 3951, "train_loss": 3.7390549182891846, "perplexity": 42.058222919038236, "lr": 0.0026291804804649314, "grad_norm": 0.137458, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:25.196397+00:00", "epoch": 0, "step": 3952, "train_loss": 3.708247184753418, "perplexity": 40.782260074203535, "lr": 0.0026291804804649314, "grad_norm": 0.165225, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:25.500037+00:00", "epoch": 0, "step": 3953, "train_loss": 3.647554636001587, "perplexity": 38.380696427616094, "lr": 0.0026291804804649314, "grad_norm": 0.151585, "tokens_per_sec": 107918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:25.804510+00:00", "epoch": 0, "step": 3954, "train_loss": 3.629331111907959, "perplexity": 37.687599397918234, "lr": 0.0026291804804649314, "grad_norm": 0.13854, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:26.109684+00:00", "epoch": 0, "step": 3955, "train_loss": 3.6172845363616943, "perplexity": 37.23631654858254, "lr": 0.0026291804804649314, "grad_norm": 0.140964, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:26.415342+00:00", "epoch": 0, "step": 3956, "train_loss": 3.546505928039551, "perplexity": 34.69188951369595, "lr": 0.0026291804804649314, "grad_norm": 0.143991, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:26.721573+00:00", "epoch": 0, "step": 3957, "train_loss": 3.690408945083618, "perplexity": 40.06122644950228, "lr": 0.0026291804804649314, "grad_norm": 0.159913, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:27.027256+00:00", "epoch": 0, "step": 3958, "train_loss": 3.6248507499694824, "perplexity": 37.5191230114458, "lr": 0.0026291804804649314, "grad_norm": 0.148249, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:27.331790+00:00", "epoch": 0, "step": 3959, "train_loss": 3.5962471961975098, "perplexity": 36.46114584449246, "lr": 0.0026291804804649314, "grad_norm": 0.139029, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:27.636749+00:00", "epoch": 0, "step": 3960, "train_loss": 3.69812273979187, "perplexity": 40.37144547078116, "lr": 0.0026291804804649314, "grad_norm": 0.157028, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:27.941611+00:00", "epoch": 0, "step": 3961, "train_loss": 3.6111252307891846, "perplexity": 37.00767156656352, "lr": 0.0026291804804649314, "grad_norm": 0.132179, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:28.247833+00:00", "epoch": 0, "step": 3962, "train_loss": 3.633989095687866, "perplexity": 37.863557110559746, "lr": 0.0026291804804649314, "grad_norm": 0.176106, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:28.552579+00:00", "epoch": 0, "step": 3963, "train_loss": 3.6189119815826416, "perplexity": 37.29696595240019, "lr": 0.0026291804804649314, "grad_norm": 0.193227, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:28.857020+00:00", "epoch": 0, "step": 3964, "train_loss": 3.7163844108581543, "perplexity": 41.11546840264529, "lr": 0.0026291804804649314, "grad_norm": 0.173496, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:29.162459+00:00", "epoch": 0, "step": 3965, "train_loss": 3.5651566982269287, "perplexity": 35.34499146785981, "lr": 0.0026291804804649314, "grad_norm": 0.147973, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:29.468535+00:00", "epoch": 0, "step": 3966, "train_loss": 3.6243982315063477, "perplexity": 37.502148756434046, "lr": 0.0026291804804649314, "grad_norm": 0.174406, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:29.774464+00:00", "epoch": 0, "step": 3967, "train_loss": 3.5676445960998535, "perplexity": 35.433035674034414, "lr": 0.0026291804804649314, "grad_norm": 0.179457, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:30.078289+00:00", "epoch": 0, "step": 3968, "train_loss": 3.5734715461730957, "perplexity": 35.64010490900657, "lr": 0.0026291804804649314, "grad_norm": 0.183825, "tokens_per_sec": 107851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:30.383617+00:00", "epoch": 0, "step": 3969, "train_loss": 3.639010429382324, "perplexity": 38.054160807567406, "lr": 0.0026291804804649314, "grad_norm": 0.177083, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:30.688951+00:00", "epoch": 0, "step": 3970, "train_loss": 3.717954635620117, "perplexity": 41.18007964304012, "lr": 0.0026291804804649314, "grad_norm": 0.191813, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:30.994450+00:00", "epoch": 0, "step": 3971, "train_loss": 3.608893871307373, "perplexity": 36.92518620918016, "lr": 0.0026291804804649314, "grad_norm": 0.15384, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:31.300108+00:00", "epoch": 0, "step": 3972, "train_loss": 3.5745131969451904, "perplexity": 35.67724879392223, "lr": 0.0026291804804649314, "grad_norm": 0.151006, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:31.604035+00:00", "epoch": 0, "step": 3973, "train_loss": 3.5714378356933594, "perplexity": 35.56769690756638, "lr": 0.0026291804804649314, "grad_norm": 0.153518, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:31.908420+00:00", "epoch": 0, "step": 3974, "train_loss": 3.661343574523926, "perplexity": 38.9135910733967, "lr": 0.0026291804804649314, "grad_norm": 0.139568, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:32.214954+00:00", "epoch": 0, "step": 3975, "train_loss": 3.484234571456909, "perplexity": 32.59746652699459, "lr": 0.0026291804804649314, "grad_norm": 0.164217, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:32.519977+00:00", "epoch": 0, "step": 3976, "train_loss": 3.76509428024292, "perplexity": 43.16777553637584, "lr": 0.0026291804804649314, "grad_norm": 0.16025, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:32.825012+00:00", "epoch": 0, "step": 3977, "train_loss": 3.621204137802124, "perplexity": 37.3825544785387, "lr": 0.0026291804804649314, "grad_norm": 0.152308, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:33.129790+00:00", "epoch": 0, "step": 3978, "train_loss": 3.5645737648010254, "perplexity": 35.32439369504334, "lr": 0.0026291804804649314, "grad_norm": 0.152761, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:33.434493+00:00", "epoch": 0, "step": 3979, "train_loss": 3.6552207469940186, "perplexity": 38.676057796402404, "lr": 0.0026291804804649314, "grad_norm": 0.143179, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:33.739472+00:00", "epoch": 0, "step": 3980, "train_loss": 3.6656692028045654, "perplexity": 39.08228138620778, "lr": 0.0026291804804649314, "grad_norm": 0.150598, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:34.044042+00:00", "epoch": 0, "step": 3981, "train_loss": 3.6766140460968018, "perplexity": 39.512380220213586, "lr": 0.0026291804804649314, "grad_norm": 0.149487, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:34.348747+00:00", "epoch": 0, "step": 3982, "train_loss": 3.6210732460021973, "perplexity": 37.37766172891453, "lr": 0.0026291804804649314, "grad_norm": 0.170271, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:34.652771+00:00", "epoch": 0, "step": 3983, "train_loss": 3.6683461666107178, "perplexity": 39.187043398436465, "lr": 0.0026291804804649314, "grad_norm": 0.148698, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:34.957854+00:00", "epoch": 0, "step": 3984, "train_loss": 3.69435977935791, "perplexity": 40.219814787865566, "lr": 0.0026291804804649314, "grad_norm": 0.154601, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:35.262282+00:00", "epoch": 0, "step": 3985, "train_loss": 3.622447967529297, "perplexity": 37.42908094057658, "lr": 0.0026291804804649314, "grad_norm": 0.153426, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:35.567806+00:00", "epoch": 0, "step": 3986, "train_loss": 3.5580978393554688, "perplexity": 35.09637466905153, "lr": 0.0026291804804649314, "grad_norm": 0.149079, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:35.873138+00:00", "epoch": 0, "step": 3987, "train_loss": 3.616563558578491, "perplexity": 37.20947966718283, "lr": 0.0026291804804649314, "grad_norm": 0.157925, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:36.177398+00:00", "epoch": 0, "step": 3988, "train_loss": 3.644559144973755, "perplexity": 38.26589941836997, "lr": 0.0026291804804649314, "grad_norm": 0.16143, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:36.481778+00:00", "epoch": 0, "step": 3989, "train_loss": 3.6139307022094727, "perplexity": 37.111641305347675, "lr": 0.0026291804804649314, "grad_norm": 0.176899, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:36.786919+00:00", "epoch": 0, "step": 3990, "train_loss": 3.5888631343841553, "perplexity": 36.1929060581718, "lr": 0.0026291804804649314, "grad_norm": 0.167089, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:37.092133+00:00", "epoch": 0, "step": 3991, "train_loss": 3.681227684020996, "perplexity": 39.6950972069784, "lr": 0.0026291804804649314, "grad_norm": 0.170064, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:37.397441+00:00", "epoch": 0, "step": 3992, "train_loss": 3.678622245788574, "perplexity": 39.59180869742104, "lr": 0.0026291804804649314, "grad_norm": 0.153144, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:37.702354+00:00", "epoch": 0, "step": 3993, "train_loss": 3.644693374633789, "perplexity": 38.271036181785085, "lr": 0.0026291804804649314, "grad_norm": 0.193233, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:38.007694+00:00", "epoch": 0, "step": 3994, "train_loss": 3.697493553161621, "perplexity": 40.34605228641088, "lr": 0.0026291804804649314, "grad_norm": 0.171689, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:38.313396+00:00", "epoch": 0, "step": 3995, "train_loss": 3.6371097564697266, "perplexity": 37.98190098780344, "lr": 0.0026291804804649314, "grad_norm": 0.152309, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:38.618326+00:00", "epoch": 0, "step": 3996, "train_loss": 3.733478307723999, "perplexity": 41.82433335041066, "lr": 0.0026291804804649314, "grad_norm": 0.173564, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:38.924000+00:00", "epoch": 0, "step": 3997, "train_loss": 3.6345393657684326, "perplexity": 37.88439802672193, "lr": 0.0026291804804649314, "grad_norm": 0.155212, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:39.228354+00:00", "epoch": 0, "step": 3998, "train_loss": 3.675504684448242, "perplexity": 39.4685710055782, "lr": 0.0026291804804649314, "grad_norm": 0.175029, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:39.532182+00:00", "epoch": 0, "step": 3999, "train_loss": 3.608707904815674, "perplexity": 36.91832000030768, "lr": 0.0026291804804649314, "grad_norm": 0.14858, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:39.837284+00:00", "epoch": 0, "step": 4000, "train_loss": 3.677903175354004, "perplexity": 39.56334963159823, "lr": 0.0026291804804649314, "grad_norm": 0.163679, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:18:42.910490+00:00", "step": 4000, "epoch": 0, "val_loss": 3.5909712076187135, "val_ppl": 36.26928383139858, "eval_train_loss": 3.677903175354004, "eval_train_ppl": 39.56334963159823} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:18:43.782231+00:00", "step": 4000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5910_epoch_0000_step_0004000.pt", "val_loss": 3.5909712076187135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:44.860409+00:00", "epoch": 0, "step": 4001, "train_loss": 3.673165798187256, "perplexity": 39.37636637721466, "lr": 0.0026291804804649314, "grad_norm": 0.149764, "tokens_per_sec": 6523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:45.165497+00:00", "epoch": 0, "step": 4002, "train_loss": 3.6003165245056152, "perplexity": 36.60982051528273, "lr": 0.0026291804804649314, "grad_norm": 0.150632, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:45.470113+00:00", "epoch": 0, "step": 4003, "train_loss": 3.6289167404174805, "perplexity": 37.67198596628704, "lr": 0.0026291804804649314, "grad_norm": 0.158936, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:45.773540+00:00", "epoch": 0, "step": 4004, "train_loss": 3.6704938411712646, "perplexity": 39.271294854616706, "lr": 0.0026291804804649314, "grad_norm": 0.16736, "tokens_per_sec": 107993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:46.077046+00:00", "epoch": 0, "step": 4005, "train_loss": 3.7450907230377197, "perplexity": 42.3128457946166, "lr": 0.0026291804804649314, "grad_norm": 0.184304, "tokens_per_sec": 107964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:46.381506+00:00", "epoch": 0, "step": 4006, "train_loss": 3.6169064044952393, "perplexity": 37.22223897246392, "lr": 0.0026291804804649314, "grad_norm": 0.194724, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:46.685820+00:00", "epoch": 0, "step": 4007, "train_loss": 3.6395998001098633, "perplexity": 38.076595426512895, "lr": 0.0026291804804649314, "grad_norm": 0.171893, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:46.990597+00:00", "epoch": 0, "step": 4008, "train_loss": 3.5897507667541504, "perplexity": 36.22504631540933, "lr": 0.0026291804804649314, "grad_norm": 0.161539, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:47.295781+00:00", "epoch": 0, "step": 4009, "train_loss": 3.696904182434082, "perplexity": 40.322280510104164, "lr": 0.0026291804804649314, "grad_norm": 0.155437, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:47.600054+00:00", "epoch": 0, "step": 4010, "train_loss": 3.6710591316223145, "perplexity": 39.293500818416696, "lr": 0.0026291804804649314, "grad_norm": 0.195602, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:47.903933+00:00", "epoch": 0, "step": 4011, "train_loss": 3.5826826095581055, "perplexity": 35.96990474619123, "lr": 0.0026291804804649314, "grad_norm": 0.167535, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:48.209211+00:00", "epoch": 0, "step": 4012, "train_loss": 3.5584075450897217, "perplexity": 35.10724590089255, "lr": 0.0026291804804649314, "grad_norm": 0.158707, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:48.513710+00:00", "epoch": 0, "step": 4013, "train_loss": 3.769756317138672, "perplexity": 43.369495145406326, "lr": 0.0026291804804649314, "grad_norm": 0.168315, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:48.817825+00:00", "epoch": 0, "step": 4014, "train_loss": 3.690281867980957, "perplexity": 40.05613590836849, "lr": 0.0026291804804649314, "grad_norm": 0.139743, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:49.189902+00:00", "epoch": 0, "step": 4015, "train_loss": 3.621638298034668, "perplexity": 37.39878802080843, "lr": 0.0026291804804649314, "grad_norm": 0.153652, "tokens_per_sec": 88067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:49.493933+00:00", "epoch": 0, "step": 4016, "train_loss": 3.5803709030151367, "perplexity": 35.88684891937607, "lr": 0.0026291804804649314, "grad_norm": 0.146723, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:49.797895+00:00", "epoch": 0, "step": 4017, "train_loss": 3.62477707862854, "perplexity": 37.51635902915696, "lr": 0.0026291804804649314, "grad_norm": 0.139189, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:50.102499+00:00", "epoch": 0, "step": 4018, "train_loss": 3.5905661582946777, "perplexity": 36.254595957356955, "lr": 0.0026291804804649314, "grad_norm": 0.152917, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:50.406377+00:00", "epoch": 0, "step": 4019, "train_loss": 3.6088473796844482, "perplexity": 36.92346953725224, "lr": 0.0026291804804649314, "grad_norm": 0.151694, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:50.712254+00:00", "epoch": 0, "step": 4020, "train_loss": 3.5827059745788574, "perplexity": 35.97074519358057, "lr": 0.0026291804804649314, "grad_norm": 0.153885, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:51.016747+00:00", "epoch": 0, "step": 4021, "train_loss": 3.5983753204345703, "perplexity": 36.53882231595007, "lr": 0.0026291804804649314, "grad_norm": 0.14882, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:51.322023+00:00", "epoch": 0, "step": 4022, "train_loss": 3.521009683609009, "perplexity": 33.81855727426005, "lr": 0.0026291804804649314, "grad_norm": 0.129796, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:51.625577+00:00", "epoch": 0, "step": 4023, "train_loss": 3.708571195602417, "perplexity": 40.79547610986812, "lr": 0.0026291804804649314, "grad_norm": 0.150281, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:51.928977+00:00", "epoch": 0, "step": 4024, "train_loss": 3.63230562210083, "perplexity": 37.799868436394846, "lr": 0.0026291804804649314, "grad_norm": 0.157521, "tokens_per_sec": 108003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:52.234537+00:00", "epoch": 0, "step": 4025, "train_loss": 3.621175765991211, "perplexity": 37.38149388281717, "lr": 0.0026291804804649314, "grad_norm": 0.17239, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:52.540081+00:00", "epoch": 0, "step": 4026, "train_loss": 3.6377975940704346, "perplexity": 38.008035354518874, "lr": 0.0026291804804649314, "grad_norm": 0.195182, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:52.845061+00:00", "epoch": 0, "step": 4027, "train_loss": 3.7187769412994385, "perplexity": 41.21395618293217, "lr": 0.0026291804804649314, "grad_norm": 0.185394, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:53.149686+00:00", "epoch": 0, "step": 4028, "train_loss": 3.7120938301086426, "perplexity": 40.93943707377309, "lr": 0.0026291804804649314, "grad_norm": 0.170112, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:53.454611+00:00", "epoch": 0, "step": 4029, "train_loss": 3.6092076301574707, "perplexity": 36.936773630877795, "lr": 0.0026291804804649314, "grad_norm": 0.158423, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:53.760748+00:00", "epoch": 0, "step": 4030, "train_loss": 3.6869912147521973, "perplexity": 39.924541689626565, "lr": 0.0026291804804649314, "grad_norm": 0.163222, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:54.067478+00:00", "epoch": 0, "step": 4031, "train_loss": 3.5418570041656494, "perplexity": 34.53098386914874, "lr": 0.0026291804804649314, "grad_norm": 0.149507, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:54.371406+00:00", "epoch": 0, "step": 4032, "train_loss": 3.5694172382354736, "perplexity": 35.495901468879836, "lr": 0.0026291804804649314, "grad_norm": 0.152348, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:54.674680+00:00", "epoch": 0, "step": 4033, "train_loss": 3.6674444675445557, "perplexity": 39.15172440394366, "lr": 0.0026291804804649314, "grad_norm": 0.161814, "tokens_per_sec": 108043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:54.978716+00:00", "epoch": 0, "step": 4034, "train_loss": 3.571770191192627, "perplexity": 35.5795199918546, "lr": 0.0026291804804649314, "grad_norm": 0.158922, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:55.283609+00:00", "epoch": 0, "step": 4035, "train_loss": 3.611582040786743, "perplexity": 37.02458090280437, "lr": 0.0026291804804649314, "grad_norm": 0.159276, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:55.589272+00:00", "epoch": 0, "step": 4036, "train_loss": 3.6918821334838867, "perplexity": 40.12028767708427, "lr": 0.0026291804804649314, "grad_norm": 0.160898, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:55.893800+00:00", "epoch": 0, "step": 4037, "train_loss": 3.5872817039489746, "perplexity": 36.13571472896735, "lr": 0.0026291804804649314, "grad_norm": 0.155716, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:56.198515+00:00", "epoch": 0, "step": 4038, "train_loss": 3.6283838748931885, "perplexity": 37.65191721118272, "lr": 0.0026291804804649314, "grad_norm": 0.15123, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:56.503004+00:00", "epoch": 0, "step": 4039, "train_loss": 3.6302437782287598, "perplexity": 37.72201130150462, "lr": 0.0026291804804649314, "grad_norm": 0.154267, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:56.808022+00:00", "epoch": 0, "step": 4040, "train_loss": 3.6525139808654785, "perplexity": 38.57151230709841, "lr": 0.0026291804804649314, "grad_norm": 0.164977, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:57.113387+00:00", "epoch": 0, "step": 4041, "train_loss": 3.6226139068603516, "perplexity": 37.43529241257936, "lr": 0.0026291804804649314, "grad_norm": 0.160456, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:57.417657+00:00", "epoch": 0, "step": 4042, "train_loss": 3.5625510215759277, "perplexity": 35.25301373302084, "lr": 0.0026291804804649314, "grad_norm": 0.173113, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:57.722111+00:00", "epoch": 0, "step": 4043, "train_loss": 3.626525640487671, "perplexity": 37.58201608963597, "lr": 0.0026291804804649314, "grad_norm": 0.170202, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:58.026088+00:00", "epoch": 0, "step": 4044, "train_loss": 3.6505846977233887, "perplexity": 38.49716867665562, "lr": 0.0026291804804649314, "grad_norm": 0.149681, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:58.330730+00:00", "epoch": 0, "step": 4045, "train_loss": 3.597541570663452, "perplexity": 36.50837077745285, "lr": 0.0026291804804649314, "grad_norm": 0.148728, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:58.636094+00:00", "epoch": 0, "step": 4046, "train_loss": 3.6041998863220215, "perplexity": 36.752266099235584, "lr": 0.0026291804804649314, "grad_norm": 0.161611, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:58.941167+00:00", "epoch": 0, "step": 4047, "train_loss": 3.6159956455230713, "perplexity": 37.18835391725685, "lr": 0.0026291804804649314, "grad_norm": 0.190049, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:59.246988+00:00", "epoch": 0, "step": 4048, "train_loss": 3.6127355098724365, "perplexity": 37.06731225219396, "lr": 0.0026291804804649314, "grad_norm": 0.212745, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:59.551736+00:00", "epoch": 0, "step": 4049, "train_loss": 3.5569984912872314, "perplexity": 35.057812737733045, "lr": 0.0026291804804649314, "grad_norm": 0.155467, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:18:59.857426+00:00", "epoch": 0, "step": 4050, "train_loss": 3.5033087730407715, "perplexity": 33.22520494709969, "lr": 0.0026291804804649314, "grad_norm": 0.179372, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:00.162529+00:00", "epoch": 0, "step": 4051, "train_loss": 3.6513803005218506, "perplexity": 38.527809319061795, "lr": 0.0026291804804649314, "grad_norm": 0.184211, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:00.467311+00:00", "epoch": 0, "step": 4052, "train_loss": 3.584486722946167, "perplexity": 36.034857106004715, "lr": 0.0026291804804649314, "grad_norm": 0.144268, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:00.772549+00:00", "epoch": 0, "step": 4053, "train_loss": 3.5551767349243164, "perplexity": 34.99400408386612, "lr": 0.0026291804804649314, "grad_norm": 0.175567, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:01.077113+00:00", "epoch": 0, "step": 4054, "train_loss": 3.5482847690582275, "perplexity": 34.75365578968383, "lr": 0.0026291804804649314, "grad_norm": 0.164678, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:01.382232+00:00", "epoch": 0, "step": 4055, "train_loss": 3.5646557807922363, "perplexity": 35.327290979016354, "lr": 0.0026291804804649314, "grad_norm": 0.151739, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:01.686806+00:00", "epoch": 0, "step": 4056, "train_loss": 3.7130253314971924, "perplexity": 40.97758998323908, "lr": 0.0026291804804649314, "grad_norm": 0.145958, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:01.993225+00:00", "epoch": 0, "step": 4057, "train_loss": 3.606842041015625, "perplexity": 36.849499668109964, "lr": 0.0026291804804649314, "grad_norm": 0.16635, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:02.298261+00:00", "epoch": 0, "step": 4058, "train_loss": 3.6157338619232178, "perplexity": 37.17861989025567, "lr": 0.0026291804804649314, "grad_norm": 0.151263, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:02.602299+00:00", "epoch": 0, "step": 4059, "train_loss": 3.6133298873901367, "perplexity": 37.0893507781916, "lr": 0.0026291804804649314, "grad_norm": 0.151525, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:02.906486+00:00", "epoch": 0, "step": 4060, "train_loss": 3.6702277660369873, "perplexity": 39.26084712956658, "lr": 0.0026291804804649314, "grad_norm": 0.148287, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:03.210583+00:00", "epoch": 0, "step": 4061, "train_loss": 3.6831154823303223, "perplexity": 39.77010432125028, "lr": 0.0026291804804649314, "grad_norm": 0.153726, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:03.515163+00:00", "epoch": 0, "step": 4062, "train_loss": 3.639155864715576, "perplexity": 38.059695629595666, "lr": 0.0026291804804649314, "grad_norm": 0.150782, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:03.822236+00:00", "epoch": 0, "step": 4063, "train_loss": 3.59466552734375, "perplexity": 36.40352196869565, "lr": 0.0026291804804649314, "grad_norm": 0.148101, "tokens_per_sec": 106711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:04.127464+00:00", "epoch": 0, "step": 4064, "train_loss": 3.7312679290771484, "perplexity": 41.7319878339313, "lr": 0.0026291804804649314, "grad_norm": 0.140375, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:04.431557+00:00", "epoch": 0, "step": 4065, "train_loss": 3.6166982650756836, "perplexity": 37.21449236346495, "lr": 0.0026291804804649314, "grad_norm": 0.160629, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:04.734832+00:00", "epoch": 0, "step": 4066, "train_loss": 3.5620431900024414, "perplexity": 35.23511568456837, "lr": 0.0026291804804649314, "grad_norm": 0.173539, "tokens_per_sec": 108047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:05.039030+00:00", "epoch": 0, "step": 4067, "train_loss": 3.6347854137420654, "perplexity": 37.89372055293602, "lr": 0.0026291804804649314, "grad_norm": 0.157603, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:05.345048+00:00", "epoch": 0, "step": 4068, "train_loss": 3.8491766452789307, "perplexity": 46.95438719534348, "lr": 0.0026291804804649314, "grad_norm": 0.150803, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:05.650789+00:00", "epoch": 0, "step": 4069, "train_loss": 3.6755287647247314, "perplexity": 39.46952143112386, "lr": 0.0026291804804649314, "grad_norm": 0.149406, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:05.955777+00:00", "epoch": 0, "step": 4070, "train_loss": 3.646864175796509, "perplexity": 38.354205230700714, "lr": 0.0026291804804649314, "grad_norm": 0.15236, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:06.259514+00:00", "epoch": 0, "step": 4071, "train_loss": 3.6087584495544434, "perplexity": 36.920186074307615, "lr": 0.0026291804804649314, "grad_norm": 0.155022, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:06.563950+00:00", "epoch": 0, "step": 4072, "train_loss": 3.6512694358825684, "perplexity": 38.523538184142595, "lr": 0.0026291804804649314, "grad_norm": 0.15534, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:06.869385+00:00", "epoch": 0, "step": 4073, "train_loss": 3.6926302909851074, "perplexity": 40.15031520251958, "lr": 0.0026291804804649314, "grad_norm": 0.159666, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:07.174310+00:00", "epoch": 0, "step": 4074, "train_loss": 3.6674251556396484, "perplexity": 39.15096831686578, "lr": 0.0026291804804649314, "grad_norm": 0.177523, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:07.478535+00:00", "epoch": 0, "step": 4075, "train_loss": 3.807269811630249, "perplexity": 45.02733778427354, "lr": 0.0026291804804649314, "grad_norm": 0.177826, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:07.782693+00:00", "epoch": 0, "step": 4076, "train_loss": 3.662351369857788, "perplexity": 38.952827776866734, "lr": 0.0026291804804649314, "grad_norm": 0.162646, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:08.087125+00:00", "epoch": 0, "step": 4077, "train_loss": 3.7113497257232666, "perplexity": 40.908985190206806, "lr": 0.0026291804804649314, "grad_norm": 0.146541, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:08.392426+00:00", "epoch": 0, "step": 4078, "train_loss": 3.5572354793548584, "perplexity": 35.06612200558877, "lr": 0.0026291804804649314, "grad_norm": 0.155438, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:08.697390+00:00", "epoch": 0, "step": 4079, "train_loss": 3.6874771118164062, "perplexity": 39.94394562099833, "lr": 0.0026291804804649314, "grad_norm": 0.14502, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:09.003513+00:00", "epoch": 0, "step": 4080, "train_loss": 3.663984775543213, "perplexity": 39.01650553887232, "lr": 0.0026291804804649314, "grad_norm": 0.156629, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:09.308115+00:00", "epoch": 0, "step": 4081, "train_loss": 3.665001392364502, "perplexity": 39.056190543514795, "lr": 0.0026291804804649314, "grad_norm": 0.163735, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:09.613373+00:00", "epoch": 0, "step": 4082, "train_loss": 3.652989625930786, "perplexity": 38.58986302045589, "lr": 0.0026291804804649314, "grad_norm": 0.177456, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:09.919553+00:00", "epoch": 0, "step": 4083, "train_loss": 3.486745595932007, "perplexity": 32.67942241683002, "lr": 0.0026291804804649314, "grad_norm": 0.174471, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:10.225141+00:00", "epoch": 0, "step": 4084, "train_loss": 3.669926404953003, "perplexity": 39.24901722074407, "lr": 0.0026291804804649314, "grad_norm": 0.136068, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:10.530034+00:00", "epoch": 0, "step": 4085, "train_loss": 3.7950003147125244, "perplexity": 44.478250403389005, "lr": 0.0026291804804649314, "grad_norm": 0.154543, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:10.834324+00:00", "epoch": 0, "step": 4086, "train_loss": 3.5921261310577393, "perplexity": 36.311196275585495, "lr": 0.0026291804804649314, "grad_norm": 0.151506, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:11.139143+00:00", "epoch": 0, "step": 4087, "train_loss": 3.677987813949585, "perplexity": 39.56669835966135, "lr": 0.0026291804804649314, "grad_norm": 0.16287, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:11.443652+00:00", "epoch": 0, "step": 4088, "train_loss": 3.6418986320495605, "perplexity": 38.16422780770066, "lr": 0.0026291804804649314, "grad_norm": 0.181068, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:11.749865+00:00", "epoch": 0, "step": 4089, "train_loss": 3.6143856048583984, "perplexity": 37.128527329740876, "lr": 0.0026291804804649314, "grad_norm": 0.173826, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:12.056334+00:00", "epoch": 0, "step": 4090, "train_loss": 3.7167670726776123, "perplexity": 41.13120473324654, "lr": 0.0026291804804649314, "grad_norm": 0.174212, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:12.361471+00:00", "epoch": 0, "step": 4091, "train_loss": 3.610457181930542, "perplexity": 36.98295689003734, "lr": 0.0026291804804649314, "grad_norm": 0.151067, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:12.666236+00:00", "epoch": 0, "step": 4092, "train_loss": 3.543670892715454, "perplexity": 34.593676066542876, "lr": 0.0026291804804649314, "grad_norm": 0.166367, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:12.970291+00:00", "epoch": 0, "step": 4093, "train_loss": 3.627243995666504, "perplexity": 37.60902302464939, "lr": 0.0026291804804649314, "grad_norm": 0.157642, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:13.275560+00:00", "epoch": 0, "step": 4094, "train_loss": 3.6405699253082275, "perplexity": 38.11355241475837, "lr": 0.0026291804804649314, "grad_norm": 0.148954, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:13.581473+00:00", "epoch": 0, "step": 4095, "train_loss": 3.565270185470581, "perplexity": 35.34900290113736, "lr": 0.0026291804804649314, "grad_norm": 0.158175, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:13.886615+00:00", "epoch": 0, "step": 4096, "train_loss": 3.632538318634033, "perplexity": 37.80866535820244, "lr": 0.0026291804804649314, "grad_norm": 0.140776, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:14.191658+00:00", "epoch": 0, "step": 4097, "train_loss": 3.5429186820983887, "perplexity": 34.56766412057977, "lr": 0.0026291804804649314, "grad_norm": 0.15679, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:14.496132+00:00", "epoch": 0, "step": 4098, "train_loss": 3.6193575859069824, "perplexity": 37.31358934516616, "lr": 0.0026291804804649314, "grad_norm": 0.144572, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:14.800087+00:00", "epoch": 0, "step": 4099, "train_loss": 3.6649935245513916, "perplexity": 39.055883257915625, "lr": 0.0026291804804649314, "grad_norm": 0.149047, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:15.105462+00:00", "epoch": 0, "step": 4100, "train_loss": 3.6024882793426514, "perplexity": 36.689414468061926, "lr": 0.0026291804804649314, "grad_norm": 0.151293, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:15.411545+00:00", "epoch": 0, "step": 4101, "train_loss": 3.609624147415161, "perplexity": 36.95216163900146, "lr": 0.0026291804804649314, "grad_norm": 0.176598, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:15.716867+00:00", "epoch": 0, "step": 4102, "train_loss": 3.5701942443847656, "perplexity": 35.52349272049374, "lr": 0.0026291804804649314, "grad_norm": 0.188652, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:16.021878+00:00", "epoch": 0, "step": 4103, "train_loss": 3.6287848949432373, "perplexity": 37.66701941284762, "lr": 0.0026291804804649314, "grad_norm": 0.16188, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:16.326197+00:00", "epoch": 0, "step": 4104, "train_loss": 3.5276670455932617, "perplexity": 34.044450744813574, "lr": 0.0026291804804649314, "grad_norm": 0.146028, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:16.631831+00:00", "epoch": 0, "step": 4105, "train_loss": 3.597990036010742, "perplexity": 36.52474718848452, "lr": 0.0026291804804649314, "grad_norm": 0.151113, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:16.937292+00:00", "epoch": 0, "step": 4106, "train_loss": 3.5094313621520996, "perplexity": 33.42925323964339, "lr": 0.0026291804804649314, "grad_norm": 0.161082, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:17.242954+00:00", "epoch": 0, "step": 4107, "train_loss": 3.749039649963379, "perplexity": 42.48026647913728, "lr": 0.0026291804804649314, "grad_norm": 0.159432, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:17.547518+00:00", "epoch": 0, "step": 4108, "train_loss": 3.644509792327881, "perplexity": 38.26401094158797, "lr": 0.0026291804804649314, "grad_norm": 0.159505, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:17.852851+00:00", "epoch": 0, "step": 4109, "train_loss": 3.523761034011841, "perplexity": 33.91173209495278, "lr": 0.0026291804804649314, "grad_norm": 0.1709, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:18.158167+00:00", "epoch": 0, "step": 4110, "train_loss": 3.585592269897461, "perplexity": 36.07471736205097, "lr": 0.0026291804804649314, "grad_norm": 0.159889, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:18.463580+00:00", "epoch": 0, "step": 4111, "train_loss": 3.6400091648101807, "perplexity": 38.09218583145178, "lr": 0.0026291804804649314, "grad_norm": 0.174969, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:18.768460+00:00", "epoch": 0, "step": 4112, "train_loss": 3.588900089263916, "perplexity": 36.194243587377336, "lr": 0.0026291804804649314, "grad_norm": 0.159413, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:19.073806+00:00", "epoch": 0, "step": 4113, "train_loss": 3.502352714538574, "perplexity": 33.1934548872985, "lr": 0.0026291804804649314, "grad_norm": 0.195606, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:19.378676+00:00", "epoch": 0, "step": 4114, "train_loss": 3.620166301727295, "perplexity": 37.3437776404129, "lr": 0.0026291804804649314, "grad_norm": 0.145126, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:19.683835+00:00", "epoch": 0, "step": 4115, "train_loss": 3.6521992683410645, "perplexity": 38.55937527902732, "lr": 0.0026291804804649314, "grad_norm": 0.149022, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:19.989136+00:00", "epoch": 0, "step": 4116, "train_loss": 3.595283031463623, "perplexity": 36.42600823545587, "lr": 0.0026291804804649314, "grad_norm": 0.161956, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:20.294135+00:00", "epoch": 0, "step": 4117, "train_loss": 3.7634708881378174, "perplexity": 43.09775416182644, "lr": 0.0026291804804649314, "grad_norm": 0.146316, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:20.599243+00:00", "epoch": 0, "step": 4118, "train_loss": 3.559213638305664, "perplexity": 35.13555702281232, "lr": 0.0026291804804649314, "grad_norm": 0.173093, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:20.904239+00:00", "epoch": 0, "step": 4119, "train_loss": 3.6079626083374023, "perplexity": 36.89081515733222, "lr": 0.0026291804804649314, "grad_norm": 0.148237, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:21.209331+00:00", "epoch": 0, "step": 4120, "train_loss": 3.6014180183410645, "perplexity": 36.65016822419408, "lr": 0.0026291804804649314, "grad_norm": 0.169838, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:21.513992+00:00", "epoch": 0, "step": 4121, "train_loss": 3.631071090698242, "perplexity": 37.75323210472747, "lr": 0.0026291804804649314, "grad_norm": 0.158519, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:21.819964+00:00", "epoch": 0, "step": 4122, "train_loss": 3.664457321166992, "perplexity": 39.03494697468661, "lr": 0.0026291804804649314, "grad_norm": 0.158216, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:22.124810+00:00", "epoch": 0, "step": 4123, "train_loss": 3.7184810638427734, "perplexity": 41.20176370622611, "lr": 0.0026291804804649314, "grad_norm": 0.155415, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:22.429657+00:00", "epoch": 0, "step": 4124, "train_loss": 3.677213191986084, "perplexity": 39.53606099380896, "lr": 0.0026291804804649314, "grad_norm": 0.141108, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:22.734094+00:00", "epoch": 0, "step": 4125, "train_loss": 3.692636728286743, "perplexity": 40.15057366304121, "lr": 0.0026291804804649314, "grad_norm": 0.137296, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:23.038733+00:00", "epoch": 0, "step": 4126, "train_loss": 3.591247797012329, "perplexity": 36.279316918079104, "lr": 0.0026291804804649314, "grad_norm": 0.149754, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:23.345088+00:00", "epoch": 0, "step": 4127, "train_loss": 3.613656997680664, "perplexity": 37.101485071017954, "lr": 0.0026291804804649314, "grad_norm": 0.139429, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:23.650111+00:00", "epoch": 0, "step": 4128, "train_loss": 3.6152424812316895, "perplexity": 37.16035552205037, "lr": 0.0026291804804649314, "grad_norm": 0.148982, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:23.954562+00:00", "epoch": 0, "step": 4129, "train_loss": 3.6214380264282227, "perplexity": 37.39129885541102, "lr": 0.0026291804804649314, "grad_norm": 0.140002, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:24.259523+00:00", "epoch": 0, "step": 4130, "train_loss": 3.6453351974487305, "perplexity": 38.29560729026353, "lr": 0.0026291804804649314, "grad_norm": 0.164809, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:24.564145+00:00", "epoch": 0, "step": 4131, "train_loss": 3.578712224960327, "perplexity": 35.82737352951501, "lr": 0.0026291804804649314, "grad_norm": 0.164146, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:24.868382+00:00", "epoch": 0, "step": 4132, "train_loss": 3.7542076110839844, "perplexity": 42.7003711008822, "lr": 0.0026291804804649314, "grad_norm": 0.155404, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:25.173585+00:00", "epoch": 0, "step": 4133, "train_loss": 3.6815032958984375, "perplexity": 39.706039155041, "lr": 0.0026291804804649314, "grad_norm": 0.158246, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:25.478477+00:00", "epoch": 0, "step": 4134, "train_loss": 3.7267520427703857, "perplexity": 41.543955806694825, "lr": 0.0026291804804649314, "grad_norm": 0.177406, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:25.784126+00:00", "epoch": 0, "step": 4135, "train_loss": 3.6551802158355713, "perplexity": 38.67449024274334, "lr": 0.0026291804804649314, "grad_norm": 0.162966, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:26.088332+00:00", "epoch": 0, "step": 4136, "train_loss": 3.675306797027588, "perplexity": 39.460761444597416, "lr": 0.0026291804804649314, "grad_norm": 0.154341, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:26.392981+00:00", "epoch": 0, "step": 4137, "train_loss": 3.545348644256592, "perplexity": 34.65176437512594, "lr": 0.0026291804804649314, "grad_norm": 0.147989, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:26.698301+00:00", "epoch": 0, "step": 4138, "train_loss": 3.6930370330810547, "perplexity": 40.166649347545, "lr": 0.0026291804804649314, "grad_norm": 0.161992, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:27.003076+00:00", "epoch": 0, "step": 4139, "train_loss": 3.65316104888916, "perplexity": 38.596478775968095, "lr": 0.0026291804804649314, "grad_norm": 0.130525, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:27.308241+00:00", "epoch": 0, "step": 4140, "train_loss": 3.6432387828826904, "perplexity": 38.215407916265505, "lr": 0.0026291804804649314, "grad_norm": 0.153577, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:27.613324+00:00", "epoch": 0, "step": 4141, "train_loss": 3.6362459659576416, "perplexity": 37.94910674781497, "lr": 0.0026291804804649314, "grad_norm": 0.17995, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:27.918615+00:00", "epoch": 0, "step": 4142, "train_loss": 3.5787203311920166, "perplexity": 35.827663955682794, "lr": 0.0026291804804649314, "grad_norm": 0.222654, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:28.224418+00:00", "epoch": 0, "step": 4143, "train_loss": 3.632298469543457, "perplexity": 37.79959807163406, "lr": 0.0026291804804649314, "grad_norm": 0.223859, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:28.530164+00:00", "epoch": 0, "step": 4144, "train_loss": 3.6449663639068604, "perplexity": 38.281485190300764, "lr": 0.0026291804804649314, "grad_norm": 0.186931, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:28.835377+00:00", "epoch": 0, "step": 4145, "train_loss": 3.5006027221679688, "perplexity": 33.1354173918903, "lr": 0.0026291804804649314, "grad_norm": 0.176642, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:29.140465+00:00", "epoch": 0, "step": 4146, "train_loss": 3.7209904193878174, "perplexity": 41.30528340999846, "lr": 0.0026291804804649314, "grad_norm": 0.175571, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:29.446576+00:00", "epoch": 0, "step": 4147, "train_loss": 3.674650192260742, "perplexity": 39.43485982502356, "lr": 0.0026291804804649314, "grad_norm": 0.189553, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:29.752653+00:00", "epoch": 0, "step": 4148, "train_loss": 3.6744625568389893, "perplexity": 39.42746114261767, "lr": 0.0026291804804649314, "grad_norm": 0.164671, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:30.059421+00:00", "epoch": 0, "step": 4149, "train_loss": 3.6368558406829834, "perplexity": 37.97225800783635, "lr": 0.0026291804804649314, "grad_norm": 0.187239, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:30.365872+00:00", "epoch": 0, "step": 4150, "train_loss": 3.6212239265441895, "perplexity": 37.38329423958647, "lr": 0.0026291804804649314, "grad_norm": 0.173391, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:30.671327+00:00", "epoch": 0, "step": 4151, "train_loss": 3.7066235542297363, "perplexity": 40.71609847745079, "lr": 0.0026291804804649314, "grad_norm": 0.145925, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:30.976853+00:00", "epoch": 0, "step": 4152, "train_loss": 3.5753633975982666, "perplexity": 35.707594512293646, "lr": 0.0026291804804649314, "grad_norm": 0.153144, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:31.281589+00:00", "epoch": 0, "step": 4153, "train_loss": 3.6914660930633545, "perplexity": 40.103599487448655, "lr": 0.0026291804804649314, "grad_norm": 0.17483, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:31.587300+00:00", "epoch": 0, "step": 4154, "train_loss": 3.5742204189300537, "perplexity": 35.666804808794026, "lr": 0.0026291804804649314, "grad_norm": 0.149794, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:31.892369+00:00", "epoch": 0, "step": 4155, "train_loss": 3.508035659790039, "perplexity": 33.38262849663556, "lr": 0.0026291804804649314, "grad_norm": 0.14584, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:32.197601+00:00", "epoch": 0, "step": 4156, "train_loss": 3.6898465156555176, "perplexity": 40.03870117185457, "lr": 0.0026291804804649314, "grad_norm": 0.158868, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:32.502697+00:00", "epoch": 0, "step": 4157, "train_loss": 3.6600608825683594, "perplexity": 38.86370892172125, "lr": 0.0026291804804649314, "grad_norm": 0.161067, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:32.808437+00:00", "epoch": 0, "step": 4158, "train_loss": 3.5897293090820312, "perplexity": 36.22426901858251, "lr": 0.0026291804804649314, "grad_norm": 0.165462, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:33.113262+00:00", "epoch": 0, "step": 4159, "train_loss": 3.5791401863098145, "perplexity": 35.84270954201552, "lr": 0.0026291804804649314, "grad_norm": 0.162329, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:33.418924+00:00", "epoch": 0, "step": 4160, "train_loss": 3.703491687774658, "perplexity": 40.58878056985695, "lr": 0.0026291804804649314, "grad_norm": 0.154325, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:33.723275+00:00", "epoch": 0, "step": 4161, "train_loss": 3.6558644771575928, "perplexity": 38.70096275659111, "lr": 0.0026291804804649314, "grad_norm": 0.154358, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:34.028444+00:00", "epoch": 0, "step": 4162, "train_loss": 3.6644628047943115, "perplexity": 39.03516102837514, "lr": 0.0026291804804649314, "grad_norm": 0.158418, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:34.333608+00:00", "epoch": 0, "step": 4163, "train_loss": 3.6415467262268066, "perplexity": 38.150799956522306, "lr": 0.0026291804804649314, "grad_norm": 0.159601, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:34.639631+00:00", "epoch": 0, "step": 4164, "train_loss": 3.586637020111084, "perplexity": 36.112426125410714, "lr": 0.0026291804804649314, "grad_norm": 0.157949, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:34.946663+00:00", "epoch": 0, "step": 4165, "train_loss": 3.6691102981567383, "perplexity": 39.21699889800601, "lr": 0.0026291804804649314, "grad_norm": 0.153889, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:35.251261+00:00", "epoch": 0, "step": 4166, "train_loss": 3.679297924041748, "perplexity": 39.618569061241544, "lr": 0.0026291804804649314, "grad_norm": 0.153605, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:35.556359+00:00", "epoch": 0, "step": 4167, "train_loss": 3.7027502059936523, "perplexity": 40.55869588355292, "lr": 0.0026291804804649314, "grad_norm": 0.153865, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:35.861778+00:00", "epoch": 0, "step": 4168, "train_loss": 3.6969246864318848, "perplexity": 40.323107286531226, "lr": 0.0026291804804649314, "grad_norm": 0.170962, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:36.167761+00:00", "epoch": 0, "step": 4169, "train_loss": 3.606239080429077, "perplexity": 36.827287569359086, "lr": 0.0026291804804649314, "grad_norm": 0.177244, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:36.473292+00:00", "epoch": 0, "step": 4170, "train_loss": 3.656426191329956, "perplexity": 38.72270774251722, "lr": 0.0026291804804649314, "grad_norm": 0.168499, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:36.777917+00:00", "epoch": 0, "step": 4171, "train_loss": 3.531485080718994, "perplexity": 34.174682109352915, "lr": 0.0026291804804649314, "grad_norm": 0.168933, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:37.083741+00:00", "epoch": 0, "step": 4172, "train_loss": 3.571354866027832, "perplexity": 35.56474599007039, "lr": 0.0026291804804649314, "grad_norm": 0.160682, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:37.388797+00:00", "epoch": 0, "step": 4173, "train_loss": 3.5561392307281494, "perplexity": 35.027701880348545, "lr": 0.0026291804804649314, "grad_norm": 0.158005, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:37.694703+00:00", "epoch": 0, "step": 4174, "train_loss": 3.530160903930664, "perplexity": 34.12945873703316, "lr": 0.0026291804804649314, "grad_norm": 0.151738, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:37.999856+00:00", "epoch": 0, "step": 4175, "train_loss": 3.6434526443481445, "perplexity": 38.223581593391536, "lr": 0.0026291804804649314, "grad_norm": 0.155919, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:38.305361+00:00", "epoch": 0, "step": 4176, "train_loss": 3.589195489883423, "perplexity": 36.2049369686936, "lr": 0.0026291804804649314, "grad_norm": 0.154283, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:38.610936+00:00", "epoch": 0, "step": 4177, "train_loss": 3.662652015686035, "perplexity": 38.96454054264513, "lr": 0.0026291804804649314, "grad_norm": 0.157446, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:38.916388+00:00", "epoch": 0, "step": 4178, "train_loss": 3.6589560508728027, "perplexity": 38.82079477511759, "lr": 0.0026291804804649314, "grad_norm": 0.176379, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:39.221501+00:00", "epoch": 0, "step": 4179, "train_loss": 3.664440631866455, "perplexity": 39.03429551416132, "lr": 0.0026291804804649314, "grad_norm": 0.190081, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:39.526159+00:00", "epoch": 0, "step": 4180, "train_loss": 3.632181406021118, "perplexity": 37.795173376531054, "lr": 0.0026291804804649314, "grad_norm": 0.146232, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:39.830937+00:00", "epoch": 0, "step": 4181, "train_loss": 3.667569398880005, "perplexity": 39.1566159867082, "lr": 0.0026291804804649314, "grad_norm": 0.157438, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:40.135131+00:00", "epoch": 0, "step": 4182, "train_loss": 3.639193058013916, "perplexity": 38.06111122153505, "lr": 0.0026291804804649314, "grad_norm": 0.163622, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:40.440432+00:00", "epoch": 0, "step": 4183, "train_loss": 3.622243642807007, "perplexity": 37.42143403526033, "lr": 0.0026291804804649314, "grad_norm": 0.135392, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:40.744787+00:00", "epoch": 0, "step": 4184, "train_loss": 3.5875766277313232, "perplexity": 36.146373582330504, "lr": 0.0026291804804649314, "grad_norm": 0.157149, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:41.049937+00:00", "epoch": 0, "step": 4185, "train_loss": 3.6376683712005615, "perplexity": 38.003124164438, "lr": 0.0026291804804649314, "grad_norm": 0.147604, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:41.354320+00:00", "epoch": 0, "step": 4186, "train_loss": 3.6379497051239014, "perplexity": 38.01381723654979, "lr": 0.0026291804804649314, "grad_norm": 0.153953, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:41.659482+00:00", "epoch": 0, "step": 4187, "train_loss": 3.639666795730591, "perplexity": 38.07914647711234, "lr": 0.0026291804804649314, "grad_norm": 0.148937, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:41.963970+00:00", "epoch": 0, "step": 4188, "train_loss": 3.617624521255493, "perplexity": 37.248978486021535, "lr": 0.0026291804804649314, "grad_norm": 0.133848, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:42.268795+00:00", "epoch": 0, "step": 4189, "train_loss": 3.5555551052093506, "perplexity": 35.007247280423854, "lr": 0.0026291804804649314, "grad_norm": 0.141729, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:42.572952+00:00", "epoch": 0, "step": 4190, "train_loss": 3.69317364692688, "perplexity": 40.17213704282532, "lr": 0.0026291804804649314, "grad_norm": 0.152481, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:42.878273+00:00", "epoch": 0, "step": 4191, "train_loss": 3.579676389694214, "perplexity": 35.861933677740716, "lr": 0.0026291804804649314, "grad_norm": 0.179019, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:43.183644+00:00", "epoch": 0, "step": 4192, "train_loss": 3.7055835723876953, "perplexity": 40.673776485221694, "lr": 0.0026291804804649314, "grad_norm": 0.180749, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:43.489597+00:00", "epoch": 0, "step": 4193, "train_loss": 3.6387641429901123, "perplexity": 38.044789739624086, "lr": 0.0026291804804649314, "grad_norm": 0.157902, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:43.794757+00:00", "epoch": 0, "step": 4194, "train_loss": 3.5960700511932373, "perplexity": 36.45468750670408, "lr": 0.0026291804804649314, "grad_norm": 0.186022, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:44.099955+00:00", "epoch": 0, "step": 4195, "train_loss": 3.6463265419006348, "perplexity": 38.33359025207158, "lr": 0.0026291804804649314, "grad_norm": 0.185402, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:44.405338+00:00", "epoch": 0, "step": 4196, "train_loss": 3.6749701499938965, "perplexity": 39.447479332127216, "lr": 0.0026291804804649314, "grad_norm": 0.151606, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:44.710764+00:00", "epoch": 0, "step": 4197, "train_loss": 3.533379077911377, "perplexity": 34.239470196178424, "lr": 0.0026291804804649314, "grad_norm": 0.166108, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:45.016933+00:00", "epoch": 0, "step": 4198, "train_loss": 3.5824711322784424, "perplexity": 35.962298732863616, "lr": 0.0026291804804649314, "grad_norm": 0.163436, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:45.322452+00:00", "epoch": 0, "step": 4199, "train_loss": 3.5795891284942627, "perplexity": 35.8588044589071, "lr": 0.0026291804804649314, "grad_norm": 0.146518, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:45.627889+00:00", "epoch": 0, "step": 4200, "train_loss": 3.6107101440429688, "perplexity": 36.99231336030238, "lr": 0.0026291804804649314, "grad_norm": 0.147345, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:45.932406+00:00", "epoch": 0, "step": 4201, "train_loss": 3.655665874481201, "perplexity": 38.69327740499968, "lr": 0.0026291804804649314, "grad_norm": 0.141643, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:46.237326+00:00", "epoch": 0, "step": 4202, "train_loss": 3.635770559310913, "perplexity": 37.93106977801703, "lr": 0.0026291804804649314, "grad_norm": 0.162499, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:46.542746+00:00", "epoch": 0, "step": 4203, "train_loss": 3.6633145809173584, "perplexity": 38.99036564692652, "lr": 0.0026291804804649314, "grad_norm": 0.191839, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:46.848934+00:00", "epoch": 0, "step": 4204, "train_loss": 3.625288963317871, "perplexity": 37.53556799491172, "lr": 0.0026291804804649314, "grad_norm": 0.170663, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:47.154518+00:00", "epoch": 0, "step": 4205, "train_loss": 3.695875644683838, "perplexity": 40.28082884337989, "lr": 0.0026291804804649314, "grad_norm": 0.14736, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:47.459776+00:00", "epoch": 0, "step": 4206, "train_loss": 3.6772191524505615, "perplexity": 39.536296647798395, "lr": 0.0026291804804649314, "grad_norm": 0.171489, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:47.763461+00:00", "epoch": 0, "step": 4207, "train_loss": 3.690485715866089, "perplexity": 40.06430209926205, "lr": 0.0026291804804649314, "grad_norm": 0.189404, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:48.068728+00:00", "epoch": 0, "step": 4208, "train_loss": 3.6403727531433105, "perplexity": 38.10603822393505, "lr": 0.0026291804804649314, "grad_norm": 0.163503, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:48.374895+00:00", "epoch": 0, "step": 4209, "train_loss": 3.5577445030212402, "perplexity": 35.08397603525416, "lr": 0.0026291804804649314, "grad_norm": 0.158967, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:48.680216+00:00", "epoch": 0, "step": 4210, "train_loss": 3.676928997039795, "perplexity": 39.52482664152703, "lr": 0.0026291804804649314, "grad_norm": 0.148272, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:48.985426+00:00", "epoch": 0, "step": 4211, "train_loss": 3.671903133392334, "perplexity": 39.326678601742145, "lr": 0.0026291804804649314, "grad_norm": 0.1698, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:49.290288+00:00", "epoch": 0, "step": 4212, "train_loss": 3.616731882095337, "perplexity": 37.21574342481448, "lr": 0.0026291804804649314, "grad_norm": 0.174159, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:49.594489+00:00", "epoch": 0, "step": 4213, "train_loss": 3.630988359451294, "perplexity": 37.75010886195579, "lr": 0.0026291804804649314, "grad_norm": 0.135907, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:49.899672+00:00", "epoch": 0, "step": 4214, "train_loss": 3.6170551776885986, "perplexity": 37.2277770557688, "lr": 0.0026291804804649314, "grad_norm": 0.147833, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:50.205573+00:00", "epoch": 0, "step": 4215, "train_loss": 3.63834285736084, "perplexity": 38.02876539208876, "lr": 0.0026291804804649314, "grad_norm": 0.159884, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:50.511280+00:00", "epoch": 0, "step": 4216, "train_loss": 3.5532302856445312, "perplexity": 34.9259562771067, "lr": 0.0026291804804649314, "grad_norm": 0.1516, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:50.817293+00:00", "epoch": 0, "step": 4217, "train_loss": 3.6059064865112305, "perplexity": 36.81504107417052, "lr": 0.0026291804804649314, "grad_norm": 0.138332, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:51.122945+00:00", "epoch": 0, "step": 4218, "train_loss": 3.6719846725463867, "perplexity": 39.32988539658493, "lr": 0.0026291804804649314, "grad_norm": 0.141745, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:51.427711+00:00", "epoch": 0, "step": 4219, "train_loss": 3.5919089317321777, "perplexity": 36.303310364682375, "lr": 0.0026291804804649314, "grad_norm": 0.14228, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:51.733397+00:00", "epoch": 0, "step": 4220, "train_loss": 3.566728115081787, "perplexity": 35.400576845648686, "lr": 0.0026291804804649314, "grad_norm": 0.147144, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:52.039565+00:00", "epoch": 0, "step": 4221, "train_loss": 3.691326856613159, "perplexity": 40.09801599333794, "lr": 0.0026291804804649314, "grad_norm": 0.139375, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:52.345036+00:00", "epoch": 0, "step": 4222, "train_loss": 3.580911159515381, "perplexity": 35.906242260995775, "lr": 0.0026291804804649314, "grad_norm": 0.159488, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:52.651104+00:00", "epoch": 0, "step": 4223, "train_loss": 3.5337767601013184, "perplexity": 34.25308933153277, "lr": 0.0026291804804649314, "grad_norm": 0.147714, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:52.957265+00:00", "epoch": 0, "step": 4224, "train_loss": 3.713737726211548, "perplexity": 41.00679260241089, "lr": 0.0026291804804649314, "grad_norm": 0.155622, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:53.262138+00:00", "epoch": 0, "step": 4225, "train_loss": 3.7340500354766846, "perplexity": 41.84825231944233, "lr": 0.0026291804804649314, "grad_norm": 0.137562, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:53.566655+00:00", "epoch": 0, "step": 4226, "train_loss": 3.606888771057129, "perplexity": 36.85122168699355, "lr": 0.0026291804804649314, "grad_norm": 0.174398, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:53.873159+00:00", "epoch": 0, "step": 4227, "train_loss": 3.589818000793457, "perplexity": 36.22748195347516, "lr": 0.0026291804804649314, "grad_norm": 0.192588, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:54.178632+00:00", "epoch": 0, "step": 4228, "train_loss": 3.466259479522705, "perplexity": 32.01675884202226, "lr": 0.0026291804804649314, "grad_norm": 0.174085, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:54.483421+00:00", "epoch": 0, "step": 4229, "train_loss": 3.638744831085205, "perplexity": 38.04405502935677, "lr": 0.0026291804804649314, "grad_norm": 0.194427, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:54.788447+00:00", "epoch": 0, "step": 4230, "train_loss": 3.7049612998962402, "perplexity": 40.64847418627012, "lr": 0.0026291804804649314, "grad_norm": 0.159554, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:55.093634+00:00", "epoch": 0, "step": 4231, "train_loss": 3.533097505569458, "perplexity": 34.22983066554552, "lr": 0.0026291804804649314, "grad_norm": 0.180293, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:55.399072+00:00", "epoch": 0, "step": 4232, "train_loss": 3.663691282272339, "perplexity": 39.00505613728699, "lr": 0.0026291804804649314, "grad_norm": 0.170501, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:55.704638+00:00", "epoch": 0, "step": 4233, "train_loss": 3.7240164279937744, "perplexity": 41.43046285458742, "lr": 0.0026291804804649314, "grad_norm": 0.171549, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:56.010303+00:00", "epoch": 0, "step": 4234, "train_loss": 3.5676465034484863, "perplexity": 35.43310325725102, "lr": 0.0026291804804649314, "grad_norm": 0.165266, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:56.315270+00:00", "epoch": 0, "step": 4235, "train_loss": 3.629214286804199, "perplexity": 37.683196797380255, "lr": 0.0026291804804649314, "grad_norm": 0.162487, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:56.620319+00:00", "epoch": 0, "step": 4236, "train_loss": 3.5848729610443115, "perplexity": 36.048777828863955, "lr": 0.0026291804804649314, "grad_norm": 0.160318, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:56.925597+00:00", "epoch": 0, "step": 4237, "train_loss": 3.6011528968811035, "perplexity": 36.6404527660318, "lr": 0.0026291804804649314, "grad_norm": 0.152522, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:57.231018+00:00", "epoch": 0, "step": 4238, "train_loss": 3.6921701431274414, "perplexity": 40.13184437097722, "lr": 0.0026291804804649314, "grad_norm": 0.148714, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:57.535821+00:00", "epoch": 0, "step": 4239, "train_loss": 3.567148447036743, "perplexity": 35.41545996702742, "lr": 0.0026291804804649314, "grad_norm": 0.154912, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:57.840599+00:00", "epoch": 0, "step": 4240, "train_loss": 3.669959545135498, "perplexity": 39.2503179618908, "lr": 0.0026291804804649314, "grad_norm": 0.154654, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:58.145111+00:00", "epoch": 0, "step": 4241, "train_loss": 3.5424861907958984, "perplexity": 34.55271713895298, "lr": 0.0026291804804649314, "grad_norm": 0.139803, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:58.449922+00:00", "epoch": 0, "step": 4242, "train_loss": 3.696885585784912, "perplexity": 40.32153065777218, "lr": 0.0026291804804649314, "grad_norm": 0.176913, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:58.754700+00:00", "epoch": 0, "step": 4243, "train_loss": 3.5243327617645264, "perplexity": 33.93112591679792, "lr": 0.0026291804804649314, "grad_norm": 0.187248, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:59.059917+00:00", "epoch": 0, "step": 4244, "train_loss": 3.6871345043182373, "perplexity": 39.930262869762544, "lr": 0.0026291804804649314, "grad_norm": 0.188572, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:59.365197+00:00", "epoch": 0, "step": 4245, "train_loss": 3.633840322494507, "perplexity": 37.85792444726151, "lr": 0.0026291804804649314, "grad_norm": 0.152777, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:59.668843+00:00", "epoch": 0, "step": 4246, "train_loss": 3.760795831680298, "perplexity": 42.98261930103572, "lr": 0.0026291804804649314, "grad_norm": 0.162762, "tokens_per_sec": 107915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:19:59.973942+00:00", "epoch": 0, "step": 4247, "train_loss": 3.7352278232574463, "perplexity": 41.8975697166833, "lr": 0.0026291804804649314, "grad_norm": 0.170014, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:00.279641+00:00", "epoch": 0, "step": 4248, "train_loss": 3.7205498218536377, "perplexity": 41.28708841260926, "lr": 0.0026291804804649314, "grad_norm": 0.164479, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:00.585078+00:00", "epoch": 0, "step": 4249, "train_loss": 3.632847785949707, "perplexity": 37.82036771503513, "lr": 0.0026291804804649314, "grad_norm": 0.134373, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:00.890634+00:00", "epoch": 0, "step": 4250, "train_loss": 3.5578057765960693, "perplexity": 35.08612582174694, "lr": 0.0026291804804649314, "grad_norm": 0.166496, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:01.194591+00:00", "epoch": 0, "step": 4251, "train_loss": 3.69008207321167, "perplexity": 40.04813370136231, "lr": 0.0026291804804649314, "grad_norm": 0.153919, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:01.499339+00:00", "epoch": 0, "step": 4252, "train_loss": 3.5737032890319824, "perplexity": 35.64836520590464, "lr": 0.0026291804804649314, "grad_norm": 0.143315, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:01.805070+00:00", "epoch": 0, "step": 4253, "train_loss": 3.5802252292633057, "perplexity": 35.88162152820868, "lr": 0.0026291804804649314, "grad_norm": 0.153521, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:02.110753+00:00", "epoch": 0, "step": 4254, "train_loss": 3.6787948608398438, "perplexity": 39.59864342938103, "lr": 0.0026291804804649314, "grad_norm": 0.171522, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:02.415712+00:00", "epoch": 0, "step": 4255, "train_loss": 3.594770669937134, "perplexity": 36.40734973063062, "lr": 0.0026291804804649314, "grad_norm": 0.167396, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:02.721146+00:00", "epoch": 0, "step": 4256, "train_loss": 3.5534543991088867, "perplexity": 34.93378453133971, "lr": 0.0026291804804649314, "grad_norm": 0.145394, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:03.026090+00:00", "epoch": 0, "step": 4257, "train_loss": 3.6989502906799316, "perplexity": 40.404868724146866, "lr": 0.0026291804804649314, "grad_norm": 0.144832, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:03.332249+00:00", "epoch": 0, "step": 4258, "train_loss": 3.635206699371338, "perplexity": 37.90968799603563, "lr": 0.0026291804804649314, "grad_norm": 0.155118, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:03.635659+00:00", "epoch": 0, "step": 4259, "train_loss": 3.7067978382110596, "perplexity": 40.72319525960715, "lr": 0.0026291804804649314, "grad_norm": 0.160323, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:03.940803+00:00", "epoch": 0, "step": 4260, "train_loss": 3.5833518505096436, "perplexity": 35.99398533643009, "lr": 0.0026291804804649314, "grad_norm": 0.149, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:04.246764+00:00", "epoch": 0, "step": 4261, "train_loss": 3.6402387619018555, "perplexity": 38.10093269062251, "lr": 0.0026291804804649314, "grad_norm": 0.151959, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:04.551320+00:00", "epoch": 0, "step": 4262, "train_loss": 3.6148898601531982, "perplexity": 37.14725430742703, "lr": 0.0026291804804649314, "grad_norm": 0.174616, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:04.855606+00:00", "epoch": 0, "step": 4263, "train_loss": 3.685957193374634, "perplexity": 39.883280196339534, "lr": 0.0026291804804649314, "grad_norm": 0.148977, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:05.161611+00:00", "epoch": 0, "step": 4264, "train_loss": 3.6421470642089844, "perplexity": 38.173710207045374, "lr": 0.0026291804804649314, "grad_norm": 0.14746, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:05.467180+00:00", "epoch": 0, "step": 4265, "train_loss": 3.6654982566833496, "perplexity": 39.075600992806486, "lr": 0.0026291804804649314, "grad_norm": 0.152268, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:05.772833+00:00", "epoch": 0, "step": 4266, "train_loss": 3.637993335723877, "perplexity": 38.01547583838582, "lr": 0.0026291804804649314, "grad_norm": 0.153249, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:06.078142+00:00", "epoch": 0, "step": 4267, "train_loss": 3.615208387374878, "perplexity": 37.15908860380732, "lr": 0.0026291804804649314, "grad_norm": 0.145642, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:06.383536+00:00", "epoch": 0, "step": 4268, "train_loss": 3.702815294265747, "perplexity": 40.56133586490137, "lr": 0.0026291804804649314, "grad_norm": 0.137597, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:06.688252+00:00", "epoch": 0, "step": 4269, "train_loss": 3.5897789001464844, "perplexity": 36.22606546318561, "lr": 0.0026291804804649314, "grad_norm": 0.157346, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:06.994542+00:00", "epoch": 0, "step": 4270, "train_loss": 3.5972471237182617, "perplexity": 36.49762258166792, "lr": 0.0026291804804649314, "grad_norm": 0.143109, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:07.299873+00:00", "epoch": 0, "step": 4271, "train_loss": 3.7320199012756348, "perplexity": 41.763380930458595, "lr": 0.0026291804804649314, "grad_norm": 0.156842, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:07.604658+00:00", "epoch": 0, "step": 4272, "train_loss": 3.55718994140625, "perplexity": 35.06452520268482, "lr": 0.0026291804804649314, "grad_norm": 0.161335, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:07.909758+00:00", "epoch": 0, "step": 4273, "train_loss": 3.5500054359436035, "perplexity": 34.81350673134689, "lr": 0.0026291804804649314, "grad_norm": 0.155443, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:08.215619+00:00", "epoch": 0, "step": 4274, "train_loss": 3.5707595348358154, "perplexity": 35.54357948861078, "lr": 0.0026291804804649314, "grad_norm": 0.151421, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:08.520758+00:00", "epoch": 0, "step": 4275, "train_loss": 3.660881519317627, "perplexity": 38.895614999334576, "lr": 0.0026291804804649314, "grad_norm": 0.144716, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:08.826515+00:00", "epoch": 0, "step": 4276, "train_loss": 3.6035451889038086, "perplexity": 36.7282123603265, "lr": 0.0026291804804649314, "grad_norm": 0.14957, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:09.132700+00:00", "epoch": 0, "step": 4277, "train_loss": 3.549649477005005, "perplexity": 34.80111675773366, "lr": 0.0026291804804649314, "grad_norm": 0.164618, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:09.437255+00:00", "epoch": 0, "step": 4278, "train_loss": 3.6850767135620117, "perplexity": 39.848179228380005, "lr": 0.0026291804804649314, "grad_norm": 0.153903, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:09.742639+00:00", "epoch": 0, "step": 4279, "train_loss": 3.5830483436584473, "perplexity": 35.983062572929256, "lr": 0.0026291804804649314, "grad_norm": 0.13372, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:10.049068+00:00", "epoch": 0, "step": 4280, "train_loss": 3.630497694015503, "perplexity": 37.7315907318146, "lr": 0.0026291804804649314, "grad_norm": 0.141776, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:10.354926+00:00", "epoch": 0, "step": 4281, "train_loss": 3.6161587238311768, "perplexity": 37.194419025625265, "lr": 0.0026291804804649314, "grad_norm": 0.148152, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:10.660303+00:00", "epoch": 0, "step": 4282, "train_loss": 3.614288330078125, "perplexity": 37.12491583605948, "lr": 0.0026291804804649314, "grad_norm": 0.147472, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:10.964441+00:00", "epoch": 0, "step": 4283, "train_loss": 3.6005125045776367, "perplexity": 36.61699601364838, "lr": 0.0026291804804649314, "grad_norm": 0.185094, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:11.269177+00:00", "epoch": 0, "step": 4284, "train_loss": 3.606595516204834, "perplexity": 36.84041647183916, "lr": 0.0026291804804649314, "grad_norm": 0.180449, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:11.574316+00:00", "epoch": 0, "step": 4285, "train_loss": 3.643463611602783, "perplexity": 38.224000803442856, "lr": 0.0026291804804649314, "grad_norm": 0.150137, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:11.879308+00:00", "epoch": 0, "step": 4286, "train_loss": 3.535407543182373, "perplexity": 34.30899426210624, "lr": 0.0026291804804649314, "grad_norm": 0.154772, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:12.185186+00:00", "epoch": 0, "step": 4287, "train_loss": 3.5753719806671143, "perplexity": 35.70790099435101, "lr": 0.0026291804804649314, "grad_norm": 0.158887, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:12.490257+00:00", "epoch": 0, "step": 4288, "train_loss": 3.5966992378234863, "perplexity": 36.477631525971375, "lr": 0.0026291804804649314, "grad_norm": 0.171564, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:12.795208+00:00", "epoch": 0, "step": 4289, "train_loss": 3.6257145404815674, "perplexity": 37.55154567510349, "lr": 0.0026291804804649314, "grad_norm": 0.162411, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:13.100835+00:00", "epoch": 0, "step": 4290, "train_loss": 3.628072500228882, "perplexity": 37.64019518316656, "lr": 0.0026291804804649314, "grad_norm": 0.165086, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:13.405300+00:00", "epoch": 0, "step": 4291, "train_loss": 3.5877342224121094, "perplexity": 36.15207050742754, "lr": 0.0026291804804649314, "grad_norm": 0.154769, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:13.711081+00:00", "epoch": 0, "step": 4292, "train_loss": 3.6651220321655273, "perplexity": 39.06090255879335, "lr": 0.0026291804804649314, "grad_norm": 0.159735, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:14.014729+00:00", "epoch": 0, "step": 4293, "train_loss": 3.6308653354644775, "perplexity": 37.74546497872121, "lr": 0.0026291804804649314, "grad_norm": 0.166796, "tokens_per_sec": 107915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:14.320020+00:00", "epoch": 0, "step": 4294, "train_loss": 3.5825068950653076, "perplexity": 35.963584867886134, "lr": 0.0026291804804649314, "grad_norm": 0.191899, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:14.625187+00:00", "epoch": 0, "step": 4295, "train_loss": 3.617142677307129, "perplexity": 37.23103461457541, "lr": 0.0026291804804649314, "grad_norm": 0.195368, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:14.931050+00:00", "epoch": 0, "step": 4296, "train_loss": 3.511582612991333, "perplexity": 33.501245357517874, "lr": 0.0026291804804649314, "grad_norm": 0.167693, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:15.237524+00:00", "epoch": 0, "step": 4297, "train_loss": 3.5807924270629883, "perplexity": 35.901979277878155, "lr": 0.0026291804804649314, "grad_norm": 0.157427, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:15.542371+00:00", "epoch": 0, "step": 4298, "train_loss": 3.708284378051758, "perplexity": 40.78377692917769, "lr": 0.0026291804804649314, "grad_norm": 0.149305, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:15.846189+00:00", "epoch": 0, "step": 4299, "train_loss": 3.650836706161499, "perplexity": 38.50687151055215, "lr": 0.0026291804804649314, "grad_norm": 0.138902, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:16.150633+00:00", "epoch": 0, "step": 4300, "train_loss": 3.6453936100006104, "perplexity": 38.29784429974523, "lr": 0.0026291804804649314, "grad_norm": 0.144615, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:16.454935+00:00", "epoch": 0, "step": 4301, "train_loss": 3.6337356567382812, "perplexity": 37.85396222632814, "lr": 0.0026291804804649314, "grad_norm": 0.156012, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:16.760617+00:00", "epoch": 0, "step": 4302, "train_loss": 3.625458002090454, "perplexity": 37.54191349755667, "lr": 0.0026291804804649314, "grad_norm": 0.146716, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:17.066372+00:00", "epoch": 0, "step": 4303, "train_loss": 3.6106467247009277, "perplexity": 36.9899674065187, "lr": 0.0026291804804649314, "grad_norm": 0.146641, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:17.371493+00:00", "epoch": 0, "step": 4304, "train_loss": 3.6227123737335205, "perplexity": 37.438978730256494, "lr": 0.0026291804804649314, "grad_norm": 0.151821, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:17.676911+00:00", "epoch": 0, "step": 4305, "train_loss": 3.5994982719421387, "perplexity": 36.57987668827499, "lr": 0.0026291804804649314, "grad_norm": 0.152059, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:17.982597+00:00", "epoch": 0, "step": 4306, "train_loss": 3.7398529052734375, "perplexity": 42.09179822805655, "lr": 0.0026291804804649314, "grad_norm": 0.167318, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:18.287683+00:00", "epoch": 0, "step": 4307, "train_loss": 3.683651924133301, "perplexity": 39.79144439105802, "lr": 0.0026291804804649314, "grad_norm": 0.151762, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:18.593617+00:00", "epoch": 0, "step": 4308, "train_loss": 3.7239155769348145, "perplexity": 41.426284759221545, "lr": 0.0026291804804649314, "grad_norm": 0.143246, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:18.899165+00:00", "epoch": 0, "step": 4309, "train_loss": 3.653892755508423, "perplexity": 38.62473040966191, "lr": 0.0026291804804649314, "grad_norm": 0.152274, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:19.204131+00:00", "epoch": 0, "step": 4310, "train_loss": 3.680952548980713, "perplexity": 39.684177197116874, "lr": 0.0026291804804649314, "grad_norm": 0.175304, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:19.507388+00:00", "epoch": 0, "step": 4311, "train_loss": 3.5788612365722656, "perplexity": 35.83271262197967, "lr": 0.0026291804804649314, "grad_norm": 0.188238, "tokens_per_sec": 107967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:19.813831+00:00", "epoch": 0, "step": 4312, "train_loss": 3.6406588554382324, "perplexity": 38.11694200864584, "lr": 0.0026291804804649314, "grad_norm": 0.159301, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:20.120310+00:00", "epoch": 0, "step": 4313, "train_loss": 3.5638699531555176, "perplexity": 35.29954072232199, "lr": 0.0026291804804649314, "grad_norm": 0.151396, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:20.425716+00:00", "epoch": 0, "step": 4314, "train_loss": 3.6675851345062256, "perplexity": 39.15723214542925, "lr": 0.0026291804804649314, "grad_norm": 0.171709, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:20.731039+00:00", "epoch": 0, "step": 4315, "train_loss": 3.6543161869049072, "perplexity": 38.64108879628094, "lr": 0.0026291804804649314, "grad_norm": 0.166212, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:21.035432+00:00", "epoch": 0, "step": 4316, "train_loss": 3.686753511428833, "perplexity": 39.91505262121938, "lr": 0.0026291804804649314, "grad_norm": 0.157997, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:21.341038+00:00", "epoch": 0, "step": 4317, "train_loss": 3.6844711303710938, "perplexity": 39.824055146157214, "lr": 0.0026291804804649314, "grad_norm": 0.16008, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:21.647187+00:00", "epoch": 0, "step": 4318, "train_loss": 3.5609912872314453, "perplexity": 35.19807125572788, "lr": 0.0026291804804649314, "grad_norm": 0.169915, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:21.953407+00:00", "epoch": 0, "step": 4319, "train_loss": 3.622006416320801, "perplexity": 37.412557732844, "lr": 0.0026291804804649314, "grad_norm": 0.158376, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:22.258573+00:00", "epoch": 0, "step": 4320, "train_loss": 3.6205005645751953, "perplexity": 37.356262364351785, "lr": 0.0026291804804649314, "grad_norm": 0.144703, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:22.562893+00:00", "epoch": 0, "step": 4321, "train_loss": 3.595381021499634, "perplexity": 36.42957779620242, "lr": 0.0026291804804649314, "grad_norm": 0.163878, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:22.866285+00:00", "epoch": 0, "step": 4322, "train_loss": 3.61152720451355, "perplexity": 37.022550668436864, "lr": 0.0026291804804649314, "grad_norm": 0.178896, "tokens_per_sec": 108006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:23.170494+00:00", "epoch": 0, "step": 4323, "train_loss": 3.577855348587036, "perplexity": 35.79668704875826, "lr": 0.0026291804804649314, "grad_norm": 0.130411, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:23.475685+00:00", "epoch": 0, "step": 4324, "train_loss": 3.7873799800872803, "perplexity": 44.14059939254822, "lr": 0.0026291804804649314, "grad_norm": 0.158158, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:23.781571+00:00", "epoch": 0, "step": 4325, "train_loss": 3.772791624069214, "perplexity": 43.501334860374676, "lr": 0.0026291804804649314, "grad_norm": 0.179216, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:24.086382+00:00", "epoch": 0, "step": 4326, "train_loss": 3.7433900833129883, "perplexity": 42.240948041607695, "lr": 0.0026291804804649314, "grad_norm": 0.152938, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:24.391168+00:00", "epoch": 0, "step": 4327, "train_loss": 3.6891627311706543, "perplexity": 40.01133268733809, "lr": 0.0026291804804649314, "grad_norm": 0.160557, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:24.696694+00:00", "epoch": 0, "step": 4328, "train_loss": 3.565957546234131, "perplexity": 35.373308771255964, "lr": 0.0026291804804649314, "grad_norm": 0.169599, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:25.002467+00:00", "epoch": 0, "step": 4329, "train_loss": 3.6369595527648926, "perplexity": 37.97619639399472, "lr": 0.0026291804804649314, "grad_norm": 0.163134, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:25.307196+00:00", "epoch": 0, "step": 4330, "train_loss": 3.6201765537261963, "perplexity": 37.34416049074273, "lr": 0.0026291804804649314, "grad_norm": 0.156719, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:25.611853+00:00", "epoch": 0, "step": 4331, "train_loss": 3.5862293243408203, "perplexity": 36.09770624284624, "lr": 0.0026291804804649314, "grad_norm": 0.156414, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:25.917416+00:00", "epoch": 0, "step": 4332, "train_loss": 3.655604124069214, "perplexity": 38.69088815294819, "lr": 0.0026291804804649314, "grad_norm": 0.152967, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:26.222175+00:00", "epoch": 0, "step": 4333, "train_loss": 3.5466842651367188, "perplexity": 34.6980769162724, "lr": 0.0026291804804649314, "grad_norm": 0.146138, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:26.527458+00:00", "epoch": 0, "step": 4334, "train_loss": 3.744885206222534, "perplexity": 42.30415068683349, "lr": 0.0026291804804649314, "grad_norm": 0.166456, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:26.832671+00:00", "epoch": 0, "step": 4335, "train_loss": 3.7219481468200684, "perplexity": 41.34486156253309, "lr": 0.0026291804804649314, "grad_norm": 0.168131, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:27.137872+00:00", "epoch": 0, "step": 4336, "train_loss": 3.7209227085113525, "perplexity": 41.302486687741464, "lr": 0.0026291804804649314, "grad_norm": 0.167073, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:27.441716+00:00", "epoch": 0, "step": 4337, "train_loss": 3.714789628982544, "perplexity": 41.04995045613248, "lr": 0.0026291804804649314, "grad_norm": 0.165901, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:27.746105+00:00", "epoch": 0, "step": 4338, "train_loss": 3.575026035308838, "perplexity": 35.69555014823034, "lr": 0.0026291804804649314, "grad_norm": 0.151006, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:28.051472+00:00", "epoch": 0, "step": 4339, "train_loss": 3.579885244369507, "perplexity": 35.86942439246252, "lr": 0.0026291804804649314, "grad_norm": 0.136674, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:28.356718+00:00", "epoch": 0, "step": 4340, "train_loss": 3.5047507286071777, "perplexity": 33.273148774445296, "lr": 0.0026291804804649314, "grad_norm": 0.156805, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:28.661759+00:00", "epoch": 0, "step": 4341, "train_loss": 3.539203643798828, "perplexity": 34.439482172294404, "lr": 0.0026291804804649314, "grad_norm": 0.169459, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:28.968098+00:00", "epoch": 0, "step": 4342, "train_loss": 3.7116172313690186, "perplexity": 40.91993003854626, "lr": 0.0026291804804649314, "grad_norm": 0.162429, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:29.271633+00:00", "epoch": 0, "step": 4343, "train_loss": 3.643174409866333, "perplexity": 38.21294795436504, "lr": 0.0026291804804649314, "grad_norm": 0.169099, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:29.577156+00:00", "epoch": 0, "step": 4344, "train_loss": 3.710023880004883, "perplexity": 40.85478212773824, "lr": 0.0026291804804649314, "grad_norm": 0.14757, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:29.883621+00:00", "epoch": 0, "step": 4345, "train_loss": 3.6268677711486816, "perplexity": 37.59487624944488, "lr": 0.0026291804804649314, "grad_norm": 0.182575, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:30.189123+00:00", "epoch": 0, "step": 4346, "train_loss": 3.6826725006103516, "perplexity": 39.75249079355806, "lr": 0.0026291804804649314, "grad_norm": 0.163701, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:30.495160+00:00", "epoch": 0, "step": 4347, "train_loss": 3.5978517532348633, "perplexity": 36.51969679425434, "lr": 0.0026291804804649314, "grad_norm": 0.154072, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:30.801198+00:00", "epoch": 0, "step": 4348, "train_loss": 3.661489486694336, "perplexity": 38.91926945419103, "lr": 0.0026291804804649314, "grad_norm": 0.155895, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:31.106236+00:00", "epoch": 0, "step": 4349, "train_loss": 3.7008872032165527, "perplexity": 40.48320526193945, "lr": 0.0026291804804649314, "grad_norm": 0.155112, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:31.410625+00:00", "epoch": 0, "step": 4350, "train_loss": 3.5593032836914062, "perplexity": 35.13870690455897, "lr": 0.0026291804804649314, "grad_norm": 0.177744, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:31.715413+00:00", "epoch": 0, "step": 4351, "train_loss": 3.614423990249634, "perplexity": 37.12995255014207, "lr": 0.0026291804804649314, "grad_norm": 0.175787, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:32.020945+00:00", "epoch": 0, "step": 4352, "train_loss": 3.678394079208374, "perplexity": 39.582776200322776, "lr": 0.0026291804804649314, "grad_norm": 0.153926, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:32.326013+00:00", "epoch": 0, "step": 4353, "train_loss": 3.6405551433563232, "perplexity": 38.11298902622367, "lr": 0.0026291804804649314, "grad_norm": 0.152525, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:32.630173+00:00", "epoch": 0, "step": 4354, "train_loss": 3.5884461402893066, "perplexity": 36.17781697631779, "lr": 0.0026291804804649314, "grad_norm": 0.176827, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:32.935235+00:00", "epoch": 0, "step": 4355, "train_loss": 3.581584930419922, "perplexity": 35.93044299428209, "lr": 0.0026291804804649314, "grad_norm": 0.156881, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:33.239605+00:00", "epoch": 0, "step": 4356, "train_loss": 3.7243309020996094, "perplexity": 41.44349371117404, "lr": 0.0026291804804649314, "grad_norm": 0.159466, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:33.545032+00:00", "epoch": 0, "step": 4357, "train_loss": 3.5813279151916504, "perplexity": 35.92120950989789, "lr": 0.0026291804804649314, "grad_norm": 0.179281, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:33.849049+00:00", "epoch": 0, "step": 4358, "train_loss": 3.5376148223876953, "perplexity": 34.38480743133181, "lr": 0.0026291804804649314, "grad_norm": 0.165566, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:34.153793+00:00", "epoch": 0, "step": 4359, "train_loss": 3.6810522079467773, "perplexity": 39.68813227826183, "lr": 0.0026291804804649314, "grad_norm": 0.148317, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:34.458559+00:00", "epoch": 0, "step": 4360, "train_loss": 3.670973539352417, "perplexity": 39.29013774241814, "lr": 0.0026291804804649314, "grad_norm": 0.169384, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:34.763239+00:00", "epoch": 0, "step": 4361, "train_loss": 3.594855546951294, "perplexity": 36.410440008914186, "lr": 0.0026291804804649314, "grad_norm": 0.174399, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:35.068942+00:00", "epoch": 0, "step": 4362, "train_loss": 3.5829811096191406, "perplexity": 35.98064336761323, "lr": 0.0026291804804649314, "grad_norm": 0.176247, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:35.373827+00:00", "epoch": 0, "step": 4363, "train_loss": 3.5256893634796143, "perplexity": 33.97718817741809, "lr": 0.0026291804804649314, "grad_norm": 0.174922, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:35.679530+00:00", "epoch": 0, "step": 4364, "train_loss": 3.6419458389282227, "perplexity": 38.16602946429697, "lr": 0.0026291804804649314, "grad_norm": 0.193427, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:35.983736+00:00", "epoch": 0, "step": 4365, "train_loss": 3.590130567550659, "perplexity": 36.238807229891215, "lr": 0.0026291804804649314, "grad_norm": 0.185351, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:36.289262+00:00", "epoch": 0, "step": 4366, "train_loss": 3.5938122272491455, "perplexity": 36.37247208927236, "lr": 0.0026291804804649314, "grad_norm": 0.154779, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:36.594609+00:00", "epoch": 0, "step": 4367, "train_loss": 3.626840353012085, "perplexity": 37.593845482123456, "lr": 0.0026291804804649314, "grad_norm": 0.166977, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:36.900085+00:00", "epoch": 0, "step": 4368, "train_loss": 3.6589462757110596, "perplexity": 38.8204152974244, "lr": 0.0026291804804649314, "grad_norm": 0.159341, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:37.205438+00:00", "epoch": 0, "step": 4369, "train_loss": 3.558018207550049, "perplexity": 35.09357999264698, "lr": 0.0026291804804649314, "grad_norm": 0.146601, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:37.509936+00:00", "epoch": 0, "step": 4370, "train_loss": 3.5992228984832764, "perplexity": 36.56980494791522, "lr": 0.0026291804804649314, "grad_norm": 0.155173, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:37.814991+00:00", "epoch": 0, "step": 4371, "train_loss": 3.460148572921753, "perplexity": 31.8217040070098, "lr": 0.0026291804804649314, "grad_norm": 0.151053, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:38.121027+00:00", "epoch": 0, "step": 4372, "train_loss": 3.6549339294433594, "perplexity": 38.66496641491356, "lr": 0.0026291804804649314, "grad_norm": 0.143189, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:38.426231+00:00", "epoch": 0, "step": 4373, "train_loss": 3.571665048599243, "perplexity": 35.57577926550959, "lr": 0.0026291804804649314, "grad_norm": 0.179031, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:38.731016+00:00", "epoch": 0, "step": 4374, "train_loss": 3.7301244735717773, "perplexity": 41.68429643438554, "lr": 0.0026291804804649314, "grad_norm": 0.150642, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:39.035284+00:00", "epoch": 0, "step": 4375, "train_loss": 3.7338478565216064, "perplexity": 41.83979233876038, "lr": 0.0026291804804649314, "grad_norm": 0.170397, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:39.340443+00:00", "epoch": 0, "step": 4376, "train_loss": 3.6002590656280518, "perplexity": 36.607717016521036, "lr": 0.0026291804804649314, "grad_norm": 0.152253, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:39.646796+00:00", "epoch": 0, "step": 4377, "train_loss": 3.622779607772827, "perplexity": 37.44149598864582, "lr": 0.0026291804804649314, "grad_norm": 0.197921, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:39.953991+00:00", "epoch": 0, "step": 4378, "train_loss": 3.575005292892456, "perplexity": 35.6948097439451, "lr": 0.0026291804804649314, "grad_norm": 0.18084, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:40.259414+00:00", "epoch": 0, "step": 4379, "train_loss": 3.629692316055298, "perplexity": 37.701214773941125, "lr": 0.0026291804804649314, "grad_norm": 0.180572, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:40.564795+00:00", "epoch": 0, "step": 4380, "train_loss": 3.632854700088501, "perplexity": 37.820629211210765, "lr": 0.0026291804804649314, "grad_norm": 0.166956, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:40.869953+00:00", "epoch": 0, "step": 4381, "train_loss": 3.5897860527038574, "perplexity": 36.22632457312388, "lr": 0.0026291804804649314, "grad_norm": 0.142622, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:41.177104+00:00", "epoch": 0, "step": 4382, "train_loss": 3.5335123538970947, "perplexity": 34.24403379942188, "lr": 0.0026291804804649314, "grad_norm": 0.179473, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:41.483905+00:00", "epoch": 0, "step": 4383, "train_loss": 3.6577064990997314, "perplexity": 38.772316476554025, "lr": 0.0026291804804649314, "grad_norm": 0.158032, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:41.789380+00:00", "epoch": 0, "step": 4384, "train_loss": 3.73844313621521, "perplexity": 42.032500321307516, "lr": 0.0026291804804649314, "grad_norm": 0.149124, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:42.094394+00:00", "epoch": 0, "step": 4385, "train_loss": 3.5573413372039795, "perplexity": 35.06983422632168, "lr": 0.0026291804804649314, "grad_norm": 0.176126, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:42.398076+00:00", "epoch": 0, "step": 4386, "train_loss": 3.6883482933044434, "perplexity": 39.978759209257476, "lr": 0.0026291804804649314, "grad_norm": 0.169596, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:42.703834+00:00", "epoch": 0, "step": 4387, "train_loss": 3.6201536655426025, "perplexity": 37.34330576052291, "lr": 0.0026291804804649314, "grad_norm": 0.172991, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:43.010164+00:00", "epoch": 0, "step": 4388, "train_loss": 3.655919313430786, "perplexity": 38.703085031345964, "lr": 0.0026291804804649314, "grad_norm": 0.166768, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:43.315661+00:00", "epoch": 0, "step": 4389, "train_loss": 3.766322612762451, "perplexity": 43.22083249798763, "lr": 0.0026291804804649314, "grad_norm": 0.154417, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:43.620982+00:00", "epoch": 0, "step": 4390, "train_loss": 3.6516308784484863, "perplexity": 38.537464747306856, "lr": 0.0026291804804649314, "grad_norm": 0.176272, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:43.925356+00:00", "epoch": 0, "step": 4391, "train_loss": 3.65208101272583, "perplexity": 38.5548156859847, "lr": 0.0026291804804649314, "grad_norm": 0.145174, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:44.230794+00:00", "epoch": 0, "step": 4392, "train_loss": 3.5336544513702393, "perplexity": 34.24890013583392, "lr": 0.0026291804804649314, "grad_norm": 0.150665, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:44.537424+00:00", "epoch": 0, "step": 4393, "train_loss": 3.6149160861968994, "perplexity": 37.14822854571703, "lr": 0.0026291804804649314, "grad_norm": 0.143948, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:44.842274+00:00", "epoch": 0, "step": 4394, "train_loss": 3.589538812637329, "perplexity": 36.217369081350164, "lr": 0.0026291804804649314, "grad_norm": 0.143615, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:45.147132+00:00", "epoch": 0, "step": 4395, "train_loss": 3.7289156913757324, "perplexity": 41.63393964033063, "lr": 0.0026291804804649314, "grad_norm": 0.155855, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:45.451970+00:00", "epoch": 0, "step": 4396, "train_loss": 3.6558523178100586, "perplexity": 38.700492180995994, "lr": 0.0026291804804649314, "grad_norm": 0.148355, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:45.757875+00:00", "epoch": 0, "step": 4397, "train_loss": 3.5017282962799072, "perplexity": 33.17273475768549, "lr": 0.0026291804804649314, "grad_norm": 0.164298, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:46.062231+00:00", "epoch": 0, "step": 4398, "train_loss": 3.6928467750549316, "perplexity": 40.15900804705658, "lr": 0.0026291804804649314, "grad_norm": 0.151775, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:46.367615+00:00", "epoch": 0, "step": 4399, "train_loss": 3.704230785369873, "perplexity": 40.618790728820876, "lr": 0.0026291804804649314, "grad_norm": 0.178498, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:46.673133+00:00", "epoch": 0, "step": 4400, "train_loss": 3.660367012023926, "perplexity": 38.87560806902156, "lr": 0.0026291804804649314, "grad_norm": 0.153116, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:46.978133+00:00", "epoch": 0, "step": 4401, "train_loss": 3.623353958129883, "perplexity": 37.46300670198722, "lr": 0.0026291804804649314, "grad_norm": 0.152032, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:47.283984+00:00", "epoch": 0, "step": 4402, "train_loss": 3.6391260623931885, "perplexity": 38.05856137917827, "lr": 0.0026291804804649314, "grad_norm": 0.145503, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:47.588681+00:00", "epoch": 0, "step": 4403, "train_loss": 3.5829153060913086, "perplexity": 35.97827579224424, "lr": 0.0026291804804649314, "grad_norm": 0.133864, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:47.892606+00:00", "epoch": 0, "step": 4404, "train_loss": 3.5575037002563477, "perplexity": 35.07552873392902, "lr": 0.0026291804804649314, "grad_norm": 0.142647, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:48.197441+00:00", "epoch": 0, "step": 4405, "train_loss": 3.5472354888916016, "perplexity": 34.7172085929503, "lr": 0.0026291804804649314, "grad_norm": 0.14695, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:48.502845+00:00", "epoch": 0, "step": 4406, "train_loss": 3.5614967346191406, "perplexity": 35.21586652580353, "lr": 0.0026291804804649314, "grad_norm": 0.149355, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:48.806268+00:00", "epoch": 0, "step": 4407, "train_loss": 3.628180980682373, "perplexity": 37.644278630092614, "lr": 0.0026291804804649314, "grad_norm": 0.171383, "tokens_per_sec": 107994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:49.111197+00:00", "epoch": 0, "step": 4408, "train_loss": 3.6654913425445557, "perplexity": 39.07533081961177, "lr": 0.0026291804804649314, "grad_norm": 0.175887, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:49.416866+00:00", "epoch": 0, "step": 4409, "train_loss": 3.6987099647521973, "perplexity": 40.395159553315224, "lr": 0.0026291804804649314, "grad_norm": 0.195083, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:49.721975+00:00", "epoch": 0, "step": 4410, "train_loss": 3.6375370025634766, "perplexity": 37.998132073720804, "lr": 0.0026291804804649314, "grad_norm": 0.19625, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:50.027637+00:00", "epoch": 0, "step": 4411, "train_loss": 3.642526626586914, "perplexity": 38.188202261411284, "lr": 0.0026291804804649314, "grad_norm": 0.16092, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:50.332284+00:00", "epoch": 0, "step": 4412, "train_loss": 3.6060991287231445, "perplexity": 36.822133888280526, "lr": 0.0026291804804649314, "grad_norm": 0.169856, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:50.636380+00:00", "epoch": 0, "step": 4413, "train_loss": 3.673736333847046, "perplexity": 39.39883840832379, "lr": 0.0026291804804649314, "grad_norm": 0.142487, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:50.939471+00:00", "epoch": 0, "step": 4414, "train_loss": 3.697383165359497, "perplexity": 40.3415988201833, "lr": 0.0026291804804649314, "grad_norm": 0.163121, "tokens_per_sec": 108111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:51.244387+00:00", "epoch": 0, "step": 4415, "train_loss": 3.635746717453003, "perplexity": 37.930165441621575, "lr": 0.0026291804804649314, "grad_norm": 0.153214, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:51.549939+00:00", "epoch": 0, "step": 4416, "train_loss": 3.623060703277588, "perplexity": 37.452022104212325, "lr": 0.0026291804804649314, "grad_norm": 0.149472, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:51.856033+00:00", "epoch": 0, "step": 4417, "train_loss": 3.5801076889038086, "perplexity": 35.87740423737071, "lr": 0.0026291804804649314, "grad_norm": 0.144874, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:52.161480+00:00", "epoch": 0, "step": 4418, "train_loss": 3.533487319946289, "perplexity": 34.24317654669461, "lr": 0.0026291804804649314, "grad_norm": 0.157873, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:52.466134+00:00", "epoch": 0, "step": 4419, "train_loss": 3.601357936859131, "perplexity": 36.64796629392234, "lr": 0.0026291804804649314, "grad_norm": 0.172068, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:52.770529+00:00", "epoch": 0, "step": 4420, "train_loss": 3.712620973587036, "perplexity": 40.9610237201657, "lr": 0.0026291804804649314, "grad_norm": 0.160736, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:53.075260+00:00", "epoch": 0, "step": 4421, "train_loss": 3.700117588043213, "perplexity": 40.45206075908267, "lr": 0.0026291804804649314, "grad_norm": 0.143943, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:53.381017+00:00", "epoch": 0, "step": 4422, "train_loss": 3.605738639831543, "perplexity": 36.80886231032067, "lr": 0.0026291804804649314, "grad_norm": 0.167323, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:53.685354+00:00", "epoch": 0, "step": 4423, "train_loss": 3.537097692489624, "perplexity": 34.3670306162271, "lr": 0.0026291804804649314, "grad_norm": 0.167895, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:53.990067+00:00", "epoch": 0, "step": 4424, "train_loss": 3.608504056930542, "perplexity": 36.910795045852076, "lr": 0.0026291804804649314, "grad_norm": 0.169279, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:54.295366+00:00", "epoch": 0, "step": 4425, "train_loss": 3.6579620838165283, "perplexity": 38.78222735456077, "lr": 0.0026291804804649314, "grad_norm": 0.163858, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:54.601179+00:00", "epoch": 0, "step": 4426, "train_loss": 3.6441256999969482, "perplexity": 38.24931685055907, "lr": 0.0026291804804649314, "grad_norm": 0.179317, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:54.906260+00:00", "epoch": 0, "step": 4427, "train_loss": 3.6105613708496094, "perplexity": 36.98681030507772, "lr": 0.0026291804804649314, "grad_norm": 0.166196, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:55.211056+00:00", "epoch": 0, "step": 4428, "train_loss": 3.6074016094207764, "perplexity": 36.8701252540448, "lr": 0.0026291804804649314, "grad_norm": 0.170874, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:55.516229+00:00", "epoch": 0, "step": 4429, "train_loss": 3.7276318073272705, "perplexity": 41.58052078849545, "lr": 0.0026291804804649314, "grad_norm": 0.174585, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:55.821313+00:00", "epoch": 0, "step": 4430, "train_loss": 3.6892318725585938, "perplexity": 40.014099222053325, "lr": 0.0026291804804649314, "grad_norm": 0.161726, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:56.125031+00:00", "epoch": 0, "step": 4431, "train_loss": 3.736614227294922, "perplexity": 41.9556969611115, "lr": 0.0026291804804649314, "grad_norm": 0.15163, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:56.430497+00:00", "epoch": 0, "step": 4432, "train_loss": 3.6117591857910156, "perplexity": 37.03114020330309, "lr": 0.0026291804804649314, "grad_norm": 0.148601, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:56.735053+00:00", "epoch": 0, "step": 4433, "train_loss": 3.5587613582611084, "perplexity": 35.11966950459111, "lr": 0.0026291804804649314, "grad_norm": 0.141433, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:57.038885+00:00", "epoch": 0, "step": 4434, "train_loss": 3.6081645488739014, "perplexity": 36.89826566059097, "lr": 0.0026291804804649314, "grad_norm": 0.144208, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:57.343247+00:00", "epoch": 0, "step": 4435, "train_loss": 3.671408176422119, "perplexity": 39.307218404430245, "lr": 0.0026291804804649314, "grad_norm": 0.148766, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:57.647610+00:00", "epoch": 0, "step": 4436, "train_loss": 3.551643133163452, "perplexity": 34.870567425851064, "lr": 0.0026291804804649314, "grad_norm": 0.135695, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:57.952023+00:00", "epoch": 0, "step": 4437, "train_loss": 3.575303792953491, "perplexity": 35.70546623723513, "lr": 0.0026291804804649314, "grad_norm": 0.136355, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:58.256784+00:00", "epoch": 0, "step": 4438, "train_loss": 3.598665475845337, "perplexity": 36.54942579120155, "lr": 0.0026291804804649314, "grad_norm": 0.146547, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:58.560799+00:00", "epoch": 0, "step": 4439, "train_loss": 3.6118741035461426, "perplexity": 37.035395983332705, "lr": 0.0026291804804649314, "grad_norm": 0.130217, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:58.864722+00:00", "epoch": 0, "step": 4440, "train_loss": 3.5859601497650146, "perplexity": 36.08799096569223, "lr": 0.0026291804804649314, "grad_norm": 0.141861, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:59.168826+00:00", "epoch": 0, "step": 4441, "train_loss": 3.644244432449341, "perplexity": 38.25385855536961, "lr": 0.0026291804804649314, "grad_norm": 0.125439, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:59.474850+00:00", "epoch": 0, "step": 4442, "train_loss": 3.581324338912964, "perplexity": 35.92108104587164, "lr": 0.0026291804804649314, "grad_norm": 0.148607, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:20:59.779946+00:00", "epoch": 0, "step": 4443, "train_loss": 3.628286838531494, "perplexity": 37.64826378338626, "lr": 0.0026291804804649314, "grad_norm": 0.161834, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:00.084408+00:00", "epoch": 0, "step": 4444, "train_loss": 3.6731367111206055, "perplexity": 39.375221050878565, "lr": 0.0026291804804649314, "grad_norm": 0.167075, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:00.388545+00:00", "epoch": 0, "step": 4445, "train_loss": 3.6565747261047363, "perplexity": 38.7284598383732, "lr": 0.0026291804804649314, "grad_norm": 0.171422, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:00.692140+00:00", "epoch": 0, "step": 4446, "train_loss": 3.600931167602539, "perplexity": 36.63232940550043, "lr": 0.0026291804804649314, "grad_norm": 0.175358, "tokens_per_sec": 107932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:00.997548+00:00", "epoch": 0, "step": 4447, "train_loss": 3.6285898685455322, "perplexity": 37.659674066030796, "lr": 0.0026291804804649314, "grad_norm": 0.138578, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:01.303959+00:00", "epoch": 0, "step": 4448, "train_loss": 3.5875301361083984, "perplexity": 36.14469311782388, "lr": 0.0026291804804649314, "grad_norm": 0.1693, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:01.608379+00:00", "epoch": 0, "step": 4449, "train_loss": 3.693563461303711, "perplexity": 40.187799771972806, "lr": 0.0026291804804649314, "grad_norm": 0.168292, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:01.913031+00:00", "epoch": 0, "step": 4450, "train_loss": 3.5597712993621826, "perplexity": 35.155156219009335, "lr": 0.0026291804804649314, "grad_norm": 0.162254, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:02.217343+00:00", "epoch": 0, "step": 4451, "train_loss": 3.5197601318359375, "perplexity": 33.776325626865194, "lr": 0.0026291804804649314, "grad_norm": 0.15438, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:02.521479+00:00", "epoch": 0, "step": 4452, "train_loss": 3.5544347763061523, "perplexity": 34.96804961071521, "lr": 0.0026291804804649314, "grad_norm": 0.174717, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:02.826025+00:00", "epoch": 0, "step": 4453, "train_loss": 3.65856671333313, "perplexity": 38.80568332431013, "lr": 0.0026291804804649314, "grad_norm": 0.155415, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:03.131438+00:00", "epoch": 0, "step": 4454, "train_loss": 3.6531386375427246, "perplexity": 38.59561378660388, "lr": 0.0026291804804649314, "grad_norm": 0.156513, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:03.437442+00:00", "epoch": 0, "step": 4455, "train_loss": 3.7221217155456543, "perplexity": 41.35203836027983, "lr": 0.0026291804804649314, "grad_norm": 0.185702, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:03.742145+00:00", "epoch": 0, "step": 4456, "train_loss": 3.5317325592041016, "perplexity": 34.18314065452117, "lr": 0.0026291804804649314, "grad_norm": 0.16791, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:04.046749+00:00", "epoch": 0, "step": 4457, "train_loss": 3.500375270843506, "perplexity": 33.127881554368486, "lr": 0.0026291804804649314, "grad_norm": 0.164988, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:04.352276+00:00", "epoch": 0, "step": 4458, "train_loss": 3.525495767593384, "perplexity": 33.970610970241935, "lr": 0.0026291804804649314, "grad_norm": 0.157097, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:04.657057+00:00", "epoch": 0, "step": 4459, "train_loss": 3.6382925510406494, "perplexity": 38.026852352959885, "lr": 0.0026291804804649314, "grad_norm": 0.165321, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:04.961304+00:00", "epoch": 0, "step": 4460, "train_loss": 3.5912206172943115, "perplexity": 36.27833086987572, "lr": 0.0026291804804649314, "grad_norm": 0.178718, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:05.264344+00:00", "epoch": 0, "step": 4461, "train_loss": 3.587599277496338, "perplexity": 36.1471922984701, "lr": 0.0026291804804649314, "grad_norm": 0.158974, "tokens_per_sec": 108074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:05.569051+00:00", "epoch": 0, "step": 4462, "train_loss": 3.581078052520752, "perplexity": 35.91223526175934, "lr": 0.0026291804804649314, "grad_norm": 0.171889, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:05.873732+00:00", "epoch": 0, "step": 4463, "train_loss": 3.6224403381347656, "perplexity": 37.42879538044047, "lr": 0.0026291804804649314, "grad_norm": 0.186382, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:06.179812+00:00", "epoch": 0, "step": 4464, "train_loss": 3.6415181159973145, "perplexity": 38.14970846899417, "lr": 0.0026291804804649314, "grad_norm": 0.141088, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:06.485755+00:00", "epoch": 0, "step": 4465, "train_loss": 3.597679615020752, "perplexity": 36.51341089990512, "lr": 0.0026291804804649314, "grad_norm": 0.153292, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:06.789614+00:00", "epoch": 0, "step": 4466, "train_loss": 3.637937545776367, "perplexity": 38.01335501614506, "lr": 0.0026291804804649314, "grad_norm": 0.157804, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:07.094874+00:00", "epoch": 0, "step": 4467, "train_loss": 3.620483636856079, "perplexity": 37.3556300133874, "lr": 0.0026291804804649314, "grad_norm": 0.144121, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:07.399920+00:00", "epoch": 0, "step": 4468, "train_loss": 3.467702627182007, "perplexity": 32.062997108901925, "lr": 0.0026291804804649314, "grad_norm": 0.173578, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:07.705608+00:00", "epoch": 0, "step": 4469, "train_loss": 3.6362946033477783, "perplexity": 37.95095253821206, "lr": 0.0026291804804649314, "grad_norm": 0.144517, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:08.010746+00:00", "epoch": 0, "step": 4470, "train_loss": 3.6709342002868652, "perplexity": 39.288592135515515, "lr": 0.0026291804804649314, "grad_norm": 0.147076, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:08.315576+00:00", "epoch": 0, "step": 4471, "train_loss": 3.6700427532196045, "perplexity": 39.25358404152921, "lr": 0.0026291804804649314, "grad_norm": 0.158028, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:08.620361+00:00", "epoch": 0, "step": 4472, "train_loss": 3.5398707389831543, "perplexity": 34.46246424976655, "lr": 0.0026291804804649314, "grad_norm": 0.143141, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:08.925344+00:00", "epoch": 0, "step": 4473, "train_loss": 3.6248133182525635, "perplexity": 37.51771863253851, "lr": 0.0026291804804649314, "grad_norm": 0.141602, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:09.230332+00:00", "epoch": 0, "step": 4474, "train_loss": 3.5039050579071045, "perplexity": 33.2450225418692, "lr": 0.0026291804804649314, "grad_norm": 0.157953, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:09.535744+00:00", "epoch": 0, "step": 4475, "train_loss": 3.6851580142974854, "perplexity": 39.85141904635657, "lr": 0.0026291804804649314, "grad_norm": 0.15416, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:09.840059+00:00", "epoch": 0, "step": 4476, "train_loss": 3.6294548511505127, "perplexity": 37.69226312145826, "lr": 0.0026291804804649314, "grad_norm": 0.147542, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:10.144175+00:00", "epoch": 0, "step": 4477, "train_loss": 3.578362464904785, "perplexity": 35.81484473652279, "lr": 0.0026291804804649314, "grad_norm": 0.132158, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:10.448499+00:00", "epoch": 0, "step": 4478, "train_loss": 3.597947359085083, "perplexity": 36.523188457825185, "lr": 0.0026291804804649314, "grad_norm": 0.167487, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:10.753489+00:00", "epoch": 0, "step": 4479, "train_loss": 3.638425350189209, "perplexity": 38.03190262190304, "lr": 0.0026291804804649314, "grad_norm": 0.152842, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:11.059347+00:00", "epoch": 0, "step": 4480, "train_loss": 3.54213809967041, "perplexity": 34.54069173784453, "lr": 0.0026291804804649314, "grad_norm": 0.154842, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:11.363537+00:00", "epoch": 0, "step": 4481, "train_loss": 3.5981647968292236, "perplexity": 36.53113084098793, "lr": 0.0026291804804649314, "grad_norm": 0.168358, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:11.667454+00:00", "epoch": 0, "step": 4482, "train_loss": 3.6284780502319336, "perplexity": 37.65546326021305, "lr": 0.0026291804804649314, "grad_norm": 0.155537, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:11.973098+00:00", "epoch": 0, "step": 4483, "train_loss": 3.6366238594055176, "perplexity": 37.963450176580906, "lr": 0.0026291804804649314, "grad_norm": 0.168317, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:12.277992+00:00", "epoch": 0, "step": 4484, "train_loss": 3.6073718070983887, "perplexity": 36.869026455058965, "lr": 0.0026291804804649314, "grad_norm": 0.188138, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:12.582779+00:00", "epoch": 0, "step": 4485, "train_loss": 3.673082113265991, "perplexity": 39.37307130697046, "lr": 0.0026291804804649314, "grad_norm": 0.181644, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:12.888575+00:00", "epoch": 0, "step": 4486, "train_loss": 3.573157787322998, "perplexity": 35.628924264777716, "lr": 0.0026291804804649314, "grad_norm": 0.145107, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:13.192726+00:00", "epoch": 0, "step": 4487, "train_loss": 3.633056163787842, "perplexity": 37.8282494626593, "lr": 0.0026291804804649314, "grad_norm": 0.209143, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:13.497452+00:00", "epoch": 0, "step": 4488, "train_loss": 3.5750815868377686, "perplexity": 35.69753314569586, "lr": 0.0026291804804649314, "grad_norm": 0.183004, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:13.802006+00:00", "epoch": 0, "step": 4489, "train_loss": 3.634730577468872, "perplexity": 37.89164265949595, "lr": 0.0026291804804649314, "grad_norm": 0.170235, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:14.106639+00:00", "epoch": 0, "step": 4490, "train_loss": 3.6675732135772705, "perplexity": 39.15676535762904, "lr": 0.0026291804804649314, "grad_norm": 0.163211, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:14.411137+00:00", "epoch": 0, "step": 4491, "train_loss": 3.578246593475342, "perplexity": 35.81069505968701, "lr": 0.0026291804804649314, "grad_norm": 0.145818, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:14.716181+00:00", "epoch": 0, "step": 4492, "train_loss": 3.6163330078125, "perplexity": 37.200901981977374, "lr": 0.0026291804804649314, "grad_norm": 0.161667, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:15.021246+00:00", "epoch": 0, "step": 4493, "train_loss": 3.5938262939453125, "perplexity": 36.372983733384636, "lr": 0.0026291804804649314, "grad_norm": 0.164129, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:15.325133+00:00", "epoch": 0, "step": 4494, "train_loss": 3.595730781555176, "perplexity": 36.442321635868915, "lr": 0.0026291804804649314, "grad_norm": 0.2214, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:15.629474+00:00", "epoch": 0, "step": 4495, "train_loss": 3.7313356399536133, "perplexity": 41.73481363907195, "lr": 0.0026291804804649314, "grad_norm": 0.183426, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:15.934888+00:00", "epoch": 0, "step": 4496, "train_loss": 3.5154426097869873, "perplexity": 33.63080995584409, "lr": 0.0026291804804649314, "grad_norm": 0.168654, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:16.240555+00:00", "epoch": 0, "step": 4497, "train_loss": 3.6325292587280273, "perplexity": 37.80832281679978, "lr": 0.0026291804804649314, "grad_norm": 0.160798, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:16.545199+00:00", "epoch": 0, "step": 4498, "train_loss": 3.672131061553955, "perplexity": 39.335643280911114, "lr": 0.0026291804804649314, "grad_norm": 0.149621, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:16.849945+00:00", "epoch": 0, "step": 4499, "train_loss": 3.648268699645996, "perplexity": 38.40811247481853, "lr": 0.0026291804804649314, "grad_norm": 0.155556, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:17.155637+00:00", "epoch": 0, "step": 4500, "train_loss": 3.724403142929077, "perplexity": 41.44648773167974, "lr": 0.0026291804804649314, "grad_norm": 0.156823, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:21:20.229116+00:00", "step": 4500, "epoch": 0, "val_loss": 3.575733006000519, "val_ppl": 35.72079477856399, "eval_train_loss": 3.724403142929077, "eval_train_ppl": 41.44648773167974} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:21:21.132328+00:00", "step": 4500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5757_epoch_0000_step_0004500.pt", "val_loss": 3.575733006000519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:22.173654+00:00", "epoch": 0, "step": 4501, "train_loss": 3.6073215007781982, "perplexity": 36.86717175666092, "lr": 0.0026291804804649314, "grad_norm": 0.178394, "tokens_per_sec": 6530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:22.478675+00:00", "epoch": 0, "step": 4502, "train_loss": 3.5733895301818848, "perplexity": 35.63718197034108, "lr": 0.0026291804804649314, "grad_norm": 0.146, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:22.782392+00:00", "epoch": 0, "step": 4503, "train_loss": 3.6653318405151367, "perplexity": 39.06909872207514, "lr": 0.0026291804804649314, "grad_norm": 0.154141, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:23.086311+00:00", "epoch": 0, "step": 4504, "train_loss": 3.6709556579589844, "perplexity": 39.2894351862885, "lr": 0.0026291804804649314, "grad_norm": 0.159463, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:23.391592+00:00", "epoch": 0, "step": 4505, "train_loss": 3.640918016433716, "perplexity": 38.12682171344335, "lr": 0.0026291804804649314, "grad_norm": 0.152763, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:23.695959+00:00", "epoch": 0, "step": 4506, "train_loss": 3.5263054370880127, "perplexity": 33.998127075630805, "lr": 0.0026291804804649314, "grad_norm": 0.148069, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:24.001495+00:00", "epoch": 0, "step": 4507, "train_loss": 3.578742027282715, "perplexity": 35.82844128436194, "lr": 0.0026291804804649314, "grad_norm": 0.153073, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:24.307025+00:00", "epoch": 0, "step": 4508, "train_loss": 3.673149347305298, "perplexity": 39.37571860658766, "lr": 0.0026291804804649314, "grad_norm": 0.134851, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:24.612608+00:00", "epoch": 0, "step": 4509, "train_loss": 3.596506357192993, "perplexity": 36.47059637589765, "lr": 0.0026291804804649314, "grad_norm": 0.156738, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:24.918344+00:00", "epoch": 0, "step": 4510, "train_loss": 3.661299228668213, "perplexity": 38.91186545516398, "lr": 0.0026291804804649314, "grad_norm": 0.150029, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:25.223198+00:00", "epoch": 0, "step": 4511, "train_loss": 3.639732837677002, "perplexity": 38.081661381107025, "lr": 0.0026291804804649314, "grad_norm": 0.146346, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:25.529195+00:00", "epoch": 0, "step": 4512, "train_loss": 3.642029047012329, "perplexity": 38.169205318613194, "lr": 0.0026291804804649314, "grad_norm": 0.16341, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:25.834538+00:00", "epoch": 0, "step": 4513, "train_loss": 3.6542282104492188, "perplexity": 38.637689439778555, "lr": 0.0026291804804649314, "grad_norm": 0.195898, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:26.141250+00:00", "epoch": 0, "step": 4514, "train_loss": 3.61765193939209, "perplexity": 37.24999979760292, "lr": 0.0026291804804649314, "grad_norm": 0.151055, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:26.528708+00:00", "epoch": 0, "step": 4515, "train_loss": 3.5029778480529785, "perplexity": 33.21421171562597, "lr": 0.0026291804804649314, "grad_norm": 0.167801, "tokens_per_sec": 84572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:26.834674+00:00", "epoch": 0, "step": 4516, "train_loss": 3.6371583938598633, "perplexity": 37.983748373265556, "lr": 0.0026291804804649314, "grad_norm": 0.152725, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:27.140633+00:00", "epoch": 0, "step": 4517, "train_loss": 3.633862257003784, "perplexity": 37.858754851363734, "lr": 0.0026291804804649314, "grad_norm": 0.178625, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:27.446155+00:00", "epoch": 0, "step": 4518, "train_loss": 3.652914047241211, "perplexity": 38.586946559390455, "lr": 0.0026291804804649314, "grad_norm": 0.157434, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:27.752077+00:00", "epoch": 0, "step": 4519, "train_loss": 3.605790615081787, "perplexity": 36.81077550986953, "lr": 0.0026291804804649314, "grad_norm": 0.16357, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:28.056561+00:00", "epoch": 0, "step": 4520, "train_loss": 3.729901075363159, "perplexity": 41.67498527732121, "lr": 0.0026291804804649314, "grad_norm": 0.173326, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:28.361685+00:00", "epoch": 0, "step": 4521, "train_loss": 3.6027941703796387, "perplexity": 36.700639147776954, "lr": 0.0026291804804649314, "grad_norm": 0.14333, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:28.666516+00:00", "epoch": 0, "step": 4522, "train_loss": 3.642965793609619, "perplexity": 38.204976943674914, "lr": 0.0026291804804649314, "grad_norm": 0.170338, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:28.970830+00:00", "epoch": 0, "step": 4523, "train_loss": 3.633474349975586, "perplexity": 37.844072022248476, "lr": 0.0026291804804649314, "grad_norm": 0.148285, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:29.275290+00:00", "epoch": 0, "step": 4524, "train_loss": 3.674859046936035, "perplexity": 39.44309684000714, "lr": 0.0026291804804649314, "grad_norm": 0.157861, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:29.580246+00:00", "epoch": 0, "step": 4525, "train_loss": 3.6801860332489014, "perplexity": 39.65377030616022, "lr": 0.0026291804804649314, "grad_norm": 0.168334, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:29.885037+00:00", "epoch": 0, "step": 4526, "train_loss": 3.6764838695526123, "perplexity": 39.50723696987636, "lr": 0.0026291804804649314, "grad_norm": 0.155643, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:30.190550+00:00", "epoch": 0, "step": 4527, "train_loss": 3.614485740661621, "perplexity": 37.13224541080093, "lr": 0.0026291804804649314, "grad_norm": 0.153573, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:30.495458+00:00", "epoch": 0, "step": 4528, "train_loss": 3.63295578956604, "perplexity": 37.824452672110496, "lr": 0.0026291804804649314, "grad_norm": 0.166397, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:30.799859+00:00", "epoch": 0, "step": 4529, "train_loss": 3.579817295074463, "perplexity": 35.866987173166024, "lr": 0.0026291804804649314, "grad_norm": 0.17946, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:31.104321+00:00", "epoch": 0, "step": 4530, "train_loss": 3.597837448120117, "perplexity": 36.51917437953781, "lr": 0.0026291804804649314, "grad_norm": 0.148399, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:31.420444+00:00", "epoch": 0, "step": 4531, "train_loss": 3.7100136280059814, "perplexity": 40.85436328670374, "lr": 0.0026291804804649314, "grad_norm": 0.149024, "tokens_per_sec": 103575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:31.724446+00:00", "epoch": 0, "step": 4532, "train_loss": 3.691504716873169, "perplexity": 40.10514847116177, "lr": 0.0026291804804649314, "grad_norm": 0.152008, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:32.029649+00:00", "epoch": 0, "step": 4533, "train_loss": 3.6564478874206543, "perplexity": 38.72354788301033, "lr": 0.0026291804804649314, "grad_norm": 0.165397, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:32.334493+00:00", "epoch": 0, "step": 4534, "train_loss": 3.6178417205810547, "perplexity": 37.25706981771064, "lr": 0.0026291804804649314, "grad_norm": 0.163473, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:32.638213+00:00", "epoch": 0, "step": 4535, "train_loss": 3.5994210243225098, "perplexity": 36.57705108901132, "lr": 0.0026291804804649314, "grad_norm": 0.157105, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:32.943083+00:00", "epoch": 0, "step": 4536, "train_loss": 3.534219264984131, "perplexity": 34.268249844866595, "lr": 0.0026291804804649314, "grad_norm": 0.156937, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:33.248458+00:00", "epoch": 0, "step": 4537, "train_loss": 3.548032522201538, "perplexity": 34.74489039482055, "lr": 0.0026291804804649314, "grad_norm": 0.170788, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:33.553796+00:00", "epoch": 0, "step": 4538, "train_loss": 3.5798332691192627, "perplexity": 35.86756011860208, "lr": 0.0026291804804649314, "grad_norm": 0.153171, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:33.857922+00:00", "epoch": 0, "step": 4539, "train_loss": 3.563119888305664, "perplexity": 35.27307370484081, "lr": 0.0026291804804649314, "grad_norm": 0.148959, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:34.163144+00:00", "epoch": 0, "step": 4540, "train_loss": 3.552558660507202, "perplexity": 34.90250700236838, "lr": 0.0026291804804649314, "grad_norm": 0.152384, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:34.471193+00:00", "epoch": 0, "step": 4541, "train_loss": 3.5924017429351807, "perplexity": 36.32120545182372, "lr": 0.0026291804804649314, "grad_norm": 0.142832, "tokens_per_sec": 106307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:34.775964+00:00", "epoch": 0, "step": 4542, "train_loss": 3.737616539001465, "perplexity": 41.99777072932462, "lr": 0.0026291804804649314, "grad_norm": 0.165741, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:35.082185+00:00", "epoch": 0, "step": 4543, "train_loss": 3.493461847305298, "perplexity": 32.89964433759011, "lr": 0.0026291804804649314, "grad_norm": 0.156402, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:35.387621+00:00", "epoch": 0, "step": 4544, "train_loss": 3.721615791320801, "perplexity": 41.33112265365375, "lr": 0.0026291804804649314, "grad_norm": 0.152233, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:35.691666+00:00", "epoch": 0, "step": 4545, "train_loss": 3.7065236568450928, "perplexity": 40.712031248856135, "lr": 0.0026291804804649314, "grad_norm": 0.162561, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:35.996365+00:00", "epoch": 0, "step": 4546, "train_loss": 3.636934995651245, "perplexity": 37.97526381967469, "lr": 0.0026291804804649314, "grad_norm": 0.152933, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:36.300978+00:00", "epoch": 0, "step": 4547, "train_loss": 3.5389468669891357, "perplexity": 34.430640047209664, "lr": 0.0026291804804649314, "grad_norm": 0.145573, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:36.605759+00:00", "epoch": 0, "step": 4548, "train_loss": 3.5149612426757812, "perplexity": 33.61462508575273, "lr": 0.0026291804804649314, "grad_norm": 0.148818, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:36.911521+00:00", "epoch": 0, "step": 4549, "train_loss": 3.616238594055176, "perplexity": 37.197389870843814, "lr": 0.0026291804804649314, "grad_norm": 0.15264, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:37.216494+00:00", "epoch": 0, "step": 4550, "train_loss": 3.669078826904297, "perplexity": 39.21576470935453, "lr": 0.0026291804804649314, "grad_norm": 0.146996, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:37.520076+00:00", "epoch": 0, "step": 4551, "train_loss": 3.5366556644439697, "perplexity": 34.35184278182054, "lr": 0.0026291804804649314, "grad_norm": 0.179816, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:37.825213+00:00", "epoch": 0, "step": 4552, "train_loss": 3.590108633041382, "perplexity": 36.23801235815542, "lr": 0.0026291804804649314, "grad_norm": 0.166071, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:38.130458+00:00", "epoch": 0, "step": 4553, "train_loss": 3.603530168533325, "perplexity": 36.72766069311279, "lr": 0.0026291804804649314, "grad_norm": 0.134716, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:38.435181+00:00", "epoch": 0, "step": 4554, "train_loss": 3.518918991088867, "perplexity": 33.747926928419716, "lr": 0.0026291804804649314, "grad_norm": 0.153076, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:38.739520+00:00", "epoch": 0, "step": 4555, "train_loss": 3.6727538108825684, "perplexity": 39.360147155448665, "lr": 0.0026291804804649314, "grad_norm": 0.151858, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:39.043740+00:00", "epoch": 0, "step": 4556, "train_loss": 3.6069090366363525, "perplexity": 36.85196850591346, "lr": 0.0026291804804649314, "grad_norm": 0.164948, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:39.348204+00:00", "epoch": 0, "step": 4557, "train_loss": 3.5948827266693115, "perplexity": 36.41142964785551, "lr": 0.0026291804804649314, "grad_norm": 0.163663, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:39.653007+00:00", "epoch": 0, "step": 4558, "train_loss": 3.708791971206665, "perplexity": 40.804483750053876, "lr": 0.0026291804804649314, "grad_norm": 0.172217, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:39.959109+00:00", "epoch": 0, "step": 4559, "train_loss": 3.6026835441589355, "perplexity": 36.69657931933652, "lr": 0.0026291804804649314, "grad_norm": 0.167013, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:40.264439+00:00", "epoch": 0, "step": 4560, "train_loss": 3.6975338459014893, "perplexity": 40.34767797215181, "lr": 0.0026291804804649314, "grad_norm": 0.140233, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:40.568558+00:00", "epoch": 0, "step": 4561, "train_loss": 3.5794284343719482, "perplexity": 35.853042622756334, "lr": 0.0026291804804649314, "grad_norm": 0.165972, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:40.872660+00:00", "epoch": 0, "step": 4562, "train_loss": 3.6166210174560547, "perplexity": 37.21161774354437, "lr": 0.0026291804804649314, "grad_norm": 0.169502, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:41.177594+00:00", "epoch": 0, "step": 4563, "train_loss": 3.6249866485595703, "perplexity": 37.52422215383967, "lr": 0.0026291804804649314, "grad_norm": 0.173366, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:41.483042+00:00", "epoch": 0, "step": 4564, "train_loss": 3.559379816055298, "perplexity": 35.141396255772364, "lr": 0.0026291804804649314, "grad_norm": 0.162959, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:41.787981+00:00", "epoch": 0, "step": 4565, "train_loss": 3.5586001873016357, "perplexity": 35.11400968987183, "lr": 0.0026291804804649314, "grad_norm": 0.155451, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:42.093954+00:00", "epoch": 0, "step": 4566, "train_loss": 3.562404155731201, "perplexity": 35.247836649557456, "lr": 0.0026291804804649314, "grad_norm": 0.153679, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:42.399931+00:00", "epoch": 0, "step": 4567, "train_loss": 3.465078353881836, "perplexity": 31.978965351040728, "lr": 0.0026291804804649314, "grad_norm": 0.158556, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:42.704145+00:00", "epoch": 0, "step": 4568, "train_loss": 3.5794901847839355, "perplexity": 35.855256631266556, "lr": 0.0026291804804649314, "grad_norm": 0.151534, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:43.008949+00:00", "epoch": 0, "step": 4569, "train_loss": 3.6692593097686768, "perplexity": 39.22284312164497, "lr": 0.0026291804804649314, "grad_norm": 0.162959, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:43.312584+00:00", "epoch": 0, "step": 4570, "train_loss": 3.6283700466156006, "perplexity": 37.65139655361971, "lr": 0.0026291804804649314, "grad_norm": 0.166964, "tokens_per_sec": 107978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:43.617195+00:00", "epoch": 0, "step": 4571, "train_loss": 3.5473973751068115, "perplexity": 34.72282928539611, "lr": 0.0026291804804649314, "grad_norm": 0.164821, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:43.921356+00:00", "epoch": 0, "step": 4572, "train_loss": 3.589139938354492, "perplexity": 36.20292578495284, "lr": 0.0026291804804649314, "grad_norm": 0.172936, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:44.225427+00:00", "epoch": 0, "step": 4573, "train_loss": 3.657243251800537, "perplexity": 38.75435946525238, "lr": 0.0026291804804649314, "grad_norm": 0.157034, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:44.530762+00:00", "epoch": 0, "step": 4574, "train_loss": 3.641782760620117, "perplexity": 38.15980592026114, "lr": 0.0026291804804649314, "grad_norm": 0.144128, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:44.836760+00:00", "epoch": 0, "step": 4575, "train_loss": 3.595283269882202, "perplexity": 36.42601692009403, "lr": 0.0026291804804649314, "grad_norm": 0.140266, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:45.142162+00:00", "epoch": 0, "step": 4576, "train_loss": 3.6625123023986816, "perplexity": 38.95909705886806, "lr": 0.0026291804804649314, "grad_norm": 0.173627, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:45.447517+00:00", "epoch": 0, "step": 4577, "train_loss": 3.6799533367156982, "perplexity": 39.64454408477803, "lr": 0.0026291804804649314, "grad_norm": 0.137814, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:45.752357+00:00", "epoch": 0, "step": 4578, "train_loss": 3.5884835720062256, "perplexity": 36.17917119946689, "lr": 0.0026291804804649314, "grad_norm": 0.147437, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:46.056653+00:00", "epoch": 0, "step": 4579, "train_loss": 3.6050634384155273, "perplexity": 36.78401730300278, "lr": 0.0026291804804649314, "grad_norm": 0.181442, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:46.362669+00:00", "epoch": 0, "step": 4580, "train_loss": 3.583425283432007, "perplexity": 35.9966285770101, "lr": 0.0026291804804649314, "grad_norm": 0.163711, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:46.667038+00:00", "epoch": 0, "step": 4581, "train_loss": 3.5536997318267822, "perplexity": 34.94235598303094, "lr": 0.0026291804804649314, "grad_norm": 0.144259, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:46.971860+00:00", "epoch": 0, "step": 4582, "train_loss": 3.492767810821533, "perplexity": 32.87681870594374, "lr": 0.0026291804804649314, "grad_norm": 0.143972, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:47.277470+00:00", "epoch": 0, "step": 4583, "train_loss": 3.5403525829315186, "perplexity": 34.47907378089069, "lr": 0.0026291804804649314, "grad_norm": 0.165094, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:47.583148+00:00", "epoch": 0, "step": 4584, "train_loss": 3.533855438232422, "perplexity": 34.255784406606246, "lr": 0.0026291804804649314, "grad_norm": 0.170458, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:47.889252+00:00", "epoch": 0, "step": 4585, "train_loss": 3.5717391967773438, "perplexity": 35.57841724252602, "lr": 0.0026291804804649314, "grad_norm": 0.154002, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:48.194504+00:00", "epoch": 0, "step": 4586, "train_loss": 3.642186164855957, "perplexity": 38.17520285299334, "lr": 0.0026291804804649314, "grad_norm": 0.170219, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:48.497489+00:00", "epoch": 0, "step": 4587, "train_loss": 3.544651985168457, "perplexity": 34.62763231543946, "lr": 0.0026291804804649314, "grad_norm": 0.157709, "tokens_per_sec": 108150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:48.801195+00:00", "epoch": 0, "step": 4588, "train_loss": 3.6087636947631836, "perplexity": 36.92037972889818, "lr": 0.0026291804804649314, "grad_norm": 0.148366, "tokens_per_sec": 107895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:49.106790+00:00", "epoch": 0, "step": 4589, "train_loss": 3.571977138519287, "perplexity": 35.586883840338864, "lr": 0.0026291804804649314, "grad_norm": 0.170691, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:49.411951+00:00", "epoch": 0, "step": 4590, "train_loss": 3.720217227935791, "perplexity": 41.2733588614268, "lr": 0.0026291804804649314, "grad_norm": 0.176921, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:49.717834+00:00", "epoch": 0, "step": 4591, "train_loss": 3.6318366527557373, "perplexity": 37.78214561290492, "lr": 0.0026291804804649314, "grad_norm": 0.186306, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:50.022416+00:00", "epoch": 0, "step": 4592, "train_loss": 3.7206907272338867, "perplexity": 41.29290639538432, "lr": 0.0026291804804649314, "grad_norm": 0.165908, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:50.327947+00:00", "epoch": 0, "step": 4593, "train_loss": 3.5572633743286133, "perplexity": 35.067100187784916, "lr": 0.0026291804804649314, "grad_norm": 0.161622, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:50.633362+00:00", "epoch": 0, "step": 4594, "train_loss": 3.6668100357055664, "perplexity": 39.12689318112253, "lr": 0.0026291804804649314, "grad_norm": 0.166779, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:50.939769+00:00", "epoch": 0, "step": 4595, "train_loss": 3.5789389610290527, "perplexity": 35.83549780834057, "lr": 0.0026291804804649314, "grad_norm": 0.165249, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:51.245408+00:00", "epoch": 0, "step": 4596, "train_loss": 3.6012487411499023, "perplexity": 36.64396471173286, "lr": 0.0026291804804649314, "grad_norm": 0.156285, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:51.550241+00:00", "epoch": 0, "step": 4597, "train_loss": 3.560917615890503, "perplexity": 35.195478262135715, "lr": 0.0026291804804649314, "grad_norm": 0.150756, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:51.855311+00:00", "epoch": 0, "step": 4598, "train_loss": 3.6097729206085205, "perplexity": 36.95765953904996, "lr": 0.0026291804804649314, "grad_norm": 0.148404, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:52.160052+00:00", "epoch": 0, "step": 4599, "train_loss": 3.67914080619812, "perplexity": 39.612344766089734, "lr": 0.0026291804804649314, "grad_norm": 0.146854, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:52.466770+00:00", "epoch": 0, "step": 4600, "train_loss": 3.5401434898376465, "perplexity": 34.471865198338676, "lr": 0.0026291804804649314, "grad_norm": 0.169158, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:52.771975+00:00", "epoch": 0, "step": 4601, "train_loss": 3.678520679473877, "perplexity": 39.587787707521464, "lr": 0.0026291804804649314, "grad_norm": 0.163112, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:53.077095+00:00", "epoch": 0, "step": 4602, "train_loss": 3.553982973098755, "perplexity": 34.95225450215369, "lr": 0.0026291804804649314, "grad_norm": 0.152558, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:53.382459+00:00", "epoch": 0, "step": 4603, "train_loss": 3.6266229152679443, "perplexity": 37.5856720498068, "lr": 0.0026291804804649314, "grad_norm": 0.143997, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:53.687234+00:00", "epoch": 0, "step": 4604, "train_loss": 3.485853672027588, "perplexity": 32.65028785363198, "lr": 0.0026291804804649314, "grad_norm": 0.148153, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:53.993448+00:00", "epoch": 0, "step": 4605, "train_loss": 3.58304762840271, "perplexity": 35.983036835846505, "lr": 0.0026291804804649314, "grad_norm": 0.160685, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:54.298808+00:00", "epoch": 0, "step": 4606, "train_loss": 3.625972270965576, "perplexity": 37.56122510043352, "lr": 0.0026291804804649314, "grad_norm": 0.16556, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:54.602611+00:00", "epoch": 0, "step": 4607, "train_loss": 3.5611793994903564, "perplexity": 35.20469306722354, "lr": 0.0026291804804649314, "grad_norm": 0.16124, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:54.907587+00:00", "epoch": 0, "step": 4608, "train_loss": 3.631931781768799, "perplexity": 37.78573996208915, "lr": 0.0026291804804649314, "grad_norm": 0.183465, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:55.212847+00:00", "epoch": 0, "step": 4609, "train_loss": 3.6025500297546387, "perplexity": 36.69168012447279, "lr": 0.0026291804804649314, "grad_norm": 0.153463, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:55.516093+00:00", "epoch": 0, "step": 4610, "train_loss": 3.544542074203491, "perplexity": 34.6238265681077, "lr": 0.0026291804804649314, "grad_norm": 0.146604, "tokens_per_sec": 108125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:55.820653+00:00", "epoch": 0, "step": 4611, "train_loss": 3.5465598106384277, "perplexity": 34.69375885322488, "lr": 0.0026291804804649314, "grad_norm": 0.135685, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:56.127006+00:00", "epoch": 0, "step": 4612, "train_loss": 3.632225275039673, "perplexity": 37.796831450061944, "lr": 0.0026291804804649314, "grad_norm": 0.148605, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:56.431638+00:00", "epoch": 0, "step": 4613, "train_loss": 3.661193609237671, "perplexity": 38.9077558231256, "lr": 0.0026291804804649314, "grad_norm": 0.189581, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:56.736575+00:00", "epoch": 0, "step": 4614, "train_loss": 3.5641636848449707, "perplexity": 35.309910838996664, "lr": 0.0026291804804649314, "grad_norm": 0.162499, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:57.041113+00:00", "epoch": 0, "step": 4615, "train_loss": 3.6049208641052246, "perplexity": 36.77877322095021, "lr": 0.0026291804804649314, "grad_norm": 0.150588, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:57.345059+00:00", "epoch": 0, "step": 4616, "train_loss": 3.6423425674438477, "perplexity": 38.181174020453646, "lr": 0.0026291804804649314, "grad_norm": 0.178316, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:57.649253+00:00", "epoch": 0, "step": 4617, "train_loss": 3.5502302646636963, "perplexity": 34.821334687448946, "lr": 0.0026291804804649314, "grad_norm": 0.196749, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:57.954617+00:00", "epoch": 0, "step": 4618, "train_loss": 3.5866477489471436, "perplexity": 36.11281357178874, "lr": 0.0026291804804649314, "grad_norm": 0.143281, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:58.259334+00:00", "epoch": 0, "step": 4619, "train_loss": 3.5877437591552734, "perplexity": 36.15241528208283, "lr": 0.0026291804804649314, "grad_norm": 0.185586, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:58.563916+00:00", "epoch": 0, "step": 4620, "train_loss": 3.6888859272003174, "perplexity": 40.00025892429326, "lr": 0.0026291804804649314, "grad_norm": 0.183213, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:58.868191+00:00", "epoch": 0, "step": 4621, "train_loss": 3.653857469558716, "perplexity": 38.62336752341273, "lr": 0.0026291804804649314, "grad_norm": 0.173632, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:59.172843+00:00", "epoch": 0, "step": 4622, "train_loss": 3.7288122177124023, "perplexity": 41.629631846952606, "lr": 0.0026291804804649314, "grad_norm": 0.150841, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:59.478309+00:00", "epoch": 0, "step": 4623, "train_loss": 3.689699649810791, "perplexity": 40.032821285972986, "lr": 0.0026291804804649314, "grad_norm": 0.172934, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:21:59.783377+00:00", "epoch": 0, "step": 4624, "train_loss": 3.6139261722564697, "perplexity": 37.111473191737474, "lr": 0.0026291804804649314, "grad_norm": 0.183948, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:00.087313+00:00", "epoch": 0, "step": 4625, "train_loss": 3.6441709995269775, "perplexity": 38.25104956588164, "lr": 0.0026291804804649314, "grad_norm": 0.164817, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:00.391623+00:00", "epoch": 0, "step": 4626, "train_loss": 3.586313009262085, "perplexity": 36.10072720295365, "lr": 0.0026291804804649314, "grad_norm": 0.163166, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:00.695891+00:00", "epoch": 0, "step": 4627, "train_loss": 3.6440298557281494, "perplexity": 38.24565104842934, "lr": 0.0026291804804649314, "grad_norm": 0.162117, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:01.001282+00:00", "epoch": 0, "step": 4628, "train_loss": 3.520538091659546, "perplexity": 33.80261247492207, "lr": 0.0026291804804649314, "grad_norm": 0.173259, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:01.307228+00:00", "epoch": 0, "step": 4629, "train_loss": 3.5645742416381836, "perplexity": 35.324410539030865, "lr": 0.0026291804804649314, "grad_norm": 0.150299, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:01.611962+00:00", "epoch": 0, "step": 4630, "train_loss": 3.604034185409546, "perplexity": 36.7461767197292, "lr": 0.0026291804804649314, "grad_norm": 0.154858, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:01.915268+00:00", "epoch": 0, "step": 4631, "train_loss": 3.4599742889404297, "perplexity": 31.816158477005526, "lr": 0.0026291804804649314, "grad_norm": 0.179951, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:02.218502+00:00", "epoch": 0, "step": 4632, "train_loss": 3.6363778114318848, "perplexity": 37.95411049564475, "lr": 0.0026291804804649314, "grad_norm": 0.171012, "tokens_per_sec": 108062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:02.523382+00:00", "epoch": 0, "step": 4633, "train_loss": 3.6173436641693115, "perplexity": 37.23851831543598, "lr": 0.0026291804804649314, "grad_norm": 0.174023, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:02.828066+00:00", "epoch": 0, "step": 4634, "train_loss": 3.570047378540039, "perplexity": 35.5182759158223, "lr": 0.0026291804804649314, "grad_norm": 0.16305, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:03.133395+00:00", "epoch": 0, "step": 4635, "train_loss": 3.5866851806640625, "perplexity": 36.114165361703265, "lr": 0.0026291804804649314, "grad_norm": 0.180463, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:03.437512+00:00", "epoch": 0, "step": 4636, "train_loss": 3.633037805557251, "perplexity": 37.827555009307304, "lr": 0.0026291804804649314, "grad_norm": 0.171871, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:03.741583+00:00", "epoch": 0, "step": 4637, "train_loss": 3.5706467628479004, "perplexity": 35.53957139449885, "lr": 0.0026291804804649314, "grad_norm": 0.155837, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:04.046243+00:00", "epoch": 0, "step": 4638, "train_loss": 3.5813562870025635, "perplexity": 35.92222867411957, "lr": 0.0026291804804649314, "grad_norm": 0.158909, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:04.350917+00:00", "epoch": 0, "step": 4639, "train_loss": 3.601064920425415, "perplexity": 36.63722941065439, "lr": 0.0026291804804649314, "grad_norm": 0.14406, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:04.655971+00:00", "epoch": 0, "step": 4640, "train_loss": 3.6137351989746094, "perplexity": 37.10438656860675, "lr": 0.0026291804804649314, "grad_norm": 0.148886, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:04.959866+00:00", "epoch": 0, "step": 4641, "train_loss": 3.5893115997314453, "perplexity": 36.20914096248052, "lr": 0.0026291804804649314, "grad_norm": 0.146297, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:05.264031+00:00", "epoch": 0, "step": 4642, "train_loss": 3.6360509395599365, "perplexity": 37.94170639188566, "lr": 0.0026291804804649314, "grad_norm": 0.15345, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:05.568110+00:00", "epoch": 0, "step": 4643, "train_loss": 3.605454444885254, "perplexity": 36.79840290399903, "lr": 0.0026291804804649314, "grad_norm": 0.143912, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:05.872762+00:00", "epoch": 0, "step": 4644, "train_loss": 3.5988025665283203, "perplexity": 36.554436720413925, "lr": 0.0026291804804649314, "grad_norm": 0.162197, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:06.178975+00:00", "epoch": 0, "step": 4645, "train_loss": 3.554746389389038, "perplexity": 34.978947810379516, "lr": 0.0026291804804649314, "grad_norm": 0.15947, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:06.483274+00:00", "epoch": 0, "step": 4646, "train_loss": 3.635636806488037, "perplexity": 37.92599672963435, "lr": 0.0026291804804649314, "grad_norm": 0.143462, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:06.787507+00:00", "epoch": 0, "step": 4647, "train_loss": 3.548792839050293, "perplexity": 34.771317565634405, "lr": 0.0026291804804649314, "grad_norm": 0.158825, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:07.091259+00:00", "epoch": 0, "step": 4648, "train_loss": 3.634119987487793, "perplexity": 37.86851346406752, "lr": 0.0026291804804649314, "grad_norm": 0.177836, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:07.396565+00:00", "epoch": 0, "step": 4649, "train_loss": 3.690086841583252, "perplexity": 40.04832466620026, "lr": 0.0026291804804649314, "grad_norm": 0.16441, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:07.701596+00:00", "epoch": 0, "step": 4650, "train_loss": 3.599100351333618, "perplexity": 36.56532369714316, "lr": 0.0026291804804649314, "grad_norm": 0.153396, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:08.007815+00:00", "epoch": 0, "step": 4651, "train_loss": 3.5652477741241455, "perplexity": 35.34821069126447, "lr": 0.0026291804804649314, "grad_norm": 0.164746, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:08.312008+00:00", "epoch": 0, "step": 4652, "train_loss": 3.620917558670044, "perplexity": 37.37184295344439, "lr": 0.0026291804804649314, "grad_norm": 0.160954, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:08.616513+00:00", "epoch": 0, "step": 4653, "train_loss": 3.645493268966675, "perplexity": 38.301661213502314, "lr": 0.0026291804804649314, "grad_norm": 0.182292, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:08.921094+00:00", "epoch": 0, "step": 4654, "train_loss": 3.6414339542388916, "perplexity": 38.14649785755335, "lr": 0.0026291804804649314, "grad_norm": 0.159738, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:09.227143+00:00", "epoch": 0, "step": 4655, "train_loss": 3.5857021808624268, "perplexity": 36.07868258695395, "lr": 0.0026291804804649314, "grad_norm": 0.161522, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:09.532332+00:00", "epoch": 0, "step": 4656, "train_loss": 3.5937376022338867, "perplexity": 36.36975789426237, "lr": 0.0026291804804649314, "grad_norm": 0.155103, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:09.837605+00:00", "epoch": 0, "step": 4657, "train_loss": 3.6550941467285156, "perplexity": 38.67116170714641, "lr": 0.0026291804804649314, "grad_norm": 0.159371, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:10.142747+00:00", "epoch": 0, "step": 4658, "train_loss": 3.619690179824829, "perplexity": 37.326001682054844, "lr": 0.0026291804804649314, "grad_norm": 0.155968, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:10.447644+00:00", "epoch": 0, "step": 4659, "train_loss": 3.674699544906616, "perplexity": 39.436806087721806, "lr": 0.0026291804804649314, "grad_norm": 0.148707, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:10.751856+00:00", "epoch": 0, "step": 4660, "train_loss": 3.585134267807007, "perplexity": 36.05819884913208, "lr": 0.0026291804804649314, "grad_norm": 0.1597, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:11.056043+00:00", "epoch": 0, "step": 4661, "train_loss": 3.5403079986572266, "perplexity": 34.47753659067542, "lr": 0.0026291804804649314, "grad_norm": 0.165573, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:11.361543+00:00", "epoch": 0, "step": 4662, "train_loss": 3.6470863819122314, "perplexity": 38.362728716616935, "lr": 0.0026291804804649314, "grad_norm": 0.14745, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:11.667152+00:00", "epoch": 0, "step": 4663, "train_loss": 3.6470413208007812, "perplexity": 38.36100008836995, "lr": 0.0026291804804649314, "grad_norm": 0.162217, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:11.972276+00:00", "epoch": 0, "step": 4664, "train_loss": 3.5738346576690674, "perplexity": 35.65304859067425, "lr": 0.0026291804804649314, "grad_norm": 0.168058, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:12.277734+00:00", "epoch": 0, "step": 4665, "train_loss": 3.6248345375061035, "perplexity": 37.51851473896878, "lr": 0.0026291804804649314, "grad_norm": 0.145466, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:12.582987+00:00", "epoch": 0, "step": 4666, "train_loss": 3.5974669456481934, "perplexity": 36.50564644137959, "lr": 0.0026291804804649314, "grad_norm": 0.14974, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:12.887930+00:00", "epoch": 0, "step": 4667, "train_loss": 3.5463197231292725, "perplexity": 34.68543031490759, "lr": 0.0026291804804649314, "grad_norm": 0.16692, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:13.192429+00:00", "epoch": 0, "step": 4668, "train_loss": 3.6303017139434814, "perplexity": 37.72419681649919, "lr": 0.0026291804804649314, "grad_norm": 0.144924, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:13.497508+00:00", "epoch": 0, "step": 4669, "train_loss": 3.656423807144165, "perplexity": 38.72261542049769, "lr": 0.0026291804804649314, "grad_norm": 0.152075, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:13.802949+00:00", "epoch": 0, "step": 4670, "train_loss": 3.6590304374694824, "perplexity": 38.82368262932881, "lr": 0.0026291804804649314, "grad_norm": 0.156705, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:14.108574+00:00", "epoch": 0, "step": 4671, "train_loss": 3.6210386753082275, "perplexity": 37.37636957954498, "lr": 0.0026291804804649314, "grad_norm": 0.167279, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:14.413572+00:00", "epoch": 0, "step": 4672, "train_loss": 3.6159322261810303, "perplexity": 37.18599553110427, "lr": 0.0026291804804649314, "grad_norm": 0.168298, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:14.718249+00:00", "epoch": 0, "step": 4673, "train_loss": 3.7374868392944336, "perplexity": 41.99232398399334, "lr": 0.0026291804804649314, "grad_norm": 0.152305, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:15.021575+00:00", "epoch": 0, "step": 4674, "train_loss": 3.6407055854797363, "perplexity": 38.118723256546474, "lr": 0.0026291804804649314, "grad_norm": 0.154927, "tokens_per_sec": 108026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:15.326149+00:00", "epoch": 0, "step": 4675, "train_loss": 3.594890594482422, "perplexity": 36.41171612730605, "lr": 0.0026291804804649314, "grad_norm": 0.155934, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:15.631007+00:00", "epoch": 0, "step": 4676, "train_loss": 3.6569325923919678, "perplexity": 38.74232192874521, "lr": 0.0026291804804649314, "grad_norm": 0.151223, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:15.936237+00:00", "epoch": 0, "step": 4677, "train_loss": 3.620738983154297, "perplexity": 37.36516985315846, "lr": 0.0026291804804649314, "grad_norm": 0.166475, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:16.241093+00:00", "epoch": 0, "step": 4678, "train_loss": 3.5974128246307373, "perplexity": 36.50367077211439, "lr": 0.0026291804804649314, "grad_norm": 0.130098, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:16.546234+00:00", "epoch": 0, "step": 4679, "train_loss": 3.6227831840515137, "perplexity": 37.441629890109354, "lr": 0.0026291804804649314, "grad_norm": 0.162587, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:16.850543+00:00", "epoch": 0, "step": 4680, "train_loss": 3.6623342037200928, "perplexity": 38.9521591130007, "lr": 0.0026291804804649314, "grad_norm": 0.160294, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:17.155517+00:00", "epoch": 0, "step": 4681, "train_loss": 3.5517055988311768, "perplexity": 34.872745707162444, "lr": 0.0026291804804649314, "grad_norm": 0.153161, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:17.459850+00:00", "epoch": 0, "step": 4682, "train_loss": 3.616760492324829, "perplexity": 37.216808191006116, "lr": 0.0026291804804649314, "grad_norm": 0.170081, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:17.764190+00:00", "epoch": 0, "step": 4683, "train_loss": 3.683565616607666, "perplexity": 39.78801023814995, "lr": 0.0026291804804649314, "grad_norm": 0.172754, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:18.069778+00:00", "epoch": 0, "step": 4684, "train_loss": 3.5944936275482178, "perplexity": 36.39726474853542, "lr": 0.0026291804804649314, "grad_norm": 0.164266, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:18.374323+00:00", "epoch": 0, "step": 4685, "train_loss": 3.5754311084747314, "perplexity": 35.7100123866718, "lr": 0.0026291804804649314, "grad_norm": 0.174896, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:18.678767+00:00", "epoch": 0, "step": 4686, "train_loss": 3.5004384517669678, "perplexity": 33.12997467063924, "lr": 0.0026291804804649314, "grad_norm": 0.148768, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:18.982811+00:00", "epoch": 0, "step": 4687, "train_loss": 3.597076177597046, "perplexity": 36.491383987900875, "lr": 0.0026291804804649314, "grad_norm": 0.150558, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:19.288259+00:00", "epoch": 0, "step": 4688, "train_loss": 3.623774290084839, "perplexity": 37.478756910758925, "lr": 0.0026291804804649314, "grad_norm": 0.159542, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:19.594049+00:00", "epoch": 0, "step": 4689, "train_loss": 3.6215944290161133, "perplexity": 37.39714740866911, "lr": 0.0026291804804649314, "grad_norm": 0.141672, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:19.898505+00:00", "epoch": 0, "step": 4690, "train_loss": 3.6726765632629395, "perplexity": 39.357106795204466, "lr": 0.0026291804804649314, "grad_norm": 0.162707, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:20.203320+00:00", "epoch": 0, "step": 4691, "train_loss": 3.6269571781158447, "perplexity": 37.598237643575, "lr": 0.0026291804804649314, "grad_norm": 0.14255, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:20.508567+00:00", "epoch": 0, "step": 4692, "train_loss": 3.6952972412109375, "perplexity": 40.25753700877394, "lr": 0.0026291804804649314, "grad_norm": 0.152471, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:20.813264+00:00", "epoch": 0, "step": 4693, "train_loss": 3.623929262161255, "perplexity": 37.48456552161349, "lr": 0.0026291804804649314, "grad_norm": 0.160637, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:21.118320+00:00", "epoch": 0, "step": 4694, "train_loss": 3.590916156768799, "perplexity": 36.26728723145563, "lr": 0.0026291804804649314, "grad_norm": 0.151096, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:21.423413+00:00", "epoch": 0, "step": 4695, "train_loss": 3.583904504776001, "perplexity": 36.013883063764986, "lr": 0.0026291804804649314, "grad_norm": 0.143916, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:21.727304+00:00", "epoch": 0, "step": 4696, "train_loss": 3.6741890907287598, "perplexity": 39.416680542314374, "lr": 0.0026291804804649314, "grad_norm": 0.14479, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:22.031941+00:00", "epoch": 0, "step": 4697, "train_loss": 3.5989983081817627, "perplexity": 36.5615926466318, "lr": 0.0026291804804649314, "grad_norm": 0.152877, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:22.336692+00:00", "epoch": 0, "step": 4698, "train_loss": 3.5426840782165527, "perplexity": 34.559555363601206, "lr": 0.0026291804804649314, "grad_norm": 0.184889, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:22.642298+00:00", "epoch": 0, "step": 4699, "train_loss": 3.6036009788513184, "perplexity": 36.73026148252589, "lr": 0.0026291804804649314, "grad_norm": 0.209791, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:22.948736+00:00", "epoch": 0, "step": 4700, "train_loss": 3.6117959022521973, "perplexity": 37.03249988068599, "lr": 0.0026291804804649314, "grad_norm": 0.159536, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:23.254412+00:00", "epoch": 0, "step": 4701, "train_loss": 3.585766077041626, "perplexity": 36.08098795057298, "lr": 0.0026291804804649314, "grad_norm": 0.172814, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:23.558915+00:00", "epoch": 0, "step": 4702, "train_loss": 3.4400882720947266, "perplexity": 31.18971122794251, "lr": 0.0026291804804649314, "grad_norm": 0.203846, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:23.864001+00:00", "epoch": 0, "step": 4703, "train_loss": 3.6224560737609863, "perplexity": 37.429384350608366, "lr": 0.0026291804804649314, "grad_norm": 0.203623, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:24.170205+00:00", "epoch": 0, "step": 4704, "train_loss": 3.626835346221924, "perplexity": 37.59365725809898, "lr": 0.0026291804804649314, "grad_norm": 0.200952, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:24.474932+00:00", "epoch": 0, "step": 4705, "train_loss": 3.6510398387908936, "perplexity": 38.514694307117566, "lr": 0.0026291804804649314, "grad_norm": 0.165658, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:24.778744+00:00", "epoch": 0, "step": 4706, "train_loss": 3.669265031814575, "perplexity": 39.22306755719569, "lr": 0.0026291804804649314, "grad_norm": 0.171215, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:25.083483+00:00", "epoch": 0, "step": 4707, "train_loss": 3.5597431659698486, "perplexity": 35.15416719911917, "lr": 0.0026291804804649314, "grad_norm": 0.16913, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:25.388368+00:00", "epoch": 0, "step": 4708, "train_loss": 3.6758105754852295, "perplexity": 39.4806459344036, "lr": 0.0026291804804649314, "grad_norm": 0.161738, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:25.693731+00:00", "epoch": 0, "step": 4709, "train_loss": 3.564727783203125, "perplexity": 35.32983472071368, "lr": 0.0026291804804649314, "grad_norm": 0.172804, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:26.001038+00:00", "epoch": 0, "step": 4710, "train_loss": 3.6009645462036133, "perplexity": 36.63355216181691, "lr": 0.0026291804804649314, "grad_norm": 0.182032, "tokens_per_sec": 106697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:26.306911+00:00", "epoch": 0, "step": 4711, "train_loss": 3.660041570663452, "perplexity": 38.86295839671727, "lr": 0.0026291804804649314, "grad_norm": 0.165414, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:26.611474+00:00", "epoch": 0, "step": 4712, "train_loss": 3.571397066116333, "perplexity": 35.56624685716683, "lr": 0.0026291804804649314, "grad_norm": 0.158221, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:26.916037+00:00", "epoch": 0, "step": 4713, "train_loss": 3.7104642391204834, "perplexity": 40.87277686524663, "lr": 0.0026291804804649314, "grad_norm": 0.160652, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:27.221010+00:00", "epoch": 0, "step": 4714, "train_loss": 3.647414445877075, "perplexity": 38.375316210140554, "lr": 0.0026291804804649314, "grad_norm": 0.188644, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:27.526834+00:00", "epoch": 0, "step": 4715, "train_loss": 3.56022047996521, "perplexity": 35.17095078031904, "lr": 0.0026291804804649314, "grad_norm": 0.148235, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:27.831622+00:00", "epoch": 0, "step": 4716, "train_loss": 3.5727992057800293, "perplexity": 35.61615068046319, "lr": 0.0026291804804649314, "grad_norm": 0.162964, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:28.136358+00:00", "epoch": 0, "step": 4717, "train_loss": 3.5457091331481934, "perplexity": 34.66425820306536, "lr": 0.0026291804804649314, "grad_norm": 0.167565, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:28.440271+00:00", "epoch": 0, "step": 4718, "train_loss": 3.676938772201538, "perplexity": 39.52521300498869, "lr": 0.0026291804804649314, "grad_norm": 0.178627, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:28.745064+00:00", "epoch": 0, "step": 4719, "train_loss": 3.6050429344177246, "perplexity": 36.78326309132502, "lr": 0.0026291804804649314, "grad_norm": 0.166679, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:29.049968+00:00", "epoch": 0, "step": 4720, "train_loss": 3.7402591705322266, "perplexity": 42.10890213748341, "lr": 0.0026291804804649314, "grad_norm": 0.149144, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:29.355786+00:00", "epoch": 0, "step": 4721, "train_loss": 3.6461782455444336, "perplexity": 38.32790594180875, "lr": 0.0026291804804649314, "grad_norm": 0.15444, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:29.660613+00:00", "epoch": 0, "step": 4722, "train_loss": 3.510890245437622, "perplexity": 33.478058210163944, "lr": 0.0026291804804649314, "grad_norm": 0.139264, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:29.965536+00:00", "epoch": 0, "step": 4723, "train_loss": 3.575066089630127, "perplexity": 35.696979937899, "lr": 0.0026291804804649314, "grad_norm": 0.144438, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:30.269132+00:00", "epoch": 0, "step": 4724, "train_loss": 3.5706140995025635, "perplexity": 35.53841057216354, "lr": 0.0026291804804649314, "grad_norm": 0.146077, "tokens_per_sec": 107932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:30.573340+00:00", "epoch": 0, "step": 4725, "train_loss": 3.577521800994873, "perplexity": 35.7847491410266, "lr": 0.0026291804804649314, "grad_norm": 0.151789, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:30.877678+00:00", "epoch": 0, "step": 4726, "train_loss": 3.4324228763580322, "perplexity": 30.951543738687253, "lr": 0.0026291804804649314, "grad_norm": 0.16445, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:31.182110+00:00", "epoch": 0, "step": 4727, "train_loss": 3.6370439529418945, "perplexity": 37.979401726955665, "lr": 0.0026291804804649314, "grad_norm": 0.182464, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:31.486154+00:00", "epoch": 0, "step": 4728, "train_loss": 3.652402400970459, "perplexity": 38.56720874190458, "lr": 0.0026291804804649314, "grad_norm": 0.159012, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:31.790713+00:00", "epoch": 0, "step": 4729, "train_loss": 3.583019495010376, "perplexity": 35.98202452519377, "lr": 0.0026291804804649314, "grad_norm": 0.192157, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:32.095399+00:00", "epoch": 0, "step": 4730, "train_loss": 3.513676881790161, "perplexity": 33.57147948929706, "lr": 0.0026291804804649314, "grad_norm": 0.158106, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:32.400916+00:00", "epoch": 0, "step": 4731, "train_loss": 3.506598711013794, "perplexity": 33.33469381755675, "lr": 0.0026291804804649314, "grad_norm": 0.160501, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:32.706098+00:00", "epoch": 0, "step": 4732, "train_loss": 3.6754987239837646, "perplexity": 39.468335755263844, "lr": 0.0026291804804649314, "grad_norm": 0.172957, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:33.011262+00:00", "epoch": 0, "step": 4733, "train_loss": 3.4626224040985107, "perplexity": 31.900522982702967, "lr": 0.0026291804804649314, "grad_norm": 0.169675, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:33.316290+00:00", "epoch": 0, "step": 4734, "train_loss": 3.652735948562622, "perplexity": 38.580074887133485, "lr": 0.0026291804804649314, "grad_norm": 0.148387, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:33.620582+00:00", "epoch": 0, "step": 4735, "train_loss": 3.628303289413452, "perplexity": 37.64888313562412, "lr": 0.0026291804804649314, "grad_norm": 0.155598, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:33.925953+00:00", "epoch": 0, "step": 4736, "train_loss": 3.5573859214782715, "perplexity": 35.07139782428588, "lr": 0.0026291804804649314, "grad_norm": 0.164981, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:34.230992+00:00", "epoch": 0, "step": 4737, "train_loss": 3.5407917499542236, "perplexity": 34.494219178504835, "lr": 0.0026291804804649314, "grad_norm": 0.155989, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:34.536318+00:00", "epoch": 0, "step": 4738, "train_loss": 3.6629273891448975, "perplexity": 38.97527182043407, "lr": 0.0026291804804649314, "grad_norm": 0.149738, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:34.840467+00:00", "epoch": 0, "step": 4739, "train_loss": 3.63128662109375, "perplexity": 37.76136995071978, "lr": 0.0026291804804649314, "grad_norm": 0.151349, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:35.145453+00:00", "epoch": 0, "step": 4740, "train_loss": 3.6145851612091064, "perplexity": 37.13593730249094, "lr": 0.0026291804804649314, "grad_norm": 0.181941, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:35.451649+00:00", "epoch": 0, "step": 4741, "train_loss": 3.602595806121826, "perplexity": 36.693359774738745, "lr": 0.0026291804804649314, "grad_norm": 0.178743, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:35.757188+00:00", "epoch": 0, "step": 4742, "train_loss": 3.56990909576416, "perplexity": 35.51336468961105, "lr": 0.0026291804804649314, "grad_norm": 0.163041, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:36.063738+00:00", "epoch": 0, "step": 4743, "train_loss": 3.6754090785980225, "perplexity": 39.464797759665316, "lr": 0.0026291804804649314, "grad_norm": 0.168335, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:36.368368+00:00", "epoch": 0, "step": 4744, "train_loss": 3.5352139472961426, "perplexity": 34.3023528248546, "lr": 0.0026291804804649314, "grad_norm": 0.163815, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:36.672682+00:00", "epoch": 0, "step": 4745, "train_loss": 3.5720536708831787, "perplexity": 35.58960749290516, "lr": 0.0026291804804649314, "grad_norm": 0.155129, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:36.977586+00:00", "epoch": 0, "step": 4746, "train_loss": 3.688318967819214, "perplexity": 39.97758682993517, "lr": 0.0026291804804649314, "grad_norm": 0.154603, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:37.284182+00:00", "epoch": 0, "step": 4747, "train_loss": 3.5673954486846924, "perplexity": 35.42420872443642, "lr": 0.0026291804804649314, "grad_norm": 0.169714, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:37.589072+00:00", "epoch": 0, "step": 4748, "train_loss": 3.544093370437622, "perplexity": 34.608294211711794, "lr": 0.0026291804804649314, "grad_norm": 0.136022, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:37.892946+00:00", "epoch": 0, "step": 4749, "train_loss": 3.594538688659668, "perplexity": 36.39890488669168, "lr": 0.0026291804804649314, "grad_norm": 0.178203, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:38.197377+00:00", "epoch": 0, "step": 4750, "train_loss": 3.7321577072143555, "perplexity": 41.76913656894334, "lr": 0.0026291804804649314, "grad_norm": 0.169229, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:38.501831+00:00", "epoch": 0, "step": 4751, "train_loss": 3.508443593978882, "perplexity": 33.396249190097166, "lr": 0.0026291804804649314, "grad_norm": 0.138589, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:38.805562+00:00", "epoch": 0, "step": 4752, "train_loss": 3.6321520805358887, "perplexity": 37.794065030983916, "lr": 0.0026291804804649314, "grad_norm": 0.179653, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:39.111254+00:00", "epoch": 0, "step": 4753, "train_loss": 3.6174933910369873, "perplexity": 37.24409433957014, "lr": 0.0026291804804649314, "grad_norm": 0.156058, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:39.415581+00:00", "epoch": 0, "step": 4754, "train_loss": 3.574322462081909, "perplexity": 35.67044454767539, "lr": 0.0026291804804649314, "grad_norm": 0.142926, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:39.720731+00:00", "epoch": 0, "step": 4755, "train_loss": 3.596129894256592, "perplexity": 36.45686913215504, "lr": 0.0026291804804649314, "grad_norm": 0.156573, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:40.027030+00:00", "epoch": 0, "step": 4756, "train_loss": 3.6308066844940186, "perplexity": 37.74325123548951, "lr": 0.0026291804804649314, "grad_norm": 0.160344, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:40.332316+00:00", "epoch": 0, "step": 4757, "train_loss": 3.694807529449463, "perplexity": 40.23782724585799, "lr": 0.0026291804804649314, "grad_norm": 0.150097, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:40.637299+00:00", "epoch": 0, "step": 4758, "train_loss": 3.741243362426758, "perplexity": 42.15036577839746, "lr": 0.0026291804804649314, "grad_norm": 0.142321, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:40.942653+00:00", "epoch": 0, "step": 4759, "train_loss": 3.6716790199279785, "perplexity": 39.31786595111283, "lr": 0.0026291804804649314, "grad_norm": 0.143228, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:41.247975+00:00", "epoch": 0, "step": 4760, "train_loss": 3.636401653289795, "perplexity": 37.955015402941584, "lr": 0.0026291804804649314, "grad_norm": 0.169522, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:41.552625+00:00", "epoch": 0, "step": 4761, "train_loss": 3.6532883644104004, "perplexity": 38.6013930196046, "lr": 0.0026291804804649314, "grad_norm": 0.158904, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:41.857668+00:00", "epoch": 0, "step": 4762, "train_loss": 3.69105863571167, "perplexity": 40.08726230958612, "lr": 0.0026291804804649314, "grad_norm": 0.168884, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:42.163887+00:00", "epoch": 0, "step": 4763, "train_loss": 3.559650182723999, "perplexity": 35.15089860251258, "lr": 0.0026291804804649314, "grad_norm": 0.218066, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:42.468780+00:00", "epoch": 0, "step": 4764, "train_loss": 3.579406499862671, "perplexity": 35.852256212485095, "lr": 0.0026291804804649314, "grad_norm": 0.184732, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:42.773132+00:00", "epoch": 0, "step": 4765, "train_loss": 3.495960235595703, "perplexity": 32.981943188195245, "lr": 0.0026291804804649314, "grad_norm": 0.180307, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:43.078182+00:00", "epoch": 0, "step": 4766, "train_loss": 3.5306684970855713, "perplexity": 34.14678701415414, "lr": 0.0026291804804649314, "grad_norm": 0.18806, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:43.382072+00:00", "epoch": 0, "step": 4767, "train_loss": 3.596503734588623, "perplexity": 36.47050072807764, "lr": 0.0026291804804649314, "grad_norm": 0.148099, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:43.687514+00:00", "epoch": 0, "step": 4768, "train_loss": 3.566141128540039, "perplexity": 35.37980328096808, "lr": 0.0026291804804649314, "grad_norm": 0.161075, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:43.995906+00:00", "epoch": 0, "step": 4769, "train_loss": 3.561480760574341, "perplexity": 35.21530399046698, "lr": 0.0026291804804649314, "grad_norm": 0.157397, "tokens_per_sec": 106255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:44.301279+00:00", "epoch": 0, "step": 4770, "train_loss": 3.5453433990478516, "perplexity": 34.651582619865245, "lr": 0.0026291804804649314, "grad_norm": 0.14942, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:44.604611+00:00", "epoch": 0, "step": 4771, "train_loss": 3.6524393558502197, "perplexity": 38.568634014801574, "lr": 0.0026291804804649314, "grad_norm": 0.151039, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:44.908012+00:00", "epoch": 0, "step": 4772, "train_loss": 3.7152247428894043, "perplexity": 41.067815746888236, "lr": 0.0026291804804649314, "grad_norm": 0.137906, "tokens_per_sec": 108002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:45.211648+00:00", "epoch": 0, "step": 4773, "train_loss": 3.6244215965270996, "perplexity": 37.50302500515473, "lr": 0.0026291804804649314, "grad_norm": 0.160779, "tokens_per_sec": 107919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:45.517159+00:00", "epoch": 0, "step": 4774, "train_loss": 3.5951650142669678, "perplexity": 36.42170959374039, "lr": 0.0026291804804649314, "grad_norm": 0.164882, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:45.822086+00:00", "epoch": 0, "step": 4775, "train_loss": 3.6918282508850098, "perplexity": 40.1181259499568, "lr": 0.0026291804804649314, "grad_norm": 0.150775, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:46.126021+00:00", "epoch": 0, "step": 4776, "train_loss": 3.6099393367767334, "perplexity": 36.96381040292397, "lr": 0.0026291804804649314, "grad_norm": 0.162005, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:46.430641+00:00", "epoch": 0, "step": 4777, "train_loss": 3.665799140930176, "perplexity": 39.08735999454096, "lr": 0.0026291804804649314, "grad_norm": 0.165718, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:46.736131+00:00", "epoch": 0, "step": 4778, "train_loss": 3.601738214492798, "perplexity": 36.66190534601019, "lr": 0.0026291804804649314, "grad_norm": 0.138665, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:47.041337+00:00", "epoch": 0, "step": 4779, "train_loss": 3.6582956314086914, "perplexity": 38.79516523069237, "lr": 0.0026291804804649314, "grad_norm": 0.157467, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:47.346726+00:00", "epoch": 0, "step": 4780, "train_loss": 3.5749638080596924, "perplexity": 35.693328981447145, "lr": 0.0026291804804649314, "grad_norm": 0.148146, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:47.650735+00:00", "epoch": 0, "step": 4781, "train_loss": 3.5993354320526123, "perplexity": 36.57392051016104, "lr": 0.0026291804804649314, "grad_norm": 0.151544, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:47.954799+00:00", "epoch": 0, "step": 4782, "train_loss": 3.6409647464752197, "perplexity": 38.12860342303379, "lr": 0.0026291804804649314, "grad_norm": 0.141429, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:48.259517+00:00", "epoch": 0, "step": 4783, "train_loss": 3.6511292457580566, "perplexity": 38.51813794306701, "lr": 0.0026291804804649314, "grad_norm": 0.182102, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:48.566064+00:00", "epoch": 0, "step": 4784, "train_loss": 3.5730044841766357, "perplexity": 35.62346265723793, "lr": 0.0026291804804649314, "grad_norm": 0.190126, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:48.871418+00:00", "epoch": 0, "step": 4785, "train_loss": 3.68339204788208, "perplexity": 39.78110488321347, "lr": 0.0026291804804649314, "grad_norm": 0.198854, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:49.176427+00:00", "epoch": 0, "step": 4786, "train_loss": 3.6704041957855225, "perplexity": 39.26777452203401, "lr": 0.0026291804804649314, "grad_norm": 0.245927, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:49.480562+00:00", "epoch": 0, "step": 4787, "train_loss": 3.5870842933654785, "perplexity": 36.12858186051268, "lr": 0.0026291804804649314, "grad_norm": 0.200039, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:49.784870+00:00", "epoch": 0, "step": 4788, "train_loss": 3.60975980758667, "perplexity": 36.957174915630326, "lr": 0.0026291804804649314, "grad_norm": 0.187545, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:50.089164+00:00", "epoch": 0, "step": 4789, "train_loss": 3.678267002105713, "perplexity": 39.57774645539743, "lr": 0.0026291804804649314, "grad_norm": 0.172679, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:50.399740+00:00", "epoch": 0, "step": 4790, "train_loss": 3.5794782638549805, "perplexity": 35.854829205847246, "lr": 0.0026291804804649314, "grad_norm": 0.163944, "tokens_per_sec": 105574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:50.704186+00:00", "epoch": 0, "step": 4791, "train_loss": 3.6560428142547607, "perplexity": 38.707865189408345, "lr": 0.0026291804804649314, "grad_norm": 0.159269, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:51.009584+00:00", "epoch": 0, "step": 4792, "train_loss": 3.6374902725219727, "perplexity": 37.99635646091948, "lr": 0.0026291804804649314, "grad_norm": 0.147917, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:51.314358+00:00", "epoch": 0, "step": 4793, "train_loss": 3.6216933727264404, "perplexity": 37.400847804251796, "lr": 0.0026291804804649314, "grad_norm": 0.147835, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:51.618308+00:00", "epoch": 0, "step": 4794, "train_loss": 3.582152843475342, "perplexity": 35.95085415727994, "lr": 0.0026291804804649314, "grad_norm": 0.160099, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:51.922960+00:00", "epoch": 0, "step": 4795, "train_loss": 3.585850238800049, "perplexity": 36.08402471775258, "lr": 0.0026291804804649314, "grad_norm": 0.161012, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:52.227504+00:00", "epoch": 0, "step": 4796, "train_loss": 3.697962522506714, "perplexity": 40.36497778552132, "lr": 0.0026291804804649314, "grad_norm": 0.165724, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:52.531614+00:00", "epoch": 0, "step": 4797, "train_loss": 3.584181547164917, "perplexity": 36.02386181816827, "lr": 0.0026291804804649314, "grad_norm": 0.140837, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:52.836122+00:00", "epoch": 0, "step": 4798, "train_loss": 3.691608190536499, "perplexity": 40.10929851249876, "lr": 0.0026291804804649314, "grad_norm": 0.165252, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:53.140744+00:00", "epoch": 0, "step": 4799, "train_loss": 3.68615984916687, "perplexity": 39.891363593130635, "lr": 0.0026291804804649314, "grad_norm": 0.142252, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:53.450100+00:00", "epoch": 0, "step": 4800, "train_loss": 3.612842082977295, "perplexity": 37.071262841258964, "lr": 0.0026291804804649314, "grad_norm": 0.148722, "tokens_per_sec": 105978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:53.755629+00:00", "epoch": 0, "step": 4801, "train_loss": 3.5874547958374023, "perplexity": 36.1419700694282, "lr": 0.0026291804804649314, "grad_norm": 0.134444, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:54.060896+00:00", "epoch": 0, "step": 4802, "train_loss": 3.5494492053985596, "perplexity": 34.79414778004197, "lr": 0.0026291804804649314, "grad_norm": 0.145042, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:54.365133+00:00", "epoch": 0, "step": 4803, "train_loss": 3.6055479049682617, "perplexity": 36.80184224650711, "lr": 0.0026291804804649314, "grad_norm": 0.136183, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:54.669528+00:00", "epoch": 0, "step": 4804, "train_loss": 3.5453362464904785, "perplexity": 34.65133477331886, "lr": 0.0026291804804649314, "grad_norm": 0.133088, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:54.975102+00:00", "epoch": 0, "step": 4805, "train_loss": 3.6980881690979004, "perplexity": 40.37004982601901, "lr": 0.0026291804804649314, "grad_norm": 0.138898, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:55.280545+00:00", "epoch": 0, "step": 4806, "train_loss": 3.5277059078216553, "perplexity": 34.04577381374249, "lr": 0.0026291804804649314, "grad_norm": 0.125288, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:55.586721+00:00", "epoch": 0, "step": 4807, "train_loss": 3.6559150218963623, "perplexity": 38.70291893608064, "lr": 0.0026291804804649314, "grad_norm": 0.142519, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:55.891985+00:00", "epoch": 0, "step": 4808, "train_loss": 3.701810121536255, "perplexity": 40.52058520037766, "lr": 0.0026291804804649314, "grad_norm": 0.152597, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:56.196423+00:00", "epoch": 0, "step": 4809, "train_loss": 3.6701955795288086, "perplexity": 39.25958348032568, "lr": 0.0026291804804649314, "grad_norm": 0.136448, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:56.502744+00:00", "epoch": 0, "step": 4810, "train_loss": 3.5191447734832764, "perplexity": 33.755547476428376, "lr": 0.0026291804804649314, "grad_norm": 0.145485, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:56.806837+00:00", "epoch": 0, "step": 4811, "train_loss": 3.612492561340332, "perplexity": 37.05830789693496, "lr": 0.0026291804804649314, "grad_norm": 0.138571, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:57.113170+00:00", "epoch": 0, "step": 4812, "train_loss": 3.5438477993011475, "perplexity": 34.599796457015124, "lr": 0.0026291804804649314, "grad_norm": 0.157198, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:57.418113+00:00", "epoch": 0, "step": 4813, "train_loss": 3.6697285175323486, "perplexity": 39.24125110239695, "lr": 0.0026291804804649314, "grad_norm": 0.135607, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:57.722439+00:00", "epoch": 0, "step": 4814, "train_loss": 3.7123684883117676, "perplexity": 40.9506829703149, "lr": 0.0026291804804649314, "grad_norm": 0.138096, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:58.027839+00:00", "epoch": 0, "step": 4815, "train_loss": 3.538578987121582, "perplexity": 34.41797603746924, "lr": 0.0026291804804649314, "grad_norm": 0.141814, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:58.334068+00:00", "epoch": 0, "step": 4816, "train_loss": 3.615252733230591, "perplexity": 37.1607364919272, "lr": 0.0026291804804649314, "grad_norm": 0.166134, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:58.640811+00:00", "epoch": 0, "step": 4817, "train_loss": 3.587437868118286, "perplexity": 36.14135827348873, "lr": 0.0026291804804649314, "grad_norm": 0.163657, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:58.945890+00:00", "epoch": 0, "step": 4818, "train_loss": 3.5257160663604736, "perplexity": 33.978095478339654, "lr": 0.0026291804804649314, "grad_norm": 0.191231, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:59.251077+00:00", "epoch": 0, "step": 4819, "train_loss": 3.725005626678467, "perplexity": 41.47146609077951, "lr": 0.0026291804804649314, "grad_norm": 0.149344, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:59.556749+00:00", "epoch": 0, "step": 4820, "train_loss": 3.5809895992279053, "perplexity": 35.90905884678121, "lr": 0.0026291804804649314, "grad_norm": 0.157716, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:22:59.863548+00:00", "epoch": 0, "step": 4821, "train_loss": 3.6251237392425537, "perplexity": 37.52936672771168, "lr": 0.0026291804804649314, "grad_norm": 0.19006, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:00.168417+00:00", "epoch": 0, "step": 4822, "train_loss": 3.5641727447509766, "perplexity": 35.3102307449191, "lr": 0.0026291804804649314, "grad_norm": 0.155017, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:00.474869+00:00", "epoch": 0, "step": 4823, "train_loss": 3.4893717765808105, "perplexity": 32.76535727443544, "lr": 0.0026291804804649314, "grad_norm": 0.166782, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:00.780918+00:00", "epoch": 0, "step": 4824, "train_loss": 3.4823713302612305, "perplexity": 32.536786133146705, "lr": 0.0026291804804649314, "grad_norm": 0.189613, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:01.085765+00:00", "epoch": 0, "step": 4825, "train_loss": 3.51309871673584, "perplexity": 33.55207524300357, "lr": 0.0026291804804649314, "grad_norm": 0.161866, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:01.391356+00:00", "epoch": 0, "step": 4826, "train_loss": 3.5711491107940674, "perplexity": 35.5574291102144, "lr": 0.0026291804804649314, "grad_norm": 0.163732, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:01.695305+00:00", "epoch": 0, "step": 4827, "train_loss": 3.553215742111206, "perplexity": 34.925448333991326, "lr": 0.0026291804804649314, "grad_norm": 0.192909, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:01.999468+00:00", "epoch": 0, "step": 4828, "train_loss": 3.6554317474365234, "perplexity": 38.684219322724346, "lr": 0.0026291804804649314, "grad_norm": 0.149478, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:02.303698+00:00", "epoch": 0, "step": 4829, "train_loss": 3.598392963409424, "perplexity": 36.53946697516021, "lr": 0.0026291804804649314, "grad_norm": 0.192078, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:02.609436+00:00", "epoch": 0, "step": 4830, "train_loss": 3.67325496673584, "perplexity": 39.37987766719901, "lr": 0.0026291804804649314, "grad_norm": 0.178858, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:02.914350+00:00", "epoch": 0, "step": 4831, "train_loss": 3.492438316345215, "perplexity": 32.86598776024759, "lr": 0.0026291804804649314, "grad_norm": 0.175171, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:03.219019+00:00", "epoch": 0, "step": 4832, "train_loss": 3.7063968181610107, "perplexity": 40.706867715861016, "lr": 0.0026291804804649314, "grad_norm": 0.179062, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:03.523642+00:00", "epoch": 0, "step": 4833, "train_loss": 3.5901732444763184, "perplexity": 36.240353823775045, "lr": 0.0026291804804649314, "grad_norm": 0.174169, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:03.828876+00:00", "epoch": 0, "step": 4834, "train_loss": 3.6532206535339355, "perplexity": 38.598779373937596, "lr": 0.0026291804804649314, "grad_norm": 0.18403, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:04.134881+00:00", "epoch": 0, "step": 4835, "train_loss": 3.592242956161499, "perplexity": 36.31543858265719, "lr": 0.0026291804804649314, "grad_norm": 0.157668, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:04.440336+00:00", "epoch": 0, "step": 4836, "train_loss": 3.6392006874084473, "perplexity": 38.06140160587658, "lr": 0.0026291804804649314, "grad_norm": 0.16071, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:04.745871+00:00", "epoch": 0, "step": 4837, "train_loss": 3.5433971881866455, "perplexity": 34.58420891639537, "lr": 0.0026291804804649314, "grad_norm": 0.175111, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:05.049598+00:00", "epoch": 0, "step": 4838, "train_loss": 3.5746774673461914, "perplexity": 35.68310999128566, "lr": 0.0026291804804649314, "grad_norm": 0.167596, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:05.354764+00:00", "epoch": 0, "step": 4839, "train_loss": 3.570732355117798, "perplexity": 35.542613437271505, "lr": 0.0026291804804649314, "grad_norm": 0.16154, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:05.659241+00:00", "epoch": 0, "step": 4840, "train_loss": 3.563508987426758, "perplexity": 35.28680109730262, "lr": 0.0026291804804649314, "grad_norm": 0.178735, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:05.964273+00:00", "epoch": 0, "step": 4841, "train_loss": 3.6538994312286377, "perplexity": 38.62498825841616, "lr": 0.0026291804804649314, "grad_norm": 0.207666, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:06.268726+00:00", "epoch": 0, "step": 4842, "train_loss": 3.48665452003479, "perplexity": 32.67644624464372, "lr": 0.0026291804804649314, "grad_norm": 0.184229, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:06.572688+00:00", "epoch": 0, "step": 4843, "train_loss": 3.500732421875, "perplexity": 33.139715324532666, "lr": 0.0026291804804649314, "grad_norm": 0.148185, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:06.876893+00:00", "epoch": 0, "step": 4844, "train_loss": 3.511981725692749, "perplexity": 33.5146187986309, "lr": 0.0026291804804649314, "grad_norm": 0.160257, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:07.182219+00:00", "epoch": 0, "step": 4845, "train_loss": 3.5341386795043945, "perplexity": 34.26548843277947, "lr": 0.0026291804804649314, "grad_norm": 0.147754, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:07.487806+00:00", "epoch": 0, "step": 4846, "train_loss": 3.6058132648468018, "perplexity": 36.81160927472709, "lr": 0.0026291804804649314, "grad_norm": 0.160975, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:07.792450+00:00", "epoch": 0, "step": 4847, "train_loss": 3.5937108993530273, "perplexity": 36.36878672991694, "lr": 0.0026291804804649314, "grad_norm": 0.142388, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:08.097053+00:00", "epoch": 0, "step": 4848, "train_loss": 3.569934368133545, "perplexity": 35.51426220782274, "lr": 0.0026291804804649314, "grad_norm": 0.165796, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:08.401725+00:00", "epoch": 0, "step": 4849, "train_loss": 3.623002529144287, "perplexity": 37.44984342865794, "lr": 0.0026291804804649314, "grad_norm": 0.148347, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:08.707144+00:00", "epoch": 0, "step": 4850, "train_loss": 3.545497417449951, "perplexity": 34.65692001226836, "lr": 0.0026291804804649314, "grad_norm": 0.15038, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:09.011564+00:00", "epoch": 0, "step": 4851, "train_loss": 3.507523775100708, "perplexity": 33.36554481302958, "lr": 0.0026291804804649314, "grad_norm": 0.159023, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:09.316706+00:00", "epoch": 0, "step": 4852, "train_loss": 3.485048532485962, "perplexity": 32.62401039576527, "lr": 0.0026291804804649314, "grad_norm": 0.136841, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:09.620866+00:00", "epoch": 0, "step": 4853, "train_loss": 3.631755828857422, "perplexity": 37.779092036012415, "lr": 0.0026291804804649314, "grad_norm": 0.152005, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:09.925049+00:00", "epoch": 0, "step": 4854, "train_loss": 3.58335542678833, "perplexity": 35.99411406118287, "lr": 0.0026291804804649314, "grad_norm": 0.125225, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:10.229849+00:00", "epoch": 0, "step": 4855, "train_loss": 3.5068001747131348, "perplexity": 33.341410224823036, "lr": 0.0026291804804649314, "grad_norm": 0.159009, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:10.535684+00:00", "epoch": 0, "step": 4856, "train_loss": 3.640104055404663, "perplexity": 38.09580059311119, "lr": 0.0026291804804649314, "grad_norm": 0.144332, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:10.841590+00:00", "epoch": 0, "step": 4857, "train_loss": 3.6464192867279053, "perplexity": 38.337145659148426, "lr": 0.0026291804804649314, "grad_norm": 0.160273, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:11.147039+00:00", "epoch": 0, "step": 4858, "train_loss": 3.539726734161377, "perplexity": 34.45750184605786, "lr": 0.0026291804804649314, "grad_norm": 0.192152, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:11.451212+00:00", "epoch": 0, "step": 4859, "train_loss": 3.678496837615967, "perplexity": 39.586843872363396, "lr": 0.0026291804804649314, "grad_norm": 0.189482, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:11.756046+00:00", "epoch": 0, "step": 4860, "train_loss": 3.572474241256714, "perplexity": 35.60457857539863, "lr": 0.0026291804804649314, "grad_norm": 0.159845, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:12.061470+00:00", "epoch": 0, "step": 4861, "train_loss": 3.625697612762451, "perplexity": 37.55091001846604, "lr": 0.0026291804804649314, "grad_norm": 0.163772, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:12.366267+00:00", "epoch": 0, "step": 4862, "train_loss": 3.59004807472229, "perplexity": 36.23581791148646, "lr": 0.0026291804804649314, "grad_norm": 0.14369, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:12.672213+00:00", "epoch": 0, "step": 4863, "train_loss": 3.6284377574920654, "perplexity": 37.65394604899379, "lr": 0.0026291804804649314, "grad_norm": 0.156061, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:12.978455+00:00", "epoch": 0, "step": 4864, "train_loss": 3.565704345703125, "perplexity": 35.36435336449636, "lr": 0.0026291804804649314, "grad_norm": 0.153242, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:13.282463+00:00", "epoch": 0, "step": 4865, "train_loss": 3.6367909908294678, "perplexity": 37.96979559231139, "lr": 0.0026291804804649314, "grad_norm": 0.163455, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:13.586853+00:00", "epoch": 0, "step": 4866, "train_loss": 3.642930746078491, "perplexity": 38.20363797702011, "lr": 0.0026291804804649314, "grad_norm": 0.148633, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:13.891957+00:00", "epoch": 0, "step": 4867, "train_loss": 3.6362433433532715, "perplexity": 37.94900722245229, "lr": 0.0026291804804649314, "grad_norm": 0.149005, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:14.197602+00:00", "epoch": 0, "step": 4868, "train_loss": 3.5851738452911377, "perplexity": 36.05962597016556, "lr": 0.0026291804804649314, "grad_norm": 0.141277, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:14.501753+00:00", "epoch": 0, "step": 4869, "train_loss": 3.5602219104766846, "perplexity": 35.171001092803685, "lr": 0.0026291804804649314, "grad_norm": 0.1598, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:14.806165+00:00", "epoch": 0, "step": 4870, "train_loss": 3.6290440559387207, "perplexity": 37.676782500146615, "lr": 0.0026291804804649314, "grad_norm": 0.158519, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:15.109942+00:00", "epoch": 0, "step": 4871, "train_loss": 3.6399030685424805, "perplexity": 38.088144607089724, "lr": 0.0026291804804649314, "grad_norm": 0.142212, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:15.414905+00:00", "epoch": 0, "step": 4872, "train_loss": 3.5392658710479736, "perplexity": 34.44162531321217, "lr": 0.0026291804804649314, "grad_norm": 0.148576, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:15.720409+00:00", "epoch": 0, "step": 4873, "train_loss": 3.6770262718200684, "perplexity": 39.52867159735951, "lr": 0.0026291804804649314, "grad_norm": 0.16019, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:16.025241+00:00", "epoch": 0, "step": 4874, "train_loss": 3.627800941467285, "perplexity": 37.62997504612305, "lr": 0.0026291804804649314, "grad_norm": 0.162219, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:16.331013+00:00", "epoch": 0, "step": 4875, "train_loss": 3.5712921619415283, "perplexity": 35.56251600508378, "lr": 0.0026291804804649314, "grad_norm": 0.168574, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:16.635323+00:00", "epoch": 0, "step": 4876, "train_loss": 3.553704261779785, "perplexity": 34.942514270619874, "lr": 0.0026291804804649314, "grad_norm": 0.192747, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:16.940787+00:00", "epoch": 0, "step": 4877, "train_loss": 3.6478946208953857, "perplexity": 38.3937475030735, "lr": 0.0026291804804649314, "grad_norm": 0.176865, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:17.246393+00:00", "epoch": 0, "step": 4878, "train_loss": 3.48783278465271, "perplexity": 32.71497043648228, "lr": 0.0026291804804649314, "grad_norm": 0.17122, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:17.552188+00:00", "epoch": 0, "step": 4879, "train_loss": 3.496420383453369, "perplexity": 32.99712325096379, "lr": 0.0026291804804649314, "grad_norm": 0.177097, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:17.856933+00:00", "epoch": 0, "step": 4880, "train_loss": 3.5361690521240234, "perplexity": 34.335130818363474, "lr": 0.0026291804804649314, "grad_norm": 0.196916, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:18.162130+00:00", "epoch": 0, "step": 4881, "train_loss": 3.587949752807617, "perplexity": 36.15986321723518, "lr": 0.0026291804804649314, "grad_norm": 0.211494, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:18.466613+00:00", "epoch": 0, "step": 4882, "train_loss": 3.467075824737549, "perplexity": 32.04290624109772, "lr": 0.0026291804804649314, "grad_norm": 0.206127, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:18.772376+00:00", "epoch": 0, "step": 4883, "train_loss": 3.6953022480010986, "perplexity": 40.257738570318736, "lr": 0.0026291804804649314, "grad_norm": 0.164551, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:19.077794+00:00", "epoch": 0, "step": 4884, "train_loss": 3.746561288833618, "perplexity": 42.375115392918154, "lr": 0.0026291804804649314, "grad_norm": 0.166795, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:19.382530+00:00", "epoch": 0, "step": 4885, "train_loss": 3.614088535308838, "perplexity": 37.11749921299113, "lr": 0.0026291804804649314, "grad_norm": 0.163188, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:19.686861+00:00", "epoch": 0, "step": 4886, "train_loss": 3.6620264053344727, "perplexity": 38.94017154628078, "lr": 0.0026291804804649314, "grad_norm": 0.165181, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:19.991977+00:00", "epoch": 0, "step": 4887, "train_loss": 3.5303378105163574, "perplexity": 34.135496997138134, "lr": 0.0026291804804649314, "grad_norm": 0.146447, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:20.297514+00:00", "epoch": 0, "step": 4888, "train_loss": 3.711738348007202, "perplexity": 40.9248864230515, "lr": 0.0026291804804649314, "grad_norm": 0.16309, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:20.603747+00:00", "epoch": 0, "step": 4889, "train_loss": 3.5304622650146484, "perplexity": 34.13974557766282, "lr": 0.0026291804804649314, "grad_norm": 0.156857, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:20.909216+00:00", "epoch": 0, "step": 4890, "train_loss": 3.617363929748535, "perplexity": 37.239272983225945, "lr": 0.0026291804804649314, "grad_norm": 0.155961, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:21.212710+00:00", "epoch": 0, "step": 4891, "train_loss": 3.6282315254211426, "perplexity": 37.6461813984092, "lr": 0.0026291804804649314, "grad_norm": 0.134662, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:21.516094+00:00", "epoch": 0, "step": 4892, "train_loss": 3.587826728820801, "perplexity": 36.15541496032622, "lr": 0.0026291804804649314, "grad_norm": 0.163372, "tokens_per_sec": 108009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:21.821045+00:00", "epoch": 0, "step": 4893, "train_loss": 3.682098388671875, "perplexity": 39.72967496405604, "lr": 0.0026291804804649314, "grad_norm": 0.161697, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:22.126415+00:00", "epoch": 0, "step": 4894, "train_loss": 3.7108852863311768, "perplexity": 40.88998985742591, "lr": 0.0026291804804649314, "grad_norm": 0.141938, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:22.432047+00:00", "epoch": 0, "step": 4895, "train_loss": 3.601500988006592, "perplexity": 36.653209202545895, "lr": 0.0026291804804649314, "grad_norm": 0.15017, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:22.736606+00:00", "epoch": 0, "step": 4896, "train_loss": 3.5874762535095215, "perplexity": 36.1427456002922, "lr": 0.0026291804804649314, "grad_norm": 0.139054, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:23.041261+00:00", "epoch": 0, "step": 4897, "train_loss": 3.7048189640045166, "perplexity": 40.642688861189086, "lr": 0.0026291804804649314, "grad_norm": 0.143802, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:23.345738+00:00", "epoch": 0, "step": 4898, "train_loss": 3.548225164413452, "perplexity": 34.75158437210951, "lr": 0.0026291804804649314, "grad_norm": 0.139589, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:23.651131+00:00", "epoch": 0, "step": 4899, "train_loss": 3.608880043029785, "perplexity": 36.9246756009857, "lr": 0.0026291804804649314, "grad_norm": 0.140862, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:23.956058+00:00", "epoch": 0, "step": 4900, "train_loss": 3.5722270011901855, "perplexity": 35.59577678514537, "lr": 0.0026291804804649314, "grad_norm": 0.140508, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:24.260659+00:00", "epoch": 0, "step": 4901, "train_loss": 3.7187600135803223, "perplexity": 41.21325853056309, "lr": 0.0026291804804649314, "grad_norm": 0.152639, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:24.564714+00:00", "epoch": 0, "step": 4902, "train_loss": 3.601609706878662, "perplexity": 36.65719431473268, "lr": 0.0026291804804649314, "grad_norm": 0.1378, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:24.869047+00:00", "epoch": 0, "step": 4903, "train_loss": 3.5982916355133057, "perplexity": 36.53576469542161, "lr": 0.0026291804804649314, "grad_norm": 0.146854, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:25.173556+00:00", "epoch": 0, "step": 4904, "train_loss": 3.5602331161499023, "perplexity": 35.171395209756845, "lr": 0.0026291804804649314, "grad_norm": 0.144376, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:25.479488+00:00", "epoch": 0, "step": 4905, "train_loss": 3.6070704460144043, "perplexity": 36.857917239307746, "lr": 0.0026291804804649314, "grad_norm": 0.156948, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:25.785047+00:00", "epoch": 0, "step": 4906, "train_loss": 3.609220266342163, "perplexity": 36.93724037372025, "lr": 0.0026291804804649314, "grad_norm": 0.163363, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:26.090772+00:00", "epoch": 0, "step": 4907, "train_loss": 3.6487011909484863, "perplexity": 38.42472724202113, "lr": 0.0026291804804649314, "grad_norm": 0.154086, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:26.395427+00:00", "epoch": 0, "step": 4908, "train_loss": 3.6426711082458496, "perplexity": 38.19372015483338, "lr": 0.0026291804804649314, "grad_norm": 0.148088, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:26.699762+00:00", "epoch": 0, "step": 4909, "train_loss": 3.6952648162841797, "perplexity": 40.25623168224765, "lr": 0.0026291804804649314, "grad_norm": 0.150097, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:27.005892+00:00", "epoch": 0, "step": 4910, "train_loss": 3.5916285514831543, "perplexity": 36.29313306030619, "lr": 0.0026291804804649314, "grad_norm": 0.13981, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:27.311673+00:00", "epoch": 0, "step": 4911, "train_loss": 3.5865468978881836, "perplexity": 36.10917173994239, "lr": 0.0026291804804649314, "grad_norm": 0.15075, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:27.615977+00:00", "epoch": 0, "step": 4912, "train_loss": 3.5381503105163574, "perplexity": 34.403225018277105, "lr": 0.0026291804804649314, "grad_norm": 0.152976, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:27.920518+00:00", "epoch": 0, "step": 4913, "train_loss": 3.648798942565918, "perplexity": 38.4284835048458, "lr": 0.0026291804804649314, "grad_norm": 0.150327, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:28.225872+00:00", "epoch": 0, "step": 4914, "train_loss": 3.570652484893799, "perplexity": 35.5397747541394, "lr": 0.0026291804804649314, "grad_norm": 0.156398, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:28.530989+00:00", "epoch": 0, "step": 4915, "train_loss": 3.564560651779175, "perplexity": 35.32393048853399, "lr": 0.0026291804804649314, "grad_norm": 0.152914, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:28.836855+00:00", "epoch": 0, "step": 4916, "train_loss": 3.5947842597961426, "perplexity": 36.40784450474229, "lr": 0.0026291804804649314, "grad_norm": 0.154845, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:29.141470+00:00", "epoch": 0, "step": 4917, "train_loss": 3.5511815547943115, "perplexity": 34.8544756403034, "lr": 0.0026291804804649314, "grad_norm": 0.15157, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:29.444812+00:00", "epoch": 0, "step": 4918, "train_loss": 3.6031241416931152, "perplexity": 36.71275130410414, "lr": 0.0026291804804649314, "grad_norm": 0.16767, "tokens_per_sec": 108022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:29.748700+00:00", "epoch": 0, "step": 4919, "train_loss": 3.64163875579834, "perplexity": 38.1543111198589, "lr": 0.0026291804804649314, "grad_norm": 0.158481, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:30.054270+00:00", "epoch": 0, "step": 4920, "train_loss": 3.7466609477996826, "perplexity": 42.37933866354497, "lr": 0.0026291804804649314, "grad_norm": 0.152824, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:30.359598+00:00", "epoch": 0, "step": 4921, "train_loss": 3.563164710998535, "perplexity": 35.27465477442374, "lr": 0.0026291804804649314, "grad_norm": 0.162136, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:30.664056+00:00", "epoch": 0, "step": 4922, "train_loss": 3.68237042427063, "perplexity": 39.74048432017139, "lr": 0.0026291804804649314, "grad_norm": 0.171059, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:30.968629+00:00", "epoch": 0, "step": 4923, "train_loss": 3.562005043029785, "perplexity": 35.233771597210406, "lr": 0.0026291804804649314, "grad_norm": 0.165239, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:31.273601+00:00", "epoch": 0, "step": 4924, "train_loss": 3.660731792449951, "perplexity": 38.889791716696216, "lr": 0.0026291804804649314, "grad_norm": 0.171875, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:31.577763+00:00", "epoch": 0, "step": 4925, "train_loss": 3.555828094482422, "perplexity": 35.016805187954915, "lr": 0.0026291804804649314, "grad_norm": 0.192422, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:31.882435+00:00", "epoch": 0, "step": 4926, "train_loss": 3.625365734100342, "perplexity": 37.53844974045269, "lr": 0.0026291804804649314, "grad_norm": 0.190375, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:32.187353+00:00", "epoch": 0, "step": 4927, "train_loss": 3.6876585483551025, "perplexity": 39.95119356973515, "lr": 0.0026291804804649314, "grad_norm": 0.160342, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:32.494719+00:00", "epoch": 0, "step": 4928, "train_loss": 3.5812981128692627, "perplexity": 35.920138990383585, "lr": 0.0026291804804649314, "grad_norm": 0.159944, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:32.801510+00:00", "epoch": 0, "step": 4929, "train_loss": 3.596385955810547, "perplexity": 36.466205530012616, "lr": 0.0026291804804649314, "grad_norm": 0.158417, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:33.106283+00:00", "epoch": 0, "step": 4930, "train_loss": 3.5629019737243652, "perplexity": 35.26538802519454, "lr": 0.0026291804804649314, "grad_norm": 0.148582, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:33.410800+00:00", "epoch": 0, "step": 4931, "train_loss": 3.669239044189453, "perplexity": 39.22204825606455, "lr": 0.0026291804804649314, "grad_norm": 0.149528, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:33.715695+00:00", "epoch": 0, "step": 4932, "train_loss": 3.628783702850342, "perplexity": 37.66697451028815, "lr": 0.0026291804804649314, "grad_norm": 0.150308, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:34.020683+00:00", "epoch": 0, "step": 4933, "train_loss": 3.619196653366089, "perplexity": 37.30758485759466, "lr": 0.0026291804804649314, "grad_norm": 0.154649, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:34.325517+00:00", "epoch": 0, "step": 4934, "train_loss": 3.5518782138824463, "perplexity": 34.87876578751389, "lr": 0.0026291804804649314, "grad_norm": 0.147797, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:34.630064+00:00", "epoch": 0, "step": 4935, "train_loss": 3.5623867511749268, "perplexity": 35.24722318193952, "lr": 0.0026291804804649314, "grad_norm": 0.171008, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:34.934958+00:00", "epoch": 0, "step": 4936, "train_loss": 3.562267303466797, "perplexity": 35.24301323335185, "lr": 0.0026291804804649314, "grad_norm": 0.161057, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:35.240370+00:00", "epoch": 0, "step": 4937, "train_loss": 3.613426923751831, "perplexity": 37.09294996847206, "lr": 0.0026291804804649314, "grad_norm": 0.182482, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:35.547798+00:00", "epoch": 0, "step": 4938, "train_loss": 3.5883069038391113, "perplexity": 36.172780056175775, "lr": 0.0026291804804649314, "grad_norm": 0.176079, "tokens_per_sec": 106586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:35.853190+00:00", "epoch": 0, "step": 4939, "train_loss": 3.706536293029785, "perplexity": 40.71254569685252, "lr": 0.0026291804804649314, "grad_norm": 0.161658, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:36.158722+00:00", "epoch": 0, "step": 4940, "train_loss": 3.6421775817871094, "perplexity": 38.17487519400514, "lr": 0.0026291804804649314, "grad_norm": 0.192509, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:36.463006+00:00", "epoch": 0, "step": 4941, "train_loss": 3.54707932472229, "perplexity": 34.71178743221612, "lr": 0.0026291804804649314, "grad_norm": 0.167731, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:36.768180+00:00", "epoch": 0, "step": 4942, "train_loss": 3.542337656021118, "perplexity": 34.54758524003686, "lr": 0.0026291804804649314, "grad_norm": 0.180977, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:37.074271+00:00", "epoch": 0, "step": 4943, "train_loss": 3.586400270462036, "perplexity": 36.10387753317729, "lr": 0.0026291804804649314, "grad_norm": 0.173279, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:37.380710+00:00", "epoch": 0, "step": 4944, "train_loss": 3.6117565631866455, "perplexity": 37.03104308540031, "lr": 0.0026291804804649314, "grad_norm": 0.150259, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:37.686345+00:00", "epoch": 0, "step": 4945, "train_loss": 3.548614263534546, "perplexity": 34.76510881404892, "lr": 0.0026291804804649314, "grad_norm": 0.171192, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:37.990430+00:00", "epoch": 0, "step": 4946, "train_loss": 3.7154712677001953, "perplexity": 41.077941230435016, "lr": 0.0026291804804649314, "grad_norm": 0.150851, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:38.296261+00:00", "epoch": 0, "step": 4947, "train_loss": 3.6235952377319336, "perplexity": 37.47204685189393, "lr": 0.0026291804804649314, "grad_norm": 0.152967, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:38.601621+00:00", "epoch": 0, "step": 4948, "train_loss": 3.688469171524048, "perplexity": 39.98359206258016, "lr": 0.0026291804804649314, "grad_norm": 0.173727, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:38.907719+00:00", "epoch": 0, "step": 4949, "train_loss": 3.6066856384277344, "perplexity": 36.84373676167788, "lr": 0.0026291804804649314, "grad_norm": 0.154879, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:39.212909+00:00", "epoch": 0, "step": 4950, "train_loss": 3.7433221340179443, "perplexity": 42.23807789647955, "lr": 0.0026291804804649314, "grad_norm": 0.154646, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:39.517456+00:00", "epoch": 0, "step": 4951, "train_loss": 3.629040002822876, "perplexity": 37.67662979209196, "lr": 0.0026291804804649314, "grad_norm": 0.150374, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:39.821793+00:00", "epoch": 0, "step": 4952, "train_loss": 3.562964916229248, "perplexity": 35.26760778691046, "lr": 0.0026291804804649314, "grad_norm": 0.155577, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:40.126662+00:00", "epoch": 0, "step": 4953, "train_loss": 3.532588481903076, "perplexity": 34.212411305449905, "lr": 0.0026291804804649314, "grad_norm": 0.150346, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:40.431851+00:00", "epoch": 0, "step": 4954, "train_loss": 3.5588107109069824, "perplexity": 35.121402795974284, "lr": 0.0026291804804649314, "grad_norm": 0.198767, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:40.737316+00:00", "epoch": 0, "step": 4955, "train_loss": 3.7504913806915283, "perplexity": 42.54198117304502, "lr": 0.0026291804804649314, "grad_norm": 0.206617, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:41.042077+00:00", "epoch": 0, "step": 4956, "train_loss": 3.644237518310547, "perplexity": 38.25359406379653, "lr": 0.0026291804804649314, "grad_norm": 0.204284, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:41.346235+00:00", "epoch": 0, "step": 4957, "train_loss": 3.605971336364746, "perplexity": 36.81742860160592, "lr": 0.0026291804804649314, "grad_norm": 0.175716, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:41.651481+00:00", "epoch": 0, "step": 4958, "train_loss": 3.6872870922088623, "perplexity": 39.93635620921903, "lr": 0.0026291804804649314, "grad_norm": 0.165864, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:41.956531+00:00", "epoch": 0, "step": 4959, "train_loss": 3.6207711696624756, "perplexity": 37.36637252685837, "lr": 0.0026291804804649314, "grad_norm": 0.145494, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:42.262945+00:00", "epoch": 0, "step": 4960, "train_loss": 3.5598320960998535, "perplexity": 35.15729360279209, "lr": 0.0026291804804649314, "grad_norm": 0.141943, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:42.568054+00:00", "epoch": 0, "step": 4961, "train_loss": 3.6220695972442627, "perplexity": 37.414921567464475, "lr": 0.0026291804804649314, "grad_norm": 0.151508, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:42.873422+00:00", "epoch": 0, "step": 4962, "train_loss": 3.641545057296753, "perplexity": 38.15073628555882, "lr": 0.0026291804804649314, "grad_norm": 0.164209, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:43.179172+00:00", "epoch": 0, "step": 4963, "train_loss": 3.6594882011413574, "perplexity": 38.84145876917003, "lr": 0.0026291804804649314, "grad_norm": 0.147776, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:43.484358+00:00", "epoch": 0, "step": 4964, "train_loss": 3.6679937839508057, "perplexity": 39.17323699656081, "lr": 0.0026291804804649314, "grad_norm": 0.146259, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:43.789789+00:00", "epoch": 0, "step": 4965, "train_loss": 3.622950553894043, "perplexity": 37.447897014257265, "lr": 0.0026291804804649314, "grad_norm": 0.149976, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:44.096047+00:00", "epoch": 0, "step": 4966, "train_loss": 3.6483607292175293, "perplexity": 38.41164731960561, "lr": 0.0026291804804649314, "grad_norm": 0.156568, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:44.400550+00:00", "epoch": 0, "step": 4967, "train_loss": 3.6602280139923096, "perplexity": 38.87020481155186, "lr": 0.0026291804804649314, "grad_norm": 0.144483, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:44.704275+00:00", "epoch": 0, "step": 4968, "train_loss": 3.6895792484283447, "perplexity": 40.02800156910304, "lr": 0.0026291804804649314, "grad_norm": 0.171454, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:45.009936+00:00", "epoch": 0, "step": 4969, "train_loss": 3.566890239715576, "perplexity": 35.40631661647226, "lr": 0.0026291804804649314, "grad_norm": 0.174558, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:45.314979+00:00", "epoch": 0, "step": 4970, "train_loss": 3.6136856079101562, "perplexity": 37.1025465682051, "lr": 0.0026291804804649314, "grad_norm": 0.166005, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:45.619608+00:00", "epoch": 0, "step": 4971, "train_loss": 3.6178040504455566, "perplexity": 37.25566636527663, "lr": 0.0026291804804649314, "grad_norm": 0.185754, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:45.923879+00:00", "epoch": 0, "step": 4972, "train_loss": 3.6526031494140625, "perplexity": 38.574951826213706, "lr": 0.0026291804804649314, "grad_norm": 0.194093, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:46.228133+00:00", "epoch": 0, "step": 4973, "train_loss": 3.6153016090393066, "perplexity": 37.16255279736206, "lr": 0.0026291804804649314, "grad_norm": 0.167325, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:46.534297+00:00", "epoch": 0, "step": 4974, "train_loss": 3.567617893218994, "perplexity": 35.43208952253687, "lr": 0.0026291804804649314, "grad_norm": 0.158447, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:46.839944+00:00", "epoch": 0, "step": 4975, "train_loss": 3.5894691944122314, "perplexity": 36.21484778016229, "lr": 0.0026291804804649314, "grad_norm": 0.159886, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:47.145679+00:00", "epoch": 0, "step": 4976, "train_loss": 3.7035415172576904, "perplexity": 40.59080313820101, "lr": 0.0026291804804649314, "grad_norm": 0.164058, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:47.450645+00:00", "epoch": 0, "step": 4977, "train_loss": 3.5748214721679688, "perplexity": 35.68824890118544, "lr": 0.0026291804804649314, "grad_norm": 0.158017, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:47.755574+00:00", "epoch": 0, "step": 4978, "train_loss": 3.752951145172119, "perplexity": 42.646753231736284, "lr": 0.0026291804804649314, "grad_norm": 0.161622, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:48.059806+00:00", "epoch": 0, "step": 4979, "train_loss": 3.519519567489624, "perplexity": 33.76820122442986, "lr": 0.0026291804804649314, "grad_norm": 0.164426, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:48.365340+00:00", "epoch": 0, "step": 4980, "train_loss": 3.6144628524780273, "perplexity": 37.13139553087686, "lr": 0.0026291804804649314, "grad_norm": 0.144629, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:48.670522+00:00", "epoch": 0, "step": 4981, "train_loss": 3.6046857833862305, "perplexity": 36.77012825666973, "lr": 0.0026291804804649314, "grad_norm": 0.132971, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:48.976355+00:00", "epoch": 0, "step": 4982, "train_loss": 3.5808351039886475, "perplexity": 35.90351149667377, "lr": 0.0026291804804649314, "grad_norm": 0.147291, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:49.281823+00:00", "epoch": 0, "step": 4983, "train_loss": 3.6847500801086426, "perplexity": 39.83516560544617, "lr": 0.0026291804804649314, "grad_norm": 0.152852, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:49.585521+00:00", "epoch": 0, "step": 4984, "train_loss": 3.6264498233795166, "perplexity": 37.57916683786983, "lr": 0.0026291804804649314, "grad_norm": 0.143226, "tokens_per_sec": 107896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:49.890953+00:00", "epoch": 0, "step": 4985, "train_loss": 3.5697901248931885, "perplexity": 35.50913988500212, "lr": 0.0026291804804649314, "grad_norm": 0.148306, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:50.196906+00:00", "epoch": 0, "step": 4986, "train_loss": 3.6628780364990234, "perplexity": 38.973348335111005, "lr": 0.0026291804804649314, "grad_norm": 0.150767, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:50.502316+00:00", "epoch": 0, "step": 4987, "train_loss": 3.5166079998016357, "perplexity": 33.67002581239663, "lr": 0.0026291804804649314, "grad_norm": 0.140576, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:50.807469+00:00", "epoch": 0, "step": 4988, "train_loss": 3.5540285110473633, "perplexity": 34.95384619236384, "lr": 0.0026291804804649314, "grad_norm": 0.144129, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:51.111706+00:00", "epoch": 0, "step": 4989, "train_loss": 3.6411852836608887, "perplexity": 38.13701312521813, "lr": 0.0026291804804649314, "grad_norm": 0.144185, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:51.416661+00:00", "epoch": 0, "step": 4990, "train_loss": 3.5467686653137207, "perplexity": 34.7010055636932, "lr": 0.0026291804804649314, "grad_norm": 0.152254, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:51.721563+00:00", "epoch": 0, "step": 4991, "train_loss": 3.5469696521759033, "perplexity": 34.70798071084902, "lr": 0.0026291804804649314, "grad_norm": 0.157032, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:52.028954+00:00", "epoch": 0, "step": 4992, "train_loss": 3.5264177322387695, "perplexity": 34.00194511480584, "lr": 0.0026291804804649314, "grad_norm": 0.177092, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:52.333510+00:00", "epoch": 0, "step": 4993, "train_loss": 3.5420401096343994, "perplexity": 34.53730726004256, "lr": 0.0026291804804649314, "grad_norm": 0.155221, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:52.637285+00:00", "epoch": 0, "step": 4994, "train_loss": 3.627392530441284, "perplexity": 37.614609687310626, "lr": 0.0026291804804649314, "grad_norm": 0.172034, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:52.942088+00:00", "epoch": 0, "step": 4995, "train_loss": 3.6344797611236572, "perplexity": 37.88214000792989, "lr": 0.0026291804804649314, "grad_norm": 0.168304, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:53.247100+00:00", "epoch": 0, "step": 4996, "train_loss": 3.6826229095458984, "perplexity": 39.75051947410526, "lr": 0.0026291804804649314, "grad_norm": 0.158931, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:53.553237+00:00", "epoch": 0, "step": 4997, "train_loss": 3.6108314990997314, "perplexity": 36.996802836994846, "lr": 0.0026291804804649314, "grad_norm": 0.161091, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:53.858321+00:00", "epoch": 0, "step": 4998, "train_loss": 3.581968307495117, "perplexity": 35.944220543257174, "lr": 0.0026291804804649314, "grad_norm": 0.178558, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:54.162165+00:00", "epoch": 0, "step": 4999, "train_loss": 3.6469740867614746, "perplexity": 38.35842101008406, "lr": 0.0026291804804649314, "grad_norm": 0.159857, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:23:54.466751+00:00", "epoch": 0, "step": 5000, "train_loss": 3.6726670265197754, "perplexity": 39.35673145837503, "lr": 0.0026291804804649314, "grad_norm": 0.161445, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:23:57.539134+00:00", "step": 5000, "epoch": 0, "val_loss": 3.5620506167411805, "val_ppl": 35.235377367538725, "eval_train_loss": 3.6726670265197754, "eval_train_ppl": 39.35673145837503} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:23:58.434643+00:00", "step": 5000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5621_epoch_0000_step_0005000.pt", "val_loss": 3.5620506167411805} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T21:24:00.321517+00:00", "step": 5000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0005000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:01.378789+00:00", "epoch": 0, "step": 5001, "train_loss": 3.598498821258545, "perplexity": 36.543335169277974, "lr": 0.0026291804804649314, "grad_norm": 0.175197, "tokens_per_sec": 4741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:01.681698+00:00", "epoch": 0, "step": 5002, "train_loss": 3.5859248638153076, "perplexity": 36.08671758912427, "lr": 0.0026291804804649314, "grad_norm": 0.163192, "tokens_per_sec": 108172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:01.985361+00:00", "epoch": 0, "step": 5003, "train_loss": 3.613353729248047, "perplexity": 37.09023506776435, "lr": 0.0026291804804649314, "grad_norm": 0.149267, "tokens_per_sec": 107909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:02.289655+00:00", "epoch": 0, "step": 5004, "train_loss": 3.6091208457946777, "perplexity": 36.93356823560571, "lr": 0.0026291804804649314, "grad_norm": 0.174515, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:02.594423+00:00", "epoch": 0, "step": 5005, "train_loss": 3.606175661087036, "perplexity": 36.82495208107063, "lr": 0.0026291804804649314, "grad_norm": 0.160964, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:02.898611+00:00", "epoch": 0, "step": 5006, "train_loss": 3.507100820541382, "perplexity": 33.35143568769628, "lr": 0.0026291804804649314, "grad_norm": 0.152556, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:03.202496+00:00", "epoch": 0, "step": 5007, "train_loss": 3.6302437782287598, "perplexity": 37.72201130150462, "lr": 0.0026291804804649314, "grad_norm": 0.146304, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:03.506252+00:00", "epoch": 0, "step": 5008, "train_loss": 3.582542657852173, "perplexity": 35.964871048905245, "lr": 0.0026291804804649314, "grad_norm": 0.146965, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:03.810229+00:00", "epoch": 0, "step": 5009, "train_loss": 3.5402941703796387, "perplexity": 34.4770598290253, "lr": 0.0026291804804649314, "grad_norm": 0.150141, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:04.115009+00:00", "epoch": 0, "step": 5010, "train_loss": 3.6918911933898926, "perplexity": 40.12065116476614, "lr": 0.0026291804804649314, "grad_norm": 0.165475, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:04.419314+00:00", "epoch": 0, "step": 5011, "train_loss": 3.62438702583313, "perplexity": 37.50172852196463, "lr": 0.0026291804804649314, "grad_norm": 0.145827, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:04.722983+00:00", "epoch": 0, "step": 5012, "train_loss": 3.567941904067993, "perplexity": 35.44357176402983, "lr": 0.0026291804804649314, "grad_norm": 0.159, "tokens_per_sec": 107907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:05.027773+00:00", "epoch": 0, "step": 5013, "train_loss": 3.594961166381836, "perplexity": 36.414285861948535, "lr": 0.0026291804804649314, "grad_norm": 0.152876, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:05.334286+00:00", "epoch": 0, "step": 5014, "train_loss": 3.540195941925049, "perplexity": 34.47367336704568, "lr": 0.0026291804804649314, "grad_norm": 0.164018, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:05.814238+00:00", "epoch": 0, "step": 5015, "train_loss": 3.6590752601623535, "perplexity": 38.82542285033183, "lr": 0.0026291804804649314, "grad_norm": 0.147835, "tokens_per_sec": 68273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:06.117912+00:00", "epoch": 0, "step": 5016, "train_loss": 3.669510841369629, "perplexity": 39.23271014705158, "lr": 0.0026291804804649314, "grad_norm": 0.17252, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:06.422197+00:00", "epoch": 0, "step": 5017, "train_loss": 3.555241823196411, "perplexity": 34.996281857252974, "lr": 0.0026291804804649314, "grad_norm": 0.176994, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:06.726925+00:00", "epoch": 0, "step": 5018, "train_loss": 3.5213699340820312, "perplexity": 33.830742620271586, "lr": 0.0026291804804649314, "grad_norm": 0.150146, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:07.032508+00:00", "epoch": 0, "step": 5019, "train_loss": 3.632368564605713, "perplexity": 37.80224772967703, "lr": 0.0026291804804649314, "grad_norm": 0.173208, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:07.337011+00:00", "epoch": 0, "step": 5020, "train_loss": 3.6670942306518555, "perplexity": 39.138014426654635, "lr": 0.0026291804804649314, "grad_norm": 0.155813, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:07.641430+00:00", "epoch": 0, "step": 5021, "train_loss": 3.695540189743042, "perplexity": 40.2673187064728, "lr": 0.0026291804804649314, "grad_norm": 0.130873, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:07.947039+00:00", "epoch": 0, "step": 5022, "train_loss": 3.566976308822632, "perplexity": 35.4093641376744, "lr": 0.0026291804804649314, "grad_norm": 0.140603, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:08.252786+00:00", "epoch": 0, "step": 5023, "train_loss": 3.6081957817077637, "perplexity": 36.899418115989285, "lr": 0.0026291804804649314, "grad_norm": 0.134061, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:08.556724+00:00", "epoch": 0, "step": 5024, "train_loss": 3.470177173614502, "perplexity": 32.14243673199424, "lr": 0.0026291804804649314, "grad_norm": 0.153656, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:08.861198+00:00", "epoch": 0, "step": 5025, "train_loss": 3.630013942718506, "perplexity": 37.7133424400336, "lr": 0.0026291804804649314, "grad_norm": 0.158619, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:09.165788+00:00", "epoch": 0, "step": 5026, "train_loss": 3.6033527851104736, "perplexity": 36.72114639272724, "lr": 0.0026291804804649314, "grad_norm": 0.160591, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:09.470732+00:00", "epoch": 0, "step": 5027, "train_loss": 3.619842529296875, "perplexity": 37.33168871190169, "lr": 0.0026291804804649314, "grad_norm": 0.166499, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:09.776525+00:00", "epoch": 0, "step": 5028, "train_loss": 3.6028811931610107, "perplexity": 36.70383307844407, "lr": 0.0026291804804649314, "grad_norm": 0.161668, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:10.081118+00:00", "epoch": 0, "step": 5029, "train_loss": 3.557093620300293, "perplexity": 35.06114791149226, "lr": 0.0026291804804649314, "grad_norm": 0.155975, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:10.385904+00:00", "epoch": 0, "step": 5030, "train_loss": 3.616513252258301, "perplexity": 37.20760784226728, "lr": 0.0026291804804649314, "grad_norm": 0.163637, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:10.690091+00:00", "epoch": 0, "step": 5031, "train_loss": 3.558588981628418, "perplexity": 35.11361621595845, "lr": 0.0026291804804649314, "grad_norm": 0.187865, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:10.994760+00:00", "epoch": 0, "step": 5032, "train_loss": 3.58245587348938, "perplexity": 35.96174999591959, "lr": 0.0026291804804649314, "grad_norm": 0.162278, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:11.299389+00:00", "epoch": 0, "step": 5033, "train_loss": 3.597399950027466, "perplexity": 36.50320080486057, "lr": 0.0026291804804649314, "grad_norm": 0.141752, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:11.603526+00:00", "epoch": 0, "step": 5034, "train_loss": 3.5323517322540283, "perplexity": 34.2043124878142, "lr": 0.0026291804804649314, "grad_norm": 0.152353, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:11.907516+00:00", "epoch": 0, "step": 5035, "train_loss": 3.5753185749053955, "perplexity": 35.70599403762071, "lr": 0.0026291804804649314, "grad_norm": 0.14558, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:12.212589+00:00", "epoch": 0, "step": 5036, "train_loss": 3.5927884578704834, "perplexity": 36.3352541206801, "lr": 0.0026291804804649314, "grad_norm": 0.169243, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:12.518290+00:00", "epoch": 0, "step": 5037, "train_loss": 3.581477642059326, "perplexity": 35.92658828274429, "lr": 0.0026291804804649314, "grad_norm": 0.144476, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:12.822858+00:00", "epoch": 0, "step": 5038, "train_loss": 3.666388511657715, "perplexity": 39.11040373032315, "lr": 0.0026291804804649314, "grad_norm": 0.158085, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:13.127296+00:00", "epoch": 0, "step": 5039, "train_loss": 3.683358669281006, "perplexity": 39.77977706774372, "lr": 0.0026291804804649314, "grad_norm": 0.139172, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:13.432093+00:00", "epoch": 0, "step": 5040, "train_loss": 3.4790005683898926, "perplexity": 32.42729700958299, "lr": 0.0026291804804649314, "grad_norm": 0.1608, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:13.736452+00:00", "epoch": 0, "step": 5041, "train_loss": 3.6105074882507324, "perplexity": 36.98481741330588, "lr": 0.0026291804804649314, "grad_norm": 0.159031, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:14.040399+00:00", "epoch": 0, "step": 5042, "train_loss": 3.598745822906494, "perplexity": 36.552362548129175, "lr": 0.0026291804804649314, "grad_norm": 0.165981, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:14.345552+00:00", "epoch": 0, "step": 5043, "train_loss": 3.500091314315796, "perplexity": 33.11847601159776, "lr": 0.0026291804804649314, "grad_norm": 0.147318, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:14.650222+00:00", "epoch": 0, "step": 5044, "train_loss": 3.725344181060791, "perplexity": 41.48550881434463, "lr": 0.0026291804804649314, "grad_norm": 0.160161, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:14.954429+00:00", "epoch": 0, "step": 5045, "train_loss": 3.620162010192871, "perplexity": 37.343617378649526, "lr": 0.0026291804804649314, "grad_norm": 0.145809, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:15.259244+00:00", "epoch": 0, "step": 5046, "train_loss": 3.5461530685424805, "perplexity": 34.679650310496264, "lr": 0.0026291804804649314, "grad_norm": 0.163442, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:15.563987+00:00", "epoch": 0, "step": 5047, "train_loss": 3.588491439819336, "perplexity": 36.17945585154417, "lr": 0.0026291804804649314, "grad_norm": 0.172307, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:15.868986+00:00", "epoch": 0, "step": 5048, "train_loss": 3.6444153785705566, "perplexity": 38.260398463081195, "lr": 0.0026291804804649314, "grad_norm": 0.212416, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:16.173561+00:00", "epoch": 0, "step": 5049, "train_loss": 3.698969841003418, "perplexity": 40.405658660122576, "lr": 0.0026291804804649314, "grad_norm": 0.228365, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:16.477301+00:00", "epoch": 0, "step": 5050, "train_loss": 3.591991424560547, "perplexity": 36.30630525096014, "lr": 0.0026291804804649314, "grad_norm": 0.179081, "tokens_per_sec": 107939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:16.780900+00:00", "epoch": 0, "step": 5051, "train_loss": 3.611600637435913, "perplexity": 37.025269442348325, "lr": 0.0026291804804649314, "grad_norm": 0.163691, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:17.086248+00:00", "epoch": 0, "step": 5052, "train_loss": 3.5944581031799316, "perplexity": 36.39597178166393, "lr": 0.0026291804804649314, "grad_norm": 0.183984, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:17.391459+00:00", "epoch": 0, "step": 5053, "train_loss": 3.6256015300750732, "perplexity": 37.547302199445234, "lr": 0.0026291804804649314, "grad_norm": 0.159884, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:17.697570+00:00", "epoch": 0, "step": 5054, "train_loss": 3.6263327598571777, "perplexity": 37.574767945713155, "lr": 0.0026291804804649314, "grad_norm": 0.145589, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:18.002146+00:00", "epoch": 0, "step": 5055, "train_loss": 3.7137699127197266, "perplexity": 41.00811248911753, "lr": 0.0026291804804649314, "grad_norm": 0.149684, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:18.307589+00:00", "epoch": 0, "step": 5056, "train_loss": 3.622880458831787, "perplexity": 37.44527219357926, "lr": 0.0026291804804649314, "grad_norm": 0.153484, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:18.612578+00:00", "epoch": 0, "step": 5057, "train_loss": 3.540074110031128, "perplexity": 34.469473629965115, "lr": 0.0026291804804649314, "grad_norm": 0.159892, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:18.917802+00:00", "epoch": 0, "step": 5058, "train_loss": 3.6158885955810547, "perplexity": 37.18437311920222, "lr": 0.0026291804804649314, "grad_norm": 0.160229, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:19.223238+00:00", "epoch": 0, "step": 5059, "train_loss": 3.622689962387085, "perplexity": 37.43813968173612, "lr": 0.0026291804804649314, "grad_norm": 0.137153, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:19.528599+00:00", "epoch": 0, "step": 5060, "train_loss": 3.627187728881836, "perplexity": 37.606906945382335, "lr": 0.0026291804804649314, "grad_norm": 0.176375, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:19.833258+00:00", "epoch": 0, "step": 5061, "train_loss": 3.627964496612549, "perplexity": 37.63613012549172, "lr": 0.0026291804804649314, "grad_norm": 0.166546, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:20.137878+00:00", "epoch": 0, "step": 5062, "train_loss": 3.6383349895477295, "perplexity": 38.02846619004687, "lr": 0.0026291804804649314, "grad_norm": 0.14469, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:20.443081+00:00", "epoch": 0, "step": 5063, "train_loss": 3.605660915374756, "perplexity": 36.80600147267261, "lr": 0.0026291804804649314, "grad_norm": 0.151454, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:20.749117+00:00", "epoch": 0, "step": 5064, "train_loss": 3.683708429336548, "perplexity": 39.793692878235845, "lr": 0.0026291804804649314, "grad_norm": 0.172874, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:21.054885+00:00", "epoch": 0, "step": 5065, "train_loss": 3.675931930541992, "perplexity": 39.485437401160475, "lr": 0.0026291804804649314, "grad_norm": 0.156178, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:21.359683+00:00", "epoch": 0, "step": 5066, "train_loss": 3.488161087036133, "perplexity": 32.725712602493054, "lr": 0.0026291804804649314, "grad_norm": 0.171949, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:21.663974+00:00", "epoch": 0, "step": 5067, "train_loss": 3.621581554412842, "perplexity": 37.39666593833212, "lr": 0.0026291804804649314, "grad_norm": 0.167407, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:21.968582+00:00", "epoch": 0, "step": 5068, "train_loss": 3.6294877529144287, "perplexity": 37.69350328380259, "lr": 0.0026291804804649314, "grad_norm": 0.16046, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:22.273937+00:00", "epoch": 0, "step": 5069, "train_loss": 3.6316921710968018, "perplexity": 37.776687180159826, "lr": 0.0026291804804649314, "grad_norm": 0.158763, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:22.578867+00:00", "epoch": 0, "step": 5070, "train_loss": 3.54245924949646, "perplexity": 34.551786256393775, "lr": 0.0026291804804649314, "grad_norm": 0.143259, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:22.883058+00:00", "epoch": 0, "step": 5071, "train_loss": 3.7212305068969727, "perplexity": 41.31520148316242, "lr": 0.0026291804804649314, "grad_norm": 0.148725, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:23.187482+00:00", "epoch": 0, "step": 5072, "train_loss": 3.6069653034210205, "perplexity": 36.85404210602683, "lr": 0.0026291804804649314, "grad_norm": 0.160346, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:23.491886+00:00", "epoch": 0, "step": 5073, "train_loss": 3.572402000427246, "perplexity": 35.602006564012726, "lr": 0.0026291804804649314, "grad_norm": 0.145781, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:23.798563+00:00", "epoch": 0, "step": 5074, "train_loss": 3.669139862060547, "perplexity": 39.21815832272757, "lr": 0.0026291804804649314, "grad_norm": 0.155493, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:24.104053+00:00", "epoch": 0, "step": 5075, "train_loss": 3.577028512954712, "perplexity": 35.76710130534616, "lr": 0.0026291804804649314, "grad_norm": 0.150792, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:24.408702+00:00", "epoch": 0, "step": 5076, "train_loss": 3.591365098953247, "perplexity": 36.28357280197561, "lr": 0.0026291804804649314, "grad_norm": 0.161394, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:24.712396+00:00", "epoch": 0, "step": 5077, "train_loss": 3.6054115295410156, "perplexity": 36.7968237217568, "lr": 0.0026291804804649314, "grad_norm": 0.16823, "tokens_per_sec": 107895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:25.016744+00:00", "epoch": 0, "step": 5078, "train_loss": 3.508227825164795, "perplexity": 33.38904409835947, "lr": 0.0026291804804649314, "grad_norm": 0.167941, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:25.321456+00:00", "epoch": 0, "step": 5079, "train_loss": 3.7069694995880127, "perplexity": 40.730186459421674, "lr": 0.0026291804804649314, "grad_norm": 0.157957, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:25.626819+00:00", "epoch": 0, "step": 5080, "train_loss": 3.56477952003479, "perplexity": 35.33166262170987, "lr": 0.0026291804804649314, "grad_norm": 0.186505, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:25.932598+00:00", "epoch": 0, "step": 5081, "train_loss": 3.5638229846954346, "perplexity": 35.297882796188055, "lr": 0.0026291804804649314, "grad_norm": 0.167987, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:26.237561+00:00", "epoch": 0, "step": 5082, "train_loss": 3.660794734954834, "perplexity": 38.89223961463884, "lr": 0.0026291804804649314, "grad_norm": 0.176847, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:26.542405+00:00", "epoch": 0, "step": 5083, "train_loss": 3.55025577545166, "perplexity": 34.82222301846574, "lr": 0.0026291804804649314, "grad_norm": 0.171504, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:26.847349+00:00", "epoch": 0, "step": 5084, "train_loss": 3.6415724754333496, "perplexity": 38.15178232199768, "lr": 0.0026291804804649314, "grad_norm": 0.158922, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:27.153719+00:00", "epoch": 0, "step": 5085, "train_loss": 3.60447096824646, "perplexity": 36.762230324756935, "lr": 0.0026291804804649314, "grad_norm": 0.170194, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:27.458924+00:00", "epoch": 0, "step": 5086, "train_loss": 3.596597671508789, "perplexity": 36.47392681550848, "lr": 0.0026291804804649314, "grad_norm": 0.177453, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:27.762628+00:00", "epoch": 0, "step": 5087, "train_loss": 3.6617608070373535, "perplexity": 38.92983047637463, "lr": 0.0026291804804649314, "grad_norm": 0.158455, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:28.066931+00:00", "epoch": 0, "step": 5088, "train_loss": 3.507859230041504, "perplexity": 33.37673932741206, "lr": 0.0026291804804649314, "grad_norm": 0.1646, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:28.374423+00:00", "epoch": 0, "step": 5089, "train_loss": 3.606764793395996, "perplexity": 36.84665324191732, "lr": 0.0026291804804649314, "grad_norm": 0.143019, "tokens_per_sec": 106565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:28.679706+00:00", "epoch": 0, "step": 5090, "train_loss": 3.5399887561798096, "perplexity": 34.466531653194195, "lr": 0.0026291804804649314, "grad_norm": 0.164707, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:28.984376+00:00", "epoch": 0, "step": 5091, "train_loss": 3.552273750305176, "perplexity": 34.89256433849755, "lr": 0.0026291804804649314, "grad_norm": 0.13975, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:29.290422+00:00", "epoch": 0, "step": 5092, "train_loss": 3.6975858211517334, "perplexity": 40.34977510731027, "lr": 0.0026291804804649314, "grad_norm": 0.157568, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:29.594571+00:00", "epoch": 0, "step": 5093, "train_loss": 3.633999824523926, "perplexity": 37.86396334463583, "lr": 0.0026291804804649314, "grad_norm": 0.15311, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:29.899046+00:00", "epoch": 0, "step": 5094, "train_loss": 3.6266002655029297, "perplexity": 37.58482075280783, "lr": 0.0026291804804649314, "grad_norm": 0.167123, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:30.204193+00:00", "epoch": 0, "step": 5095, "train_loss": 3.5627481937408447, "perplexity": 35.25996533136676, "lr": 0.0026291804804649314, "grad_norm": 0.154596, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:30.509644+00:00", "epoch": 0, "step": 5096, "train_loss": 3.5463199615478516, "perplexity": 34.68543858455959, "lr": 0.0026291804804649314, "grad_norm": 0.149457, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:30.814355+00:00", "epoch": 0, "step": 5097, "train_loss": 3.6147360801696777, "perplexity": 37.14154224248369, "lr": 0.0026291804804649314, "grad_norm": 0.165786, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:31.118981+00:00", "epoch": 0, "step": 5098, "train_loss": 3.6265432834625244, "perplexity": 37.582679154049984, "lr": 0.0026291804804649314, "grad_norm": 0.185705, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:31.423913+00:00", "epoch": 0, "step": 5099, "train_loss": 3.5867958068847656, "perplexity": 36.118160756324706, "lr": 0.0026291804804649314, "grad_norm": 0.175288, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:31.728856+00:00", "epoch": 0, "step": 5100, "train_loss": 3.7185235023498535, "perplexity": 41.20351228467013, "lr": 0.0026291804804649314, "grad_norm": 0.180957, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:32.034420+00:00", "epoch": 0, "step": 5101, "train_loss": 3.523016929626465, "perplexity": 33.886507612373165, "lr": 0.0026291804804649314, "grad_norm": 0.187643, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:32.340346+00:00", "epoch": 0, "step": 5102, "train_loss": 3.601475715637207, "perplexity": 36.652282900808764, "lr": 0.0026291804804649314, "grad_norm": 0.226935, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:32.643923+00:00", "epoch": 0, "step": 5103, "train_loss": 3.6537365913391113, "perplexity": 38.618699081673526, "lr": 0.0026291804804649314, "grad_norm": 0.146034, "tokens_per_sec": 107940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:32.948260+00:00", "epoch": 0, "step": 5104, "train_loss": 3.559577703475952, "perplexity": 35.14835098413954, "lr": 0.0026291804804649314, "grad_norm": 0.182652, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:33.253572+00:00", "epoch": 0, "step": 5105, "train_loss": 3.6947619915008545, "perplexity": 40.23599493946882, "lr": 0.0026291804804649314, "grad_norm": 0.153368, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:33.559154+00:00", "epoch": 0, "step": 5106, "train_loss": 3.6119699478149414, "perplexity": 37.038945783891705, "lr": 0.0026291804804649314, "grad_norm": 0.144818, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:33.864235+00:00", "epoch": 0, "step": 5107, "train_loss": 3.663268566131592, "perplexity": 38.98857155488201, "lr": 0.0026291804804649314, "grad_norm": 0.140982, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:34.169590+00:00", "epoch": 0, "step": 5108, "train_loss": 3.636991024017334, "perplexity": 37.97739157126497, "lr": 0.0026291804804649314, "grad_norm": 0.150138, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:34.474929+00:00", "epoch": 0, "step": 5109, "train_loss": 3.719388723373413, "perplexity": 41.239177856819786, "lr": 0.0026291804804649314, "grad_norm": 0.152685, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:34.779594+00:00", "epoch": 0, "step": 5110, "train_loss": 3.5994908809661865, "perplexity": 36.579606328285166, "lr": 0.0026291804804649314, "grad_norm": 0.142035, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:35.084698+00:00", "epoch": 0, "step": 5111, "train_loss": 3.6090691089630127, "perplexity": 36.931657459232305, "lr": 0.0026291804804649314, "grad_norm": 0.147448, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:35.389799+00:00", "epoch": 0, "step": 5112, "train_loss": 3.5892810821533203, "perplexity": 36.20803596405338, "lr": 0.0026291804804649314, "grad_norm": 0.158937, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:35.694899+00:00", "epoch": 0, "step": 5113, "train_loss": 3.600548028945923, "perplexity": 36.61829683240555, "lr": 0.0026291804804649314, "grad_norm": 0.153964, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:35.999843+00:00", "epoch": 0, "step": 5114, "train_loss": 3.556097984313965, "perplexity": 35.02625714304418, "lr": 0.0026291804804649314, "grad_norm": 0.177353, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:36.308208+00:00", "epoch": 0, "step": 5115, "train_loss": 3.596395969390869, "perplexity": 36.46657068911901, "lr": 0.0026291804804649314, "grad_norm": 0.180401, "tokens_per_sec": 106271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:36.611979+00:00", "epoch": 0, "step": 5116, "train_loss": 3.5852129459381104, "perplexity": 36.06103595243602, "lr": 0.0026291804804649314, "grad_norm": 0.180188, "tokens_per_sec": 107862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:36.917019+00:00", "epoch": 0, "step": 5117, "train_loss": 3.5689010620117188, "perplexity": 35.47758405641395, "lr": 0.0026291804804649314, "grad_norm": 0.180807, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:37.223916+00:00", "epoch": 0, "step": 5118, "train_loss": 3.5073869228363037, "perplexity": 33.36097897509859, "lr": 0.0026291804804649314, "grad_norm": 0.149883, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:37.528392+00:00", "epoch": 0, "step": 5119, "train_loss": 3.6971452236175537, "perplexity": 40.33200101179215, "lr": 0.0026291804804649314, "grad_norm": 0.172203, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:37.832737+00:00", "epoch": 0, "step": 5120, "train_loss": 3.5950071811676025, "perplexity": 36.415961496063, "lr": 0.0026291804804649314, "grad_norm": 0.150676, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:38.137986+00:00", "epoch": 0, "step": 5121, "train_loss": 3.6603784561157227, "perplexity": 38.87605296759469, "lr": 0.0026291804804649314, "grad_norm": 0.164561, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:38.443432+00:00", "epoch": 0, "step": 5122, "train_loss": 3.59710431098938, "perplexity": 36.492410628764794, "lr": 0.0026291804804649314, "grad_norm": 0.165405, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:38.749899+00:00", "epoch": 0, "step": 5123, "train_loss": 3.6803371906280518, "perplexity": 39.65976471919159, "lr": 0.0026291804804649314, "grad_norm": 0.15937, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:39.055174+00:00", "epoch": 0, "step": 5124, "train_loss": 3.6500532627105713, "perplexity": 38.476715368609895, "lr": 0.0026291804804649314, "grad_norm": 0.146459, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:39.360204+00:00", "epoch": 0, "step": 5125, "train_loss": 3.6244678497314453, "perplexity": 37.50475968035071, "lr": 0.0026291804804649314, "grad_norm": 0.149215, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:39.665375+00:00", "epoch": 0, "step": 5126, "train_loss": 3.6907196044921875, "perplexity": 40.07367377975767, "lr": 0.0026291804804649314, "grad_norm": 0.141413, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:39.970038+00:00", "epoch": 0, "step": 5127, "train_loss": 3.55288028717041, "perplexity": 34.91373438464911, "lr": 0.0026291804804649314, "grad_norm": 0.136471, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:40.275935+00:00", "epoch": 0, "step": 5128, "train_loss": 3.5308189392089844, "perplexity": 34.15192451573892, "lr": 0.0026291804804649314, "grad_norm": 0.148663, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:40.580813+00:00", "epoch": 0, "step": 5129, "train_loss": 3.5766005516052246, "perplexity": 35.75179764332577, "lr": 0.0026291804804649314, "grad_norm": 0.142296, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:40.885205+00:00", "epoch": 0, "step": 5130, "train_loss": 3.6463849544525146, "perplexity": 38.3358294802998, "lr": 0.0026291804804649314, "grad_norm": 0.142855, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:41.189587+00:00", "epoch": 0, "step": 5131, "train_loss": 3.5787768363952637, "perplexity": 35.82968846231351, "lr": 0.0026291804804649314, "grad_norm": 0.14671, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:41.494576+00:00", "epoch": 0, "step": 5132, "train_loss": 3.5934178829193115, "perplexity": 36.358131638865395, "lr": 0.0026291804804649314, "grad_norm": 0.156431, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:41.800102+00:00", "epoch": 0, "step": 5133, "train_loss": 3.6302592754364014, "perplexity": 37.72259589187617, "lr": 0.0026291804804649314, "grad_norm": 0.14882, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:42.105832+00:00", "epoch": 0, "step": 5134, "train_loss": 3.5582895278930664, "perplexity": 35.10310288662731, "lr": 0.0026291804804649314, "grad_norm": 0.162398, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:42.412148+00:00", "epoch": 0, "step": 5135, "train_loss": 3.523411989212036, "perplexity": 33.89989744674403, "lr": 0.0026291804804649314, "grad_norm": 0.149497, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:42.717561+00:00", "epoch": 0, "step": 5136, "train_loss": 3.591783285140991, "perplexity": 36.29874926403564, "lr": 0.0026291804804649314, "grad_norm": 0.149324, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:43.022308+00:00", "epoch": 0, "step": 5137, "train_loss": 3.6252694129943848, "perplexity": 37.534834169588464, "lr": 0.0026291804804649314, "grad_norm": 0.173863, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:43.328477+00:00", "epoch": 0, "step": 5138, "train_loss": 3.534393548965454, "perplexity": 34.27422277236021, "lr": 0.0026291804804649314, "grad_norm": 0.156294, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:43.634411+00:00", "epoch": 0, "step": 5139, "train_loss": 3.634903907775879, "perplexity": 37.898210998780804, "lr": 0.0026291804804649314, "grad_norm": 0.177353, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:43.940042+00:00", "epoch": 0, "step": 5140, "train_loss": 3.6567232608795166, "perplexity": 38.73421278867889, "lr": 0.0026291804804649314, "grad_norm": 0.186733, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:44.245063+00:00", "epoch": 0, "step": 5141, "train_loss": 3.630014181137085, "perplexity": 37.71335143159619, "lr": 0.0026291804804649314, "grad_norm": 0.153615, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:44.550302+00:00", "epoch": 0, "step": 5142, "train_loss": 3.5116219520568848, "perplexity": 33.502563291128034, "lr": 0.0026291804804649314, "grad_norm": 0.156989, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:44.855189+00:00", "epoch": 0, "step": 5143, "train_loss": 3.549253463745117, "perplexity": 34.78733778254721, "lr": 0.0026291804804649314, "grad_norm": 0.16508, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:45.162264+00:00", "epoch": 0, "step": 5144, "train_loss": 3.5326154232025146, "perplexity": 34.213333044683765, "lr": 0.0026291804804649314, "grad_norm": 0.162734, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:45.468160+00:00", "epoch": 0, "step": 5145, "train_loss": 3.738978385925293, "perplexity": 42.05500422698579, "lr": 0.0026291804804649314, "grad_norm": 0.16143, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:45.773627+00:00", "epoch": 0, "step": 5146, "train_loss": 3.7198705673217773, "perplexity": 41.259053493198614, "lr": 0.0026291804804649314, "grad_norm": 0.157278, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:46.078348+00:00", "epoch": 0, "step": 5147, "train_loss": 3.6196060180664062, "perplexity": 37.32286039230848, "lr": 0.0026291804804649314, "grad_norm": 0.171461, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:46.382654+00:00", "epoch": 0, "step": 5148, "train_loss": 3.535661220550537, "perplexity": 34.31769878149811, "lr": 0.0026291804804649314, "grad_norm": 0.148038, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:46.688868+00:00", "epoch": 0, "step": 5149, "train_loss": 3.583880662918091, "perplexity": 36.01302443611787, "lr": 0.0026291804804649314, "grad_norm": 0.153041, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:46.994992+00:00", "epoch": 0, "step": 5150, "train_loss": 3.5104174613952637, "perplexity": 33.46223405947075, "lr": 0.0026291804804649314, "grad_norm": 0.144252, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:47.300366+00:00", "epoch": 0, "step": 5151, "train_loss": 3.5363197326660156, "perplexity": 34.34030484428728, "lr": 0.0026291804804649314, "grad_norm": 0.149525, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:47.605120+00:00", "epoch": 0, "step": 5152, "train_loss": 3.608405351638794, "perplexity": 36.90715193485855, "lr": 0.0026291804804649314, "grad_norm": 0.152991, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:47.909993+00:00", "epoch": 0, "step": 5153, "train_loss": 3.616478204727173, "perplexity": 37.206303830324565, "lr": 0.0026291804804649314, "grad_norm": 0.152733, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:48.214756+00:00", "epoch": 0, "step": 5154, "train_loss": 3.5725972652435303, "perplexity": 35.60895906205083, "lr": 0.0026291804804649314, "grad_norm": 0.147261, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:48.519709+00:00", "epoch": 0, "step": 5155, "train_loss": 3.571054458618164, "perplexity": 35.554063681454686, "lr": 0.0026291804804649314, "grad_norm": 0.169839, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:48.824707+00:00", "epoch": 0, "step": 5156, "train_loss": 3.638777017593384, "perplexity": 38.04527955435161, "lr": 0.0026291804804649314, "grad_norm": 0.156653, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:49.129892+00:00", "epoch": 0, "step": 5157, "train_loss": 3.564591646194458, "perplexity": 35.325025350072195, "lr": 0.0026291804804649314, "grad_norm": 0.178698, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:49.435941+00:00", "epoch": 0, "step": 5158, "train_loss": 3.533202648162842, "perplexity": 34.2334298679242, "lr": 0.0026291804804649314, "grad_norm": 0.187263, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:49.740004+00:00", "epoch": 0, "step": 5159, "train_loss": 3.693124771118164, "perplexity": 40.170173645121224, "lr": 0.0026291804804649314, "grad_norm": 0.155786, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:50.045704+00:00", "epoch": 0, "step": 5160, "train_loss": 3.522289752960205, "perplexity": 33.86187509191891, "lr": 0.0026291804804649314, "grad_norm": 0.178533, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:50.351467+00:00", "epoch": 0, "step": 5161, "train_loss": 3.571249008178711, "perplexity": 35.56098138181553, "lr": 0.0026291804804649314, "grad_norm": 0.151504, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:50.657000+00:00", "epoch": 0, "step": 5162, "train_loss": 3.6500563621520996, "perplexity": 38.4768346251242, "lr": 0.0026291804804649314, "grad_norm": 0.147748, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:50.961356+00:00", "epoch": 0, "step": 5163, "train_loss": 3.5468339920043945, "perplexity": 34.70327253959594, "lr": 0.0026291804804649314, "grad_norm": 0.163687, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:51.266039+00:00", "epoch": 0, "step": 5164, "train_loss": 3.6990113258361816, "perplexity": 40.40733491688417, "lr": 0.0026291804804649314, "grad_norm": 0.173509, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:51.570715+00:00", "epoch": 0, "step": 5165, "train_loss": 3.615586519241333, "perplexity": 37.173142296243874, "lr": 0.0026291804804649314, "grad_norm": 0.167274, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:51.876325+00:00", "epoch": 0, "step": 5166, "train_loss": 3.610267400741577, "perplexity": 36.97593888647091, "lr": 0.0026291804804649314, "grad_norm": 0.164206, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:52.182165+00:00", "epoch": 0, "step": 5167, "train_loss": 3.7123351097106934, "perplexity": 40.949316116616274, "lr": 0.0026291804804649314, "grad_norm": 0.185973, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:52.487057+00:00", "epoch": 0, "step": 5168, "train_loss": 3.6279969215393066, "perplexity": 37.63735049403955, "lr": 0.0026291804804649314, "grad_norm": 0.182413, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:52.791861+00:00", "epoch": 0, "step": 5169, "train_loss": 3.6310408115386963, "perplexity": 37.752088985895625, "lr": 0.0026291804804649314, "grad_norm": 0.158262, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:53.099114+00:00", "epoch": 0, "step": 5170, "train_loss": 3.5085067749023438, "perplexity": 33.39835926261862, "lr": 0.0026291804804649314, "grad_norm": 0.150036, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:53.404643+00:00", "epoch": 0, "step": 5171, "train_loss": 3.607187032699585, "perplexity": 36.86221463220587, "lr": 0.0026291804804649314, "grad_norm": 0.164875, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:53.709978+00:00", "epoch": 0, "step": 5172, "train_loss": 3.5311758518218994, "perplexity": 34.164115943861724, "lr": 0.0026291804804649314, "grad_norm": 0.154637, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:54.014023+00:00", "epoch": 0, "step": 5173, "train_loss": 3.6279149055480957, "perplexity": 37.6342637560149, "lr": 0.0026291804804649314, "grad_norm": 0.167935, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:54.318749+00:00", "epoch": 0, "step": 5174, "train_loss": 3.67620849609375, "perplexity": 39.49635922317198, "lr": 0.0026291804804649314, "grad_norm": 0.172761, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:54.623076+00:00", "epoch": 0, "step": 5175, "train_loss": 3.5343446731567383, "perplexity": 34.27254763294133, "lr": 0.0026291804804649314, "grad_norm": 0.169296, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:54.927637+00:00", "epoch": 0, "step": 5176, "train_loss": 3.6167075634002686, "perplexity": 37.21483839750297, "lr": 0.0026291804804649314, "grad_norm": 0.190877, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:55.233746+00:00", "epoch": 0, "step": 5177, "train_loss": 3.58398699760437, "perplexity": 36.01685407338128, "lr": 0.0026291804804649314, "grad_norm": 0.16756, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:55.540292+00:00", "epoch": 0, "step": 5178, "train_loss": 3.5288281440734863, "perplexity": 34.084002662225565, "lr": 0.0026291804804649314, "grad_norm": 0.191647, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:55.845813+00:00", "epoch": 0, "step": 5179, "train_loss": 3.6130857467651367, "perplexity": 37.080296866170606, "lr": 0.0026291804804649314, "grad_norm": 0.167541, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:56.150456+00:00", "epoch": 0, "step": 5180, "train_loss": 3.57015323638916, "perplexity": 35.5220360031291, "lr": 0.0026291804804649314, "grad_norm": 0.166426, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:56.454968+00:00", "epoch": 0, "step": 5181, "train_loss": 3.7098259925842285, "perplexity": 40.846698280153845, "lr": 0.0026291804804649314, "grad_norm": 0.154496, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:56.760740+00:00", "epoch": 0, "step": 5182, "train_loss": 3.5860109329223633, "perplexity": 36.08982367435081, "lr": 0.0026291804804649314, "grad_norm": 0.153114, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:57.068213+00:00", "epoch": 0, "step": 5183, "train_loss": 3.5231378078460693, "perplexity": 33.89060400065904, "lr": 0.0026291804804649314, "grad_norm": 0.17048, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:57.372534+00:00", "epoch": 0, "step": 5184, "train_loss": 3.496459722518921, "perplexity": 32.99842135249126, "lr": 0.0026291804804649314, "grad_norm": 0.14551, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:57.677103+00:00", "epoch": 0, "step": 5185, "train_loss": 3.5488076210021973, "perplexity": 34.7718315573772, "lr": 0.0026291804804649314, "grad_norm": 0.172825, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:57.980582+00:00", "epoch": 0, "step": 5186, "train_loss": 3.5869200229644775, "perplexity": 36.122647491316805, "lr": 0.0026291804804649314, "grad_norm": 0.166758, "tokens_per_sec": 107972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:58.285670+00:00", "epoch": 0, "step": 5187, "train_loss": 3.6400768756866455, "perplexity": 38.094765174064676, "lr": 0.0026291804804649314, "grad_norm": 0.145614, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:58.591805+00:00", "epoch": 0, "step": 5188, "train_loss": 3.546600103378296, "perplexity": 34.69515678798852, "lr": 0.0026291804804649314, "grad_norm": 0.159305, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:58.897716+00:00", "epoch": 0, "step": 5189, "train_loss": 3.6898353099823, "perplexity": 40.038252513766935, "lr": 0.0026291804804649314, "grad_norm": 0.164677, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:59.202453+00:00", "epoch": 0, "step": 5190, "train_loss": 3.5628511905670166, "perplexity": 35.2635971829182, "lr": 0.0026291804804649314, "grad_norm": 0.152398, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:59.507263+00:00", "epoch": 0, "step": 5191, "train_loss": 3.5766282081604004, "perplexity": 35.75278642856306, "lr": 0.0026291804804649314, "grad_norm": 0.180722, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:24:59.812258+00:00", "epoch": 0, "step": 5192, "train_loss": 3.5412821769714355, "perplexity": 34.5111402244575, "lr": 0.0026291804804649314, "grad_norm": 0.189085, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:00.118130+00:00", "epoch": 0, "step": 5193, "train_loss": 3.5739588737487793, "perplexity": 35.65747754766811, "lr": 0.0026291804804649314, "grad_norm": 0.184561, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:00.422500+00:00", "epoch": 0, "step": 5194, "train_loss": 3.5576632022857666, "perplexity": 35.081123798145214, "lr": 0.0026291804804649314, "grad_norm": 0.167438, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:00.727109+00:00", "epoch": 0, "step": 5195, "train_loss": 3.644014596939087, "perplexity": 38.24506747055979, "lr": 0.0026291804804649314, "grad_norm": 0.162235, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:01.032141+00:00", "epoch": 0, "step": 5196, "train_loss": 3.5764472484588623, "perplexity": 35.746317200354255, "lr": 0.0026291804804649314, "grad_norm": 0.141917, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:01.336768+00:00", "epoch": 0, "step": 5197, "train_loss": 3.593289375305176, "perplexity": 36.35345964231404, "lr": 0.0026291804804649314, "grad_norm": 0.149101, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:01.642128+00:00", "epoch": 0, "step": 5198, "train_loss": 3.539658546447754, "perplexity": 34.45515234789418, "lr": 0.0026291804804649314, "grad_norm": 0.14495, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:01.947356+00:00", "epoch": 0, "step": 5199, "train_loss": 3.591613292694092, "perplexity": 36.29257927526946, "lr": 0.0026291804804649314, "grad_norm": 0.155218, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:02.254295+00:00", "epoch": 0, "step": 5200, "train_loss": 3.6022050380706787, "perplexity": 36.679023983216595, "lr": 0.0026291804804649314, "grad_norm": 0.161595, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:02.559883+00:00", "epoch": 0, "step": 5201, "train_loss": 3.498589515686035, "perplexity": 33.06877605870545, "lr": 0.0026291804804649314, "grad_norm": 0.152853, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:02.864099+00:00", "epoch": 0, "step": 5202, "train_loss": 3.5018789768218994, "perplexity": 33.17773361994432, "lr": 0.0026291804804649314, "grad_norm": 0.163522, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:03.168966+00:00", "epoch": 0, "step": 5203, "train_loss": 3.4678826332092285, "perplexity": 32.06876916111882, "lr": 0.0026291804804649314, "grad_norm": 0.1547, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:03.474459+00:00", "epoch": 0, "step": 5204, "train_loss": 3.631114959716797, "perplexity": 37.75488833829558, "lr": 0.0026291804804649314, "grad_norm": 0.141012, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:03.780956+00:00", "epoch": 0, "step": 5205, "train_loss": 3.6108322143554688, "perplexity": 36.996829299179794, "lr": 0.0026291804804649314, "grad_norm": 0.160302, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:04.085651+00:00", "epoch": 0, "step": 5206, "train_loss": 3.6973259449005127, "perplexity": 40.33929052142423, "lr": 0.0026291804804649314, "grad_norm": 0.143586, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:04.390457+00:00", "epoch": 0, "step": 5207, "train_loss": 3.568936586380005, "perplexity": 35.478844397562156, "lr": 0.0026291804804649314, "grad_norm": 0.144712, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:04.694611+00:00", "epoch": 0, "step": 5208, "train_loss": 3.68735933303833, "perplexity": 39.93924134892869, "lr": 0.0026291804804649314, "grad_norm": 0.153318, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:04.999829+00:00", "epoch": 0, "step": 5209, "train_loss": 3.592003583908081, "perplexity": 36.306746714627316, "lr": 0.0026291804804649314, "grad_norm": 0.139387, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:05.306841+00:00", "epoch": 0, "step": 5210, "train_loss": 3.566765308380127, "perplexity": 35.401893534350556, "lr": 0.0026291804804649314, "grad_norm": 0.153293, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:05.610492+00:00", "epoch": 0, "step": 5211, "train_loss": 3.548123598098755, "perplexity": 34.74805496099262, "lr": 0.0026291804804649314, "grad_norm": 0.169345, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:05.915827+00:00", "epoch": 0, "step": 5212, "train_loss": 3.6084790229797363, "perplexity": 36.909871034390584, "lr": 0.0026291804804649314, "grad_norm": 0.146484, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:06.221096+00:00", "epoch": 0, "step": 5213, "train_loss": 3.605377674102783, "perplexity": 36.795577970252, "lr": 0.0026291804804649314, "grad_norm": 0.170875, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:06.527623+00:00", "epoch": 0, "step": 5214, "train_loss": 3.6482317447662354, "perplexity": 38.406693133866135, "lr": 0.0026291804804649314, "grad_norm": 0.159307, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:06.833950+00:00", "epoch": 0, "step": 5215, "train_loss": 3.497736692428589, "perplexity": 33.040586259554026, "lr": 0.0026291804804649314, "grad_norm": 0.160614, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:07.139139+00:00", "epoch": 0, "step": 5216, "train_loss": 3.6965155601501465, "perplexity": 40.306613417846606, "lr": 0.0026291804804649314, "grad_norm": 0.146314, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:07.443663+00:00", "epoch": 0, "step": 5217, "train_loss": 3.5248308181762695, "perplexity": 33.94802974079798, "lr": 0.0026291804804649314, "grad_norm": 0.148637, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:07.749638+00:00", "epoch": 0, "step": 5218, "train_loss": 3.674182653427124, "perplexity": 39.416426806068934, "lr": 0.0026291804804649314, "grad_norm": 0.144895, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:08.055398+00:00", "epoch": 0, "step": 5219, "train_loss": 3.677968740463257, "perplexity": 39.565943691978234, "lr": 0.0026291804804649314, "grad_norm": 0.162021, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:08.361018+00:00", "epoch": 0, "step": 5220, "train_loss": 3.576608657836914, "perplexity": 35.752087456855435, "lr": 0.0026291804804649314, "grad_norm": 0.161668, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:08.667004+00:00", "epoch": 0, "step": 5221, "train_loss": 3.600297212600708, "perplexity": 36.60911351673703, "lr": 0.0026291804804649314, "grad_norm": 0.144079, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:08.972333+00:00", "epoch": 0, "step": 5222, "train_loss": 3.555100440979004, "perplexity": 34.991334355075644, "lr": 0.0026291804804649314, "grad_norm": 0.152646, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:09.277256+00:00", "epoch": 0, "step": 5223, "train_loss": 3.525214672088623, "perplexity": 33.96106332616701, "lr": 0.0026291804804649314, "grad_norm": 0.156116, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:09.582662+00:00", "epoch": 0, "step": 5224, "train_loss": 3.616764545440674, "perplexity": 37.21695903534678, "lr": 0.0026291804804649314, "grad_norm": 0.151214, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:09.888023+00:00", "epoch": 0, "step": 5225, "train_loss": 3.618347406387329, "perplexity": 37.2759149535581, "lr": 0.0026291804804649314, "grad_norm": 0.148099, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:10.194013+00:00", "epoch": 0, "step": 5226, "train_loss": 3.6804652214050293, "perplexity": 39.664842714746264, "lr": 0.0026291804804649314, "grad_norm": 0.174845, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:10.498509+00:00", "epoch": 0, "step": 5227, "train_loss": 3.567781686782837, "perplexity": 35.43789354607203, "lr": 0.0026291804804649314, "grad_norm": 0.176235, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:10.803826+00:00", "epoch": 0, "step": 5228, "train_loss": 3.576444625854492, "perplexity": 35.74622345202948, "lr": 0.0026291804804649314, "grad_norm": 0.19585, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:11.109465+00:00", "epoch": 0, "step": 5229, "train_loss": 3.5805907249450684, "perplexity": 35.894738502884664, "lr": 0.0026291804804649314, "grad_norm": 0.192268, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:11.415559+00:00", "epoch": 0, "step": 5230, "train_loss": 3.6293323040008545, "perplexity": 37.687644325064504, "lr": 0.0026291804804649314, "grad_norm": 0.150126, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:11.721547+00:00", "epoch": 0, "step": 5231, "train_loss": 3.5103909969329834, "perplexity": 33.4613485111575, "lr": 0.0026291804804649314, "grad_norm": 0.176418, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:12.026754+00:00", "epoch": 0, "step": 5232, "train_loss": 3.6036462783813477, "perplexity": 36.731925383795584, "lr": 0.0026291804804649314, "grad_norm": 0.169364, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:12.332052+00:00", "epoch": 0, "step": 5233, "train_loss": 3.5843145847320557, "perplexity": 36.0286546639107, "lr": 0.0026291804804649314, "grad_norm": 0.157088, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:12.637670+00:00", "epoch": 0, "step": 5234, "train_loss": 3.5470945835113525, "perplexity": 34.71231709609953, "lr": 0.0026291804804649314, "grad_norm": 0.162143, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:12.944336+00:00", "epoch": 0, "step": 5235, "train_loss": 3.6229381561279297, "perplexity": 37.44743274686659, "lr": 0.0026291804804649314, "grad_norm": 0.15329, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:13.251016+00:00", "epoch": 0, "step": 5236, "train_loss": 3.4379160404205322, "perplexity": 31.12203348178808, "lr": 0.0026291804804649314, "grad_norm": 0.169623, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:13.557743+00:00", "epoch": 0, "step": 5237, "train_loss": 3.5532712936401367, "perplexity": 34.92738854993535, "lr": 0.0026291804804649314, "grad_norm": 0.162449, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:13.863893+00:00", "epoch": 0, "step": 5238, "train_loss": 3.530569314956665, "perplexity": 34.14340043106981, "lr": 0.0026291804804649314, "grad_norm": 0.16629, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:14.170067+00:00", "epoch": 0, "step": 5239, "train_loss": 3.579406261444092, "perplexity": 35.852247664642135, "lr": 0.0026291804804649314, "grad_norm": 0.130414, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:14.476652+00:00", "epoch": 0, "step": 5240, "train_loss": 3.589851140975952, "perplexity": 36.228682558732466, "lr": 0.0026291804804649314, "grad_norm": 0.165481, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:14.782757+00:00", "epoch": 0, "step": 5241, "train_loss": 3.557401180267334, "perplexity": 35.07193297543028, "lr": 0.0026291804804649314, "grad_norm": 0.160674, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:15.087926+00:00", "epoch": 0, "step": 5242, "train_loss": 3.5707764625549316, "perplexity": 35.544181165433244, "lr": 0.0026291804804649314, "grad_norm": 0.158206, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:15.393381+00:00", "epoch": 0, "step": 5243, "train_loss": 3.618831157684326, "perplexity": 37.29395158803456, "lr": 0.0026291804804649314, "grad_norm": 0.166818, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:15.699290+00:00", "epoch": 0, "step": 5244, "train_loss": 3.531975507736206, "perplexity": 34.19144640726027, "lr": 0.0026291804804649314, "grad_norm": 0.164915, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:16.006182+00:00", "epoch": 0, "step": 5245, "train_loss": 3.642860174179077, "perplexity": 38.20094196885587, "lr": 0.0026291804804649314, "grad_norm": 0.171631, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:16.313900+00:00", "epoch": 0, "step": 5246, "train_loss": 3.5865020751953125, "perplexity": 36.107553265900115, "lr": 0.0026291804804649314, "grad_norm": 0.162523, "tokens_per_sec": 106488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:16.620185+00:00", "epoch": 0, "step": 5247, "train_loss": 3.429983615875244, "perplexity": 30.876136867054424, "lr": 0.0026291804804649314, "grad_norm": 0.151289, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:16.926228+00:00", "epoch": 0, "step": 5248, "train_loss": 3.625328540802002, "perplexity": 37.5370535876562, "lr": 0.0026291804804649314, "grad_norm": 0.15985, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:17.231698+00:00", "epoch": 0, "step": 5249, "train_loss": 3.6587936878204346, "perplexity": 38.81449222404711, "lr": 0.0026291804804649314, "grad_norm": 0.195438, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:17.538256+00:00", "epoch": 0, "step": 5250, "train_loss": 3.5442886352539062, "perplexity": 34.61505265374445, "lr": 0.0026291804804649314, "grad_norm": 0.184063, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:17.842766+00:00", "epoch": 0, "step": 5251, "train_loss": 3.6052651405334473, "perplexity": 36.79143746550448, "lr": 0.0026291804804649314, "grad_norm": 0.145812, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:18.147535+00:00", "epoch": 0, "step": 5252, "train_loss": 3.681737184524536, "perplexity": 39.71532703210885, "lr": 0.0026291804804649314, "grad_norm": 0.154495, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:18.452431+00:00", "epoch": 0, "step": 5253, "train_loss": 3.694761037826538, "perplexity": 40.23595656745214, "lr": 0.0026291804804649314, "grad_norm": 0.141974, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:18.758679+00:00", "epoch": 0, "step": 5254, "train_loss": 3.45215106010437, "perplexity": 31.568224475293626, "lr": 0.0026291804804649314, "grad_norm": 0.15297, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:19.064403+00:00", "epoch": 0, "step": 5255, "train_loss": 3.6298296451568604, "perplexity": 37.70639260341904, "lr": 0.0026291804804649314, "grad_norm": 0.15244, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:19.369317+00:00", "epoch": 0, "step": 5256, "train_loss": 3.4946680068969727, "perplexity": 32.93935050034133, "lr": 0.0026291804804649314, "grad_norm": 0.151105, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:19.673788+00:00", "epoch": 0, "step": 5257, "train_loss": 3.607421398162842, "perplexity": 36.870854874662484, "lr": 0.0026291804804649314, "grad_norm": 0.161983, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:19.978001+00:00", "epoch": 0, "step": 5258, "train_loss": 3.580972909927368, "perplexity": 35.908459554707, "lr": 0.0026291804804649314, "grad_norm": 0.150706, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:20.283806+00:00", "epoch": 0, "step": 5259, "train_loss": 3.578028440475464, "perplexity": 35.80288370119865, "lr": 0.0026291804804649314, "grad_norm": 0.154469, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:20.590283+00:00", "epoch": 0, "step": 5260, "train_loss": 3.588548183441162, "perplexity": 36.18150886315199, "lr": 0.0026291804804649314, "grad_norm": 0.173919, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:20.895566+00:00", "epoch": 0, "step": 5261, "train_loss": 3.5828707218170166, "perplexity": 35.97667176268528, "lr": 0.0026291804804649314, "grad_norm": 0.171287, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:21.201697+00:00", "epoch": 0, "step": 5262, "train_loss": 3.4959404468536377, "perplexity": 32.98129052348641, "lr": 0.0026291804804649314, "grad_norm": 0.173422, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:21.507787+00:00", "epoch": 0, "step": 5263, "train_loss": 3.585437536239624, "perplexity": 36.06913582091538, "lr": 0.0026291804804649314, "grad_norm": 0.159458, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:21.814142+00:00", "epoch": 0, "step": 5264, "train_loss": 3.5440423488616943, "perplexity": 34.606528487046354, "lr": 0.0026291804804649314, "grad_norm": 0.16498, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:22.120573+00:00", "epoch": 0, "step": 5265, "train_loss": 3.638948917388916, "perplexity": 38.05182009227043, "lr": 0.0026291804804649314, "grad_norm": 0.161547, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:22.427064+00:00", "epoch": 0, "step": 5266, "train_loss": 3.553619623184204, "perplexity": 34.93955691044111, "lr": 0.0026291804804649314, "grad_norm": 0.16759, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:22.732256+00:00", "epoch": 0, "step": 5267, "train_loss": 3.6255385875701904, "perplexity": 37.54493895256832, "lr": 0.0026291804804649314, "grad_norm": 0.153433, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:23.037732+00:00", "epoch": 0, "step": 5268, "train_loss": 3.704854726791382, "perplexity": 40.6441423829993, "lr": 0.0026291804804649314, "grad_norm": 0.152276, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:23.342528+00:00", "epoch": 0, "step": 5269, "train_loss": 3.6922645568847656, "perplexity": 40.135633548065044, "lr": 0.0026291804804649314, "grad_norm": 0.182451, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:23.648546+00:00", "epoch": 0, "step": 5270, "train_loss": 3.633881092071533, "perplexity": 37.85946793029168, "lr": 0.0026291804804649314, "grad_norm": 0.143728, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:23.955807+00:00", "epoch": 0, "step": 5271, "train_loss": 3.6162803173065186, "perplexity": 37.198941899268306, "lr": 0.0026291804804649314, "grad_norm": 0.158895, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:24.261626+00:00", "epoch": 0, "step": 5272, "train_loss": 3.5603137016296387, "perplexity": 35.17422962771775, "lr": 0.0026291804804649314, "grad_norm": 0.168216, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:24.567307+00:00", "epoch": 0, "step": 5273, "train_loss": 3.531076669692993, "perplexity": 34.16072764214248, "lr": 0.0026291804804649314, "grad_norm": 0.143943, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:24.872928+00:00", "epoch": 0, "step": 5274, "train_loss": 3.561615228652954, "perplexity": 35.22003964312228, "lr": 0.0026291804804649314, "grad_norm": 0.148363, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:25.178224+00:00", "epoch": 0, "step": 5275, "train_loss": 3.700218677520752, "perplexity": 40.45615024346863, "lr": 0.0026291804804649314, "grad_norm": 0.144071, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:25.483693+00:00", "epoch": 0, "step": 5276, "train_loss": 3.66396427154541, "perplexity": 39.01570555272997, "lr": 0.0026291804804649314, "grad_norm": 0.150662, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:25.789709+00:00", "epoch": 0, "step": 5277, "train_loss": 3.551393985748291, "perplexity": 34.86188059630721, "lr": 0.0026291804804649314, "grad_norm": 0.171485, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:26.095419+00:00", "epoch": 0, "step": 5278, "train_loss": 3.5056540966033936, "perplexity": 33.303220252942054, "lr": 0.0026291804804649314, "grad_norm": 0.155312, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:26.401379+00:00", "epoch": 0, "step": 5279, "train_loss": 3.672450542449951, "perplexity": 39.348212275140945, "lr": 0.0026291804804649314, "grad_norm": 0.157647, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:26.707626+00:00", "epoch": 0, "step": 5280, "train_loss": 3.559159517288208, "perplexity": 35.133655502173944, "lr": 0.0026291804804649314, "grad_norm": 0.146391, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:27.014142+00:00", "epoch": 0, "step": 5281, "train_loss": 3.506455183029175, "perplexity": 33.329909699470846, "lr": 0.0026291804804649314, "grad_norm": 0.158578, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:27.319683+00:00", "epoch": 0, "step": 5282, "train_loss": 3.594341278076172, "perplexity": 36.39172006684245, "lr": 0.0026291804804649314, "grad_norm": 0.166166, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:27.625046+00:00", "epoch": 0, "step": 5283, "train_loss": 3.540752410888672, "perplexity": 34.492862234846044, "lr": 0.0026291804804649314, "grad_norm": 0.16897, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:27.931324+00:00", "epoch": 0, "step": 5284, "train_loss": 3.4888110160827637, "perplexity": 32.74698890698031, "lr": 0.0026291804804649314, "grad_norm": 0.149502, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:28.236882+00:00", "epoch": 0, "step": 5285, "train_loss": 3.6083831787109375, "perplexity": 36.90633360431374, "lr": 0.0026291804804649314, "grad_norm": 0.165522, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:28.541787+00:00", "epoch": 0, "step": 5286, "train_loss": 3.656832695007324, "perplexity": 38.73845186541733, "lr": 0.0026291804804649314, "grad_norm": 0.146953, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:28.849695+00:00", "epoch": 0, "step": 5287, "train_loss": 3.592376470565796, "perplexity": 36.32028754050199, "lr": 0.0026291804804649314, "grad_norm": 0.157691, "tokens_per_sec": 106421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:29.155223+00:00", "epoch": 0, "step": 5288, "train_loss": 3.5710597038269043, "perplexity": 35.55425017042934, "lr": 0.0026291804804649314, "grad_norm": 0.168331, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:29.459351+00:00", "epoch": 0, "step": 5289, "train_loss": 3.5799880027770996, "perplexity": 35.87311046677868, "lr": 0.0026291804804649314, "grad_norm": 0.150073, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:29.765587+00:00", "epoch": 0, "step": 5290, "train_loss": 3.4978432655334473, "perplexity": 33.044107685059345, "lr": 0.0026291804804649314, "grad_norm": 0.1669, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:30.071509+00:00", "epoch": 0, "step": 5291, "train_loss": 3.455787181854248, "perplexity": 31.683219323849045, "lr": 0.0026291804804649314, "grad_norm": 0.179466, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:30.378529+00:00", "epoch": 0, "step": 5292, "train_loss": 3.5570452213287354, "perplexity": 35.05945102905573, "lr": 0.0026291804804649314, "grad_norm": 0.170746, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:30.684653+00:00", "epoch": 0, "step": 5293, "train_loss": 3.5597174167633057, "perplexity": 35.15326201886101, "lr": 0.0026291804804649314, "grad_norm": 0.160807, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:30.990823+00:00", "epoch": 0, "step": 5294, "train_loss": 3.5771753787994385, "perplexity": 35.77235465665229, "lr": 0.0026291804804649314, "grad_norm": 0.169303, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:31.296830+00:00", "epoch": 0, "step": 5295, "train_loss": 3.7553677558898926, "perplexity": 42.749938461723815, "lr": 0.0026291804804649314, "grad_norm": 0.1567, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:31.603250+00:00", "epoch": 0, "step": 5296, "train_loss": 3.5255818367004395, "perplexity": 33.97353491622319, "lr": 0.0026291804804649314, "grad_norm": 0.17294, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:31.908306+00:00", "epoch": 0, "step": 5297, "train_loss": 3.577662944793701, "perplexity": 35.78980029292147, "lr": 0.0026291804804649314, "grad_norm": 0.16047, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:32.214454+00:00", "epoch": 0, "step": 5298, "train_loss": 3.5080068111419678, "perplexity": 33.38166546682549, "lr": 0.0026291804804649314, "grad_norm": 0.145394, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:32.520720+00:00", "epoch": 0, "step": 5299, "train_loss": 3.6049938201904297, "perplexity": 36.781456554144604, "lr": 0.0026291804804649314, "grad_norm": 0.170922, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:32.826373+00:00", "epoch": 0, "step": 5300, "train_loss": 3.579859972000122, "perplexity": 35.86851789857433, "lr": 0.0026291804804649314, "grad_norm": 0.151729, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:33.132114+00:00", "epoch": 0, "step": 5301, "train_loss": 3.4938807487487793, "perplexity": 32.91342893308505, "lr": 0.0026291804804649314, "grad_norm": 0.145957, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:33.437163+00:00", "epoch": 0, "step": 5302, "train_loss": 3.5953824520111084, "perplexity": 36.42962990916875, "lr": 0.0026291804804649314, "grad_norm": 0.163304, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:33.743912+00:00", "epoch": 0, "step": 5303, "train_loss": 3.6104209423065186, "perplexity": 36.981616665869126, "lr": 0.0026291804804649314, "grad_norm": 0.143654, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:34.050869+00:00", "epoch": 0, "step": 5304, "train_loss": 3.6426842212677, "perplexity": 38.19422099320407, "lr": 0.0026291804804649314, "grad_norm": 0.159519, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:34.357787+00:00", "epoch": 0, "step": 5305, "train_loss": 3.575077772140503, "perplexity": 35.69739697067351, "lr": 0.0026291804804649314, "grad_norm": 0.155883, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:34.664893+00:00", "epoch": 0, "step": 5306, "train_loss": 3.5292844772338867, "perplexity": 34.09955987224495, "lr": 0.0026291804804649314, "grad_norm": 0.137561, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:34.971194+00:00", "epoch": 0, "step": 5307, "train_loss": 3.652970790863037, "perplexity": 38.58913618461649, "lr": 0.0026291804804649314, "grad_norm": 0.148144, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:35.276410+00:00", "epoch": 0, "step": 5308, "train_loss": 3.607118844985962, "perplexity": 36.85970116776569, "lr": 0.0026291804804649314, "grad_norm": 0.145621, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:35.583117+00:00", "epoch": 0, "step": 5309, "train_loss": 3.6095430850982666, "perplexity": 36.949166332569625, "lr": 0.0026291804804649314, "grad_norm": 0.153638, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:35.889333+00:00", "epoch": 0, "step": 5310, "train_loss": 3.583616256713867, "perplexity": 36.00350362775997, "lr": 0.0026291804804649314, "grad_norm": 0.156618, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:36.196050+00:00", "epoch": 0, "step": 5311, "train_loss": 3.54445743560791, "perplexity": 34.62089618006705, "lr": 0.0026291804804649314, "grad_norm": 0.147218, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:36.502519+00:00", "epoch": 0, "step": 5312, "train_loss": 3.6158411502838135, "perplexity": 37.182608937418244, "lr": 0.0026291804804649314, "grad_norm": 0.183609, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:36.807117+00:00", "epoch": 0, "step": 5313, "train_loss": 3.6245808601379395, "perplexity": 37.508998347989916, "lr": 0.0026291804804649314, "grad_norm": 0.151196, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:37.113483+00:00", "epoch": 0, "step": 5314, "train_loss": 3.597791910171509, "perplexity": 36.51751140911611, "lr": 0.0026291804804649314, "grad_norm": 0.15652, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:37.420236+00:00", "epoch": 0, "step": 5315, "train_loss": 3.577441930770874, "perplexity": 35.7818911192339, "lr": 0.0026291804804649314, "grad_norm": 0.178367, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:37.725628+00:00", "epoch": 0, "step": 5316, "train_loss": 3.659593343734741, "perplexity": 38.845542875578836, "lr": 0.0026291804804649314, "grad_norm": 0.15691, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:38.032254+00:00", "epoch": 0, "step": 5317, "train_loss": 3.6182687282562256, "perplexity": 37.27298226960494, "lr": 0.0026291804804649314, "grad_norm": 0.146284, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:38.338068+00:00", "epoch": 0, "step": 5318, "train_loss": 3.530442953109741, "perplexity": 34.139086280508835, "lr": 0.0026291804804649314, "grad_norm": 0.155149, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:38.644423+00:00", "epoch": 0, "step": 5319, "train_loss": 3.576024293899536, "perplexity": 35.73120132940385, "lr": 0.0026291804804649314, "grad_norm": 0.169506, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:38.950921+00:00", "epoch": 0, "step": 5320, "train_loss": 3.4986276626586914, "perplexity": 33.070037556462545, "lr": 0.0026291804804649314, "grad_norm": 0.185062, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:39.257905+00:00", "epoch": 0, "step": 5321, "train_loss": 3.5849897861480713, "perplexity": 36.05298947708255, "lr": 0.0026291804804649314, "grad_norm": 0.194735, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:39.563321+00:00", "epoch": 0, "step": 5322, "train_loss": 3.586932897567749, "perplexity": 36.12311255906615, "lr": 0.0026291804804649314, "grad_norm": 0.176321, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:39.867966+00:00", "epoch": 0, "step": 5323, "train_loss": 3.647397756576538, "perplexity": 38.374675758299475, "lr": 0.0026291804804649314, "grad_norm": 0.152926, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:40.173323+00:00", "epoch": 0, "step": 5324, "train_loss": 3.512953281402588, "perplexity": 33.5471959405794, "lr": 0.0026291804804649314, "grad_norm": 0.158409, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:40.480163+00:00", "epoch": 0, "step": 5325, "train_loss": 3.5357134342193604, "perplexity": 34.31949068123749, "lr": 0.0026291804804649314, "grad_norm": 0.17955, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:40.786229+00:00", "epoch": 0, "step": 5326, "train_loss": 3.634575605392456, "perplexity": 37.88577096794006, "lr": 0.0026291804804649314, "grad_norm": 0.154831, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:41.092695+00:00", "epoch": 0, "step": 5327, "train_loss": 3.5306272506713867, "perplexity": 34.145378610679884, "lr": 0.0026291804804649314, "grad_norm": 0.191456, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:41.398065+00:00", "epoch": 0, "step": 5328, "train_loss": 3.6087448596954346, "perplexity": 36.919684337593566, "lr": 0.0026291804804649314, "grad_norm": 0.150082, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:41.702660+00:00", "epoch": 0, "step": 5329, "train_loss": 3.5819084644317627, "perplexity": 35.94206959535027, "lr": 0.0026291804804649314, "grad_norm": 0.148129, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:42.007715+00:00", "epoch": 0, "step": 5330, "train_loss": 3.611476421356201, "perplexity": 37.02067059415928, "lr": 0.0026291804804649314, "grad_norm": 0.156836, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:42.314271+00:00", "epoch": 0, "step": 5331, "train_loss": 3.538151264190674, "perplexity": 34.403257827764854, "lr": 0.0026291804804649314, "grad_norm": 0.146331, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:42.620584+00:00", "epoch": 0, "step": 5332, "train_loss": 3.684499979019165, "perplexity": 39.825204032880734, "lr": 0.0026291804804649314, "grad_norm": 0.160483, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:42.925340+00:00", "epoch": 0, "step": 5333, "train_loss": 3.667996644973755, "perplexity": 39.17334907225118, "lr": 0.0026291804804649314, "grad_norm": 0.138498, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:43.231536+00:00", "epoch": 0, "step": 5334, "train_loss": 3.528724193572998, "perplexity": 34.080459797234674, "lr": 0.0026291804804649314, "grad_norm": 0.148347, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:43.536629+00:00", "epoch": 0, "step": 5335, "train_loss": 3.6473734378814697, "perplexity": 38.37374254760864, "lr": 0.0026291804804649314, "grad_norm": 0.155952, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:43.842975+00:00", "epoch": 0, "step": 5336, "train_loss": 3.5473086833953857, "perplexity": 34.71974979480591, "lr": 0.0026291804804649314, "grad_norm": 0.156236, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:44.150203+00:00", "epoch": 0, "step": 5337, "train_loss": 3.603255271911621, "perplexity": 36.71756577085874, "lr": 0.0026291804804649314, "grad_norm": 0.163258, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:44.456486+00:00", "epoch": 0, "step": 5338, "train_loss": 3.549088954925537, "perplexity": 34.781615429373964, "lr": 0.0026291804804649314, "grad_norm": 0.166094, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:44.761639+00:00", "epoch": 0, "step": 5339, "train_loss": 3.586426258087158, "perplexity": 36.104815799403674, "lr": 0.0026291804804649314, "grad_norm": 0.169573, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:45.068209+00:00", "epoch": 0, "step": 5340, "train_loss": 3.59800386428833, "perplexity": 36.52525226631964, "lr": 0.0026291804804649314, "grad_norm": 0.153381, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:45.373866+00:00", "epoch": 0, "step": 5341, "train_loss": 3.544203519821167, "perplexity": 34.612106503941725, "lr": 0.0026291804804649314, "grad_norm": 0.152237, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:45.680488+00:00", "epoch": 0, "step": 5342, "train_loss": 3.5011191368103027, "perplexity": 33.152533425716186, "lr": 0.0026291804804649314, "grad_norm": 0.157567, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:45.987123+00:00", "epoch": 0, "step": 5343, "train_loss": 3.5031399726867676, "perplexity": 33.21959699406831, "lr": 0.0026291804804649314, "grad_norm": 0.158213, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:46.294680+00:00", "epoch": 0, "step": 5344, "train_loss": 3.5411064624786377, "perplexity": 34.50507664970167, "lr": 0.0026291804804649314, "grad_norm": 0.156898, "tokens_per_sec": 106543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:46.601168+00:00", "epoch": 0, "step": 5345, "train_loss": 3.643331527709961, "perplexity": 38.21895236203373, "lr": 0.0026291804804649314, "grad_norm": 0.168872, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:46.907036+00:00", "epoch": 0, "step": 5346, "train_loss": 3.468538522720337, "perplexity": 32.089809629802836, "lr": 0.0026291804804649314, "grad_norm": 0.158433, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:47.212713+00:00", "epoch": 0, "step": 5347, "train_loss": 3.5059263706207275, "perplexity": 33.31228908905867, "lr": 0.0026291804804649314, "grad_norm": 0.163766, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:47.519445+00:00", "epoch": 0, "step": 5348, "train_loss": 3.6530933380126953, "perplexity": 38.59386546303757, "lr": 0.0026291804804649314, "grad_norm": 0.165532, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:47.826650+00:00", "epoch": 0, "step": 5349, "train_loss": 3.591785430908203, "perplexity": 36.29882715278521, "lr": 0.0026291804804649314, "grad_norm": 0.174386, "tokens_per_sec": 106668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:48.132651+00:00", "epoch": 0, "step": 5350, "train_loss": 3.5932583808898926, "perplexity": 36.35233290555027, "lr": 0.0026291804804649314, "grad_norm": 0.148486, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:48.437995+00:00", "epoch": 0, "step": 5351, "train_loss": 3.5697569847106934, "perplexity": 35.50796312512521, "lr": 0.0026291804804649314, "grad_norm": 0.164545, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:48.742866+00:00", "epoch": 0, "step": 5352, "train_loss": 3.5338821411132812, "perplexity": 34.256699146949046, "lr": 0.0026291804804649314, "grad_norm": 0.161276, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:49.049390+00:00", "epoch": 0, "step": 5353, "train_loss": 3.6587443351745605, "perplexity": 38.812576673426726, "lr": 0.0026291804804649314, "grad_norm": 0.152256, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:49.356392+00:00", "epoch": 0, "step": 5354, "train_loss": 3.6269822120666504, "perplexity": 37.59917888778803, "lr": 0.0026291804804649314, "grad_norm": 0.155305, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:49.663472+00:00", "epoch": 0, "step": 5355, "train_loss": 3.625622510910034, "perplexity": 37.54808998146004, "lr": 0.0026291804804649314, "grad_norm": 0.17669, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:49.969211+00:00", "epoch": 0, "step": 5356, "train_loss": 3.6369922161102295, "perplexity": 37.977436843870635, "lr": 0.0026291804804649314, "grad_norm": 0.182326, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:50.274661+00:00", "epoch": 0, "step": 5357, "train_loss": 3.5560567378997803, "perplexity": 35.024812465328814, "lr": 0.0026291804804649314, "grad_norm": 0.159185, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:50.580603+00:00", "epoch": 0, "step": 5358, "train_loss": 3.6923952102661133, "perplexity": 40.140877746879326, "lr": 0.0026291804804649314, "grad_norm": 0.168616, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:50.887328+00:00", "epoch": 0, "step": 5359, "train_loss": 3.599332571029663, "perplexity": 36.573815871484804, "lr": 0.0026291804804649314, "grad_norm": 0.143725, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:51.194674+00:00", "epoch": 0, "step": 5360, "train_loss": 3.5920960903167725, "perplexity": 36.310105476728275, "lr": 0.0026291804804649314, "grad_norm": 0.149873, "tokens_per_sec": 106686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:51.501144+00:00", "epoch": 0, "step": 5361, "train_loss": 3.5648000240325928, "perplexity": 35.33238706946965, "lr": 0.0026291804804649314, "grad_norm": 0.152873, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:51.806443+00:00", "epoch": 0, "step": 5362, "train_loss": 3.5304994583129883, "perplexity": 34.141015371019094, "lr": 0.0026291804804649314, "grad_norm": 0.173204, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:52.111736+00:00", "epoch": 0, "step": 5363, "train_loss": 3.5130980014801025, "perplexity": 33.552051244697836, "lr": 0.0026291804804649314, "grad_norm": 0.15065, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:52.417837+00:00", "epoch": 0, "step": 5364, "train_loss": 3.524496555328369, "perplexity": 33.93668407201963, "lr": 0.0026291804804649314, "grad_norm": 0.140419, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:52.725219+00:00", "epoch": 0, "step": 5365, "train_loss": 3.5785868167877197, "perplexity": 35.82288076579188, "lr": 0.0026291804804649314, "grad_norm": 0.145496, "tokens_per_sec": 106604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:53.031450+00:00", "epoch": 0, "step": 5366, "train_loss": 3.57175350189209, "perplexity": 35.57892619950749, "lr": 0.0026291804804649314, "grad_norm": 0.138119, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:53.336387+00:00", "epoch": 0, "step": 5367, "train_loss": 3.5354719161987305, "perplexity": 34.311202906642876, "lr": 0.0026291804804649314, "grad_norm": 0.149835, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:53.642049+00:00", "epoch": 0, "step": 5368, "train_loss": 3.5450637340545654, "perplexity": 34.64189314021179, "lr": 0.0026291804804649314, "grad_norm": 0.166291, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:53.947829+00:00", "epoch": 0, "step": 5369, "train_loss": 3.4657628536224365, "perplexity": 32.00086243794837, "lr": 0.0026291804804649314, "grad_norm": 0.149307, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:54.254081+00:00", "epoch": 0, "step": 5370, "train_loss": 3.6247658729553223, "perplexity": 37.515938635452756, "lr": 0.0026291804804649314, "grad_norm": 0.14852, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:54.561140+00:00", "epoch": 0, "step": 5371, "train_loss": 3.563596248626709, "perplexity": 35.28988040025869, "lr": 0.0026291804804649314, "grad_norm": 0.165449, "tokens_per_sec": 106652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:54.865905+00:00", "epoch": 0, "step": 5372, "train_loss": 3.648956298828125, "perplexity": 38.43453094316107, "lr": 0.0026291804804649314, "grad_norm": 0.150897, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:55.171331+00:00", "epoch": 0, "step": 5373, "train_loss": 3.5411784648895264, "perplexity": 34.50756118785364, "lr": 0.0026291804804649314, "grad_norm": 0.15772, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:55.477013+00:00", "epoch": 0, "step": 5374, "train_loss": 3.714186668395996, "perplexity": 41.02520641451827, "lr": 0.0026291804804649314, "grad_norm": 0.190644, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:55.783445+00:00", "epoch": 0, "step": 5375, "train_loss": 3.52235746383667, "perplexity": 33.864167986786214, "lr": 0.0026291804804649314, "grad_norm": 0.167998, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:56.089553+00:00", "epoch": 0, "step": 5376, "train_loss": 3.5128493309020996, "perplexity": 33.54370887401569, "lr": 0.0026291804804649314, "grad_norm": 0.163906, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:56.396388+00:00", "epoch": 0, "step": 5377, "train_loss": 3.5577311515808105, "perplexity": 35.083507616765125, "lr": 0.0026291804804649314, "grad_norm": 0.190406, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:56.702068+00:00", "epoch": 0, "step": 5378, "train_loss": 3.557621717453003, "perplexity": 35.07966849377802, "lr": 0.0026291804804649314, "grad_norm": 0.179482, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:57.006898+00:00", "epoch": 0, "step": 5379, "train_loss": 3.5680434703826904, "perplexity": 35.447171819812716, "lr": 0.0026291804804649314, "grad_norm": 0.137302, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:57.311821+00:00", "epoch": 0, "step": 5380, "train_loss": 3.6387665271759033, "perplexity": 38.04488044557934, "lr": 0.0026291804804649314, "grad_norm": 0.167561, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:57.618244+00:00", "epoch": 0, "step": 5381, "train_loss": 3.655179023742676, "perplexity": 38.674444139185766, "lr": 0.0026291804804649314, "grad_norm": 0.140884, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:57.924084+00:00", "epoch": 0, "step": 5382, "train_loss": 3.5974526405334473, "perplexity": 36.50512422765354, "lr": 0.0026291804804649314, "grad_norm": 0.173355, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:58.228681+00:00", "epoch": 0, "step": 5383, "train_loss": 3.602090358734131, "perplexity": 36.674817898261104, "lr": 0.0026291804804649314, "grad_norm": 0.157435, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:58.533254+00:00", "epoch": 0, "step": 5384, "train_loss": 3.467876672744751, "perplexity": 32.06857801692906, "lr": 0.0026291804804649314, "grad_norm": 0.158308, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:58.837914+00:00", "epoch": 0, "step": 5385, "train_loss": 3.6041035652160645, "perplexity": 36.74872625080227, "lr": 0.0026291804804649314, "grad_norm": 0.16682, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:59.143849+00:00", "epoch": 0, "step": 5386, "train_loss": 3.7286343574523926, "perplexity": 41.62222824823071, "lr": 0.0026291804804649314, "grad_norm": 0.186287, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:59.450918+00:00", "epoch": 0, "step": 5387, "train_loss": 3.638723373413086, "perplexity": 38.04323870125615, "lr": 0.0026291804804649314, "grad_norm": 0.172358, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:25:59.757781+00:00", "epoch": 0, "step": 5388, "train_loss": 3.6461687088012695, "perplexity": 38.327540420156716, "lr": 0.0026291804804649314, "grad_norm": 0.175585, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:00.063261+00:00", "epoch": 0, "step": 5389, "train_loss": 3.555765390396118, "perplexity": 35.01460956001848, "lr": 0.0026291804804649314, "grad_norm": 0.182607, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:00.369122+00:00", "epoch": 0, "step": 5390, "train_loss": 3.5290794372558594, "perplexity": 34.09256881598551, "lr": 0.0026291804804649314, "grad_norm": 0.155055, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:00.674942+00:00", "epoch": 0, "step": 5391, "train_loss": 3.655817747116089, "perplexity": 38.699154301250175, "lr": 0.0026291804804649314, "grad_norm": 0.166506, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:00.981306+00:00", "epoch": 0, "step": 5392, "train_loss": 3.5963778495788574, "perplexity": 36.46590992769987, "lr": 0.0026291804804649314, "grad_norm": 0.182971, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:01.288540+00:00", "epoch": 0, "step": 5393, "train_loss": 3.7003750801086426, "perplexity": 40.462478184903105, "lr": 0.0026291804804649314, "grad_norm": 0.164196, "tokens_per_sec": 106655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:01.595563+00:00", "epoch": 0, "step": 5394, "train_loss": 3.6252150535583496, "perplexity": 37.53279385262708, "lr": 0.0026291804804649314, "grad_norm": 0.159219, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:01.901876+00:00", "epoch": 0, "step": 5395, "train_loss": 3.571096897125244, "perplexity": 35.55557257485532, "lr": 0.0026291804804649314, "grad_norm": 0.163319, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:02.208650+00:00", "epoch": 0, "step": 5396, "train_loss": 3.5445210933685303, "perplexity": 34.62310013893733, "lr": 0.0026291804804649314, "grad_norm": 0.15831, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:02.513696+00:00", "epoch": 0, "step": 5397, "train_loss": 3.728487968444824, "perplexity": 41.616135657499704, "lr": 0.0026291804804649314, "grad_norm": 0.151404, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:02.819260+00:00", "epoch": 0, "step": 5398, "train_loss": 3.583698034286499, "perplexity": 36.00644802728417, "lr": 0.0026291804804649314, "grad_norm": 0.139658, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:03.125408+00:00", "epoch": 0, "step": 5399, "train_loss": 3.5931317806243896, "perplexity": 36.347730981861304, "lr": 0.0026291804804649314, "grad_norm": 0.142392, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:03.430868+00:00", "epoch": 0, "step": 5400, "train_loss": 3.642487049102783, "perplexity": 38.18669089835047, "lr": 0.0026291804804649314, "grad_norm": 0.136147, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:03.735946+00:00", "epoch": 0, "step": 5401, "train_loss": 3.6399552822113037, "perplexity": 38.09013338077846, "lr": 0.0026291804804649314, "grad_norm": 0.15037, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:04.041283+00:00", "epoch": 0, "step": 5402, "train_loss": 3.516207695007324, "perplexity": 33.656550236988004, "lr": 0.0026291804804649314, "grad_norm": 0.146897, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:04.348295+00:00", "epoch": 0, "step": 5403, "train_loss": 3.6107687950134277, "perplexity": 36.99448305900733, "lr": 0.0026291804804649314, "grad_norm": 0.144682, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:04.654432+00:00", "epoch": 0, "step": 5404, "train_loss": 3.6391701698303223, "perplexity": 38.060240081803066, "lr": 0.0026291804804649314, "grad_norm": 0.13786, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:04.961379+00:00", "epoch": 0, "step": 5405, "train_loss": 3.5236308574676514, "perplexity": 33.90731787018168, "lr": 0.0026291804804649314, "grad_norm": 0.146444, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:05.266966+00:00", "epoch": 0, "step": 5406, "train_loss": 3.6066417694091797, "perplexity": 36.84212049855846, "lr": 0.0026291804804649314, "grad_norm": 0.137798, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:05.572379+00:00", "epoch": 0, "step": 5407, "train_loss": 3.6766207218170166, "perplexity": 39.5126439946894, "lr": 0.0026291804804649314, "grad_norm": 0.151788, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:05.878279+00:00", "epoch": 0, "step": 5408, "train_loss": 3.5572798252105713, "perplexity": 35.06767707725587, "lr": 0.0026291804804649314, "grad_norm": 0.164191, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:06.185010+00:00", "epoch": 0, "step": 5409, "train_loss": 3.6564624309539795, "perplexity": 38.724111064314755, "lr": 0.0026291804804649314, "grad_norm": 0.152279, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:06.491754+00:00", "epoch": 0, "step": 5410, "train_loss": 3.5889859199523926, "perplexity": 36.19735029754696, "lr": 0.0026291804804649314, "grad_norm": 0.136173, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:06.797074+00:00", "epoch": 0, "step": 5411, "train_loss": 3.7174479961395264, "perplexity": 41.15922147311144, "lr": 0.0026291804804649314, "grad_norm": 0.155773, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:07.102930+00:00", "epoch": 0, "step": 5412, "train_loss": 3.638730525970459, "perplexity": 38.04351080867675, "lr": 0.0026291804804649314, "grad_norm": 0.171313, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:07.408391+00:00", "epoch": 0, "step": 5413, "train_loss": 3.5626955032348633, "perplexity": 35.25810751489763, "lr": 0.0026291804804649314, "grad_norm": 0.143835, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:07.713897+00:00", "epoch": 0, "step": 5414, "train_loss": 3.5542068481445312, "perplexity": 34.96008031569985, "lr": 0.0026291804804649314, "grad_norm": 0.173715, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:08.020264+00:00", "epoch": 0, "step": 5415, "train_loss": 3.567999839782715, "perplexity": 35.44562527217742, "lr": 0.0026291804804649314, "grad_norm": 0.164624, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:08.326929+00:00", "epoch": 0, "step": 5416, "train_loss": 3.710343360900879, "perplexity": 40.86783653534402, "lr": 0.0026291804804649314, "grad_norm": 0.152165, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:08.632984+00:00", "epoch": 0, "step": 5417, "train_loss": 3.5769968032836914, "perplexity": 35.76596716031219, "lr": 0.0026291804804649314, "grad_norm": 0.154825, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:08.939165+00:00", "epoch": 0, "step": 5418, "train_loss": 3.4898569583892822, "perplexity": 32.781258286862126, "lr": 0.0026291804804649314, "grad_norm": 0.162976, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:09.244371+00:00", "epoch": 0, "step": 5419, "train_loss": 3.5293259620666504, "perplexity": 34.1009745161265, "lr": 0.0026291804804649314, "grad_norm": 0.194462, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:09.551249+00:00", "epoch": 0, "step": 5420, "train_loss": 3.5225532054901123, "perplexity": 33.87079726381206, "lr": 0.0026291804804649314, "grad_norm": 0.236563, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:09.858892+00:00", "epoch": 0, "step": 5421, "train_loss": 3.5887391567230225, "perplexity": 36.18841922446843, "lr": 0.0026291804804649314, "grad_norm": 0.222049, "tokens_per_sec": 106453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:10.164657+00:00", "epoch": 0, "step": 5422, "train_loss": 3.5942797660827637, "perplexity": 36.389481608444314, "lr": 0.0026291804804649314, "grad_norm": 0.170339, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:10.470515+00:00", "epoch": 0, "step": 5423, "train_loss": 3.505070924758911, "perplexity": 33.28380441449536, "lr": 0.0026291804804649314, "grad_norm": 0.159422, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:10.775705+00:00", "epoch": 0, "step": 5424, "train_loss": 3.733703374862671, "perplexity": 41.833747692834514, "lr": 0.0026291804804649314, "grad_norm": 0.150264, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:11.082208+00:00", "epoch": 0, "step": 5425, "train_loss": 3.613199472427368, "perplexity": 37.08451408728604, "lr": 0.0026291804804649314, "grad_norm": 0.155946, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:11.389407+00:00", "epoch": 0, "step": 5426, "train_loss": 3.6451494693756104, "perplexity": 38.28849538137356, "lr": 0.0026291804804649314, "grad_norm": 0.16592, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:11.695650+00:00", "epoch": 0, "step": 5427, "train_loss": 3.558199167251587, "perplexity": 35.099931091037455, "lr": 0.0026291804804649314, "grad_norm": 0.15812, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:12.002085+00:00", "epoch": 0, "step": 5428, "train_loss": 3.6356067657470703, "perplexity": 37.9248574217036, "lr": 0.0026291804804649314, "grad_norm": 0.144559, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:12.308050+00:00", "epoch": 0, "step": 5429, "train_loss": 3.603181838989258, "perplexity": 36.714869591697486, "lr": 0.0026291804804649314, "grad_norm": 0.159638, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:12.615276+00:00", "epoch": 0, "step": 5430, "train_loss": 3.6211328506469727, "perplexity": 37.3798896775618, "lr": 0.0026291804804649314, "grad_norm": 0.1416, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:12.921844+00:00", "epoch": 0, "step": 5431, "train_loss": 3.59401273727417, "perplexity": 36.37976586577439, "lr": 0.0026291804804649314, "grad_norm": 0.13976, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:13.229259+00:00", "epoch": 0, "step": 5432, "train_loss": 3.561081647872925, "perplexity": 35.201251919726644, "lr": 0.0026291804804649314, "grad_norm": 0.163105, "tokens_per_sec": 106592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:13.535370+00:00", "epoch": 0, "step": 5433, "train_loss": 3.594618558883667, "perplexity": 36.401812191480296, "lr": 0.0026291804804649314, "grad_norm": 0.158652, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:13.841068+00:00", "epoch": 0, "step": 5434, "train_loss": 3.524927854537964, "perplexity": 33.95132409392415, "lr": 0.0026291804804649314, "grad_norm": 0.159942, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:14.147383+00:00", "epoch": 0, "step": 5435, "train_loss": 3.5502188205718994, "perplexity": 34.82093619117851, "lr": 0.0026291804804649314, "grad_norm": 0.161887, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:14.454876+00:00", "epoch": 0, "step": 5436, "train_loss": 3.5734219551086426, "perplexity": 35.638337522090566, "lr": 0.0026291804804649314, "grad_norm": 0.176594, "tokens_per_sec": 106566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:14.761676+00:00", "epoch": 0, "step": 5437, "train_loss": 3.545733690261841, "perplexity": 34.66510946764581, "lr": 0.0026291804804649314, "grad_norm": 0.181562, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:15.067282+00:00", "epoch": 0, "step": 5438, "train_loss": 3.6494956016540527, "perplexity": 38.45526438461118, "lr": 0.0026291804804649314, "grad_norm": 0.177769, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:15.373242+00:00", "epoch": 0, "step": 5439, "train_loss": 3.5900514125823975, "perplexity": 36.23593886177938, "lr": 0.0026291804804649314, "grad_norm": 0.172504, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:15.679099+00:00", "epoch": 0, "step": 5440, "train_loss": 3.657210111618042, "perplexity": 38.753075159988384, "lr": 0.0026291804804649314, "grad_norm": 0.177135, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:15.985753+00:00", "epoch": 0, "step": 5441, "train_loss": 3.686265230178833, "perplexity": 39.895567606902375, "lr": 0.0026291804804649314, "grad_norm": 0.157358, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:16.292668+00:00", "epoch": 0, "step": 5442, "train_loss": 3.631821870803833, "perplexity": 37.781587123173416, "lr": 0.0026291804804649314, "grad_norm": 0.151713, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:16.599506+00:00", "epoch": 0, "step": 5443, "train_loss": 3.671679973602295, "perplexity": 39.31790344756964, "lr": 0.0026291804804649314, "grad_norm": 0.137133, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:16.905496+00:00", "epoch": 0, "step": 5444, "train_loss": 3.62284517288208, "perplexity": 37.44395092489911, "lr": 0.0026291804804649314, "grad_norm": 0.149835, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:17.212147+00:00", "epoch": 0, "step": 5445, "train_loss": 3.6374330520629883, "perplexity": 37.99418235416534, "lr": 0.0026291804804649314, "grad_norm": 0.141088, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:17.518491+00:00", "epoch": 0, "step": 5446, "train_loss": 3.595768928527832, "perplexity": 36.44371182663151, "lr": 0.0026291804804649314, "grad_norm": 0.16042, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:17.824582+00:00", "epoch": 0, "step": 5447, "train_loss": 3.579786539077759, "perplexity": 35.86588406519042, "lr": 0.0026291804804649314, "grad_norm": 0.155479, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:18.131450+00:00", "epoch": 0, "step": 5448, "train_loss": 3.607790470123291, "perplexity": 36.88446538482886, "lr": 0.0026291804804649314, "grad_norm": 0.146355, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:18.437320+00:00", "epoch": 0, "step": 5449, "train_loss": 3.61061954498291, "perplexity": 36.988962043297924, "lr": 0.0026291804804649314, "grad_norm": 0.162395, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:18.742478+00:00", "epoch": 0, "step": 5450, "train_loss": 3.626023769378662, "perplexity": 37.56315949372842, "lr": 0.0026291804804649314, "grad_norm": 0.158238, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:19.049476+00:00", "epoch": 0, "step": 5451, "train_loss": 3.517971992492676, "perplexity": 33.715982816895824, "lr": 0.0026291804804649314, "grad_norm": 0.146646, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:19.356790+00:00", "epoch": 0, "step": 5452, "train_loss": 3.4930903911590576, "perplexity": 32.887425831951795, "lr": 0.0026291804804649314, "grad_norm": 0.146645, "tokens_per_sec": 106628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:19.663458+00:00", "epoch": 0, "step": 5453, "train_loss": 3.6124460697174072, "perplexity": 37.0565850361076, "lr": 0.0026291804804649314, "grad_norm": 0.151689, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:19.971271+00:00", "epoch": 0, "step": 5454, "train_loss": 3.7079615592956543, "perplexity": 40.77061328589, "lr": 0.0026291804804649314, "grad_norm": 0.149774, "tokens_per_sec": 106455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:20.276663+00:00", "epoch": 0, "step": 5455, "train_loss": 3.6111104488372803, "perplexity": 37.007124524985514, "lr": 0.0026291804804649314, "grad_norm": 0.161549, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:20.583090+00:00", "epoch": 0, "step": 5456, "train_loss": 3.5322537422180176, "perplexity": 34.200960970212144, "lr": 0.0026291804804649314, "grad_norm": 0.21114, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:20.890686+00:00", "epoch": 0, "step": 5457, "train_loss": 3.611722230911255, "perplexity": 37.02977174725519, "lr": 0.0026291804804649314, "grad_norm": 0.173868, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:21.197060+00:00", "epoch": 0, "step": 5458, "train_loss": 3.7142333984375, "perplexity": 41.02712356891074, "lr": 0.0026291804804649314, "grad_norm": 0.163875, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:21.501790+00:00", "epoch": 0, "step": 5459, "train_loss": 3.530486822128296, "perplexity": 34.14058396156897, "lr": 0.0026291804804649314, "grad_norm": 0.185497, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:21.808056+00:00", "epoch": 0, "step": 5460, "train_loss": 3.6039788722991943, "perplexity": 36.744144230613465, "lr": 0.0026291804804649314, "grad_norm": 0.166758, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:22.114176+00:00", "epoch": 0, "step": 5461, "train_loss": 3.6178970336914062, "perplexity": 37.259130679120645, "lr": 0.0026291804804649314, "grad_norm": 0.148641, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:22.420553+00:00", "epoch": 0, "step": 5462, "train_loss": 3.511953115463257, "perplexity": 33.513659951412215, "lr": 0.0026291804804649314, "grad_norm": 0.144627, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:22.727732+00:00", "epoch": 0, "step": 5463, "train_loss": 3.5576834678649902, "perplexity": 35.08183474464264, "lr": 0.0026291804804649314, "grad_norm": 0.140469, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:23.033870+00:00", "epoch": 0, "step": 5464, "train_loss": 3.6336729526519775, "perplexity": 37.851588702629364, "lr": 0.0026291804804649314, "grad_norm": 0.139459, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:23.340122+00:00", "epoch": 0, "step": 5465, "train_loss": 3.6218302249908447, "perplexity": 37.4059665452121, "lr": 0.0026291804804649314, "grad_norm": 0.131229, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:23.645374+00:00", "epoch": 0, "step": 5466, "train_loss": 3.636564254760742, "perplexity": 37.96118744605398, "lr": 0.0026291804804649314, "grad_norm": 0.168348, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:23.952484+00:00", "epoch": 0, "step": 5467, "train_loss": 3.5697975158691406, "perplexity": 35.50940233317096, "lr": 0.0026291804804649314, "grad_norm": 0.167591, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:24.257830+00:00", "epoch": 0, "step": 5468, "train_loss": 3.5996387004852295, "perplexity": 36.58501390776245, "lr": 0.0026291804804649314, "grad_norm": 0.160084, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:24.563843+00:00", "epoch": 0, "step": 5469, "train_loss": 3.5249650478363037, "perplexity": 33.952586879133634, "lr": 0.0026291804804649314, "grad_norm": 0.149597, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:24.870064+00:00", "epoch": 0, "step": 5470, "train_loss": 3.6638143062591553, "perplexity": 39.00985498997994, "lr": 0.0026291804804649314, "grad_norm": 0.145962, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:25.177720+00:00", "epoch": 0, "step": 5471, "train_loss": 3.4815051555633545, "perplexity": 32.50861579422663, "lr": 0.0026291804804649314, "grad_norm": 0.161951, "tokens_per_sec": 106450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:25.484670+00:00", "epoch": 0, "step": 5472, "train_loss": 3.576430320739746, "perplexity": 35.74571210185873, "lr": 0.0026291804804649314, "grad_norm": 0.145576, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:25.790206+00:00", "epoch": 0, "step": 5473, "train_loss": 3.5047550201416016, "perplexity": 33.273291567615054, "lr": 0.0026291804804649314, "grad_norm": 0.157536, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:26.095258+00:00", "epoch": 0, "step": 5474, "train_loss": 3.6251518726348877, "perplexity": 37.53042257096203, "lr": 0.0026291804804649314, "grad_norm": 0.182281, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:26.401457+00:00", "epoch": 0, "step": 5475, "train_loss": 3.5564959049224854, "perplexity": 35.040197586011566, "lr": 0.0026291804804649314, "grad_norm": 0.147667, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:26.706749+00:00", "epoch": 0, "step": 5476, "train_loss": 3.5213210582733154, "perplexity": 33.8290891557741, "lr": 0.0026291804804649314, "grad_norm": 0.145218, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:27.013670+00:00", "epoch": 0, "step": 5477, "train_loss": 3.553025007247925, "perplexity": 34.91878746862812, "lr": 0.0026291804804649314, "grad_norm": 0.164608, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:27.319688+00:00", "epoch": 0, "step": 5478, "train_loss": 3.5880942344665527, "perplexity": 36.16508803169553, "lr": 0.0026291804804649314, "grad_norm": 0.130306, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:27.625626+00:00", "epoch": 0, "step": 5479, "train_loss": 3.576566219329834, "perplexity": 35.75057022383355, "lr": 0.0026291804804649314, "grad_norm": 0.170663, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:27.930983+00:00", "epoch": 0, "step": 5480, "train_loss": 3.514659881591797, "perplexity": 33.604496472160804, "lr": 0.0026291804804649314, "grad_norm": 0.182106, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:28.236046+00:00", "epoch": 0, "step": 5481, "train_loss": 3.6582233905792236, "perplexity": 38.79236273700521, "lr": 0.0026291804804649314, "grad_norm": 0.163453, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:28.541166+00:00", "epoch": 0, "step": 5482, "train_loss": 3.6483569145202637, "perplexity": 38.4115007910791, "lr": 0.0026291804804649314, "grad_norm": 0.203666, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:28.847410+00:00", "epoch": 0, "step": 5483, "train_loss": 3.5554099082946777, "perplexity": 35.0021647051236, "lr": 0.0026291804804649314, "grad_norm": 0.191689, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:29.153005+00:00", "epoch": 0, "step": 5484, "train_loss": 3.629615545272827, "perplexity": 37.698320533280814, "lr": 0.0026291804804649314, "grad_norm": 0.142326, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:29.458045+00:00", "epoch": 0, "step": 5485, "train_loss": 3.5683202743530273, "perplexity": 35.456985095823875, "lr": 0.0026291804804649314, "grad_norm": 0.167753, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:29.763158+00:00", "epoch": 0, "step": 5486, "train_loss": 3.663053512573242, "perplexity": 38.98018782534191, "lr": 0.0026291804804649314, "grad_norm": 0.154803, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:30.069857+00:00", "epoch": 0, "step": 5487, "train_loss": 3.453500270843506, "perplexity": 31.610845408616125, "lr": 0.0026291804804649314, "grad_norm": 0.148396, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:30.375436+00:00", "epoch": 0, "step": 5488, "train_loss": 3.6255476474761963, "perplexity": 37.54527910772712, "lr": 0.0026291804804649314, "grad_norm": 0.175353, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:30.682107+00:00", "epoch": 0, "step": 5489, "train_loss": 3.5394039154052734, "perplexity": 34.44638011342201, "lr": 0.0026291804804649314, "grad_norm": 0.154144, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:30.988222+00:00", "epoch": 0, "step": 5490, "train_loss": 3.6175620555877686, "perplexity": 37.24665177637884, "lr": 0.0026291804804649314, "grad_norm": 0.149607, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:31.293658+00:00", "epoch": 0, "step": 5491, "train_loss": 3.473773956298828, "perplexity": 32.25825425223017, "lr": 0.0026291804804649314, "grad_norm": 0.150073, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:31.599634+00:00", "epoch": 0, "step": 5492, "train_loss": 3.53617000579834, "perplexity": 34.3351635629115, "lr": 0.0026291804804649314, "grad_norm": 0.152667, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:31.906285+00:00", "epoch": 0, "step": 5493, "train_loss": 3.5782828330993652, "perplexity": 35.81199284932756, "lr": 0.0026291804804649314, "grad_norm": 0.160684, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:32.213789+00:00", "epoch": 0, "step": 5494, "train_loss": 3.44097638130188, "perplexity": 31.217423401549294, "lr": 0.0026291804804649314, "grad_norm": 0.16761, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:32.519279+00:00", "epoch": 0, "step": 5495, "train_loss": 3.6082520484924316, "perplexity": 36.90149438601476, "lr": 0.0026291804804649314, "grad_norm": 0.198278, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:32.824184+00:00", "epoch": 0, "step": 5496, "train_loss": 3.561204433441162, "perplexity": 35.205574390809375, "lr": 0.0026291804804649314, "grad_norm": 0.156186, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:33.130134+00:00", "epoch": 0, "step": 5497, "train_loss": 3.509981870651245, "perplexity": 33.44766139412975, "lr": 0.0026291804804649314, "grad_norm": 0.165779, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:33.435891+00:00", "epoch": 0, "step": 5498, "train_loss": 3.5504024028778076, "perplexity": 34.82732928574994, "lr": 0.0026291804804649314, "grad_norm": 0.187599, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:33.742189+00:00", "epoch": 0, "step": 5499, "train_loss": 3.5156233310699463, "perplexity": 33.636888308193534, "lr": 0.0026291804804649314, "grad_norm": 0.16982, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:34.050218+00:00", "epoch": 0, "step": 5500, "train_loss": 3.6282196044921875, "perplexity": 37.645732623630224, "lr": 0.0026291804804649314, "grad_norm": 0.140197, "tokens_per_sec": 106460} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:26:37.123906+00:00", "step": 5500, "epoch": 0, "val_loss": 3.550050139427185, "val_ppl": 34.815063051159676, "eval_train_loss": 3.6282196044921875, "eval_train_ppl": 37.645732623630224} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:26:38.084319+00:00", "step": 5500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5501_epoch_0000_step_0005500.pt", "val_loss": 3.550050139427185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:39.148027+00:00", "epoch": 0, "step": 5501, "train_loss": 3.6491897106170654, "perplexity": 38.443503062844115, "lr": 0.0026291804804649314, "grad_norm": 0.160862, "tokens_per_sec": 6428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:39.453617+00:00", "epoch": 0, "step": 5502, "train_loss": 3.5709118843078613, "perplexity": 35.54899494669111, "lr": 0.0026291804804649314, "grad_norm": 0.152677, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:39.758896+00:00", "epoch": 0, "step": 5503, "train_loss": 3.697587013244629, "perplexity": 40.34982320801918, "lr": 0.0026291804804649314, "grad_norm": 0.149926, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:40.065768+00:00", "epoch": 0, "step": 5504, "train_loss": 3.6343069076538086, "perplexity": 37.87559251447902, "lr": 0.0026291804804649314, "grad_norm": 0.144052, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:40.369884+00:00", "epoch": 0, "step": 5505, "train_loss": 3.646507501602173, "perplexity": 38.34052771480414, "lr": 0.0026291804804649314, "grad_norm": 0.142772, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:40.673256+00:00", "epoch": 0, "step": 5506, "train_loss": 3.6867752075195312, "perplexity": 39.91591863121575, "lr": 0.0026291804804649314, "grad_norm": 0.149419, "tokens_per_sec": 108011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:40.978553+00:00", "epoch": 0, "step": 5507, "train_loss": 3.5426368713378906, "perplexity": 34.557923953371734, "lr": 0.0026291804804649314, "grad_norm": 0.139848, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:41.283699+00:00", "epoch": 0, "step": 5508, "train_loss": 3.650480270385742, "perplexity": 38.49314872972362, "lr": 0.0026291804804649314, "grad_norm": 0.142882, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:41.588585+00:00", "epoch": 0, "step": 5509, "train_loss": 3.553788423538208, "perplexity": 34.945455217820516, "lr": 0.0026291804804649314, "grad_norm": 0.15851, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:41.893329+00:00", "epoch": 0, "step": 5510, "train_loss": 3.5237176418304443, "perplexity": 33.91026062284767, "lr": 0.0026291804804649314, "grad_norm": 0.164427, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:42.198555+00:00", "epoch": 0, "step": 5511, "train_loss": 3.6658942699432373, "perplexity": 39.09107851338713, "lr": 0.0026291804804649314, "grad_norm": 0.154675, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:42.503748+00:00", "epoch": 0, "step": 5512, "train_loss": 3.499021530151367, "perplexity": 33.08306533467835, "lr": 0.0026291804804649314, "grad_norm": 0.163312, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:42.810275+00:00", "epoch": 0, "step": 5513, "train_loss": 3.6240642070770264, "perplexity": 37.4896242144653, "lr": 0.0026291804804649314, "grad_norm": 0.158537, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:43.116161+00:00", "epoch": 0, "step": 5514, "train_loss": 3.6254258155822754, "perplexity": 37.540705173896804, "lr": 0.0026291804804649314, "grad_norm": 0.154201, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:43.420991+00:00", "epoch": 0, "step": 5515, "train_loss": 3.6206142902374268, "perplexity": 37.36051097161092, "lr": 0.0026291804804649314, "grad_norm": 0.171042, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:43.725444+00:00", "epoch": 0, "step": 5516, "train_loss": 3.6171884536743164, "perplexity": 37.232738955095655, "lr": 0.0026291804804649314, "grad_norm": 0.139481, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:44.029609+00:00", "epoch": 0, "step": 5517, "train_loss": 3.617879867553711, "perplexity": 37.25849108924266, "lr": 0.0026291804804649314, "grad_norm": 0.163107, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:44.334940+00:00", "epoch": 0, "step": 5518, "train_loss": 3.6060566902160645, "perplexity": 36.82057124504916, "lr": 0.0026291804804649314, "grad_norm": 0.159881, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:44.639846+00:00", "epoch": 0, "step": 5519, "train_loss": 3.6045703887939453, "perplexity": 36.765885427515755, "lr": 0.0026291804804649314, "grad_norm": 0.132159, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:44.945076+00:00", "epoch": 0, "step": 5520, "train_loss": 3.5647354125976562, "perplexity": 35.33010426698972, "lr": 0.0026291804804649314, "grad_norm": 0.148056, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:45.250421+00:00", "epoch": 0, "step": 5521, "train_loss": 3.595432996749878, "perplexity": 36.431471281831435, "lr": 0.0026291804804649314, "grad_norm": 0.140304, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:45.555647+00:00", "epoch": 0, "step": 5522, "train_loss": 3.6003973484039307, "perplexity": 36.61277958327349, "lr": 0.0026291804804649314, "grad_norm": 0.151828, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:45.860810+00:00", "epoch": 0, "step": 5523, "train_loss": 3.477365493774414, "perplexity": 32.3743192824653, "lr": 0.0026291804804649314, "grad_norm": 0.14099, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:46.167000+00:00", "epoch": 0, "step": 5524, "train_loss": 3.556755304336548, "perplexity": 35.04928817172934, "lr": 0.0026291804804649314, "grad_norm": 0.150891, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:46.473031+00:00", "epoch": 0, "step": 5525, "train_loss": 3.529146671295166, "perplexity": 34.094861074155375, "lr": 0.0026291804804649314, "grad_norm": 0.155969, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:46.777737+00:00", "epoch": 0, "step": 5526, "train_loss": 3.507855176925659, "perplexity": 33.3766040478952, "lr": 0.0026291804804649314, "grad_norm": 0.162301, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:47.082552+00:00", "epoch": 0, "step": 5527, "train_loss": 3.5884368419647217, "perplexity": 36.17748058479672, "lr": 0.0026291804804649314, "grad_norm": 0.161495, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:47.388632+00:00", "epoch": 0, "step": 5528, "train_loss": 3.6149680614471436, "perplexity": 37.150159384369296, "lr": 0.0026291804804649314, "grad_norm": 0.144106, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:47.694459+00:00", "epoch": 0, "step": 5529, "train_loss": 3.585099220275879, "perplexity": 36.056935120430914, "lr": 0.0026291804804649314, "grad_norm": 0.177298, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:47.999905+00:00", "epoch": 0, "step": 5530, "train_loss": 3.613205909729004, "perplexity": 37.084752812257605, "lr": 0.0026291804804649314, "grad_norm": 0.160841, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:48.305011+00:00", "epoch": 0, "step": 5531, "train_loss": 3.5435569286346436, "perplexity": 34.58973385468764, "lr": 0.0026291804804649314, "grad_norm": 0.146321, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:48.610379+00:00", "epoch": 0, "step": 5532, "train_loss": 3.561046838760376, "perplexity": 35.20002661671269, "lr": 0.0026291804804649314, "grad_norm": 0.144271, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:48.915378+00:00", "epoch": 0, "step": 5533, "train_loss": 3.563931703567505, "perplexity": 35.30172055080652, "lr": 0.0026291804804649314, "grad_norm": 0.146027, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:49.220738+00:00", "epoch": 0, "step": 5534, "train_loss": 3.6125223636627197, "perplexity": 37.05941233703141, "lr": 0.0026291804804649314, "grad_norm": 0.18171, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:49.527326+00:00", "epoch": 0, "step": 5535, "train_loss": 3.551377773284912, "perplexity": 34.86131540392632, "lr": 0.0026291804804649314, "grad_norm": 0.167937, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:49.832407+00:00", "epoch": 0, "step": 5536, "train_loss": 3.4851179122924805, "perplexity": 32.62627392181498, "lr": 0.0026291804804649314, "grad_norm": 0.178243, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:50.137025+00:00", "epoch": 0, "step": 5537, "train_loss": 3.559236526489258, "perplexity": 35.13636122109541, "lr": 0.0026291804804649314, "grad_norm": 0.175834, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:50.442219+00:00", "epoch": 0, "step": 5538, "train_loss": 3.579371213912964, "perplexity": 35.85099115389503, "lr": 0.0026291804804649314, "grad_norm": 0.162244, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:50.747310+00:00", "epoch": 0, "step": 5539, "train_loss": 3.7011146545410156, "perplexity": 40.492414267855374, "lr": 0.0026291804804649314, "grad_norm": 0.167673, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:51.053211+00:00", "epoch": 0, "step": 5540, "train_loss": 3.5090689659118652, "perplexity": 33.41714079883957, "lr": 0.0026291804804649314, "grad_norm": 0.168267, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:51.359378+00:00", "epoch": 0, "step": 5541, "train_loss": 3.8016982078552246, "perplexity": 44.77716088954256, "lr": 0.0026291804804649314, "grad_norm": 0.158135, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:51.664416+00:00", "epoch": 0, "step": 5542, "train_loss": 3.467252492904663, "perplexity": 32.04856770269764, "lr": 0.0026291804804649314, "grad_norm": 0.175416, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:51.971242+00:00", "epoch": 0, "step": 5543, "train_loss": 3.5467495918273926, "perplexity": 34.70034370085005, "lr": 0.0026291804804649314, "grad_norm": 0.153612, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:52.276124+00:00", "epoch": 0, "step": 5544, "train_loss": 3.521804094314575, "perplexity": 33.845433772280956, "lr": 0.0026291804804649314, "grad_norm": 0.169291, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:52.582499+00:00", "epoch": 0, "step": 5545, "train_loss": 3.6107289791107178, "perplexity": 36.99301011959245, "lr": 0.0026291804804649314, "grad_norm": 0.158661, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:52.887818+00:00", "epoch": 0, "step": 5546, "train_loss": 3.520490884780884, "perplexity": 33.801016796760294, "lr": 0.0026291804804649314, "grad_norm": 0.144578, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:53.194673+00:00", "epoch": 0, "step": 5547, "train_loss": 3.6488897800445557, "perplexity": 38.431974409945354, "lr": 0.0026291804804649314, "grad_norm": 0.158966, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:53.499267+00:00", "epoch": 0, "step": 5548, "train_loss": 3.5911824703216553, "perplexity": 36.27694698777563, "lr": 0.0026291804804649314, "grad_norm": 0.155978, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:53.804888+00:00", "epoch": 0, "step": 5549, "train_loss": 3.6994011402130127, "perplexity": 40.42308934741646, "lr": 0.0026291804804649314, "grad_norm": 0.160817, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:54.110504+00:00", "epoch": 0, "step": 5550, "train_loss": 3.5198240280151367, "perplexity": 33.778483873971275, "lr": 0.0026291804804649314, "grad_norm": 0.150189, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:54.416063+00:00", "epoch": 0, "step": 5551, "train_loss": 3.6954996585845947, "perplexity": 40.26568665847268, "lr": 0.0026291804804649314, "grad_norm": 0.156609, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:54.721579+00:00", "epoch": 0, "step": 5552, "train_loss": 3.585767984390259, "perplexity": 36.08105676966165, "lr": 0.0026291804804649314, "grad_norm": 0.15126, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:55.027118+00:00", "epoch": 0, "step": 5553, "train_loss": 3.475412368774414, "perplexity": 32.311149899048736, "lr": 0.0026291804804649314, "grad_norm": 0.143097, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:55.331754+00:00", "epoch": 0, "step": 5554, "train_loss": 3.5390264987945557, "perplexity": 34.43338193040749, "lr": 0.0026291804804649314, "grad_norm": 0.157061, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:55.636858+00:00", "epoch": 0, "step": 5555, "train_loss": 3.6403579711914062, "perplexity": 38.10547494647395, "lr": 0.0026291804804649314, "grad_norm": 0.175131, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:55.942925+00:00", "epoch": 0, "step": 5556, "train_loss": 3.5951731204986572, "perplexity": 36.42200483775353, "lr": 0.0026291804804649314, "grad_norm": 0.169507, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:56.248688+00:00", "epoch": 0, "step": 5557, "train_loss": 3.584895133972168, "perplexity": 36.04957714467562, "lr": 0.0026291804804649314, "grad_norm": 0.169577, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:56.554344+00:00", "epoch": 0, "step": 5558, "train_loss": 3.59262752532959, "perplexity": 36.32940706641371, "lr": 0.0026291804804649314, "grad_norm": 0.177945, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:56.860896+00:00", "epoch": 0, "step": 5559, "train_loss": 3.5301918983459473, "perplexity": 34.130516576044116, "lr": 0.0026291804804649314, "grad_norm": 0.190856, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:57.165764+00:00", "epoch": 0, "step": 5560, "train_loss": 3.662837505340576, "perplexity": 38.971768732166204, "lr": 0.0026291804804649314, "grad_norm": 0.172513, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:57.471331+00:00", "epoch": 0, "step": 5561, "train_loss": 3.4953134059906006, "perplexity": 32.96061638905342, "lr": 0.0026291804804649314, "grad_norm": 0.182529, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:57.776515+00:00", "epoch": 0, "step": 5562, "train_loss": 3.5348424911499023, "perplexity": 34.28961337128795, "lr": 0.0026291804804649314, "grad_norm": 0.202463, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:58.082489+00:00", "epoch": 0, "step": 5563, "train_loss": 3.5450522899627686, "perplexity": 34.64149669747515, "lr": 0.0026291804804649314, "grad_norm": 0.199889, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:58.387610+00:00", "epoch": 0, "step": 5564, "train_loss": 3.5123095512390137, "perplexity": 33.525607547944375, "lr": 0.0026291804804649314, "grad_norm": 0.188763, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:58.692770+00:00", "epoch": 0, "step": 5565, "train_loss": 3.734844923019409, "perplexity": 41.88153019822929, "lr": 0.0026291804804649314, "grad_norm": 0.169827, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:58.997906+00:00", "epoch": 0, "step": 5566, "train_loss": 3.6059107780456543, "perplexity": 36.815199067525626, "lr": 0.0026291804804649314, "grad_norm": 0.170574, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:59.304009+00:00", "epoch": 0, "step": 5567, "train_loss": 3.517049551010132, "perplexity": 33.68489613574721, "lr": 0.0026291804804649314, "grad_norm": 0.183195, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:59.608662+00:00", "epoch": 0, "step": 5568, "train_loss": 3.5733041763305664, "perplexity": 35.634140329419516, "lr": 0.0026291804804649314, "grad_norm": 0.157874, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:26:59.914994+00:00", "epoch": 0, "step": 5569, "train_loss": 3.5582354068756104, "perplexity": 35.10120312239227, "lr": 0.0026291804804649314, "grad_norm": 0.148732, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:00.220893+00:00", "epoch": 0, "step": 5570, "train_loss": 3.640421152114868, "perplexity": 38.107882561626894, "lr": 0.0026291804804649314, "grad_norm": 0.155428, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:00.526115+00:00", "epoch": 0, "step": 5571, "train_loss": 3.493346929550171, "perplexity": 32.895863801547875, "lr": 0.0026291804804649314, "grad_norm": 0.147654, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:00.831986+00:00", "epoch": 0, "step": 5572, "train_loss": 3.5634710788726807, "perplexity": 35.28546345104929, "lr": 0.0026291804804649314, "grad_norm": 0.133575, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:01.138408+00:00", "epoch": 0, "step": 5573, "train_loss": 3.5595853328704834, "perplexity": 35.14861914579927, "lr": 0.0026291804804649314, "grad_norm": 0.155162, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:01.445142+00:00", "epoch": 0, "step": 5574, "train_loss": 3.592716693878174, "perplexity": 36.33264665134514, "lr": 0.0026291804804649314, "grad_norm": 0.147278, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:01.751105+00:00", "epoch": 0, "step": 5575, "train_loss": 3.635754346847534, "perplexity": 37.930454826922286, "lr": 0.0026291804804649314, "grad_norm": 0.172996, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:02.055970+00:00", "epoch": 0, "step": 5576, "train_loss": 3.64477276802063, "perplexity": 38.27407476958576, "lr": 0.0026291804804649314, "grad_norm": 0.14899, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:02.361119+00:00", "epoch": 0, "step": 5577, "train_loss": 3.7413382530212402, "perplexity": 42.15436564143551, "lr": 0.0026291804804649314, "grad_norm": 0.157046, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:02.666535+00:00", "epoch": 0, "step": 5578, "train_loss": 3.6603477001190186, "perplexity": 38.8748573142246, "lr": 0.0026291804804649314, "grad_norm": 0.162933, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:02.972647+00:00", "epoch": 0, "step": 5579, "train_loss": 3.5705549716949463, "perplexity": 35.53630932598186, "lr": 0.0026291804804649314, "grad_norm": 0.155788, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:03.278566+00:00", "epoch": 0, "step": 5580, "train_loss": 3.5346767902374268, "perplexity": 34.28393202177929, "lr": 0.0026291804804649314, "grad_norm": 0.169461, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:03.583816+00:00", "epoch": 0, "step": 5581, "train_loss": 3.662283420562744, "perplexity": 38.95018104960197, "lr": 0.0026291804804649314, "grad_norm": 0.149588, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:03.888982+00:00", "epoch": 0, "step": 5582, "train_loss": 3.5666933059692383, "perplexity": 35.39934460443171, "lr": 0.0026291804804649314, "grad_norm": 0.158333, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:04.194683+00:00", "epoch": 0, "step": 5583, "train_loss": 3.531883478164673, "perplexity": 34.1882999278841, "lr": 0.0026291804804649314, "grad_norm": 0.16102, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:04.499700+00:00", "epoch": 0, "step": 5584, "train_loss": 3.557236433029175, "perplexity": 35.06615544726465, "lr": 0.0026291804804649314, "grad_norm": 0.148863, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:04.804977+00:00", "epoch": 0, "step": 5585, "train_loss": 3.669013023376465, "perplexity": 39.213184258592335, "lr": 0.0026291804804649314, "grad_norm": 0.171699, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:05.111236+00:00", "epoch": 0, "step": 5586, "train_loss": 3.6828811168670654, "perplexity": 39.76078467447158, "lr": 0.0026291804804649314, "grad_norm": 0.144503, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:05.416696+00:00", "epoch": 0, "step": 5587, "train_loss": 3.5788702964782715, "perplexity": 35.833037264458575, "lr": 0.0026291804804649314, "grad_norm": 0.139004, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:05.722292+00:00", "epoch": 0, "step": 5588, "train_loss": 3.6366705894470215, "perplexity": 37.96522425163427, "lr": 0.0026291804804649314, "grad_norm": 0.169022, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:06.027848+00:00", "epoch": 0, "step": 5589, "train_loss": 3.6581976413726807, "perplexity": 38.79136387730479, "lr": 0.0026291804804649314, "grad_norm": 0.143677, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:06.332902+00:00", "epoch": 0, "step": 5590, "train_loss": 3.540780782699585, "perplexity": 34.49384087369403, "lr": 0.0026291804804649314, "grad_norm": 0.124152, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:06.639571+00:00", "epoch": 0, "step": 5591, "train_loss": 3.5905370712280273, "perplexity": 36.25354143284455, "lr": 0.0026291804804649314, "grad_norm": 0.144224, "tokens_per_sec": 106797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:06.945059+00:00", "epoch": 0, "step": 5592, "train_loss": 3.562129497528076, "perplexity": 35.238156871455324, "lr": 0.0026291804804649314, "grad_norm": 0.168735, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:07.250191+00:00", "epoch": 0, "step": 5593, "train_loss": 3.6882736682891846, "perplexity": 39.975775905057404, "lr": 0.0026291804804649314, "grad_norm": 0.163289, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:07.555930+00:00", "epoch": 0, "step": 5594, "train_loss": 3.560800313949585, "perplexity": 35.19135000635497, "lr": 0.0026291804804649314, "grad_norm": 0.170107, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:07.861752+00:00", "epoch": 0, "step": 5595, "train_loss": 3.6104605197906494, "perplexity": 36.98308033417982, "lr": 0.0026291804804649314, "grad_norm": 0.170766, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:08.168409+00:00", "epoch": 0, "step": 5596, "train_loss": 3.6417572498321533, "perplexity": 38.15883244596067, "lr": 0.0026291804804649314, "grad_norm": 0.16636, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:08.473435+00:00", "epoch": 0, "step": 5597, "train_loss": 3.559460401535034, "perplexity": 35.14422825615576, "lr": 0.0026291804804649314, "grad_norm": 0.161797, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:08.778541+00:00", "epoch": 0, "step": 5598, "train_loss": 3.620434045791626, "perplexity": 37.353777553864816, "lr": 0.0026291804804649314, "grad_norm": 0.159722, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:09.085942+00:00", "epoch": 0, "step": 5599, "train_loss": 3.5232045650482178, "perplexity": 33.89286651808007, "lr": 0.0026291804804649314, "grad_norm": 0.158621, "tokens_per_sec": 106605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:09.393044+00:00", "epoch": 0, "step": 5600, "train_loss": 3.5452561378479004, "perplexity": 34.64855901310933, "lr": 0.0026291804804649314, "grad_norm": 0.138117, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:09.699362+00:00", "epoch": 0, "step": 5601, "train_loss": 3.5648820400238037, "perplexity": 35.33528500905407, "lr": 0.0026291804804649314, "grad_norm": 0.159837, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:10.006381+00:00", "epoch": 0, "step": 5602, "train_loss": 3.5588972568511963, "perplexity": 35.12444254247835, "lr": 0.0026291804804649314, "grad_norm": 0.182693, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:10.312581+00:00", "epoch": 0, "step": 5603, "train_loss": 3.510319948196411, "perplexity": 33.45897120907494, "lr": 0.0026291804804649314, "grad_norm": 0.161993, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:10.616890+00:00", "epoch": 0, "step": 5604, "train_loss": 3.5262458324432373, "perplexity": 33.99610068973504, "lr": 0.0026291804804649314, "grad_norm": 0.151055, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:10.921378+00:00", "epoch": 0, "step": 5605, "train_loss": 3.556370496749878, "perplexity": 35.03580353439525, "lr": 0.0026291804804649314, "grad_norm": 0.179871, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:11.227602+00:00", "epoch": 0, "step": 5606, "train_loss": 3.5164551734924316, "perplexity": 33.66488053979744, "lr": 0.0026291804804649314, "grad_norm": 0.17643, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:11.534731+00:00", "epoch": 0, "step": 5607, "train_loss": 3.7686843872070312, "perplexity": 43.32303099304673, "lr": 0.0026291804804649314, "grad_norm": 0.170089, "tokens_per_sec": 106698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:11.840262+00:00", "epoch": 0, "step": 5608, "train_loss": 3.6157920360565186, "perplexity": 37.18078278715682, "lr": 0.0026291804804649314, "grad_norm": 0.158308, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:12.145570+00:00", "epoch": 0, "step": 5609, "train_loss": 3.602396011352539, "perplexity": 36.68602936570126, "lr": 0.0026291804804649314, "grad_norm": 0.168543, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:12.450577+00:00", "epoch": 0, "step": 5610, "train_loss": 3.689129114151001, "perplexity": 40.00998764818902, "lr": 0.0026291804804649314, "grad_norm": 0.164318, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:12.755793+00:00", "epoch": 0, "step": 5611, "train_loss": 3.575094223022461, "perplexity": 35.69798422916772, "lr": 0.0026291804804649314, "grad_norm": 0.191694, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:13.061827+00:00", "epoch": 0, "step": 5612, "train_loss": 3.573176145553589, "perplexity": 35.62957835478902, "lr": 0.0026291804804649314, "grad_norm": 0.147377, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:13.367413+00:00", "epoch": 0, "step": 5613, "train_loss": 3.5467100143432617, "perplexity": 34.69897037572445, "lr": 0.0026291804804649314, "grad_norm": 0.165009, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:13.672901+00:00", "epoch": 0, "step": 5614, "train_loss": 3.598144769668579, "perplexity": 36.53039923348808, "lr": 0.0026291804804649314, "grad_norm": 0.15946, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:13.978519+00:00", "epoch": 0, "step": 5615, "train_loss": 3.4835011959075928, "perplexity": 32.573569106037226, "lr": 0.0026291804804649314, "grad_norm": 0.150754, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:14.282548+00:00", "epoch": 0, "step": 5616, "train_loss": 3.5441665649414062, "perplexity": 34.61082744134155, "lr": 0.0026291804804649314, "grad_norm": 0.145203, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:14.588383+00:00", "epoch": 0, "step": 5617, "train_loss": 3.650512218475342, "perplexity": 38.494378531933016, "lr": 0.0026291804804649314, "grad_norm": 0.141903, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:14.894084+00:00", "epoch": 0, "step": 5618, "train_loss": 3.6090121269226074, "perplexity": 36.929553077991265, "lr": 0.0026291804804649314, "grad_norm": 0.155503, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:15.199451+00:00", "epoch": 0, "step": 5619, "train_loss": 3.6006522178649902, "perplexity": 36.62211225192937, "lr": 0.0026291804804649314, "grad_norm": 0.174809, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:15.504110+00:00", "epoch": 0, "step": 5620, "train_loss": 3.6331803798675537, "perplexity": 37.83294863136003, "lr": 0.0026291804804649314, "grad_norm": 0.174895, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:15.810735+00:00", "epoch": 0, "step": 5621, "train_loss": 3.5741171836853027, "perplexity": 35.66312292752333, "lr": 0.0026291804804649314, "grad_norm": 0.159796, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:16.117152+00:00", "epoch": 0, "step": 5622, "train_loss": 3.597088575363159, "perplexity": 36.491836402349165, "lr": 0.0026291804804649314, "grad_norm": 0.165206, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:16.424115+00:00", "epoch": 0, "step": 5623, "train_loss": 3.596027135848999, "perplexity": 36.453123074809945, "lr": 0.0026291804804649314, "grad_norm": 0.155092, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:16.729444+00:00", "epoch": 0, "step": 5624, "train_loss": 3.5964598655700684, "perplexity": 36.46890083809756, "lr": 0.0026291804804649314, "grad_norm": 0.171724, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:17.035467+00:00", "epoch": 0, "step": 5625, "train_loss": 3.545776605606079, "perplexity": 34.666597164673966, "lr": 0.0026291804804649314, "grad_norm": 0.166052, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:17.341696+00:00", "epoch": 0, "step": 5626, "train_loss": 3.742600202560425, "perplexity": 42.207595903619584, "lr": 0.0026291804804649314, "grad_norm": 0.152221, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:17.648042+00:00", "epoch": 0, "step": 5627, "train_loss": 3.5208096504211426, "perplexity": 33.81179311698997, "lr": 0.0026291804804649314, "grad_norm": 0.146172, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:17.954616+00:00", "epoch": 0, "step": 5628, "train_loss": 3.5474236011505127, "perplexity": 34.723739939775754, "lr": 0.0026291804804649314, "grad_norm": 0.153268, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:18.260479+00:00", "epoch": 0, "step": 5629, "train_loss": 3.6559629440307617, "perplexity": 38.70477370700549, "lr": 0.0026291804804649314, "grad_norm": 0.147647, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:18.566047+00:00", "epoch": 0, "step": 5630, "train_loss": 3.5063345432281494, "perplexity": 33.325889028327865, "lr": 0.0026291804804649314, "grad_norm": 0.158762, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:18.870706+00:00", "epoch": 0, "step": 5631, "train_loss": 3.5547239780426025, "perplexity": 34.978163893846535, "lr": 0.0026291804804649314, "grad_norm": 0.14491, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:19.176372+00:00", "epoch": 0, "step": 5632, "train_loss": 3.6721444129943848, "perplexity": 39.33616847191516, "lr": 0.0026291804804649314, "grad_norm": 0.151072, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:19.482383+00:00", "epoch": 0, "step": 5633, "train_loss": 3.622755527496338, "perplexity": 37.440594397925565, "lr": 0.0026291804804649314, "grad_norm": 0.165429, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:19.789743+00:00", "epoch": 0, "step": 5634, "train_loss": 3.585973024368286, "perplexity": 36.08845558724968, "lr": 0.0026291804804649314, "grad_norm": 0.147197, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:20.095497+00:00", "epoch": 0, "step": 5635, "train_loss": 3.6613805294036865, "perplexity": 38.91502914704763, "lr": 0.0026291804804649314, "grad_norm": 0.14825, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:20.400258+00:00", "epoch": 0, "step": 5636, "train_loss": 3.4925050735473633, "perplexity": 32.86818187487197, "lr": 0.0026291804804649314, "grad_norm": 0.159142, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:20.705649+00:00", "epoch": 0, "step": 5637, "train_loss": 3.4313149452209473, "perplexity": 30.917270549311045, "lr": 0.0026291804804649314, "grad_norm": 0.139643, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:21.011456+00:00", "epoch": 0, "step": 5638, "train_loss": 3.473116636276245, "perplexity": 32.23705722319562, "lr": 0.0026291804804649314, "grad_norm": 0.150551, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:21.315674+00:00", "epoch": 0, "step": 5639, "train_loss": 3.6486592292785645, "perplexity": 38.423114910128064, "lr": 0.0026291804804649314, "grad_norm": 0.167428, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:21.621446+00:00", "epoch": 0, "step": 5640, "train_loss": 3.636749267578125, "perplexity": 37.96821140203547, "lr": 0.0026291804804649314, "grad_norm": 0.150281, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:21.926497+00:00", "epoch": 0, "step": 5641, "train_loss": 3.4924614429473877, "perplexity": 32.866747847660626, "lr": 0.0026291804804649314, "grad_norm": 0.14671, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:22.231940+00:00", "epoch": 0, "step": 5642, "train_loss": 3.6131694316864014, "perplexity": 37.083400057737684, "lr": 0.0026291804804649314, "grad_norm": 0.166704, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:22.537187+00:00", "epoch": 0, "step": 5643, "train_loss": 3.50243878364563, "perplexity": 33.19631194127102, "lr": 0.0026291804804649314, "grad_norm": 0.177137, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:22.843011+00:00", "epoch": 0, "step": 5644, "train_loss": 3.5899362564086914, "perplexity": 36.23176630996157, "lr": 0.0026291804804649314, "grad_norm": 0.189115, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:23.148851+00:00", "epoch": 0, "step": 5645, "train_loss": 3.678990364074707, "perplexity": 39.60638584907706, "lr": 0.0026291804804649314, "grad_norm": 0.197572, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:23.455038+00:00", "epoch": 0, "step": 5646, "train_loss": 3.6453640460968018, "perplexity": 38.296712082696736, "lr": 0.0026291804804649314, "grad_norm": 0.163487, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:23.759870+00:00", "epoch": 0, "step": 5647, "train_loss": 3.5909698009490967, "perplexity": 36.26923281253488, "lr": 0.0026291804804649314, "grad_norm": 0.145618, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:24.065723+00:00", "epoch": 0, "step": 5648, "train_loss": 3.656533718109131, "perplexity": 38.72687169442586, "lr": 0.0026291804804649314, "grad_norm": 0.165554, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:24.371500+00:00", "epoch": 0, "step": 5649, "train_loss": 3.5781168937683105, "perplexity": 35.80605072422017, "lr": 0.0026291804804649314, "grad_norm": 0.164297, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:24.677578+00:00", "epoch": 0, "step": 5650, "train_loss": 3.5863499641418457, "perplexity": 36.102061325637735, "lr": 0.0026291804804649314, "grad_norm": 0.168509, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:24.983430+00:00", "epoch": 0, "step": 5651, "train_loss": 3.614978313446045, "perplexity": 37.1505402497148, "lr": 0.0026291804804649314, "grad_norm": 0.146871, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:25.289304+00:00", "epoch": 0, "step": 5652, "train_loss": 3.5158028602600098, "perplexity": 33.64292765361094, "lr": 0.0026291804804649314, "grad_norm": 0.151028, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:25.593869+00:00", "epoch": 0, "step": 5653, "train_loss": 3.5217158794403076, "perplexity": 33.84244823328261, "lr": 0.0026291804804649314, "grad_norm": 0.160069, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:25.899990+00:00", "epoch": 0, "step": 5654, "train_loss": 3.6466221809387207, "perplexity": 38.34492483320986, "lr": 0.0026291804804649314, "grad_norm": 0.163696, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:26.206434+00:00", "epoch": 0, "step": 5655, "train_loss": 3.611687421798706, "perplexity": 37.028482796196535, "lr": 0.0026291804804649314, "grad_norm": 0.158702, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:26.512798+00:00", "epoch": 0, "step": 5656, "train_loss": 3.5994701385498047, "perplexity": 36.57884758672872, "lr": 0.0026291804804649314, "grad_norm": 0.16831, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:26.818277+00:00", "epoch": 0, "step": 5657, "train_loss": 3.5625319480895996, "perplexity": 35.25234134155782, "lr": 0.0026291804804649314, "grad_norm": 0.160013, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:27.123579+00:00", "epoch": 0, "step": 5658, "train_loss": 3.649144172668457, "perplexity": 38.44175246443695, "lr": 0.0026291804804649314, "grad_norm": 0.167969, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:27.429899+00:00", "epoch": 0, "step": 5659, "train_loss": 3.5651824474334717, "perplexity": 35.34590158506273, "lr": 0.0026291804804649314, "grad_norm": 0.162323, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:27.735296+00:00", "epoch": 0, "step": 5660, "train_loss": 3.671595573425293, "perplexity": 39.31458514959376, "lr": 0.0026291804804649314, "grad_norm": 0.152837, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:28.039646+00:00", "epoch": 0, "step": 5661, "train_loss": 3.529987096786499, "perplexity": 34.12352730875542, "lr": 0.0026291804804649314, "grad_norm": 0.170199, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:28.345167+00:00", "epoch": 0, "step": 5662, "train_loss": 3.5958240032196045, "perplexity": 36.445719008099346, "lr": 0.0026291804804649314, "grad_norm": 0.134279, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:28.650195+00:00", "epoch": 0, "step": 5663, "train_loss": 3.52876615524292, "perplexity": 34.081889900244015, "lr": 0.0026291804804649314, "grad_norm": 0.170896, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:28.956431+00:00", "epoch": 0, "step": 5664, "train_loss": 3.6001172065734863, "perplexity": 36.60252424872454, "lr": 0.0026291804804649314, "grad_norm": 0.183046, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:29.262301+00:00", "epoch": 0, "step": 5665, "train_loss": 3.6153783798217773, "perplexity": 37.16540590513518, "lr": 0.0026291804804649314, "grad_norm": 0.165765, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:29.568096+00:00", "epoch": 0, "step": 5666, "train_loss": 3.6227612495422363, "perplexity": 37.440808635338115, "lr": 0.0026291804804649314, "grad_norm": 0.171786, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:29.872330+00:00", "epoch": 0, "step": 5667, "train_loss": 3.5690901279449463, "perplexity": 35.48429229308177, "lr": 0.0026291804804649314, "grad_norm": 0.174883, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:30.177202+00:00", "epoch": 0, "step": 5668, "train_loss": 3.5733957290649414, "perplexity": 35.63740288174929, "lr": 0.0026291804804649314, "grad_norm": 0.178986, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:30.483158+00:00", "epoch": 0, "step": 5669, "train_loss": 3.526081085205078, "perplexity": 33.990500387368066, "lr": 0.0026291804804649314, "grad_norm": 0.155953, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:30.788593+00:00", "epoch": 0, "step": 5670, "train_loss": 3.5860846042633057, "perplexity": 36.09248255799583, "lr": 0.0026291804804649314, "grad_norm": 0.156446, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:31.094545+00:00", "epoch": 0, "step": 5671, "train_loss": 3.603532075881958, "perplexity": 36.727730745633, "lr": 0.0026291804804649314, "grad_norm": 0.169657, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:31.400748+00:00", "epoch": 0, "step": 5672, "train_loss": 3.622799873352051, "perplexity": 37.44225476993758, "lr": 0.0026291804804649314, "grad_norm": 0.153434, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:31.707240+00:00", "epoch": 0, "step": 5673, "train_loss": 3.7232563495635986, "perplexity": 41.39898441797527, "lr": 0.0026291804804649314, "grad_norm": 0.173557, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:32.014440+00:00", "epoch": 0, "step": 5674, "train_loss": 3.519084930419922, "perplexity": 33.75352750150353, "lr": 0.0026291804804649314, "grad_norm": 0.16654, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:32.319551+00:00", "epoch": 0, "step": 5675, "train_loss": 3.544426918029785, "perplexity": 34.61983965028456, "lr": 0.0026291804804649314, "grad_norm": 0.162471, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:32.624277+00:00", "epoch": 0, "step": 5676, "train_loss": 3.6378509998321533, "perplexity": 38.01006525680217, "lr": 0.0026291804804649314, "grad_norm": 0.161182, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:32.929724+00:00", "epoch": 0, "step": 5677, "train_loss": 3.581512212753296, "perplexity": 35.92783031130196, "lr": 0.0026291804804649314, "grad_norm": 0.153478, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:33.235566+00:00", "epoch": 0, "step": 5678, "train_loss": 3.6211495399475098, "perplexity": 37.380513526980465, "lr": 0.0026291804804649314, "grad_norm": 0.159249, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:33.541126+00:00", "epoch": 0, "step": 5679, "train_loss": 3.625852108001709, "perplexity": 37.55671190346396, "lr": 0.0026291804804649314, "grad_norm": 0.169633, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:33.846900+00:00", "epoch": 0, "step": 5680, "train_loss": 3.5612123012542725, "perplexity": 35.20585138277878, "lr": 0.0026291804804649314, "grad_norm": 0.166653, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:34.152085+00:00", "epoch": 0, "step": 5681, "train_loss": 3.546426296234131, "perplexity": 34.68912704589215, "lr": 0.0026291804804649314, "grad_norm": 0.158845, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:34.456980+00:00", "epoch": 0, "step": 5682, "train_loss": 3.7318196296691895, "perplexity": 41.755017748550955, "lr": 0.0026291804804649314, "grad_norm": 0.20364, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:34.761988+00:00", "epoch": 0, "step": 5683, "train_loss": 3.6047964096069336, "perplexity": 36.774196222001194, "lr": 0.0026291804804649314, "grad_norm": 0.184297, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:35.068626+00:00", "epoch": 0, "step": 5684, "train_loss": 3.5663933753967285, "perplexity": 35.38872885081214, "lr": 0.0026291804804649314, "grad_norm": 0.178794, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:35.374843+00:00", "epoch": 0, "step": 5685, "train_loss": 3.6589629650115967, "perplexity": 38.82106318840868, "lr": 0.0026291804804649314, "grad_norm": 0.192562, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:35.680269+00:00", "epoch": 0, "step": 5686, "train_loss": 3.543086051940918, "perplexity": 34.57345018927344, "lr": 0.0026291804804649314, "grad_norm": 0.163584, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:35.984911+00:00", "epoch": 0, "step": 5687, "train_loss": 3.659665822982788, "perplexity": 38.848358473351404, "lr": 0.0026291804804649314, "grad_norm": 0.151509, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:36.290286+00:00", "epoch": 0, "step": 5688, "train_loss": 3.5800423622131348, "perplexity": 35.875060561835035, "lr": 0.0026291804804649314, "grad_norm": 0.144472, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:36.596283+00:00", "epoch": 0, "step": 5689, "train_loss": 3.6723902225494385, "perplexity": 39.34583886647376, "lr": 0.0026291804804649314, "grad_norm": 0.151154, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:36.901515+00:00", "epoch": 0, "step": 5690, "train_loss": 3.6038625240325928, "perplexity": 36.73986936181609, "lr": 0.0026291804804649314, "grad_norm": 0.169142, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:37.207311+00:00", "epoch": 0, "step": 5691, "train_loss": 3.5319790840148926, "perplexity": 34.19156868561997, "lr": 0.0026291804804649314, "grad_norm": 0.157596, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:37.512830+00:00", "epoch": 0, "step": 5692, "train_loss": 3.5959925651550293, "perplexity": 36.451862886830796, "lr": 0.0026291804804649314, "grad_norm": 0.151958, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:37.818200+00:00", "epoch": 0, "step": 5693, "train_loss": 3.6249642372131348, "perplexity": 37.52338119492081, "lr": 0.0026291804804649314, "grad_norm": 0.169673, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:38.123847+00:00", "epoch": 0, "step": 5694, "train_loss": 3.5417017936706543, "perplexity": 34.52562471395907, "lr": 0.0026291804804649314, "grad_norm": 0.188024, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:38.430083+00:00", "epoch": 0, "step": 5695, "train_loss": 3.5351319313049316, "perplexity": 34.29953959875314, "lr": 0.0026291804804649314, "grad_norm": 0.156001, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:38.735556+00:00", "epoch": 0, "step": 5696, "train_loss": 3.489701747894287, "perplexity": 32.776170686371564, "lr": 0.0026291804804649314, "grad_norm": 0.156174, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:39.041387+00:00", "epoch": 0, "step": 5697, "train_loss": 3.517307996749878, "perplexity": 33.693602978722055, "lr": 0.0026291804804649314, "grad_norm": 0.173838, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:39.346643+00:00", "epoch": 0, "step": 5698, "train_loss": 3.6235387325286865, "perplexity": 37.46992954609044, "lr": 0.0026291804804649314, "grad_norm": 0.166593, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:39.652026+00:00", "epoch": 0, "step": 5699, "train_loss": 3.5191550254821777, "perplexity": 33.75589354003794, "lr": 0.0026291804804649314, "grad_norm": 0.152256, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:39.958679+00:00", "epoch": 0, "step": 5700, "train_loss": 3.502985954284668, "perplexity": 33.21448095881279, "lr": 0.0026291804804649314, "grad_norm": 0.168252, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:40.264343+00:00", "epoch": 0, "step": 5701, "train_loss": 3.6560842990875244, "perplexity": 38.70947101203078, "lr": 0.0026291804804649314, "grad_norm": 0.17152, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:40.569717+00:00", "epoch": 0, "step": 5702, "train_loss": 3.602407217025757, "perplexity": 36.68644045966128, "lr": 0.0026291804804649314, "grad_norm": 0.223347, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:40.875532+00:00", "epoch": 0, "step": 5703, "train_loss": 3.6059587001800537, "perplexity": 36.816963372717574, "lr": 0.0026291804804649314, "grad_norm": 0.191298, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:41.180391+00:00", "epoch": 0, "step": 5704, "train_loss": 3.5516393184661865, "perplexity": 34.870434405446574, "lr": 0.0026291804804649314, "grad_norm": 0.160594, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:41.486685+00:00", "epoch": 0, "step": 5705, "train_loss": 3.590559244155884, "perplexity": 36.25434528891517, "lr": 0.0026291804804649314, "grad_norm": 0.172078, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:41.792491+00:00", "epoch": 0, "step": 5706, "train_loss": 3.6266071796417236, "perplexity": 37.58508062037344, "lr": 0.0026291804804649314, "grad_norm": 0.151945, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:42.097596+00:00", "epoch": 0, "step": 5707, "train_loss": 3.581571578979492, "perplexity": 35.92996327431532, "lr": 0.0026291804804649314, "grad_norm": 0.182698, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:42.403699+00:00", "epoch": 0, "step": 5708, "train_loss": 3.5581414699554443, "perplexity": 35.09790597834104, "lr": 0.0026291804804649314, "grad_norm": 0.188766, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:42.708393+00:00", "epoch": 0, "step": 5709, "train_loss": 3.4815316200256348, "perplexity": 32.509476128647194, "lr": 0.0026291804804649314, "grad_norm": 0.159038, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:43.016025+00:00", "epoch": 0, "step": 5710, "train_loss": 3.609879732131958, "perplexity": 36.96160725379497, "lr": 0.0026291804804649314, "grad_norm": 0.172065, "tokens_per_sec": 106573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:43.322570+00:00", "epoch": 0, "step": 5711, "train_loss": 3.640273094177246, "perplexity": 38.10224080479142, "lr": 0.0026291804804649314, "grad_norm": 0.193708, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:43.629252+00:00", "epoch": 0, "step": 5712, "train_loss": 3.519791841506958, "perplexity": 33.77739668002039, "lr": 0.0026291804804649314, "grad_norm": 0.165231, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:43.934832+00:00", "epoch": 0, "step": 5713, "train_loss": 3.6450884342193604, "perplexity": 38.28615850839181, "lr": 0.0026291804804649314, "grad_norm": 0.170636, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:44.239959+00:00", "epoch": 0, "step": 5714, "train_loss": 3.583789110183716, "perplexity": 36.00972749618187, "lr": 0.0026291804804649314, "grad_norm": 0.162469, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:44.545498+00:00", "epoch": 0, "step": 5715, "train_loss": 3.5000767707824707, "perplexity": 33.1179943554407, "lr": 0.0026291804804649314, "grad_norm": 0.152096, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:44.851251+00:00", "epoch": 0, "step": 5716, "train_loss": 3.673725128173828, "perplexity": 39.39839692028902, "lr": 0.0026291804804649314, "grad_norm": 0.154928, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:45.156395+00:00", "epoch": 0, "step": 5717, "train_loss": 3.623537540435791, "perplexity": 37.46988487848026, "lr": 0.0026291804804649314, "grad_norm": 0.149081, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:45.461895+00:00", "epoch": 0, "step": 5718, "train_loss": 3.6403145790100098, "perplexity": 38.10382150266639, "lr": 0.0026291804804649314, "grad_norm": 0.147704, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:45.766403+00:00", "epoch": 0, "step": 5719, "train_loss": 3.4969561100006104, "perplexity": 33.0148054218532, "lr": 0.0026291804804649314, "grad_norm": 0.156505, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:46.071382+00:00", "epoch": 0, "step": 5720, "train_loss": 3.6278281211853027, "perplexity": 37.63099783213327, "lr": 0.0026291804804649314, "grad_norm": 0.16327, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:46.376801+00:00", "epoch": 0, "step": 5721, "train_loss": 3.6390976905822754, "perplexity": 38.05748160418885, "lr": 0.0026291804804649314, "grad_norm": 0.15239, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:46.683999+00:00", "epoch": 0, "step": 5722, "train_loss": 3.5995664596557617, "perplexity": 36.582371071473155, "lr": 0.0026291804804649314, "grad_norm": 0.145003, "tokens_per_sec": 106669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:46.990291+00:00", "epoch": 0, "step": 5723, "train_loss": 3.5220046043395996, "perplexity": 33.852220801464554, "lr": 0.0026291804804649314, "grad_norm": 0.136551, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:47.295742+00:00", "epoch": 0, "step": 5724, "train_loss": 3.6843104362487793, "perplexity": 39.8176561687213, "lr": 0.0026291804804649314, "grad_norm": 0.157484, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:47.600377+00:00", "epoch": 0, "step": 5725, "train_loss": 3.6364901065826416, "perplexity": 37.95837279751814, "lr": 0.0026291804804649314, "grad_norm": 0.146809, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:47.905723+00:00", "epoch": 0, "step": 5726, "train_loss": 3.560187339782715, "perplexity": 35.16978522790507, "lr": 0.0026291804804649314, "grad_norm": 0.163514, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:48.212641+00:00", "epoch": 0, "step": 5727, "train_loss": 3.51300311088562, "perplexity": 33.548867621659454, "lr": 0.0026291804804649314, "grad_norm": 0.143589, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:48.517679+00:00", "epoch": 0, "step": 5728, "train_loss": 3.5953426361083984, "perplexity": 36.4281794594442, "lr": 0.0026291804804649314, "grad_norm": 0.146542, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:48.823569+00:00", "epoch": 0, "step": 5729, "train_loss": 3.6181278228759766, "perplexity": 37.26773067586282, "lr": 0.0026291804804649314, "grad_norm": 0.145414, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:49.128715+00:00", "epoch": 0, "step": 5730, "train_loss": 3.547283172607422, "perplexity": 34.718864077928444, "lr": 0.0026291804804649314, "grad_norm": 0.146552, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:49.433596+00:00", "epoch": 0, "step": 5731, "train_loss": 3.5901310443878174, "perplexity": 36.23882450990519, "lr": 0.0026291804804649314, "grad_norm": 0.163522, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:49.739356+00:00", "epoch": 0, "step": 5732, "train_loss": 3.586219310760498, "perplexity": 36.09734477737511, "lr": 0.0026291804804649314, "grad_norm": 0.164081, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:50.044812+00:00", "epoch": 0, "step": 5733, "train_loss": 3.6493422985076904, "perplexity": 38.44936952344876, "lr": 0.0026291804804649314, "grad_norm": 0.136809, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:50.351738+00:00", "epoch": 0, "step": 5734, "train_loss": 3.6321182250976562, "perplexity": 37.79278551800907, "lr": 0.0026291804804649314, "grad_norm": 0.160643, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:50.657526+00:00", "epoch": 0, "step": 5735, "train_loss": 3.621782064437866, "perplexity": 37.404165096558316, "lr": 0.0026291804804649314, "grad_norm": 0.175731, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:50.963196+00:00", "epoch": 0, "step": 5736, "train_loss": 3.6186206340789795, "perplexity": 37.286101157268114, "lr": 0.0026291804804649314, "grad_norm": 0.157703, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:51.269711+00:00", "epoch": 0, "step": 5737, "train_loss": 3.546077013015747, "perplexity": 34.67701283172366, "lr": 0.0026291804804649314, "grad_norm": 0.137952, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:51.575942+00:00", "epoch": 0, "step": 5738, "train_loss": 3.4328737258911133, "perplexity": 30.96550137389021, "lr": 0.0026291804804649314, "grad_norm": 0.157979, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:51.881596+00:00", "epoch": 0, "step": 5739, "train_loss": 3.559831380844116, "perplexity": 35.15726845634512, "lr": 0.0026291804804649314, "grad_norm": 0.166389, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:52.187938+00:00", "epoch": 0, "step": 5740, "train_loss": 3.549407958984375, "perplexity": 34.79271267580809, "lr": 0.0026291804804649314, "grad_norm": 0.177996, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:52.492452+00:00", "epoch": 0, "step": 5741, "train_loss": 3.6016054153442383, "perplexity": 36.65703699945896, "lr": 0.0026291804804649314, "grad_norm": 0.172709, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:52.797495+00:00", "epoch": 0, "step": 5742, "train_loss": 3.5806784629821777, "perplexity": 35.89788797494569, "lr": 0.0026291804804649314, "grad_norm": 0.159896, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:53.102890+00:00", "epoch": 0, "step": 5743, "train_loss": 3.6167187690734863, "perplexity": 37.2152554171574, "lr": 0.0026291804804649314, "grad_norm": 0.158065, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:53.408539+00:00", "epoch": 0, "step": 5744, "train_loss": 3.6322076320648193, "perplexity": 37.79616460739769, "lr": 0.0026291804804649314, "grad_norm": 0.139753, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:53.714147+00:00", "epoch": 0, "step": 5745, "train_loss": 3.6547281742095947, "perplexity": 38.657011714099085, "lr": 0.0026291804804649314, "grad_norm": 0.155633, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:54.019964+00:00", "epoch": 0, "step": 5746, "train_loss": 3.646860361099243, "perplexity": 38.354058921297955, "lr": 0.0026291804804649314, "grad_norm": 0.169211, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:54.325126+00:00", "epoch": 0, "step": 5747, "train_loss": 3.506803035736084, "perplexity": 33.34150561549931, "lr": 0.0026291804804649314, "grad_norm": 0.1473, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:54.630739+00:00", "epoch": 0, "step": 5748, "train_loss": 3.6083176136016846, "perplexity": 36.90391391584328, "lr": 0.0026291804804649314, "grad_norm": 0.176876, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:54.936372+00:00", "epoch": 0, "step": 5749, "train_loss": 3.5533955097198486, "perplexity": 34.93172736268517, "lr": 0.0026291804804649314, "grad_norm": 0.145666, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:55.243083+00:00", "epoch": 0, "step": 5750, "train_loss": 3.678298234939575, "perplexity": 39.578982599881165, "lr": 0.0026291804804649314, "grad_norm": 0.138893, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:55.548114+00:00", "epoch": 0, "step": 5751, "train_loss": 3.414093017578125, "perplexity": 30.389374296077413, "lr": 0.0026291804804649314, "grad_norm": 0.165069, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:55.852448+00:00", "epoch": 0, "step": 5752, "train_loss": 3.520275592803955, "perplexity": 33.79374049232501, "lr": 0.0026291804804649314, "grad_norm": 0.144569, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:56.157693+00:00", "epoch": 0, "step": 5753, "train_loss": 3.583865165710449, "perplexity": 36.01246633912486, "lr": 0.0026291804804649314, "grad_norm": 0.146364, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:56.463402+00:00", "epoch": 0, "step": 5754, "train_loss": 3.596928119659424, "perplexity": 36.48598154879352, "lr": 0.0026291804804649314, "grad_norm": 0.149109, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:56.768752+00:00", "epoch": 0, "step": 5755, "train_loss": 3.5380451679229736, "perplexity": 34.399607964134475, "lr": 0.0026291804804649314, "grad_norm": 0.150782, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:57.075760+00:00", "epoch": 0, "step": 5756, "train_loss": 3.6648175716400146, "perplexity": 39.04901186608843, "lr": 0.0026291804804649314, "grad_norm": 0.152265, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:57.381081+00:00", "epoch": 0, "step": 5757, "train_loss": 3.449141502380371, "perplexity": 31.47336090179298, "lr": 0.0026291804804649314, "grad_norm": 0.157293, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:57.686225+00:00", "epoch": 0, "step": 5758, "train_loss": 3.6749486923217773, "perplexity": 39.44663289013115, "lr": 0.0026291804804649314, "grad_norm": 0.16892, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:57.991787+00:00", "epoch": 0, "step": 5759, "train_loss": 3.434283494949341, "perplexity": 31.00918636523904, "lr": 0.0026291804804649314, "grad_norm": 0.167569, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:58.298180+00:00", "epoch": 0, "step": 5760, "train_loss": 3.4816396236419678, "perplexity": 32.51298745924897, "lr": 0.0026291804804649314, "grad_norm": 0.170695, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:58.603414+00:00", "epoch": 0, "step": 5761, "train_loss": 3.596620559692383, "perplexity": 36.47476164699567, "lr": 0.0026291804804649314, "grad_norm": 0.179358, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:58.907795+00:00", "epoch": 0, "step": 5762, "train_loss": 3.522364854812622, "perplexity": 33.86441827696239, "lr": 0.0026291804804649314, "grad_norm": 0.166307, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:59.213319+00:00", "epoch": 0, "step": 5763, "train_loss": 3.6220321655273438, "perplexity": 37.41352108892314, "lr": 0.0026291804804649314, "grad_norm": 0.205586, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:59.519502+00:00", "epoch": 0, "step": 5764, "train_loss": 3.5772476196289062, "perplexity": 35.77493897457021, "lr": 0.0026291804804649314, "grad_norm": 0.181674, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:27:59.825357+00:00", "epoch": 0, "step": 5765, "train_loss": 3.6898386478424072, "perplexity": 40.038386156075816, "lr": 0.0026291804804649314, "grad_norm": 0.207987, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:00.130153+00:00", "epoch": 0, "step": 5766, "train_loss": 3.5136876106262207, "perplexity": 33.57183967412896, "lr": 0.0026291804804649314, "grad_norm": 0.196249, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:00.436083+00:00", "epoch": 0, "step": 5767, "train_loss": 3.5452001094818115, "perplexity": 34.646617765343464, "lr": 0.0026291804804649314, "grad_norm": 0.162135, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:00.740728+00:00", "epoch": 0, "step": 5768, "train_loss": 3.530766248703003, "perplexity": 34.15012508096293, "lr": 0.0026291804804649314, "grad_norm": 0.165739, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:01.045897+00:00", "epoch": 0, "step": 5769, "train_loss": 3.6581625938415527, "perplexity": 38.79000435959581, "lr": 0.0026291804804649314, "grad_norm": 0.166134, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:01.352185+00:00", "epoch": 0, "step": 5770, "train_loss": 3.5466275215148926, "perplexity": 34.69610807757782, "lr": 0.0026291804804649314, "grad_norm": 0.158333, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:01.657599+00:00", "epoch": 0, "step": 5771, "train_loss": 3.581303596496582, "perplexity": 35.92033596357913, "lr": 0.0026291804804649314, "grad_norm": 0.17753, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:01.965076+00:00", "epoch": 0, "step": 5772, "train_loss": 3.5535221099853516, "perplexity": 34.93615000859193, "lr": 0.0026291804804649314, "grad_norm": 0.185064, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:02.269786+00:00", "epoch": 0, "step": 5773, "train_loss": 3.665893077850342, "perplexity": 39.09103191321793, "lr": 0.0026291804804649314, "grad_norm": 0.15723, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:02.574823+00:00", "epoch": 0, "step": 5774, "train_loss": 3.4605612754821777, "perplexity": 31.83483961609388, "lr": 0.0026291804804649314, "grad_norm": 0.154531, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:02.880441+00:00", "epoch": 0, "step": 5775, "train_loss": 3.566879987716675, "perplexity": 35.40595363281386, "lr": 0.0026291804804649314, "grad_norm": 0.14439, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:03.186443+00:00", "epoch": 0, "step": 5776, "train_loss": 3.5911502838134766, "perplexity": 36.27577937831545, "lr": 0.0026291804804649314, "grad_norm": 0.165874, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:03.492633+00:00", "epoch": 0, "step": 5777, "train_loss": 3.5344231128692627, "perplexity": 34.27523606716377, "lr": 0.0026291804804649314, "grad_norm": 0.153526, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:03.797340+00:00", "epoch": 0, "step": 5778, "train_loss": 3.554957389831543, "perplexity": 34.986329162552764, "lr": 0.0026291804804649314, "grad_norm": 0.161386, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:04.103151+00:00", "epoch": 0, "step": 5779, "train_loss": 3.4533493518829346, "perplexity": 31.606075092658894, "lr": 0.0026291804804649314, "grad_norm": 0.148963, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:04.409971+00:00", "epoch": 0, "step": 5780, "train_loss": 3.5682406425476074, "perplexity": 35.45416170450331, "lr": 0.0026291804804649314, "grad_norm": 0.166957, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:04.715605+00:00", "epoch": 0, "step": 5781, "train_loss": 3.5915706157684326, "perplexity": 36.29103045261148, "lr": 0.0026291804804649314, "grad_norm": 0.193558, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:05.022377+00:00", "epoch": 0, "step": 5782, "train_loss": 3.5299270153045654, "perplexity": 34.12147717825396, "lr": 0.0026291804804649314, "grad_norm": 0.20036, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:05.328249+00:00", "epoch": 0, "step": 5783, "train_loss": 3.636287212371826, "perplexity": 37.95067204467105, "lr": 0.0026291804804649314, "grad_norm": 0.170586, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:05.634339+00:00", "epoch": 0, "step": 5784, "train_loss": 3.6194674968719482, "perplexity": 37.3176907431676, "lr": 0.0026291804804649314, "grad_norm": 0.153189, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:05.940144+00:00", "epoch": 0, "step": 5785, "train_loss": 3.6294121742248535, "perplexity": 37.69065456587145, "lr": 0.0026291804804649314, "grad_norm": 0.175808, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:06.245758+00:00", "epoch": 0, "step": 5786, "train_loss": 3.5684256553649902, "perplexity": 35.46072178567905, "lr": 0.0026291804804649314, "grad_norm": 0.143449, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:06.550115+00:00", "epoch": 0, "step": 5787, "train_loss": 3.6251049041748047, "perplexity": 37.5286598662037, "lr": 0.0026291804804649314, "grad_norm": 0.151951, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:06.855570+00:00", "epoch": 0, "step": 5788, "train_loss": 3.5744667053222656, "perplexity": 35.67559013928138, "lr": 0.0026291804804649314, "grad_norm": 0.15884, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:07.162846+00:00", "epoch": 0, "step": 5789, "train_loss": 3.538945436477661, "perplexity": 34.43059079381923, "lr": 0.0026291804804649314, "grad_norm": 0.14607, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:07.469222+00:00", "epoch": 0, "step": 5790, "train_loss": 3.584380865097046, "perplexity": 36.03104273543218, "lr": 0.0026291804804649314, "grad_norm": 0.164679, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:07.775310+00:00", "epoch": 0, "step": 5791, "train_loss": 3.6234853267669678, "perplexity": 37.46792848939594, "lr": 0.0026291804804649314, "grad_norm": 0.143021, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:08.081077+00:00", "epoch": 0, "step": 5792, "train_loss": 3.656196355819702, "perplexity": 38.71380891189762, "lr": 0.0026291804804649314, "grad_norm": 0.154478, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:08.386139+00:00", "epoch": 0, "step": 5793, "train_loss": 3.6621053218841553, "perplexity": 38.94324469152268, "lr": 0.0026291804804649314, "grad_norm": 0.147581, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:08.691650+00:00", "epoch": 0, "step": 5794, "train_loss": 3.4943325519561768, "perplexity": 32.928302685595, "lr": 0.0026291804804649314, "grad_norm": 0.162371, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:08.997170+00:00", "epoch": 0, "step": 5795, "train_loss": 3.570589780807495, "perplexity": 35.537546334902224, "lr": 0.0026291804804649314, "grad_norm": 0.150064, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:09.302746+00:00", "epoch": 0, "step": 5796, "train_loss": 3.5644118785858154, "perplexity": 35.31867562549427, "lr": 0.0026291804804649314, "grad_norm": 0.166983, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:09.607292+00:00", "epoch": 0, "step": 5797, "train_loss": 3.5702881813049316, "perplexity": 35.52682984473058, "lr": 0.0026291804804649314, "grad_norm": 0.165364, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:09.913259+00:00", "epoch": 0, "step": 5798, "train_loss": 3.447767734527588, "perplexity": 31.430153495635064, "lr": 0.0026291804804649314, "grad_norm": 0.171281, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:10.219563+00:00", "epoch": 0, "step": 5799, "train_loss": 3.578045606613159, "perplexity": 35.80349830370532, "lr": 0.0026291804804649314, "grad_norm": 0.150086, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:10.526070+00:00", "epoch": 0, "step": 5800, "train_loss": 3.5571630001068115, "perplexity": 35.063580531537056, "lr": 0.0026291804804649314, "grad_norm": 0.182354, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:10.832549+00:00", "epoch": 0, "step": 5801, "train_loss": 3.6119418144226074, "perplexity": 37.03790376735613, "lr": 0.0026291804804649314, "grad_norm": 0.190949, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:11.138118+00:00", "epoch": 0, "step": 5802, "train_loss": 3.5879881381988525, "perplexity": 36.16125125437179, "lr": 0.0026291804804649314, "grad_norm": 0.159761, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:11.443188+00:00", "epoch": 0, "step": 5803, "train_loss": 3.5937442779541016, "perplexity": 36.370000689400776, "lr": 0.0026291804804649314, "grad_norm": 0.180482, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:11.748203+00:00", "epoch": 0, "step": 5804, "train_loss": 3.623035192489624, "perplexity": 37.45106668580439, "lr": 0.0026291804804649314, "grad_norm": 0.169817, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:12.054520+00:00", "epoch": 0, "step": 5805, "train_loss": 3.6314587593078613, "perplexity": 37.76787068500184, "lr": 0.0026291804804649314, "grad_norm": 0.160055, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:12.360042+00:00", "epoch": 0, "step": 5806, "train_loss": 3.6389834880828857, "perplexity": 38.05313559283658, "lr": 0.0026291804804649314, "grad_norm": 0.158883, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:12.666047+00:00", "epoch": 0, "step": 5807, "train_loss": 3.6685006618499756, "perplexity": 39.19309807777964, "lr": 0.0026291804804649314, "grad_norm": 0.169086, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:12.971391+00:00", "epoch": 0, "step": 5808, "train_loss": 3.5635035037994385, "perplexity": 35.28660759816665, "lr": 0.0026291804804649314, "grad_norm": 0.154451, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:13.277480+00:00", "epoch": 0, "step": 5809, "train_loss": 3.6132140159606934, "perplexity": 37.085053431074485, "lr": 0.0026291804804649314, "grad_norm": 0.178077, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:13.583224+00:00", "epoch": 0, "step": 5810, "train_loss": 3.6085567474365234, "perplexity": 36.91273994555765, "lr": 0.0026291804804649314, "grad_norm": 0.144247, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:13.889022+00:00", "epoch": 0, "step": 5811, "train_loss": 3.6821322441101074, "perplexity": 39.73102005238192, "lr": 0.0026291804804649314, "grad_norm": 0.162964, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:14.195415+00:00", "epoch": 0, "step": 5812, "train_loss": 3.5185813903808594, "perplexity": 33.736535527375615, "lr": 0.0026291804804649314, "grad_norm": 0.160459, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:14.500028+00:00", "epoch": 0, "step": 5813, "train_loss": 3.548996686935425, "perplexity": 34.77840634767547, "lr": 0.0026291804804649314, "grad_norm": 0.157725, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:14.805579+00:00", "epoch": 0, "step": 5814, "train_loss": 3.4772419929504395, "perplexity": 32.37032127424204, "lr": 0.0026291804804649314, "grad_norm": 0.157225, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:15.111118+00:00", "epoch": 0, "step": 5815, "train_loss": 3.629507541656494, "perplexity": 37.69424919819695, "lr": 0.0026291804804649314, "grad_norm": 0.151036, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:15.415839+00:00", "epoch": 0, "step": 5816, "train_loss": 3.501399278640747, "perplexity": 33.16182213813163, "lr": 0.0026291804804649314, "grad_norm": 0.138378, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:15.722356+00:00", "epoch": 0, "step": 5817, "train_loss": 3.5204079151153564, "perplexity": 33.798212454041476, "lr": 0.0026291804804649314, "grad_norm": 0.153301, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:16.027930+00:00", "epoch": 0, "step": 5818, "train_loss": 3.511272430419922, "perplexity": 33.49085546655226, "lr": 0.0026291804804649314, "grad_norm": 0.141282, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:16.333794+00:00", "epoch": 0, "step": 5819, "train_loss": 3.6368942260742188, "perplexity": 37.97371561579132, "lr": 0.0026291804804649314, "grad_norm": 0.151238, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:16.639619+00:00", "epoch": 0, "step": 5820, "train_loss": 3.639230489730835, "perplexity": 38.06253594094059, "lr": 0.0026291804804649314, "grad_norm": 0.164067, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:16.946234+00:00", "epoch": 0, "step": 5821, "train_loss": 3.709583282470703, "perplexity": 40.83678557638252, "lr": 0.0026291804804649314, "grad_norm": 0.134086, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:17.252755+00:00", "epoch": 0, "step": 5822, "train_loss": 3.478965997695923, "perplexity": 32.42617599479905, "lr": 0.0026291804804649314, "grad_norm": 0.197612, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:17.558421+00:00", "epoch": 0, "step": 5823, "train_loss": 3.589711904525757, "perplexity": 36.22363855674035, "lr": 0.0026291804804649314, "grad_norm": 0.195795, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:17.864039+00:00", "epoch": 0, "step": 5824, "train_loss": 3.6969246864318848, "perplexity": 40.323107286531226, "lr": 0.0026291804804649314, "grad_norm": 0.186409, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:18.168259+00:00", "epoch": 0, "step": 5825, "train_loss": 3.6244847774505615, "perplexity": 37.50539455576158, "lr": 0.0026291804804649314, "grad_norm": 0.176008, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:18.474161+00:00", "epoch": 0, "step": 5826, "train_loss": 3.590954303741455, "perplexity": 36.26867074505823, "lr": 0.0026291804804649314, "grad_norm": 0.161693, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:18.781132+00:00", "epoch": 0, "step": 5827, "train_loss": 3.6114416122436523, "perplexity": 37.019381959898176, "lr": 0.0026291804804649314, "grad_norm": 0.172076, "tokens_per_sec": 106747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:19.087680+00:00", "epoch": 0, "step": 5828, "train_loss": 3.5383799076080322, "perplexity": 34.41112480553492, "lr": 0.0026291804804649314, "grad_norm": 0.17615, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:19.392293+00:00", "epoch": 0, "step": 5829, "train_loss": 3.6032934188842773, "perplexity": 36.71896646155209, "lr": 0.0026291804804649314, "grad_norm": 0.154291, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:19.697961+00:00", "epoch": 0, "step": 5830, "train_loss": 3.7032620906829834, "perplexity": 40.57946257361684, "lr": 0.0026291804804649314, "grad_norm": 0.15376, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:20.003658+00:00", "epoch": 0, "step": 5831, "train_loss": 3.6398251056671143, "perplexity": 38.08517526156965, "lr": 0.0026291804804649314, "grad_norm": 0.168463, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:20.309927+00:00", "epoch": 0, "step": 5832, "train_loss": 3.577840805053711, "perplexity": 35.79616644223297, "lr": 0.0026291804804649314, "grad_norm": 0.163216, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:20.615093+00:00", "epoch": 0, "step": 5833, "train_loss": 3.587531089782715, "perplexity": 36.144727588105816, "lr": 0.0026291804804649314, "grad_norm": 0.158399, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:20.920595+00:00", "epoch": 0, "step": 5834, "train_loss": 3.5474307537078857, "perplexity": 34.7239883042061, "lr": 0.0026291804804649314, "grad_norm": 0.160864, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:21.226296+00:00", "epoch": 0, "step": 5835, "train_loss": 3.6080095767974854, "perplexity": 36.892547902803244, "lr": 0.0026291804804649314, "grad_norm": 0.173353, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:21.530814+00:00", "epoch": 0, "step": 5836, "train_loss": 3.607025146484375, "perplexity": 36.85624763079547, "lr": 0.0026291804804649314, "grad_norm": 0.15566, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:21.836368+00:00", "epoch": 0, "step": 5837, "train_loss": 3.6438705921173096, "perplexity": 38.23956039296721, "lr": 0.0026291804804649314, "grad_norm": 0.144775, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:22.142850+00:00", "epoch": 0, "step": 5838, "train_loss": 3.56020450592041, "perplexity": 35.17038896246289, "lr": 0.0026291804804649314, "grad_norm": 0.159345, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:22.449526+00:00", "epoch": 0, "step": 5839, "train_loss": 3.5771572589874268, "perplexity": 35.77170647418318, "lr": 0.0026291804804649314, "grad_norm": 0.143872, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:22.754962+00:00", "epoch": 0, "step": 5840, "train_loss": 3.5726397037506104, "perplexity": 35.6104702851789, "lr": 0.0026291804804649314, "grad_norm": 0.161664, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:23.060296+00:00", "epoch": 0, "step": 5841, "train_loss": 3.6054041385650635, "perplexity": 36.7965517583226, "lr": 0.0026291804804649314, "grad_norm": 0.179257, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:23.365983+00:00", "epoch": 0, "step": 5842, "train_loss": 3.5158865451812744, "perplexity": 33.64574317716953, "lr": 0.0026291804804649314, "grad_norm": 0.207934, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:23.671753+00:00", "epoch": 0, "step": 5843, "train_loss": 3.6104135513305664, "perplexity": 36.98134333663976, "lr": 0.0026291804804649314, "grad_norm": 0.158602, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:23.978190+00:00", "epoch": 0, "step": 5844, "train_loss": 3.563983201980591, "perplexity": 35.30353858020651, "lr": 0.0026291804804649314, "grad_norm": 0.194433, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:24.283852+00:00", "epoch": 0, "step": 5845, "train_loss": 3.574063301086426, "perplexity": 35.66120135754599, "lr": 0.0026291804804649314, "grad_norm": 0.190378, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:24.588290+00:00", "epoch": 0, "step": 5846, "train_loss": 3.563037395477295, "perplexity": 35.27016404924014, "lr": 0.0026291804804649314, "grad_norm": 0.166631, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:24.893344+00:00", "epoch": 0, "step": 5847, "train_loss": 3.5630006790161133, "perplexity": 35.26886907740449, "lr": 0.0026291804804649314, "grad_norm": 0.208474, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:25.199185+00:00", "epoch": 0, "step": 5848, "train_loss": 3.6051440238952637, "perplexity": 36.78698168012504, "lr": 0.0026291804804649314, "grad_norm": 0.175054, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:25.505070+00:00", "epoch": 0, "step": 5849, "train_loss": 3.569135904312134, "perplexity": 35.48591667225397, "lr": 0.0026291804804649314, "grad_norm": 0.161405, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:25.811430+00:00", "epoch": 0, "step": 5850, "train_loss": 3.609375238418579, "perplexity": 36.9429650581297, "lr": 0.0026291804804649314, "grad_norm": 0.21612, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:26.117210+00:00", "epoch": 0, "step": 5851, "train_loss": 3.57523512840271, "perplexity": 35.703014621606016, "lr": 0.0026291804804649314, "grad_norm": 0.172273, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:26.422773+00:00", "epoch": 0, "step": 5852, "train_loss": 3.6846024990081787, "perplexity": 39.82928712165618, "lr": 0.0026291804804649314, "grad_norm": 0.180176, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:26.728674+00:00", "epoch": 0, "step": 5853, "train_loss": 3.6540896892547607, "perplexity": 38.63233767156148, "lr": 0.0026291804804649314, "grad_norm": 0.156182, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:27.033781+00:00", "epoch": 0, "step": 5854, "train_loss": 3.5918021202087402, "perplexity": 36.29943295987594, "lr": 0.0026291804804649314, "grad_norm": 0.149314, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:27.339617+00:00", "epoch": 0, "step": 5855, "train_loss": 3.541304349899292, "perplexity": 34.511905445963514, "lr": 0.0026291804804649314, "grad_norm": 0.1568, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:27.644843+00:00", "epoch": 0, "step": 5856, "train_loss": 3.625878095626831, "perplexity": 37.55768792589593, "lr": 0.0026291804804649314, "grad_norm": 0.173299, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:27.950033+00:00", "epoch": 0, "step": 5857, "train_loss": 3.6123766899108887, "perplexity": 37.05401414659249, "lr": 0.0026291804804649314, "grad_norm": 0.154178, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:28.255645+00:00", "epoch": 0, "step": 5858, "train_loss": 3.5605528354644775, "perplexity": 35.182641981935305, "lr": 0.0026291804804649314, "grad_norm": 0.156432, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:28.563167+00:00", "epoch": 0, "step": 5859, "train_loss": 3.6805450916290283, "perplexity": 39.66801088113817, "lr": 0.0026291804804649314, "grad_norm": 0.158941, "tokens_per_sec": 106554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:28.869334+00:00", "epoch": 0, "step": 5860, "train_loss": 3.5476536750793457, "perplexity": 34.731729886151406, "lr": 0.0026291804804649314, "grad_norm": 0.159706, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:29.175265+00:00", "epoch": 0, "step": 5861, "train_loss": 3.5417160987854004, "perplexity": 34.5261186105149, "lr": 0.0026291804804649314, "grad_norm": 0.170147, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:29.480141+00:00", "epoch": 0, "step": 5862, "train_loss": 3.5484395027160645, "perplexity": 34.759033766033625, "lr": 0.0026291804804649314, "grad_norm": 0.150322, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:29.785798+00:00", "epoch": 0, "step": 5863, "train_loss": 3.547075033187866, "perplexity": 34.711638465705086, "lr": 0.0026291804804649314, "grad_norm": 0.156962, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:30.091801+00:00", "epoch": 0, "step": 5864, "train_loss": 3.5286765098571777, "perplexity": 34.07883475301905, "lr": 0.0026291804804649314, "grad_norm": 0.157832, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:30.397191+00:00", "epoch": 0, "step": 5865, "train_loss": 3.6745223999023438, "perplexity": 39.42982067327281, "lr": 0.0026291804804649314, "grad_norm": 0.164594, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:30.703530+00:00", "epoch": 0, "step": 5866, "train_loss": 3.5968263149261475, "perplexity": 36.482267292241275, "lr": 0.0026291804804649314, "grad_norm": 0.149249, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:31.008757+00:00", "epoch": 0, "step": 5867, "train_loss": 3.5289878845214844, "perplexity": 34.089447690964484, "lr": 0.0026291804804649314, "grad_norm": 0.154966, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:31.314882+00:00", "epoch": 0, "step": 5868, "train_loss": 3.589599132537842, "perplexity": 36.21955377533958, "lr": 0.0026291804804649314, "grad_norm": 0.140644, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:31.622060+00:00", "epoch": 0, "step": 5869, "train_loss": 3.5715293884277344, "perplexity": 35.570953376540736, "lr": 0.0026291804804649314, "grad_norm": 0.144833, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:31.929514+00:00", "epoch": 0, "step": 5870, "train_loss": 3.686603307723999, "perplexity": 39.90905768267929, "lr": 0.0026291804804649314, "grad_norm": 0.1581, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:32.236862+00:00", "epoch": 0, "step": 5871, "train_loss": 3.4994828701019287, "perplexity": 33.098331395555405, "lr": 0.0026291804804649314, "grad_norm": 0.16622, "tokens_per_sec": 106560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:32.542385+00:00", "epoch": 0, "step": 5872, "train_loss": 3.6504712104797363, "perplexity": 38.49279998699405, "lr": 0.0026291804804649314, "grad_norm": 0.137797, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:32.848612+00:00", "epoch": 0, "step": 5873, "train_loss": 3.5892179012298584, "perplexity": 36.205748379171055, "lr": 0.0026291804804649314, "grad_norm": 0.163825, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:33.154302+00:00", "epoch": 0, "step": 5874, "train_loss": 3.5403692722320557, "perplexity": 34.47964921731706, "lr": 0.0026291804804649314, "grad_norm": 0.161695, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:33.459920+00:00", "epoch": 0, "step": 5875, "train_loss": 3.626818895339966, "perplexity": 37.59303881436804, "lr": 0.0026291804804649314, "grad_norm": 0.175933, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:33.765662+00:00", "epoch": 0, "step": 5876, "train_loss": 3.499807834625244, "perplexity": 33.10908892685324, "lr": 0.0026291804804649314, "grad_norm": 0.137537, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:34.070151+00:00", "epoch": 0, "step": 5877, "train_loss": 3.5739059448242188, "perplexity": 35.655590285674776, "lr": 0.0026291804804649314, "grad_norm": 0.171222, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:34.375584+00:00", "epoch": 0, "step": 5878, "train_loss": 3.5880978107452393, "perplexity": 36.165217368360324, "lr": 0.0026291804804649314, "grad_norm": 0.15119, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:34.682271+00:00", "epoch": 0, "step": 5879, "train_loss": 3.6183359622955322, "perplexity": 37.27548836700652, "lr": 0.0026291804804649314, "grad_norm": 0.159157, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:34.989336+00:00", "epoch": 0, "step": 5880, "train_loss": 3.55334210395813, "perplexity": 34.929861856992034, "lr": 0.0026291804804649314, "grad_norm": 0.132843, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:35.295677+00:00", "epoch": 0, "step": 5881, "train_loss": 3.540294647216797, "perplexity": 34.477076268972446, "lr": 0.0026291804804649314, "grad_norm": 0.149539, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:35.601253+00:00", "epoch": 0, "step": 5882, "train_loss": 3.641231060028076, "perplexity": 38.13875893909259, "lr": 0.0026291804804649314, "grad_norm": 0.147936, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:35.907656+00:00", "epoch": 0, "step": 5883, "train_loss": 3.408696174621582, "perplexity": 30.225809379535715, "lr": 0.0026291804804649314, "grad_norm": 0.156886, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:36.213633+00:00", "epoch": 0, "step": 5884, "train_loss": 3.6325414180755615, "perplexity": 37.80878254413158, "lr": 0.0026291804804649314, "grad_norm": 0.180004, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:36.519865+00:00", "epoch": 0, "step": 5885, "train_loss": 3.486370086669922, "perplexity": 32.667153294761576, "lr": 0.0026291804804649314, "grad_norm": 0.176044, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:36.825114+00:00", "epoch": 0, "step": 5886, "train_loss": 3.6053099632263184, "perplexity": 36.793086593765196, "lr": 0.0026291804804649314, "grad_norm": 0.169897, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:37.131758+00:00", "epoch": 0, "step": 5887, "train_loss": 3.6325225830078125, "perplexity": 37.80807041985733, "lr": 0.0026291804804649314, "grad_norm": 0.183821, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:37.437662+00:00", "epoch": 0, "step": 5888, "train_loss": 3.6346585750579834, "perplexity": 37.88891446809129, "lr": 0.0026291804804649314, "grad_norm": 0.205788, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:37.744107+00:00", "epoch": 0, "step": 5889, "train_loss": 3.6204473972320557, "perplexity": 37.35427628393003, "lr": 0.0026291804804649314, "grad_norm": 0.158021, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:38.051038+00:00", "epoch": 0, "step": 5890, "train_loss": 3.565164804458618, "perplexity": 35.34527798371099, "lr": 0.0026291804804649314, "grad_norm": 0.219754, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:38.356949+00:00", "epoch": 0, "step": 5891, "train_loss": 3.5218071937561035, "perplexity": 33.8455386743865, "lr": 0.0026291804804649314, "grad_norm": 0.155468, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:38.662473+00:00", "epoch": 0, "step": 5892, "train_loss": 3.525582790374756, "perplexity": 33.973567315926324, "lr": 0.0026291804804649314, "grad_norm": 0.163235, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:38.968043+00:00", "epoch": 0, "step": 5893, "train_loss": 3.6626992225646973, "perplexity": 38.966379980399374, "lr": 0.0026291804804649314, "grad_norm": 0.150281, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:39.273707+00:00", "epoch": 0, "step": 5894, "train_loss": 3.66436505317688, "perplexity": 39.031345464739715, "lr": 0.0026291804804649314, "grad_norm": 0.159773, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:39.579824+00:00", "epoch": 0, "step": 5895, "train_loss": 3.5122106075286865, "perplexity": 33.52229056404266, "lr": 0.0026291804804649314, "grad_norm": 0.156098, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:39.884769+00:00", "epoch": 0, "step": 5896, "train_loss": 3.6262059211730957, "perplexity": 37.57000231383186, "lr": 0.0026291804804649314, "grad_norm": 0.152464, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:40.189921+00:00", "epoch": 0, "step": 5897, "train_loss": 3.5586049556732178, "perplexity": 35.114177126916964, "lr": 0.0026291804804649314, "grad_norm": 0.147976, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:40.495302+00:00", "epoch": 0, "step": 5898, "train_loss": 3.588785171508789, "perplexity": 36.19008446513883, "lr": 0.0026291804804649314, "grad_norm": 0.146819, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:40.801795+00:00", "epoch": 0, "step": 5899, "train_loss": 3.5619280338287354, "perplexity": 35.231058377082526, "lr": 0.0026291804804649314, "grad_norm": 0.148916, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:41.109250+00:00", "epoch": 0, "step": 5900, "train_loss": 3.637124538421631, "perplexity": 37.98246243858674, "lr": 0.0026291804804649314, "grad_norm": 0.14519, "tokens_per_sec": 106649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:41.413896+00:00", "epoch": 0, "step": 5901, "train_loss": 3.5104877948760986, "perplexity": 33.46458765763607, "lr": 0.0026291804804649314, "grad_norm": 0.167878, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:41.718855+00:00", "epoch": 0, "step": 5902, "train_loss": 3.6946659088134766, "perplexity": 40.23212914266675, "lr": 0.0026291804804649314, "grad_norm": 0.171694, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:42.024434+00:00", "epoch": 0, "step": 5903, "train_loss": 3.622166633605957, "perplexity": 37.418552351482646, "lr": 0.0026291804804649314, "grad_norm": 0.147155, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:42.328981+00:00", "epoch": 0, "step": 5904, "train_loss": 3.54608154296875, "perplexity": 34.67716991731786, "lr": 0.0026291804804649314, "grad_norm": 0.16917, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:42.634847+00:00", "epoch": 0, "step": 5905, "train_loss": 3.6433751583099365, "perplexity": 38.220619914233616, "lr": 0.0026291804804649314, "grad_norm": 0.177787, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:42.940912+00:00", "epoch": 0, "step": 5906, "train_loss": 3.5533714294433594, "perplexity": 34.9308862071597, "lr": 0.0026291804804649314, "grad_norm": 0.172344, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:43.246340+00:00", "epoch": 0, "step": 5907, "train_loss": 3.6664626598358154, "perplexity": 39.113303803020756, "lr": 0.0026291804804649314, "grad_norm": 0.169943, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:43.550779+00:00", "epoch": 0, "step": 5908, "train_loss": 3.5847644805908203, "perplexity": 36.04486745320077, "lr": 0.0026291804804649314, "grad_norm": 0.154657, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:43.857230+00:00", "epoch": 0, "step": 5909, "train_loss": 3.6370394229888916, "perplexity": 37.97922968244044, "lr": 0.0026291804804649314, "grad_norm": 0.153692, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:44.162956+00:00", "epoch": 0, "step": 5910, "train_loss": 3.600399971008301, "perplexity": 36.61287560423514, "lr": 0.0026291804804649314, "grad_norm": 0.149772, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:44.467944+00:00", "epoch": 0, "step": 5911, "train_loss": 3.611527681350708, "perplexity": 37.02256832216892, "lr": 0.0026291804804649314, "grad_norm": 0.160062, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:44.772657+00:00", "epoch": 0, "step": 5912, "train_loss": 3.574030876159668, "perplexity": 35.66004506445034, "lr": 0.0026291804804649314, "grad_norm": 0.138308, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:45.078214+00:00", "epoch": 0, "step": 5913, "train_loss": 3.598515272140503, "perplexity": 36.54393634431611, "lr": 0.0026291804804649314, "grad_norm": 0.157693, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:45.382908+00:00", "epoch": 0, "step": 5914, "train_loss": 3.61653995513916, "perplexity": 37.208601405852, "lr": 0.0026291804804649314, "grad_norm": 0.149515, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:45.689455+00:00", "epoch": 0, "step": 5915, "train_loss": 3.613138198852539, "perplexity": 37.08224185615168, "lr": 0.0026291804804649314, "grad_norm": 0.152286, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:45.995275+00:00", "epoch": 0, "step": 5916, "train_loss": 3.567427635192871, "perplexity": 35.42534892436968, "lr": 0.0026291804804649314, "grad_norm": 0.181313, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:46.302002+00:00", "epoch": 0, "step": 5917, "train_loss": 3.6328415870666504, "perplexity": 37.820133271725155, "lr": 0.0026291804804649314, "grad_norm": 0.195915, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:46.607564+00:00", "epoch": 0, "step": 5918, "train_loss": 3.5850470066070557, "perplexity": 36.05505250471132, "lr": 0.0026291804804649314, "grad_norm": 0.183078, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:46.912816+00:00", "epoch": 0, "step": 5919, "train_loss": 3.559889316558838, "perplexity": 35.15930537682546, "lr": 0.0026291804804649314, "grad_norm": 0.134981, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:47.217780+00:00", "epoch": 0, "step": 5920, "train_loss": 3.6845321655273438, "perplexity": 39.82648588776516, "lr": 0.0026291804804649314, "grad_norm": 0.168741, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:47.523273+00:00", "epoch": 0, "step": 5921, "train_loss": 3.4319190979003906, "perplexity": 30.93595494469994, "lr": 0.0026291804804649314, "grad_norm": 0.163354, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:47.829474+00:00", "epoch": 0, "step": 5922, "train_loss": 3.5950398445129395, "perplexity": 36.41715098261532, "lr": 0.0026291804804649314, "grad_norm": 0.153798, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:48.135150+00:00", "epoch": 0, "step": 5923, "train_loss": 3.5640039443969727, "perplexity": 35.30427086849817, "lr": 0.0026291804804649314, "grad_norm": 0.159155, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:48.439527+00:00", "epoch": 0, "step": 5924, "train_loss": 3.57039213180542, "perplexity": 35.530523068426604, "lr": 0.0026291804804649314, "grad_norm": 0.163297, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:48.744543+00:00", "epoch": 0, "step": 5925, "train_loss": 3.539506673812866, "perplexity": 34.449919950460696, "lr": 0.0026291804804649314, "grad_norm": 0.181862, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:49.050142+00:00", "epoch": 0, "step": 5926, "train_loss": 3.5924644470214844, "perplexity": 36.32348301123042, "lr": 0.0026291804804649314, "grad_norm": 0.144535, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:49.357584+00:00", "epoch": 0, "step": 5927, "train_loss": 3.5238423347473145, "perplexity": 33.91448925579137, "lr": 0.0026291804804649314, "grad_norm": 0.161451, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:49.664000+00:00", "epoch": 0, "step": 5928, "train_loss": 3.544710874557495, "perplexity": 34.629671575595154, "lr": 0.0026291804804649314, "grad_norm": 0.14775, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:49.969466+00:00", "epoch": 0, "step": 5929, "train_loss": 3.6314167976379395, "perplexity": 37.76628591532853, "lr": 0.0026291804804649314, "grad_norm": 0.150842, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:50.275546+00:00", "epoch": 0, "step": 5930, "train_loss": 3.5723958015441895, "perplexity": 35.60178587202147, "lr": 0.0026291804804649314, "grad_norm": 0.171053, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:50.582164+00:00", "epoch": 0, "step": 5931, "train_loss": 3.518322229385376, "perplexity": 33.72779346609376, "lr": 0.0026291804804649314, "grad_norm": 0.179531, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:50.888791+00:00", "epoch": 0, "step": 5932, "train_loss": 3.561882972717285, "perplexity": 35.229470862202355, "lr": 0.0026291804804649314, "grad_norm": 0.14857, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:51.195749+00:00", "epoch": 0, "step": 5933, "train_loss": 3.567857503890991, "perplexity": 35.440580446535, "lr": 0.0026291804804649314, "grad_norm": 0.166642, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:51.502189+00:00", "epoch": 0, "step": 5934, "train_loss": 3.5094079971313477, "perplexity": 33.42847217357257, "lr": 0.0026291804804649314, "grad_norm": 0.156296, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:51.807933+00:00", "epoch": 0, "step": 5935, "train_loss": 3.5982723236083984, "perplexity": 36.53505912702106, "lr": 0.0026291804804649314, "grad_norm": 0.146784, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:52.113662+00:00", "epoch": 0, "step": 5936, "train_loss": 3.661292791366577, "perplexity": 38.911614968555064, "lr": 0.0026291804804649314, "grad_norm": 0.201798, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:52.420342+00:00", "epoch": 0, "step": 5937, "train_loss": 3.603804111480713, "perplexity": 36.73772335496841, "lr": 0.0026291804804649314, "grad_norm": 0.163757, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:52.726371+00:00", "epoch": 0, "step": 5938, "train_loss": 3.476402997970581, "perplexity": 32.34317412694912, "lr": 0.0026291804804649314, "grad_norm": 0.163524, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:53.031981+00:00", "epoch": 0, "step": 5939, "train_loss": 3.6737937927246094, "perplexity": 39.40110228639536, "lr": 0.0026291804804649314, "grad_norm": 0.150572, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:53.337971+00:00", "epoch": 0, "step": 5940, "train_loss": 3.505903720855713, "perplexity": 33.31153458208344, "lr": 0.0026291804804649314, "grad_norm": 0.151167, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:53.643146+00:00", "epoch": 0, "step": 5941, "train_loss": 3.606588363647461, "perplexity": 36.840152969589056, "lr": 0.0026291804804649314, "grad_norm": 0.15932, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:53.949021+00:00", "epoch": 0, "step": 5942, "train_loss": 3.459897756576538, "perplexity": 31.8137236043618, "lr": 0.0026291804804649314, "grad_norm": 0.16909, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:54.254253+00:00", "epoch": 0, "step": 5943, "train_loss": 3.589966058731079, "perplexity": 36.232846116832114, "lr": 0.0026291804804649314, "grad_norm": 0.157309, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:54.559237+00:00", "epoch": 0, "step": 5944, "train_loss": 3.6829044818878174, "perplexity": 39.76171369688389, "lr": 0.0026291804804649314, "grad_norm": 0.156083, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:54.865766+00:00", "epoch": 0, "step": 5945, "train_loss": 3.7185165882110596, "perplexity": 41.20322739885227, "lr": 0.0026291804804649314, "grad_norm": 0.164267, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:55.171440+00:00", "epoch": 0, "step": 5946, "train_loss": 3.5206892490386963, "perplexity": 33.80772237542207, "lr": 0.0026291804804649314, "grad_norm": 0.1832, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:55.476624+00:00", "epoch": 0, "step": 5947, "train_loss": 3.546592950820923, "perplexity": 34.69490862977651, "lr": 0.0026291804804649314, "grad_norm": 0.169242, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:55.782054+00:00", "epoch": 0, "step": 5948, "train_loss": 3.5442323684692383, "perplexity": 34.613105030824265, "lr": 0.0026291804804649314, "grad_norm": 0.161332, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:56.089533+00:00", "epoch": 0, "step": 5949, "train_loss": 3.545767307281494, "perplexity": 34.66627482489988, "lr": 0.0026291804804649314, "grad_norm": 0.153651, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:56.397986+00:00", "epoch": 0, "step": 5950, "train_loss": 3.6508431434631348, "perplexity": 38.50711939169695, "lr": 0.0026291804804649314, "grad_norm": 0.164654, "tokens_per_sec": 106321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:56.703037+00:00", "epoch": 0, "step": 5951, "train_loss": 3.5606532096862793, "perplexity": 35.18617358948338, "lr": 0.0026291804804649314, "grad_norm": 0.174332, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:57.008721+00:00", "epoch": 0, "step": 5952, "train_loss": 3.4559357166290283, "perplexity": 31.687925733219704, "lr": 0.0026291804804649314, "grad_norm": 0.174903, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:57.315265+00:00", "epoch": 0, "step": 5953, "train_loss": 3.566518545150757, "perplexity": 35.39315872653522, "lr": 0.0026291804804649314, "grad_norm": 0.154157, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:57.621601+00:00", "epoch": 0, "step": 5954, "train_loss": 3.4929556846618652, "perplexity": 32.88299598038789, "lr": 0.0026291804804649314, "grad_norm": 0.158214, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:57.928826+00:00", "epoch": 0, "step": 5955, "train_loss": 3.560424566268921, "perplexity": 35.17812942216811, "lr": 0.0026291804804649314, "grad_norm": 0.169967, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:58.234941+00:00", "epoch": 0, "step": 5956, "train_loss": 3.509833812713623, "perplexity": 33.44270956895301, "lr": 0.0026291804804649314, "grad_norm": 0.158251, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:58.541230+00:00", "epoch": 0, "step": 5957, "train_loss": 3.5306589603424072, "perplexity": 34.146461366569326, "lr": 0.0026291804804649314, "grad_norm": 0.175386, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:58.846493+00:00", "epoch": 0, "step": 5958, "train_loss": 3.620816469192505, "perplexity": 37.3680652443121, "lr": 0.0026291804804649314, "grad_norm": 0.164642, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:59.151356+00:00", "epoch": 0, "step": 5959, "train_loss": 3.6236796379089355, "perplexity": 37.47520963274861, "lr": 0.0026291804804649314, "grad_norm": 0.158715, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:59.457424+00:00", "epoch": 0, "step": 5960, "train_loss": 3.57878041267395, "perplexity": 35.82981659949384, "lr": 0.0026291804804649314, "grad_norm": 0.164792, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:28:59.763549+00:00", "epoch": 0, "step": 5961, "train_loss": 3.4663264751434326, "perplexity": 32.01890389650839, "lr": 0.0026291804804649314, "grad_norm": 0.146745, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:00.069802+00:00", "epoch": 0, "step": 5962, "train_loss": 3.588665723800659, "perplexity": 36.185761900657795, "lr": 0.0026291804804649314, "grad_norm": 0.154321, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:00.375288+00:00", "epoch": 0, "step": 5963, "train_loss": 3.595489501953125, "perplexity": 36.433529907681795, "lr": 0.0026291804804649314, "grad_norm": 0.152362, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:00.680257+00:00", "epoch": 0, "step": 5964, "train_loss": 3.5545098781585693, "perplexity": 34.97067587463379, "lr": 0.0026291804804649314, "grad_norm": 0.145132, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:00.986993+00:00", "epoch": 0, "step": 5965, "train_loss": 3.585183620452881, "perplexity": 36.05997846056464, "lr": 0.0026291804804649314, "grad_norm": 0.156386, "tokens_per_sec": 106828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:01.293757+00:00", "epoch": 0, "step": 5966, "train_loss": 3.548722982406616, "perplexity": 34.76888864293223, "lr": 0.0026291804804649314, "grad_norm": 0.176368, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:01.599188+00:00", "epoch": 0, "step": 5967, "train_loss": 3.6597697734832764, "perplexity": 38.85239698955711, "lr": 0.0026291804804649314, "grad_norm": 0.155326, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:01.904000+00:00", "epoch": 0, "step": 5968, "train_loss": 3.5172204971313477, "perplexity": 33.69065493029294, "lr": 0.0026291804804649314, "grad_norm": 0.148747, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:02.209734+00:00", "epoch": 0, "step": 5969, "train_loss": 3.615705966949463, "perplexity": 37.17758280809434, "lr": 0.0026291804804649314, "grad_norm": 0.174921, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:02.514785+00:00", "epoch": 0, "step": 5970, "train_loss": 3.5033493041992188, "perplexity": 33.22655163043698, "lr": 0.0026291804804649314, "grad_norm": 0.147977, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:02.820536+00:00", "epoch": 0, "step": 5971, "train_loss": 3.578582763671875, "perplexity": 35.822735571800486, "lr": 0.0026291804804649314, "grad_norm": 0.155241, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:03.126420+00:00", "epoch": 0, "step": 5972, "train_loss": 3.545409679412842, "perplexity": 34.653879415524166, "lr": 0.0026291804804649314, "grad_norm": 0.179089, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:03.431542+00:00", "epoch": 0, "step": 5973, "train_loss": 3.625822067260742, "perplexity": 37.55558368895639, "lr": 0.0026291804804649314, "grad_norm": 0.161741, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:03.736063+00:00", "epoch": 0, "step": 5974, "train_loss": 3.56406831741333, "perplexity": 35.306543584054275, "lr": 0.0026291804804649314, "grad_norm": 0.151717, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:04.041694+00:00", "epoch": 0, "step": 5975, "train_loss": 3.4707581996917725, "perplexity": 32.161117752477026, "lr": 0.0026291804804649314, "grad_norm": 0.165212, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:04.347790+00:00", "epoch": 0, "step": 5976, "train_loss": 3.638700246810913, "perplexity": 38.04235890058278, "lr": 0.0026291804804649314, "grad_norm": 0.159694, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:04.655464+00:00", "epoch": 0, "step": 5977, "train_loss": 3.5980029106140137, "perplexity": 36.52521743314126, "lr": 0.0026291804804649314, "grad_norm": 0.157511, "tokens_per_sec": 106503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:04.960919+00:00", "epoch": 0, "step": 5978, "train_loss": 3.608396291732788, "perplexity": 36.90681756104578, "lr": 0.0026291804804649314, "grad_norm": 0.170349, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:05.267530+00:00", "epoch": 0, "step": 5979, "train_loss": 3.546571969985962, "perplexity": 34.694180709260785, "lr": 0.0026291804804649314, "grad_norm": 0.148797, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:05.572607+00:00", "epoch": 0, "step": 5980, "train_loss": 3.641848564147949, "perplexity": 38.16231705273186, "lr": 0.0026291804804649314, "grad_norm": 0.139698, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:05.877239+00:00", "epoch": 0, "step": 5981, "train_loss": 3.596552610397339, "perplexity": 36.47228329685689, "lr": 0.0026291804804649314, "grad_norm": 0.165864, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:06.183620+00:00", "epoch": 0, "step": 5982, "train_loss": 3.529400110244751, "perplexity": 34.10350313500341, "lr": 0.0026291804804649314, "grad_norm": 0.17373, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:06.489282+00:00", "epoch": 0, "step": 5983, "train_loss": 3.5548219680786133, "perplexity": 34.981591573322554, "lr": 0.0026291804804649314, "grad_norm": 0.178401, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:06.794909+00:00", "epoch": 0, "step": 5984, "train_loss": 3.610381603240967, "perplexity": 36.9801618722422, "lr": 0.0026291804804649314, "grad_norm": 0.161497, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:07.100821+00:00", "epoch": 0, "step": 5985, "train_loss": 3.5467123985290527, "perplexity": 34.6990531046152, "lr": 0.0026291804804649314, "grad_norm": 0.179206, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:07.405464+00:00", "epoch": 0, "step": 5986, "train_loss": 3.5429162979125977, "perplexity": 34.567581704944395, "lr": 0.0026291804804649314, "grad_norm": 0.171365, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:07.711536+00:00", "epoch": 0, "step": 5987, "train_loss": 3.458035469055176, "perplexity": 31.754532436478804, "lr": 0.0026291804804649314, "grad_norm": 0.144518, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:08.018242+00:00", "epoch": 0, "step": 5988, "train_loss": 3.5041568279266357, "perplexity": 33.253393695602426, "lr": 0.0026291804804649314, "grad_norm": 0.151595, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:08.324248+00:00", "epoch": 0, "step": 5989, "train_loss": 3.6119892597198486, "perplexity": 37.03966108339742, "lr": 0.0026291804804649314, "grad_norm": 0.178855, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:08.630652+00:00", "epoch": 0, "step": 5990, "train_loss": 3.636665105819702, "perplexity": 37.965016065064184, "lr": 0.0026291804804649314, "grad_norm": 0.157003, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:08.936587+00:00", "epoch": 0, "step": 5991, "train_loss": 3.5247232913970947, "perplexity": 33.944379614747284, "lr": 0.0026291804804649314, "grad_norm": 0.169965, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:09.241526+00:00", "epoch": 0, "step": 5992, "train_loss": 3.5552191734313965, "perplexity": 34.99548920866921, "lr": 0.0026291804804649314, "grad_norm": 0.1738, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:09.547137+00:00", "epoch": 0, "step": 5993, "train_loss": 3.5124764442443848, "perplexity": 33.5312032042699, "lr": 0.0026291804804649314, "grad_norm": 0.151825, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:09.852620+00:00", "epoch": 0, "step": 5994, "train_loss": 3.5605430603027344, "perplexity": 35.18229806760029, "lr": 0.0026291804804649314, "grad_norm": 0.160772, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:10.158540+00:00", "epoch": 0, "step": 5995, "train_loss": 3.571803331375122, "perplexity": 35.58069912317843, "lr": 0.0026291804804649314, "grad_norm": 0.161767, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:10.463484+00:00", "epoch": 0, "step": 5996, "train_loss": 3.5666801929473877, "perplexity": 35.39888041509588, "lr": 0.0026291804804649314, "grad_norm": 0.161522, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:10.769317+00:00", "epoch": 0, "step": 5997, "train_loss": 3.5829617977142334, "perplexity": 35.979948519559464, "lr": 0.0026291804804649314, "grad_norm": 0.150682, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:11.074577+00:00", "epoch": 0, "step": 5998, "train_loss": 3.650334358215332, "perplexity": 38.48753252059314, "lr": 0.0026291804804649314, "grad_norm": 0.176636, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:11.381503+00:00", "epoch": 0, "step": 5999, "train_loss": 3.5793073177337646, "perplexity": 35.84870048572304, "lr": 0.0026291804804649314, "grad_norm": 0.215542, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:11.688785+00:00", "epoch": 0, "step": 6000, "train_loss": 3.593423366546631, "perplexity": 36.358331013855974, "lr": 0.0026291804804649314, "grad_norm": 0.204439, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:29:14.762964+00:00", "step": 6000, "epoch": 0, "val_loss": 3.5405242681503295, "val_ppl": 34.48499383639563, "eval_train_loss": 3.593423366546631, "eval_train_ppl": 36.358331013855974} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:29:15.776913+00:00", "step": 6000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5405_epoch_0000_step_0006000.pt", "val_loss": 3.5405242681503295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:17.070757+00:00", "epoch": 0, "step": 6001, "train_loss": 3.595486640930176, "perplexity": 36.43342567066572, "lr": 0.0026291804804649314, "grad_norm": 0.173972, "tokens_per_sec": 6088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:17.374468+00:00", "epoch": 0, "step": 6002, "train_loss": 3.5040764808654785, "perplexity": 33.250721990478745, "lr": 0.0026291804804649314, "grad_norm": 0.183485, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:17.679019+00:00", "epoch": 0, "step": 6003, "train_loss": 3.6459105014801025, "perplexity": 38.31764524617607, "lr": 0.0026291804804649314, "grad_norm": 0.154343, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:17.984351+00:00", "epoch": 0, "step": 6004, "train_loss": 3.625492572784424, "perplexity": 37.54321136999328, "lr": 0.0026291804804649314, "grad_norm": 0.164709, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:18.290225+00:00", "epoch": 0, "step": 6005, "train_loss": 3.5085766315460205, "perplexity": 33.400692441394085, "lr": 0.0026291804804649314, "grad_norm": 0.168582, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:18.595457+00:00", "epoch": 0, "step": 6006, "train_loss": 3.625145196914673, "perplexity": 37.53017202919768, "lr": 0.0026291804804649314, "grad_norm": 0.193512, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:18.900538+00:00", "epoch": 0, "step": 6007, "train_loss": 3.6088943481445312, "perplexity": 36.92520381648522, "lr": 0.0026291804804649314, "grad_norm": 0.184597, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:19.205625+00:00", "epoch": 0, "step": 6008, "train_loss": 3.5962066650390625, "perplexity": 36.45966806196139, "lr": 0.0026291804804649314, "grad_norm": 0.188366, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:19.510705+00:00", "epoch": 0, "step": 6009, "train_loss": 3.6182820796966553, "perplexity": 37.273479920929525, "lr": 0.0026291804804649314, "grad_norm": 0.167492, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:19.816561+00:00", "epoch": 0, "step": 6010, "train_loss": 3.5796520709991455, "perplexity": 35.86106157291531, "lr": 0.0026291804804649314, "grad_norm": 0.160349, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:20.121959+00:00", "epoch": 0, "step": 6011, "train_loss": 3.48752498626709, "perplexity": 32.70490237094297, "lr": 0.0026291804804649314, "grad_norm": 0.148485, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:20.426303+00:00", "epoch": 0, "step": 6012, "train_loss": 3.5816996097564697, "perplexity": 35.93456370992249, "lr": 0.0026291804804649314, "grad_norm": 0.172988, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:20.731392+00:00", "epoch": 0, "step": 6013, "train_loss": 3.517340898513794, "perplexity": 33.69471157593004, "lr": 0.0026291804804649314, "grad_norm": 0.158032, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:21.131268+00:00", "epoch": 0, "step": 6014, "train_loss": 3.6051456928253174, "perplexity": 36.78704307507558, "lr": 0.0026291804804649314, "grad_norm": 0.180776, "tokens_per_sec": 81945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:21.437624+00:00", "epoch": 0, "step": 6015, "train_loss": 3.66064715385437, "perplexity": 38.886500278636184, "lr": 0.0026291804804649314, "grad_norm": 0.152096, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:21.742112+00:00", "epoch": 0, "step": 6016, "train_loss": 3.5876195430755615, "perplexity": 36.1479248496821, "lr": 0.0026291804804649314, "grad_norm": 0.139558, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:22.047056+00:00", "epoch": 0, "step": 6017, "train_loss": 3.5813584327697754, "perplexity": 35.92230575494274, "lr": 0.0026291804804649314, "grad_norm": 0.140645, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:22.352061+00:00", "epoch": 0, "step": 6018, "train_loss": 3.6231205463409424, "perplexity": 37.45426341500664, "lr": 0.0026291804804649314, "grad_norm": 0.130967, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:22.657037+00:00", "epoch": 0, "step": 6019, "train_loss": 3.5880565643310547, "perplexity": 36.16372571358859, "lr": 0.0026291804804649314, "grad_norm": 0.143099, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:22.963451+00:00", "epoch": 0, "step": 6020, "train_loss": 3.6287169456481934, "perplexity": 37.66446005238647, "lr": 0.0026291804804649314, "grad_norm": 0.139565, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:23.269293+00:00", "epoch": 0, "step": 6021, "train_loss": 3.5579111576080322, "perplexity": 35.08982342801721, "lr": 0.0026291804804649314, "grad_norm": 0.149267, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:23.575352+00:00", "epoch": 0, "step": 6022, "train_loss": 3.5334794521331787, "perplexity": 34.242907128841104, "lr": 0.0026291804804649314, "grad_norm": 0.157165, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:23.880632+00:00", "epoch": 0, "step": 6023, "train_loss": 3.6789348125457764, "perplexity": 39.6041857148987, "lr": 0.0026291804804649314, "grad_norm": 0.156695, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:24.186177+00:00", "epoch": 0, "step": 6024, "train_loss": 3.5456089973449707, "perplexity": 34.66078724351364, "lr": 0.0026291804804649314, "grad_norm": 0.154779, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:24.491478+00:00", "epoch": 0, "step": 6025, "train_loss": 3.5510120391845703, "perplexity": 34.84856776336585, "lr": 0.0026291804804649314, "grad_norm": 0.145169, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:24.796675+00:00", "epoch": 0, "step": 6026, "train_loss": 3.549574851989746, "perplexity": 34.79851982076403, "lr": 0.0026291804804649314, "grad_norm": 0.143883, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:25.101767+00:00", "epoch": 0, "step": 6027, "train_loss": 3.6945362091064453, "perplexity": 40.226911385681795, "lr": 0.0026291804804649314, "grad_norm": 0.159883, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:25.407220+00:00", "epoch": 0, "step": 6028, "train_loss": 3.6064882278442383, "perplexity": 36.836464135975845, "lr": 0.0026291804804649314, "grad_norm": 0.142357, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:25.713177+00:00", "epoch": 0, "step": 6029, "train_loss": 3.586179494857788, "perplexity": 36.09590755761966, "lr": 0.0026291804804649314, "grad_norm": 0.150208, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:26.020613+00:00", "epoch": 0, "step": 6030, "train_loss": 3.552497625350952, "perplexity": 34.90037678740964, "lr": 0.0026291804804649314, "grad_norm": 0.170013, "tokens_per_sec": 106647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:26.326432+00:00", "epoch": 0, "step": 6031, "train_loss": 3.5271859169006348, "perplexity": 34.02807492249041, "lr": 0.0026291804804649314, "grad_norm": 0.151496, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:26.631708+00:00", "epoch": 0, "step": 6032, "train_loss": 3.5510504245758057, "perplexity": 34.849905464947376, "lr": 0.0026291804804649314, "grad_norm": 0.165542, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:26.936623+00:00", "epoch": 0, "step": 6033, "train_loss": 3.5911366939544678, "perplexity": 36.27528639893803, "lr": 0.0026291804804649314, "grad_norm": 0.161677, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:27.242224+00:00", "epoch": 0, "step": 6034, "train_loss": 3.539257287979126, "perplexity": 34.44132969963952, "lr": 0.0026291804804649314, "grad_norm": 0.162179, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:27.547805+00:00", "epoch": 0, "step": 6035, "train_loss": 3.601532459259033, "perplexity": 36.65436274309708, "lr": 0.0026291804804649314, "grad_norm": 0.189484, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:27.854616+00:00", "epoch": 0, "step": 6036, "train_loss": 3.6498067378997803, "perplexity": 38.46723107273895, "lr": 0.0026291804804649314, "grad_norm": 0.165506, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:28.161215+00:00", "epoch": 0, "step": 6037, "train_loss": 3.5404207706451416, "perplexity": 34.48142491025781, "lr": 0.0026291804804649314, "grad_norm": 0.151688, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:28.466762+00:00", "epoch": 0, "step": 6038, "train_loss": 3.677302598953247, "perplexity": 39.53959595113854, "lr": 0.0026291804804649314, "grad_norm": 0.146831, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:28.772520+00:00", "epoch": 0, "step": 6039, "train_loss": 3.6879055500030518, "perplexity": 39.96106279919221, "lr": 0.0026291804804649314, "grad_norm": 0.166693, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:29.078450+00:00", "epoch": 0, "step": 6040, "train_loss": 3.6947991847991943, "perplexity": 40.237491476663, "lr": 0.0026291804804649314, "grad_norm": 0.165106, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:29.385300+00:00", "epoch": 0, "step": 6041, "train_loss": 3.577058792114258, "perplexity": 35.76818431950937, "lr": 0.0026291804804649314, "grad_norm": 0.140379, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:29.692185+00:00", "epoch": 0, "step": 6042, "train_loss": 3.5992116928100586, "perplexity": 36.56939516092731, "lr": 0.0026291804804649314, "grad_norm": 0.158432, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:29.998035+00:00", "epoch": 0, "step": 6043, "train_loss": 3.4827871322631836, "perplexity": 32.550317807011574, "lr": 0.0026291804804649314, "grad_norm": 0.160123, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:30.304215+00:00", "epoch": 0, "step": 6044, "train_loss": 3.521912097930908, "perplexity": 33.84908939893162, "lr": 0.0026291804804649314, "grad_norm": 0.145914, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:30.610604+00:00", "epoch": 0, "step": 6045, "train_loss": 3.5876452922821045, "perplexity": 36.14885564204869, "lr": 0.0026291804804649314, "grad_norm": 0.13598, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:30.916093+00:00", "epoch": 0, "step": 6046, "train_loss": 3.4375498294830322, "perplexity": 31.110638339370052, "lr": 0.0026291804804649314, "grad_norm": 0.153202, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:31.222600+00:00", "epoch": 0, "step": 6047, "train_loss": 3.56921124458313, "perplexity": 35.48859029154694, "lr": 0.0026291804804649314, "grad_norm": 0.148142, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:31.528252+00:00", "epoch": 0, "step": 6048, "train_loss": 3.527299404144287, "perplexity": 34.03193689405827, "lr": 0.0026291804804649314, "grad_norm": 0.156682, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:31.834294+00:00", "epoch": 0, "step": 6049, "train_loss": 3.625547409057617, "perplexity": 37.545270156236086, "lr": 0.0026291804804649314, "grad_norm": 0.145591, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:32.139664+00:00", "epoch": 0, "step": 6050, "train_loss": 3.6443326473236084, "perplexity": 38.25723326353987, "lr": 0.0026291804804649314, "grad_norm": 0.16456, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:32.446061+00:00", "epoch": 0, "step": 6051, "train_loss": 3.5825929641723633, "perplexity": 35.966680354733214, "lr": 0.0026291804804649314, "grad_norm": 0.161627, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:32.752216+00:00", "epoch": 0, "step": 6052, "train_loss": 3.5948984622955322, "perplexity": 36.41200260901056, "lr": 0.0026291804804649314, "grad_norm": 0.174016, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:33.058878+00:00", "epoch": 0, "step": 6053, "train_loss": 3.5831871032714844, "perplexity": 35.988055915196874, "lr": 0.0026291804804649314, "grad_norm": 0.18854, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:33.363772+00:00", "epoch": 0, "step": 6054, "train_loss": 3.575104236602783, "perplexity": 35.6983416955899, "lr": 0.0026291804804649314, "grad_norm": 0.157608, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:33.669298+00:00", "epoch": 0, "step": 6055, "train_loss": 3.5893170833587646, "perplexity": 36.209339520459515, "lr": 0.0026291804804649314, "grad_norm": 0.168041, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:33.975979+00:00", "epoch": 0, "step": 6056, "train_loss": 3.557175636291504, "perplexity": 35.064023604216, "lr": 0.0026291804804649314, "grad_norm": 0.1916, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:34.281710+00:00", "epoch": 0, "step": 6057, "train_loss": 3.549626111984253, "perplexity": 34.80030363841774, "lr": 0.0026291804804649314, "grad_norm": 0.170317, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:34.586576+00:00", "epoch": 0, "step": 6058, "train_loss": 3.541013479232788, "perplexity": 34.501868404835314, "lr": 0.0026291804804649314, "grad_norm": 0.157552, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:34.891702+00:00", "epoch": 0, "step": 6059, "train_loss": 3.4511373043060303, "perplexity": 31.536238220553464, "lr": 0.0026291804804649314, "grad_norm": 0.154876, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:35.197793+00:00", "epoch": 0, "step": 6060, "train_loss": 3.570563316345215, "perplexity": 35.53660586529228, "lr": 0.0026291804804649314, "grad_norm": 0.158274, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:35.503622+00:00", "epoch": 0, "step": 6061, "train_loss": 3.5438594818115234, "perplexity": 34.60020067185736, "lr": 0.0026291804804649314, "grad_norm": 0.147115, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:35.809912+00:00", "epoch": 0, "step": 6062, "train_loss": 3.5132956504821777, "perplexity": 33.55868342954469, "lr": 0.0026291804804649314, "grad_norm": 0.156457, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:36.116499+00:00", "epoch": 0, "step": 6063, "train_loss": 3.566016435623169, "perplexity": 35.37539194513557, "lr": 0.0026291804804649314, "grad_norm": 0.146808, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:36.422765+00:00", "epoch": 0, "step": 6064, "train_loss": 3.5385706424713135, "perplexity": 34.41768883269457, "lr": 0.0026291804804649314, "grad_norm": 0.156507, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:36.729067+00:00", "epoch": 0, "step": 6065, "train_loss": 3.5621564388275146, "perplexity": 35.2391062459799, "lr": 0.0026291804804649314, "grad_norm": 0.160519, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:37.034713+00:00", "epoch": 0, "step": 6066, "train_loss": 3.670698404312134, "perplexity": 39.27932913576904, "lr": 0.0026291804804649314, "grad_norm": 0.138818, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:37.341498+00:00", "epoch": 0, "step": 6067, "train_loss": 3.562753200531006, "perplexity": 35.260141871056206, "lr": 0.0026291804804649314, "grad_norm": 0.171522, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:37.647741+00:00", "epoch": 0, "step": 6068, "train_loss": 3.6186134815216064, "perplexity": 37.28583446724413, "lr": 0.0026291804804649314, "grad_norm": 0.154042, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:37.954477+00:00", "epoch": 0, "step": 6069, "train_loss": 3.604428291320801, "perplexity": 36.760661459263716, "lr": 0.0026291804804649314, "grad_norm": 0.166451, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:38.260408+00:00", "epoch": 0, "step": 6070, "train_loss": 3.511491537094116, "perplexity": 33.49819434047823, "lr": 0.0026291804804649314, "grad_norm": 0.155461, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:38.564991+00:00", "epoch": 0, "step": 6071, "train_loss": 3.4916040897369385, "perplexity": 32.83858151185094, "lr": 0.0026291804804649314, "grad_norm": 0.153747, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:38.871583+00:00", "epoch": 0, "step": 6072, "train_loss": 3.6462037563323975, "perplexity": 38.328883729362346, "lr": 0.0026291804804649314, "grad_norm": 0.143297, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:39.178575+00:00", "epoch": 0, "step": 6073, "train_loss": 3.745870590209961, "perplexity": 42.34585706454569, "lr": 0.0026291804804649314, "grad_norm": 0.154485, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:39.484271+00:00", "epoch": 0, "step": 6074, "train_loss": 3.6123175621032715, "perplexity": 37.051823288743535, "lr": 0.0026291804804649314, "grad_norm": 0.143963, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:39.789547+00:00", "epoch": 0, "step": 6075, "train_loss": 3.578977346420288, "perplexity": 35.83687339434509, "lr": 0.0026291804804649314, "grad_norm": 0.142034, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:40.095429+00:00", "epoch": 0, "step": 6076, "train_loss": 3.5190317630767822, "perplexity": 33.751732963830506, "lr": 0.0026291804804649314, "grad_norm": 0.150235, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:40.401878+00:00", "epoch": 0, "step": 6077, "train_loss": 3.6060776710510254, "perplexity": 36.8213437794818, "lr": 0.0026291804804649314, "grad_norm": 0.148481, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:40.707569+00:00", "epoch": 0, "step": 6078, "train_loss": 3.479881763458252, "perplexity": 32.455884377462006, "lr": 0.0026291804804649314, "grad_norm": 0.148066, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:41.014277+00:00", "epoch": 0, "step": 6079, "train_loss": 3.594877004623413, "perplexity": 36.41122130057993, "lr": 0.0026291804804649314, "grad_norm": 0.152532, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:41.320329+00:00", "epoch": 0, "step": 6080, "train_loss": 3.5770630836486816, "perplexity": 35.768337820233036, "lr": 0.0026291804804649314, "grad_norm": 0.140517, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:41.626535+00:00", "epoch": 0, "step": 6081, "train_loss": 3.538083076477051, "perplexity": 34.40091202825065, "lr": 0.0026291804804649314, "grad_norm": 0.152695, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:41.931201+00:00", "epoch": 0, "step": 6082, "train_loss": 3.5958092212677, "perplexity": 36.445180273215634, "lr": 0.0026291804804649314, "grad_norm": 0.145346, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:42.237040+00:00", "epoch": 0, "step": 6083, "train_loss": 3.6607460975646973, "perplexity": 38.89034804360833, "lr": 0.0026291804804649314, "grad_norm": 0.168959, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:42.543602+00:00", "epoch": 0, "step": 6084, "train_loss": 3.572667360305786, "perplexity": 35.611455161734256, "lr": 0.0026291804804649314, "grad_norm": 0.167027, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:42.849641+00:00", "epoch": 0, "step": 6085, "train_loss": 3.5357532501220703, "perplexity": 34.32085716994332, "lr": 0.0026291804804649314, "grad_norm": 0.149935, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:43.155005+00:00", "epoch": 0, "step": 6086, "train_loss": 3.587831735610962, "perplexity": 36.15559598335528, "lr": 0.0026291804804649314, "grad_norm": 0.149903, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:43.460326+00:00", "epoch": 0, "step": 6087, "train_loss": 3.5335605144500732, "perplexity": 34.245683050739984, "lr": 0.0026291804804649314, "grad_norm": 0.150634, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:43.766777+00:00", "epoch": 0, "step": 6088, "train_loss": 3.5689585208892822, "perplexity": 35.479622617138645, "lr": 0.0026291804804649314, "grad_norm": 0.16984, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:44.074170+00:00", "epoch": 0, "step": 6089, "train_loss": 3.6199684143066406, "perplexity": 37.33638850771047, "lr": 0.0026291804804649314, "grad_norm": 0.170842, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:44.380347+00:00", "epoch": 0, "step": 6090, "train_loss": 3.6340086460113525, "perplexity": 37.86429736258566, "lr": 0.0026291804804649314, "grad_norm": 0.17616, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:44.685999+00:00", "epoch": 0, "step": 6091, "train_loss": 3.679729461669922, "perplexity": 39.63566965407513, "lr": 0.0026291804804649314, "grad_norm": 0.160064, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:44.992417+00:00", "epoch": 0, "step": 6092, "train_loss": 3.592078447341919, "perplexity": 36.30946486410159, "lr": 0.0026291804804649314, "grad_norm": 0.188521, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:45.299591+00:00", "epoch": 0, "step": 6093, "train_loss": 3.703011989593506, "perplexity": 40.56931487484492, "lr": 0.0026291804804649314, "grad_norm": 0.187214, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:45.605689+00:00", "epoch": 0, "step": 6094, "train_loss": 3.5656986236572266, "perplexity": 35.36415100862219, "lr": 0.0026291804804649314, "grad_norm": 0.161046, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:45.912227+00:00", "epoch": 0, "step": 6095, "train_loss": 3.5119121074676514, "perplexity": 33.512285651571034, "lr": 0.0026291804804649314, "grad_norm": 0.170533, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:46.218239+00:00", "epoch": 0, "step": 6096, "train_loss": 3.586740493774414, "perplexity": 36.116163003764555, "lr": 0.0026291804804649314, "grad_norm": 0.162813, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:46.524394+00:00", "epoch": 0, "step": 6097, "train_loss": 3.5662307739257812, "perplexity": 35.382975059246206, "lr": 0.0026291804804649314, "grad_norm": 0.171986, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:46.831354+00:00", "epoch": 0, "step": 6098, "train_loss": 3.5546531677246094, "perplexity": 34.97568716662835, "lr": 0.0026291804804649314, "grad_norm": 0.138952, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:47.137184+00:00", "epoch": 0, "step": 6099, "train_loss": 3.63272762298584, "perplexity": 37.815823380591766, "lr": 0.0026291804804649314, "grad_norm": 0.158824, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:47.444274+00:00", "epoch": 0, "step": 6100, "train_loss": 3.5856740474700928, "perplexity": 36.077667585499626, "lr": 0.0026291804804649314, "grad_norm": 0.146268, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:47.749765+00:00", "epoch": 0, "step": 6101, "train_loss": 3.494264602661133, "perplexity": 32.926065306655545, "lr": 0.0026291804804649314, "grad_norm": 0.151091, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:48.057120+00:00", "epoch": 0, "step": 6102, "train_loss": 3.6137123107910156, "perplexity": 37.10353732631368, "lr": 0.0026291804804649314, "grad_norm": 0.154359, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:48.363075+00:00", "epoch": 0, "step": 6103, "train_loss": 3.6399428844451904, "perplexity": 38.08966115114089, "lr": 0.0026291804804649314, "grad_norm": 0.152299, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:48.669447+00:00", "epoch": 0, "step": 6104, "train_loss": 3.5645763874053955, "perplexity": 35.3244863370741, "lr": 0.0026291804804649314, "grad_norm": 0.180165, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:48.975556+00:00", "epoch": 0, "step": 6105, "train_loss": 3.6293561458587646, "perplexity": 37.688542879237026, "lr": 0.0026291804804649314, "grad_norm": 0.169641, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:49.281029+00:00", "epoch": 0, "step": 6106, "train_loss": 3.5558652877807617, "perplexity": 35.01810760265758, "lr": 0.0026291804804649314, "grad_norm": 0.156353, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:49.587551+00:00", "epoch": 0, "step": 6107, "train_loss": 3.616145610809326, "perplexity": 37.193931297593664, "lr": 0.0026291804804649314, "grad_norm": 0.151519, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:49.893410+00:00", "epoch": 0, "step": 6108, "train_loss": 3.597621440887451, "perplexity": 36.51128682565585, "lr": 0.0026291804804649314, "grad_norm": 0.168193, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:50.201329+00:00", "epoch": 0, "step": 6109, "train_loss": 3.5242691040039062, "perplexity": 33.928966006055006, "lr": 0.0026291804804649314, "grad_norm": 0.177932, "tokens_per_sec": 106426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:50.506964+00:00", "epoch": 0, "step": 6110, "train_loss": 3.595465660095215, "perplexity": 36.43266127499352, "lr": 0.0026291804804649314, "grad_norm": 0.190045, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:50.812192+00:00", "epoch": 0, "step": 6111, "train_loss": 3.635183811187744, "perplexity": 37.908820322066575, "lr": 0.0026291804804649314, "grad_norm": 0.183651, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:51.117480+00:00", "epoch": 0, "step": 6112, "train_loss": 3.577178478240967, "perplexity": 35.7724655311457, "lr": 0.0026291804804649314, "grad_norm": 0.2005, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:51.423282+00:00", "epoch": 0, "step": 6113, "train_loss": 3.6349048614501953, "perplexity": 37.898247141348506, "lr": 0.0026291804804649314, "grad_norm": 0.159877, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:51.729941+00:00", "epoch": 0, "step": 6114, "train_loss": 3.5131866931915283, "perplexity": 33.55502716551238, "lr": 0.0026291804804649314, "grad_norm": 0.163272, "tokens_per_sec": 106855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:52.035431+00:00", "epoch": 0, "step": 6115, "train_loss": 3.5631768703460693, "perplexity": 35.27508369381797, "lr": 0.0026291804804649314, "grad_norm": 0.147292, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:52.341217+00:00", "epoch": 0, "step": 6116, "train_loss": 3.6942973136901855, "perplexity": 40.2173025087455, "lr": 0.0026291804804649314, "grad_norm": 0.139193, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:52.646117+00:00", "epoch": 0, "step": 6117, "train_loss": 3.6724817752838135, "perplexity": 39.349441250509805, "lr": 0.0026291804804649314, "grad_norm": 0.149489, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:52.951611+00:00", "epoch": 0, "step": 6118, "train_loss": 3.5147762298583984, "perplexity": 33.60840652453485, "lr": 0.0026291804804649314, "grad_norm": 0.1633, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:53.259761+00:00", "epoch": 0, "step": 6119, "train_loss": 3.5365383625030518, "perplexity": 34.34781348031519, "lr": 0.0026291804804649314, "grad_norm": 0.152645, "tokens_per_sec": 106337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:53.566918+00:00", "epoch": 0, "step": 6120, "train_loss": 3.635063409805298, "perplexity": 37.90425632245434, "lr": 0.0026291804804649314, "grad_norm": 0.151483, "tokens_per_sec": 106767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:53.872155+00:00", "epoch": 0, "step": 6121, "train_loss": 3.634970188140869, "perplexity": 37.900722989285185, "lr": 0.0026291804804649314, "grad_norm": 0.155584, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:54.177469+00:00", "epoch": 0, "step": 6122, "train_loss": 3.597221612930298, "perplexity": 36.49669151043328, "lr": 0.0026291804804649314, "grad_norm": 0.143018, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:54.482968+00:00", "epoch": 0, "step": 6123, "train_loss": 3.499058246612549, "perplexity": 33.084280050062375, "lr": 0.0026291804804649314, "grad_norm": 0.162887, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:54.788663+00:00", "epoch": 0, "step": 6124, "train_loss": 3.66794490814209, "perplexity": 39.17132241971121, "lr": 0.0026291804804649314, "grad_norm": 0.175081, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:55.095502+00:00", "epoch": 0, "step": 6125, "train_loss": 3.6323463916778564, "perplexity": 37.80140955245776, "lr": 0.0026291804804649314, "grad_norm": 0.144377, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:55.401135+00:00", "epoch": 0, "step": 6126, "train_loss": 3.550201177597046, "perplexity": 34.82032185169632, "lr": 0.0026291804804649314, "grad_norm": 0.147715, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:55.705838+00:00", "epoch": 0, "step": 6127, "train_loss": 3.624805212020874, "perplexity": 37.51741450645147, "lr": 0.0026291804804649314, "grad_norm": 0.154651, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:56.011816+00:00", "epoch": 0, "step": 6128, "train_loss": 3.5653164386749268, "perplexity": 35.35063794360465, "lr": 0.0026291804804649314, "grad_norm": 0.17091, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:56.317913+00:00", "epoch": 0, "step": 6129, "train_loss": 3.6430578231811523, "perplexity": 38.20849309312586, "lr": 0.0026291804804649314, "grad_norm": 0.169598, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:56.624913+00:00", "epoch": 0, "step": 6130, "train_loss": 3.6876168251037598, "perplexity": 39.94952671081803, "lr": 0.0026291804804649314, "grad_norm": 0.168722, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:56.930502+00:00", "epoch": 0, "step": 6131, "train_loss": 3.7463643550872803, "perplexity": 42.36677112435333, "lr": 0.0026291804804649314, "grad_norm": 0.161098, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:57.236712+00:00", "epoch": 0, "step": 6132, "train_loss": 3.554590940475464, "perplexity": 34.973510793544555, "lr": 0.0026291804804649314, "grad_norm": 0.163051, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:57.542448+00:00", "epoch": 0, "step": 6133, "train_loss": 3.6021320819854736, "perplexity": 36.676348122828976, "lr": 0.0026291804804649314, "grad_norm": 0.155336, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:57.848105+00:00", "epoch": 0, "step": 6134, "train_loss": 3.6355392932891846, "perplexity": 37.92229862468353, "lr": 0.0026291804804649314, "grad_norm": 0.207186, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:58.154112+00:00", "epoch": 0, "step": 6135, "train_loss": 3.6139121055603027, "perplexity": 37.110951159591416, "lr": 0.0026291804804649314, "grad_norm": 0.180523, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:58.460914+00:00", "epoch": 0, "step": 6136, "train_loss": 3.5781540870666504, "perplexity": 35.80738249411343, "lr": 0.0026291804804649314, "grad_norm": 0.193314, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:58.765858+00:00", "epoch": 0, "step": 6137, "train_loss": 3.633490562438965, "perplexity": 37.84468557285381, "lr": 0.0026291804804649314, "grad_norm": 0.209208, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:59.071229+00:00", "epoch": 0, "step": 6138, "train_loss": 3.5750367641448975, "perplexity": 35.695933121990365, "lr": 0.0026291804804649314, "grad_norm": 0.179926, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:59.377359+00:00", "epoch": 0, "step": 6139, "train_loss": 3.6639227867126465, "perplexity": 39.01408702628234, "lr": 0.0026291804804649314, "grad_norm": 0.175786, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:59.683055+00:00", "epoch": 0, "step": 6140, "train_loss": 3.582392692565918, "perplexity": 35.95947797112061, "lr": 0.0026291804804649314, "grad_norm": 0.153599, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:29:59.990988+00:00", "epoch": 0, "step": 6141, "train_loss": 3.554626703262329, "perplexity": 34.97476156612242, "lr": 0.0026291804804649314, "grad_norm": 0.174543, "tokens_per_sec": 106357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:00.298955+00:00", "epoch": 0, "step": 6142, "train_loss": 3.614985466003418, "perplexity": 37.15080597203568, "lr": 0.0026291804804649314, "grad_norm": 0.161767, "tokens_per_sec": 106401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:00.604983+00:00", "epoch": 0, "step": 6143, "train_loss": 3.5633044242858887, "perplexity": 35.27958345669572, "lr": 0.0026291804804649314, "grad_norm": 0.160956, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:00.910164+00:00", "epoch": 0, "step": 6144, "train_loss": 3.612070322036743, "perplexity": 37.04266372584076, "lr": 0.0026291804804649314, "grad_norm": 0.16303, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:01.216670+00:00", "epoch": 0, "step": 6145, "train_loss": 3.621859312057495, "perplexity": 37.407054590878076, "lr": 0.0026291804804649314, "grad_norm": 0.154714, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:01.523937+00:00", "epoch": 0, "step": 6146, "train_loss": 3.6542587280273438, "perplexity": 38.63886858647686, "lr": 0.0026291804804649314, "grad_norm": 0.157154, "tokens_per_sec": 106643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:01.830110+00:00", "epoch": 0, "step": 6147, "train_loss": 3.628512144088745, "perplexity": 37.656747102071044, "lr": 0.0026291804804649314, "grad_norm": 0.146218, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:02.137435+00:00", "epoch": 0, "step": 6148, "train_loss": 3.5504415035247803, "perplexity": 34.82869108348075, "lr": 0.0026291804804649314, "grad_norm": 0.149976, "tokens_per_sec": 106625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:02.443382+00:00", "epoch": 0, "step": 6149, "train_loss": 3.5721938610076904, "perplexity": 35.594597154153355, "lr": 0.0026291804804649314, "grad_norm": 0.151311, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:02.748556+00:00", "epoch": 0, "step": 6150, "train_loss": 3.594877243041992, "perplexity": 36.41122998169261, "lr": 0.0026291804804649314, "grad_norm": 0.168617, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:03.055156+00:00", "epoch": 0, "step": 6151, "train_loss": 3.538917064666748, "perplexity": 34.429613949465086, "lr": 0.0026291804804649314, "grad_norm": 0.153621, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:03.362222+00:00", "epoch": 0, "step": 6152, "train_loss": 3.579714298248291, "perplexity": 35.86329317756102, "lr": 0.0026291804804649314, "grad_norm": 0.173303, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:03.668550+00:00", "epoch": 0, "step": 6153, "train_loss": 3.6413819789886475, "perplexity": 38.144515235305356, "lr": 0.0026291804804649314, "grad_norm": 0.15218, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:03.974606+00:00", "epoch": 0, "step": 6154, "train_loss": 3.592104434967041, "perplexity": 36.310408473123886, "lr": 0.0026291804804649314, "grad_norm": 0.162496, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:04.281775+00:00", "epoch": 0, "step": 6155, "train_loss": 3.6298258304595947, "perplexity": 37.70624876522063, "lr": 0.0026291804804649314, "grad_norm": 0.15901, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:04.587169+00:00", "epoch": 0, "step": 6156, "train_loss": 3.582317352294922, "perplexity": 35.95676887635864, "lr": 0.0026291804804649314, "grad_norm": 0.145579, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:04.894246+00:00", "epoch": 0, "step": 6157, "train_loss": 3.609851837158203, "perplexity": 36.96057622511101, "lr": 0.0026291804804649314, "grad_norm": 0.158077, "tokens_per_sec": 106709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:05.200256+00:00", "epoch": 0, "step": 6158, "train_loss": 3.5185465812683105, "perplexity": 33.73536120895204, "lr": 0.0026291804804649314, "grad_norm": 0.14683, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:05.505957+00:00", "epoch": 0, "step": 6159, "train_loss": 3.6390187740325928, "perplexity": 38.05447835755553, "lr": 0.0026291804804649314, "grad_norm": 0.14688, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:05.811073+00:00", "epoch": 0, "step": 6160, "train_loss": 3.561281442642212, "perplexity": 35.20828564836026, "lr": 0.0026291804804649314, "grad_norm": 0.148113, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:06.115778+00:00", "epoch": 0, "step": 6161, "train_loss": 3.585864543914795, "perplexity": 36.08454090755873, "lr": 0.0026291804804649314, "grad_norm": 0.169184, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:06.421205+00:00", "epoch": 0, "step": 6162, "train_loss": 3.5749664306640625, "perplexity": 35.693422591050464, "lr": 0.0026291804804649314, "grad_norm": 0.174219, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:06.727267+00:00", "epoch": 0, "step": 6163, "train_loss": 3.6396384239196777, "perplexity": 38.0780661180947, "lr": 0.0026291804804649314, "grad_norm": 0.154855, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:07.034011+00:00", "epoch": 0, "step": 6164, "train_loss": 3.5665693283081055, "perplexity": 35.39495614852289, "lr": 0.0026291804804649314, "grad_norm": 0.157769, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:07.339712+00:00", "epoch": 0, "step": 6165, "train_loss": 3.567087411880493, "perplexity": 35.413298444859755, "lr": 0.0026291804804649314, "grad_norm": 0.162528, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:07.645617+00:00", "epoch": 0, "step": 6166, "train_loss": 3.607537031173706, "perplexity": 36.87511860913426, "lr": 0.0026291804804649314, "grad_norm": 0.157627, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:07.951330+00:00", "epoch": 0, "step": 6167, "train_loss": 3.5992391109466553, "perplexity": 36.570397839344814, "lr": 0.0026291804804649314, "grad_norm": 0.13088, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:08.258330+00:00", "epoch": 0, "step": 6168, "train_loss": 3.6021692752838135, "perplexity": 36.67771226255499, "lr": 0.0026291804804649314, "grad_norm": 0.156355, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:08.563626+00:00", "epoch": 0, "step": 6169, "train_loss": 3.4856553077697754, "perplexity": 32.64381184584, "lr": 0.0026291804804649314, "grad_norm": 0.151965, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:08.871260+00:00", "epoch": 0, "step": 6170, "train_loss": 3.6581575870513916, "perplexity": 38.78981014666983, "lr": 0.0026291804804649314, "grad_norm": 0.16023, "tokens_per_sec": 106570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:09.177796+00:00", "epoch": 0, "step": 6171, "train_loss": 3.611595630645752, "perplexity": 37.02508406505764, "lr": 0.0026291804804649314, "grad_norm": 0.149352, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:09.484728+00:00", "epoch": 0, "step": 6172, "train_loss": 3.530501365661621, "perplexity": 34.14108048990019, "lr": 0.0026291804804649314, "grad_norm": 0.174539, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:09.789727+00:00", "epoch": 0, "step": 6173, "train_loss": 3.5298194885253906, "perplexity": 34.11780840296163, "lr": 0.0026291804804649314, "grad_norm": 0.149274, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:10.096332+00:00", "epoch": 0, "step": 6174, "train_loss": 3.4998114109039307, "perplexity": 33.10920733439403, "lr": 0.0026291804804649314, "grad_norm": 0.150882, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:10.403355+00:00", "epoch": 0, "step": 6175, "train_loss": 3.607941150665283, "perplexity": 36.890023574809156, "lr": 0.0026291804804649314, "grad_norm": 0.158053, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:10.710235+00:00", "epoch": 0, "step": 6176, "train_loss": 3.5856354236602783, "perplexity": 36.07627415543821, "lr": 0.0026291804804649314, "grad_norm": 0.154665, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:11.015651+00:00", "epoch": 0, "step": 6177, "train_loss": 3.5346760749816895, "perplexity": 34.283907500008986, "lr": 0.0026291804804649314, "grad_norm": 0.158295, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:11.321421+00:00", "epoch": 0, "step": 6178, "train_loss": 3.54909610748291, "perplexity": 34.78186420776355, "lr": 0.0026291804804649314, "grad_norm": 0.155652, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:11.627790+00:00", "epoch": 0, "step": 6179, "train_loss": 3.558215856552124, "perplexity": 35.10051688922454, "lr": 0.0026291804804649314, "grad_norm": 0.158166, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:11.933996+00:00", "epoch": 0, "step": 6180, "train_loss": 3.6261866092681885, "perplexity": 37.569276772525626, "lr": 0.0026291804804649314, "grad_norm": 0.152886, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:12.240682+00:00", "epoch": 0, "step": 6181, "train_loss": 3.529733180999756, "perplexity": 34.11486390640622, "lr": 0.0026291804804649314, "grad_norm": 0.179494, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:12.546299+00:00", "epoch": 0, "step": 6182, "train_loss": 3.5669519901275635, "perplexity": 35.40850303861582, "lr": 0.0026291804804649314, "grad_norm": 0.176188, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:12.853412+00:00", "epoch": 0, "step": 6183, "train_loss": 3.561417579650879, "perplexity": 35.213079125326125, "lr": 0.0026291804804649314, "grad_norm": 0.158315, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:13.158446+00:00", "epoch": 0, "step": 6184, "train_loss": 3.6873347759246826, "perplexity": 39.938260568482505, "lr": 0.0026291804804649314, "grad_norm": 0.187081, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:13.463586+00:00", "epoch": 0, "step": 6185, "train_loss": 3.723358631134033, "perplexity": 41.403218987671465, "lr": 0.0026291804804649314, "grad_norm": 0.176513, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:13.769677+00:00", "epoch": 0, "step": 6186, "train_loss": 3.6325125694274902, "perplexity": 37.80769182760289, "lr": 0.0026291804804649314, "grad_norm": 0.180847, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:14.076518+00:00", "epoch": 0, "step": 6187, "train_loss": 3.522263288497925, "perplexity": 33.86097896746058, "lr": 0.0026291804804649314, "grad_norm": 0.159344, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:14.383308+00:00", "epoch": 0, "step": 6188, "train_loss": 3.504070520401001, "perplexity": 33.250523801322124, "lr": 0.0026291804804649314, "grad_norm": 0.161669, "tokens_per_sec": 106809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:14.689101+00:00", "epoch": 0, "step": 6189, "train_loss": 3.6765425205230713, "perplexity": 39.50955417561731, "lr": 0.0026291804804649314, "grad_norm": 0.183899, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:14.996326+00:00", "epoch": 0, "step": 6190, "train_loss": 3.5837175846099854, "perplexity": 36.00715197187185, "lr": 0.0026291804804649314, "grad_norm": 0.178835, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:15.301625+00:00", "epoch": 0, "step": 6191, "train_loss": 3.5800487995147705, "perplexity": 35.875291501164384, "lr": 0.0026291804804649314, "grad_norm": 0.159707, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:15.606754+00:00", "epoch": 0, "step": 6192, "train_loss": 3.545074224472046, "perplexity": 34.642256550039306, "lr": 0.0026291804804649314, "grad_norm": 0.166257, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:15.913173+00:00", "epoch": 0, "step": 6193, "train_loss": 3.5744364261627197, "perplexity": 35.67450992874968, "lr": 0.0026291804804649314, "grad_norm": 0.161169, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:16.220173+00:00", "epoch": 0, "step": 6194, "train_loss": 3.5763678550720215, "perplexity": 35.74347929182171, "lr": 0.0026291804804649314, "grad_norm": 0.141475, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:16.526078+00:00", "epoch": 0, "step": 6195, "train_loss": 3.577340841293335, "perplexity": 35.77827412937818, "lr": 0.0026291804804649314, "grad_norm": 0.158325, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:16.831181+00:00", "epoch": 0, "step": 6196, "train_loss": 3.556995153427124, "perplexity": 35.057695719853754, "lr": 0.0026291804804649314, "grad_norm": 0.141692, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:17.138536+00:00", "epoch": 0, "step": 6197, "train_loss": 3.6597683429718018, "perplexity": 38.85234141079715, "lr": 0.0026291804804649314, "grad_norm": 0.164951, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:17.446286+00:00", "epoch": 0, "step": 6198, "train_loss": 3.6502859592437744, "perplexity": 38.48566980867839, "lr": 0.0026291804804649314, "grad_norm": 0.154687, "tokens_per_sec": 106477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:17.752370+00:00", "epoch": 0, "step": 6199, "train_loss": 3.6291441917419434, "perplexity": 37.68055548392734, "lr": 0.0026291804804649314, "grad_norm": 0.140562, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:18.059098+00:00", "epoch": 0, "step": 6200, "train_loss": 3.5271835327148438, "perplexity": 34.027993793334396, "lr": 0.0026291804804649314, "grad_norm": 0.165646, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:18.364665+00:00", "epoch": 0, "step": 6201, "train_loss": 3.6006486415863037, "perplexity": 36.621981281284064, "lr": 0.0026291804804649314, "grad_norm": 0.161799, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:18.671264+00:00", "epoch": 0, "step": 6202, "train_loss": 3.562481641769409, "perplexity": 35.25056797059309, "lr": 0.0026291804804649314, "grad_norm": 0.15129, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:18.977362+00:00", "epoch": 0, "step": 6203, "train_loss": 3.5327024459838867, "perplexity": 34.21631051363726, "lr": 0.0026291804804649314, "grad_norm": 0.161016, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:19.284432+00:00", "epoch": 0, "step": 6204, "train_loss": 3.5273780822753906, "perplexity": 34.03461456858676, "lr": 0.0026291804804649314, "grad_norm": 0.151533, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:19.590721+00:00", "epoch": 0, "step": 6205, "train_loss": 3.5814082622528076, "perplexity": 35.92409578946571, "lr": 0.0026291804804649314, "grad_norm": 0.179402, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:19.897342+00:00", "epoch": 0, "step": 6206, "train_loss": 3.608198404312134, "perplexity": 36.89951488869139, "lr": 0.0026291804804649314, "grad_norm": 0.166753, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:20.203542+00:00", "epoch": 0, "step": 6207, "train_loss": 3.571141004562378, "perplexity": 35.55714087462401, "lr": 0.0026291804804649314, "grad_norm": 0.151085, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:20.509292+00:00", "epoch": 0, "step": 6208, "train_loss": 3.5861992835998535, "perplexity": 36.09662185729146, "lr": 0.0026291804804649314, "grad_norm": 0.15595, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:20.815432+00:00", "epoch": 0, "step": 6209, "train_loss": 3.6181931495666504, "perplexity": 37.27016533289998, "lr": 0.0026291804804649314, "grad_norm": 0.180507, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:21.122350+00:00", "epoch": 0, "step": 6210, "train_loss": 3.637906312942505, "perplexity": 38.01216776988392, "lr": 0.0026291804804649314, "grad_norm": 0.150823, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:21.427932+00:00", "epoch": 0, "step": 6211, "train_loss": 3.5967371463775635, "perplexity": 36.47901436644926, "lr": 0.0026291804804649314, "grad_norm": 0.1598, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:21.733671+00:00", "epoch": 0, "step": 6212, "train_loss": 3.6133270263671875, "perplexity": 37.08924466485965, "lr": 0.0026291804804649314, "grad_norm": 0.163218, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:22.038814+00:00", "epoch": 0, "step": 6213, "train_loss": 3.4839670658111572, "perplexity": 32.588747686882876, "lr": 0.0026291804804649314, "grad_norm": 0.155808, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:22.344474+00:00", "epoch": 0, "step": 6214, "train_loss": 3.6008822917938232, "perplexity": 36.630539014529354, "lr": 0.0026291804804649314, "grad_norm": 0.170131, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:22.650674+00:00", "epoch": 0, "step": 6215, "train_loss": 3.5904181003570557, "perplexity": 36.24922857400185, "lr": 0.0026291804804649314, "grad_norm": 0.168531, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:22.955631+00:00", "epoch": 0, "step": 6216, "train_loss": 3.5497281551361084, "perplexity": 34.80385495227708, "lr": 0.0026291804804649314, "grad_norm": 0.147291, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:23.261133+00:00", "epoch": 0, "step": 6217, "train_loss": 3.5404069423675537, "perplexity": 34.48094809483929, "lr": 0.0026291804804649314, "grad_norm": 0.172533, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:23.565990+00:00", "epoch": 0, "step": 6218, "train_loss": 3.6521334648132324, "perplexity": 38.556838019584184, "lr": 0.0026291804804649314, "grad_norm": 0.193142, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:23.873758+00:00", "epoch": 0, "step": 6219, "train_loss": 3.5002901554107666, "perplexity": 33.12506198039058, "lr": 0.0026291804804649314, "grad_norm": 0.207281, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:24.180785+00:00", "epoch": 0, "step": 6220, "train_loss": 3.5448410511016846, "perplexity": 34.63417983999606, "lr": 0.0026291804804649314, "grad_norm": 0.142336, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:24.487636+00:00", "epoch": 0, "step": 6221, "train_loss": 3.5353128910064697, "perplexity": 34.30574699482916, "lr": 0.0026291804804649314, "grad_norm": 0.215098, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:24.793756+00:00", "epoch": 0, "step": 6222, "train_loss": 3.5364835262298584, "perplexity": 34.34593002587287, "lr": 0.0026291804804649314, "grad_norm": 0.196712, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:25.099757+00:00", "epoch": 0, "step": 6223, "train_loss": 3.512075185775757, "perplexity": 33.51775122406188, "lr": 0.0026291804804649314, "grad_norm": 0.155821, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:25.406233+00:00", "epoch": 0, "step": 6224, "train_loss": 3.493067979812622, "perplexity": 32.886688788717194, "lr": 0.0026291804804649314, "grad_norm": 0.182275, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:25.712164+00:00", "epoch": 0, "step": 6225, "train_loss": 3.5790610313415527, "perplexity": 35.83987252576276, "lr": 0.0026291804804649314, "grad_norm": 0.172232, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:26.019237+00:00", "epoch": 0, "step": 6226, "train_loss": 3.6433002948760986, "perplexity": 38.21775869448513, "lr": 0.0026291804804649314, "grad_norm": 0.181307, "tokens_per_sec": 106712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:26.325090+00:00", "epoch": 0, "step": 6227, "train_loss": 3.4732794761657715, "perplexity": 32.242307129467974, "lr": 0.0026291804804649314, "grad_norm": 0.173968, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:26.631794+00:00", "epoch": 0, "step": 6228, "train_loss": 3.7019639015197754, "perplexity": 40.526816934447716, "lr": 0.0026291804804649314, "grad_norm": 0.165616, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:26.937196+00:00", "epoch": 0, "step": 6229, "train_loss": 3.551124334335327, "perplexity": 34.852481308268445, "lr": 0.0026291804804649314, "grad_norm": 0.153296, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:27.243595+00:00", "epoch": 0, "step": 6230, "train_loss": 3.6085081100463867, "perplexity": 36.910944649883504, "lr": 0.0026291804804649314, "grad_norm": 0.149532, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:27.549746+00:00", "epoch": 0, "step": 6231, "train_loss": 3.570291519165039, "perplexity": 35.52694842851657, "lr": 0.0026291804804649314, "grad_norm": 0.141912, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:27.856083+00:00", "epoch": 0, "step": 6232, "train_loss": 3.4537250995635986, "perplexity": 31.617953233522044, "lr": 0.0026291804804649314, "grad_norm": 0.160106, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:28.161988+00:00", "epoch": 0, "step": 6233, "train_loss": 3.4845430850982666, "perplexity": 32.60752484157571, "lr": 0.0026291804804649314, "grad_norm": 0.146553, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:28.466758+00:00", "epoch": 0, "step": 6234, "train_loss": 3.607919454574585, "perplexity": 36.889223214194196, "lr": 0.0026291804804649314, "grad_norm": 0.148731, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:28.772611+00:00", "epoch": 0, "step": 6235, "train_loss": 3.6793220043182373, "perplexity": 39.61952309882535, "lr": 0.0026291804804649314, "grad_norm": 0.171555, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:29.078604+00:00", "epoch": 0, "step": 6236, "train_loss": 3.645689010620117, "perplexity": 38.30915917780586, "lr": 0.0026291804804649314, "grad_norm": 0.179019, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:29.385577+00:00", "epoch": 0, "step": 6237, "train_loss": 3.5320041179656982, "perplexity": 34.192424646382406, "lr": 0.0026291804804649314, "grad_norm": 0.156566, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:29.691434+00:00", "epoch": 0, "step": 6238, "train_loss": 3.590284585952759, "perplexity": 36.24438910291931, "lr": 0.0026291804804649314, "grad_norm": 0.165266, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:29.997472+00:00", "epoch": 0, "step": 6239, "train_loss": 3.5641350746154785, "perplexity": 35.308900628795456, "lr": 0.0026291804804649314, "grad_norm": 0.151452, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:30.303464+00:00", "epoch": 0, "step": 6240, "train_loss": 3.669970989227295, "perplexity": 39.25076714870288, "lr": 0.0026291804804649314, "grad_norm": 0.144713, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:30.609663+00:00", "epoch": 0, "step": 6241, "train_loss": 3.5492429733276367, "perplexity": 34.78697285076499, "lr": 0.0026291804804649314, "grad_norm": 0.159827, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:30.916177+00:00", "epoch": 0, "step": 6242, "train_loss": 3.624297857284546, "perplexity": 37.49838469634719, "lr": 0.0026291804804649314, "grad_norm": 0.191732, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:31.223600+00:00", "epoch": 0, "step": 6243, "train_loss": 3.547459363937378, "perplexity": 34.72498177969208, "lr": 0.0026291804804649314, "grad_norm": 0.172969, "tokens_per_sec": 106590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:31.528057+00:00", "epoch": 0, "step": 6244, "train_loss": 3.6302099227905273, "perplexity": 37.720734227899065, "lr": 0.0026291804804649314, "grad_norm": 0.167726, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:31.833017+00:00", "epoch": 0, "step": 6245, "train_loss": 3.5626792907714844, "perplexity": 35.257535898754405, "lr": 0.0026291804804649314, "grad_norm": 0.161028, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:32.139501+00:00", "epoch": 0, "step": 6246, "train_loss": 3.6523869037628174, "perplexity": 38.56661106249374, "lr": 0.0026291804804649314, "grad_norm": 0.14056, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:32.446110+00:00", "epoch": 0, "step": 6247, "train_loss": 3.515507936477661, "perplexity": 33.63300701712578, "lr": 0.0026291804804649314, "grad_norm": 0.150695, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:32.753589+00:00", "epoch": 0, "step": 6248, "train_loss": 3.6134912967681885, "perplexity": 37.09533783040324, "lr": 0.0026291804804649314, "grad_norm": 0.156549, "tokens_per_sec": 106575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:33.059502+00:00", "epoch": 0, "step": 6249, "train_loss": 3.4655206203460693, "perplexity": 31.993111702974314, "lr": 0.0026291804804649314, "grad_norm": 0.14677, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:33.365571+00:00", "epoch": 0, "step": 6250, "train_loss": 3.6190853118896484, "perplexity": 37.303431207255265, "lr": 0.0026291804804649314, "grad_norm": 0.162729, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:33.671110+00:00", "epoch": 0, "step": 6251, "train_loss": 3.5654823780059814, "perplexity": 35.35650449154943, "lr": 0.0026291804804649314, "grad_norm": 0.170038, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:33.977009+00:00", "epoch": 0, "step": 6252, "train_loss": 3.5360682010650635, "perplexity": 34.33166825866525, "lr": 0.0026291804804649314, "grad_norm": 0.159807, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:34.284139+00:00", "epoch": 0, "step": 6253, "train_loss": 3.54787540435791, "perplexity": 34.73943178139864, "lr": 0.0026291804804649314, "grad_norm": 0.146835, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:34.590885+00:00", "epoch": 0, "step": 6254, "train_loss": 3.6315174102783203, "perplexity": 37.77008587223049, "lr": 0.0026291804804649314, "grad_norm": 0.179859, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:34.897129+00:00", "epoch": 0, "step": 6255, "train_loss": 3.511014699935913, "perplexity": 33.48222496438252, "lr": 0.0026291804804649314, "grad_norm": 0.164017, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:35.202222+00:00", "epoch": 0, "step": 6256, "train_loss": 3.5187528133392334, "perplexity": 33.74231923981739, "lr": 0.0026291804804649314, "grad_norm": 0.150765, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:35.508162+00:00", "epoch": 0, "step": 6257, "train_loss": 3.7013723850250244, "perplexity": 40.502851742353215, "lr": 0.0026291804804649314, "grad_norm": 0.165235, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:35.816063+00:00", "epoch": 0, "step": 6258, "train_loss": 3.462019443511963, "perplexity": 31.8812940223893, "lr": 0.0026291804804649314, "grad_norm": 0.141444, "tokens_per_sec": 106424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:36.121903+00:00", "epoch": 0, "step": 6259, "train_loss": 3.5711090564727783, "perplexity": 35.556004910047484, "lr": 0.0026291804804649314, "grad_norm": 0.13991, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:36.429123+00:00", "epoch": 0, "step": 6260, "train_loss": 3.545236110687256, "perplexity": 34.64786510780037, "lr": 0.0026291804804649314, "grad_norm": 0.140283, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:36.734564+00:00", "epoch": 0, "step": 6261, "train_loss": 3.6267547607421875, "perplexity": 37.59062787725749, "lr": 0.0026291804804649314, "grad_norm": 0.141996, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:37.039882+00:00", "epoch": 0, "step": 6262, "train_loss": 3.6635355949401855, "perplexity": 38.998984016844865, "lr": 0.0026291804804649314, "grad_norm": 0.142709, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:37.345344+00:00", "epoch": 0, "step": 6263, "train_loss": 3.4799633026123047, "perplexity": 32.45853091071476, "lr": 0.0026291804804649314, "grad_norm": 0.139289, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:37.652524+00:00", "epoch": 0, "step": 6264, "train_loss": 3.5532033443450928, "perplexity": 34.92501533913557, "lr": 0.0026291804804649314, "grad_norm": 0.142558, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:37.959227+00:00", "epoch": 0, "step": 6265, "train_loss": 3.583493232727051, "perplexity": 35.999074605647884, "lr": 0.0026291804804649314, "grad_norm": 0.151917, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:38.265289+00:00", "epoch": 0, "step": 6266, "train_loss": 3.5624537467956543, "perplexity": 35.24958467063933, "lr": 0.0026291804804649314, "grad_norm": 0.147521, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:38.570937+00:00", "epoch": 0, "step": 6267, "train_loss": 3.6246135234832764, "perplexity": 37.51022353736548, "lr": 0.0026291804804649314, "grad_norm": 0.148298, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:38.877820+00:00", "epoch": 0, "step": 6268, "train_loss": 3.613455057144165, "perplexity": 37.093993533665795, "lr": 0.0026291804804649314, "grad_norm": 0.153362, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:39.184725+00:00", "epoch": 0, "step": 6269, "train_loss": 3.5798182487487793, "perplexity": 35.86702137860681, "lr": 0.0026291804804649314, "grad_norm": 0.163762, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:39.491122+00:00", "epoch": 0, "step": 6270, "train_loss": 3.5895767211914062, "perplexity": 36.21874205546808, "lr": 0.0026291804804649314, "grad_norm": 0.173221, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:39.797657+00:00", "epoch": 0, "step": 6271, "train_loss": 3.558687448501587, "perplexity": 35.11707391418446, "lr": 0.0026291804804649314, "grad_norm": 0.151756, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:40.105119+00:00", "epoch": 0, "step": 6272, "train_loss": 3.539637565612793, "perplexity": 34.45442945761266, "lr": 0.0026291804804649314, "grad_norm": 0.167482, "tokens_per_sec": 106577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:40.411341+00:00", "epoch": 0, "step": 6273, "train_loss": 3.6320841312408447, "perplexity": 37.79149703815585, "lr": 0.0026291804804649314, "grad_norm": 0.172095, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:40.717225+00:00", "epoch": 0, "step": 6274, "train_loss": 3.5817201137542725, "perplexity": 35.93530051969159, "lr": 0.0026291804804649314, "grad_norm": 0.203017, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:41.023393+00:00", "epoch": 0, "step": 6275, "train_loss": 3.6091532707214355, "perplexity": 36.934765823266396, "lr": 0.0026291804804649314, "grad_norm": 0.173509, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:41.330309+00:00", "epoch": 0, "step": 6276, "train_loss": 3.5862619876861572, "perplexity": 36.09888533394755, "lr": 0.0026291804804649314, "grad_norm": 0.173634, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:41.636666+00:00", "epoch": 0, "step": 6277, "train_loss": 3.601454973220825, "perplexity": 36.65152265178022, "lr": 0.0026291804804649314, "grad_norm": 0.168966, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:41.942466+00:00", "epoch": 0, "step": 6278, "train_loss": 3.5794198513031006, "perplexity": 35.85273489494373, "lr": 0.0026291804804649314, "grad_norm": 0.151197, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:42.247780+00:00", "epoch": 0, "step": 6279, "train_loss": 3.5824780464172363, "perplexity": 35.962547382048, "lr": 0.0026291804804649314, "grad_norm": 0.156848, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:42.554884+00:00", "epoch": 0, "step": 6280, "train_loss": 3.6146109104156494, "perplexity": 37.136893535721775, "lr": 0.0026291804804649314, "grad_norm": 0.155278, "tokens_per_sec": 106757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:42.861387+00:00", "epoch": 0, "step": 6281, "train_loss": 3.6967461109161377, "perplexity": 40.31590720974886, "lr": 0.0026291804804649314, "grad_norm": 0.152567, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:43.166859+00:00", "epoch": 0, "step": 6282, "train_loss": 3.614523410797119, "perplexity": 37.133644213863285, "lr": 0.0026291804804649314, "grad_norm": 0.187201, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:43.472780+00:00", "epoch": 0, "step": 6283, "train_loss": 3.5046262741088867, "perplexity": 33.26900803907993, "lr": 0.0026291804804649314, "grad_norm": 0.167398, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:43.778285+00:00", "epoch": 0, "step": 6284, "train_loss": 3.549633741378784, "perplexity": 34.80056914467683, "lr": 0.0026291804804649314, "grad_norm": 0.173527, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:44.084177+00:00", "epoch": 0, "step": 6285, "train_loss": 3.442228317260742, "perplexity": 31.256530090876222, "lr": 0.0026291804804649314, "grad_norm": 0.194502, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:44.390934+00:00", "epoch": 0, "step": 6286, "train_loss": 3.5546669960021973, "perplexity": 34.97617082348338, "lr": 0.0026291804804649314, "grad_norm": 0.177976, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:44.697297+00:00", "epoch": 0, "step": 6287, "train_loss": 3.5861101150512695, "perplexity": 36.09340331741006, "lr": 0.0026291804804649314, "grad_norm": 0.178528, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:45.004932+00:00", "epoch": 0, "step": 6288, "train_loss": 3.5813872814178467, "perplexity": 35.923342079847586, "lr": 0.0026291804804649314, "grad_norm": 0.187996, "tokens_per_sec": 106518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:45.311243+00:00", "epoch": 0, "step": 6289, "train_loss": 3.562692165374756, "perplexity": 35.25798982846351, "lr": 0.0026291804804649314, "grad_norm": 0.147341, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:45.616991+00:00", "epoch": 0, "step": 6290, "train_loss": 3.559643030166626, "perplexity": 35.15064718459275, "lr": 0.0026291804804649314, "grad_norm": 0.167898, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:45.922819+00:00", "epoch": 0, "step": 6291, "train_loss": 3.594975471496582, "perplexity": 36.41480677621205, "lr": 0.0026291804804649314, "grad_norm": 0.155076, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:46.229909+00:00", "epoch": 0, "step": 6292, "train_loss": 3.6159372329711914, "perplexity": 37.18618171404692, "lr": 0.0026291804804649314, "grad_norm": 0.149218, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:46.535229+00:00", "epoch": 0, "step": 6293, "train_loss": 3.610200881958008, "perplexity": 36.973479373797645, "lr": 0.0026291804804649314, "grad_norm": 0.149639, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:46.840387+00:00", "epoch": 0, "step": 6294, "train_loss": 3.6201376914978027, "perplexity": 37.342709241648144, "lr": 0.0026291804804649314, "grad_norm": 0.149646, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:47.146785+00:00", "epoch": 0, "step": 6295, "train_loss": 3.5407967567443848, "perplexity": 34.49439188425439, "lr": 0.0026291804804649314, "grad_norm": 0.145952, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:47.454112+00:00", "epoch": 0, "step": 6296, "train_loss": 3.5466675758361816, "perplexity": 34.69749783447094, "lr": 0.0026291804804649314, "grad_norm": 0.144883, "tokens_per_sec": 106625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:47.760755+00:00", "epoch": 0, "step": 6297, "train_loss": 3.6091561317443848, "perplexity": 36.93487149463021, "lr": 0.0026291804804649314, "grad_norm": 0.154729, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:48.068063+00:00", "epoch": 0, "step": 6298, "train_loss": 3.5569469928741455, "perplexity": 35.05600736249817, "lr": 0.0026291804804649314, "grad_norm": 0.175558, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:48.374969+00:00", "epoch": 0, "step": 6299, "train_loss": 3.6174962520599365, "perplexity": 37.2442008959312, "lr": 0.0026291804804649314, "grad_norm": 0.181233, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:48.681866+00:00", "epoch": 0, "step": 6300, "train_loss": 3.581779718399048, "perplexity": 35.93744249434915, "lr": 0.0026291804804649314, "grad_norm": 0.193883, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:48.987788+00:00", "epoch": 0, "step": 6301, "train_loss": 3.4749763011932373, "perplexity": 32.29706312568303, "lr": 0.0026291804804649314, "grad_norm": 0.155706, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:49.295561+00:00", "epoch": 0, "step": 6302, "train_loss": 3.5641489028930664, "perplexity": 35.309388893450596, "lr": 0.0026291804804649314, "grad_norm": 0.156461, "tokens_per_sec": 106469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:49.603443+00:00", "epoch": 0, "step": 6303, "train_loss": 3.637256145477295, "perplexity": 37.98746152758562, "lr": 0.0026291804804649314, "grad_norm": 0.166298, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:49.909824+00:00", "epoch": 0, "step": 6304, "train_loss": 3.6410789489746094, "perplexity": 38.13295805349305, "lr": 0.0026291804804649314, "grad_norm": 0.176275, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:50.216205+00:00", "epoch": 0, "step": 6305, "train_loss": 3.615586042404175, "perplexity": 37.17312457071256, "lr": 0.0026291804804649314, "grad_norm": 0.14773, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:50.522437+00:00", "epoch": 0, "step": 6306, "train_loss": 3.5545918941497803, "perplexity": 34.97354414689946, "lr": 0.0026291804804649314, "grad_norm": 0.147957, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:50.830272+00:00", "epoch": 0, "step": 6307, "train_loss": 3.6497764587402344, "perplexity": 38.46606633494574, "lr": 0.0026291804804649314, "grad_norm": 0.133763, "tokens_per_sec": 106446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:51.136766+00:00", "epoch": 0, "step": 6308, "train_loss": 3.5726590156555176, "perplexity": 35.61115799783525, "lr": 0.0026291804804649314, "grad_norm": 0.157546, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:51.442994+00:00", "epoch": 0, "step": 6309, "train_loss": 3.494131326675415, "perplexity": 32.92167734525743, "lr": 0.0026291804804649314, "grad_norm": 0.145096, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:51.748626+00:00", "epoch": 0, "step": 6310, "train_loss": 3.5734469890594482, "perplexity": 35.63922970164624, "lr": 0.0026291804804649314, "grad_norm": 0.152654, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:52.056651+00:00", "epoch": 0, "step": 6311, "train_loss": 3.490234375, "perplexity": 32.79363281328884, "lr": 0.0026291804804649314, "grad_norm": 0.142211, "tokens_per_sec": 106321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:52.362412+00:00", "epoch": 0, "step": 6312, "train_loss": 3.563725233078003, "perplexity": 35.29443253968964, "lr": 0.0026291804804649314, "grad_norm": 0.17487, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:52.668440+00:00", "epoch": 0, "step": 6313, "train_loss": 3.557046413421631, "perplexity": 35.059492823203136, "lr": 0.0026291804804649314, "grad_norm": 0.181726, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:52.974541+00:00", "epoch": 0, "step": 6314, "train_loss": 3.5765814781188965, "perplexity": 35.75111573840539, "lr": 0.0026291804804649314, "grad_norm": 0.156413, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:53.281271+00:00", "epoch": 0, "step": 6315, "train_loss": 3.592667579650879, "perplexity": 36.3308622452995, "lr": 0.0026291804804649314, "grad_norm": 0.166927, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:53.588345+00:00", "epoch": 0, "step": 6316, "train_loss": 3.702333450317383, "perplexity": 40.54179633855661, "lr": 0.0026291804804649314, "grad_norm": 0.15415, "tokens_per_sec": 106711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:53.894438+00:00", "epoch": 0, "step": 6317, "train_loss": 3.5431571006774902, "perplexity": 34.57590667649245, "lr": 0.0026291804804649314, "grad_norm": 0.162772, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:54.200674+00:00", "epoch": 0, "step": 6318, "train_loss": 3.545301675796509, "perplexity": 34.65013687333501, "lr": 0.0026291804804649314, "grad_norm": 0.151533, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:54.506272+00:00", "epoch": 0, "step": 6319, "train_loss": 3.5894577503204346, "perplexity": 36.214433336491346, "lr": 0.0026291804804649314, "grad_norm": 0.167027, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:54.812732+00:00", "epoch": 0, "step": 6320, "train_loss": 3.5738790035247803, "perplexity": 35.65462969068013, "lr": 0.0026291804804649314, "grad_norm": 0.158146, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:55.119533+00:00", "epoch": 0, "step": 6321, "train_loss": 3.469867467880249, "perplexity": 32.13248357637972, "lr": 0.0026291804804649314, "grad_norm": 0.146091, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:55.427651+00:00", "epoch": 0, "step": 6322, "train_loss": 3.5175724029541016, "perplexity": 33.70251295426677, "lr": 0.0026291804804649314, "grad_norm": 0.144275, "tokens_per_sec": 106349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:55.733250+00:00", "epoch": 0, "step": 6323, "train_loss": 3.673109531402588, "perplexity": 39.37415085801736, "lr": 0.0026291804804649314, "grad_norm": 0.132678, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:56.038855+00:00", "epoch": 0, "step": 6324, "train_loss": 3.492733955383301, "perplexity": 32.8757056656801, "lr": 0.0026291804804649314, "grad_norm": 0.142666, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:56.344636+00:00", "epoch": 0, "step": 6325, "train_loss": 3.64090895652771, "perplexity": 38.12647628958708, "lr": 0.0026291804804649314, "grad_norm": 0.144943, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:56.652169+00:00", "epoch": 0, "step": 6326, "train_loss": 3.5973987579345703, "perplexity": 36.50315728968017, "lr": 0.0026291804804649314, "grad_norm": 0.165889, "tokens_per_sec": 106552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:56.958836+00:00", "epoch": 0, "step": 6327, "train_loss": 3.6074540615081787, "perplexity": 36.87205921979699, "lr": 0.0026291804804649314, "grad_norm": 0.141872, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:57.264721+00:00", "epoch": 0, "step": 6328, "train_loss": 3.529540777206421, "perplexity": 34.10830070859385, "lr": 0.0026291804804649314, "grad_norm": 0.161742, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:57.570505+00:00", "epoch": 0, "step": 6329, "train_loss": 3.60567045211792, "perplexity": 36.806352483729306, "lr": 0.0026291804804649314, "grad_norm": 0.174168, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:57.876487+00:00", "epoch": 0, "step": 6330, "train_loss": 3.6567635536193848, "perplexity": 38.73577352768179, "lr": 0.0026291804804649314, "grad_norm": 0.155363, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:58.181225+00:00", "epoch": 0, "step": 6331, "train_loss": 3.56404447555542, "perplexity": 35.30570182049349, "lr": 0.0026291804804649314, "grad_norm": 0.155288, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:58.487514+00:00", "epoch": 0, "step": 6332, "train_loss": 3.609872341156006, "perplexity": 36.96133407245414, "lr": 0.0026291804804649314, "grad_norm": 0.147296, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:58.794867+00:00", "epoch": 0, "step": 6333, "train_loss": 3.654975414276123, "perplexity": 38.66657045785059, "lr": 0.0026291804804649314, "grad_norm": 0.17832, "tokens_per_sec": 106614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:59.100729+00:00", "epoch": 0, "step": 6334, "train_loss": 3.6171414852142334, "perplexity": 37.230990231750006, "lr": 0.0026291804804649314, "grad_norm": 0.173892, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:59.406184+00:00", "epoch": 0, "step": 6335, "train_loss": 3.580249786376953, "perplexity": 35.88250268808573, "lr": 0.0026291804804649314, "grad_norm": 0.166517, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:30:59.711514+00:00", "epoch": 0, "step": 6336, "train_loss": 3.603733777999878, "perplexity": 36.73513955387183, "lr": 0.0026291804804649314, "grad_norm": 0.142826, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:00.019275+00:00", "epoch": 0, "step": 6337, "train_loss": 3.6159448623657227, "perplexity": 37.186465423180586, "lr": 0.0026291804804649314, "grad_norm": 0.156789, "tokens_per_sec": 106472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:00.324973+00:00", "epoch": 0, "step": 6338, "train_loss": 3.5375406742095947, "perplexity": 34.382257955027114, "lr": 0.0026291804804649314, "grad_norm": 0.168143, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:00.631197+00:00", "epoch": 0, "step": 6339, "train_loss": 3.561563014984131, "perplexity": 35.2182007236452, "lr": 0.0026291804804649314, "grad_norm": 0.171708, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:00.937807+00:00", "epoch": 0, "step": 6340, "train_loss": 3.5575459003448486, "perplexity": 35.077008955578336, "lr": 0.0026291804804649314, "grad_norm": 0.153708, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:01.243851+00:00", "epoch": 0, "step": 6341, "train_loss": 3.601788282394409, "perplexity": 36.66374097663264, "lr": 0.0026291804804649314, "grad_norm": 0.152377, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:01.549983+00:00", "epoch": 0, "step": 6342, "train_loss": 3.5669314861297607, "perplexity": 35.40777703019038, "lr": 0.0026291804804649314, "grad_norm": 0.154833, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:01.856503+00:00", "epoch": 0, "step": 6343, "train_loss": 3.627547025680542, "perplexity": 37.62042141436438, "lr": 0.0026291804804649314, "grad_norm": 0.152446, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:02.163465+00:00", "epoch": 0, "step": 6344, "train_loss": 3.604464054107666, "perplexity": 36.76197614647281, "lr": 0.0026291804804649314, "grad_norm": 0.151491, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:02.470225+00:00", "epoch": 0, "step": 6345, "train_loss": 3.4576382637023926, "perplexity": 31.741921870877785, "lr": 0.0026291804804649314, "grad_norm": 0.142843, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:02.776741+00:00", "epoch": 0, "step": 6346, "train_loss": 3.691502809524536, "perplexity": 40.105071976734614, "lr": 0.0026291804804649314, "grad_norm": 0.170386, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:03.081927+00:00", "epoch": 0, "step": 6347, "train_loss": 3.510573148727417, "perplexity": 33.46744411097835, "lr": 0.0026291804804649314, "grad_norm": 0.180268, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:03.388355+00:00", "epoch": 0, "step": 6348, "train_loss": 3.5996131896972656, "perplexity": 36.58408060713466, "lr": 0.0026291804804649314, "grad_norm": 0.168186, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:03.694967+00:00", "epoch": 0, "step": 6349, "train_loss": 3.6667730808258057, "perplexity": 39.125447278206366, "lr": 0.0026291804804649314, "grad_norm": 0.162709, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:04.001536+00:00", "epoch": 0, "step": 6350, "train_loss": 3.6500093936920166, "perplexity": 38.475027469892964, "lr": 0.0026291804804649314, "grad_norm": 0.168397, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:04.307168+00:00", "epoch": 0, "step": 6351, "train_loss": 3.5358822345733643, "perplexity": 34.32528431238345, "lr": 0.0026291804804649314, "grad_norm": 0.177173, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:04.613184+00:00", "epoch": 0, "step": 6352, "train_loss": 3.6013197898864746, "perplexity": 36.64656831161878, "lr": 0.0026291804804649314, "grad_norm": 0.185306, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:04.918870+00:00", "epoch": 0, "step": 6353, "train_loss": 3.6065242290496826, "perplexity": 36.837790316960955, "lr": 0.0026291804804649314, "grad_norm": 0.1907, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:05.225654+00:00", "epoch": 0, "step": 6354, "train_loss": 3.566636323928833, "perplexity": 35.397327535016046, "lr": 0.0026291804804649314, "grad_norm": 0.182544, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:05.532692+00:00", "epoch": 0, "step": 6355, "train_loss": 3.5731446743011475, "perplexity": 35.62845706497853, "lr": 0.0026291804804649314, "grad_norm": 0.216899, "tokens_per_sec": 106725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:05.838765+00:00", "epoch": 0, "step": 6356, "train_loss": 3.5061399936676025, "perplexity": 33.31940612190631, "lr": 0.0026291804804649314, "grad_norm": 0.177197, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:06.143985+00:00", "epoch": 0, "step": 6357, "train_loss": 3.5391786098480225, "perplexity": 34.438620026783425, "lr": 0.0026291804804649314, "grad_norm": 0.172111, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:06.449686+00:00", "epoch": 0, "step": 6358, "train_loss": 3.5453431606292725, "perplexity": 34.65157435828514, "lr": 0.0026291804804649314, "grad_norm": 0.170103, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:06.756713+00:00", "epoch": 0, "step": 6359, "train_loss": 3.59029483795166, "perplexity": 36.24476068226128, "lr": 0.0026291804804649314, "grad_norm": 0.158818, "tokens_per_sec": 106727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:07.063808+00:00", "epoch": 0, "step": 6360, "train_loss": 3.5206899642944336, "perplexity": 33.80774655659811, "lr": 0.0026291804804649314, "grad_norm": 0.157989, "tokens_per_sec": 106760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:07.370109+00:00", "epoch": 0, "step": 6361, "train_loss": 3.6308846473693848, "perplexity": 37.74619392259018, "lr": 0.0026291804804649314, "grad_norm": 0.157125, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:07.675771+00:00", "epoch": 0, "step": 6362, "train_loss": 3.532768964767456, "perplexity": 34.21858661669182, "lr": 0.0026291804804649314, "grad_norm": 0.141502, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:07.981256+00:00", "epoch": 0, "step": 6363, "train_loss": 3.6528258323669434, "perplexity": 38.58354276688612, "lr": 0.0026291804804649314, "grad_norm": 0.185194, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:08.287459+00:00", "epoch": 0, "step": 6364, "train_loss": 3.5883476734161377, "perplexity": 36.17425483518139, "lr": 0.0026291804804649314, "grad_norm": 0.151168, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:08.593350+00:00", "epoch": 0, "step": 6365, "train_loss": 3.5311388969421387, "perplexity": 34.162853436392936, "lr": 0.0026291804804649314, "grad_norm": 0.147865, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:08.899736+00:00", "epoch": 0, "step": 6366, "train_loss": 3.608024835586548, "perplexity": 36.89311084270455, "lr": 0.0026291804804649314, "grad_norm": 0.140033, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:09.205200+00:00", "epoch": 0, "step": 6367, "train_loss": 3.6077165603637695, "perplexity": 36.88173936360321, "lr": 0.0026291804804649314, "grad_norm": 0.140226, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:09.511533+00:00", "epoch": 0, "step": 6368, "train_loss": 3.691459894180298, "perplexity": 40.10335089069579, "lr": 0.0026291804804649314, "grad_norm": 0.139164, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:09.817725+00:00", "epoch": 0, "step": 6369, "train_loss": 3.519680976867676, "perplexity": 33.77365216869235, "lr": 0.0026291804804649314, "grad_norm": 0.146994, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:10.123071+00:00", "epoch": 0, "step": 6370, "train_loss": 3.5542778968811035, "perplexity": 34.96256427347673, "lr": 0.0026291804804649314, "grad_norm": 0.164645, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:10.429575+00:00", "epoch": 0, "step": 6371, "train_loss": 3.5978875160217285, "perplexity": 36.52100286374137, "lr": 0.0026291804804649314, "grad_norm": 0.158041, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:10.735613+00:00", "epoch": 0, "step": 6372, "train_loss": 3.541520357131958, "perplexity": 34.51936107235855, "lr": 0.0026291804804649314, "grad_norm": 0.155166, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:11.044455+00:00", "epoch": 0, "step": 6373, "train_loss": 3.614854097366333, "perplexity": 37.145925841843564, "lr": 0.0026291804804649314, "grad_norm": 0.176619, "tokens_per_sec": 106100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:11.351367+00:00", "epoch": 0, "step": 6374, "train_loss": 3.654660701751709, "perplexity": 38.65440351849589, "lr": 0.0026291804804649314, "grad_norm": 0.164756, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:11.657323+00:00", "epoch": 0, "step": 6375, "train_loss": 3.6042823791503906, "perplexity": 36.755298022669336, "lr": 0.0026291804804649314, "grad_norm": 0.1777, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:11.964602+00:00", "epoch": 0, "step": 6376, "train_loss": 3.6440093517303467, "perplexity": 38.24486686772372, "lr": 0.0026291804804649314, "grad_norm": 0.1819, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:12.270840+00:00", "epoch": 0, "step": 6377, "train_loss": 3.5286803245544434, "perplexity": 34.07896475370476, "lr": 0.0026291804804649314, "grad_norm": 0.169226, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:12.577687+00:00", "epoch": 0, "step": 6378, "train_loss": 3.5604238510131836, "perplexity": 35.178104260818216, "lr": 0.0026291804804649314, "grad_norm": 0.15048, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:12.884373+00:00", "epoch": 0, "step": 6379, "train_loss": 3.58085298538208, "perplexity": 35.904153507228465, "lr": 0.0026291804804649314, "grad_norm": 0.156933, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:13.190927+00:00", "epoch": 0, "step": 6380, "train_loss": 3.510164737701416, "perplexity": 33.45377842858899, "lr": 0.0026291804804649314, "grad_norm": 0.162891, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:13.496041+00:00", "epoch": 0, "step": 6381, "train_loss": 3.592000722885132, "perplexity": 36.30664284034035, "lr": 0.0026291804804649314, "grad_norm": 0.155195, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:13.801464+00:00", "epoch": 0, "step": 6382, "train_loss": 3.5164544582366943, "perplexity": 33.664856460807094, "lr": 0.0026291804804649314, "grad_norm": 0.159109, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:14.107871+00:00", "epoch": 0, "step": 6383, "train_loss": 3.519923210144043, "perplexity": 33.78183426205969, "lr": 0.0026291804804649314, "grad_norm": 0.162461, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:14.415074+00:00", "epoch": 0, "step": 6384, "train_loss": 3.512669324874878, "perplexity": 33.53767134765487, "lr": 0.0026291804804649314, "grad_norm": 0.158982, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:14.720364+00:00", "epoch": 0, "step": 6385, "train_loss": 3.5640153884887695, "perplexity": 35.30467489612668, "lr": 0.0026291804804649314, "grad_norm": 0.155617, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:15.025665+00:00", "epoch": 0, "step": 6386, "train_loss": 3.6236419677734375, "perplexity": 37.473797963112986, "lr": 0.0026291804804649314, "grad_norm": 0.171948, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:15.331361+00:00", "epoch": 0, "step": 6387, "train_loss": 3.590496301651001, "perplexity": 36.25206342142378, "lr": 0.0026291804804649314, "grad_norm": 0.177003, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:15.638285+00:00", "epoch": 0, "step": 6388, "train_loss": 3.6973860263824463, "perplexity": 40.34171423858844, "lr": 0.0026291804804649314, "grad_norm": 0.129743, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:15.946658+00:00", "epoch": 0, "step": 6389, "train_loss": 3.6102328300476074, "perplexity": 36.97466062469874, "lr": 0.0026291804804649314, "grad_norm": 0.18876, "tokens_per_sec": 106265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:16.252973+00:00", "epoch": 0, "step": 6390, "train_loss": 3.5431900024414062, "perplexity": 34.577044303525966, "lr": 0.0026291804804649314, "grad_norm": 0.155591, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:16.558148+00:00", "epoch": 0, "step": 6391, "train_loss": 3.561918020248413, "perplexity": 35.23070558981597, "lr": 0.0026291804804649314, "grad_norm": 0.18181, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:16.863657+00:00", "epoch": 0, "step": 6392, "train_loss": 3.6170825958251953, "perplexity": 37.2287977860385, "lr": 0.0026291804804649314, "grad_norm": 0.170185, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:17.169535+00:00", "epoch": 0, "step": 6393, "train_loss": 3.6231422424316406, "perplexity": 37.45507603491804, "lr": 0.0026291804804649314, "grad_norm": 0.153358, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:17.477028+00:00", "epoch": 0, "step": 6394, "train_loss": 3.5357375144958496, "perplexity": 34.32031711401239, "lr": 0.0026291804804649314, "grad_norm": 0.200144, "tokens_per_sec": 106565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:17.783843+00:00", "epoch": 0, "step": 6395, "train_loss": 3.5991594791412354, "perplexity": 36.56748578848742, "lr": 0.0026291804804649314, "grad_norm": 0.164777, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:18.089023+00:00", "epoch": 0, "step": 6396, "train_loss": 3.5599615573883057, "perplexity": 35.16184540595519, "lr": 0.0026291804804649314, "grad_norm": 0.178458, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:18.395533+00:00", "epoch": 0, "step": 6397, "train_loss": 3.540759325027466, "perplexity": 34.49310072410741, "lr": 0.0026291804804649314, "grad_norm": 0.163585, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:18.701716+00:00", "epoch": 0, "step": 6398, "train_loss": 3.5156168937683105, "perplexity": 33.636671778094346, "lr": 0.0026291804804649314, "grad_norm": 0.177368, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:19.008177+00:00", "epoch": 0, "step": 6399, "train_loss": 3.5937414169311523, "perplexity": 36.36989663414299, "lr": 0.0026291804804649314, "grad_norm": 0.16301, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:19.316772+00:00", "epoch": 0, "step": 6400, "train_loss": 3.6210529804229736, "perplexity": 37.376904256624904, "lr": 0.0026291804804649314, "grad_norm": 0.200193, "tokens_per_sec": 106242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:19.624120+00:00", "epoch": 0, "step": 6401, "train_loss": 3.656942129135132, "perplexity": 38.74269140608083, "lr": 0.0026291804804649314, "grad_norm": 0.167678, "tokens_per_sec": 106559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:19.930860+00:00", "epoch": 0, "step": 6402, "train_loss": 3.559237003326416, "perplexity": 35.13637797542204, "lr": 0.0026291804804649314, "grad_norm": 0.215019, "tokens_per_sec": 106827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:20.237714+00:00", "epoch": 0, "step": 6403, "train_loss": 3.6228229999542236, "perplexity": 37.443120692080974, "lr": 0.0026291804804649314, "grad_norm": 0.182604, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:20.544750+00:00", "epoch": 0, "step": 6404, "train_loss": 3.6383352279663086, "perplexity": 38.028475256740826, "lr": 0.0026291804804649314, "grad_norm": 0.170046, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:20.851335+00:00", "epoch": 0, "step": 6405, "train_loss": 3.5232656002044678, "perplexity": 33.89493523761543, "lr": 0.0026291804804649314, "grad_norm": 0.166603, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:21.158548+00:00", "epoch": 0, "step": 6406, "train_loss": 3.5840816497802734, "perplexity": 36.02026330833173, "lr": 0.0026291804804649314, "grad_norm": 0.153447, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:21.464828+00:00", "epoch": 0, "step": 6407, "train_loss": 3.6025378704071045, "perplexity": 36.69123398029496, "lr": 0.0026291804804649314, "grad_norm": 0.162608, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:21.771425+00:00", "epoch": 0, "step": 6408, "train_loss": 3.6288938522338867, "perplexity": 37.67112373282342, "lr": 0.0026291804804649314, "grad_norm": 0.170118, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:22.077223+00:00", "epoch": 0, "step": 6409, "train_loss": 3.6893882751464844, "perplexity": 40.02035802015712, "lr": 0.0026291804804649314, "grad_norm": 0.17531, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:22.384661+00:00", "epoch": 0, "step": 6410, "train_loss": 3.4927847385406494, "perplexity": 32.87737524020664, "lr": 0.0026291804804649314, "grad_norm": 0.182258, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:22.691647+00:00", "epoch": 0, "step": 6411, "train_loss": 3.530296564102173, "perplexity": 34.13408905932699, "lr": 0.0026291804804649314, "grad_norm": 0.160109, "tokens_per_sec": 106688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:22.996732+00:00", "epoch": 0, "step": 6412, "train_loss": 3.625537633895874, "perplexity": 37.54490314694141, "lr": 0.0026291804804649314, "grad_norm": 0.158716, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:23.302468+00:00", "epoch": 0, "step": 6413, "train_loss": 3.443962812423706, "perplexity": 31.310791435546165, "lr": 0.0026291804804649314, "grad_norm": 0.148847, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:23.608054+00:00", "epoch": 0, "step": 6414, "train_loss": 3.586737632751465, "perplexity": 36.11605967474118, "lr": 0.0026291804804649314, "grad_norm": 0.169235, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:23.912969+00:00", "epoch": 0, "step": 6415, "train_loss": 3.5877957344055176, "perplexity": 36.154294361746445, "lr": 0.0026291804804649314, "grad_norm": 0.141842, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:24.219560+00:00", "epoch": 0, "step": 6416, "train_loss": 3.482964038848877, "perplexity": 32.55607668197448, "lr": 0.0026291804804649314, "grad_norm": 0.147608, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:24.526666+00:00", "epoch": 0, "step": 6417, "train_loss": 3.6191322803497314, "perplexity": 37.30518333312188, "lr": 0.0026291804804649314, "grad_norm": 0.14427, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:24.833138+00:00", "epoch": 0, "step": 6418, "train_loss": 3.639984369277954, "perplexity": 38.09124132714021, "lr": 0.0026291804804649314, "grad_norm": 0.133135, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:25.138767+00:00", "epoch": 0, "step": 6419, "train_loss": 3.54734468460083, "perplexity": 34.720999770151415, "lr": 0.0026291804804649314, "grad_norm": 0.157944, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:25.448777+00:00", "epoch": 0, "step": 6420, "train_loss": 3.650186777114868, "perplexity": 38.481852907301715, "lr": 0.0026291804804649314, "grad_norm": 0.150854, "tokens_per_sec": 105755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:25.756116+00:00", "epoch": 0, "step": 6421, "train_loss": 3.638890266418457, "perplexity": 38.04958838154093, "lr": 0.0026291804804649314, "grad_norm": 0.153881, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:26.062650+00:00", "epoch": 0, "step": 6422, "train_loss": 3.5737462043762207, "perplexity": 35.64989510059673, "lr": 0.0026291804804649314, "grad_norm": 0.154914, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:26.369706+00:00", "epoch": 0, "step": 6423, "train_loss": 3.582512140274048, "perplexity": 35.96377350489053, "lr": 0.0026291804804649314, "grad_norm": 0.176104, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:26.675528+00:00", "epoch": 0, "step": 6424, "train_loss": 3.5806164741516113, "perplexity": 35.89566277581977, "lr": 0.0026291804804649314, "grad_norm": 0.152996, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:26.981443+00:00", "epoch": 0, "step": 6425, "train_loss": 3.7368054389953613, "perplexity": 41.96372014830974, "lr": 0.0026291804804649314, "grad_norm": 0.145783, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:27.287522+00:00", "epoch": 0, "step": 6426, "train_loss": 3.545821189880371, "perplexity": 34.66814278420563, "lr": 0.0026291804804649314, "grad_norm": 0.159927, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:27.594182+00:00", "epoch": 0, "step": 6427, "train_loss": 3.6630778312683105, "perplexity": 38.98113578416986, "lr": 0.0026291804804649314, "grad_norm": 0.139159, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:27.900448+00:00", "epoch": 0, "step": 6428, "train_loss": 3.696120262145996, "perplexity": 40.29068344275959, "lr": 0.0026291804804649314, "grad_norm": 0.159604, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:28.207906+00:00", "epoch": 0, "step": 6429, "train_loss": 3.506899118423462, "perplexity": 33.34470931086744, "lr": 0.0026291804804649314, "grad_norm": 0.155095, "tokens_per_sec": 106577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:28.514030+00:00", "epoch": 0, "step": 6430, "train_loss": 3.6319799423217773, "perplexity": 37.78755978804199, "lr": 0.0026291804804649314, "grad_norm": 0.142873, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:28.820302+00:00", "epoch": 0, "step": 6431, "train_loss": 3.5760819911956787, "perplexity": 35.73326298258382, "lr": 0.0026291804804649314, "grad_norm": 0.179086, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:29.126728+00:00", "epoch": 0, "step": 6432, "train_loss": 3.514315128326416, "perplexity": 33.59291320906916, "lr": 0.0026291804804649314, "grad_norm": 0.155447, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:29.433397+00:00", "epoch": 0, "step": 6433, "train_loss": 3.5251705646514893, "perplexity": 33.95956542373592, "lr": 0.0026291804804649314, "grad_norm": 0.164955, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:29.739162+00:00", "epoch": 0, "step": 6434, "train_loss": 3.5620052814483643, "perplexity": 35.23377999759717, "lr": 0.0026291804804649314, "grad_norm": 0.182305, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:30.045025+00:00", "epoch": 0, "step": 6435, "train_loss": 3.637099027633667, "perplexity": 37.98149348840051, "lr": 0.0026291804804649314, "grad_norm": 0.195328, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:30.352174+00:00", "epoch": 0, "step": 6436, "train_loss": 3.523650884628296, "perplexity": 33.90799694428363, "lr": 0.0026291804804649314, "grad_norm": 0.228722, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:30.658285+00:00", "epoch": 0, "step": 6437, "train_loss": 3.5820345878601074, "perplexity": 35.94660301826885, "lr": 0.0026291804804649314, "grad_norm": 0.207459, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:30.964876+00:00", "epoch": 0, "step": 6438, "train_loss": 3.4836299419403076, "perplexity": 32.57776309380484, "lr": 0.0026291804804649314, "grad_norm": 0.157145, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:31.271677+00:00", "epoch": 0, "step": 6439, "train_loss": 3.6148979663848877, "perplexity": 37.14755543289757, "lr": 0.0026291804804649314, "grad_norm": 0.164128, "tokens_per_sec": 106806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:31.579062+00:00", "epoch": 0, "step": 6440, "train_loss": 3.5982413291931152, "perplexity": 36.533926761774666, "lr": 0.0026291804804649314, "grad_norm": 0.172055, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:31.885219+00:00", "epoch": 0, "step": 6441, "train_loss": 3.54740047454834, "perplexity": 34.72293690694196, "lr": 0.0026291804804649314, "grad_norm": 0.176452, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:32.192488+00:00", "epoch": 0, "step": 6442, "train_loss": 3.580753803253174, "perplexity": 35.90059263343748, "lr": 0.0026291804804649314, "grad_norm": 0.180718, "tokens_per_sec": 106644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:32.499415+00:00", "epoch": 0, "step": 6443, "train_loss": 3.6318767070770264, "perplexity": 37.78365898141256, "lr": 0.0026291804804649314, "grad_norm": 0.16827, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:32.806591+00:00", "epoch": 0, "step": 6444, "train_loss": 3.5917322635650635, "perplexity": 36.29689729188966, "lr": 0.0026291804804649314, "grad_norm": 0.167093, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:33.113995+00:00", "epoch": 0, "step": 6445, "train_loss": 3.5062801837921143, "perplexity": 33.324077501032164, "lr": 0.0026291804804649314, "grad_norm": 0.162405, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:33.421291+00:00", "epoch": 0, "step": 6446, "train_loss": 3.5339109897613525, "perplexity": 34.25768742066193, "lr": 0.0026291804804649314, "grad_norm": 0.189873, "tokens_per_sec": 106634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:33.727239+00:00", "epoch": 0, "step": 6447, "train_loss": 3.5604584217071533, "perplexity": 35.179320413316546, "lr": 0.0026291804804649314, "grad_norm": 0.163736, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:34.034354+00:00", "epoch": 0, "step": 6448, "train_loss": 3.527759313583374, "perplexity": 34.04759210277944, "lr": 0.0026291804804649314, "grad_norm": 0.154183, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:34.339966+00:00", "epoch": 0, "step": 6449, "train_loss": 3.566084861755371, "perplexity": 35.37781262919959, "lr": 0.0026291804804649314, "grad_norm": 0.160531, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:34.646435+00:00", "epoch": 0, "step": 6450, "train_loss": 3.6105191707611084, "perplexity": 36.98524949134294, "lr": 0.0026291804804649314, "grad_norm": 0.167738, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:34.952985+00:00", "epoch": 0, "step": 6451, "train_loss": 3.5105390548706055, "perplexity": 33.46630309618189, "lr": 0.0026291804804649314, "grad_norm": 0.153915, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:35.261093+00:00", "epoch": 0, "step": 6452, "train_loss": 3.585991859436035, "perplexity": 36.089135322157034, "lr": 0.0026291804804649314, "grad_norm": 0.14407, "tokens_per_sec": 106353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:35.568088+00:00", "epoch": 0, "step": 6453, "train_loss": 3.4781079292297363, "perplexity": 32.398364049681895, "lr": 0.0026291804804649314, "grad_norm": 0.153091, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:35.874475+00:00", "epoch": 0, "step": 6454, "train_loss": 3.602771282196045, "perplexity": 36.69979914642322, "lr": 0.0026291804804649314, "grad_norm": 0.157387, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:36.180233+00:00", "epoch": 0, "step": 6455, "train_loss": 3.6200335025787354, "perplexity": 37.3388187478139, "lr": 0.0026291804804649314, "grad_norm": 0.156978, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:36.485957+00:00", "epoch": 0, "step": 6456, "train_loss": 3.589041233062744, "perplexity": 36.19935254095305, "lr": 0.0026291804804649314, "grad_norm": 0.16078, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:36.792864+00:00", "epoch": 0, "step": 6457, "train_loss": 3.5664408206939697, "perplexity": 35.390407919403096, "lr": 0.0026291804804649314, "grad_norm": 0.154552, "tokens_per_sec": 106770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:37.098963+00:00", "epoch": 0, "step": 6458, "train_loss": 3.538133144378662, "perplexity": 34.402634452848154, "lr": 0.0026291804804649314, "grad_norm": 0.158917, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:37.406647+00:00", "epoch": 0, "step": 6459, "train_loss": 3.5570011138916016, "perplexity": 35.05790468062651, "lr": 0.0026291804804649314, "grad_norm": 0.152274, "tokens_per_sec": 106503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:37.712198+00:00", "epoch": 0, "step": 6460, "train_loss": 3.5629634857177734, "perplexity": 35.267557336228926, "lr": 0.0026291804804649314, "grad_norm": 0.17924, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:38.018345+00:00", "epoch": 0, "step": 6461, "train_loss": 3.6535701751708984, "perplexity": 38.61227284048103, "lr": 0.0026291804804649314, "grad_norm": 0.179295, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:38.326122+00:00", "epoch": 0, "step": 6462, "train_loss": 3.6331565380096436, "perplexity": 37.83204663432713, "lr": 0.0026291804804649314, "grad_norm": 0.152583, "tokens_per_sec": 106466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:38.633041+00:00", "epoch": 0, "step": 6463, "train_loss": 3.562729597091675, "perplexity": 35.259309620258776, "lr": 0.0026291804804649314, "grad_norm": 0.192673, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:38.938154+00:00", "epoch": 0, "step": 6464, "train_loss": 3.5548319816589355, "perplexity": 34.98194186605342, "lr": 0.0026291804804649314, "grad_norm": 0.165316, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:39.244660+00:00", "epoch": 0, "step": 6465, "train_loss": 3.6710033416748047, "perplexity": 39.29130869721829, "lr": 0.0026291804804649314, "grad_norm": 0.176215, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:39.549901+00:00", "epoch": 0, "step": 6466, "train_loss": 3.4704110622406006, "perplexity": 32.149955361587644, "lr": 0.0026291804804649314, "grad_norm": 0.159451, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:39.856107+00:00", "epoch": 0, "step": 6467, "train_loss": 3.617999792098999, "perplexity": 37.26295956477925, "lr": 0.0026291804804649314, "grad_norm": 0.15691, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:40.163802+00:00", "epoch": 0, "step": 6468, "train_loss": 3.608684539794922, "perplexity": 36.917457413071965, "lr": 0.0026291804804649314, "grad_norm": 0.172177, "tokens_per_sec": 106496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:40.469249+00:00", "epoch": 0, "step": 6469, "train_loss": 3.5597753524780273, "perplexity": 35.1552987072188, "lr": 0.0026291804804649314, "grad_norm": 0.150438, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:40.775398+00:00", "epoch": 0, "step": 6470, "train_loss": 3.5757436752319336, "perplexity": 35.721175894022906, "lr": 0.0026291804804649314, "grad_norm": 0.153138, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:41.080840+00:00", "epoch": 0, "step": 6471, "train_loss": 3.5451862812042236, "perplexity": 34.64613866560809, "lr": 0.0026291804804649314, "grad_norm": 0.162238, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:41.387278+00:00", "epoch": 0, "step": 6472, "train_loss": 3.4923834800720215, "perplexity": 32.86418556137738, "lr": 0.0026291804804649314, "grad_norm": 0.163213, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:41.693714+00:00", "epoch": 0, "step": 6473, "train_loss": 3.5437588691711426, "perplexity": 34.596719629431426, "lr": 0.0026291804804649314, "grad_norm": 0.151234, "tokens_per_sec": 106931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:42.000185+00:00", "epoch": 0, "step": 6474, "train_loss": 3.4940617084503174, "perplexity": 32.91938547629226, "lr": 0.0026291804804649314, "grad_norm": 0.17169, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:42.306092+00:00", "epoch": 0, "step": 6475, "train_loss": 3.6150217056274414, "perplexity": 37.152152327671836, "lr": 0.0026291804804649314, "grad_norm": 0.17899, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:42.612333+00:00", "epoch": 0, "step": 6476, "train_loss": 3.449136734008789, "perplexity": 31.473210825471075, "lr": 0.0026291804804649314, "grad_norm": 0.155585, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:42.917246+00:00", "epoch": 0, "step": 6477, "train_loss": 3.5037834644317627, "perplexity": 33.240980409793934, "lr": 0.0026291804804649314, "grad_norm": 0.15149, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:43.223627+00:00", "epoch": 0, "step": 6478, "train_loss": 3.492419958114624, "perplexity": 32.86538440440398, "lr": 0.0026291804804649314, "grad_norm": 0.159006, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:43.529188+00:00", "epoch": 0, "step": 6479, "train_loss": 3.6575443744659424, "perplexity": 38.766031038470054, "lr": 0.0026291804804649314, "grad_norm": 0.13941, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:43.837113+00:00", "epoch": 0, "step": 6480, "train_loss": 3.53190279006958, "perplexity": 34.18896017545655, "lr": 0.0026291804804649314, "grad_norm": 0.158695, "tokens_per_sec": 106470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:44.143649+00:00", "epoch": 0, "step": 6481, "train_loss": 3.554783821105957, "perplexity": 34.98025715695747, "lr": 0.0026291804804649314, "grad_norm": 0.158483, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:44.449454+00:00", "epoch": 0, "step": 6482, "train_loss": 3.56623911857605, "perplexity": 35.38327031903046, "lr": 0.0026291804804649314, "grad_norm": 0.151801, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:44.755669+00:00", "epoch": 0, "step": 6483, "train_loss": 3.5872340202331543, "perplexity": 36.133991684896145, "lr": 0.0026291804804649314, "grad_norm": 0.188306, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:45.062130+00:00", "epoch": 0, "step": 6484, "train_loss": 3.5969700813293457, "perplexity": 36.48751259363042, "lr": 0.0026291804804649314, "grad_norm": 0.155633, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:45.369855+00:00", "epoch": 0, "step": 6485, "train_loss": 3.5119001865386963, "perplexity": 33.511886156375844, "lr": 0.0026291804804649314, "grad_norm": 0.187298, "tokens_per_sec": 106484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:45.676037+00:00", "epoch": 0, "step": 6486, "train_loss": 3.5647995471954346, "perplexity": 35.33237022167862, "lr": 0.0026291804804649314, "grad_norm": 0.212809, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:45.982670+00:00", "epoch": 0, "step": 6487, "train_loss": 3.538820505142212, "perplexity": 34.42628960281341, "lr": 0.0026291804804649314, "grad_norm": 0.16492, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:46.288448+00:00", "epoch": 0, "step": 6488, "train_loss": 3.51733136177063, "perplexity": 33.69439023965201, "lr": 0.0026291804804649314, "grad_norm": 0.15321, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:46.594981+00:00", "epoch": 0, "step": 6489, "train_loss": 3.581936836242676, "perplexity": 35.94308935141875, "lr": 0.0026291804804649314, "grad_norm": 0.151608, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:46.902352+00:00", "epoch": 0, "step": 6490, "train_loss": 3.57993483543396, "perplexity": 35.871203239506556, "lr": 0.0026291804804649314, "grad_norm": 0.141162, "tokens_per_sec": 106665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:47.209537+00:00", "epoch": 0, "step": 6491, "train_loss": 3.631716728210449, "perplexity": 37.777614877950874, "lr": 0.0026291804804649314, "grad_norm": 0.142696, "tokens_per_sec": 106613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:47.516387+00:00", "epoch": 0, "step": 6492, "train_loss": 3.5583503246307373, "perplexity": 35.105237105641066, "lr": 0.0026291804804649314, "grad_norm": 0.138253, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:47.822910+00:00", "epoch": 0, "step": 6493, "train_loss": 3.614393711090088, "perplexity": 37.128828303405584, "lr": 0.0026291804804649314, "grad_norm": 0.147188, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:48.130055+00:00", "epoch": 0, "step": 6494, "train_loss": 3.553570032119751, "perplexity": 34.93782426358466, "lr": 0.0026291804804649314, "grad_norm": 0.152779, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:48.437109+00:00", "epoch": 0, "step": 6495, "train_loss": 3.563702344894409, "perplexity": 35.29362472348259, "lr": 0.0026291804804649314, "grad_norm": 0.153966, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:48.743221+00:00", "epoch": 0, "step": 6496, "train_loss": 3.6234989166259766, "perplexity": 37.468437676721344, "lr": 0.0026291804804649314, "grad_norm": 0.143121, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:49.048731+00:00", "epoch": 0, "step": 6497, "train_loss": 3.5842084884643555, "perplexity": 36.02483236089022, "lr": 0.0026291804804649314, "grad_norm": 0.15124, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:49.353447+00:00", "epoch": 0, "step": 6498, "train_loss": 3.4647891521453857, "perplexity": 31.969718315925082, "lr": 0.0026291804804649314, "grad_norm": 0.154282, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:49.658799+00:00", "epoch": 0, "step": 6499, "train_loss": 3.554572820663452, "perplexity": 34.972877085844935, "lr": 0.0026291804804649314, "grad_norm": 0.177659, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:49.967407+00:00", "epoch": 0, "step": 6500, "train_loss": 3.568650960922241, "perplexity": 35.468712183468206, "lr": 0.0026291804804649314, "grad_norm": 0.127775, "tokens_per_sec": 106246} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:31:53.040154+00:00", "step": 6500, "epoch": 0, "val_loss": 3.528679847717285, "val_ppl": 34.078948503591924, "eval_train_loss": 3.568650960922241, "eval_train_ppl": 35.468712183468206} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:31:53.991205+00:00", "step": 6500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5287_epoch_0000_step_0006500.pt", "val_loss": 3.528679847717285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:55.093554+00:00", "epoch": 0, "step": 6501, "train_loss": 3.535290241241455, "perplexity": 34.30496998652063, "lr": 0.0026291804804649314, "grad_norm": 0.160788, "tokens_per_sec": 6392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:55.398451+00:00", "epoch": 0, "step": 6502, "train_loss": 3.6277620792388916, "perplexity": 37.62851268985376, "lr": 0.0026291804804649314, "grad_norm": 0.17103, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:55.703553+00:00", "epoch": 0, "step": 6503, "train_loss": 3.632485866546631, "perplexity": 37.8066822667916, "lr": 0.0026291804804649314, "grad_norm": 0.137274, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:56.009457+00:00", "epoch": 0, "step": 6504, "train_loss": 3.5360875129699707, "perplexity": 34.33233127498001, "lr": 0.0026291804804649314, "grad_norm": 0.174197, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:56.315682+00:00", "epoch": 0, "step": 6505, "train_loss": 3.685481309890747, "perplexity": 39.864304917379904, "lr": 0.0026291804804649314, "grad_norm": 0.168443, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:56.621824+00:00", "epoch": 0, "step": 6506, "train_loss": 3.5681567192077637, "perplexity": 35.451186397692304, "lr": 0.0026291804804649314, "grad_norm": 0.204318, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:56.927009+00:00", "epoch": 0, "step": 6507, "train_loss": 3.5288617610931396, "perplexity": 34.08514848407238, "lr": 0.0026291804804649314, "grad_norm": 0.186186, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:57.232386+00:00", "epoch": 0, "step": 6508, "train_loss": 3.560153007507324, "perplexity": 35.16857778988036, "lr": 0.0026291804804649314, "grad_norm": 0.179723, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:57.538127+00:00", "epoch": 0, "step": 6509, "train_loss": 3.569227695465088, "perplexity": 35.48917411495888, "lr": 0.0026291804804649314, "grad_norm": 0.160965, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:57.843080+00:00", "epoch": 0, "step": 6510, "train_loss": 3.550344705581665, "perplexity": 34.82531990098687, "lr": 0.0026291804804649314, "grad_norm": 0.174647, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:58.148084+00:00", "epoch": 0, "step": 6511, "train_loss": 3.60884428024292, "perplexity": 36.92335509529474, "lr": 0.0026291804804649314, "grad_norm": 0.174179, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:58.454431+00:00", "epoch": 0, "step": 6512, "train_loss": 3.552877426147461, "perplexity": 34.913634495796686, "lr": 0.0026291804804649314, "grad_norm": 0.175637, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:58.760772+00:00", "epoch": 0, "step": 6513, "train_loss": 3.558558464050293, "perplexity": 35.11254464978321, "lr": 0.0026291804804649314, "grad_norm": 0.188087, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:59.067255+00:00", "epoch": 0, "step": 6514, "train_loss": 3.518197536468506, "perplexity": 33.7235881113413, "lr": 0.0026291804804649314, "grad_norm": 0.177107, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:59.372213+00:00", "epoch": 0, "step": 6515, "train_loss": 3.6221697330474854, "perplexity": 37.41866832827747, "lr": 0.0026291804804649314, "grad_norm": 0.176926, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:59.677248+00:00", "epoch": 0, "step": 6516, "train_loss": 3.5619490146636963, "perplexity": 35.231797561858166, "lr": 0.0026291804804649314, "grad_norm": 0.162585, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:31:59.983475+00:00", "epoch": 0, "step": 6517, "train_loss": 3.6250016689300537, "perplexity": 37.524785785791494, "lr": 0.0026291804804649314, "grad_norm": 0.185287, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:00.290300+00:00", "epoch": 0, "step": 6518, "train_loss": 3.649567127227783, "perplexity": 38.45801501782839, "lr": 0.0026291804804649314, "grad_norm": 0.182882, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:00.596413+00:00", "epoch": 0, "step": 6519, "train_loss": 3.503220796585083, "perplexity": 33.222282039904314, "lr": 0.0026291804804649314, "grad_norm": 0.161432, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:00.903429+00:00", "epoch": 0, "step": 6520, "train_loss": 3.5106422901153564, "perplexity": 33.4697581765129, "lr": 0.0026291804804649314, "grad_norm": 0.175524, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:01.208107+00:00", "epoch": 0, "step": 6521, "train_loss": 3.618934392929077, "perplexity": 37.29780183699176, "lr": 0.0026291804804649314, "grad_norm": 0.151393, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:01.513456+00:00", "epoch": 0, "step": 6522, "train_loss": 3.5229175090789795, "perplexity": 33.88313876470309, "lr": 0.0026291804804649314, "grad_norm": 0.172646, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:01.819123+00:00", "epoch": 0, "step": 6523, "train_loss": 3.496154546737671, "perplexity": 32.988352569927436, "lr": 0.0026291804804649314, "grad_norm": 0.179857, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:02.124942+00:00", "epoch": 0, "step": 6524, "train_loss": 3.5516605377197266, "perplexity": 34.87117433788565, "lr": 0.0026291804804649314, "grad_norm": 0.146306, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:02.431400+00:00", "epoch": 0, "step": 6525, "train_loss": 3.5432565212249756, "perplexity": 34.57934440295152, "lr": 0.0026291804804649314, "grad_norm": 0.167487, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:02.737380+00:00", "epoch": 0, "step": 6526, "train_loss": 3.5807156562805176, "perplexity": 35.89922316063274, "lr": 0.0026291804804649314, "grad_norm": 0.151885, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:03.042061+00:00", "epoch": 0, "step": 6527, "train_loss": 3.484849452972412, "perplexity": 32.617516270090775, "lr": 0.0026291804804649314, "grad_norm": 0.164346, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:03.347837+00:00", "epoch": 0, "step": 6528, "train_loss": 3.490596294403076, "perplexity": 32.805503613308176, "lr": 0.0026291804804649314, "grad_norm": 0.173963, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:03.652962+00:00", "epoch": 0, "step": 6529, "train_loss": 3.5483222007751465, "perplexity": 34.75495670303681, "lr": 0.0026291804804649314, "grad_norm": 0.179952, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:03.958309+00:00", "epoch": 0, "step": 6530, "train_loss": 3.6269187927246094, "perplexity": 37.59679444821228, "lr": 0.0026291804804649314, "grad_norm": 0.185572, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:04.275070+00:00", "epoch": 0, "step": 6531, "train_loss": 3.626138687133789, "perplexity": 37.56747641573366, "lr": 0.0026291804804649314, "grad_norm": 0.189391, "tokens_per_sec": 103393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:04.579429+00:00", "epoch": 0, "step": 6532, "train_loss": 3.4674429893493652, "perplexity": 32.05467342244329, "lr": 0.0026291804804649314, "grad_norm": 0.202947, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:04.884275+00:00", "epoch": 0, "step": 6533, "train_loss": 3.564626932144165, "perplexity": 35.32627184913192, "lr": 0.0026291804804649314, "grad_norm": 0.167836, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:05.190171+00:00", "epoch": 0, "step": 6534, "train_loss": 3.623788595199585, "perplexity": 37.47929305251185, "lr": 0.0026291804804649314, "grad_norm": 0.168245, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:05.494988+00:00", "epoch": 0, "step": 6535, "train_loss": 3.5466067790985107, "perplexity": 34.69538840392115, "lr": 0.0026291804804649314, "grad_norm": 0.174075, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:05.800542+00:00", "epoch": 0, "step": 6536, "train_loss": 3.5371086597442627, "perplexity": 34.367407530269894, "lr": 0.0026291804804649314, "grad_norm": 0.158978, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:06.106660+00:00", "epoch": 0, "step": 6537, "train_loss": 3.536863327026367, "perplexity": 34.3589771149441, "lr": 0.0026291804804649314, "grad_norm": 0.160071, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:06.412474+00:00", "epoch": 0, "step": 6538, "train_loss": 3.593780517578125, "perplexity": 36.37131874843433, "lr": 0.0026291804804649314, "grad_norm": 0.157004, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:06.718893+00:00", "epoch": 0, "step": 6539, "train_loss": 3.4853343963623047, "perplexity": 32.633337754952855, "lr": 0.0026291804804649314, "grad_norm": 0.164531, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:07.026564+00:00", "epoch": 0, "step": 6540, "train_loss": 3.524939775466919, "perplexity": 33.951728827659004, "lr": 0.0026291804804649314, "grad_norm": 0.143815, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:07.333562+00:00", "epoch": 0, "step": 6541, "train_loss": 3.522289514541626, "perplexity": 33.86186701861973, "lr": 0.0026291804804649314, "grad_norm": 0.155629, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:07.638955+00:00", "epoch": 0, "step": 6542, "train_loss": 3.5882656574249268, "perplexity": 36.17128808947672, "lr": 0.0026291804804649314, "grad_norm": 0.149766, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:07.944246+00:00", "epoch": 0, "step": 6543, "train_loss": 3.7007718086242676, "perplexity": 40.47853398849888, "lr": 0.0026291804804649314, "grad_norm": 0.161768, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:08.251103+00:00", "epoch": 0, "step": 6544, "train_loss": 3.5779058933258057, "perplexity": 35.798496428680885, "lr": 0.0026291804804649314, "grad_norm": 0.155507, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:08.558047+00:00", "epoch": 0, "step": 6545, "train_loss": 3.5388782024383545, "perplexity": 34.428275963942994, "lr": 0.0026291804804649314, "grad_norm": 0.161752, "tokens_per_sec": 106755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:08.863570+00:00", "epoch": 0, "step": 6546, "train_loss": 3.6050665378570557, "perplexity": 36.78413131309026, "lr": 0.0026291804804649314, "grad_norm": 0.139532, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:09.169992+00:00", "epoch": 0, "step": 6547, "train_loss": 3.6727499961853027, "perplexity": 39.35999700868932, "lr": 0.0026291804804649314, "grad_norm": 0.160948, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:09.475926+00:00", "epoch": 0, "step": 6548, "train_loss": 3.5255849361419678, "perplexity": 33.973640215371354, "lr": 0.0026291804804649314, "grad_norm": 0.177384, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:09.781475+00:00", "epoch": 0, "step": 6549, "train_loss": 3.545441150665283, "perplexity": 34.6549700336728, "lr": 0.0026291804804649314, "grad_norm": 0.141696, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:10.087209+00:00", "epoch": 0, "step": 6550, "train_loss": 3.546595573425293, "perplexity": 34.69499962091483, "lr": 0.0026291804804649314, "grad_norm": 0.14746, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:10.393321+00:00", "epoch": 0, "step": 6551, "train_loss": 3.5450828075408936, "perplexity": 34.642553888188345, "lr": 0.0026291804804649314, "grad_norm": 0.167502, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:10.699316+00:00", "epoch": 0, "step": 6552, "train_loss": 3.570991039276123, "perplexity": 35.55180893762706, "lr": 0.0026291804804649314, "grad_norm": 0.170856, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:11.005819+00:00", "epoch": 0, "step": 6553, "train_loss": 3.55729603767395, "perplexity": 35.06824561529496, "lr": 0.0026291804804649314, "grad_norm": 0.148723, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:11.312028+00:00", "epoch": 0, "step": 6554, "train_loss": 3.70861554145813, "perplexity": 40.797285260279295, "lr": 0.0026291804804649314, "grad_norm": 0.150968, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:11.618259+00:00", "epoch": 0, "step": 6555, "train_loss": 3.6897783279418945, "perplexity": 40.03597111744437, "lr": 0.0026291804804649314, "grad_norm": 0.153823, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:11.926201+00:00", "epoch": 0, "step": 6556, "train_loss": 3.668076753616333, "perplexity": 39.17648732176937, "lr": 0.0026291804804649314, "grad_norm": 0.148987, "tokens_per_sec": 106410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:12.234497+00:00", "epoch": 0, "step": 6557, "train_loss": 3.66143536567688, "perplexity": 38.91716316072741, "lr": 0.0026291804804649314, "grad_norm": 0.145585, "tokens_per_sec": 106287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:12.540834+00:00", "epoch": 0, "step": 6558, "train_loss": 3.573427677154541, "perplexity": 35.63854144687705, "lr": 0.0026291804804649314, "grad_norm": 0.163428, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:12.845793+00:00", "epoch": 0, "step": 6559, "train_loss": 3.5540406703948975, "perplexity": 34.954271210911315, "lr": 0.0026291804804649314, "grad_norm": 0.139481, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:13.151493+00:00", "epoch": 0, "step": 6560, "train_loss": 3.5810043811798096, "perplexity": 35.9095896566852, "lr": 0.0026291804804649314, "grad_norm": 0.158019, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:13.456982+00:00", "epoch": 0, "step": 6561, "train_loss": 3.605618476867676, "perplexity": 36.80443951406236, "lr": 0.0026291804804649314, "grad_norm": 0.177683, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:13.762615+00:00", "epoch": 0, "step": 6562, "train_loss": 3.6597347259521484, "perplexity": 38.85103533282572, "lr": 0.0026291804804649314, "grad_norm": 0.186368, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:14.068935+00:00", "epoch": 0, "step": 6563, "train_loss": 3.453714370727539, "perplexity": 31.617614011504994, "lr": 0.0026291804804649314, "grad_norm": 0.17377, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:14.374286+00:00", "epoch": 0, "step": 6564, "train_loss": 3.629049062728882, "perplexity": 37.67697114036278, "lr": 0.0026291804804649314, "grad_norm": 0.159264, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:14.679806+00:00", "epoch": 0, "step": 6565, "train_loss": 3.6017158031463623, "perplexity": 36.66108371255548, "lr": 0.0026291804804649314, "grad_norm": 0.155014, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:14.986207+00:00", "epoch": 0, "step": 6566, "train_loss": 3.6249172687530518, "perplexity": 37.52161882087729, "lr": 0.0026291804804649314, "grad_norm": 0.157341, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:15.292372+00:00", "epoch": 0, "step": 6567, "train_loss": 3.528536319732666, "perplexity": 34.07405757179645, "lr": 0.0026291804804649314, "grad_norm": 0.174395, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:15.599328+00:00", "epoch": 0, "step": 6568, "train_loss": 3.7835323810577393, "perplexity": 43.97109037564439, "lr": 0.0026291804804649314, "grad_norm": 0.15505, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:15.906225+00:00", "epoch": 0, "step": 6569, "train_loss": 3.5082955360412598, "perplexity": 33.39130497634184, "lr": 0.0026291804804649314, "grad_norm": 0.157203, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:16.212026+00:00", "epoch": 0, "step": 6570, "train_loss": 3.4845595359802246, "perplexity": 32.60806126853016, "lr": 0.0026291804804649314, "grad_norm": 0.169489, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:16.517570+00:00", "epoch": 0, "step": 6571, "train_loss": 3.4759631156921387, "perplexity": 32.328950066507524, "lr": 0.0026291804804649314, "grad_norm": 0.16275, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:16.823578+00:00", "epoch": 0, "step": 6572, "train_loss": 3.6856579780578613, "perplexity": 39.87134829321477, "lr": 0.0026291804804649314, "grad_norm": 0.155182, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:17.129121+00:00", "epoch": 0, "step": 6573, "train_loss": 3.6189424991607666, "perplexity": 37.2981041828404, "lr": 0.0026291804804649314, "grad_norm": 0.164772, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:17.435668+00:00", "epoch": 0, "step": 6574, "train_loss": 3.644176483154297, "perplexity": 38.25125932095714, "lr": 0.0026291804804649314, "grad_norm": 0.165819, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:17.741688+00:00", "epoch": 0, "step": 6575, "train_loss": 3.6138358116149902, "perplexity": 37.10811992671749, "lr": 0.0026291804804649314, "grad_norm": 0.147899, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:18.046326+00:00", "epoch": 0, "step": 6576, "train_loss": 3.5466668605804443, "perplexity": 34.697473016895415, "lr": 0.0026291804804649314, "grad_norm": 0.172446, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:18.352205+00:00", "epoch": 0, "step": 6577, "train_loss": 3.637342929840088, "perplexity": 37.99075838828433, "lr": 0.0026291804804649314, "grad_norm": 0.14773, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:18.659682+00:00", "epoch": 0, "step": 6578, "train_loss": 3.5594937801361084, "perplexity": 35.14540134090864, "lr": 0.0026291804804649314, "grad_norm": 0.154095, "tokens_per_sec": 106570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:18.965410+00:00", "epoch": 0, "step": 6579, "train_loss": 3.536695718765259, "perplexity": 34.35321874912473, "lr": 0.0026291804804649314, "grad_norm": 0.16088, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:19.270864+00:00", "epoch": 0, "step": 6580, "train_loss": 3.59673810005188, "perplexity": 36.47904915556494, "lr": 0.0026291804804649314, "grad_norm": 0.162741, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:19.575732+00:00", "epoch": 0, "step": 6581, "train_loss": 3.4981114864349365, "perplexity": 33.05297199415488, "lr": 0.0026291804804649314, "grad_norm": 0.154847, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:19.881290+00:00", "epoch": 0, "step": 6582, "train_loss": 3.5899007320404053, "perplexity": 36.23047922221315, "lr": 0.0026291804804649314, "grad_norm": 0.151484, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:20.187453+00:00", "epoch": 0, "step": 6583, "train_loss": 3.6158976554870605, "perplexity": 37.18471000765365, "lr": 0.0026291804804649314, "grad_norm": 0.174457, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:20.495064+00:00", "epoch": 0, "step": 6584, "train_loss": 3.573981761932373, "perplexity": 35.6582936919007, "lr": 0.0026291804804649314, "grad_norm": 0.153694, "tokens_per_sec": 106524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:20.801904+00:00", "epoch": 0, "step": 6585, "train_loss": 3.565551519393921, "perplexity": 35.35894917385605, "lr": 0.0026291804804649314, "grad_norm": 0.157446, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:21.108059+00:00", "epoch": 0, "step": 6586, "train_loss": 3.535222053527832, "perplexity": 34.30263088880111, "lr": 0.0026291804804649314, "grad_norm": 0.16768, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:21.413266+00:00", "epoch": 0, "step": 6587, "train_loss": 3.7329797744750977, "perplexity": 41.80348772617248, "lr": 0.0026291804804649314, "grad_norm": 0.15208, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:21.718382+00:00", "epoch": 0, "step": 6588, "train_loss": 3.53082275390625, "perplexity": 34.152054795240474, "lr": 0.0026291804804649314, "grad_norm": 0.151091, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:22.024369+00:00", "epoch": 0, "step": 6589, "train_loss": 3.4708824157714844, "perplexity": 32.16511292857081, "lr": 0.0026291804804649314, "grad_norm": 0.197311, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:22.331759+00:00", "epoch": 0, "step": 6590, "train_loss": 3.591043472290039, "perplexity": 36.27190491397857, "lr": 0.0026291804804649314, "grad_norm": 0.171352, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:22.636485+00:00", "epoch": 0, "step": 6591, "train_loss": 3.4001054763793945, "perplexity": 29.96726071886699, "lr": 0.0026291804804649314, "grad_norm": 0.184341, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:22.943162+00:00", "epoch": 0, "step": 6592, "train_loss": 3.5217554569244385, "perplexity": 33.84378765874589, "lr": 0.0026291804804649314, "grad_norm": 0.169273, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:23.248505+00:00", "epoch": 0, "step": 6593, "train_loss": 3.514979124069214, "perplexity": 33.61522616746305, "lr": 0.0026291804804649314, "grad_norm": 0.163773, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:23.554389+00:00", "epoch": 0, "step": 6594, "train_loss": 3.6775717735290527, "perplexity": 39.550240437654516, "lr": 0.0026291804804649314, "grad_norm": 0.166533, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:23.860301+00:00", "epoch": 0, "step": 6595, "train_loss": 3.5916879177093506, "perplexity": 36.29528771060892, "lr": 0.0026291804804649314, "grad_norm": 0.166019, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:24.165528+00:00", "epoch": 0, "step": 6596, "train_loss": 3.5461626052856445, "perplexity": 34.67998104299134, "lr": 0.0026291804804649314, "grad_norm": 0.138462, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:24.469429+00:00", "epoch": 0, "step": 6597, "train_loss": 3.49068546295166, "perplexity": 32.80842896287359, "lr": 0.0026291804804649314, "grad_norm": 0.146649, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:24.774270+00:00", "epoch": 0, "step": 6598, "train_loss": 3.517850399017334, "perplexity": 33.7118834226054, "lr": 0.0026291804804649314, "grad_norm": 0.154179, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:25.081026+00:00", "epoch": 0, "step": 6599, "train_loss": 3.518554449081421, "perplexity": 33.7356266335134, "lr": 0.0026291804804649314, "grad_norm": 0.155793, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:25.388892+00:00", "epoch": 0, "step": 6600, "train_loss": 3.514941692352295, "perplexity": 33.6139679153824, "lr": 0.0026291804804649314, "grad_norm": 0.163712, "tokens_per_sec": 106504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:25.694995+00:00", "epoch": 0, "step": 6601, "train_loss": 3.579801082611084, "perplexity": 35.86640568566365, "lr": 0.0026291804804649314, "grad_norm": 0.196905, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:26.000364+00:00", "epoch": 0, "step": 6602, "train_loss": 3.5729172229766846, "perplexity": 35.62035424676383, "lr": 0.0026291804804649314, "grad_norm": 0.182138, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:26.305029+00:00", "epoch": 0, "step": 6603, "train_loss": 3.533080816268921, "perplexity": 34.229259398381245, "lr": 0.0026291804804649314, "grad_norm": 0.176489, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:26.609189+00:00", "epoch": 0, "step": 6604, "train_loss": 3.4742612838745117, "perplexity": 32.27397842016982, "lr": 0.0026291804804649314, "grad_norm": 0.187066, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:26.915062+00:00", "epoch": 0, "step": 6605, "train_loss": 3.531156301498413, "perplexity": 34.16344803087238, "lr": 0.0026291804804649314, "grad_norm": 0.178751, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:27.221178+00:00", "epoch": 0, "step": 6606, "train_loss": 3.503013849258423, "perplexity": 33.21540748881013, "lr": 0.0026291804804649314, "grad_norm": 0.162789, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:27.528191+00:00", "epoch": 0, "step": 6607, "train_loss": 3.557135581970215, "perplexity": 35.06261916667595, "lr": 0.0026291804804649314, "grad_norm": 0.166181, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:27.834309+00:00", "epoch": 0, "step": 6608, "train_loss": 3.575151205062866, "perplexity": 35.700018431103395, "lr": 0.0026291804804649314, "grad_norm": 0.157865, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:28.139277+00:00", "epoch": 0, "step": 6609, "train_loss": 3.5036497116088867, "perplexity": 33.236534632153244, "lr": 0.0026291804804649314, "grad_norm": 0.171156, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:28.444954+00:00", "epoch": 0, "step": 6610, "train_loss": 3.498198986053467, "perplexity": 33.05586424312916, "lr": 0.0026291804804649314, "grad_norm": 0.164814, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:28.751076+00:00", "epoch": 0, "step": 6611, "train_loss": 3.5001649856567383, "perplexity": 33.12091598401241, "lr": 0.0026291804804649314, "grad_norm": 0.153174, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:29.058298+00:00", "epoch": 0, "step": 6612, "train_loss": 3.5445914268493652, "perplexity": 34.62553538772616, "lr": 0.0026291804804649314, "grad_norm": 0.150641, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:29.363860+00:00", "epoch": 0, "step": 6613, "train_loss": 3.5598533153533936, "perplexity": 35.15803962223379, "lr": 0.0026291804804649314, "grad_norm": 0.177154, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:29.668737+00:00", "epoch": 0, "step": 6614, "train_loss": 3.548447847366333, "perplexity": 34.75932381922427, "lr": 0.0026291804804649314, "grad_norm": 0.160937, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:29.973955+00:00", "epoch": 0, "step": 6615, "train_loss": 3.545480489730835, "perplexity": 34.656333354626355, "lr": 0.0026291804804649314, "grad_norm": 0.134003, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:30.279123+00:00", "epoch": 0, "step": 6616, "train_loss": 3.5833792686462402, "perplexity": 35.994972237966145, "lr": 0.0026291804804649314, "grad_norm": 0.140542, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:30.584813+00:00", "epoch": 0, "step": 6617, "train_loss": 3.707907199859619, "perplexity": 40.76839707858141, "lr": 0.0026291804804649314, "grad_norm": 0.155035, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:30.888865+00:00", "epoch": 0, "step": 6618, "train_loss": 3.6534385681152344, "perplexity": 38.60719152731566, "lr": 0.0026291804804649314, "grad_norm": 0.159549, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:31.194459+00:00", "epoch": 0, "step": 6619, "train_loss": 3.542492151260376, "perplexity": 34.552923089809866, "lr": 0.0026291804804649314, "grad_norm": 0.146758, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:31.500991+00:00", "epoch": 0, "step": 6620, "train_loss": 3.5479280948638916, "perplexity": 34.74126226786092, "lr": 0.0026291804804649314, "grad_norm": 0.130016, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:31.807216+00:00", "epoch": 0, "step": 6621, "train_loss": 3.6181280612945557, "perplexity": 37.26773956118328, "lr": 0.0026291804804649314, "grad_norm": 0.16646, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:32.115211+00:00", "epoch": 0, "step": 6622, "train_loss": 3.565704584121704, "perplexity": 35.36436179601625, "lr": 0.0026291804804649314, "grad_norm": 0.140666, "tokens_per_sec": 106392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:32.420570+00:00", "epoch": 0, "step": 6623, "train_loss": 3.548645496368408, "perplexity": 34.7661946438734, "lr": 0.0026291804804649314, "grad_norm": 0.183365, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:32.727290+00:00", "epoch": 0, "step": 6624, "train_loss": 3.5108680725097656, "perplexity": 33.47731591182397, "lr": 0.0026291804804649314, "grad_norm": 0.155945, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:33.032913+00:00", "epoch": 0, "step": 6625, "train_loss": 3.595921754837036, "perplexity": 36.44928181021284, "lr": 0.0026291804804649314, "grad_norm": 0.144979, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:33.338891+00:00", "epoch": 0, "step": 6626, "train_loss": 3.3992230892181396, "perplexity": 29.940829655681043, "lr": 0.0026291804804649314, "grad_norm": 0.146316, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:33.645312+00:00", "epoch": 0, "step": 6627, "train_loss": 3.5992064476013184, "perplexity": 36.56920334731924, "lr": 0.0026291804804649314, "grad_norm": 0.153947, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:33.951423+00:00", "epoch": 0, "step": 6628, "train_loss": 3.460566520690918, "perplexity": 31.8350065969108, "lr": 0.0026291804804649314, "grad_norm": 0.131373, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:34.258538+00:00", "epoch": 0, "step": 6629, "train_loss": 3.523582935333252, "perplexity": 33.905692998071565, "lr": 0.0026291804804649314, "grad_norm": 0.15013, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:34.563640+00:00", "epoch": 0, "step": 6630, "train_loss": 3.604814291000366, "perplexity": 36.77485380175121, "lr": 0.0026291804804649314, "grad_norm": 0.159622, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:34.868405+00:00", "epoch": 0, "step": 6631, "train_loss": 3.6404075622558594, "perplexity": 38.1073646843947, "lr": 0.0026291804804649314, "grad_norm": 0.170234, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:35.174805+00:00", "epoch": 0, "step": 6632, "train_loss": 3.600519895553589, "perplexity": 36.61726664998549, "lr": 0.0026291804804649314, "grad_norm": 0.159902, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:35.480853+00:00", "epoch": 0, "step": 6633, "train_loss": 3.5876383781433105, "perplexity": 36.148605704707585, "lr": 0.0026291804804649314, "grad_norm": 0.157006, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:35.788315+00:00", "epoch": 0, "step": 6634, "train_loss": 3.609300374984741, "perplexity": 36.940199484430764, "lr": 0.0026291804804649314, "grad_norm": 0.159953, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:36.094093+00:00", "epoch": 0, "step": 6635, "train_loss": 3.4784133434295654, "perplexity": 32.408260481292345, "lr": 0.0026291804804649314, "grad_norm": 0.155109, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:36.400766+00:00", "epoch": 0, "step": 6636, "train_loss": 3.5594375133514404, "perplexity": 35.14342387781259, "lr": 0.0026291804804649314, "grad_norm": 0.169603, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:36.706770+00:00", "epoch": 0, "step": 6637, "train_loss": 3.6530284881591797, "perplexity": 38.591362737667254, "lr": 0.0026291804804649314, "grad_norm": 0.198473, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:37.013366+00:00", "epoch": 0, "step": 6638, "train_loss": 3.5501389503479004, "perplexity": 34.818155146267884, "lr": 0.0026291804804649314, "grad_norm": 0.155074, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:37.319600+00:00", "epoch": 0, "step": 6639, "train_loss": 3.535576581954956, "perplexity": 34.31479430258691, "lr": 0.0026291804804649314, "grad_norm": 0.158267, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:37.627290+00:00", "epoch": 0, "step": 6640, "train_loss": 3.5486953258514404, "perplexity": 34.76792706854206, "lr": 0.0026291804804649314, "grad_norm": 0.160177, "tokens_per_sec": 106556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:37.934751+00:00", "epoch": 0, "step": 6641, "train_loss": 3.572970390319824, "perplexity": 35.62224813670692, "lr": 0.0026291804804649314, "grad_norm": 0.1486, "tokens_per_sec": 106520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:38.240474+00:00", "epoch": 0, "step": 6642, "train_loss": 3.541856288909912, "perplexity": 34.53095917067325, "lr": 0.0026291804804649314, "grad_norm": 0.135856, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:38.545950+00:00", "epoch": 0, "step": 6643, "train_loss": 3.5172982215881348, "perplexity": 33.693273619913, "lr": 0.0026291804804649314, "grad_norm": 0.173187, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:38.852697+00:00", "epoch": 0, "step": 6644, "train_loss": 3.5573890209198, "perplexity": 35.07150652620121, "lr": 0.0026291804804649314, "grad_norm": 0.165857, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:39.158707+00:00", "epoch": 0, "step": 6645, "train_loss": 3.5767617225646973, "perplexity": 35.757560259225514, "lr": 0.0026291804804649314, "grad_norm": 0.191394, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:39.463950+00:00", "epoch": 0, "step": 6646, "train_loss": 3.7139697074890137, "perplexity": 41.016306514025594, "lr": 0.0026291804804649314, "grad_norm": 0.146641, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:39.769510+00:00", "epoch": 0, "step": 6647, "train_loss": 3.5835368633270264, "perplexity": 36.000645301136444, "lr": 0.0026291804804649314, "grad_norm": 0.167809, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:40.075700+00:00", "epoch": 0, "step": 6648, "train_loss": 3.5605993270874023, "perplexity": 35.18427771808354, "lr": 0.0026291804804649314, "grad_norm": 0.173895, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:40.381831+00:00", "epoch": 0, "step": 6649, "train_loss": 3.734103202819824, "perplexity": 41.850477338981854, "lr": 0.0026291804804649314, "grad_norm": 0.164875, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:40.687787+00:00", "epoch": 0, "step": 6650, "train_loss": 3.524325132369995, "perplexity": 33.93086704383893, "lr": 0.0026291804804649314, "grad_norm": 0.184045, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:40.994286+00:00", "epoch": 0, "step": 6651, "train_loss": 3.6783604621887207, "perplexity": 39.5814455677234, "lr": 0.0026291804804649314, "grad_norm": 0.163603, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:41.299619+00:00", "epoch": 0, "step": 6652, "train_loss": 3.4615731239318848, "perplexity": 31.867067951552016, "lr": 0.0026291804804649314, "grad_norm": 0.14598, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:41.605162+00:00", "epoch": 0, "step": 6653, "train_loss": 3.667393445968628, "perplexity": 39.149726872223326, "lr": 0.0026291804804649314, "grad_norm": 0.149159, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:41.911487+00:00", "epoch": 0, "step": 6654, "train_loss": 3.4689550399780273, "perplexity": 32.103178373272556, "lr": 0.0026291804804649314, "grad_norm": 0.165694, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:42.218173+00:00", "epoch": 0, "step": 6655, "train_loss": 3.613576650619507, "perplexity": 37.09850419548182, "lr": 0.0026291804804649314, "grad_norm": 0.149412, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:42.525038+00:00", "epoch": 0, "step": 6656, "train_loss": 3.58817195892334, "perplexity": 36.16789905275858, "lr": 0.0026291804804649314, "grad_norm": 0.154797, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:42.831554+00:00", "epoch": 0, "step": 6657, "train_loss": 3.5341124534606934, "perplexity": 34.26458979636628, "lr": 0.0026291804804649314, "grad_norm": 0.16487, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:43.137410+00:00", "epoch": 0, "step": 6658, "train_loss": 3.6185221672058105, "perplexity": 37.28242989222645, "lr": 0.0026291804804649314, "grad_norm": 0.138682, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:43.442263+00:00", "epoch": 0, "step": 6659, "train_loss": 3.5625455379486084, "perplexity": 35.25282041916167, "lr": 0.0026291804804649314, "grad_norm": 0.158715, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:43.748120+00:00", "epoch": 0, "step": 6660, "train_loss": 3.565502882003784, "perplexity": 35.357229448672065, "lr": 0.0026291804804649314, "grad_norm": 0.151653, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:44.054790+00:00", "epoch": 0, "step": 6661, "train_loss": 3.566211700439453, "perplexity": 35.38230018899125, "lr": 0.0026291804804649314, "grad_norm": 0.167009, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:44.362653+00:00", "epoch": 0, "step": 6662, "train_loss": 3.5008771419525146, "perplexity": 33.14451165375744, "lr": 0.0026291804804649314, "grad_norm": 0.162906, "tokens_per_sec": 106440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:44.667939+00:00", "epoch": 0, "step": 6663, "train_loss": 3.5066921710968018, "perplexity": 33.33780942639824, "lr": 0.0026291804804649314, "grad_norm": 0.157851, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:44.974033+00:00", "epoch": 0, "step": 6664, "train_loss": 3.518089771270752, "perplexity": 33.71995407801419, "lr": 0.0026291804804649314, "grad_norm": 0.183401, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:45.279593+00:00", "epoch": 0, "step": 6665, "train_loss": 3.6011860370635986, "perplexity": 36.64166705744398, "lr": 0.0026291804804649314, "grad_norm": 0.132746, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:45.584680+00:00", "epoch": 0, "step": 6666, "train_loss": 3.551795244216919, "perplexity": 34.87587202803128, "lr": 0.0026291804804649314, "grad_norm": 0.165824, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:45.891640+00:00", "epoch": 0, "step": 6667, "train_loss": 3.5166587829589844, "perplexity": 33.67173572603243, "lr": 0.0026291804804649314, "grad_norm": 0.158231, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:46.197592+00:00", "epoch": 0, "step": 6668, "train_loss": 3.5681352615356445, "perplexity": 35.45042570591971, "lr": 0.0026291804804649314, "grad_norm": 0.15639, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:46.503876+00:00", "epoch": 0, "step": 6669, "train_loss": 3.6234078407287598, "perplexity": 37.46502536053486, "lr": 0.0026291804804649314, "grad_norm": 0.152206, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:46.809621+00:00", "epoch": 0, "step": 6670, "train_loss": 3.5918679237365723, "perplexity": 36.301821669214895, "lr": 0.0026291804804649314, "grad_norm": 0.153583, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:47.115047+00:00", "epoch": 0, "step": 6671, "train_loss": 3.6348650455474854, "perplexity": 37.896738218467206, "lr": 0.0026291804804649314, "grad_norm": 0.172146, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:47.420076+00:00", "epoch": 0, "step": 6672, "train_loss": 3.6764981746673584, "perplexity": 39.507802129476836, "lr": 0.0026291804804649314, "grad_norm": 0.165723, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:47.726635+00:00", "epoch": 0, "step": 6673, "train_loss": 3.4442968368530273, "perplexity": 31.32125175169033, "lr": 0.0026291804804649314, "grad_norm": 0.145107, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:48.032147+00:00", "epoch": 0, "step": 6674, "train_loss": 3.5780441761016846, "perplexity": 35.80344708642679, "lr": 0.0026291804804649314, "grad_norm": 0.175187, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:48.337230+00:00", "epoch": 0, "step": 6675, "train_loss": 3.573878288269043, "perplexity": 35.65460418851081, "lr": 0.0026291804804649314, "grad_norm": 0.191391, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:48.642469+00:00", "epoch": 0, "step": 6676, "train_loss": 3.4325296878814697, "perplexity": 30.954849896791472, "lr": 0.0026291804804649314, "grad_norm": 0.163211, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:48.949327+00:00", "epoch": 0, "step": 6677, "train_loss": 3.5734059810638428, "perplexity": 35.63776823823729, "lr": 0.0026291804804649314, "grad_norm": 0.15313, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:49.255431+00:00", "epoch": 0, "step": 6678, "train_loss": 3.514841079711914, "perplexity": 33.61058609544653, "lr": 0.0026291804804649314, "grad_norm": 0.155014, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:49.562559+00:00", "epoch": 0, "step": 6679, "train_loss": 3.622131586074829, "perplexity": 37.41724094658523, "lr": 0.0026291804804649314, "grad_norm": 0.145587, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:49.868225+00:00", "epoch": 0, "step": 6680, "train_loss": 3.590393304824829, "perplexity": 36.24832976622981, "lr": 0.0026291804804649314, "grad_norm": 0.148401, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:50.174517+00:00", "epoch": 0, "step": 6681, "train_loss": 3.5947625637054443, "perplexity": 36.407054605414686, "lr": 0.0026291804804649314, "grad_norm": 0.149059, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:50.480791+00:00", "epoch": 0, "step": 6682, "train_loss": 3.6141302585601807, "perplexity": 37.11904790804808, "lr": 0.0026291804804649314, "grad_norm": 0.157432, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:50.786906+00:00", "epoch": 0, "step": 6683, "train_loss": 3.63870906829834, "perplexity": 38.04269449225371, "lr": 0.0026291804804649314, "grad_norm": 0.169593, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:51.094012+00:00", "epoch": 0, "step": 6684, "train_loss": 3.649446487426758, "perplexity": 38.45337573039579, "lr": 0.0026291804804649314, "grad_norm": 0.168248, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:51.399538+00:00", "epoch": 0, "step": 6685, "train_loss": 3.617332935333252, "perplexity": 37.23811879162109, "lr": 0.0026291804804649314, "grad_norm": 0.173862, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:51.705527+00:00", "epoch": 0, "step": 6686, "train_loss": 3.559018135070801, "perplexity": 35.12868857917899, "lr": 0.0026291804804649314, "grad_norm": 0.148167, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:52.012162+00:00", "epoch": 0, "step": 6687, "train_loss": 3.623016357421875, "perplexity": 37.450361299069115, "lr": 0.0026291804804649314, "grad_norm": 0.176055, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:52.319284+00:00", "epoch": 0, "step": 6688, "train_loss": 3.5779831409454346, "perplexity": 35.80126188412735, "lr": 0.0026291804804649314, "grad_norm": 0.164631, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:52.626754+00:00", "epoch": 0, "step": 6689, "train_loss": 3.490797519683838, "perplexity": 32.81210557420174, "lr": 0.0026291804804649314, "grad_norm": 0.168511, "tokens_per_sec": 106573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:52.933459+00:00", "epoch": 0, "step": 6690, "train_loss": 3.5776474475860596, "perplexity": 35.789245655252564, "lr": 0.0026291804804649314, "grad_norm": 0.160263, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:53.240044+00:00", "epoch": 0, "step": 6691, "train_loss": 3.380009412765503, "perplexity": 29.371047574771698, "lr": 0.0026291804804649314, "grad_norm": 0.147776, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:53.545236+00:00", "epoch": 0, "step": 6692, "train_loss": 3.5267086029052734, "perplexity": 34.01183672176323, "lr": 0.0026291804804649314, "grad_norm": 0.142726, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:53.850719+00:00", "epoch": 0, "step": 6693, "train_loss": 3.5510315895080566, "perplexity": 34.84924907079853, "lr": 0.0026291804804649314, "grad_norm": 0.1449, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:54.157306+00:00", "epoch": 0, "step": 6694, "train_loss": 3.5164635181427, "perplexity": 33.66516146262398, "lr": 0.0026291804804649314, "grad_norm": 0.149162, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:54.464463+00:00", "epoch": 0, "step": 6695, "train_loss": 3.582198143005371, "perplexity": 35.9524827509644, "lr": 0.0026291804804649314, "grad_norm": 0.141056, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:54.770072+00:00", "epoch": 0, "step": 6696, "train_loss": 3.5274617671966553, "perplexity": 34.03746287180556, "lr": 0.0026291804804649314, "grad_norm": 0.141714, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:55.075978+00:00", "epoch": 0, "step": 6697, "train_loss": 3.5737173557281494, "perplexity": 35.648866664153765, "lr": 0.0026291804804649314, "grad_norm": 0.145394, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:55.382024+00:00", "epoch": 0, "step": 6698, "train_loss": 3.5411388874053955, "perplexity": 34.506195492423906, "lr": 0.0026291804804649314, "grad_norm": 0.153096, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:55.686971+00:00", "epoch": 0, "step": 6699, "train_loss": 3.6056764125823975, "perplexity": 36.80657186733964, "lr": 0.0026291804804649314, "grad_norm": 0.179436, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:55.993588+00:00", "epoch": 0, "step": 6700, "train_loss": 3.500824451446533, "perplexity": 33.14276529867647, "lr": 0.0026291804804649314, "grad_norm": 0.182282, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:56.299403+00:00", "epoch": 0, "step": 6701, "train_loss": 3.6201395988464355, "perplexity": 37.342780467281486, "lr": 0.0026291804804649314, "grad_norm": 0.149029, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:56.605271+00:00", "epoch": 0, "step": 6702, "train_loss": 3.48421049118042, "perplexity": 32.59668158043867, "lr": 0.0026291804804649314, "grad_norm": 0.175915, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:56.911216+00:00", "epoch": 0, "step": 6703, "train_loss": 3.65024471282959, "perplexity": 38.484082445538036, "lr": 0.0026291804804649314, "grad_norm": 0.201978, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:57.217208+00:00", "epoch": 0, "step": 6704, "train_loss": 3.5281147956848145, "perplexity": 34.05969756387902, "lr": 0.0026291804804649314, "grad_norm": 0.203724, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:57.524167+00:00", "epoch": 0, "step": 6705, "train_loss": 3.6040713787078857, "perplexity": 36.74754345665935, "lr": 0.0026291804804649314, "grad_norm": 0.162903, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:57.829928+00:00", "epoch": 0, "step": 6706, "train_loss": 3.571367025375366, "perplexity": 35.56517843680598, "lr": 0.0026291804804649314, "grad_norm": 0.150655, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:58.135839+00:00", "epoch": 0, "step": 6707, "train_loss": 3.534827947616577, "perplexity": 34.289114682779534, "lr": 0.0026291804804649314, "grad_norm": 0.14986, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:58.442048+00:00", "epoch": 0, "step": 6708, "train_loss": 3.5728533267974854, "perplexity": 35.61807831493819, "lr": 0.0026291804804649314, "grad_norm": 0.168325, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:58.748295+00:00", "epoch": 0, "step": 6709, "train_loss": 3.482192277908325, "perplexity": 32.530960866563, "lr": 0.0026291804804649314, "grad_norm": 0.157859, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:59.055371+00:00", "epoch": 0, "step": 6710, "train_loss": 3.5934245586395264, "perplexity": 36.35837435638991, "lr": 0.0026291804804649314, "grad_norm": 0.153308, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:59.364247+00:00", "epoch": 0, "step": 6711, "train_loss": 3.6203126907348633, "perplexity": 37.34924475911381, "lr": 0.0026291804804649314, "grad_norm": 0.158992, "tokens_per_sec": 106034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:59.670607+00:00", "epoch": 0, "step": 6712, "train_loss": 3.606038808822632, "perplexity": 36.81991284781486, "lr": 0.0026291804804649314, "grad_norm": 0.170024, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:32:59.975316+00:00", "epoch": 0, "step": 6713, "train_loss": 3.6301631927490234, "perplexity": 37.71897157760772, "lr": 0.0026291804804649314, "grad_norm": 0.1827, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:00.280543+00:00", "epoch": 0, "step": 6714, "train_loss": 3.498924493789673, "perplexity": 33.079855230135536, "lr": 0.0026291804804649314, "grad_norm": 0.179353, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:00.588460+00:00", "epoch": 0, "step": 6715, "train_loss": 3.562138557434082, "perplexity": 35.23847612729061, "lr": 0.0026291804804649314, "grad_norm": 0.163142, "tokens_per_sec": 106418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:00.894089+00:00", "epoch": 0, "step": 6716, "train_loss": 3.6606602668762207, "perplexity": 38.887010201507344, "lr": 0.0026291804804649314, "grad_norm": 0.168873, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:01.201707+00:00", "epoch": 0, "step": 6717, "train_loss": 3.583336591720581, "perplexity": 35.993436115990555, "lr": 0.0026291804804649314, "grad_norm": 0.163424, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:01.507780+00:00", "epoch": 0, "step": 6718, "train_loss": 3.607996702194214, "perplexity": 36.89207292894287, "lr": 0.0026291804804649314, "grad_norm": 0.166205, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:01.813906+00:00", "epoch": 0, "step": 6719, "train_loss": 3.593902587890625, "perplexity": 36.37575887767845, "lr": 0.0026291804804649314, "grad_norm": 0.166041, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:02.121335+00:00", "epoch": 0, "step": 6720, "train_loss": 3.4814565181732178, "perplexity": 32.50703469844794, "lr": 0.0026291804804649314, "grad_norm": 0.158359, "tokens_per_sec": 106652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:02.428904+00:00", "epoch": 0, "step": 6721, "train_loss": 3.6425833702087402, "perplexity": 38.190369259799716, "lr": 0.0026291804804649314, "grad_norm": 0.178672, "tokens_per_sec": 106474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:02.735769+00:00", "epoch": 0, "step": 6722, "train_loss": 3.634552478790283, "perplexity": 37.884894808918204, "lr": 0.0026291804804649314, "grad_norm": 0.200957, "tokens_per_sec": 106783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:03.042105+00:00", "epoch": 0, "step": 6723, "train_loss": 3.5737290382385254, "perplexity": 35.64928313484117, "lr": 0.0026291804804649314, "grad_norm": 0.155326, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:03.347736+00:00", "epoch": 0, "step": 6724, "train_loss": 3.499258279800415, "perplexity": 33.090898666018404, "lr": 0.0026291804804649314, "grad_norm": 0.171176, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:03.654702+00:00", "epoch": 0, "step": 6725, "train_loss": 3.6841371059417725, "perplexity": 39.81075516024752, "lr": 0.0026291804804649314, "grad_norm": 0.156904, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:03.960943+00:00", "epoch": 0, "step": 6726, "train_loss": 3.685530424118042, "perplexity": 39.86626286999384, "lr": 0.0026291804804649314, "grad_norm": 0.171279, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:04.267954+00:00", "epoch": 0, "step": 6727, "train_loss": 3.535954236984253, "perplexity": 34.32775590458761, "lr": 0.0026291804804649314, "grad_norm": 0.169969, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:04.573880+00:00", "epoch": 0, "step": 6728, "train_loss": 3.548834800720215, "perplexity": 34.77277665879762, "lr": 0.0026291804804649314, "grad_norm": 0.155456, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:04.880203+00:00", "epoch": 0, "step": 6729, "train_loss": 3.637561559677124, "perplexity": 37.999065209626046, "lr": 0.0026291804804649314, "grad_norm": 0.173826, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:05.187159+00:00", "epoch": 0, "step": 6730, "train_loss": 3.627530097961426, "perplexity": 37.61978459182763, "lr": 0.0026291804804649314, "grad_norm": 0.151755, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:05.494050+00:00", "epoch": 0, "step": 6731, "train_loss": 3.6305456161499023, "perplexity": 37.73339895350333, "lr": 0.0026291804804649314, "grad_norm": 0.15709, "tokens_per_sec": 106695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:05.800893+00:00", "epoch": 0, "step": 6732, "train_loss": 3.5854475498199463, "perplexity": 36.06949700391244, "lr": 0.0026291804804649314, "grad_norm": 0.151668, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:06.106521+00:00", "epoch": 0, "step": 6733, "train_loss": 3.61574649810791, "perplexity": 37.17908968913144, "lr": 0.0026291804804649314, "grad_norm": 0.142082, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:06.411466+00:00", "epoch": 0, "step": 6734, "train_loss": 3.625213861465454, "perplexity": 37.53274911007684, "lr": 0.0026291804804649314, "grad_norm": 0.169051, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:06.717737+00:00", "epoch": 0, "step": 6735, "train_loss": 3.521087646484375, "perplexity": 33.82119396900671, "lr": 0.0026291804804649314, "grad_norm": 0.18372, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:07.024206+00:00", "epoch": 0, "step": 6736, "train_loss": 3.6612188816070557, "perplexity": 38.908739126727845, "lr": 0.0026291804804649314, "grad_norm": 0.173535, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:07.331240+00:00", "epoch": 0, "step": 6737, "train_loss": 3.5776925086975098, "perplexity": 35.7908583947754, "lr": 0.0026291804804649314, "grad_norm": 0.154001, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:07.638533+00:00", "epoch": 0, "step": 6738, "train_loss": 3.5761525630950928, "perplexity": 35.7357848358097, "lr": 0.0026291804804649314, "grad_norm": 0.155827, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:07.944938+00:00", "epoch": 0, "step": 6739, "train_loss": 3.5786240100860596, "perplexity": 35.82421316166154, "lr": 0.0026291804804649314, "grad_norm": 0.156698, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:08.251130+00:00", "epoch": 0, "step": 6740, "train_loss": 3.582970380783081, "perplexity": 35.98025733926005, "lr": 0.0026291804804649314, "grad_norm": 0.193092, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:08.558755+00:00", "epoch": 0, "step": 6741, "train_loss": 3.555467128753662, "perplexity": 35.004167602356276, "lr": 0.0026291804804649314, "grad_norm": 0.183877, "tokens_per_sec": 106451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:08.864380+00:00", "epoch": 0, "step": 6742, "train_loss": 3.5387048721313477, "perplexity": 34.42230901744172, "lr": 0.0026291804804649314, "grad_norm": 0.156374, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:09.170956+00:00", "epoch": 0, "step": 6743, "train_loss": 3.467522382736206, "perplexity": 32.05721845255832, "lr": 0.0026291804804649314, "grad_norm": 0.186517, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:09.476297+00:00", "epoch": 0, "step": 6744, "train_loss": 3.5443503856658936, "perplexity": 34.6171902135037, "lr": 0.0026291804804649314, "grad_norm": 0.186969, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:09.783832+00:00", "epoch": 0, "step": 6745, "train_loss": 3.672825336456299, "perplexity": 39.362962513239914, "lr": 0.0026291804804649314, "grad_norm": 0.174729, "tokens_per_sec": 106551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:10.090001+00:00", "epoch": 0, "step": 6746, "train_loss": 3.6031084060668945, "perplexity": 36.71217361051728, "lr": 0.0026291804804649314, "grad_norm": 0.184085, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:10.396751+00:00", "epoch": 0, "step": 6747, "train_loss": 3.5884628295898438, "perplexity": 36.17842076381647, "lr": 0.0026291804804649314, "grad_norm": 0.142607, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:10.702376+00:00", "epoch": 0, "step": 6748, "train_loss": 3.6202869415283203, "perplexity": 37.348283058077854, "lr": 0.0026291804804649314, "grad_norm": 0.158411, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:11.009207+00:00", "epoch": 0, "step": 6749, "train_loss": 3.638225555419922, "perplexity": 38.02430480572039, "lr": 0.0026291804804649314, "grad_norm": 0.145691, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:11.314821+00:00", "epoch": 0, "step": 6750, "train_loss": 3.6215434074401855, "perplexity": 37.395239395948444, "lr": 0.0026291804804649314, "grad_norm": 0.140347, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:11.619772+00:00", "epoch": 0, "step": 6751, "train_loss": 3.6054539680480957, "perplexity": 36.79838535715735, "lr": 0.0026291804804649314, "grad_norm": 0.141013, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:11.928752+00:00", "epoch": 0, "step": 6752, "train_loss": 3.5294134616851807, "perplexity": 34.103958468933634, "lr": 0.0026291804804649314, "grad_norm": 0.154372, "tokens_per_sec": 106051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:12.234657+00:00", "epoch": 0, "step": 6753, "train_loss": 3.4241366386413574, "perplexity": 30.696131554027993, "lr": 0.0026291804804649314, "grad_norm": 0.139119, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:12.541017+00:00", "epoch": 0, "step": 6754, "train_loss": 3.616044521331787, "perplexity": 37.19017157254869, "lr": 0.0026291804804649314, "grad_norm": 0.140564, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:12.846465+00:00", "epoch": 0, "step": 6755, "train_loss": 3.5312225818634033, "perplexity": 34.16571247172034, "lr": 0.0026291804804649314, "grad_norm": 0.159465, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:13.152223+00:00", "epoch": 0, "step": 6756, "train_loss": 3.5612077713012695, "perplexity": 35.20569190228781, "lr": 0.0026291804804649314, "grad_norm": 0.174781, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:13.459046+00:00", "epoch": 0, "step": 6757, "train_loss": 3.4871907234191895, "perplexity": 32.69397216401897, "lr": 0.0026291804804649314, "grad_norm": 0.168362, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:13.765726+00:00", "epoch": 0, "step": 6758, "train_loss": 3.4620361328125, "perplexity": 31.881826103326773, "lr": 0.0026291804804649314, "grad_norm": 0.145871, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:14.071976+00:00", "epoch": 0, "step": 6759, "train_loss": 3.559349536895752, "perplexity": 35.140332219937605, "lr": 0.0026291804804649314, "grad_norm": 0.162786, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:14.378270+00:00", "epoch": 0, "step": 6760, "train_loss": 3.5665829181671143, "perplexity": 35.39543716425503, "lr": 0.0026291804804649314, "grad_norm": 0.166079, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:14.683484+00:00", "epoch": 0, "step": 6761, "train_loss": 3.6500887870788574, "perplexity": 38.47808225389581, "lr": 0.0026291804804649314, "grad_norm": 0.16805, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:14.988820+00:00", "epoch": 0, "step": 6762, "train_loss": 3.557526111602783, "perplexity": 35.07631483256362, "lr": 0.0026291804804649314, "grad_norm": 0.162956, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:15.296924+00:00", "epoch": 0, "step": 6763, "train_loss": 3.655055284500122, "perplexity": 38.66965888882953, "lr": 0.0026291804804649314, "grad_norm": 0.165547, "tokens_per_sec": 106353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:15.603811+00:00", "epoch": 0, "step": 6764, "train_loss": 3.6357932090759277, "perplexity": 37.93192891756387, "lr": 0.0026291804804649314, "grad_norm": 0.180943, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:15.910279+00:00", "epoch": 0, "step": 6765, "train_loss": 3.5864531993865967, "perplexity": 36.10578852316046, "lr": 0.0026291804804649314, "grad_norm": 0.154549, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:16.217144+00:00", "epoch": 0, "step": 6766, "train_loss": 3.5464887619018555, "perplexity": 34.691293993055055, "lr": 0.0026291804804649314, "grad_norm": 0.150387, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:16.522840+00:00", "epoch": 0, "step": 6767, "train_loss": 3.600184202194214, "perplexity": 36.60497653970224, "lr": 0.0026291804804649314, "grad_norm": 0.200425, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:16.829248+00:00", "epoch": 0, "step": 6768, "train_loss": 3.5912609100341797, "perplexity": 36.27979265267373, "lr": 0.0026291804804649314, "grad_norm": 0.165051, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:17.136098+00:00", "epoch": 0, "step": 6769, "train_loss": 3.467617988586426, "perplexity": 32.06028345669798, "lr": 0.0026291804804649314, "grad_norm": 0.153548, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:17.442195+00:00", "epoch": 0, "step": 6770, "train_loss": 3.69382905960083, "perplexity": 40.19847500075575, "lr": 0.0026291804804649314, "grad_norm": 0.143792, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:17.747649+00:00", "epoch": 0, "step": 6771, "train_loss": 3.600844621658325, "perplexity": 36.629159163151066, "lr": 0.0026291804804649314, "grad_norm": 0.161358, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:18.054230+00:00", "epoch": 0, "step": 6772, "train_loss": 3.541104316711426, "perplexity": 34.50500260991898, "lr": 0.0026291804804649314, "grad_norm": 0.15811, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:18.361089+00:00", "epoch": 0, "step": 6773, "train_loss": 3.560800552368164, "perplexity": 35.191358396627635, "lr": 0.0026291804804649314, "grad_norm": 0.151492, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:18.668055+00:00", "epoch": 0, "step": 6774, "train_loss": 3.4509365558624268, "perplexity": 31.529908005225277, "lr": 0.0026291804804649314, "grad_norm": 0.162679, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:18.975326+00:00", "epoch": 0, "step": 6775, "train_loss": 3.5905873775482178, "perplexity": 36.25536526098256, "lr": 0.0026291804804649314, "grad_norm": 0.152679, "tokens_per_sec": 106642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:19.281617+00:00", "epoch": 0, "step": 6776, "train_loss": 3.6519548892974854, "perplexity": 38.549953327086314, "lr": 0.0026291804804649314, "grad_norm": 0.153214, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:19.586434+00:00", "epoch": 0, "step": 6777, "train_loss": 3.6118545532226562, "perplexity": 37.034671936438485, "lr": 0.0026291804804649314, "grad_norm": 0.17178, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:19.891432+00:00", "epoch": 0, "step": 6778, "train_loss": 3.5656464099884033, "perplexity": 35.36230456475843, "lr": 0.0026291804804649314, "grad_norm": 0.156194, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:20.198276+00:00", "epoch": 0, "step": 6779, "train_loss": 3.561906337738037, "perplexity": 35.230294009136514, "lr": 0.0026291804804649314, "grad_norm": 0.161044, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:20.505257+00:00", "epoch": 0, "step": 6780, "train_loss": 3.5348148345947266, "perplexity": 34.288665051817475, "lr": 0.0026291804804649314, "grad_norm": 0.180559, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:20.812608+00:00", "epoch": 0, "step": 6781, "train_loss": 3.586413860321045, "perplexity": 36.10436818311656, "lr": 0.0026291804804649314, "grad_norm": 0.14951, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:21.119655+00:00", "epoch": 0, "step": 6782, "train_loss": 3.533224582672119, "perplexity": 34.23418076964454, "lr": 0.0026291804804649314, "grad_norm": 0.181668, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:21.425165+00:00", "epoch": 0, "step": 6783, "train_loss": 3.5274670124053955, "perplexity": 34.03764140587153, "lr": 0.0026291804804649314, "grad_norm": 0.156036, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:21.729961+00:00", "epoch": 0, "step": 6784, "train_loss": 3.5963501930236816, "perplexity": 36.4649014201959, "lr": 0.0026291804804649314, "grad_norm": 0.176039, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:22.036186+00:00", "epoch": 0, "step": 6785, "train_loss": 3.6250803470611572, "perplexity": 37.5277382819541, "lr": 0.0026291804804649314, "grad_norm": 0.150722, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:22.343471+00:00", "epoch": 0, "step": 6786, "train_loss": 3.447805404663086, "perplexity": 31.43133749607656, "lr": 0.0026291804804649314, "grad_norm": 0.153935, "tokens_per_sec": 106637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:22.649442+00:00", "epoch": 0, "step": 6787, "train_loss": 3.5174262523651123, "perplexity": 33.697587672073304, "lr": 0.0026291804804649314, "grad_norm": 0.143741, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:22.955062+00:00", "epoch": 0, "step": 6788, "train_loss": 3.665703773498535, "perplexity": 39.083632511151755, "lr": 0.0026291804804649314, "grad_norm": 0.184577, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:23.260800+00:00", "epoch": 0, "step": 6789, "train_loss": 3.746598243713379, "perplexity": 42.37668138914777, "lr": 0.0026291804804649314, "grad_norm": 0.164614, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:23.567289+00:00", "epoch": 0, "step": 6790, "train_loss": 3.582881212234497, "perplexity": 35.97704917497123, "lr": 0.0026291804804649314, "grad_norm": 0.159937, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:23.872828+00:00", "epoch": 0, "step": 6791, "train_loss": 3.5909695625305176, "perplexity": 36.26922416527696, "lr": 0.0026291804804649314, "grad_norm": 0.153312, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:24.179474+00:00", "epoch": 0, "step": 6792, "train_loss": 3.5740416049957275, "perplexity": 35.660427657280096, "lr": 0.0026291804804649314, "grad_norm": 0.154492, "tokens_per_sec": 106860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:24.487223+00:00", "epoch": 0, "step": 6793, "train_loss": 3.560218572616577, "perplexity": 35.17088369711813, "lr": 0.0026291804804649314, "grad_norm": 0.156657, "tokens_per_sec": 106476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:24.792647+00:00", "epoch": 0, "step": 6794, "train_loss": 3.5291945934295654, "perplexity": 34.09649501182068, "lr": 0.0026291804804649314, "grad_norm": 0.184616, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:25.097464+00:00", "epoch": 0, "step": 6795, "train_loss": 3.4637796878814697, "perplexity": 31.93746231113888, "lr": 0.0026291804804649314, "grad_norm": 0.167582, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:25.404387+00:00", "epoch": 0, "step": 6796, "train_loss": 3.5489346981048584, "perplexity": 34.77625054175565, "lr": 0.0026291804804649314, "grad_norm": 0.160953, "tokens_per_sec": 106763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:25.710299+00:00", "epoch": 0, "step": 6797, "train_loss": 3.467466354370117, "perplexity": 32.05542238930278, "lr": 0.0026291804804649314, "grad_norm": 0.14258, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:26.016667+00:00", "epoch": 0, "step": 6798, "train_loss": 3.672938585281372, "perplexity": 39.3674205749263, "lr": 0.0026291804804649314, "grad_norm": 0.139595, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:26.322710+00:00", "epoch": 0, "step": 6799, "train_loss": 3.6518959999084473, "perplexity": 38.54768321073097, "lr": 0.0026291804804649314, "grad_norm": 0.175672, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:26.629362+00:00", "epoch": 0, "step": 6800, "train_loss": 3.5781607627868652, "perplexity": 35.80762153497847, "lr": 0.0026291804804649314, "grad_norm": 0.149045, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:26.935968+00:00", "epoch": 0, "step": 6801, "train_loss": 3.5733184814453125, "perplexity": 35.63465008353184, "lr": 0.0026291804804649314, "grad_norm": 0.177479, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:27.242396+00:00", "epoch": 0, "step": 6802, "train_loss": 3.571427345275879, "perplexity": 35.567323789534086, "lr": 0.0026291804804649314, "grad_norm": 0.18751, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:27.549852+00:00", "epoch": 0, "step": 6803, "train_loss": 3.603959560394287, "perplexity": 36.743434638046004, "lr": 0.0026291804804649314, "grad_norm": 0.186801, "tokens_per_sec": 106578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:27.855826+00:00", "epoch": 0, "step": 6804, "train_loss": 3.4772250652313232, "perplexity": 32.3697733231736, "lr": 0.0026291804804649314, "grad_norm": 0.142659, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:28.161268+00:00", "epoch": 0, "step": 6805, "train_loss": 3.560108184814453, "perplexity": 35.16700147484698, "lr": 0.0026291804804649314, "grad_norm": 0.167598, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:28.466467+00:00", "epoch": 0, "step": 6806, "train_loss": 3.5721564292907715, "perplexity": 35.593264812204914, "lr": 0.0026291804804649314, "grad_norm": 0.195004, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:28.772105+00:00", "epoch": 0, "step": 6807, "train_loss": 3.6230297088623047, "perplexity": 37.45086131867506, "lr": 0.0026291804804649314, "grad_norm": 0.158092, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:29.079953+00:00", "epoch": 0, "step": 6808, "train_loss": 3.5649666786193848, "perplexity": 35.3382758645208, "lr": 0.0026291804804649314, "grad_norm": 0.139038, "tokens_per_sec": 106442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:29.386805+00:00", "epoch": 0, "step": 6809, "train_loss": 3.5902369022369385, "perplexity": 36.24266087697369, "lr": 0.0026291804804649314, "grad_norm": 0.164212, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:29.693183+00:00", "epoch": 0, "step": 6810, "train_loss": 3.5902912616729736, "perplexity": 36.24463106112793, "lr": 0.0026291804804649314, "grad_norm": 0.164316, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:29.999605+00:00", "epoch": 0, "step": 6811, "train_loss": 3.501662015914917, "perplexity": 33.17053612958177, "lr": 0.0026291804804649314, "grad_norm": 0.133177, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:30.305159+00:00", "epoch": 0, "step": 6812, "train_loss": 3.5147225856781006, "perplexity": 33.606603677472286, "lr": 0.0026291804804649314, "grad_norm": 0.154193, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:30.612004+00:00", "epoch": 0, "step": 6813, "train_loss": 3.545618772506714, "perplexity": 34.66112605997108, "lr": 0.0026291804804649314, "grad_norm": 0.142347, "tokens_per_sec": 106790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:30.919054+00:00", "epoch": 0, "step": 6814, "train_loss": 3.6301939487457275, "perplexity": 37.7201316800132, "lr": 0.0026291804804649314, "grad_norm": 0.158658, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:31.225158+00:00", "epoch": 0, "step": 6815, "train_loss": 3.637552499771118, "perplexity": 37.99872094322645, "lr": 0.0026291804804649314, "grad_norm": 0.162788, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:31.531057+00:00", "epoch": 0, "step": 6816, "train_loss": 3.5609304904937744, "perplexity": 35.195931392872225, "lr": 0.0026291804804649314, "grad_norm": 0.151233, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:31.837770+00:00", "epoch": 0, "step": 6817, "train_loss": 3.4996895790100098, "perplexity": 33.105173822668434, "lr": 0.0026291804804649314, "grad_norm": 0.178159, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:32.143982+00:00", "epoch": 0, "step": 6818, "train_loss": 3.5003905296325684, "perplexity": 33.128387049581825, "lr": 0.0026291804804649314, "grad_norm": 0.177531, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:32.452346+00:00", "epoch": 0, "step": 6819, "train_loss": 3.4772696495056152, "perplexity": 32.37121653819832, "lr": 0.0026291804804649314, "grad_norm": 0.158356, "tokens_per_sec": 106265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:32.758436+00:00", "epoch": 0, "step": 6820, "train_loss": 3.658900737762451, "perplexity": 38.81864753559806, "lr": 0.0026291804804649314, "grad_norm": 0.157367, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:33.062971+00:00", "epoch": 0, "step": 6821, "train_loss": 3.594205379486084, "perplexity": 36.38677481942818, "lr": 0.0026291804804649314, "grad_norm": 0.173001, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:33.367729+00:00", "epoch": 0, "step": 6822, "train_loss": 3.57832407951355, "perplexity": 35.81346999608069, "lr": 0.0026291804804649314, "grad_norm": 0.168558, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:33.672801+00:00", "epoch": 0, "step": 6823, "train_loss": 3.53222918510437, "perplexity": 34.20012110363914, "lr": 0.0026291804804649314, "grad_norm": 0.184025, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:33.980197+00:00", "epoch": 0, "step": 6824, "train_loss": 3.6158735752105713, "perplexity": 37.18381460033636, "lr": 0.0026291804804649314, "grad_norm": 0.178667, "tokens_per_sec": 106599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:34.288694+00:00", "epoch": 0, "step": 6825, "train_loss": 3.5813589096069336, "perplexity": 35.922322884037015, "lr": 0.0026291804804649314, "grad_norm": 0.170709, "tokens_per_sec": 106219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:34.593886+00:00", "epoch": 0, "step": 6826, "train_loss": 3.561917304992676, "perplexity": 35.23068039086068, "lr": 0.0026291804804649314, "grad_norm": 0.183861, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:34.899830+00:00", "epoch": 0, "step": 6827, "train_loss": 3.5961852073669434, "perplexity": 36.45888573075207, "lr": 0.0026291804804649314, "grad_norm": 0.198338, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:35.205595+00:00", "epoch": 0, "step": 6828, "train_loss": 3.6843631267547607, "perplexity": 39.81975423644546, "lr": 0.0026291804804649314, "grad_norm": 0.176175, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:35.511582+00:00", "epoch": 0, "step": 6829, "train_loss": 3.581787109375, "perplexity": 35.937708108103976, "lr": 0.0026291804804649314, "grad_norm": 0.157517, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:35.818072+00:00", "epoch": 0, "step": 6830, "train_loss": 3.6042404174804688, "perplexity": 36.753755741344406, "lr": 0.0026291804804649314, "grad_norm": 0.176517, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:36.124056+00:00", "epoch": 0, "step": 6831, "train_loss": 3.6007471084594727, "perplexity": 36.62558751081424, "lr": 0.0026291804804649314, "grad_norm": 0.157014, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:36.430933+00:00", "epoch": 0, "step": 6832, "train_loss": 3.623213529586792, "perplexity": 37.457746195907504, "lr": 0.0026291804804649314, "grad_norm": 0.158884, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:36.737181+00:00", "epoch": 0, "step": 6833, "train_loss": 3.530377149581909, "perplexity": 34.136839882105896, "lr": 0.0026291804804649314, "grad_norm": 0.154, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:37.042659+00:00", "epoch": 0, "step": 6834, "train_loss": 3.505174160003662, "perplexity": 33.287240653557774, "lr": 0.0026291804804649314, "grad_norm": 0.159434, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:37.349293+00:00", "epoch": 0, "step": 6835, "train_loss": 3.632108211517334, "perplexity": 37.79240707881045, "lr": 0.0026291804804649314, "grad_norm": 0.151061, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:37.655552+00:00", "epoch": 0, "step": 6836, "train_loss": 3.567923069000244, "perplexity": 35.44290418824132, "lr": 0.0026291804804649314, "grad_norm": 0.157846, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:37.961952+00:00", "epoch": 0, "step": 6837, "train_loss": 3.4861106872558594, "perplexity": 32.658680553297415, "lr": 0.0026291804804649314, "grad_norm": 0.167451, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:38.268488+00:00", "epoch": 0, "step": 6838, "train_loss": 3.503561496734619, "perplexity": 33.23360280474687, "lr": 0.0026291804804649314, "grad_norm": 0.168, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:38.576722+00:00", "epoch": 0, "step": 6839, "train_loss": 3.5762102603912354, "perplexity": 35.737846753453226, "lr": 0.0026291804804649314, "grad_norm": 0.184339, "tokens_per_sec": 106320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:38.883021+00:00", "epoch": 0, "step": 6840, "train_loss": 3.595447063446045, "perplexity": 36.431983755873276, "lr": 0.0026291804804649314, "grad_norm": 0.173623, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:39.189949+00:00", "epoch": 0, "step": 6841, "train_loss": 3.5418930053710938, "perplexity": 34.53222704857106, "lr": 0.0026291804804649314, "grad_norm": 0.170704, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:39.496798+00:00", "epoch": 0, "step": 6842, "train_loss": 3.561704397201538, "perplexity": 35.22318030296029, "lr": 0.0026291804804649314, "grad_norm": 0.187553, "tokens_per_sec": 106789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:39.803266+00:00", "epoch": 0, "step": 6843, "train_loss": 3.5714359283447266, "perplexity": 35.56762906763301, "lr": 0.0026291804804649314, "grad_norm": 0.159456, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:40.108732+00:00", "epoch": 0, "step": 6844, "train_loss": 3.6271564960479736, "perplexity": 37.60573239344802, "lr": 0.0026291804804649314, "grad_norm": 0.176737, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:40.414091+00:00", "epoch": 0, "step": 6845, "train_loss": 3.6426713466644287, "perplexity": 38.19372926092695, "lr": 0.0026291804804649314, "grad_norm": 0.187256, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:40.720225+00:00", "epoch": 0, "step": 6846, "train_loss": 3.4674243927001953, "perplexity": 32.05407731847018, "lr": 0.0026291804804649314, "grad_norm": 0.160774, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:41.026272+00:00", "epoch": 0, "step": 6847, "train_loss": 3.5952858924865723, "perplexity": 36.42611245125046, "lr": 0.0026291804804649314, "grad_norm": 0.174433, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:41.333671+00:00", "epoch": 0, "step": 6848, "train_loss": 3.6300806999206543, "perplexity": 37.71586016129564, "lr": 0.0026291804804649314, "grad_norm": 0.177662, "tokens_per_sec": 106597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:41.640484+00:00", "epoch": 0, "step": 6849, "train_loss": 3.5777251720428467, "perplexity": 35.79202746303579, "lr": 0.0026291804804649314, "grad_norm": 0.16167, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:41.946793+00:00", "epoch": 0, "step": 6850, "train_loss": 3.58651065826416, "perplexity": 36.107863180845726, "lr": 0.0026291804804649314, "grad_norm": 0.157127, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:42.251548+00:00", "epoch": 0, "step": 6851, "train_loss": 3.4661195278167725, "perplexity": 32.01227835553706, "lr": 0.0026291804804649314, "grad_norm": 0.147819, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:42.558041+00:00", "epoch": 0, "step": 6852, "train_loss": 3.4619596004486084, "perplexity": 31.87938620517668, "lr": 0.0026291804804649314, "grad_norm": 0.141644, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:42.863243+00:00", "epoch": 0, "step": 6853, "train_loss": 3.4918336868286133, "perplexity": 32.84612202026706, "lr": 0.0026291804804649314, "grad_norm": 0.147849, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:43.170094+00:00", "epoch": 0, "step": 6854, "train_loss": 3.531691312789917, "perplexity": 34.18173075162052, "lr": 0.0026291804804649314, "grad_norm": 0.138566, "tokens_per_sec": 106787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:43.475383+00:00", "epoch": 0, "step": 6855, "train_loss": 3.587473154067993, "perplexity": 36.14263357813915, "lr": 0.0026291804804649314, "grad_norm": 0.156228, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:43.780471+00:00", "epoch": 0, "step": 6856, "train_loss": 3.6675562858581543, "perplexity": 39.15610252851368, "lr": 0.0026291804804649314, "grad_norm": 0.160016, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:44.087577+00:00", "epoch": 0, "step": 6857, "train_loss": 3.5998902320861816, "perplexity": 36.594217352311574, "lr": 0.0026291804804649314, "grad_norm": 0.151744, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:44.393205+00:00", "epoch": 0, "step": 6858, "train_loss": 3.569667100906372, "perplexity": 35.50477167774769, "lr": 0.0026291804804649314, "grad_norm": 0.216933, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:44.700717+00:00", "epoch": 0, "step": 6859, "train_loss": 3.594341278076172, "perplexity": 36.39172006684245, "lr": 0.0026291804804649314, "grad_norm": 0.234729, "tokens_per_sec": 106558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:45.006908+00:00", "epoch": 0, "step": 6860, "train_loss": 3.610970973968506, "perplexity": 37.00196332108616, "lr": 0.0026291804804649314, "grad_norm": 0.184341, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:45.312975+00:00", "epoch": 0, "step": 6861, "train_loss": 3.6191048622131348, "perplexity": 37.30416050853153, "lr": 0.0026291804804649314, "grad_norm": 0.144675, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:45.619213+00:00", "epoch": 0, "step": 6862, "train_loss": 3.5243074893951416, "perplexity": 33.930268407685794, "lr": 0.0026291804804649314, "grad_norm": 0.158831, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:45.925743+00:00", "epoch": 0, "step": 6863, "train_loss": 3.559844970703125, "perplexity": 35.15774624191309, "lr": 0.0026291804804649314, "grad_norm": 0.141558, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:46.232963+00:00", "epoch": 0, "step": 6864, "train_loss": 3.4887921810150146, "perplexity": 32.7463721210343, "lr": 0.0026291804804649314, "grad_norm": 0.170909, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:46.540245+00:00", "epoch": 0, "step": 6865, "train_loss": 3.6415021419525146, "perplexity": 38.1490990687093, "lr": 0.0026291804804649314, "grad_norm": 0.179569, "tokens_per_sec": 106639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:46.846948+00:00", "epoch": 0, "step": 6866, "train_loss": 3.5755467414855957, "perplexity": 35.714141881670436, "lr": 0.0026291804804649314, "grad_norm": 0.185617, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:47.151679+00:00", "epoch": 0, "step": 6867, "train_loss": 3.558213710784912, "perplexity": 35.10044157176708, "lr": 0.0026291804804649314, "grad_norm": 0.175362, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:47.456889+00:00", "epoch": 0, "step": 6868, "train_loss": 3.4780797958374023, "perplexity": 32.397452586616446, "lr": 0.0026291804804649314, "grad_norm": 0.151544, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:47.763040+00:00", "epoch": 0, "step": 6869, "train_loss": 3.587080955505371, "perplexity": 36.12846126856181, "lr": 0.0026291804804649314, "grad_norm": 0.15306, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:48.069133+00:00", "epoch": 0, "step": 6870, "train_loss": 3.6517138481140137, "perplexity": 38.54066232051617, "lr": 0.0026291804804649314, "grad_norm": 0.16988, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:48.374866+00:00", "epoch": 0, "step": 6871, "train_loss": 3.615224838256836, "perplexity": 37.15969990861585, "lr": 0.0026291804804649314, "grad_norm": 0.159197, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:48.681046+00:00", "epoch": 0, "step": 6872, "train_loss": 3.523625373840332, "perplexity": 33.90713193559688, "lr": 0.0026291804804649314, "grad_norm": 0.198653, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:48.986423+00:00", "epoch": 0, "step": 6873, "train_loss": 3.534924268722534, "perplexity": 34.29241760729619, "lr": 0.0026291804804649314, "grad_norm": 0.161308, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:49.296657+00:00", "epoch": 0, "step": 6874, "train_loss": 3.551894187927246, "perplexity": 34.879322946931175, "lr": 0.0026291804804649314, "grad_norm": 0.175154, "tokens_per_sec": 105626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:49.603429+00:00", "epoch": 0, "step": 6875, "train_loss": 3.6598541736602783, "perplexity": 38.85567627712426, "lr": 0.0026291804804649314, "grad_norm": 0.150401, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:49.910949+00:00", "epoch": 0, "step": 6876, "train_loss": 3.617443084716797, "perplexity": 37.242220773361595, "lr": 0.0026291804804649314, "grad_norm": 0.163367, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:50.217331+00:00", "epoch": 0, "step": 6877, "train_loss": 3.473745107650757, "perplexity": 32.2573236586291, "lr": 0.0026291804804649314, "grad_norm": 0.144659, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:50.522362+00:00", "epoch": 0, "step": 6878, "train_loss": 3.5366883277893066, "perplexity": 34.35296484624937, "lr": 0.0026291804804649314, "grad_norm": 0.158239, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:50.827497+00:00", "epoch": 0, "step": 6879, "train_loss": 3.5230541229248047, "perplexity": 33.88776798679908, "lr": 0.0026291804804649314, "grad_norm": 0.169417, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:51.133450+00:00", "epoch": 0, "step": 6880, "train_loss": 3.622088670730591, "perplexity": 37.41563520726524, "lr": 0.0026291804804649314, "grad_norm": 0.144018, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:51.440258+00:00", "epoch": 0, "step": 6881, "train_loss": 3.531186819076538, "perplexity": 34.164490632475434, "lr": 0.0026291804804649314, "grad_norm": 0.174168, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:51.746383+00:00", "epoch": 0, "step": 6882, "train_loss": 3.522392988204956, "perplexity": 33.8653710113297, "lr": 0.0026291804804649314, "grad_norm": 0.166158, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:52.051129+00:00", "epoch": 0, "step": 6883, "train_loss": 3.6398861408233643, "perplexity": 38.08749986713316, "lr": 0.0026291804804649314, "grad_norm": 0.147296, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:52.356660+00:00", "epoch": 0, "step": 6884, "train_loss": 3.54872727394104, "perplexity": 34.7690378551349, "lr": 0.0026291804804649314, "grad_norm": 0.169531, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:52.662068+00:00", "epoch": 0, "step": 6885, "train_loss": 3.5097177028656006, "perplexity": 33.43882676644767, "lr": 0.0026291804804649314, "grad_norm": 0.182653, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:52.969051+00:00", "epoch": 0, "step": 6886, "train_loss": 3.6272616386413574, "perplexity": 37.60968656555028, "lr": 0.0026291804804649314, "grad_norm": 0.15069, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:53.275161+00:00", "epoch": 0, "step": 6887, "train_loss": 3.6126177310943604, "perplexity": 37.06294676653616, "lr": 0.0026291804804649314, "grad_norm": 0.168965, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:53.581902+00:00", "epoch": 0, "step": 6888, "train_loss": 3.498129367828369, "perplexity": 33.05356303263551, "lr": 0.0026291804804649314, "grad_norm": 0.169299, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:53.887258+00:00", "epoch": 0, "step": 6889, "train_loss": 3.6860485076904297, "perplexity": 39.88692227706689, "lr": 0.0026291804804649314, "grad_norm": 0.160797, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:54.192789+00:00", "epoch": 0, "step": 6890, "train_loss": 3.4868311882019043, "perplexity": 32.68221964248236, "lr": 0.0026291804804649314, "grad_norm": 0.173889, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:54.497887+00:00", "epoch": 0, "step": 6891, "train_loss": 3.630448818206787, "perplexity": 37.72974661487017, "lr": 0.0026291804804649314, "grad_norm": 0.163611, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:54.803938+00:00", "epoch": 0, "step": 6892, "train_loss": 3.559116840362549, "perplexity": 35.1321561377643, "lr": 0.0026291804804649314, "grad_norm": 0.159468, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:55.111147+00:00", "epoch": 0, "step": 6893, "train_loss": 3.611377477645874, "perplexity": 37.01700781285947, "lr": 0.0026291804804649314, "grad_norm": 0.169641, "tokens_per_sec": 106661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:55.415488+00:00", "epoch": 0, "step": 6894, "train_loss": 3.434823751449585, "perplexity": 31.025943805990693, "lr": 0.0026291804804649314, "grad_norm": 0.149292, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:55.721923+00:00", "epoch": 0, "step": 6895, "train_loss": 3.5709433555603027, "perplexity": 35.55011373568987, "lr": 0.0026291804804649314, "grad_norm": 0.14142, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:56.027037+00:00", "epoch": 0, "step": 6896, "train_loss": 3.576552629470825, "perplexity": 35.75008438192599, "lr": 0.0026291804804649314, "grad_norm": 0.155225, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:56.333786+00:00", "epoch": 0, "step": 6897, "train_loss": 3.6002442836761475, "perplexity": 36.60717588700825, "lr": 0.0026291804804649314, "grad_norm": 0.157147, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:56.640245+00:00", "epoch": 0, "step": 6898, "train_loss": 3.547849416732788, "perplexity": 34.73852899779921, "lr": 0.0026291804804649314, "grad_norm": 0.164887, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:56.947049+00:00", "epoch": 0, "step": 6899, "train_loss": 3.604076862335205, "perplexity": 36.74774496704507, "lr": 0.0026291804804649314, "grad_norm": 0.16545, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:57.252506+00:00", "epoch": 0, "step": 6900, "train_loss": 3.4945454597473145, "perplexity": 32.93531412415417, "lr": 0.0026291804804649314, "grad_norm": 0.175239, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:57.558630+00:00", "epoch": 0, "step": 6901, "train_loss": 3.5669147968292236, "perplexity": 35.407186104089256, "lr": 0.0026291804804649314, "grad_norm": 0.167654, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:57.864154+00:00", "epoch": 0, "step": 6902, "train_loss": 3.5715177059173584, "perplexity": 35.5705378209362, "lr": 0.0026291804804649314, "grad_norm": 0.150449, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:58.171177+00:00", "epoch": 0, "step": 6903, "train_loss": 3.5906026363372803, "perplexity": 36.255918478174166, "lr": 0.0026291804804649314, "grad_norm": 0.165736, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:58.478267+00:00", "epoch": 0, "step": 6904, "train_loss": 3.4532601833343506, "perplexity": 31.603256950463138, "lr": 0.0026291804804649314, "grad_norm": 0.138643, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:58.784458+00:00", "epoch": 0, "step": 6905, "train_loss": 3.623840570449829, "perplexity": 37.48124109877188, "lr": 0.0026291804804649314, "grad_norm": 0.153329, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:59.091599+00:00", "epoch": 0, "step": 6906, "train_loss": 3.514763832092285, "perplexity": 33.607989857954195, "lr": 0.0026291804804649314, "grad_norm": 0.160281, "tokens_per_sec": 106694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:59.397738+00:00", "epoch": 0, "step": 6907, "train_loss": 3.5409445762634277, "perplexity": 34.49949120555272, "lr": 0.0026291804804649314, "grad_norm": 0.147666, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:33:59.704009+00:00", "epoch": 0, "step": 6908, "train_loss": 3.578356981277466, "perplexity": 35.814648341800236, "lr": 0.0026291804804649314, "grad_norm": 0.158588, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:00.011296+00:00", "epoch": 0, "step": 6909, "train_loss": 3.6115481853485107, "perplexity": 37.023327440610906, "lr": 0.0026291804804649314, "grad_norm": 0.176025, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:00.316916+00:00", "epoch": 0, "step": 6910, "train_loss": 3.5844788551330566, "perplexity": 36.034573591598864, "lr": 0.0026291804804649314, "grad_norm": 0.156125, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:00.622667+00:00", "epoch": 0, "step": 6911, "train_loss": 3.5541458129882812, "perplexity": 34.95794658685217, "lr": 0.0026291804804649314, "grad_norm": 0.142262, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:00.928866+00:00", "epoch": 0, "step": 6912, "train_loss": 3.6034183502197266, "perplexity": 36.72355409763223, "lr": 0.0026291804804649314, "grad_norm": 0.154363, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:01.234808+00:00", "epoch": 0, "step": 6913, "train_loss": 3.5624377727508545, "perplexity": 35.249021596691925, "lr": 0.0026291804804649314, "grad_norm": 0.167454, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:01.540888+00:00", "epoch": 0, "step": 6914, "train_loss": 3.5362918376922607, "perplexity": 34.3393469357454, "lr": 0.0026291804804649314, "grad_norm": 0.150528, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:01.848894+00:00", "epoch": 0, "step": 6915, "train_loss": 3.6024703979492188, "perplexity": 36.68875841607259, "lr": 0.0026291804804649314, "grad_norm": 0.15822, "tokens_per_sec": 106388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:02.155327+00:00", "epoch": 0, "step": 6916, "train_loss": 3.5330445766448975, "perplexity": 34.228018965366594, "lr": 0.0026291804804649314, "grad_norm": 0.181318, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:02.461704+00:00", "epoch": 0, "step": 6917, "train_loss": 3.628883123397827, "perplexity": 37.67071956768082, "lr": 0.0026291804804649314, "grad_norm": 0.174362, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:02.768316+00:00", "epoch": 0, "step": 6918, "train_loss": 3.5663375854492188, "perplexity": 35.38675457056013, "lr": 0.0026291804804649314, "grad_norm": 0.177264, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:03.075347+00:00", "epoch": 0, "step": 6919, "train_loss": 3.505269765853882, "perplexity": 33.290423260637425, "lr": 0.0026291804804649314, "grad_norm": 0.160656, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:03.380970+00:00", "epoch": 0, "step": 6920, "train_loss": 3.5790443420410156, "perplexity": 35.83927438835022, "lr": 0.0026291804804649314, "grad_norm": 0.190976, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:03.688699+00:00", "epoch": 0, "step": 6921, "train_loss": 3.5236988067626953, "perplexity": 33.909621926806416, "lr": 0.0026291804804649314, "grad_norm": 0.183021, "tokens_per_sec": 106415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:03.995446+00:00", "epoch": 0, "step": 6922, "train_loss": 3.7027976512908936, "perplexity": 40.56062024858551, "lr": 0.0026291804804649314, "grad_norm": 0.155664, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:04.301009+00:00", "epoch": 0, "step": 6923, "train_loss": 3.5876193046569824, "perplexity": 36.14791623134625, "lr": 0.0026291804804649314, "grad_norm": 0.156357, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:04.605864+00:00", "epoch": 0, "step": 6924, "train_loss": 3.612220525741577, "perplexity": 37.04822808905283, "lr": 0.0026291804804649314, "grad_norm": 0.169446, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:04.913698+00:00", "epoch": 0, "step": 6925, "train_loss": 3.5074405670166016, "perplexity": 33.36276864547191, "lr": 0.0026291804804649314, "grad_norm": 0.169122, "tokens_per_sec": 106447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:05.219319+00:00", "epoch": 0, "step": 6926, "train_loss": 3.5875158309936523, "perplexity": 36.144176067539604, "lr": 0.0026291804804649314, "grad_norm": 0.145171, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:05.524916+00:00", "epoch": 0, "step": 6927, "train_loss": 3.527442693710327, "perplexity": 34.03681366491416, "lr": 0.0026291804804649314, "grad_norm": 0.150652, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:05.831079+00:00", "epoch": 0, "step": 6928, "train_loss": 3.586923122406006, "perplexity": 36.12275945152406, "lr": 0.0026291804804649314, "grad_norm": 0.158458, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:06.136590+00:00", "epoch": 0, "step": 6929, "train_loss": 3.5307321548461914, "perplexity": 34.148960791335995, "lr": 0.0026291804804649314, "grad_norm": 0.16036, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:06.442983+00:00", "epoch": 0, "step": 6930, "train_loss": 3.549468755722046, "perplexity": 34.79482802353597, "lr": 0.0026291804804649314, "grad_norm": 0.150657, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:06.749100+00:00", "epoch": 0, "step": 6931, "train_loss": 3.5074620246887207, "perplexity": 33.36348454050319, "lr": 0.0026291804804649314, "grad_norm": 0.148711, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:07.056262+00:00", "epoch": 0, "step": 6932, "train_loss": 3.577422618865967, "perplexity": 35.78120010942758, "lr": 0.0026291804804649314, "grad_norm": 0.161155, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:07.365125+00:00", "epoch": 0, "step": 6933, "train_loss": 3.656977415084839, "perplexity": 38.74405850286081, "lr": 0.0026291804804649314, "grad_norm": 0.15813, "tokens_per_sec": 106092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:07.671147+00:00", "epoch": 0, "step": 6934, "train_loss": 3.513787031173706, "perplexity": 33.57517757073445, "lr": 0.0026291804804649314, "grad_norm": 0.185839, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:07.978402+00:00", "epoch": 0, "step": 6935, "train_loss": 3.563682794570923, "perplexity": 35.29293472844708, "lr": 0.0026291804804649314, "grad_norm": 0.21487, "tokens_per_sec": 106648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:08.286399+00:00", "epoch": 0, "step": 6936, "train_loss": 3.625237464904785, "perplexity": 37.53363502249864, "lr": 0.0026291804804649314, "grad_norm": 0.158873, "tokens_per_sec": 106391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:08.593733+00:00", "epoch": 0, "step": 6937, "train_loss": 3.49772310256958, "perplexity": 33.040137245696215, "lr": 0.0026291804804649314, "grad_norm": 0.176904, "tokens_per_sec": 106621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:08.898599+00:00", "epoch": 0, "step": 6938, "train_loss": 3.4947774410247803, "perplexity": 32.94295538667908, "lr": 0.0026291804804649314, "grad_norm": 0.195393, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:09.203798+00:00", "epoch": 0, "step": 6939, "train_loss": 3.544994354248047, "perplexity": 34.63948977574174, "lr": 0.0026291804804649314, "grad_norm": 0.159421, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:09.510352+00:00", "epoch": 0, "step": 6940, "train_loss": 3.5117180347442627, "perplexity": 33.50578246209391, "lr": 0.0026291804804649314, "grad_norm": 0.157476, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:09.816109+00:00", "epoch": 0, "step": 6941, "train_loss": 3.5263376235961914, "perplexity": 33.99922137423672, "lr": 0.0026291804804649314, "grad_norm": 0.169623, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:10.124103+00:00", "epoch": 0, "step": 6942, "train_loss": 3.476386785507202, "perplexity": 32.34264976867362, "lr": 0.0026291804804649314, "grad_norm": 0.141225, "tokens_per_sec": 106393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:10.429887+00:00", "epoch": 0, "step": 6943, "train_loss": 3.543707847595215, "perplexity": 34.59495449530434, "lr": 0.0026291804804649314, "grad_norm": 0.164068, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:10.736407+00:00", "epoch": 0, "step": 6944, "train_loss": 3.633889675140381, "perplexity": 37.859792882106, "lr": 0.0026291804804649314, "grad_norm": 0.147037, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:11.044401+00:00", "epoch": 0, "step": 6945, "train_loss": 3.6557090282440186, "perplexity": 38.69494720154418, "lr": 0.0026291804804649314, "grad_norm": 0.151421, "tokens_per_sec": 106394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:11.350367+00:00", "epoch": 0, "step": 6946, "train_loss": 3.6086833477020264, "perplexity": 36.917413404059495, "lr": 0.0026291804804649314, "grad_norm": 0.161298, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:11.656866+00:00", "epoch": 0, "step": 6947, "train_loss": 3.6593074798583984, "perplexity": 38.834439925155614, "lr": 0.0026291804804649314, "grad_norm": 0.144176, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:11.962986+00:00", "epoch": 0, "step": 6948, "train_loss": 3.6849193572998047, "perplexity": 39.841909361155246, "lr": 0.0026291804804649314, "grad_norm": 0.144421, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:12.270570+00:00", "epoch": 0, "step": 6949, "train_loss": 3.5647950172424316, "perplexity": 35.33221016806455, "lr": 0.0026291804804649314, "grad_norm": 0.14179, "tokens_per_sec": 106532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:12.576168+00:00", "epoch": 0, "step": 6950, "train_loss": 3.589388847351074, "perplexity": 36.21193814046496, "lr": 0.0026291804804649314, "grad_norm": 0.160517, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:12.883290+00:00", "epoch": 0, "step": 6951, "train_loss": 3.5188372135162354, "perplexity": 33.74516721771692, "lr": 0.0026291804804649314, "grad_norm": 0.147349, "tokens_per_sec": 106636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:13.189746+00:00", "epoch": 0, "step": 6952, "train_loss": 3.6381773948669434, "perplexity": 38.02247357827114, "lr": 0.0026291804804649314, "grad_norm": 0.173273, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:13.495325+00:00", "epoch": 0, "step": 6953, "train_loss": 3.597884178161621, "perplexity": 36.520880961946276, "lr": 0.0026291804804649314, "grad_norm": 0.176599, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:13.801775+00:00", "epoch": 0, "step": 6954, "train_loss": 3.541795015335083, "perplexity": 34.52884340018365, "lr": 0.0026291804804649314, "grad_norm": 0.176358, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:14.108541+00:00", "epoch": 0, "step": 6955, "train_loss": 3.5066840648651123, "perplexity": 33.337539183486335, "lr": 0.0026291804804649314, "grad_norm": 0.148637, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:14.415310+00:00", "epoch": 0, "step": 6956, "train_loss": 3.6040470600128174, "perplexity": 36.74664981522166, "lr": 0.0026291804804649314, "grad_norm": 0.160316, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:14.721677+00:00", "epoch": 0, "step": 6957, "train_loss": 3.6126232147216797, "perplexity": 37.063150006480825, "lr": 0.0026291804804649314, "grad_norm": 0.142628, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:15.027850+00:00", "epoch": 0, "step": 6958, "train_loss": 3.446035861968994, "perplexity": 31.375767583611566, "lr": 0.0026291804804649314, "grad_norm": 0.144578, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:15.637300+00:00", "epoch": 0, "step": 6959, "train_loss": 3.5662801265716553, "perplexity": 35.38472134577585, "lr": 0.0026291804804649314, "grad_norm": 0.165621, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:15.944483+00:00", "epoch": 0, "step": 6960, "train_loss": 3.514819622039795, "perplexity": 33.60986489824799, "lr": 0.0026291804804649314, "grad_norm": 0.174988, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:16.250100+00:00", "epoch": 0, "step": 6961, "train_loss": 3.5286638736724854, "perplexity": 34.078404129289744, "lr": 0.0026291804804649314, "grad_norm": 0.164438, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:16.556729+00:00", "epoch": 0, "step": 6962, "train_loss": 3.5749130249023438, "perplexity": 35.69151640752968, "lr": 0.0026291804804649314, "grad_norm": 0.154809, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:16.862949+00:00", "epoch": 0, "step": 6963, "train_loss": 3.5460925102233887, "perplexity": 34.677550232756, "lr": 0.0026291804804649314, "grad_norm": 0.141305, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:17.169542+00:00", "epoch": 0, "step": 6964, "train_loss": 3.4739596843719482, "perplexity": 32.264246072042, "lr": 0.0026291804804649314, "grad_norm": 0.14407, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:17.474343+00:00", "epoch": 0, "step": 6965, "train_loss": 3.621713638305664, "perplexity": 37.401605761776196, "lr": 0.0026291804804649314, "grad_norm": 0.152336, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:17.780441+00:00", "epoch": 0, "step": 6966, "train_loss": 3.6093645095825195, "perplexity": 36.94256870524025, "lr": 0.0026291804804649314, "grad_norm": 0.153555, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:18.086873+00:00", "epoch": 0, "step": 6967, "train_loss": 3.5480353832244873, "perplexity": 34.74498980089154, "lr": 0.0026291804804649314, "grad_norm": 0.174687, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:18.391655+00:00", "epoch": 0, "step": 6968, "train_loss": 3.4923033714294434, "perplexity": 32.86155296153103, "lr": 0.0026291804804649314, "grad_norm": 0.16875, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:18.697377+00:00", "epoch": 0, "step": 6969, "train_loss": 3.7146215438842773, "perplexity": 41.043051151027655, "lr": 0.0026291804804649314, "grad_norm": 0.154138, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:19.004785+00:00", "epoch": 0, "step": 6970, "train_loss": 3.515423536300659, "perplexity": 33.63016850516757, "lr": 0.0026291804804649314, "grad_norm": 0.170639, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:19.311559+00:00", "epoch": 0, "step": 6971, "train_loss": 3.559434652328491, "perplexity": 35.143323331814194, "lr": 0.0026291804804649314, "grad_norm": 0.184871, "tokens_per_sec": 106726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:19.617819+00:00", "epoch": 0, "step": 6972, "train_loss": 3.6291918754577637, "perplexity": 37.6823522756655, "lr": 0.0026291804804649314, "grad_norm": 0.177428, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:19.923259+00:00", "epoch": 0, "step": 6973, "train_loss": 3.6493728160858154, "perplexity": 38.45054292299161, "lr": 0.0026291804804649314, "grad_norm": 0.164481, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:20.229965+00:00", "epoch": 0, "step": 6974, "train_loss": 3.466170072555542, "perplexity": 32.01389644867666, "lr": 0.0026291804804649314, "grad_norm": 0.157977, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:20.536978+00:00", "epoch": 0, "step": 6975, "train_loss": 3.4831812381744385, "perplexity": 32.5631486078537, "lr": 0.0026291804804649314, "grad_norm": 0.194494, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:20.844362+00:00", "epoch": 0, "step": 6976, "train_loss": 3.6040351390838623, "perplexity": 36.746211763630875, "lr": 0.0026291804804649314, "grad_norm": 0.17475, "tokens_per_sec": 106603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:21.151290+00:00", "epoch": 0, "step": 6977, "train_loss": 3.599862813949585, "perplexity": 36.59321402081636, "lr": 0.0026291804804649314, "grad_norm": 0.172567, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:21.457260+00:00", "epoch": 0, "step": 6978, "train_loss": 3.553100109100342, "perplexity": 34.921410032729646, "lr": 0.0026291804804649314, "grad_norm": 0.184057, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:21.762837+00:00", "epoch": 0, "step": 6979, "train_loss": 3.5499582290649414, "perplexity": 34.81186333314898, "lr": 0.0026291804804649314, "grad_norm": 0.158807, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:22.069262+00:00", "epoch": 0, "step": 6980, "train_loss": 3.569673538208008, "perplexity": 35.505000233408126, "lr": 0.0026291804804649314, "grad_norm": 0.153328, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:22.375219+00:00", "epoch": 0, "step": 6981, "train_loss": 3.6143956184387207, "perplexity": 37.12889912109302, "lr": 0.0026291804804649314, "grad_norm": 0.165946, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:22.680341+00:00", "epoch": 0, "step": 6982, "train_loss": 3.5930428504943848, "perplexity": 36.344498717144695, "lr": 0.0026291804804649314, "grad_norm": 0.156315, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:22.986359+00:00", "epoch": 0, "step": 6983, "train_loss": 3.6066393852233887, "perplexity": 36.84203266020297, "lr": 0.0026291804804649314, "grad_norm": 0.148697, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:23.291424+00:00", "epoch": 0, "step": 6984, "train_loss": 3.7236196994781494, "perplexity": 41.414029468569424, "lr": 0.0026291804804649314, "grad_norm": 0.172422, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:23.596239+00:00", "epoch": 0, "step": 6985, "train_loss": 3.5372211933135986, "perplexity": 34.37127523492732, "lr": 0.0026291804804649314, "grad_norm": 0.188428, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:23.902834+00:00", "epoch": 0, "step": 6986, "train_loss": 3.5635454654693604, "perplexity": 35.28808831421379, "lr": 0.0026291804804649314, "grad_norm": 0.144076, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:24.209134+00:00", "epoch": 0, "step": 6987, "train_loss": 3.579066753387451, "perplexity": 35.84007760374507, "lr": 0.0026291804804649314, "grad_norm": 0.164086, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:24.514339+00:00", "epoch": 0, "step": 6988, "train_loss": 3.443340539932251, "perplexity": 31.291313652223035, "lr": 0.0026291804804649314, "grad_norm": 0.14308, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:24.819124+00:00", "epoch": 0, "step": 6989, "train_loss": 3.544802665710449, "perplexity": 34.63285041896812, "lr": 0.0026291804804649314, "grad_norm": 0.15374, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:25.125049+00:00", "epoch": 0, "step": 6990, "train_loss": 3.595797061920166, "perplexity": 36.44473712629694, "lr": 0.0026291804804649314, "grad_norm": 0.13663, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:25.429407+00:00", "epoch": 0, "step": 6991, "train_loss": 3.532825231552124, "perplexity": 34.220512040704826, "lr": 0.0026291804804649314, "grad_norm": 0.151553, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:25.735524+00:00", "epoch": 0, "step": 6992, "train_loss": 3.4780914783477783, "perplexity": 32.39783107240327, "lr": 0.0026291804804649314, "grad_norm": 0.175284, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:26.042391+00:00", "epoch": 0, "step": 6993, "train_loss": 3.5127804279327393, "perplexity": 33.541397692495465, "lr": 0.0026291804804649314, "grad_norm": 0.146748, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:26.347824+00:00", "epoch": 0, "step": 6994, "train_loss": 3.481290340423584, "perplexity": 32.50163320139123, "lr": 0.0026291804804649314, "grad_norm": 0.16494, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:26.653733+00:00", "epoch": 0, "step": 6995, "train_loss": 3.575921058654785, "perplexity": 35.727512800485734, "lr": 0.0026291804804649314, "grad_norm": 0.166893, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:26.958753+00:00", "epoch": 0, "step": 6996, "train_loss": 3.577877998352051, "perplexity": 35.79749784449035, "lr": 0.0026291804804649314, "grad_norm": 0.169039, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:27.264561+00:00", "epoch": 0, "step": 6997, "train_loss": 3.458812713623047, "perplexity": 31.779223068407127, "lr": 0.0026291804804649314, "grad_norm": 0.148915, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:27.569925+00:00", "epoch": 0, "step": 6998, "train_loss": 3.5862877368927, "perplexity": 36.09981486356926, "lr": 0.0026291804804649314, "grad_norm": 0.150688, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:27.875399+00:00", "epoch": 0, "step": 6999, "train_loss": 3.56997013092041, "perplexity": 35.515532319523984, "lr": 0.0026291804804649314, "grad_norm": 0.166964, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:28.181459+00:00", "epoch": 0, "step": 7000, "train_loss": 3.5975074768066406, "perplexity": 36.507126087505405, "lr": 0.0026291804804649314, "grad_norm": 0.184268, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:34:31.254139+00:00", "step": 7000, "epoch": 0, "val_loss": 3.5255451083183287, "val_ppl": 33.972287146165556, "eval_train_loss": 3.5975074768066406, "eval_train_ppl": 36.507126087505405} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:34:32.323622+00:00", "step": 7000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5255_epoch_0000_step_0007000.pt", "val_loss": 3.5255451083183287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:33.505852+00:00", "epoch": 0, "step": 7001, "train_loss": 3.7231457233428955, "perplexity": 41.39440485810253, "lr": 0.0026291804804649314, "grad_norm": 0.175993, "tokens_per_sec": 6154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:33.811048+00:00", "epoch": 0, "step": 7002, "train_loss": 3.5452768802642822, "perplexity": 34.64927771540119, "lr": 0.0026291804804649314, "grad_norm": 0.166743, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:34.117037+00:00", "epoch": 0, "step": 7003, "train_loss": 3.575241804122925, "perplexity": 35.70325296573802, "lr": 0.0026291804804649314, "grad_norm": 0.168335, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:34.421293+00:00", "epoch": 0, "step": 7004, "train_loss": 3.550006151199341, "perplexity": 34.81353163191623, "lr": 0.0026291804804649314, "grad_norm": 0.195144, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:34.726023+00:00", "epoch": 0, "step": 7005, "train_loss": 3.614823341369629, "perplexity": 37.14478339943937, "lr": 0.0026291804804649314, "grad_norm": 0.166867, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:35.030746+00:00", "epoch": 0, "step": 7006, "train_loss": 3.609393835067749, "perplexity": 36.94365207987832, "lr": 0.0026291804804649314, "grad_norm": 0.149946, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:35.335473+00:00", "epoch": 0, "step": 7007, "train_loss": 3.4979307651519775, "perplexity": 33.04699915837602, "lr": 0.0026291804804649314, "grad_norm": 0.168811, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:35.640263+00:00", "epoch": 0, "step": 7008, "train_loss": 3.4952914714813232, "perplexity": 32.95989342203644, "lr": 0.0026291804804649314, "grad_norm": 0.146422, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:35.944143+00:00", "epoch": 0, "step": 7009, "train_loss": 3.6049437522888184, "perplexity": 36.77961502989774, "lr": 0.0026291804804649314, "grad_norm": 0.164859, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:36.248465+00:00", "epoch": 0, "step": 7010, "train_loss": 3.622274398803711, "perplexity": 37.42258498646142, "lr": 0.0026291804804649314, "grad_norm": 0.146296, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:36.552679+00:00", "epoch": 0, "step": 7011, "train_loss": 3.5122270584106445, "perplexity": 33.52284203982381, "lr": 0.0026291804804649314, "grad_norm": 0.146795, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:36.858296+00:00", "epoch": 0, "step": 7012, "train_loss": 3.547724962234497, "perplexity": 34.73420590062144, "lr": 0.0026291804804649314, "grad_norm": 0.159463, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:37.165178+00:00", "epoch": 0, "step": 7013, "train_loss": 3.540173292160034, "perplexity": 34.47289255528736, "lr": 0.0026291804804649314, "grad_norm": 0.137506, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:37.572968+00:00", "epoch": 0, "step": 7014, "train_loss": 3.4397048950195312, "perplexity": 31.177756099483275, "lr": 0.0026291804804649314, "grad_norm": 0.151406, "tokens_per_sec": 80354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:37.878346+00:00", "epoch": 0, "step": 7015, "train_loss": 3.4956161975860596, "perplexity": 32.97059809778966, "lr": 0.0026291804804649314, "grad_norm": 0.147038, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:38.182678+00:00", "epoch": 0, "step": 7016, "train_loss": 3.6721746921539307, "perplexity": 39.33735955606867, "lr": 0.0026291804804649314, "grad_norm": 0.170094, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:38.490067+00:00", "epoch": 0, "step": 7017, "train_loss": 3.5553512573242188, "perplexity": 35.00011185439691, "lr": 0.0026291804804649314, "grad_norm": 0.186305, "tokens_per_sec": 106602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:38.795676+00:00", "epoch": 0, "step": 7018, "train_loss": 3.528677463531494, "perplexity": 34.07886725314399, "lr": 0.0026291804804649314, "grad_norm": 0.17248, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:39.100763+00:00", "epoch": 0, "step": 7019, "train_loss": 3.6168620586395264, "perplexity": 37.22058835702439, "lr": 0.0026291804804649314, "grad_norm": 0.176575, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:39.407862+00:00", "epoch": 0, "step": 7020, "train_loss": 3.59722900390625, "perplexity": 36.496961257599416, "lr": 0.0026291804804649314, "grad_norm": 0.149073, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:39.713984+00:00", "epoch": 0, "step": 7021, "train_loss": 3.5317649841308594, "perplexity": 34.184249058323104, "lr": 0.0026291804804649314, "grad_norm": 0.169851, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:40.019204+00:00", "epoch": 0, "step": 7022, "train_loss": 3.531592607498169, "perplexity": 34.17835700042052, "lr": 0.0026291804804649314, "grad_norm": 0.173986, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:40.323529+00:00", "epoch": 0, "step": 7023, "train_loss": 3.674593210220337, "perplexity": 39.43261281026798, "lr": 0.0026291804804649314, "grad_norm": 0.163808, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:40.629194+00:00", "epoch": 0, "step": 7024, "train_loss": 3.651505947113037, "perplexity": 38.53265051110189, "lr": 0.0026291804804649314, "grad_norm": 0.172906, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:40.933982+00:00", "epoch": 0, "step": 7025, "train_loss": 3.5553691387176514, "perplexity": 35.00073771076274, "lr": 0.0026291804804649314, "grad_norm": 0.171332, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:41.239754+00:00", "epoch": 0, "step": 7026, "train_loss": 3.5805609226226807, "perplexity": 35.893668772256085, "lr": 0.0026291804804649314, "grad_norm": 0.161546, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:41.544090+00:00", "epoch": 0, "step": 7027, "train_loss": 3.453695058822632, "perplexity": 31.617003421045663, "lr": 0.0026291804804649314, "grad_norm": 0.166907, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:41.848536+00:00", "epoch": 0, "step": 7028, "train_loss": 3.574558734893799, "perplexity": 35.6788734996369, "lr": 0.0026291804804649314, "grad_norm": 0.148661, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:42.152804+00:00", "epoch": 0, "step": 7029, "train_loss": 3.5427305698394775, "perplexity": 34.56116213076794, "lr": 0.0026291804804649314, "grad_norm": 0.162235, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:42.467015+00:00", "epoch": 0, "step": 7030, "train_loss": 3.585639715194702, "perplexity": 36.076428978342854, "lr": 0.0026291804804649314, "grad_norm": 0.15693, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:42.771113+00:00", "epoch": 0, "step": 7031, "train_loss": 3.5938425064086914, "perplexity": 36.37357343383163, "lr": 0.0026291804804649314, "grad_norm": 0.14135, "tokens_per_sec": 104960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:43.076719+00:00", "epoch": 0, "step": 7032, "train_loss": 3.5740392208099365, "perplexity": 35.66034263629653, "lr": 0.0026291804804649314, "grad_norm": 0.146829, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:43.384083+00:00", "epoch": 0, "step": 7033, "train_loss": 3.53905987739563, "perplexity": 34.434531287708445, "lr": 0.0026291804804649314, "grad_norm": 0.165858, "tokens_per_sec": 106609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:43.689963+00:00", "epoch": 0, "step": 7034, "train_loss": 3.7846405506134033, "perplexity": 44.01984480842773, "lr": 0.0026291804804649314, "grad_norm": 0.169107, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:43.996529+00:00", "epoch": 0, "step": 7035, "train_loss": 3.5797338485717773, "perplexity": 35.86399432339772, "lr": 0.0026291804804649314, "grad_norm": 0.170547, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:44.302219+00:00", "epoch": 0, "step": 7036, "train_loss": 3.636002540588379, "perplexity": 37.93987009675491, "lr": 0.0026291804804649314, "grad_norm": 0.163678, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:44.607013+00:00", "epoch": 0, "step": 7037, "train_loss": 3.60870361328125, "perplexity": 36.918161564406496, "lr": 0.0026291804804649314, "grad_norm": 0.151212, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:44.912103+00:00", "epoch": 0, "step": 7038, "train_loss": 3.575995445251465, "perplexity": 35.73017054741994, "lr": 0.0026291804804649314, "grad_norm": 0.141288, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:45.218831+00:00", "epoch": 0, "step": 7039, "train_loss": 3.6259422302246094, "perplexity": 37.5600967503482, "lr": 0.0026291804804649314, "grad_norm": 0.15532, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:45.524231+00:00", "epoch": 0, "step": 7040, "train_loss": 3.5881261825561523, "perplexity": 36.166243455625036, "lr": 0.0026291804804649314, "grad_norm": 0.158815, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:45.829406+00:00", "epoch": 0, "step": 7041, "train_loss": 3.619762897491455, "perplexity": 37.32871604049135, "lr": 0.0026291804804649314, "grad_norm": 0.14001, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:46.134824+00:00", "epoch": 0, "step": 7042, "train_loss": 3.5569581985473633, "perplexity": 35.05640019086194, "lr": 0.0026291804804649314, "grad_norm": 0.152302, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:46.440130+00:00", "epoch": 0, "step": 7043, "train_loss": 3.586078643798828, "perplexity": 36.09226743067677, "lr": 0.0026291804804649314, "grad_norm": 0.170254, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:46.746220+00:00", "epoch": 0, "step": 7044, "train_loss": 3.5354526042938232, "perplexity": 34.310540298353224, "lr": 0.0026291804804649314, "grad_norm": 0.152484, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:47.050233+00:00", "epoch": 0, "step": 7045, "train_loss": 3.4823625087738037, "perplexity": 32.53649911156291, "lr": 0.0026291804804649314, "grad_norm": 0.133533, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:47.354412+00:00", "epoch": 0, "step": 7046, "train_loss": 3.6817502975463867, "perplexity": 39.71584782347459, "lr": 0.0026291804804649314, "grad_norm": 0.140665, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:47.658806+00:00", "epoch": 0, "step": 7047, "train_loss": 3.5405375957489014, "perplexity": 34.48545344161295, "lr": 0.0026291804804649314, "grad_norm": 0.166466, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:47.964229+00:00", "epoch": 0, "step": 7048, "train_loss": 3.4429879188537598, "perplexity": 31.280281620633332, "lr": 0.0026291804804649314, "grad_norm": 0.172872, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:48.270099+00:00", "epoch": 0, "step": 7049, "train_loss": 3.5528616905212402, "perplexity": 34.91308511221671, "lr": 0.0026291804804649314, "grad_norm": 0.185447, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:48.574691+00:00", "epoch": 0, "step": 7050, "train_loss": 3.6484436988830566, "perplexity": 38.41483445335196, "lr": 0.0026291804804649314, "grad_norm": 0.167202, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:48.880531+00:00", "epoch": 0, "step": 7051, "train_loss": 3.417909860610962, "perplexity": 30.505587410153563, "lr": 0.0026291804804649314, "grad_norm": 0.154412, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:49.186236+00:00", "epoch": 0, "step": 7052, "train_loss": 3.5841612815856934, "perplexity": 36.02313178113999, "lr": 0.0026291804804649314, "grad_norm": 0.163723, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:49.491435+00:00", "epoch": 0, "step": 7053, "train_loss": 3.4362292289733887, "perplexity": 31.06958073084011, "lr": 0.0026291804804649314, "grad_norm": 0.151618, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:49.797342+00:00", "epoch": 0, "step": 7054, "train_loss": 3.6557044982910156, "perplexity": 38.69477191564893, "lr": 0.0026291804804649314, "grad_norm": 0.150987, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:50.102658+00:00", "epoch": 0, "step": 7055, "train_loss": 3.6389713287353516, "perplexity": 38.05267289434921, "lr": 0.0026291804804649314, "grad_norm": 0.157034, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:50.406678+00:00", "epoch": 0, "step": 7056, "train_loss": 3.552079916000366, "perplexity": 34.885801617990765, "lr": 0.0026291804804649314, "grad_norm": 0.146897, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:50.710609+00:00", "epoch": 0, "step": 7057, "train_loss": 3.64530348777771, "perplexity": 38.294392968407806, "lr": 0.0026291804804649314, "grad_norm": 0.158632, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:51.015219+00:00", "epoch": 0, "step": 7058, "train_loss": 3.6442651748657227, "perplexity": 38.254652041061355, "lr": 0.0026291804804649314, "grad_norm": 0.159993, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:51.321332+00:00", "epoch": 0, "step": 7059, "train_loss": 3.5718963146209717, "perplexity": 35.58400768589056, "lr": 0.0026291804804649314, "grad_norm": 0.168509, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:51.626421+00:00", "epoch": 0, "step": 7060, "train_loss": 3.6467714309692383, "perplexity": 38.350648241510214, "lr": 0.0026291804804649314, "grad_norm": 0.153489, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:51.931004+00:00", "epoch": 0, "step": 7061, "train_loss": 3.5953547954559326, "perplexity": 36.42862240503124, "lr": 0.0026291804804649314, "grad_norm": 0.171657, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:52.236715+00:00", "epoch": 0, "step": 7062, "train_loss": 3.5611636638641357, "perplexity": 35.204139103690714, "lr": 0.0026291804804649314, "grad_norm": 0.168786, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:52.541374+00:00", "epoch": 0, "step": 7063, "train_loss": 3.6716928482055664, "perplexity": 39.318409653236586, "lr": 0.0026291804804649314, "grad_norm": 0.159605, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:52.846668+00:00", "epoch": 0, "step": 7064, "train_loss": 3.6187591552734375, "perplexity": 37.29126643027873, "lr": 0.0026291804804649314, "grad_norm": 0.170713, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:53.151968+00:00", "epoch": 0, "step": 7065, "train_loss": 3.483713388442993, "perplexity": 32.58048170762817, "lr": 0.0026291804804649314, "grad_norm": 0.170602, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:53.457016+00:00", "epoch": 0, "step": 7066, "train_loss": 3.611578941345215, "perplexity": 37.02446614745859, "lr": 0.0026291804804649314, "grad_norm": 0.1467, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:53.762529+00:00", "epoch": 0, "step": 7067, "train_loss": 3.645329475402832, "perplexity": 38.295388161667844, "lr": 0.0026291804804649314, "grad_norm": 0.156355, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:54.067290+00:00", "epoch": 0, "step": 7068, "train_loss": 3.4792566299438477, "perplexity": 32.43560145682529, "lr": 0.0026291804804649314, "grad_norm": 0.156409, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:54.372899+00:00", "epoch": 0, "step": 7069, "train_loss": 3.4934775829315186, "perplexity": 32.90016203816936, "lr": 0.0026291804804649314, "grad_norm": 0.156229, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:54.679016+00:00", "epoch": 0, "step": 7070, "train_loss": 3.5300769805908203, "perplexity": 34.126594599054584, "lr": 0.0026291804804649314, "grad_norm": 0.172503, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:54.984444+00:00", "epoch": 0, "step": 7071, "train_loss": 3.446542501449585, "perplexity": 31.391667813705315, "lr": 0.0026291804804649314, "grad_norm": 0.164793, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:55.289290+00:00", "epoch": 0, "step": 7072, "train_loss": 3.5142064094543457, "perplexity": 33.589261223958964, "lr": 0.0026291804804649314, "grad_norm": 0.148996, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:55.593847+00:00", "epoch": 0, "step": 7073, "train_loss": 3.631425619125366, "perplexity": 37.76661907161435, "lr": 0.0026291804804649314, "grad_norm": 0.16586, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:55.898169+00:00", "epoch": 0, "step": 7074, "train_loss": 3.6139652729034424, "perplexity": 37.11292430271888, "lr": 0.0026291804804649314, "grad_norm": 0.13896, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:56.203591+00:00", "epoch": 0, "step": 7075, "train_loss": 3.5704872608184814, "perplexity": 35.53390321279201, "lr": 0.0026291804804649314, "grad_norm": 0.156463, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:56.508014+00:00", "epoch": 0, "step": 7076, "train_loss": 3.495797872543335, "perplexity": 32.9765885739337, "lr": 0.0026291804804649314, "grad_norm": 0.156471, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:56.812105+00:00", "epoch": 0, "step": 7077, "train_loss": 3.585177421569824, "perplexity": 36.05975492966796, "lr": 0.0026291804804649314, "grad_norm": 0.147096, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:57.116794+00:00", "epoch": 0, "step": 7078, "train_loss": 3.5427262783050537, "perplexity": 34.56101381066919, "lr": 0.0026291804804649314, "grad_norm": 0.151102, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:57.421566+00:00", "epoch": 0, "step": 7079, "train_loss": 3.531010627746582, "perplexity": 34.1584716756932, "lr": 0.0026291804804649314, "grad_norm": 0.144806, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:57.726744+00:00", "epoch": 0, "step": 7080, "train_loss": 3.590261936187744, "perplexity": 36.24356818531986, "lr": 0.0026291804804649314, "grad_norm": 0.147555, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:58.032378+00:00", "epoch": 0, "step": 7081, "train_loss": 3.5938382148742676, "perplexity": 36.37341733572408, "lr": 0.0026291804804649314, "grad_norm": 0.16501, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:58.338535+00:00", "epoch": 0, "step": 7082, "train_loss": 3.5120813846588135, "perplexity": 33.517958997326026, "lr": 0.0026291804804649314, "grad_norm": 0.185129, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:58.643682+00:00", "epoch": 0, "step": 7083, "train_loss": 3.536957263946533, "perplexity": 34.36220484303338, "lr": 0.0026291804804649314, "grad_norm": 0.190482, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:58.948795+00:00", "epoch": 0, "step": 7084, "train_loss": 3.6183488368988037, "perplexity": 37.27596827722031, "lr": 0.0026291804804649314, "grad_norm": 0.185967, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:59.252142+00:00", "epoch": 0, "step": 7085, "train_loss": 3.560847759246826, "perplexity": 35.19301971002582, "lr": 0.0026291804804649314, "grad_norm": 0.182074, "tokens_per_sec": 108022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:59.556732+00:00", "epoch": 0, "step": 7086, "train_loss": 3.5226891040802, "perplexity": 33.8754005701897, "lr": 0.0026291804804649314, "grad_norm": 0.204226, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:34:59.862048+00:00", "epoch": 0, "step": 7087, "train_loss": 3.5383424758911133, "perplexity": 34.40983676215932, "lr": 0.0026291804804649314, "grad_norm": 0.179205, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:00.165996+00:00", "epoch": 0, "step": 7088, "train_loss": 3.5515754222869873, "perplexity": 34.868206389102674, "lr": 0.0026291804804649314, "grad_norm": 0.184078, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:00.470635+00:00", "epoch": 0, "step": 7089, "train_loss": 3.6470842361450195, "perplexity": 38.36264639921981, "lr": 0.0026291804804649314, "grad_norm": 0.167061, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:00.774905+00:00", "epoch": 0, "step": 7090, "train_loss": 3.7004499435424805, "perplexity": 40.465507458351105, "lr": 0.0026291804804649314, "grad_norm": 0.167277, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:01.079515+00:00", "epoch": 0, "step": 7091, "train_loss": 3.511957883834839, "perplexity": 33.51381975737694, "lr": 0.0026291804804649314, "grad_norm": 0.179245, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:01.384900+00:00", "epoch": 0, "step": 7092, "train_loss": 3.552988290786743, "perplexity": 34.91750539786021, "lr": 0.0026291804804649314, "grad_norm": 0.165594, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:01.689418+00:00", "epoch": 0, "step": 7093, "train_loss": 3.6671066284179688, "perplexity": 39.1384996536115, "lr": 0.0026291804804649314, "grad_norm": 0.15039, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:01.994422+00:00", "epoch": 0, "step": 7094, "train_loss": 3.5841140747070312, "perplexity": 36.02143128166692, "lr": 0.0026291804804649314, "grad_norm": 0.148538, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:02.298420+00:00", "epoch": 0, "step": 7095, "train_loss": 3.504328727722168, "perplexity": 33.259110438518846, "lr": 0.0026291804804649314, "grad_norm": 0.153469, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:02.602727+00:00", "epoch": 0, "step": 7096, "train_loss": 3.519425630569458, "perplexity": 33.76502929259038, "lr": 0.0026291804804649314, "grad_norm": 0.154522, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:02.907863+00:00", "epoch": 0, "step": 7097, "train_loss": 3.513127565383911, "perplexity": 33.55304318897622, "lr": 0.0026291804804649314, "grad_norm": 0.183177, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:03.214812+00:00", "epoch": 0, "step": 7098, "train_loss": 3.5706491470336914, "perplexity": 35.539656127541, "lr": 0.0026291804804649314, "grad_norm": 0.174916, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:03.520289+00:00", "epoch": 0, "step": 7099, "train_loss": 3.6517441272735596, "perplexity": 38.54182931704732, "lr": 0.0026291804804649314, "grad_norm": 0.171149, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:03.824535+00:00", "epoch": 0, "step": 7100, "train_loss": 3.5810444355010986, "perplexity": 35.911028019732804, "lr": 0.0026291804804649314, "grad_norm": 0.138053, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:04.128477+00:00", "epoch": 0, "step": 7101, "train_loss": 3.5413382053375244, "perplexity": 34.513073881425456, "lr": 0.0026291804804649314, "grad_norm": 0.175345, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:04.433736+00:00", "epoch": 0, "step": 7102, "train_loss": 3.6522881984710693, "perplexity": 38.56280452176303, "lr": 0.0026291804804649314, "grad_norm": 0.187097, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:04.738906+00:00", "epoch": 0, "step": 7103, "train_loss": 3.5226287841796875, "perplexity": 33.8733572710239, "lr": 0.0026291804804649314, "grad_norm": 0.151686, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:05.044161+00:00", "epoch": 0, "step": 7104, "train_loss": 3.7210617065429688, "perplexity": 41.30822805110178, "lr": 0.0026291804804649314, "grad_norm": 0.147462, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:05.348212+00:00", "epoch": 0, "step": 7105, "train_loss": 3.604863405227661, "perplexity": 36.77666001463459, "lr": 0.0026291804804649314, "grad_norm": 0.168464, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:05.652125+00:00", "epoch": 0, "step": 7106, "train_loss": 3.5795814990997314, "perplexity": 35.85853087898409, "lr": 0.0026291804804649314, "grad_norm": 0.188517, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:05.957123+00:00", "epoch": 0, "step": 7107, "train_loss": 3.480184555053711, "perplexity": 32.46571323444724, "lr": 0.0026291804804649314, "grad_norm": 0.176594, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:06.262855+00:00", "epoch": 0, "step": 7108, "train_loss": 3.569558620452881, "perplexity": 35.500920312917664, "lr": 0.0026291804804649314, "grad_norm": 0.155278, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:06.567730+00:00", "epoch": 0, "step": 7109, "train_loss": 3.6731038093566895, "perplexity": 39.373925557963524, "lr": 0.0026291804804649314, "grad_norm": 0.155926, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:06.871763+00:00", "epoch": 0, "step": 7110, "train_loss": 3.595095634460449, "perplexity": 36.41918275023267, "lr": 0.0026291804804649314, "grad_norm": 0.149277, "tokens_per_sec": 107862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:07.176836+00:00", "epoch": 0, "step": 7111, "train_loss": 3.6822524070739746, "perplexity": 39.73579453636121, "lr": 0.0026291804804649314, "grad_norm": 0.134217, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:07.481776+00:00", "epoch": 0, "step": 7112, "train_loss": 3.5611307621002197, "perplexity": 35.20298084447155, "lr": 0.0026291804804649314, "grad_norm": 0.1535, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:07.787533+00:00", "epoch": 0, "step": 7113, "train_loss": 3.5498294830322266, "perplexity": 34.807381732353775, "lr": 0.0026291804804649314, "grad_norm": 0.145361, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:08.094207+00:00", "epoch": 0, "step": 7114, "train_loss": 3.6802053451538086, "perplexity": 39.65453610339606, "lr": 0.0026291804804649314, "grad_norm": 0.154992, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:08.398399+00:00", "epoch": 0, "step": 7115, "train_loss": 3.588027238845825, "perplexity": 36.162665210334296, "lr": 0.0026291804804649314, "grad_norm": 0.15578, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:08.703404+00:00", "epoch": 0, "step": 7116, "train_loss": 3.5191128253936768, "perplexity": 33.75446906839975, "lr": 0.0026291804804649314, "grad_norm": 0.177441, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:09.007896+00:00", "epoch": 0, "step": 7117, "train_loss": 3.632193088531494, "perplexity": 37.79561492161536, "lr": 0.0026291804804649314, "grad_norm": 0.175778, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:09.313037+00:00", "epoch": 0, "step": 7118, "train_loss": 3.5838699340820312, "perplexity": 36.01263806035537, "lr": 0.0026291804804649314, "grad_norm": 0.150219, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:09.618734+00:00", "epoch": 0, "step": 7119, "train_loss": 3.7714195251464844, "perplexity": 43.44168765596584, "lr": 0.0026291804804649314, "grad_norm": 0.17062, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:09.923574+00:00", "epoch": 0, "step": 7120, "train_loss": 3.6078336238861084, "perplexity": 36.88605712264421, "lr": 0.0026291804804649314, "grad_norm": 0.186494, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:10.228996+00:00", "epoch": 0, "step": 7121, "train_loss": 3.621026039123535, "perplexity": 37.375897287819825, "lr": 0.0026291804804649314, "grad_norm": 0.13944, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:10.534160+00:00", "epoch": 0, "step": 7122, "train_loss": 3.618056297302246, "perplexity": 37.26506517537146, "lr": 0.0026291804804649314, "grad_norm": 0.150686, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:10.839773+00:00", "epoch": 0, "step": 7123, "train_loss": 3.6392838954925537, "perplexity": 38.06456875394694, "lr": 0.0026291804804649314, "grad_norm": 0.15106, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:11.146009+00:00", "epoch": 0, "step": 7124, "train_loss": 3.507704019546509, "perplexity": 33.37155930918672, "lr": 0.0026291804804649314, "grad_norm": 0.150261, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:11.450548+00:00", "epoch": 0, "step": 7125, "train_loss": 3.554365396499634, "perplexity": 34.9656236183573, "lr": 0.0026291804804649314, "grad_norm": 0.150451, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:11.755227+00:00", "epoch": 0, "step": 7126, "train_loss": 3.5219929218292236, "perplexity": 33.85182532485388, "lr": 0.0026291804804649314, "grad_norm": 0.162338, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:12.060416+00:00", "epoch": 0, "step": 7127, "train_loss": 3.6007933616638184, "perplexity": 36.6272816007759, "lr": 0.0026291804804649314, "grad_norm": 0.179787, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:12.365398+00:00", "epoch": 0, "step": 7128, "train_loss": 3.489797592163086, "perplexity": 32.77931224503282, "lr": 0.0026291804804649314, "grad_norm": 0.141676, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:12.670403+00:00", "epoch": 0, "step": 7129, "train_loss": 3.5426278114318848, "perplexity": 34.55761086324725, "lr": 0.0026291804804649314, "grad_norm": 0.154066, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:12.976475+00:00", "epoch": 0, "step": 7130, "train_loss": 3.6403441429138184, "perplexity": 38.10494801703204, "lr": 0.0026291804804649314, "grad_norm": 0.150731, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:13.282083+00:00", "epoch": 0, "step": 7131, "train_loss": 3.636025905609131, "perplexity": 37.940756572963274, "lr": 0.0026291804804649314, "grad_norm": 0.147916, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:13.586703+00:00", "epoch": 0, "step": 7132, "train_loss": 3.6347973346710205, "perplexity": 37.894172283979096, "lr": 0.0026291804804649314, "grad_norm": 0.1535, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:13.892469+00:00", "epoch": 0, "step": 7133, "train_loss": 3.4600164890289307, "perplexity": 31.81750115003928, "lr": 0.0026291804804649314, "grad_norm": 0.145778, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:14.198559+00:00", "epoch": 0, "step": 7134, "train_loss": 3.6153860092163086, "perplexity": 37.1656894557614, "lr": 0.0026291804804649314, "grad_norm": 0.165385, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:14.504285+00:00", "epoch": 0, "step": 7135, "train_loss": 3.5909361839294434, "perplexity": 36.268013569516384, "lr": 0.0026291804804649314, "grad_norm": 0.159665, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:14.808359+00:00", "epoch": 0, "step": 7136, "train_loss": 3.6051831245422363, "perplexity": 36.788420103030354, "lr": 0.0026291804804649314, "grad_norm": 0.14709, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:15.113579+00:00", "epoch": 0, "step": 7137, "train_loss": 3.4861998558044434, "perplexity": 32.661592810279956, "lr": 0.0026291804804649314, "grad_norm": 0.151161, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:15.419471+00:00", "epoch": 0, "step": 7138, "train_loss": 3.499063491821289, "perplexity": 33.08445358447237, "lr": 0.0026291804804649314, "grad_norm": 0.166656, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:15.725540+00:00", "epoch": 0, "step": 7139, "train_loss": 3.6986806392669678, "perplexity": 40.39397496302983, "lr": 0.0026291804804649314, "grad_norm": 0.154225, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:16.030722+00:00", "epoch": 0, "step": 7140, "train_loss": 3.6100587844848633, "perplexity": 36.96822590906614, "lr": 0.0026291804804649314, "grad_norm": 0.1624, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:16.336029+00:00", "epoch": 0, "step": 7141, "train_loss": 3.618798017501831, "perplexity": 37.29271568015219, "lr": 0.0026291804804649314, "grad_norm": 0.194797, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:16.640819+00:00", "epoch": 0, "step": 7142, "train_loss": 3.579713821411133, "perplexity": 35.86327607661429, "lr": 0.0026291804804649314, "grad_norm": 0.198601, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:16.947040+00:00", "epoch": 0, "step": 7143, "train_loss": 3.623631238937378, "perplexity": 37.473395915034864, "lr": 0.0026291804804649314, "grad_norm": 0.157404, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:17.252078+00:00", "epoch": 0, "step": 7144, "train_loss": 3.4867398738861084, "perplexity": 32.67923542421001, "lr": 0.0026291804804649314, "grad_norm": 0.153753, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:17.558143+00:00", "epoch": 0, "step": 7145, "train_loss": 3.5193138122558594, "perplexity": 33.761253955036146, "lr": 0.0026291804804649314, "grad_norm": 0.179387, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:17.863149+00:00", "epoch": 0, "step": 7146, "train_loss": 3.6296305656433105, "perplexity": 37.698886780274435, "lr": 0.0026291804804649314, "grad_norm": 0.191787, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:18.167486+00:00", "epoch": 0, "step": 7147, "train_loss": 3.5414788722991943, "perplexity": 34.5179290721407, "lr": 0.0026291804804649314, "grad_norm": 0.167249, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:18.472607+00:00", "epoch": 0, "step": 7148, "train_loss": 3.5904595851898193, "perplexity": 36.25073239837979, "lr": 0.0026291804804649314, "grad_norm": 0.167609, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:18.778625+00:00", "epoch": 0, "step": 7149, "train_loss": 3.619243860244751, "perplexity": 37.30934607379665, "lr": 0.0026291804804649314, "grad_norm": 0.155779, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:19.084725+00:00", "epoch": 0, "step": 7150, "train_loss": 3.516124725341797, "perplexity": 33.65375788011409, "lr": 0.0026291804804649314, "grad_norm": 0.1688, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:19.388692+00:00", "epoch": 0, "step": 7151, "train_loss": 3.5359833240509033, "perplexity": 34.328754412833334, "lr": 0.0026291804804649314, "grad_norm": 0.17472, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:19.693387+00:00", "epoch": 0, "step": 7152, "train_loss": 3.578852891921997, "perplexity": 35.83241361177224, "lr": 0.0026291804804649314, "grad_norm": 0.178384, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:20.000871+00:00", "epoch": 0, "step": 7153, "train_loss": 3.5454225540161133, "perplexity": 34.65432557334551, "lr": 0.0026291804804649314, "grad_norm": 0.14136, "tokens_per_sec": 106570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:20.308588+00:00", "epoch": 0, "step": 7154, "train_loss": 3.5750133991241455, "perplexity": 35.695099095515765, "lr": 0.0026291804804649314, "grad_norm": 0.167496, "tokens_per_sec": 106497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:20.614737+00:00", "epoch": 0, "step": 7155, "train_loss": 3.46571683883667, "perplexity": 31.999389958997106, "lr": 0.0026291804804649314, "grad_norm": 0.151296, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:20.920175+00:00", "epoch": 0, "step": 7156, "train_loss": 3.5844504833221436, "perplexity": 36.03355123999364, "lr": 0.0026291804804649314, "grad_norm": 0.159029, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:21.226983+00:00", "epoch": 0, "step": 7157, "train_loss": 3.580714702606201, "perplexity": 35.899188924481955, "lr": 0.0026291804804649314, "grad_norm": 0.144997, "tokens_per_sec": 106803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:21.532655+00:00", "epoch": 0, "step": 7158, "train_loss": 3.4543824195861816, "perplexity": 31.638743179330795, "lr": 0.0026291804804649314, "grad_norm": 0.159971, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:21.838908+00:00", "epoch": 0, "step": 7159, "train_loss": 3.6915299892425537, "perplexity": 40.106162036095704, "lr": 0.0026291804804649314, "grad_norm": 0.164144, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:22.145409+00:00", "epoch": 0, "step": 7160, "train_loss": 3.5745387077331543, "perplexity": 35.67815896026083, "lr": 0.0026291804804649314, "grad_norm": 0.153288, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:22.450398+00:00", "epoch": 0, "step": 7161, "train_loss": 3.538031816482544, "perplexity": 34.399148682883975, "lr": 0.0026291804804649314, "grad_norm": 0.151087, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:22.755399+00:00", "epoch": 0, "step": 7162, "train_loss": 3.5923571586608887, "perplexity": 36.319586133335584, "lr": 0.0026291804804649314, "grad_norm": 0.197539, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:23.061466+00:00", "epoch": 0, "step": 7163, "train_loss": 3.6004676818847656, "perplexity": 36.61535477806477, "lr": 0.0026291804804649314, "grad_norm": 0.196703, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:23.368612+00:00", "epoch": 0, "step": 7164, "train_loss": 3.51875901222229, "perplexity": 33.74252840515671, "lr": 0.0026291804804649314, "grad_norm": 0.158071, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:23.672990+00:00", "epoch": 0, "step": 7165, "train_loss": 3.5177929401397705, "perplexity": 33.7099464312726, "lr": 0.0026291804804649314, "grad_norm": 0.165395, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:23.977173+00:00", "epoch": 0, "step": 7166, "train_loss": 3.5103039741516113, "perplexity": 33.45843673823873, "lr": 0.0026291804804649314, "grad_norm": 0.191673, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:24.282091+00:00", "epoch": 0, "step": 7167, "train_loss": 3.5497093200683594, "perplexity": 34.803199425484586, "lr": 0.0026291804804649314, "grad_norm": 0.152957, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:24.587923+00:00", "epoch": 0, "step": 7168, "train_loss": 3.569317102432251, "perplexity": 35.492347236231076, "lr": 0.0026291804804649314, "grad_norm": 0.175923, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:24.892741+00:00", "epoch": 0, "step": 7169, "train_loss": 3.5881881713867188, "perplexity": 36.16848542825075, "lr": 0.0026291804804649314, "grad_norm": 0.178588, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:25.199371+00:00", "epoch": 0, "step": 7170, "train_loss": 3.5132265090942383, "perplexity": 33.55636321580728, "lr": 0.0026291804804649314, "grad_norm": 0.158664, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:25.504168+00:00", "epoch": 0, "step": 7171, "train_loss": 3.535006046295166, "perplexity": 34.29522207263741, "lr": 0.0026291804804649314, "grad_norm": 0.159373, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:25.808874+00:00", "epoch": 0, "step": 7172, "train_loss": 3.6490092277526855, "perplexity": 38.436565295387446, "lr": 0.0026291804804649314, "grad_norm": 0.156459, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:26.115836+00:00", "epoch": 0, "step": 7173, "train_loss": 3.539590835571289, "perplexity": 34.45281943831254, "lr": 0.0026291804804649314, "grad_norm": 0.16637, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:26.422068+00:00", "epoch": 0, "step": 7174, "train_loss": 3.603757381439209, "perplexity": 36.73600663974267, "lr": 0.0026291804804649314, "grad_norm": 0.152379, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:26.727757+00:00", "epoch": 0, "step": 7175, "train_loss": 3.4129269123077393, "perplexity": 30.353957740276513, "lr": 0.0026291804804649314, "grad_norm": 0.177251, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:27.032830+00:00", "epoch": 0, "step": 7176, "train_loss": 3.588747501373291, "perplexity": 36.188721205430596, "lr": 0.0026291804804649314, "grad_norm": 0.161556, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:27.339065+00:00", "epoch": 0, "step": 7177, "train_loss": 3.6262576580047607, "perplexity": 37.5719461169999, "lr": 0.0026291804804649314, "grad_norm": 0.160709, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:27.644405+00:00", "epoch": 0, "step": 7178, "train_loss": 3.4627175331115723, "perplexity": 31.90355779231739, "lr": 0.0026291804804649314, "grad_norm": 0.146674, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:27.950318+00:00", "epoch": 0, "step": 7179, "train_loss": 3.6199796199798584, "perplexity": 37.336806889423336, "lr": 0.0026291804804649314, "grad_norm": 0.149771, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:28.257512+00:00", "epoch": 0, "step": 7180, "train_loss": 3.561082601547241, "perplexity": 35.20128549027251, "lr": 0.0026291804804649314, "grad_norm": 0.140077, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:28.562475+00:00", "epoch": 0, "step": 7181, "train_loss": 3.580083131790161, "perplexity": 35.87652320269536, "lr": 0.0026291804804649314, "grad_norm": 0.158209, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:28.867864+00:00", "epoch": 0, "step": 7182, "train_loss": 3.5611398220062256, "perplexity": 35.20329978161389, "lr": 0.0026291804804649314, "grad_norm": 0.163466, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:29.174431+00:00", "epoch": 0, "step": 7183, "train_loss": 3.582695722579956, "perplexity": 35.97037642343068, "lr": 0.0026291804804649314, "grad_norm": 0.151896, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:29.480668+00:00", "epoch": 0, "step": 7184, "train_loss": 3.607625961303711, "perplexity": 36.87839806404513, "lr": 0.0026291804804649314, "grad_norm": 0.147301, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:29.785599+00:00", "epoch": 0, "step": 7185, "train_loss": 3.4637014865875244, "perplexity": 31.934964857914128, "lr": 0.0026291804804649314, "grad_norm": 0.146529, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:30.090004+00:00", "epoch": 0, "step": 7186, "train_loss": 3.5830020904541016, "perplexity": 35.981398279472835, "lr": 0.0026291804804649314, "grad_norm": 0.143546, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:30.395016+00:00", "epoch": 0, "step": 7187, "train_loss": 3.5774495601654053, "perplexity": 35.78216411443971, "lr": 0.0026291804804649314, "grad_norm": 0.15802, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:30.699226+00:00", "epoch": 0, "step": 7188, "train_loss": 3.5386922359466553, "perplexity": 34.42187405353559, "lr": 0.0026291804804649314, "grad_norm": 0.160191, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:31.004624+00:00", "epoch": 0, "step": 7189, "train_loss": 3.6528000831604004, "perplexity": 38.58254928406501, "lr": 0.0026291804804649314, "grad_norm": 0.13002, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:31.311591+00:00", "epoch": 0, "step": 7190, "train_loss": 3.5728209018707275, "perplexity": 35.61692342008136, "lr": 0.0026291804804649314, "grad_norm": 0.161123, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:31.617718+00:00", "epoch": 0, "step": 7191, "train_loss": 3.6220335960388184, "perplexity": 37.413574609432644, "lr": 0.0026291804804649314, "grad_norm": 0.154837, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:31.923685+00:00", "epoch": 0, "step": 7192, "train_loss": 3.5551626682281494, "perplexity": 34.99351183730516, "lr": 0.0026291804804649314, "grad_norm": 0.162234, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:32.228791+00:00", "epoch": 0, "step": 7193, "train_loss": 3.5098142623901367, "perplexity": 33.442055759553796, "lr": 0.0026291804804649314, "grad_norm": 0.144118, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:32.532966+00:00", "epoch": 0, "step": 7194, "train_loss": 3.550661563873291, "perplexity": 34.83635634075742, "lr": 0.0026291804804649314, "grad_norm": 0.149085, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:32.838174+00:00", "epoch": 0, "step": 7195, "train_loss": 3.7138986587524414, "perplexity": 41.01339246079004, "lr": 0.0026291804804649314, "grad_norm": 0.165552, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:33.143644+00:00", "epoch": 0, "step": 7196, "train_loss": 3.7201435565948486, "perplexity": 41.27031830973642, "lr": 0.0026291804804649314, "grad_norm": 0.156337, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:33.449282+00:00", "epoch": 0, "step": 7197, "train_loss": 3.5740296840667725, "perplexity": 35.6600025543893, "lr": 0.0026291804804649314, "grad_norm": 0.142211, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:33.753792+00:00", "epoch": 0, "step": 7198, "train_loss": 3.588881254196167, "perplexity": 36.19356187276734, "lr": 0.0026291804804649314, "grad_norm": 0.148171, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:34.059513+00:00", "epoch": 0, "step": 7199, "train_loss": 3.5514309406280518, "perplexity": 34.86316893671797, "lr": 0.0026291804804649314, "grad_norm": 0.155692, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:34.364631+00:00", "epoch": 0, "step": 7200, "train_loss": 3.6038897037506104, "perplexity": 36.74086795467602, "lr": 0.0026291804804649314, "grad_norm": 0.174342, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:34.669816+00:00", "epoch": 0, "step": 7201, "train_loss": 3.523684024810791, "perplexity": 33.90912068011071, "lr": 0.0026291804804649314, "grad_norm": 0.174095, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:34.976498+00:00", "epoch": 0, "step": 7202, "train_loss": 3.4859585762023926, "perplexity": 32.653713184799024, "lr": 0.0026291804804649314, "grad_norm": 0.168051, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:35.281661+00:00", "epoch": 0, "step": 7203, "train_loss": 3.569714069366455, "perplexity": 35.50643932136201, "lr": 0.0026291804804649314, "grad_norm": 0.165476, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:35.586897+00:00", "epoch": 0, "step": 7204, "train_loss": 3.7090413570404053, "perplexity": 40.81466107924234, "lr": 0.0026291804804649314, "grad_norm": 0.175402, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:35.892488+00:00", "epoch": 0, "step": 7205, "train_loss": 3.604257583618164, "perplexity": 36.75438666679154, "lr": 0.0026291804804649314, "grad_norm": 0.178803, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:36.198562+00:00", "epoch": 0, "step": 7206, "train_loss": 3.498991012573242, "perplexity": 33.082055735052734, "lr": 0.0026291804804649314, "grad_norm": 0.185749, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:36.503435+00:00", "epoch": 0, "step": 7207, "train_loss": 3.620823621749878, "perplexity": 37.368332522498534, "lr": 0.0026291804804649314, "grad_norm": 0.169433, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:36.808144+00:00", "epoch": 0, "step": 7208, "train_loss": 3.6070432662963867, "perplexity": 36.8569154651245, "lr": 0.0026291804804649314, "grad_norm": 0.187112, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:37.115720+00:00", "epoch": 0, "step": 7209, "train_loss": 3.4616341590881348, "perplexity": 31.869013022381917, "lr": 0.0026291804804649314, "grad_norm": 0.166009, "tokens_per_sec": 106536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:37.421360+00:00", "epoch": 0, "step": 7210, "train_loss": 3.578805446624756, "perplexity": 35.83071357258731, "lr": 0.0026291804804649314, "grad_norm": 0.162558, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:37.726065+00:00", "epoch": 0, "step": 7211, "train_loss": 3.5520620346069336, "perplexity": 34.885177816824054, "lr": 0.0026291804804649314, "grad_norm": 0.146811, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:38.032151+00:00", "epoch": 0, "step": 7212, "train_loss": 3.6089937686920166, "perplexity": 36.92887512296329, "lr": 0.0026291804804649314, "grad_norm": 0.157636, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:38.337639+00:00", "epoch": 0, "step": 7213, "train_loss": 3.6299586296081543, "perplexity": 37.71125645545317, "lr": 0.0026291804804649314, "grad_norm": 0.137563, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:38.641465+00:00", "epoch": 0, "step": 7214, "train_loss": 3.6398603916168213, "perplexity": 38.086519156858685, "lr": 0.0026291804804649314, "grad_norm": 0.158557, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:38.946373+00:00", "epoch": 0, "step": 7215, "train_loss": 3.5612645149230957, "perplexity": 35.207689657434635, "lr": 0.0026291804804649314, "grad_norm": 0.149234, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:39.251347+00:00", "epoch": 0, "step": 7216, "train_loss": 3.5097639560699463, "perplexity": 33.440373455104556, "lr": 0.0026291804804649314, "grad_norm": 0.167521, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:39.556254+00:00", "epoch": 0, "step": 7217, "train_loss": 3.5929815769195557, "perplexity": 36.34227182800835, "lr": 0.0026291804804649314, "grad_norm": 0.152928, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:39.862021+00:00", "epoch": 0, "step": 7218, "train_loss": 3.5759599208831787, "perplexity": 35.72890127822762, "lr": 0.0026291804804649314, "grad_norm": 0.139228, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:40.168147+00:00", "epoch": 0, "step": 7219, "train_loss": 3.6194489002227783, "perplexity": 37.316996765617866, "lr": 0.0026291804804649314, "grad_norm": 0.15849, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:40.473841+00:00", "epoch": 0, "step": 7220, "train_loss": 3.5803160667419434, "perplexity": 35.884881072279875, "lr": 0.0026291804804649314, "grad_norm": 0.168248, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:40.778970+00:00", "epoch": 0, "step": 7221, "train_loss": 3.6480472087860107, "perplexity": 38.39960637100297, "lr": 0.0026291804804649314, "grad_norm": 0.177474, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:41.085412+00:00", "epoch": 0, "step": 7222, "train_loss": 3.4562461376190186, "perplexity": 31.697763857397668, "lr": 0.0026291804804649314, "grad_norm": 0.166206, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:41.391759+00:00", "epoch": 0, "step": 7223, "train_loss": 3.5974762439727783, "perplexity": 36.505985884307506, "lr": 0.0026291804804649314, "grad_norm": 0.144617, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:41.696109+00:00", "epoch": 0, "step": 7224, "train_loss": 3.5265886783599854, "perplexity": 34.00775811227754, "lr": 0.0026291804804649314, "grad_norm": 0.184171, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:42.001444+00:00", "epoch": 0, "step": 7225, "train_loss": 3.4929659366607666, "perplexity": 32.883333098554616, "lr": 0.0026291804804649314, "grad_norm": 0.157105, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:42.307876+00:00", "epoch": 0, "step": 7226, "train_loss": 3.4762368202209473, "perplexity": 32.33779985761105, "lr": 0.0026291804804649314, "grad_norm": 0.138208, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:42.612961+00:00", "epoch": 0, "step": 7227, "train_loss": 3.40091872215271, "perplexity": 29.991641379375288, "lr": 0.0026291804804649314, "grad_norm": 0.145737, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:42.918406+00:00", "epoch": 0, "step": 7228, "train_loss": 3.427186965942383, "perplexity": 30.78990775348212, "lr": 0.0026291804804649314, "grad_norm": 0.153977, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:43.226591+00:00", "epoch": 0, "step": 7229, "train_loss": 3.5271317958831787, "perplexity": 34.026233338288186, "lr": 0.0026291804804649314, "grad_norm": 0.143298, "tokens_per_sec": 106327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:43.533209+00:00", "epoch": 0, "step": 7230, "train_loss": 3.61454176902771, "perplexity": 37.13432592812396, "lr": 0.0026291804804649314, "grad_norm": 0.157623, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:43.838202+00:00", "epoch": 0, "step": 7231, "train_loss": 3.633434534072876, "perplexity": 37.842565256355506, "lr": 0.0026291804804649314, "grad_norm": 0.1513, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:44.144349+00:00", "epoch": 0, "step": 7232, "train_loss": 3.595682382583618, "perplexity": 36.440557907662225, "lr": 0.0026291804804649314, "grad_norm": 0.143915, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:44.449848+00:00", "epoch": 0, "step": 7233, "train_loss": 3.584322214126587, "perplexity": 36.02892954178014, "lr": 0.0026291804804649314, "grad_norm": 0.165348, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:44.755422+00:00", "epoch": 0, "step": 7234, "train_loss": 3.5336506366729736, "perplexity": 34.248769486897416, "lr": 0.0026291804804649314, "grad_norm": 0.165389, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:45.061343+00:00", "epoch": 0, "step": 7235, "train_loss": 3.5148544311523438, "perplexity": 33.611034848180324, "lr": 0.0026291804804649314, "grad_norm": 0.189748, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:45.366398+00:00", "epoch": 0, "step": 7236, "train_loss": 3.564638614654541, "perplexity": 35.326684551080035, "lr": 0.0026291804804649314, "grad_norm": 0.195623, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:45.672101+00:00", "epoch": 0, "step": 7237, "train_loss": 3.6345603466033936, "perplexity": 37.885192881362855, "lr": 0.0026291804804649314, "grad_norm": 0.181658, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:45.978542+00:00", "epoch": 0, "step": 7238, "train_loss": 3.629138946533203, "perplexity": 37.68035784206672, "lr": 0.0026291804804649314, "grad_norm": 0.165063, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:46.284311+00:00", "epoch": 0, "step": 7239, "train_loss": 3.5357890129089355, "perplexity": 34.322084601391374, "lr": 0.0026291804804649314, "grad_norm": 0.197798, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:46.591247+00:00", "epoch": 0, "step": 7240, "train_loss": 3.464247465133667, "perplexity": 31.952405424252653, "lr": 0.0026291804804649314, "grad_norm": 0.217798, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:46.897235+00:00", "epoch": 0, "step": 7241, "train_loss": 3.572046995162964, "perplexity": 35.58936990743601, "lr": 0.0026291804804649314, "grad_norm": 0.144523, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:47.203746+00:00", "epoch": 0, "step": 7242, "train_loss": 3.466501235961914, "perplexity": 32.02450003533888, "lr": 0.0026291804804649314, "grad_norm": 0.156901, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:47.507882+00:00", "epoch": 0, "step": 7243, "train_loss": 3.5119683742523193, "perplexity": 33.51417133318164, "lr": 0.0026291804804649314, "grad_norm": 0.169424, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:47.812434+00:00", "epoch": 0, "step": 7244, "train_loss": 3.7992026805877686, "perplexity": 44.665557576058326, "lr": 0.0026291804804649314, "grad_norm": 0.151471, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:48.117559+00:00", "epoch": 0, "step": 7245, "train_loss": 3.5332014560699463, "perplexity": 34.23338905851999, "lr": 0.0026291804804649314, "grad_norm": 0.164781, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:48.424257+00:00", "epoch": 0, "step": 7246, "train_loss": 3.5481789112091064, "perplexity": 34.749977037148696, "lr": 0.0026291804804649314, "grad_norm": 0.15067, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:48.728436+00:00", "epoch": 0, "step": 7247, "train_loss": 3.4308533668518066, "perplexity": 30.903003099019145, "lr": 0.0026291804804649314, "grad_norm": 0.15822, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:49.033504+00:00", "epoch": 0, "step": 7248, "train_loss": 3.529574394226074, "perplexity": 34.10944734728229, "lr": 0.0026291804804649314, "grad_norm": 0.148437, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:49.339476+00:00", "epoch": 0, "step": 7249, "train_loss": 3.5559747219085693, "perplexity": 35.02193998841362, "lr": 0.0026291804804649314, "grad_norm": 0.179022, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:49.644945+00:00", "epoch": 0, "step": 7250, "train_loss": 3.4828076362609863, "perplexity": 32.550985225498714, "lr": 0.0026291804804649314, "grad_norm": 0.176074, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:49.951884+00:00", "epoch": 0, "step": 7251, "train_loss": 3.4917871952056885, "perplexity": 32.84459498624497, "lr": 0.0026291804804649314, "grad_norm": 0.169684, "tokens_per_sec": 106691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:50.257445+00:00", "epoch": 0, "step": 7252, "train_loss": 3.5802085399627686, "perplexity": 35.88102269404031, "lr": 0.0026291804804649314, "grad_norm": 0.153666, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:50.562705+00:00", "epoch": 0, "step": 7253, "train_loss": 3.5503413677215576, "perplexity": 34.82520365913485, "lr": 0.0026291804804649314, "grad_norm": 0.151241, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:50.868204+00:00", "epoch": 0, "step": 7254, "train_loss": 3.61657452583313, "perplexity": 37.20988775525912, "lr": 0.0026291804804649314, "grad_norm": 0.168139, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:51.173073+00:00", "epoch": 0, "step": 7255, "train_loss": 3.579038381576538, "perplexity": 35.839060770264965, "lr": 0.0026291804804649314, "grad_norm": 0.17953, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:51.479490+00:00", "epoch": 0, "step": 7256, "train_loss": 3.5606184005737305, "perplexity": 35.184948811323586, "lr": 0.0026291804804649314, "grad_norm": 0.173972, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:51.784656+00:00", "epoch": 0, "step": 7257, "train_loss": 3.423732042312622, "perplexity": 30.68371452400662, "lr": 0.0026291804804649314, "grad_norm": 0.187385, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:52.089820+00:00", "epoch": 0, "step": 7258, "train_loss": 3.5258240699768066, "perplexity": 33.981765433708084, "lr": 0.0026291804804649314, "grad_norm": 0.151904, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:52.394564+00:00", "epoch": 0, "step": 7259, "train_loss": 3.380044460296631, "perplexity": 29.372076975514705, "lr": 0.0026291804804649314, "grad_norm": 0.158069, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:52.700749+00:00", "epoch": 0, "step": 7260, "train_loss": 3.599726676940918, "perplexity": 36.5882326692029, "lr": 0.0026291804804649314, "grad_norm": 0.153858, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:53.007535+00:00", "epoch": 0, "step": 7261, "train_loss": 3.6127684116363525, "perplexity": 37.068531852214065, "lr": 0.0026291804804649314, "grad_norm": 0.159584, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:53.314765+00:00", "epoch": 0, "step": 7262, "train_loss": 3.5142626762390137, "perplexity": 33.59115123685938, "lr": 0.0026291804804649314, "grad_norm": 0.157424, "tokens_per_sec": 106667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:53.620449+00:00", "epoch": 0, "step": 7263, "train_loss": 3.5536372661590576, "perplexity": 34.940173353603, "lr": 0.0026291804804649314, "grad_norm": 0.142138, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:53.925737+00:00", "epoch": 0, "step": 7264, "train_loss": 3.5414061546325684, "perplexity": 34.51541910014257, "lr": 0.0026291804804649314, "grad_norm": 0.153105, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:54.231004+00:00", "epoch": 0, "step": 7265, "train_loss": 3.5979220867156982, "perplexity": 36.52226544197882, "lr": 0.0026291804804649314, "grad_norm": 0.154149, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:54.536520+00:00", "epoch": 0, "step": 7266, "train_loss": 3.4780468940734863, "perplexity": 32.39638667081531, "lr": 0.0026291804804649314, "grad_norm": 0.143707, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:54.842825+00:00", "epoch": 0, "step": 7267, "train_loss": 3.577186346054077, "perplexity": 35.7727469833262, "lr": 0.0026291804804649314, "grad_norm": 0.158335, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:55.149317+00:00", "epoch": 0, "step": 7268, "train_loss": 3.5907905101776123, "perplexity": 36.262730656708456, "lr": 0.0026291804804649314, "grad_norm": 0.168262, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:55.454312+00:00", "epoch": 0, "step": 7269, "train_loss": 3.6830785274505615, "perplexity": 39.76863464898296, "lr": 0.0026291804804649314, "grad_norm": 0.147312, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:55.760589+00:00", "epoch": 0, "step": 7270, "train_loss": 3.5874714851379395, "perplexity": 36.14257325866208, "lr": 0.0026291804804649314, "grad_norm": 0.15419, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:56.065112+00:00", "epoch": 0, "step": 7271, "train_loss": 3.491734266281128, "perplexity": 32.842856603160506, "lr": 0.0026291804804649314, "grad_norm": 0.157948, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:56.370630+00:00", "epoch": 0, "step": 7272, "train_loss": 3.6195220947265625, "perplexity": 37.31972826464298, "lr": 0.0026291804804649314, "grad_norm": 0.161588, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:56.676488+00:00", "epoch": 0, "step": 7273, "train_loss": 3.492053270339966, "perplexity": 32.853335279001946, "lr": 0.0026291804804649314, "grad_norm": 0.171808, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:56.981778+00:00", "epoch": 0, "step": 7274, "train_loss": 3.6200618743896484, "perplexity": 37.339878132747394, "lr": 0.0026291804804649314, "grad_norm": 0.173352, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:57.286992+00:00", "epoch": 0, "step": 7275, "train_loss": 3.6050193309783936, "perplexity": 36.78239489005256, "lr": 0.0026291804804649314, "grad_norm": 0.177549, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:57.591971+00:00", "epoch": 0, "step": 7276, "train_loss": 3.5232696533203125, "perplexity": 33.89507261799291, "lr": 0.0026291804804649314, "grad_norm": 0.158884, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:57.897608+00:00", "epoch": 0, "step": 7277, "train_loss": 3.572983741760254, "perplexity": 35.622723748205935, "lr": 0.0026291804804649314, "grad_norm": 0.179334, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:58.203907+00:00", "epoch": 0, "step": 7278, "train_loss": 3.647718906402588, "perplexity": 38.38700175788434, "lr": 0.0026291804804649314, "grad_norm": 0.143655, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:58.510682+00:00", "epoch": 0, "step": 7279, "train_loss": 3.639007806777954, "perplexity": 38.05406100668984, "lr": 0.0026291804804649314, "grad_norm": 0.168755, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:58.816563+00:00", "epoch": 0, "step": 7280, "train_loss": 3.474757432937622, "perplexity": 32.28999509732695, "lr": 0.0026291804804649314, "grad_norm": 0.141028, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:59.121773+00:00", "epoch": 0, "step": 7281, "train_loss": 3.656243324279785, "perplexity": 38.71562728258886, "lr": 0.0026291804804649314, "grad_norm": 0.14765, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:59.427110+00:00", "epoch": 0, "step": 7282, "train_loss": 3.588951826095581, "perplexity": 36.19611621130646, "lr": 0.0026291804804649314, "grad_norm": 0.157575, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:35:59.732027+00:00", "epoch": 0, "step": 7283, "train_loss": 3.574737310409546, "perplexity": 35.685245441792866, "lr": 0.0026291804804649314, "grad_norm": 0.151377, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:00.037352+00:00", "epoch": 0, "step": 7284, "train_loss": 3.5465025901794434, "perplexity": 34.69177371721514, "lr": 0.0026291804804649314, "grad_norm": 0.16556, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:00.342336+00:00", "epoch": 0, "step": 7285, "train_loss": 3.601750373840332, "perplexity": 36.66235113356879, "lr": 0.0026291804804649314, "grad_norm": 0.153084, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:00.648158+00:00", "epoch": 0, "step": 7286, "train_loss": 3.546060562133789, "perplexity": 34.67644236897122, "lr": 0.0026291804804649314, "grad_norm": 0.159107, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:00.953578+00:00", "epoch": 0, "step": 7287, "train_loss": 3.584594249725342, "perplexity": 36.03873202645247, "lr": 0.0026291804804649314, "grad_norm": 0.175527, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:01.259371+00:00", "epoch": 0, "step": 7288, "train_loss": 3.554926633834839, "perplexity": 34.985253139675514, "lr": 0.0026291804804649314, "grad_norm": 0.129199, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:01.565837+00:00", "epoch": 0, "step": 7289, "train_loss": 3.6868557929992676, "perplexity": 39.91913540427863, "lr": 0.0026291804804649314, "grad_norm": 0.148805, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:01.873138+00:00", "epoch": 0, "step": 7290, "train_loss": 3.6312553882598877, "perplexity": 37.76019057454343, "lr": 0.0026291804804649314, "grad_norm": 0.145619, "tokens_per_sec": 106692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:02.177572+00:00", "epoch": 0, "step": 7291, "train_loss": 3.6254332065582275, "perplexity": 37.54098263737133, "lr": 0.0026291804804649314, "grad_norm": 0.139992, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:02.481909+00:00", "epoch": 0, "step": 7292, "train_loss": 3.559328556060791, "perplexity": 35.139594954161076, "lr": 0.0026291804804649314, "grad_norm": 0.145094, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:02.787099+00:00", "epoch": 0, "step": 7293, "train_loss": 3.545804500579834, "perplexity": 34.66756420197972, "lr": 0.0026291804804649314, "grad_norm": 0.14431, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:03.092834+00:00", "epoch": 0, "step": 7294, "train_loss": 3.587083339691162, "perplexity": 36.1285474056285, "lr": 0.0026291804804649314, "grad_norm": 0.157962, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:03.399669+00:00", "epoch": 0, "step": 7295, "train_loss": 3.4856865406036377, "perplexity": 32.64483142051404, "lr": 0.0026291804804649314, "grad_norm": 0.17077, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:03.703939+00:00", "epoch": 0, "step": 7296, "train_loss": 3.5846383571624756, "perplexity": 36.04032163761629, "lr": 0.0026291804804649314, "grad_norm": 0.176639, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:04.009593+00:00", "epoch": 0, "step": 7297, "train_loss": 3.5261001586914062, "perplexity": 33.99114871089537, "lr": 0.0026291804804649314, "grad_norm": 0.188518, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:04.317702+00:00", "epoch": 0, "step": 7298, "train_loss": 3.4921181201934814, "perplexity": 32.85546588206619, "lr": 0.0026291804804649314, "grad_norm": 0.193256, "tokens_per_sec": 106352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:04.623674+00:00", "epoch": 0, "step": 7299, "train_loss": 3.569988965988159, "perplexity": 35.51620126328115, "lr": 0.0026291804804649314, "grad_norm": 0.172039, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:04.930827+00:00", "epoch": 0, "step": 7300, "train_loss": 3.5446360111236572, "perplexity": 34.627079176507486, "lr": 0.0026291804804649314, "grad_norm": 0.18545, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:05.236393+00:00", "epoch": 0, "step": 7301, "train_loss": 3.5692942142486572, "perplexity": 35.49153489016796, "lr": 0.0026291804804649314, "grad_norm": 0.186802, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:05.540652+00:00", "epoch": 0, "step": 7302, "train_loss": 3.559717893600464, "perplexity": 35.15327878124656, "lr": 0.0026291804804649314, "grad_norm": 0.173544, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:05.845543+00:00", "epoch": 0, "step": 7303, "train_loss": 3.467271089553833, "perplexity": 32.049163704209406, "lr": 0.0026291804804649314, "grad_norm": 0.176555, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:06.149918+00:00", "epoch": 0, "step": 7304, "train_loss": 3.672712564468384, "perplexity": 39.358523723997315, "lr": 0.0026291804804649314, "grad_norm": 0.156113, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:06.454816+00:00", "epoch": 0, "step": 7305, "train_loss": 3.5905697345733643, "perplexity": 36.25472561412761, "lr": 0.0026291804804649314, "grad_norm": 0.172743, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:06.760296+00:00", "epoch": 0, "step": 7306, "train_loss": 3.6754767894744873, "perplexity": 39.467470046181546, "lr": 0.0026291804804649314, "grad_norm": 0.158554, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:07.066999+00:00", "epoch": 0, "step": 7307, "train_loss": 3.5321977138519287, "perplexity": 34.19904479993076, "lr": 0.0026291804804649314, "grad_norm": 0.151067, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:07.372133+00:00", "epoch": 0, "step": 7308, "train_loss": 3.5937516689300537, "perplexity": 36.37026950019463, "lr": 0.0026291804804649314, "grad_norm": 0.167116, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:07.677942+00:00", "epoch": 0, "step": 7309, "train_loss": 3.579045057296753, "perplexity": 35.83930002260602, "lr": 0.0026291804804649314, "grad_norm": 0.178917, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:07.983966+00:00", "epoch": 0, "step": 7310, "train_loss": 3.6429336071014404, "perplexity": 38.20374727866147, "lr": 0.0026291804804649314, "grad_norm": 0.184043, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:08.288569+00:00", "epoch": 0, "step": 7311, "train_loss": 3.567695140838623, "perplexity": 35.43482667282878, "lr": 0.0026291804804649314, "grad_norm": 0.152372, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:08.593192+00:00", "epoch": 0, "step": 7312, "train_loss": 3.563666343688965, "perplexity": 35.29235413331957, "lr": 0.0026291804804649314, "grad_norm": 0.17593, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:08.898238+00:00", "epoch": 0, "step": 7313, "train_loss": 3.528681516647339, "perplexity": 34.07900537902074, "lr": 0.0026291804804649314, "grad_norm": 0.180573, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:09.203378+00:00", "epoch": 0, "step": 7314, "train_loss": 3.5818941593170166, "perplexity": 35.941555443598006, "lr": 0.0026291804804649314, "grad_norm": 0.142413, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:09.508288+00:00", "epoch": 0, "step": 7315, "train_loss": 3.4792675971984863, "perplexity": 32.43595718827652, "lr": 0.0026291804804649314, "grad_norm": 0.148275, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:09.813463+00:00", "epoch": 0, "step": 7316, "train_loss": 3.415146827697754, "perplexity": 30.421415806108353, "lr": 0.0026291804804649314, "grad_norm": 0.14262, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:10.120123+00:00", "epoch": 0, "step": 7317, "train_loss": 3.520062208175659, "perplexity": 33.78653019688179, "lr": 0.0026291804804649314, "grad_norm": 0.154917, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:10.425385+00:00", "epoch": 0, "step": 7318, "train_loss": 3.516947031021118, "perplexity": 33.68144293757939, "lr": 0.0026291804804649314, "grad_norm": 0.150347, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:10.730625+00:00", "epoch": 0, "step": 7319, "train_loss": 3.6164941787719727, "perplexity": 37.20689817023578, "lr": 0.0026291804804649314, "grad_norm": 0.176925, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:11.036985+00:00", "epoch": 0, "step": 7320, "train_loss": 3.494764804840088, "perplexity": 32.94253911604054, "lr": 0.0026291804804649314, "grad_norm": 0.161794, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:11.341873+00:00", "epoch": 0, "step": 7321, "train_loss": 3.552008628845215, "perplexity": 34.88331479707849, "lr": 0.0026291804804649314, "grad_norm": 0.155563, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:11.646459+00:00", "epoch": 0, "step": 7322, "train_loss": 3.429764747619629, "perplexity": 30.86937980031887, "lr": 0.0026291804804649314, "grad_norm": 0.161092, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:11.950995+00:00", "epoch": 0, "step": 7323, "train_loss": 3.580589771270752, "perplexity": 35.89470427101078, "lr": 0.0026291804804649314, "grad_norm": 0.142224, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:12.256055+00:00", "epoch": 0, "step": 7324, "train_loss": 3.552341938018799, "perplexity": 34.8949436638017, "lr": 0.0026291804804649314, "grad_norm": 0.177603, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:12.561358+00:00", "epoch": 0, "step": 7325, "train_loss": 3.5261518955230713, "perplexity": 33.99290735072716, "lr": 0.0026291804804649314, "grad_norm": 0.15872, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:12.866589+00:00", "epoch": 0, "step": 7326, "train_loss": 3.6159942150115967, "perplexity": 37.1883007189279, "lr": 0.0026291804804649314, "grad_norm": 0.172373, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:13.172813+00:00", "epoch": 0, "step": 7327, "train_loss": 3.620765447616577, "perplexity": 37.366158715371434, "lr": 0.0026291804804649314, "grad_norm": 0.168964, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:13.478120+00:00", "epoch": 0, "step": 7328, "train_loss": 3.5558724403381348, "perplexity": 35.01835807257705, "lr": 0.0026291804804649314, "grad_norm": 0.157224, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:13.784269+00:00", "epoch": 0, "step": 7329, "train_loss": 3.5561366081237793, "perplexity": 35.02761001666498, "lr": 0.0026291804804649314, "grad_norm": 0.166013, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:14.090098+00:00", "epoch": 0, "step": 7330, "train_loss": 3.476571798324585, "perplexity": 32.34863412700337, "lr": 0.0026291804804649314, "grad_norm": 0.166446, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:14.394272+00:00", "epoch": 0, "step": 7331, "train_loss": 3.4791269302368164, "perplexity": 32.43139484162323, "lr": 0.0026291804804649314, "grad_norm": 0.158065, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:14.699042+00:00", "epoch": 0, "step": 7332, "train_loss": 3.673837423324585, "perplexity": 39.4028214176309, "lr": 0.0026291804804649314, "grad_norm": 0.166515, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:15.005277+00:00", "epoch": 0, "step": 7333, "train_loss": 3.4314308166503906, "perplexity": 30.920853185202645, "lr": 0.0026291804804649314, "grad_norm": 0.14477, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:15.309663+00:00", "epoch": 0, "step": 7334, "train_loss": 3.6471424102783203, "perplexity": 38.364878177840474, "lr": 0.0026291804804649314, "grad_norm": 0.161855, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:15.614022+00:00", "epoch": 0, "step": 7335, "train_loss": 3.5220537185668945, "perplexity": 33.853883467961396, "lr": 0.0026291804804649314, "grad_norm": 0.16263, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:15.918971+00:00", "epoch": 0, "step": 7336, "train_loss": 3.645153284072876, "perplexity": 38.28864144067078, "lr": 0.0026291804804649314, "grad_norm": 0.164024, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:16.224874+00:00", "epoch": 0, "step": 7337, "train_loss": 3.6080257892608643, "perplexity": 36.89314602673359, "lr": 0.0026291804804649314, "grad_norm": 0.146102, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:16.530923+00:00", "epoch": 0, "step": 7338, "train_loss": 3.6993777751922607, "perplexity": 40.42214487212888, "lr": 0.0026291804804649314, "grad_norm": 0.150523, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:16.838271+00:00", "epoch": 0, "step": 7339, "train_loss": 3.562439441680908, "perplexity": 35.24908042489252, "lr": 0.0026291804804649314, "grad_norm": 0.16638, "tokens_per_sec": 106616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:17.144441+00:00", "epoch": 0, "step": 7340, "train_loss": 3.5221970081329346, "perplexity": 33.85873472379116, "lr": 0.0026291804804649314, "grad_norm": 0.157017, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:17.449486+00:00", "epoch": 0, "step": 7341, "train_loss": 3.502932071685791, "perplexity": 33.212691324473894, "lr": 0.0026291804804649314, "grad_norm": 0.160584, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:17.755281+00:00", "epoch": 0, "step": 7342, "train_loss": 3.5165109634399414, "perplexity": 33.666758754107924, "lr": 0.0026291804804649314, "grad_norm": 0.152764, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:18.061572+00:00", "epoch": 0, "step": 7343, "train_loss": 3.4647018909454346, "perplexity": 31.966928721656174, "lr": 0.0026291804804649314, "grad_norm": 0.150284, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:18.367400+00:00", "epoch": 0, "step": 7344, "train_loss": 3.7043657302856445, "perplexity": 40.624272397967886, "lr": 0.0026291804804649314, "grad_norm": 0.166875, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:18.674309+00:00", "epoch": 0, "step": 7345, "train_loss": 3.6173925399780273, "perplexity": 37.24033842261327, "lr": 0.0026291804804649314, "grad_norm": 0.144327, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:18.980779+00:00", "epoch": 0, "step": 7346, "train_loss": 3.6246514320373535, "perplexity": 37.51164552265543, "lr": 0.0026291804804649314, "grad_norm": 0.156036, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:19.286866+00:00", "epoch": 0, "step": 7347, "train_loss": 3.518345594406128, "perplexity": 33.72858152589449, "lr": 0.0026291804804649314, "grad_norm": 0.157819, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:19.593423+00:00", "epoch": 0, "step": 7348, "train_loss": 3.5514907836914062, "perplexity": 34.86525531797248, "lr": 0.0026291804804649314, "grad_norm": 0.162042, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:19.900523+00:00", "epoch": 0, "step": 7349, "train_loss": 3.578488826751709, "perplexity": 35.81937065239651, "lr": 0.0026291804804649314, "grad_norm": 0.163176, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:20.207257+00:00", "epoch": 0, "step": 7350, "train_loss": 3.529935598373413, "perplexity": 34.12177004649862, "lr": 0.0026291804804649314, "grad_norm": 0.148618, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:20.512729+00:00", "epoch": 0, "step": 7351, "train_loss": 3.588810443878174, "perplexity": 36.19099908587877, "lr": 0.0026291804804649314, "grad_norm": 0.171284, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:20.819628+00:00", "epoch": 0, "step": 7352, "train_loss": 3.4813783168792725, "perplexity": 32.504492705667055, "lr": 0.0026291804804649314, "grad_norm": 0.165649, "tokens_per_sec": 106772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:21.124755+00:00", "epoch": 0, "step": 7353, "train_loss": 3.6494290828704834, "perplexity": 38.45270647227803, "lr": 0.0026291804804649314, "grad_norm": 0.176957, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:21.431283+00:00", "epoch": 0, "step": 7354, "train_loss": 3.488914966583252, "perplexity": 32.75039314979998, "lr": 0.0026291804804649314, "grad_norm": 0.197142, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:21.737332+00:00", "epoch": 0, "step": 7355, "train_loss": 3.5524063110351562, "perplexity": 34.897190028882825, "lr": 0.0026291804804649314, "grad_norm": 0.148986, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:22.042917+00:00", "epoch": 0, "step": 7356, "train_loss": 3.508021354675293, "perplexity": 33.38215095772003, "lr": 0.0026291804804649314, "grad_norm": 0.151213, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:22.348134+00:00", "epoch": 0, "step": 7357, "train_loss": 3.460073709487915, "perplexity": 31.81932181414794, "lr": 0.0026291804804649314, "grad_norm": 0.162521, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:22.655843+00:00", "epoch": 0, "step": 7358, "train_loss": 3.5710103511810303, "perplexity": 35.55249551741011, "lr": 0.0026291804804649314, "grad_norm": 0.159274, "tokens_per_sec": 106491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:22.962630+00:00", "epoch": 0, "step": 7359, "train_loss": 3.5691990852355957, "perplexity": 35.488158776067564, "lr": 0.0026291804804649314, "grad_norm": 0.150096, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:23.269648+00:00", "epoch": 0, "step": 7360, "train_loss": 3.589939594268799, "perplexity": 36.23188724673079, "lr": 0.0026291804804649314, "grad_norm": 0.156005, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:23.575238+00:00", "epoch": 0, "step": 7361, "train_loss": 3.5872960090637207, "perplexity": 36.136231658210335, "lr": 0.0026291804804649314, "grad_norm": 0.144629, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:23.880446+00:00", "epoch": 0, "step": 7362, "train_loss": 3.5419065952301025, "perplexity": 34.5326963398567, "lr": 0.0026291804804649314, "grad_norm": 0.173415, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:24.184966+00:00", "epoch": 0, "step": 7363, "train_loss": 3.528465747833252, "perplexity": 34.07165298568197, "lr": 0.0026291804804649314, "grad_norm": 0.142128, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:24.491482+00:00", "epoch": 0, "step": 7364, "train_loss": 3.588703155517578, "perplexity": 36.187116421204614, "lr": 0.0026291804804649314, "grad_norm": 0.17426, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:24.798675+00:00", "epoch": 0, "step": 7365, "train_loss": 3.6601450443267822, "perplexity": 38.86697989744654, "lr": 0.0026291804804649314, "grad_norm": 0.163879, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:25.104339+00:00", "epoch": 0, "step": 7366, "train_loss": 3.5365383625030518, "perplexity": 34.34781348031519, "lr": 0.0026291804804649314, "grad_norm": 0.153247, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:25.409552+00:00", "epoch": 0, "step": 7367, "train_loss": 3.652545928955078, "perplexity": 38.57274461291439, "lr": 0.0026291804804649314, "grad_norm": 0.156569, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:25.713728+00:00", "epoch": 0, "step": 7368, "train_loss": 3.560889482498169, "perplexity": 35.194488107865645, "lr": 0.0026291804804649314, "grad_norm": 0.139388, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:26.018831+00:00", "epoch": 0, "step": 7369, "train_loss": 3.6039412021636963, "perplexity": 36.7427600997919, "lr": 0.0026291804804649314, "grad_norm": 0.155429, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:26.325133+00:00", "epoch": 0, "step": 7370, "train_loss": 3.6277472972869873, "perplexity": 37.62795647109996, "lr": 0.0026291804804649314, "grad_norm": 0.141132, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:26.630756+00:00", "epoch": 0, "step": 7371, "train_loss": 3.5789337158203125, "perplexity": 35.83530984416721, "lr": 0.0026291804804649314, "grad_norm": 0.157681, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:26.936180+00:00", "epoch": 0, "step": 7372, "train_loss": 3.558276891708374, "perplexity": 35.102659320138464, "lr": 0.0026291804804649314, "grad_norm": 0.19199, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:27.241435+00:00", "epoch": 0, "step": 7373, "train_loss": 3.562969923019409, "perplexity": 35.26778436486418, "lr": 0.0026291804804649314, "grad_norm": 0.164368, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:27.546983+00:00", "epoch": 0, "step": 7374, "train_loss": 3.4641590118408203, "perplexity": 31.94957925377239, "lr": 0.0026291804804649314, "grad_norm": 0.159462, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:27.853239+00:00", "epoch": 0, "step": 7375, "train_loss": 3.5993850231170654, "perplexity": 36.57573429478374, "lr": 0.0026291804804649314, "grad_norm": 0.156308, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:28.159739+00:00", "epoch": 0, "step": 7376, "train_loss": 3.575108766555786, "perplexity": 35.69850340776634, "lr": 0.0026291804804649314, "grad_norm": 0.151507, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:28.466333+00:00", "epoch": 0, "step": 7377, "train_loss": 3.5926105976104736, "perplexity": 36.32879209762026, "lr": 0.0026291804804649314, "grad_norm": 0.163335, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:28.771495+00:00", "epoch": 0, "step": 7378, "train_loss": 3.460674285888672, "perplexity": 31.8384374875542, "lr": 0.0026291804804649314, "grad_norm": 0.178111, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:29.077986+00:00", "epoch": 0, "step": 7379, "train_loss": 3.566087007522583, "perplexity": 35.3778885418314, "lr": 0.0026291804804649314, "grad_norm": 0.207941, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:29.383194+00:00", "epoch": 0, "step": 7380, "train_loss": 3.5455000400543213, "perplexity": 34.65701090377742, "lr": 0.0026291804804649314, "grad_norm": 0.177372, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:29.688044+00:00", "epoch": 0, "step": 7381, "train_loss": 3.6264758110046387, "perplexity": 37.58014344385979, "lr": 0.0026291804804649314, "grad_norm": 0.172134, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:29.993798+00:00", "epoch": 0, "step": 7382, "train_loss": 3.5823943614959717, "perplexity": 35.95953798502419, "lr": 0.0026291804804649314, "grad_norm": 0.163786, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:30.299791+00:00", "epoch": 0, "step": 7383, "train_loss": 3.554899215698242, "perplexity": 34.9842939223761, "lr": 0.0026291804804649314, "grad_norm": 0.150201, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:30.605673+00:00", "epoch": 0, "step": 7384, "train_loss": 3.479865312576294, "perplexity": 32.45535045393104, "lr": 0.0026291804804649314, "grad_norm": 0.20805, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:30.909941+00:00", "epoch": 0, "step": 7385, "train_loss": 3.6805362701416016, "perplexity": 39.66766095182239, "lr": 0.0026291804804649314, "grad_norm": 0.155741, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:31.216024+00:00", "epoch": 0, "step": 7386, "train_loss": 3.566744089126587, "perplexity": 35.40114234056576, "lr": 0.0026291804804649314, "grad_norm": 0.197223, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:31.521392+00:00", "epoch": 0, "step": 7387, "train_loss": 3.4867701530456543, "perplexity": 32.68022493897402, "lr": 0.0026291804804649314, "grad_norm": 0.15394, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:31.827729+00:00", "epoch": 0, "step": 7388, "train_loss": 3.506505012512207, "perplexity": 33.33157055302007, "lr": 0.0026291804804649314, "grad_norm": 0.204648, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:32.135136+00:00", "epoch": 0, "step": 7389, "train_loss": 3.514497756958008, "perplexity": 33.59904879708968, "lr": 0.0026291804804649314, "grad_norm": 0.187147, "tokens_per_sec": 106595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:32.441296+00:00", "epoch": 0, "step": 7390, "train_loss": 3.5857737064361572, "perplexity": 36.08126322771523, "lr": 0.0026291804804649314, "grad_norm": 0.158426, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:32.746131+00:00", "epoch": 0, "step": 7391, "train_loss": 3.5755696296691895, "perplexity": 35.71495932286155, "lr": 0.0026291804804649314, "grad_norm": 0.149307, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:33.050762+00:00", "epoch": 0, "step": 7392, "train_loss": 3.5445351600646973, "perplexity": 34.62358717499283, "lr": 0.0026291804804649314, "grad_norm": 0.146567, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:33.356561+00:00", "epoch": 0, "step": 7393, "train_loss": 3.432405948638916, "perplexity": 30.951019804083153, "lr": 0.0026291804804649314, "grad_norm": 0.143345, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:33.662799+00:00", "epoch": 0, "step": 7394, "train_loss": 3.4959311485290527, "perplexity": 32.98098385416765, "lr": 0.0026291804804649314, "grad_norm": 0.168796, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:33.969507+00:00", "epoch": 0, "step": 7395, "train_loss": 3.511179208755493, "perplexity": 33.48773353878044, "lr": 0.0026291804804649314, "grad_norm": 0.159156, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:34.275351+00:00", "epoch": 0, "step": 7396, "train_loss": 3.6544113159179688, "perplexity": 38.644764859769126, "lr": 0.0026291804804649314, "grad_norm": 0.183412, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:34.582022+00:00", "epoch": 0, "step": 7397, "train_loss": 3.594083786010742, "perplexity": 36.38235069399925, "lr": 0.0026291804804649314, "grad_norm": 0.137747, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:34.887647+00:00", "epoch": 0, "step": 7398, "train_loss": 3.5183403491973877, "perplexity": 33.72840461290785, "lr": 0.0026291804804649314, "grad_norm": 0.165765, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:35.193789+00:00", "epoch": 0, "step": 7399, "train_loss": 3.5981814861297607, "perplexity": 36.53174052509708, "lr": 0.0026291804804649314, "grad_norm": 0.205038, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:35.498485+00:00", "epoch": 0, "step": 7400, "train_loss": 3.478327751159668, "perplexity": 32.40548670342299, "lr": 0.0026291804804649314, "grad_norm": 0.190375, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:35.802429+00:00", "epoch": 0, "step": 7401, "train_loss": 3.599989175796509, "perplexity": 36.59783829908493, "lr": 0.0026291804804649314, "grad_norm": 0.19143, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:36.107263+00:00", "epoch": 0, "step": 7402, "train_loss": 3.5759148597717285, "perplexity": 35.72729133049841, "lr": 0.0026291804804649314, "grad_norm": 0.170428, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:36.411744+00:00", "epoch": 0, "step": 7403, "train_loss": 3.576963424682617, "perplexity": 35.76477336228608, "lr": 0.0026291804804649314, "grad_norm": 0.200996, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:36.716867+00:00", "epoch": 0, "step": 7404, "train_loss": 3.706605911254883, "perplexity": 40.71538013068613, "lr": 0.0026291804804649314, "grad_norm": 0.17891, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:37.022463+00:00", "epoch": 0, "step": 7405, "train_loss": 3.6353042125701904, "perplexity": 37.91338487122376, "lr": 0.0026291804804649314, "grad_norm": 0.157693, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:37.328138+00:00", "epoch": 0, "step": 7406, "train_loss": 3.576003313064575, "perplexity": 35.730451666830106, "lr": 0.0026291804804649314, "grad_norm": 0.179195, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:37.633208+00:00", "epoch": 0, "step": 7407, "train_loss": 3.6056787967681885, "perplexity": 36.80665962114992, "lr": 0.0026291804804649314, "grad_norm": 0.157459, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:37.938879+00:00", "epoch": 0, "step": 7408, "train_loss": 3.6037611961364746, "perplexity": 36.73614677675404, "lr": 0.0026291804804649314, "grad_norm": 0.16498, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:38.245365+00:00", "epoch": 0, "step": 7409, "train_loss": 3.5812926292419434, "perplexity": 35.919942018268166, "lr": 0.0026291804804649314, "grad_norm": 0.155061, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:38.551374+00:00", "epoch": 0, "step": 7410, "train_loss": 3.5135180950164795, "perplexity": 33.566149205581716, "lr": 0.0026291804804649314, "grad_norm": 0.160769, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:38.855949+00:00", "epoch": 0, "step": 7411, "train_loss": 3.556790828704834, "perplexity": 35.05053329766655, "lr": 0.0026291804804649314, "grad_norm": 0.14952, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:39.161393+00:00", "epoch": 0, "step": 7412, "train_loss": 3.554478883743286, "perplexity": 34.969591995780185, "lr": 0.0026291804804649314, "grad_norm": 0.155899, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:39.466012+00:00", "epoch": 0, "step": 7413, "train_loss": 3.4905028343200684, "perplexity": 32.802437751487446, "lr": 0.0026291804804649314, "grad_norm": 0.141302, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:39.770497+00:00", "epoch": 0, "step": 7414, "train_loss": 3.5074634552001953, "perplexity": 33.36353226738479, "lr": 0.0026291804804649314, "grad_norm": 0.164295, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:40.077152+00:00", "epoch": 0, "step": 7415, "train_loss": 3.51773738861084, "perplexity": 33.70807384422119, "lr": 0.0026291804804649314, "grad_norm": 0.143804, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:40.384781+00:00", "epoch": 0, "step": 7416, "train_loss": 3.5330071449279785, "perplexity": 34.226737775828695, "lr": 0.0026291804804649314, "grad_norm": 0.151518, "tokens_per_sec": 106517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:40.691542+00:00", "epoch": 0, "step": 7417, "train_loss": 3.5917446613311768, "perplexity": 36.29734729512243, "lr": 0.0026291804804649314, "grad_norm": 0.139941, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:40.997582+00:00", "epoch": 0, "step": 7418, "train_loss": 3.5808606147766113, "perplexity": 35.904427435225834, "lr": 0.0026291804804649314, "grad_norm": 0.151013, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:41.302755+00:00", "epoch": 0, "step": 7419, "train_loss": 3.525980234146118, "perplexity": 33.98707258226121, "lr": 0.0026291804804649314, "grad_norm": 0.146234, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:41.608442+00:00", "epoch": 0, "step": 7420, "train_loss": 3.576313018798828, "perplexity": 35.741519306366015, "lr": 0.0026291804804649314, "grad_norm": 0.148635, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:41.913884+00:00", "epoch": 0, "step": 7421, "train_loss": 3.5443615913391113, "perplexity": 34.617578124598346, "lr": 0.0026291804804649314, "grad_norm": 0.153139, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:42.218921+00:00", "epoch": 0, "step": 7422, "train_loss": 3.567223072052002, "perplexity": 35.418102944882754, "lr": 0.0026291804804649314, "grad_norm": 0.15786, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:42.524104+00:00", "epoch": 0, "step": 7423, "train_loss": 3.60017728805542, "perplexity": 36.60472344868885, "lr": 0.0026291804804649314, "grad_norm": 0.158627, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:42.829108+00:00", "epoch": 0, "step": 7424, "train_loss": 3.5919077396392822, "perplexity": 36.3032670877898, "lr": 0.0026291804804649314, "grad_norm": 0.160703, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:43.134921+00:00", "epoch": 0, "step": 7425, "train_loss": 3.542335271835327, "perplexity": 34.547502872273206, "lr": 0.0026291804804649314, "grad_norm": 0.147629, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:43.440496+00:00", "epoch": 0, "step": 7426, "train_loss": 3.4945197105407715, "perplexity": 32.93446607686655, "lr": 0.0026291804804649314, "grad_norm": 0.153138, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:43.747713+00:00", "epoch": 0, "step": 7427, "train_loss": 3.504094123840332, "perplexity": 33.25130863730577, "lr": 0.0026291804804649314, "grad_norm": 0.163665, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:44.053906+00:00", "epoch": 0, "step": 7428, "train_loss": 3.4973092079162598, "perplexity": 33.02646493918772, "lr": 0.0026291804804649314, "grad_norm": 0.15441, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:44.358069+00:00", "epoch": 0, "step": 7429, "train_loss": 3.5161967277526855, "perplexity": 33.656181119055404, "lr": 0.0026291804804649314, "grad_norm": 0.148987, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:44.662815+00:00", "epoch": 0, "step": 7430, "train_loss": 3.606221914291382, "perplexity": 36.82665539249576, "lr": 0.0026291804804649314, "grad_norm": 0.173429, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:44.967830+00:00", "epoch": 0, "step": 7431, "train_loss": 3.636298656463623, "perplexity": 37.951106358130836, "lr": 0.0026291804804649314, "grad_norm": 0.151007, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:45.273781+00:00", "epoch": 0, "step": 7432, "train_loss": 3.468719244003296, "perplexity": 32.09560946543012, "lr": 0.0026291804804649314, "grad_norm": 0.179975, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:45.579452+00:00", "epoch": 0, "step": 7433, "train_loss": 3.463615655899048, "perplexity": 31.932223975521495, "lr": 0.0026291804804649314, "grad_norm": 0.183984, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:45.884700+00:00", "epoch": 0, "step": 7434, "train_loss": 3.5854523181915283, "perplexity": 36.06966899708699, "lr": 0.0026291804804649314, "grad_norm": 0.174298, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:46.189887+00:00", "epoch": 0, "step": 7435, "train_loss": 3.551546812057495, "perplexity": 34.867208815986366, "lr": 0.0026291804804649314, "grad_norm": 0.162839, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:46.495386+00:00", "epoch": 0, "step": 7436, "train_loss": 3.593672752380371, "perplexity": 36.36739939726545, "lr": 0.0026291804804649314, "grad_norm": 0.159087, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:46.801810+00:00", "epoch": 0, "step": 7437, "train_loss": 3.504580497741699, "perplexity": 33.26748513960864, "lr": 0.0026291804804649314, "grad_norm": 0.157149, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:47.107833+00:00", "epoch": 0, "step": 7438, "train_loss": 3.6243419647216797, "perplexity": 37.50003869046925, "lr": 0.0026291804804649314, "grad_norm": 0.152089, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:47.412953+00:00", "epoch": 0, "step": 7439, "train_loss": 3.537388801574707, "perplexity": 34.37703662741652, "lr": 0.0026291804804649314, "grad_norm": 0.171352, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:47.719293+00:00", "epoch": 0, "step": 7440, "train_loss": 3.5591485500335693, "perplexity": 35.133270184540606, "lr": 0.0026291804804649314, "grad_norm": 0.1779, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:48.023325+00:00", "epoch": 0, "step": 7441, "train_loss": 3.5883162021636963, "perplexity": 36.17311640398961, "lr": 0.0026291804804649314, "grad_norm": 0.154641, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:48.329369+00:00", "epoch": 0, "step": 7442, "train_loss": 3.4768691062927246, "perplexity": 32.35825306351015, "lr": 0.0026291804804649314, "grad_norm": 0.172221, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:48.635359+00:00", "epoch": 0, "step": 7443, "train_loss": 3.562340259552002, "perplexity": 35.245584519422536, "lr": 0.0026291804804649314, "grad_norm": 0.179641, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:48.941233+00:00", "epoch": 0, "step": 7444, "train_loss": 3.6353237628936768, "perplexity": 37.91412609740804, "lr": 0.0026291804804649314, "grad_norm": 0.150972, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:49.246839+00:00", "epoch": 0, "step": 7445, "train_loss": 3.528385877609253, "perplexity": 34.06893178379894, "lr": 0.0026291804804649314, "grad_norm": 0.175959, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:49.553120+00:00", "epoch": 0, "step": 7446, "train_loss": 3.5367817878723145, "perplexity": 34.35617562723306, "lr": 0.0026291804804649314, "grad_norm": 0.195597, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:49.859189+00:00", "epoch": 0, "step": 7447, "train_loss": 3.498997688293457, "perplexity": 33.08227658233811, "lr": 0.0026291804804649314, "grad_norm": 0.203027, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:50.165073+00:00", "epoch": 0, "step": 7448, "train_loss": 3.5973010063171387, "perplexity": 36.4995892214088, "lr": 0.0026291804804649314, "grad_norm": 0.17992, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:50.471088+00:00", "epoch": 0, "step": 7449, "train_loss": 3.5044867992401123, "perplexity": 33.26436817212936, "lr": 0.0026291804804649314, "grad_norm": 0.157455, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:50.776741+00:00", "epoch": 0, "step": 7450, "train_loss": 3.594693183898926, "perplexity": 36.40452877863195, "lr": 0.0026291804804649314, "grad_norm": 0.17074, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:51.082876+00:00", "epoch": 0, "step": 7451, "train_loss": 3.50288724899292, "perplexity": 33.21120267557411, "lr": 0.0026291804804649314, "grad_norm": 0.162208, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:51.388644+00:00", "epoch": 0, "step": 7452, "train_loss": 3.50291109085083, "perplexity": 33.21199450178859, "lr": 0.0026291804804649314, "grad_norm": 0.151401, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:51.695305+00:00", "epoch": 0, "step": 7453, "train_loss": 3.5469043254852295, "perplexity": 34.70571342738708, "lr": 0.0026291804804649314, "grad_norm": 0.152971, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:52.001254+00:00", "epoch": 0, "step": 7454, "train_loss": 3.534142255783081, "perplexity": 34.265610975934564, "lr": 0.0026291804804649314, "grad_norm": 0.140585, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:52.308095+00:00", "epoch": 0, "step": 7455, "train_loss": 3.5123705863952637, "perplexity": 33.52765385108702, "lr": 0.0026291804804649314, "grad_norm": 0.149496, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:52.613968+00:00", "epoch": 0, "step": 7456, "train_loss": 3.5160486698150635, "perplexity": 33.65119842316362, "lr": 0.0026291804804649314, "grad_norm": 0.165221, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:52.920499+00:00", "epoch": 0, "step": 7457, "train_loss": 3.6245040893554688, "perplexity": 37.506118863368606, "lr": 0.0026291804804649314, "grad_norm": 0.1522, "tokens_per_sec": 106900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:53.225602+00:00", "epoch": 0, "step": 7458, "train_loss": 3.557586908340454, "perplexity": 35.07844742290157, "lr": 0.0026291804804649314, "grad_norm": 0.15449, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:53.532426+00:00", "epoch": 0, "step": 7459, "train_loss": 3.4732837677001953, "perplexity": 32.24244549873583, "lr": 0.0026291804804649314, "grad_norm": 0.150609, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:53.838834+00:00", "epoch": 0, "step": 7460, "train_loss": 3.594348669052124, "perplexity": 36.391989038164304, "lr": 0.0026291804804649314, "grad_norm": 0.135622, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:54.143665+00:00", "epoch": 0, "step": 7461, "train_loss": 3.495671510696411, "perplexity": 32.97242185455897, "lr": 0.0026291804804649314, "grad_norm": 0.160764, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:54.448521+00:00", "epoch": 0, "step": 7462, "train_loss": 3.5710959434509277, "perplexity": 35.55553866643512, "lr": 0.0026291804804649314, "grad_norm": 0.161522, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:54.753914+00:00", "epoch": 0, "step": 7463, "train_loss": 3.592522144317627, "perplexity": 36.32557883844785, "lr": 0.0026291804804649314, "grad_norm": 0.155531, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:55.059303+00:00", "epoch": 0, "step": 7464, "train_loss": 3.49173903465271, "perplexity": 32.843013210477984, "lr": 0.0026291804804649314, "grad_norm": 0.143101, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:55.364726+00:00", "epoch": 0, "step": 7465, "train_loss": 3.4882726669311523, "perplexity": 32.729364337796014, "lr": 0.0026291804804649314, "grad_norm": 0.161359, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:55.671376+00:00", "epoch": 0, "step": 7466, "train_loss": 3.4717538356781006, "perplexity": 32.193154464477125, "lr": 0.0026291804804649314, "grad_norm": 0.167264, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:55.979066+00:00", "epoch": 0, "step": 7467, "train_loss": 3.535710573196411, "perplexity": 34.319392492527506, "lr": 0.0026291804804649314, "grad_norm": 0.152803, "tokens_per_sec": 106497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:56.284710+00:00", "epoch": 0, "step": 7468, "train_loss": 3.4596383571624756, "perplexity": 31.80547221334895, "lr": 0.0026291804804649314, "grad_norm": 0.155995, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:56.589187+00:00", "epoch": 0, "step": 7469, "train_loss": 3.6285600662231445, "perplexity": 37.65855173700735, "lr": 0.0026291804804649314, "grad_norm": 0.146553, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:56.894065+00:00", "epoch": 0, "step": 7470, "train_loss": 3.5677285194396973, "perplexity": 35.436009457512164, "lr": 0.0026291804804649314, "grad_norm": 0.163713, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:57.199365+00:00", "epoch": 0, "step": 7471, "train_loss": 3.674111843109131, "perplexity": 39.413635815169286, "lr": 0.0026291804804649314, "grad_norm": 0.15435, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:57.504694+00:00", "epoch": 0, "step": 7472, "train_loss": 3.529233694076538, "perplexity": 34.09782823289987, "lr": 0.0026291804804649314, "grad_norm": 0.149087, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:57.810855+00:00", "epoch": 0, "step": 7473, "train_loss": 3.5447499752044678, "perplexity": 34.631025644630526, "lr": 0.0026291804804649314, "grad_norm": 0.183001, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:58.117294+00:00", "epoch": 0, "step": 7474, "train_loss": 3.52003812789917, "perplexity": 33.78571661768868, "lr": 0.0026291804804649314, "grad_norm": 0.20682, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:58.421889+00:00", "epoch": 0, "step": 7475, "train_loss": 3.649277448654175, "perplexity": 38.44687616831512, "lr": 0.0026291804804649314, "grad_norm": 0.181154, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:58.728192+00:00", "epoch": 0, "step": 7476, "train_loss": 3.467172145843506, "perplexity": 32.04599279791283, "lr": 0.0026291804804649314, "grad_norm": 0.161001, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:59.034988+00:00", "epoch": 0, "step": 7477, "train_loss": 3.6246378421783447, "perplexity": 37.51113574814548, "lr": 0.0026291804804649314, "grad_norm": 0.139733, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:59.340078+00:00", "epoch": 0, "step": 7478, "train_loss": 3.5530784130096436, "perplexity": 34.920652382869314, "lr": 0.0026291804804649314, "grad_norm": 0.160996, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:59.645714+00:00", "epoch": 0, "step": 7479, "train_loss": 3.5345520973205566, "perplexity": 34.279657324811474, "lr": 0.0026291804804649314, "grad_norm": 0.162638, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:36:59.951484+00:00", "epoch": 0, "step": 7480, "train_loss": 3.5503766536712646, "perplexity": 34.82643252120035, "lr": 0.0026291804804649314, "grad_norm": 0.1599, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:00.256819+00:00", "epoch": 0, "step": 7481, "train_loss": 3.625868082046509, "perplexity": 37.55731184085414, "lr": 0.0026291804804649314, "grad_norm": 0.173892, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:00.562585+00:00", "epoch": 0, "step": 7482, "train_loss": 3.5452258586883545, "perplexity": 34.64750989974615, "lr": 0.0026291804804649314, "grad_norm": 0.159004, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:00.868470+00:00", "epoch": 0, "step": 7483, "train_loss": 3.556608200073242, "perplexity": 35.04413265122227, "lr": 0.0026291804804649314, "grad_norm": 0.150383, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:01.173998+00:00", "epoch": 0, "step": 7484, "train_loss": 3.6309800148010254, "perplexity": 37.74979385181406, "lr": 0.0026291804804649314, "grad_norm": 0.151081, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:01.480000+00:00", "epoch": 0, "step": 7485, "train_loss": 3.5579516887664795, "perplexity": 35.09124568803319, "lr": 0.0026291804804649314, "grad_norm": 0.150009, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:01.785306+00:00", "epoch": 0, "step": 7486, "train_loss": 3.5277178287506104, "perplexity": 34.04617967341245, "lr": 0.0026291804804649314, "grad_norm": 0.139555, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:02.090072+00:00", "epoch": 0, "step": 7487, "train_loss": 3.564756155014038, "perplexity": 35.33083710632364, "lr": 0.0026291804804649314, "grad_norm": 0.157698, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:02.396209+00:00", "epoch": 0, "step": 7488, "train_loss": 3.5260417461395264, "perplexity": 33.989163259146046, "lr": 0.0026291804804649314, "grad_norm": 0.143429, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:02.701381+00:00", "epoch": 0, "step": 7489, "train_loss": 3.5151867866516113, "perplexity": 33.62220751699431, "lr": 0.0026291804804649314, "grad_norm": 0.14871, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:03.007639+00:00", "epoch": 0, "step": 7490, "train_loss": 3.604738473892212, "perplexity": 36.77206574437573, "lr": 0.0026291804804649314, "grad_norm": 0.153701, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:03.311677+00:00", "epoch": 0, "step": 7491, "train_loss": 3.549663782119751, "perplexity": 34.801614595262976, "lr": 0.0026291804804649314, "grad_norm": 0.141235, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:03.616934+00:00", "epoch": 0, "step": 7492, "train_loss": 3.5574231147766113, "perplexity": 35.072702269506514, "lr": 0.0026291804804649314, "grad_norm": 0.141921, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:03.923604+00:00", "epoch": 0, "step": 7493, "train_loss": 3.4881513118743896, "perplexity": 32.72539270492273, "lr": 0.0026291804804649314, "grad_norm": 0.164785, "tokens_per_sec": 106851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:04.231965+00:00", "epoch": 0, "step": 7494, "train_loss": 3.5961685180664062, "perplexity": 36.458277262528334, "lr": 0.0026291804804649314, "grad_norm": 0.169596, "tokens_per_sec": 106266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:04.538125+00:00", "epoch": 0, "step": 7495, "train_loss": 3.559126377105713, "perplexity": 35.132491185711814, "lr": 0.0026291804804649314, "grad_norm": 0.161708, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:04.842965+00:00", "epoch": 0, "step": 7496, "train_loss": 3.5636496543884277, "perplexity": 35.291765133529786, "lr": 0.0026291804804649314, "grad_norm": 0.162705, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:05.148665+00:00", "epoch": 0, "step": 7497, "train_loss": 3.5666871070861816, "perplexity": 35.39912516871435, "lr": 0.0026291804804649314, "grad_norm": 0.169907, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:05.453422+00:00", "epoch": 0, "step": 7498, "train_loss": 3.5773606300354004, "perplexity": 35.77898214342181, "lr": 0.0026291804804649314, "grad_norm": 0.158901, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:05.760572+00:00", "epoch": 0, "step": 7499, "train_loss": 3.5379693508148193, "perplexity": 34.39699998420299, "lr": 0.0026291804804649314, "grad_norm": 0.141779, "tokens_per_sec": 106684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:06.067758+00:00", "epoch": 0, "step": 7500, "train_loss": 3.416325569152832, "perplexity": 30.457295932583012, "lr": 0.0026291804804649314, "grad_norm": 0.17035, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:37:09.141116+00:00", "step": 7500, "epoch": 0, "val_loss": 3.519768536090851, "val_ppl": 33.77660949290864, "eval_train_loss": 3.416325569152832, "eval_train_ppl": 30.457295932583012} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:37:10.089066+00:00", "step": 7500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5198_epoch_0000_step_0007500.pt", "val_loss": 3.519768536090851} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T21:37:13.104242+00:00", "step": 7500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0007500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:14.279172+00:00", "epoch": 0, "step": 7501, "train_loss": 3.467510461807251, "perplexity": 32.05683630301244, "lr": 0.0026291804804649314, "grad_norm": 0.179141, "tokens_per_sec": 3990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:14.583256+00:00", "epoch": 0, "step": 7502, "train_loss": 3.576746702194214, "perplexity": 35.757023171456474, "lr": 0.0026291804804649314, "grad_norm": 0.156839, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:14.887340+00:00", "epoch": 0, "step": 7503, "train_loss": 3.544407367706299, "perplexity": 34.619162827836426, "lr": 0.0026291804804649314, "grad_norm": 0.171283, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:15.190745+00:00", "epoch": 0, "step": 7504, "train_loss": 3.5235843658447266, "perplexity": 33.905741500589144, "lr": 0.0026291804804649314, "grad_norm": 0.177836, "tokens_per_sec": 107992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:15.495901+00:00", "epoch": 0, "step": 7505, "train_loss": 3.639089584350586, "perplexity": 38.05717310267585, "lr": 0.0026291804804649314, "grad_norm": 0.174251, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:15.800336+00:00", "epoch": 0, "step": 7506, "train_loss": 3.639094352722168, "perplexity": 38.05735457385122, "lr": 0.0026291804804649314, "grad_norm": 0.16383, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:16.104396+00:00", "epoch": 0, "step": 7507, "train_loss": 3.582369327545166, "perplexity": 35.95863778698708, "lr": 0.0026291804804649314, "grad_norm": 0.18568, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:16.409528+00:00", "epoch": 0, "step": 7508, "train_loss": 3.5473082065582275, "perplexity": 34.719733239143025, "lr": 0.0026291804804649314, "grad_norm": 0.166158, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:16.714540+00:00", "epoch": 0, "step": 7509, "train_loss": 3.626476526260376, "perplexity": 37.58017032328261, "lr": 0.0026291804804649314, "grad_norm": 0.175548, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:17.020461+00:00", "epoch": 0, "step": 7510, "train_loss": 3.5946807861328125, "perplexity": 36.404077446596446, "lr": 0.0026291804804649314, "grad_norm": 0.169389, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:17.325444+00:00", "epoch": 0, "step": 7511, "train_loss": 3.613006591796875, "perplexity": 37.07736189260943, "lr": 0.0026291804804649314, "grad_norm": 0.162113, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:17.629800+00:00", "epoch": 0, "step": 7512, "train_loss": 3.557081937789917, "perplexity": 35.06073831166057, "lr": 0.0026291804804649314, "grad_norm": 0.152303, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:17.933574+00:00", "epoch": 0, "step": 7513, "train_loss": 3.6728715896606445, "perplexity": 39.36478321849509, "lr": 0.0026291804804649314, "grad_norm": 0.140906, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:18.238117+00:00", "epoch": 0, "step": 7514, "train_loss": 3.4785873889923096, "perplexity": 32.413901486107065, "lr": 0.0026291804804649314, "grad_norm": 0.145505, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:18.544366+00:00", "epoch": 0, "step": 7515, "train_loss": 3.5355451107025146, "perplexity": 34.313714390026135, "lr": 0.0026291804804649314, "grad_norm": 0.150296, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:18.850252+00:00", "epoch": 0, "step": 7516, "train_loss": 3.5950024127960205, "perplexity": 36.41578785164106, "lr": 0.0026291804804649314, "grad_norm": 0.170913, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:19.155278+00:00", "epoch": 0, "step": 7517, "train_loss": 3.586522340774536, "perplexity": 36.10828501379602, "lr": 0.0026291804804649314, "grad_norm": 0.182654, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:19.462074+00:00", "epoch": 0, "step": 7518, "train_loss": 3.52521014213562, "perplexity": 33.96090948449466, "lr": 0.0026291804804649314, "grad_norm": 0.178817, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:19.767272+00:00", "epoch": 0, "step": 7519, "train_loss": 3.480152130126953, "perplexity": 32.46466055314013, "lr": 0.0026291804804649314, "grad_norm": 0.181694, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:20.073270+00:00", "epoch": 0, "step": 7520, "train_loss": 3.4897193908691406, "perplexity": 32.776748960628005, "lr": 0.0026291804804649314, "grad_norm": 0.145125, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:20.379939+00:00", "epoch": 0, "step": 7521, "train_loss": 3.5756611824035645, "perplexity": 35.718229274729886, "lr": 0.0026291804804649314, "grad_norm": 0.145192, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:20.685317+00:00", "epoch": 0, "step": 7522, "train_loss": 3.638522148132324, "perplexity": 38.03558421003181, "lr": 0.0026291804804649314, "grad_norm": 0.158269, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:20.991391+00:00", "epoch": 0, "step": 7523, "train_loss": 3.622173309326172, "perplexity": 37.41880214810278, "lr": 0.0026291804804649314, "grad_norm": 0.158255, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:21.297015+00:00", "epoch": 0, "step": 7524, "train_loss": 3.5632994174957275, "perplexity": 35.27940681966657, "lr": 0.0026291804804649314, "grad_norm": 0.17339, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:21.601942+00:00", "epoch": 0, "step": 7525, "train_loss": 3.5293235778808594, "perplexity": 34.10089321316452, "lr": 0.0026291804804649314, "grad_norm": 0.15407, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:21.907706+00:00", "epoch": 0, "step": 7526, "train_loss": 3.518047571182251, "perplexity": 33.718531122992474, "lr": 0.0026291804804649314, "grad_norm": 0.169437, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:22.213632+00:00", "epoch": 0, "step": 7527, "train_loss": 3.5103540420532227, "perplexity": 33.460111973894826, "lr": 0.0026291804804649314, "grad_norm": 0.167578, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:22.517680+00:00", "epoch": 0, "step": 7528, "train_loss": 3.5175859928131104, "perplexity": 33.70297096977824, "lr": 0.0026291804804649314, "grad_norm": 0.14522, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:22.822752+00:00", "epoch": 0, "step": 7529, "train_loss": 3.598010540008545, "perplexity": 36.525496099498426, "lr": 0.0026291804804649314, "grad_norm": 0.189165, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:23.127696+00:00", "epoch": 0, "step": 7530, "train_loss": 3.567098379135132, "perplexity": 35.41368683365117, "lr": 0.0026291804804649314, "grad_norm": 0.162171, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:23.432489+00:00", "epoch": 0, "step": 7531, "train_loss": 3.5697786808013916, "perplexity": 35.5087335174709, "lr": 0.0026291804804649314, "grad_norm": 0.139666, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:23.736996+00:00", "epoch": 0, "step": 7532, "train_loss": 3.542581558227539, "perplexity": 34.55601249997529, "lr": 0.0026291804804649314, "grad_norm": 0.166674, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:24.042412+00:00", "epoch": 0, "step": 7533, "train_loss": 3.5844175815582275, "perplexity": 36.0323656921011, "lr": 0.0026291804804649314, "grad_norm": 0.165658, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:24.346472+00:00", "epoch": 0, "step": 7534, "train_loss": 3.5428566932678223, "perplexity": 34.56552137791927, "lr": 0.0026291804804649314, "grad_norm": 0.156268, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:24.651596+00:00", "epoch": 0, "step": 7535, "train_loss": 3.5460615158081055, "perplexity": 34.67647543901946, "lr": 0.0026291804804649314, "grad_norm": 0.18992, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:24.957639+00:00", "epoch": 0, "step": 7536, "train_loss": 3.4976375102996826, "perplexity": 33.03730938637481, "lr": 0.0026291804804649314, "grad_norm": 0.185217, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:25.265330+00:00", "epoch": 0, "step": 7537, "train_loss": 3.513572931289673, "perplexity": 33.56798989857751, "lr": 0.0026291804804649314, "grad_norm": 0.190663, "tokens_per_sec": 106496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:25.571613+00:00", "epoch": 0, "step": 7538, "train_loss": 3.542235851287842, "perplexity": 34.54406831135918, "lr": 0.0026291804804649314, "grad_norm": 0.161209, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:25.878233+00:00", "epoch": 0, "step": 7539, "train_loss": 3.540785789489746, "perplexity": 34.49401357754948, "lr": 0.0026291804804649314, "grad_norm": 0.150172, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:26.183490+00:00", "epoch": 0, "step": 7540, "train_loss": 3.5440616607666016, "perplexity": 34.60719681148695, "lr": 0.0026291804804649314, "grad_norm": 0.160369, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:26.489437+00:00", "epoch": 0, "step": 7541, "train_loss": 3.578171491622925, "perplexity": 35.80800571114048, "lr": 0.0026291804804649314, "grad_norm": 0.167923, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:26.794704+00:00", "epoch": 0, "step": 7542, "train_loss": 3.6135103702545166, "perplexity": 37.09604537456983, "lr": 0.0026291804804649314, "grad_norm": 0.178861, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:27.100735+00:00", "epoch": 0, "step": 7543, "train_loss": 3.5956246852874756, "perplexity": 36.438455446654764, "lr": 0.0026291804804649314, "grad_norm": 0.195767, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:27.406791+00:00", "epoch": 0, "step": 7544, "train_loss": 3.5164990425109863, "perplexity": 33.66635741746083, "lr": 0.0026291804804649314, "grad_norm": 0.160133, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:27.711937+00:00", "epoch": 0, "step": 7545, "train_loss": 3.5588831901550293, "perplexity": 35.12394846109212, "lr": 0.0026291804804649314, "grad_norm": 0.140312, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:28.018179+00:00", "epoch": 0, "step": 7546, "train_loss": 3.5137276649475098, "perplexity": 33.57318439831236, "lr": 0.0026291804804649314, "grad_norm": 0.164001, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:28.324641+00:00", "epoch": 0, "step": 7547, "train_loss": 3.6055171489715576, "perplexity": 36.8007103865741, "lr": 0.0026291804804649314, "grad_norm": 0.146913, "tokens_per_sec": 106921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:28.631405+00:00", "epoch": 0, "step": 7548, "train_loss": 3.497992515563965, "perplexity": 33.049039887196265, "lr": 0.0026291804804649314, "grad_norm": 0.15412, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:28.938096+00:00", "epoch": 0, "step": 7549, "train_loss": 3.5273334980010986, "perplexity": 34.03309719382119, "lr": 0.0026291804804649314, "grad_norm": 0.166639, "tokens_per_sec": 106844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:29.243956+00:00", "epoch": 0, "step": 7550, "train_loss": 3.5808396339416504, "perplexity": 35.90367413826187, "lr": 0.0026291804804649314, "grad_norm": 0.165496, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:29.549471+00:00", "epoch": 0, "step": 7551, "train_loss": 3.5487899780273438, "perplexity": 34.77121808423918, "lr": 0.0026291804804649314, "grad_norm": 0.149268, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:29.854641+00:00", "epoch": 0, "step": 7552, "train_loss": 3.6149168014526367, "perplexity": 37.14825511621013, "lr": 0.0026291804804649314, "grad_norm": 0.151943, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:30.159807+00:00", "epoch": 0, "step": 7553, "train_loss": 3.56577467918396, "perplexity": 35.36684075003818, "lr": 0.0026291804804649314, "grad_norm": 0.173868, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:30.465411+00:00", "epoch": 0, "step": 7554, "train_loss": 3.493966579437256, "perplexity": 32.91625403658906, "lr": 0.0026291804804649314, "grad_norm": 0.154627, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:30.770412+00:00", "epoch": 0, "step": 7555, "train_loss": 3.704014539718628, "perplexity": 40.610008041610115, "lr": 0.0026291804804649314, "grad_norm": 0.159555, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:31.076829+00:00", "epoch": 0, "step": 7556, "train_loss": 3.6159801483154297, "perplexity": 37.18777760607996, "lr": 0.0026291804804649314, "grad_norm": 0.152617, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:31.383324+00:00", "epoch": 0, "step": 7557, "train_loss": 3.5766873359680176, "perplexity": 35.75490047493964, "lr": 0.0026291804804649314, "grad_norm": 0.148478, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:31.690341+00:00", "epoch": 0, "step": 7558, "train_loss": 3.490002393722534, "perplexity": 32.786026186787645, "lr": 0.0026291804804649314, "grad_norm": 0.154664, "tokens_per_sec": 106740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:31.996519+00:00", "epoch": 0, "step": 7559, "train_loss": 3.570263147354126, "perplexity": 35.52594047895218, "lr": 0.0026291804804649314, "grad_norm": 0.172909, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:32.301921+00:00", "epoch": 0, "step": 7560, "train_loss": 3.5398268699645996, "perplexity": 34.4609524484438, "lr": 0.0026291804804649314, "grad_norm": 0.175051, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:32.606261+00:00", "epoch": 0, "step": 7561, "train_loss": 3.6603784561157227, "perplexity": 38.87605296759469, "lr": 0.0026291804804649314, "grad_norm": 0.168955, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:32.912772+00:00", "epoch": 0, "step": 7562, "train_loss": 3.5171778202056885, "perplexity": 33.689217147397365, "lr": 0.0026291804804649314, "grad_norm": 0.168871, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:33.219187+00:00", "epoch": 0, "step": 7563, "train_loss": 3.584092378616333, "perplexity": 36.0206497659047, "lr": 0.0026291804804649314, "grad_norm": 0.146752, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:33.524700+00:00", "epoch": 0, "step": 7564, "train_loss": 3.495058059692383, "perplexity": 32.95220109212468, "lr": 0.0026291804804649314, "grad_norm": 0.152085, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:33.830155+00:00", "epoch": 0, "step": 7565, "train_loss": 3.51377534866333, "perplexity": 33.574785330665286, "lr": 0.0026291804804649314, "grad_norm": 0.144024, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:34.136244+00:00", "epoch": 0, "step": 7566, "train_loss": 3.561021089553833, "perplexity": 35.199120255626106, "lr": 0.0026291804804649314, "grad_norm": 0.176602, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:34.443096+00:00", "epoch": 0, "step": 7567, "train_loss": 3.598317861557007, "perplexity": 36.53672289654803, "lr": 0.0026291804804649314, "grad_norm": 0.16961, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:34.750415+00:00", "epoch": 0, "step": 7568, "train_loss": 3.656886339187622, "perplexity": 38.74053001365344, "lr": 0.0026291804804649314, "grad_norm": 0.15969, "tokens_per_sec": 106627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:35.056560+00:00", "epoch": 0, "step": 7569, "train_loss": 3.538957357406616, "perplexity": 34.43100124089241, "lr": 0.0026291804804649314, "grad_norm": 0.164076, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:35.362253+00:00", "epoch": 0, "step": 7570, "train_loss": 3.5451254844665527, "perplexity": 34.64403235743331, "lr": 0.0026291804804649314, "grad_norm": 0.174456, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:35.667788+00:00", "epoch": 0, "step": 7571, "train_loss": 3.58835768699646, "perplexity": 36.17461707080141, "lr": 0.0026291804804649314, "grad_norm": 0.196215, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:35.973980+00:00", "epoch": 0, "step": 7572, "train_loss": 3.682708263397217, "perplexity": 39.75391247883523, "lr": 0.0026291804804649314, "grad_norm": 0.215493, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:36.279736+00:00", "epoch": 0, "step": 7573, "train_loss": 3.5902371406555176, "perplexity": 36.24266951789843, "lr": 0.0026291804804649314, "grad_norm": 0.18436, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:36.584641+00:00", "epoch": 0, "step": 7574, "train_loss": 3.5892348289489746, "perplexity": 36.20636126509737, "lr": 0.0026291804804649314, "grad_norm": 0.180549, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:36.890320+00:00", "epoch": 0, "step": 7575, "train_loss": 3.55792498588562, "perplexity": 35.09030866319106, "lr": 0.0026291804804649314, "grad_norm": 0.154705, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:37.195986+00:00", "epoch": 0, "step": 7576, "train_loss": 3.5437302589416504, "perplexity": 34.5957298235025, "lr": 0.0026291804804649314, "grad_norm": 0.173979, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:37.502083+00:00", "epoch": 0, "step": 7577, "train_loss": 3.623586893081665, "perplexity": 37.47173416207275, "lr": 0.0026291804804649314, "grad_norm": 0.175953, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:37.808446+00:00", "epoch": 0, "step": 7578, "train_loss": 3.59295916557312, "perplexity": 36.34145735789088, "lr": 0.0026291804804649314, "grad_norm": 0.177061, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:38.115214+00:00", "epoch": 0, "step": 7579, "train_loss": 3.458834171295166, "perplexity": 31.779904983872065, "lr": 0.0026291804804649314, "grad_norm": 0.15532, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:38.420683+00:00", "epoch": 0, "step": 7580, "train_loss": 3.479316234588623, "perplexity": 32.437534826946546, "lr": 0.0026291804804649314, "grad_norm": 0.133814, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:38.727134+00:00", "epoch": 0, "step": 7581, "train_loss": 3.4260051250457764, "perplexity": 30.7535404756974, "lr": 0.0026291804804649314, "grad_norm": 0.140158, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:39.033278+00:00", "epoch": 0, "step": 7582, "train_loss": 3.521843910217285, "perplexity": 33.84678138560725, "lr": 0.0026291804804649314, "grad_norm": 0.16315, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:39.340391+00:00", "epoch": 0, "step": 7583, "train_loss": 3.648042917251587, "perplexity": 38.39944157812397, "lr": 0.0026291804804649314, "grad_norm": 0.168793, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:39.645275+00:00", "epoch": 0, "step": 7584, "train_loss": 3.6554348468780518, "perplexity": 38.684339222386015, "lr": 0.0026291804804649314, "grad_norm": 0.157834, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:39.951267+00:00", "epoch": 0, "step": 7585, "train_loss": 3.5699024200439453, "perplexity": 35.51312761311583, "lr": 0.0026291804804649314, "grad_norm": 0.146102, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:40.256852+00:00", "epoch": 0, "step": 7586, "train_loss": 3.5675277709960938, "perplexity": 35.428896447752756, "lr": 0.0026291804804649314, "grad_norm": 0.143193, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:40.562329+00:00", "epoch": 0, "step": 7587, "train_loss": 3.5693917274475098, "perplexity": 35.49499595201415, "lr": 0.0026291804804649314, "grad_norm": 0.140025, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:40.869498+00:00", "epoch": 0, "step": 7588, "train_loss": 3.589095115661621, "perplexity": 36.201303108695974, "lr": 0.0026291804804649314, "grad_norm": 0.146937, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:41.177138+00:00", "epoch": 0, "step": 7589, "train_loss": 3.6099421977996826, "perplexity": 36.96391615738511, "lr": 0.0026291804804649314, "grad_norm": 0.171472, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:41.483247+00:00", "epoch": 0, "step": 7590, "train_loss": 3.4129068851470947, "perplexity": 30.353349842775902, "lr": 0.0026291804804649314, "grad_norm": 0.156529, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:41.789086+00:00", "epoch": 0, "step": 7591, "train_loss": 3.4819650650024414, "perplexity": 32.52357025206661, "lr": 0.0026291804804649314, "grad_norm": 0.159249, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:42.094306+00:00", "epoch": 0, "step": 7592, "train_loss": 3.575533151626587, "perplexity": 35.71365653481555, "lr": 0.0026291804804649314, "grad_norm": 0.170744, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:42.400418+00:00", "epoch": 0, "step": 7593, "train_loss": 3.536675214767456, "perplexity": 34.35251437802422, "lr": 0.0026291804804649314, "grad_norm": 0.159294, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:42.706311+00:00", "epoch": 0, "step": 7594, "train_loss": 3.542612075805664, "perplexity": 34.557067081878, "lr": 0.0026291804804649314, "grad_norm": 0.150478, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:43.012170+00:00", "epoch": 0, "step": 7595, "train_loss": 3.498293399810791, "perplexity": 33.05898531880787, "lr": 0.0026291804804649314, "grad_norm": 0.152904, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:43.317761+00:00", "epoch": 0, "step": 7596, "train_loss": 3.5210330486297607, "perplexity": 33.81934745478382, "lr": 0.0026291804804649314, "grad_norm": 0.174621, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:43.622814+00:00", "epoch": 0, "step": 7597, "train_loss": 3.5838537216186523, "perplexity": 36.012054211512464, "lr": 0.0026291804804649314, "grad_norm": 0.171031, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:43.929909+00:00", "epoch": 0, "step": 7598, "train_loss": 3.5530197620391846, "perplexity": 34.91860431277924, "lr": 0.0026291804804649314, "grad_norm": 0.149283, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:44.236076+00:00", "epoch": 0, "step": 7599, "train_loss": 3.550910472869873, "perplexity": 34.84502850250353, "lr": 0.0026291804804649314, "grad_norm": 0.166445, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:44.541341+00:00", "epoch": 0, "step": 7600, "train_loss": 3.6054046154022217, "perplexity": 36.79656930428995, "lr": 0.0026291804804649314, "grad_norm": 0.156625, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:44.846521+00:00", "epoch": 0, "step": 7601, "train_loss": 3.603908061981201, "perplexity": 36.741542458193365, "lr": 0.0026291804804649314, "grad_norm": 0.164776, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:45.150956+00:00", "epoch": 0, "step": 7602, "train_loss": 3.6981313228607178, "perplexity": 40.37179198316418, "lr": 0.0026291804804649314, "grad_norm": 0.175447, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:45.456462+00:00", "epoch": 0, "step": 7603, "train_loss": 3.5765023231506348, "perplexity": 35.74828597197031, "lr": 0.0026291804804649314, "grad_norm": 0.147419, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:45.763331+00:00", "epoch": 0, "step": 7604, "train_loss": 3.616488456726074, "perplexity": 37.206685271265826, "lr": 0.0026291804804649314, "grad_norm": 0.160589, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:46.069298+00:00", "epoch": 0, "step": 7605, "train_loss": 3.711735248565674, "perplexity": 40.92475957895555, "lr": 0.0026291804804649314, "grad_norm": 0.178478, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:46.374321+00:00", "epoch": 0, "step": 7606, "train_loss": 3.546377420425415, "perplexity": 34.687431628186935, "lr": 0.0026291804804649314, "grad_norm": 0.163781, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:46.679408+00:00", "epoch": 0, "step": 7607, "train_loss": 3.5918667316436768, "perplexity": 36.30177839409698, "lr": 0.0026291804804649314, "grad_norm": 0.177609, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:46.986348+00:00", "epoch": 0, "step": 7608, "train_loss": 3.5924932956695557, "perplexity": 36.32453090972369, "lr": 0.0026291804804649314, "grad_norm": 0.152073, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:47.291907+00:00", "epoch": 0, "step": 7609, "train_loss": 3.5225021839141846, "perplexity": 33.869069166443225, "lr": 0.0026291804804649314, "grad_norm": 0.19716, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:47.597763+00:00", "epoch": 0, "step": 7610, "train_loss": 3.710326671600342, "perplexity": 40.86715448542926, "lr": 0.0026291804804649314, "grad_norm": 0.184672, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:47.903937+00:00", "epoch": 0, "step": 7611, "train_loss": 3.5432565212249756, "perplexity": 34.57934440295152, "lr": 0.0026291804804649314, "grad_norm": 0.16012, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:48.209911+00:00", "epoch": 0, "step": 7612, "train_loss": 3.6723811626434326, "perplexity": 39.345482398486695, "lr": 0.0026291804804649314, "grad_norm": 0.18134, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:48.514496+00:00", "epoch": 0, "step": 7613, "train_loss": 3.5631372928619385, "perplexity": 35.27368762237955, "lr": 0.0026291804804649314, "grad_norm": 0.171758, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:48.820371+00:00", "epoch": 0, "step": 7614, "train_loss": 3.5432419776916504, "perplexity": 34.57884150076083, "lr": 0.0026291804804649314, "grad_norm": 0.166166, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:49.125846+00:00", "epoch": 0, "step": 7615, "train_loss": 3.5315065383911133, "perplexity": 34.175415426344, "lr": 0.0026291804804649314, "grad_norm": 0.159182, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:49.431606+00:00", "epoch": 0, "step": 7616, "train_loss": 3.536290407180786, "perplexity": 34.33929781295071, "lr": 0.0026291804804649314, "grad_norm": 0.174381, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:49.738016+00:00", "epoch": 0, "step": 7617, "train_loss": 3.4724321365356445, "perplexity": 32.214998516339456, "lr": 0.0026291804804649314, "grad_norm": 0.164198, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:50.044512+00:00", "epoch": 0, "step": 7618, "train_loss": 3.565938949584961, "perplexity": 35.37265095235939, "lr": 0.0026291804804649314, "grad_norm": 0.161833, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:50.350298+00:00", "epoch": 0, "step": 7619, "train_loss": 3.5515894889831543, "perplexity": 34.86869687301757, "lr": 0.0026291804804649314, "grad_norm": 0.149575, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:50.656124+00:00", "epoch": 0, "step": 7620, "train_loss": 3.535527229309082, "perplexity": 34.31310081848476, "lr": 0.0026291804804649314, "grad_norm": 0.178505, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:50.962215+00:00", "epoch": 0, "step": 7621, "train_loss": 3.4789044857025146, "perplexity": 32.42418145741962, "lr": 0.0026291804804649314, "grad_norm": 0.163631, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:51.268400+00:00", "epoch": 0, "step": 7622, "train_loss": 3.521749258041382, "perplexity": 33.84357786571424, "lr": 0.0026291804804649314, "grad_norm": 0.13874, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:51.574015+00:00", "epoch": 0, "step": 7623, "train_loss": 3.5684378147125244, "perplexity": 35.46115296754049, "lr": 0.0026291804804649314, "grad_norm": 0.155235, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:51.879930+00:00", "epoch": 0, "step": 7624, "train_loss": 3.546146869659424, "perplexity": 34.67943533606584, "lr": 0.0026291804804649314, "grad_norm": 0.17547, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:52.186533+00:00", "epoch": 0, "step": 7625, "train_loss": 3.549992799758911, "perplexity": 34.81306682422543, "lr": 0.0026291804804649314, "grad_norm": 0.145549, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:52.492419+00:00", "epoch": 0, "step": 7626, "train_loss": 3.561609983444214, "perplexity": 35.219854907147, "lr": 0.0026291804804649314, "grad_norm": 0.141171, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:52.797428+00:00", "epoch": 0, "step": 7627, "train_loss": 3.607666254043579, "perplexity": 36.87988402568161, "lr": 0.0026291804804649314, "grad_norm": 0.157123, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:53.103930+00:00", "epoch": 0, "step": 7628, "train_loss": 3.5793354511260986, "perplexity": 35.84970904546551, "lr": 0.0026291804804649314, "grad_norm": 0.156208, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:53.408683+00:00", "epoch": 0, "step": 7629, "train_loss": 3.548701047897339, "perplexity": 34.76812601278572, "lr": 0.0026291804804649314, "grad_norm": 0.149817, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:53.713476+00:00", "epoch": 0, "step": 7630, "train_loss": 3.5930380821228027, "perplexity": 36.34432541348304, "lr": 0.0026291804804649314, "grad_norm": 0.139955, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:54.018939+00:00", "epoch": 0, "step": 7631, "train_loss": 3.5931754112243652, "perplexity": 36.3493168897686, "lr": 0.0026291804804649314, "grad_norm": 0.137182, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:54.325200+00:00", "epoch": 0, "step": 7632, "train_loss": 3.536505937576294, "perplexity": 34.346699773034835, "lr": 0.0026291804804649314, "grad_norm": 0.167598, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:54.630065+00:00", "epoch": 0, "step": 7633, "train_loss": 3.535029172897339, "perplexity": 34.296015213766005, "lr": 0.0026291804804649314, "grad_norm": 0.16597, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:54.934893+00:00", "epoch": 0, "step": 7634, "train_loss": 3.507213830947876, "perplexity": 33.35520495998, "lr": 0.0026291804804649314, "grad_norm": 0.166044, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:55.240695+00:00", "epoch": 0, "step": 7635, "train_loss": 3.5469918251037598, "perplexity": 34.708750296933324, "lr": 0.0026291804804649314, "grad_norm": 0.149454, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:55.547177+00:00", "epoch": 0, "step": 7636, "train_loss": 3.6391334533691406, "perplexity": 38.0588426701297, "lr": 0.0026291804804649314, "grad_norm": 0.148962, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:55.854018+00:00", "epoch": 0, "step": 7637, "train_loss": 3.558688163757324, "perplexity": 35.11709903188204, "lr": 0.0026291804804649314, "grad_norm": 0.156649, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:56.160922+00:00", "epoch": 0, "step": 7638, "train_loss": 3.5351192951202393, "perplexity": 34.29910618617426, "lr": 0.0026291804804649314, "grad_norm": 0.149113, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:56.465398+00:00", "epoch": 0, "step": 7639, "train_loss": 3.5923545360565186, "perplexity": 36.31949088155517, "lr": 0.0026291804804649314, "grad_norm": 0.135049, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:56.771093+00:00", "epoch": 0, "step": 7640, "train_loss": 3.4906327724456787, "perplexity": 32.80670031569313, "lr": 0.0026291804804649314, "grad_norm": 0.17005, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:57.076188+00:00", "epoch": 0, "step": 7641, "train_loss": 3.528045177459717, "perplexity": 34.057326470723865, "lr": 0.0026291804804649314, "grad_norm": 0.195524, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:57.382427+00:00", "epoch": 0, "step": 7642, "train_loss": 3.6425201892852783, "perplexity": 38.18795643322564, "lr": 0.0026291804804649314, "grad_norm": 0.187646, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:57.687772+00:00", "epoch": 0, "step": 7643, "train_loss": 3.5100257396698, "perplexity": 33.44912874239339, "lr": 0.0026291804804649314, "grad_norm": 0.144989, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:57.994687+00:00", "epoch": 0, "step": 7644, "train_loss": 3.4800713062286377, "perplexity": 32.462036738751614, "lr": 0.0026291804804649314, "grad_norm": 0.148373, "tokens_per_sec": 106768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:58.300551+00:00", "epoch": 0, "step": 7645, "train_loss": 3.5677411556243896, "perplexity": 35.43645723630153, "lr": 0.0026291804804649314, "grad_norm": 0.158164, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:58.605094+00:00", "epoch": 0, "step": 7646, "train_loss": 3.4989001750946045, "perplexity": 33.0790507810049, "lr": 0.0026291804804649314, "grad_norm": 0.145196, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:58.910406+00:00", "epoch": 0, "step": 7647, "train_loss": 3.6503710746765137, "perplexity": 38.488945672529724, "lr": 0.0026291804804649314, "grad_norm": 0.165806, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:59.216510+00:00", "epoch": 0, "step": 7648, "train_loss": 3.6140387058258057, "perplexity": 37.11564971327409, "lr": 0.0026291804804649314, "grad_norm": 0.191762, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:59.523213+00:00", "epoch": 0, "step": 7649, "train_loss": 3.658493995666504, "perplexity": 38.80286156816397, "lr": 0.0026291804804649314, "grad_norm": 0.149602, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:37:59.828674+00:00", "epoch": 0, "step": 7650, "train_loss": 3.5549778938293457, "perplexity": 34.98704652952346, "lr": 0.0026291804804649314, "grad_norm": 0.184832, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:00.135205+00:00", "epoch": 0, "step": 7651, "train_loss": 3.5768165588378906, "perplexity": 35.75952112433139, "lr": 0.0026291804804649314, "grad_norm": 0.201012, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:00.440814+00:00", "epoch": 0, "step": 7652, "train_loss": 3.4831480979919434, "perplexity": 32.562069477047615, "lr": 0.0026291804804649314, "grad_norm": 0.162585, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:00.746329+00:00", "epoch": 0, "step": 7653, "train_loss": 3.6045784950256348, "perplexity": 36.766183461509264, "lr": 0.0026291804804649314, "grad_norm": 0.19002, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:01.052344+00:00", "epoch": 0, "step": 7654, "train_loss": 3.542454719543457, "perplexity": 34.55162973878038, "lr": 0.0026291804804649314, "grad_norm": 0.163817, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:01.359697+00:00", "epoch": 0, "step": 7655, "train_loss": 3.6139638423919678, "perplexity": 37.11287121229279, "lr": 0.0026291804804649314, "grad_norm": 0.16235, "tokens_per_sec": 106615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:01.664757+00:00", "epoch": 0, "step": 7656, "train_loss": 3.534943103790283, "perplexity": 34.29306351338792, "lr": 0.0026291804804649314, "grad_norm": 0.182628, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:01.970459+00:00", "epoch": 0, "step": 7657, "train_loss": 3.5789639949798584, "perplexity": 35.83639492365892, "lr": 0.0026291804804649314, "grad_norm": 0.170127, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:02.275366+00:00", "epoch": 0, "step": 7658, "train_loss": 3.5758206844329834, "perplexity": 35.72392685916252, "lr": 0.0026291804804649314, "grad_norm": 0.159887, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:02.580859+00:00", "epoch": 0, "step": 7659, "train_loss": 3.5592410564422607, "perplexity": 35.136520387520946, "lr": 0.0026291804804649314, "grad_norm": 0.15075, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:02.886866+00:00", "epoch": 0, "step": 7660, "train_loss": 3.5680577754974365, "perplexity": 35.44767889929993, "lr": 0.0026291804804649314, "grad_norm": 0.15864, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:03.191491+00:00", "epoch": 0, "step": 7661, "train_loss": 3.563741445541382, "perplexity": 35.295004754023154, "lr": 0.0026291804804649314, "grad_norm": 0.138577, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:03.496417+00:00", "epoch": 0, "step": 7662, "train_loss": 3.5616846084594727, "perplexity": 35.222483287427096, "lr": 0.0026291804804649314, "grad_norm": 0.16964, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:03.801789+00:00", "epoch": 0, "step": 7663, "train_loss": 3.5184223651885986, "perplexity": 33.73117099488637, "lr": 0.0026291804804649314, "grad_norm": 0.162489, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:04.108289+00:00", "epoch": 0, "step": 7664, "train_loss": 3.514089822769165, "perplexity": 33.58534539160659, "lr": 0.0026291804804649314, "grad_norm": 0.179288, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:04.414125+00:00", "epoch": 0, "step": 7665, "train_loss": 3.5344064235687256, "perplexity": 34.274664042221424, "lr": 0.0026291804804649314, "grad_norm": 0.168218, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:04.721316+00:00", "epoch": 0, "step": 7666, "train_loss": 3.601016044616699, "perplexity": 36.635438780197454, "lr": 0.0026291804804649314, "grad_norm": 0.152547, "tokens_per_sec": 106679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:05.027218+00:00", "epoch": 0, "step": 7667, "train_loss": 3.6025774478912354, "perplexity": 36.69268615576209, "lr": 0.0026291804804649314, "grad_norm": 0.153992, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:05.332592+00:00", "epoch": 0, "step": 7668, "train_loss": 3.584268808364868, "perplexity": 36.02700544073354, "lr": 0.0026291804804649314, "grad_norm": 0.163646, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:05.637032+00:00", "epoch": 0, "step": 7669, "train_loss": 3.5025784969329834, "perplexity": 33.20095023114817, "lr": 0.0026291804804649314, "grad_norm": 0.179421, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:05.943011+00:00", "epoch": 0, "step": 7670, "train_loss": 3.5661120414733887, "perplexity": 35.3787742012385, "lr": 0.0026291804804649314, "grad_norm": 0.157537, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:06.247514+00:00", "epoch": 0, "step": 7671, "train_loss": 3.5453944206237793, "perplexity": 34.65335064332219, "lr": 0.0026291804804649314, "grad_norm": 0.164987, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:06.552129+00:00", "epoch": 0, "step": 7672, "train_loss": 3.5413739681243896, "perplexity": 34.51430818720172, "lr": 0.0026291804804649314, "grad_norm": 0.16123, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:06.856657+00:00", "epoch": 0, "step": 7673, "train_loss": 3.554863929748535, "perplexity": 34.9830594901194, "lr": 0.0026291804804649314, "grad_norm": 0.146782, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:07.163484+00:00", "epoch": 0, "step": 7674, "train_loss": 3.5322563648223877, "perplexity": 34.20105066591947, "lr": 0.0026291804804649314, "grad_norm": 0.155903, "tokens_per_sec": 106797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:07.469198+00:00", "epoch": 0, "step": 7675, "train_loss": 3.469421148300171, "perplexity": 32.11814541974108, "lr": 0.0026291804804649314, "grad_norm": 0.171557, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:07.774956+00:00", "epoch": 0, "step": 7676, "train_loss": 3.6139509677886963, "perplexity": 37.11239340187548, "lr": 0.0026291804804649314, "grad_norm": 0.148291, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:08.080470+00:00", "epoch": 0, "step": 7677, "train_loss": 3.631488561630249, "perplexity": 37.76899627203236, "lr": 0.0026291804804649314, "grad_norm": 0.183303, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:08.385194+00:00", "epoch": 0, "step": 7678, "train_loss": 3.604919195175171, "perplexity": 36.77871183980146, "lr": 0.0026291804804649314, "grad_norm": 0.172615, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:08.689581+00:00", "epoch": 0, "step": 7679, "train_loss": 3.6136367321014404, "perplexity": 37.10073319555155, "lr": 0.0026291804804649314, "grad_norm": 0.158529, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:08.996534+00:00", "epoch": 0, "step": 7680, "train_loss": 3.6103572845458984, "perplexity": 36.979262573896975, "lr": 0.0026291804804649314, "grad_norm": 0.161073, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:09.302086+00:00", "epoch": 0, "step": 7681, "train_loss": 3.507542610168457, "perplexity": 33.36617326124503, "lr": 0.0026291804804649314, "grad_norm": 0.131063, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:09.606869+00:00", "epoch": 0, "step": 7682, "train_loss": 3.5733211040496826, "perplexity": 35.63474353924342, "lr": 0.0026291804804649314, "grad_norm": 0.165487, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:09.912082+00:00", "epoch": 0, "step": 7683, "train_loss": 3.543938159942627, "perplexity": 34.60292305807675, "lr": 0.0026291804804649314, "grad_norm": 0.158512, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:10.218149+00:00", "epoch": 0, "step": 7684, "train_loss": 3.552056074142456, "perplexity": 34.88496988558057, "lr": 0.0026291804804649314, "grad_norm": 0.162887, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:10.523610+00:00", "epoch": 0, "step": 7685, "train_loss": 3.481959342956543, "perplexity": 32.52338415123729, "lr": 0.0026291804804649314, "grad_norm": 0.173261, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:10.828993+00:00", "epoch": 0, "step": 7686, "train_loss": 3.526416540145874, "perplexity": 34.0019045813528, "lr": 0.0026291804804649314, "grad_norm": 0.156952, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:11.135711+00:00", "epoch": 0, "step": 7687, "train_loss": 3.6139702796936035, "perplexity": 37.113110119808304, "lr": 0.0026291804804649314, "grad_norm": 0.170067, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:11.439532+00:00", "epoch": 0, "step": 7688, "train_loss": 3.6138932704925537, "perplexity": 37.11025217889479, "lr": 0.0026291804804649314, "grad_norm": 0.154741, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:11.744090+00:00", "epoch": 0, "step": 7689, "train_loss": 3.508399724960327, "perplexity": 33.3947841615567, "lr": 0.0026291804804649314, "grad_norm": 0.169454, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:12.049589+00:00", "epoch": 0, "step": 7690, "train_loss": 3.5460283756256104, "perplexity": 34.675326273337014, "lr": 0.0026291804804649314, "grad_norm": 0.14693, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:12.355102+00:00", "epoch": 0, "step": 7691, "train_loss": 3.552922487258911, "perplexity": 34.9152077784185, "lr": 0.0026291804804649314, "grad_norm": 0.181446, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:12.660457+00:00", "epoch": 0, "step": 7692, "train_loss": 3.61450457572937, "perplexity": 37.13294480574547, "lr": 0.0026291804804649314, "grad_norm": 0.176402, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:12.967612+00:00", "epoch": 0, "step": 7693, "train_loss": 3.561723232269287, "perplexity": 35.22384374019555, "lr": 0.0026291804804649314, "grad_norm": 0.150292, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:13.274199+00:00", "epoch": 0, "step": 7694, "train_loss": 3.5531961917877197, "perplexity": 34.924765536852966, "lr": 0.0026291804804649314, "grad_norm": 0.161742, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:13.579907+00:00", "epoch": 0, "step": 7695, "train_loss": 3.6255619525909424, "perplexity": 37.54581620109451, "lr": 0.0026291804804649314, "grad_norm": 0.149496, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:13.886674+00:00", "epoch": 0, "step": 7696, "train_loss": 3.5846304893493652, "perplexity": 36.040038080216696, "lr": 0.0026291804804649314, "grad_norm": 0.143575, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:14.193302+00:00", "epoch": 0, "step": 7697, "train_loss": 3.552144765853882, "perplexity": 34.88806403047323, "lr": 0.0026291804804649314, "grad_norm": 0.167708, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:14.498422+00:00", "epoch": 0, "step": 7698, "train_loss": 3.6571099758148193, "perplexity": 38.74919478396542, "lr": 0.0026291804804649314, "grad_norm": 0.167553, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:14.806663+00:00", "epoch": 0, "step": 7699, "train_loss": 3.5155327320098877, "perplexity": 33.633840975774326, "lr": 0.0026291804804649314, "grad_norm": 0.169694, "tokens_per_sec": 106309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:15.112811+00:00", "epoch": 0, "step": 7700, "train_loss": 3.5939440727233887, "perplexity": 36.377267951253764, "lr": 0.0026291804804649314, "grad_norm": 0.138779, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:15.418329+00:00", "epoch": 0, "step": 7701, "train_loss": 3.637943983078003, "perplexity": 38.01359972036511, "lr": 0.0026291804804649314, "grad_norm": 0.178273, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:15.723783+00:00", "epoch": 0, "step": 7702, "train_loss": 3.7248854637145996, "perplexity": 41.46648305589224, "lr": 0.0026291804804649314, "grad_norm": 0.135634, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:16.029333+00:00", "epoch": 0, "step": 7703, "train_loss": 3.566922664642334, "perplexity": 35.40746468230819, "lr": 0.0026291804804649314, "grad_norm": 0.154728, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:16.335605+00:00", "epoch": 0, "step": 7704, "train_loss": 3.5415165424346924, "perplexity": 34.519229391697415, "lr": 0.0026291804804649314, "grad_norm": 0.150203, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:16.642157+00:00", "epoch": 0, "step": 7705, "train_loss": 3.5332648754119873, "perplexity": 34.23556018637494, "lr": 0.0026291804804649314, "grad_norm": 0.161568, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:16.948667+00:00", "epoch": 0, "step": 7706, "train_loss": 3.6621460914611816, "perplexity": 38.94483242350215, "lr": 0.0026291804804649314, "grad_norm": 0.161143, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:17.253600+00:00", "epoch": 0, "step": 7707, "train_loss": 3.5428051948547363, "perplexity": 34.563741354255406, "lr": 0.0026291804804649314, "grad_norm": 0.175419, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:17.559052+00:00", "epoch": 0, "step": 7708, "train_loss": 3.538634777069092, "perplexity": 34.41989626811005, "lr": 0.0026291804804649314, "grad_norm": 0.177147, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:17.864034+00:00", "epoch": 0, "step": 7709, "train_loss": 3.4881575107574463, "perplexity": 32.725595566433846, "lr": 0.0026291804804649314, "grad_norm": 0.149633, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:18.169128+00:00", "epoch": 0, "step": 7710, "train_loss": 3.602133274078369, "perplexity": 36.676391844469066, "lr": 0.0026291804804649314, "grad_norm": 0.146576, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:18.472988+00:00", "epoch": 0, "step": 7711, "train_loss": 3.555833578109741, "perplexity": 35.01699720759097, "lr": 0.0026291804804649314, "grad_norm": 0.15349, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:18.777137+00:00", "epoch": 0, "step": 7712, "train_loss": 3.482445001602173, "perplexity": 32.53918325010942, "lr": 0.0026291804804649314, "grad_norm": 0.156408, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:19.082483+00:00", "epoch": 0, "step": 7713, "train_loss": 3.5647690296173096, "perplexity": 35.33129197976279, "lr": 0.0026291804804649314, "grad_norm": 0.180193, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:19.388481+00:00", "epoch": 0, "step": 7714, "train_loss": 3.6101317405700684, "perplexity": 36.970923064491195, "lr": 0.0026291804804649314, "grad_norm": 0.195614, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:19.694553+00:00", "epoch": 0, "step": 7715, "train_loss": 3.554243803024292, "perplexity": 34.961372285136484, "lr": 0.0026291804804649314, "grad_norm": 0.191852, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:20.000996+00:00", "epoch": 0, "step": 7716, "train_loss": 3.5410823822021484, "perplexity": 34.504245767919635, "lr": 0.0026291804804649314, "grad_norm": 0.207881, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:20.305818+00:00", "epoch": 0, "step": 7717, "train_loss": 3.4476282596588135, "perplexity": 31.425770084795634, "lr": 0.0026291804804649314, "grad_norm": 0.199717, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:20.610895+00:00", "epoch": 0, "step": 7718, "train_loss": 3.560220718383789, "perplexity": 35.17095916572815, "lr": 0.0026291804804649314, "grad_norm": 0.171533, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:20.915763+00:00", "epoch": 0, "step": 7719, "train_loss": 3.6303417682647705, "perplexity": 37.725707863860634, "lr": 0.0026291804804649314, "grad_norm": 0.185396, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:21.222412+00:00", "epoch": 0, "step": 7720, "train_loss": 3.5380327701568604, "perplexity": 34.399181488484224, "lr": 0.0026291804804649314, "grad_norm": 0.172409, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:21.527773+00:00", "epoch": 0, "step": 7721, "train_loss": 3.580963134765625, "perplexity": 35.9081085454225, "lr": 0.0026291804804649314, "grad_norm": 0.199891, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:21.833642+00:00", "epoch": 0, "step": 7722, "train_loss": 3.6831634044647217, "perplexity": 39.772010235202, "lr": 0.0026291804804649314, "grad_norm": 0.207818, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:22.139052+00:00", "epoch": 0, "step": 7723, "train_loss": 3.6168935298919678, "perplexity": 37.221759753989154, "lr": 0.0026291804804649314, "grad_norm": 0.156809, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:22.444291+00:00", "epoch": 0, "step": 7724, "train_loss": 3.606346845626831, "perplexity": 36.83125648313828, "lr": 0.0026291804804649314, "grad_norm": 0.196697, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:22.750307+00:00", "epoch": 0, "step": 7725, "train_loss": 3.6119399070739746, "perplexity": 37.03783312322838, "lr": 0.0026291804804649314, "grad_norm": 0.167546, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:23.056322+00:00", "epoch": 0, "step": 7726, "train_loss": 3.5595173835754395, "perplexity": 35.146230903047176, "lr": 0.0026291804804649314, "grad_norm": 0.218197, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:23.361026+00:00", "epoch": 0, "step": 7727, "train_loss": 3.4843597412109375, "perplexity": 32.60154699923225, "lr": 0.0026291804804649314, "grad_norm": 0.157815, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:23.666253+00:00", "epoch": 0, "step": 7728, "train_loss": 3.561332941055298, "perplexity": 35.21009886588714, "lr": 0.0026291804804649314, "grad_norm": 0.181762, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:23.970441+00:00", "epoch": 0, "step": 7729, "train_loss": 3.5058629512786865, "perplexity": 33.31017651259258, "lr": 0.0026291804804649314, "grad_norm": 0.175814, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:24.274643+00:00", "epoch": 0, "step": 7730, "train_loss": 3.532898187637329, "perplexity": 34.22300872637013, "lr": 0.0026291804804649314, "grad_norm": 0.154785, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:24.579534+00:00", "epoch": 0, "step": 7731, "train_loss": 3.592158317565918, "perplexity": 36.31236502501028, "lr": 0.0026291804804649314, "grad_norm": 0.196212, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:24.884308+00:00", "epoch": 0, "step": 7732, "train_loss": 3.6114869117736816, "perplexity": 37.02105895848627, "lr": 0.0026291804804649314, "grad_norm": 0.187739, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:25.190048+00:00", "epoch": 0, "step": 7733, "train_loss": 3.515787363052368, "perplexity": 33.6424062862153, "lr": 0.0026291804804649314, "grad_norm": 0.174812, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:25.495134+00:00", "epoch": 0, "step": 7734, "train_loss": 3.5586533546447754, "perplexity": 35.11587665810445, "lr": 0.0026291804804649314, "grad_norm": 0.166988, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:25.800585+00:00", "epoch": 0, "step": 7735, "train_loss": 3.5405564308166504, "perplexity": 34.48610298358194, "lr": 0.0026291804804649314, "grad_norm": 0.170506, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:26.106503+00:00", "epoch": 0, "step": 7736, "train_loss": 3.511960029602051, "perplexity": 33.513891670309675, "lr": 0.0026291804804649314, "grad_norm": 0.174628, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:26.411113+00:00", "epoch": 0, "step": 7737, "train_loss": 3.523135185241699, "perplexity": 33.89051511912943, "lr": 0.0026291804804649314, "grad_norm": 0.162999, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:26.715898+00:00", "epoch": 0, "step": 7738, "train_loss": 3.5591163635253906, "perplexity": 35.1321393854508, "lr": 0.0026291804804649314, "grad_norm": 0.156109, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:27.021000+00:00", "epoch": 0, "step": 7739, "train_loss": 3.4865164756774902, "perplexity": 32.671935756953886, "lr": 0.0026291804804649314, "grad_norm": 0.152363, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:27.325945+00:00", "epoch": 0, "step": 7740, "train_loss": 3.5355889797210693, "perplexity": 34.315219732018086, "lr": 0.0026291804804649314, "grad_norm": 0.172365, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:27.630234+00:00", "epoch": 0, "step": 7741, "train_loss": 3.551736354827881, "perplexity": 34.87381826970826, "lr": 0.0026291804804649314, "grad_norm": 0.1651, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:27.935449+00:00", "epoch": 0, "step": 7742, "train_loss": 3.530775547027588, "perplexity": 34.150442621386844, "lr": 0.0026291804804649314, "grad_norm": 0.159343, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:28.241941+00:00", "epoch": 0, "step": 7743, "train_loss": 3.5419881343841553, "perplexity": 34.535512221504156, "lr": 0.0026291804804649314, "grad_norm": 0.14809, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:28.547937+00:00", "epoch": 0, "step": 7744, "train_loss": 3.520465850830078, "perplexity": 33.80017063436006, "lr": 0.0026291804804649314, "grad_norm": 0.179733, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:28.853316+00:00", "epoch": 0, "step": 7745, "train_loss": 3.4510159492492676, "perplexity": 31.53241137078261, "lr": 0.0026291804804649314, "grad_norm": 0.165881, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:29.159152+00:00", "epoch": 0, "step": 7746, "train_loss": 3.537623405456543, "perplexity": 34.38510255976786, "lr": 0.0026291804804649314, "grad_norm": 0.153241, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:29.465258+00:00", "epoch": 0, "step": 7747, "train_loss": 3.481700897216797, "perplexity": 32.514979707254426, "lr": 0.0026291804804649314, "grad_norm": 0.182492, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:29.770285+00:00", "epoch": 0, "step": 7748, "train_loss": 3.552802801132202, "perplexity": 34.91102916250243, "lr": 0.0026291804804649314, "grad_norm": 0.157571, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:30.074820+00:00", "epoch": 0, "step": 7749, "train_loss": 3.5292716026306152, "perplexity": 34.09912085676595, "lr": 0.0026291804804649314, "grad_norm": 0.169979, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:30.379744+00:00", "epoch": 0, "step": 7750, "train_loss": 3.5307042598724365, "perplexity": 34.148008220257, "lr": 0.0026291804804649314, "grad_norm": 0.17809, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:30.684675+00:00", "epoch": 0, "step": 7751, "train_loss": 3.4251763820648193, "perplexity": 30.72806425296709, "lr": 0.0026291804804649314, "grad_norm": 0.171089, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:30.991445+00:00", "epoch": 0, "step": 7752, "train_loss": 3.509122848510742, "perplexity": 33.41894144974428, "lr": 0.0026291804804649314, "grad_norm": 0.156133, "tokens_per_sec": 106817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:31.298295+00:00", "epoch": 0, "step": 7753, "train_loss": 3.4875669479370117, "perplexity": 32.70627475205458, "lr": 0.0026291804804649314, "grad_norm": 0.141079, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:31.602692+00:00", "epoch": 0, "step": 7754, "train_loss": 3.4266445636749268, "perplexity": 30.77321176608191, "lr": 0.0026291804804649314, "grad_norm": 0.156929, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:31.908259+00:00", "epoch": 0, "step": 7755, "train_loss": 3.5157532691955566, "perplexity": 33.64125930638517, "lr": 0.0026291804804649314, "grad_norm": 0.164903, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:32.214778+00:00", "epoch": 0, "step": 7756, "train_loss": 3.651289463043213, "perplexity": 38.5243097089561, "lr": 0.0026291804804649314, "grad_norm": 0.155985, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:32.520371+00:00", "epoch": 0, "step": 7757, "train_loss": 3.5969033241271973, "perplexity": 36.48507687067824, "lr": 0.0026291804804649314, "grad_norm": 0.140485, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:32.826566+00:00", "epoch": 0, "step": 7758, "train_loss": 3.5714499950408936, "perplexity": 35.568129390183415, "lr": 0.0026291804804649314, "grad_norm": 0.159441, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:33.132885+00:00", "epoch": 0, "step": 7759, "train_loss": 3.534548759460449, "perplexity": 34.279542904301756, "lr": 0.0026291804804649314, "grad_norm": 0.157036, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:33.438405+00:00", "epoch": 0, "step": 7760, "train_loss": 3.5245745182037354, "perplexity": 33.939329976630106, "lr": 0.0026291804804649314, "grad_norm": 0.158944, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:33.743342+00:00", "epoch": 0, "step": 7761, "train_loss": 3.559051513671875, "perplexity": 35.129861145230535, "lr": 0.0026291804804649314, "grad_norm": 0.146635, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:34.048952+00:00", "epoch": 0, "step": 7762, "train_loss": 3.45542573928833, "perplexity": 31.6717697290702, "lr": 0.0026291804804649314, "grad_norm": 0.162935, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:34.356869+00:00", "epoch": 0, "step": 7763, "train_loss": 3.53867769241333, "perplexity": 34.42137344150352, "lr": 0.0026291804804649314, "grad_norm": 0.165776, "tokens_per_sec": 106419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:34.663572+00:00", "epoch": 0, "step": 7764, "train_loss": 3.58711576461792, "perplexity": 36.12971889012454, "lr": 0.0026291804804649314, "grad_norm": 0.167942, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:34.968726+00:00", "epoch": 0, "step": 7765, "train_loss": 3.5613319873809814, "perplexity": 35.210065286936185, "lr": 0.0026291804804649314, "grad_norm": 0.15297, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:35.273628+00:00", "epoch": 0, "step": 7766, "train_loss": 3.6102042198181152, "perplexity": 36.97360278630545, "lr": 0.0026291804804649314, "grad_norm": 0.170196, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:35.577319+00:00", "epoch": 0, "step": 7767, "train_loss": 3.5049147605895996, "perplexity": 33.27860708265651, "lr": 0.0026291804804649314, "grad_norm": 0.165194, "tokens_per_sec": 107899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:35.883038+00:00", "epoch": 0, "step": 7768, "train_loss": 3.580843687057495, "perplexity": 35.90381966030731, "lr": 0.0026291804804649314, "grad_norm": 0.163628, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:36.189181+00:00", "epoch": 0, "step": 7769, "train_loss": 3.5042402744293213, "perplexity": 33.25616869078896, "lr": 0.0026291804804649314, "grad_norm": 0.169561, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:36.494549+00:00", "epoch": 0, "step": 7770, "train_loss": 3.621412515640259, "perplexity": 37.39034498608126, "lr": 0.0026291804804649314, "grad_norm": 0.173403, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:36.799017+00:00", "epoch": 0, "step": 7771, "train_loss": 3.628082752227783, "perplexity": 37.64058107238429, "lr": 0.0026291804804649314, "grad_norm": 0.155941, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:37.104277+00:00", "epoch": 0, "step": 7772, "train_loss": 3.5781142711639404, "perplexity": 35.805956819238205, "lr": 0.0026291804804649314, "grad_norm": 0.171445, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:37.408946+00:00", "epoch": 0, "step": 7773, "train_loss": 3.5746023654937744, "perplexity": 35.68043022425431, "lr": 0.0026291804804649314, "grad_norm": 0.147732, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:37.714078+00:00", "epoch": 0, "step": 7774, "train_loss": 3.6063272953033447, "perplexity": 36.83053642719832, "lr": 0.0026291804804649314, "grad_norm": 0.175187, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:38.019602+00:00", "epoch": 0, "step": 7775, "train_loss": 3.5811517238616943, "perplexity": 35.91488106174593, "lr": 0.0026291804804649314, "grad_norm": 0.145141, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:38.323889+00:00", "epoch": 0, "step": 7776, "train_loss": 3.5298519134521484, "perplexity": 34.11891468833575, "lr": 0.0026291804804649314, "grad_norm": 0.163241, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:38.629125+00:00", "epoch": 0, "step": 7777, "train_loss": 3.5391714572906494, "perplexity": 34.43837370345876, "lr": 0.0026291804804649314, "grad_norm": 0.169678, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:38.933356+00:00", "epoch": 0, "step": 7778, "train_loss": 3.551771402359009, "perplexity": 34.875040532358135, "lr": 0.0026291804804649314, "grad_norm": 0.162895, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:39.239438+00:00", "epoch": 0, "step": 7779, "train_loss": 3.6091370582580566, "perplexity": 36.934167024582095, "lr": 0.0026291804804649314, "grad_norm": 0.177647, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:39.544648+00:00", "epoch": 0, "step": 7780, "train_loss": 3.634394645690918, "perplexity": 37.87891579040731, "lr": 0.0026291804804649314, "grad_norm": 0.17438, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:39.849438+00:00", "epoch": 0, "step": 7781, "train_loss": 3.5128092765808105, "perplexity": 33.54236533043077, "lr": 0.0026291804804649314, "grad_norm": 0.16166, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:40.154665+00:00", "epoch": 0, "step": 7782, "train_loss": 3.570167064666748, "perplexity": 35.52252721509974, "lr": 0.0026291804804649314, "grad_norm": 0.150554, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:40.460710+00:00", "epoch": 0, "step": 7783, "train_loss": 3.5935490131378174, "perplexity": 36.3628996012168, "lr": 0.0026291804804649314, "grad_norm": 0.167362, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:40.766711+00:00", "epoch": 0, "step": 7784, "train_loss": 3.4691503047943115, "perplexity": 32.109447606560174, "lr": 0.0026291804804649314, "grad_norm": 0.151725, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:41.071815+00:00", "epoch": 0, "step": 7785, "train_loss": 3.593904733657837, "perplexity": 36.37583693167291, "lr": 0.0026291804804649314, "grad_norm": 0.146378, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:41.376973+00:00", "epoch": 0, "step": 7786, "train_loss": 3.564846992492676, "perplexity": 35.33404661625423, "lr": 0.0026291804804649314, "grad_norm": 0.168902, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:41.682124+00:00", "epoch": 0, "step": 7787, "train_loss": 3.5541775226593018, "perplexity": 34.959055109413335, "lr": 0.0026291804804649314, "grad_norm": 0.168758, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:41.987186+00:00", "epoch": 0, "step": 7788, "train_loss": 3.61258864402771, "perplexity": 37.061868729811835, "lr": 0.0026291804804649314, "grad_norm": 0.142719, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:42.292046+00:00", "epoch": 0, "step": 7789, "train_loss": 3.5923614501953125, "perplexity": 36.319742000424185, "lr": 0.0026291804804649314, "grad_norm": 0.178414, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:42.596641+00:00", "epoch": 0, "step": 7790, "train_loss": 3.706766128540039, "perplexity": 40.721903960956, "lr": 0.0026291804804649314, "grad_norm": 0.185972, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:42.900802+00:00", "epoch": 0, "step": 7791, "train_loss": 3.5112271308898926, "perplexity": 33.48933838090124, "lr": 0.0026291804804649314, "grad_norm": 0.14941, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:43.207357+00:00", "epoch": 0, "step": 7792, "train_loss": 3.4989917278289795, "perplexity": 33.082079397191364, "lr": 0.0026291804804649314, "grad_norm": 0.151534, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:43.512489+00:00", "epoch": 0, "step": 7793, "train_loss": 3.528327226638794, "perplexity": 34.06693366648365, "lr": 0.0026291804804649314, "grad_norm": 0.151032, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:43.817784+00:00", "epoch": 0, "step": 7794, "train_loss": 3.597532033920288, "perplexity": 36.50802260815762, "lr": 0.0026291804804649314, "grad_norm": 0.162367, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:44.123441+00:00", "epoch": 0, "step": 7795, "train_loss": 3.5334694385528564, "perplexity": 34.24256423645689, "lr": 0.0026291804804649314, "grad_norm": 0.150498, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:44.427988+00:00", "epoch": 0, "step": 7796, "train_loss": 3.5379140377044678, "perplexity": 34.395097431765635, "lr": 0.0026291804804649314, "grad_norm": 0.158226, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:44.732968+00:00", "epoch": 0, "step": 7797, "train_loss": 3.4936039447784424, "perplexity": 32.90431962608331, "lr": 0.0026291804804649314, "grad_norm": 0.164599, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:45.037590+00:00", "epoch": 0, "step": 7798, "train_loss": 3.61460542678833, "perplexity": 37.13668989139618, "lr": 0.0026291804804649314, "grad_norm": 0.162424, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:45.341900+00:00", "epoch": 0, "step": 7799, "train_loss": 3.5977747440338135, "perplexity": 36.51688454986736, "lr": 0.0026291804804649314, "grad_norm": 0.162593, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:45.646691+00:00", "epoch": 0, "step": 7800, "train_loss": 3.565255641937256, "perplexity": 35.34848880547405, "lr": 0.0026291804804649314, "grad_norm": 0.154478, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:45.951311+00:00", "epoch": 0, "step": 7801, "train_loss": 3.500962495803833, "perplexity": 33.147340786214116, "lr": 0.0026291804804649314, "grad_norm": 0.150272, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:46.257364+00:00", "epoch": 0, "step": 7802, "train_loss": 3.5124502182006836, "perplexity": 33.530323825000686, "lr": 0.0026291804804649314, "grad_norm": 0.182753, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:46.561572+00:00", "epoch": 0, "step": 7803, "train_loss": 3.6665096282958984, "perplexity": 39.115140937812505, "lr": 0.0026291804804649314, "grad_norm": 0.167986, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:46.866460+00:00", "epoch": 0, "step": 7804, "train_loss": 3.5049824714660645, "perplexity": 33.280860482598584, "lr": 0.0026291804804649314, "grad_norm": 0.172412, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:47.170680+00:00", "epoch": 0, "step": 7805, "train_loss": 3.54883074760437, "perplexity": 34.7726357209912, "lr": 0.0026291804804649314, "grad_norm": 0.163005, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:47.474797+00:00", "epoch": 0, "step": 7806, "train_loss": 3.526258707046509, "perplexity": 33.99653837886173, "lr": 0.0026291804804649314, "grad_norm": 0.194234, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:47.780165+00:00", "epoch": 0, "step": 7807, "train_loss": 3.5673165321350098, "perplexity": 35.421413278413574, "lr": 0.0026291804804649314, "grad_norm": 0.192475, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:48.086793+00:00", "epoch": 0, "step": 7808, "train_loss": 3.51163387298584, "perplexity": 33.502962675185344, "lr": 0.0026291804804649314, "grad_norm": 0.172332, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:48.391948+00:00", "epoch": 0, "step": 7809, "train_loss": 3.5998966693878174, "perplexity": 36.59445292108501, "lr": 0.0026291804804649314, "grad_norm": 0.176073, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:48.697305+00:00", "epoch": 0, "step": 7810, "train_loss": 3.6361513137817383, "perplexity": 37.94551495227602, "lr": 0.0026291804804649314, "grad_norm": 0.188446, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:49.001717+00:00", "epoch": 0, "step": 7811, "train_loss": 3.498305082321167, "perplexity": 33.05937153300285, "lr": 0.0026291804804649314, "grad_norm": 0.152391, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:49.306858+00:00", "epoch": 0, "step": 7812, "train_loss": 3.555560350418091, "perplexity": 35.00743090122482, "lr": 0.0026291804804649314, "grad_norm": 0.173361, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:49.612210+00:00", "epoch": 0, "step": 7813, "train_loss": 3.6563823223114014, "perplexity": 38.72100905259298, "lr": 0.0026291804804649314, "grad_norm": 0.162035, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:49.916657+00:00", "epoch": 0, "step": 7814, "train_loss": 3.5822813510894775, "perplexity": 35.95547441263647, "lr": 0.0026291804804649314, "grad_norm": 0.16776, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:50.221366+00:00", "epoch": 0, "step": 7815, "train_loss": 3.647144079208374, "perplexity": 38.3649422061921, "lr": 0.0026291804804649314, "grad_norm": 0.175604, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:50.527114+00:00", "epoch": 0, "step": 7816, "train_loss": 3.518725872039795, "perplexity": 33.74141019013655, "lr": 0.0026291804804649314, "grad_norm": 0.179586, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:50.832241+00:00", "epoch": 0, "step": 7817, "train_loss": 3.5181713104248047, "perplexity": 33.722703686643264, "lr": 0.0026291804804649314, "grad_norm": 0.194312, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:51.137110+00:00", "epoch": 0, "step": 7818, "train_loss": 3.516859769821167, "perplexity": 33.6785039826828, "lr": 0.0026291804804649314, "grad_norm": 0.15316, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:51.442125+00:00", "epoch": 0, "step": 7819, "train_loss": 3.5315005779266357, "perplexity": 34.17521172560143, "lr": 0.0026291804804649314, "grad_norm": 0.212068, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:51.746377+00:00", "epoch": 0, "step": 7820, "train_loss": 3.5002329349517822, "perplexity": 33.12316660336787, "lr": 0.0026291804804649314, "grad_norm": 0.136344, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:52.050378+00:00", "epoch": 0, "step": 7821, "train_loss": 3.5349676609039307, "perplexity": 34.293905662386265, "lr": 0.0026291804804649314, "grad_norm": 0.179218, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:52.354539+00:00", "epoch": 0, "step": 7822, "train_loss": 3.5712978839874268, "perplexity": 35.562719496014815, "lr": 0.0026291804804649314, "grad_norm": 0.146294, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:52.659850+00:00", "epoch": 0, "step": 7823, "train_loss": 3.469352960586548, "perplexity": 32.11595543150509, "lr": 0.0026291804804649314, "grad_norm": 0.16261, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:52.965758+00:00", "epoch": 0, "step": 7824, "train_loss": 3.459949493408203, "perplexity": 31.815369588203186, "lr": 0.0026291804804649314, "grad_norm": 0.142399, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:53.269798+00:00", "epoch": 0, "step": 7825, "train_loss": 3.673119068145752, "perplexity": 39.37452636097193, "lr": 0.0026291804804649314, "grad_norm": 0.164949, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:53.573898+00:00", "epoch": 0, "step": 7826, "train_loss": 3.6255767345428467, "perplexity": 37.54637120564582, "lr": 0.0026291804804649314, "grad_norm": 0.167034, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:53.877829+00:00", "epoch": 0, "step": 7827, "train_loss": 3.604623556137085, "perplexity": 36.767840223927315, "lr": 0.0026291804804649314, "grad_norm": 0.170784, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:54.182925+00:00", "epoch": 0, "step": 7828, "train_loss": 3.6213951110839844, "perplexity": 37.389694229380915, "lr": 0.0026291804804649314, "grad_norm": 0.148575, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:54.488149+00:00", "epoch": 0, "step": 7829, "train_loss": 3.5488250255584717, "perplexity": 34.77243675094285, "lr": 0.0026291804804649314, "grad_norm": 0.169781, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:54.792744+00:00", "epoch": 0, "step": 7830, "train_loss": 3.5800676345825195, "perplexity": 35.87596722107392, "lr": 0.0026291804804649314, "grad_norm": 0.148864, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:55.096874+00:00", "epoch": 0, "step": 7831, "train_loss": 3.556584596633911, "perplexity": 35.043305498925186, "lr": 0.0026291804804649314, "grad_norm": 0.140245, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:55.402050+00:00", "epoch": 0, "step": 7832, "train_loss": 3.5718507766723633, "perplexity": 35.582387300072085, "lr": 0.0026291804804649314, "grad_norm": 0.145836, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:55.707025+00:00", "epoch": 0, "step": 7833, "train_loss": 3.5194976329803467, "perplexity": 33.76746054363009, "lr": 0.0026291804804649314, "grad_norm": 0.162207, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:56.012473+00:00", "epoch": 0, "step": 7834, "train_loss": 3.5990450382232666, "perplexity": 36.56330121129397, "lr": 0.0026291804804649314, "grad_norm": 0.141186, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:56.317347+00:00", "epoch": 0, "step": 7835, "train_loss": 3.494142770767212, "perplexity": 32.92205410611091, "lr": 0.0026291804804649314, "grad_norm": 0.174326, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:56.621592+00:00", "epoch": 0, "step": 7836, "train_loss": 3.6680617332458496, "perplexity": 39.17589888083487, "lr": 0.0026291804804649314, "grad_norm": 0.155355, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:56.926322+00:00", "epoch": 0, "step": 7837, "train_loss": 3.4810683727264404, "perplexity": 32.49441968932906, "lr": 0.0026291804804649314, "grad_norm": 0.166919, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:57.231140+00:00", "epoch": 0, "step": 7838, "train_loss": 3.536823272705078, "perplexity": 34.3576009169971, "lr": 0.0026291804804649314, "grad_norm": 0.158581, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:57.536439+00:00", "epoch": 0, "step": 7839, "train_loss": 3.556466817855835, "perplexity": 35.0391783842718, "lr": 0.0026291804804649314, "grad_norm": 0.164689, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:57.842380+00:00", "epoch": 0, "step": 7840, "train_loss": 3.5900089740753174, "perplexity": 36.23440109526193, "lr": 0.0026291804804649314, "grad_norm": 0.144856, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:58.146415+00:00", "epoch": 0, "step": 7841, "train_loss": 3.474884510040283, "perplexity": 32.29409867707885, "lr": 0.0026291804804649314, "grad_norm": 0.135254, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:58.450489+00:00", "epoch": 0, "step": 7842, "train_loss": 3.53570294380188, "perplexity": 34.31913065734093, "lr": 0.0026291804804649314, "grad_norm": 0.146065, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:58.754402+00:00", "epoch": 0, "step": 7843, "train_loss": 3.6680219173431396, "perplexity": 39.17433908810894, "lr": 0.0026291804804649314, "grad_norm": 0.152607, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:59.059820+00:00", "epoch": 0, "step": 7844, "train_loss": 3.5857088565826416, "perplexity": 36.07892343894856, "lr": 0.0026291804804649314, "grad_norm": 0.181849, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:59.364622+00:00", "epoch": 0, "step": 7845, "train_loss": 3.5934319496154785, "perplexity": 36.358643081253504, "lr": 0.0026291804804649314, "grad_norm": 0.178492, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:59.668391+00:00", "epoch": 0, "step": 7846, "train_loss": 3.3973193168640137, "perplexity": 29.88388335551836, "lr": 0.0026291804804649314, "grad_norm": 0.168767, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:38:59.972163+00:00", "epoch": 0, "step": 7847, "train_loss": 3.5530548095703125, "perplexity": 34.91982814509686, "lr": 0.0026291804804649314, "grad_norm": 0.169723, "tokens_per_sec": 107870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:00.276639+00:00", "epoch": 0, "step": 7848, "train_loss": 3.46737003326416, "perplexity": 32.05233492426273, "lr": 0.0026291804804649314, "grad_norm": 0.15641, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:00.581413+00:00", "epoch": 0, "step": 7849, "train_loss": 3.528236150741577, "perplexity": 34.063831131219786, "lr": 0.0026291804804649314, "grad_norm": 0.163257, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:00.886602+00:00", "epoch": 0, "step": 7850, "train_loss": 3.573260545730591, "perplexity": 35.63258562441393, "lr": 0.0026291804804649314, "grad_norm": 0.172223, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:01.191735+00:00", "epoch": 0, "step": 7851, "train_loss": 3.6045641899108887, "perplexity": 36.7656575207979, "lr": 0.0026291804804649314, "grad_norm": 0.151419, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:01.495546+00:00", "epoch": 0, "step": 7852, "train_loss": 3.5375266075134277, "perplexity": 34.381774313652556, "lr": 0.0026291804804649314, "grad_norm": 0.152137, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:01.799895+00:00", "epoch": 0, "step": 7853, "train_loss": 3.4770846366882324, "perplexity": 32.36522800221952, "lr": 0.0026291804804649314, "grad_norm": 0.153456, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:02.104516+00:00", "epoch": 0, "step": 7854, "train_loss": 3.5525550842285156, "perplexity": 34.90238218149968, "lr": 0.0026291804804649314, "grad_norm": 0.157479, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:02.408823+00:00", "epoch": 0, "step": 7855, "train_loss": 3.4972987174987793, "perplexity": 33.02611847959986, "lr": 0.0026291804804649314, "grad_norm": 0.16485, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:02.714627+00:00", "epoch": 0, "step": 7856, "train_loss": 3.543645143508911, "perplexity": 34.59278531830084, "lr": 0.0026291804804649314, "grad_norm": 0.149552, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:03.019272+00:00", "epoch": 0, "step": 7857, "train_loss": 3.5037827491760254, "perplexity": 33.24095663400049, "lr": 0.0026291804804649314, "grad_norm": 0.167191, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:03.324109+00:00", "epoch": 0, "step": 7858, "train_loss": 3.5003607273101807, "perplexity": 33.1273997614226, "lr": 0.0026291804804649314, "grad_norm": 0.164551, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:03.628099+00:00", "epoch": 0, "step": 7859, "train_loss": 3.5004842281341553, "perplexity": 33.13149127523674, "lr": 0.0026291804804649314, "grad_norm": 0.190644, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:03.932925+00:00", "epoch": 0, "step": 7860, "train_loss": 3.542865037918091, "perplexity": 34.56580981630997, "lr": 0.0026291804804649314, "grad_norm": 0.160668, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:04.238744+00:00", "epoch": 0, "step": 7861, "train_loss": 3.582754611968994, "perplexity": 35.97249475929487, "lr": 0.0026291804804649314, "grad_norm": 0.151675, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:04.543011+00:00", "epoch": 0, "step": 7862, "train_loss": 3.5065507888793945, "perplexity": 33.33309638615592, "lr": 0.0026291804804649314, "grad_norm": 0.161567, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:04.848623+00:00", "epoch": 0, "step": 7863, "train_loss": 3.5455596446990967, "perplexity": 34.65907668416576, "lr": 0.0026291804804649314, "grad_norm": 0.136255, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:05.152686+00:00", "epoch": 0, "step": 7864, "train_loss": 3.5114049911499023, "perplexity": 33.49529533307005, "lr": 0.0026291804804649314, "grad_norm": 0.170735, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:05.457700+00:00", "epoch": 0, "step": 7865, "train_loss": 3.517223358154297, "perplexity": 33.690751320167756, "lr": 0.0026291804804649314, "grad_norm": 0.181808, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:05.762934+00:00", "epoch": 0, "step": 7866, "train_loss": 3.651195526123047, "perplexity": 38.52069102391725, "lr": 0.0026291804804649314, "grad_norm": 0.15121, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:06.068708+00:00", "epoch": 0, "step": 7867, "train_loss": 3.543292760848999, "perplexity": 34.58059756809864, "lr": 0.0026291804804649314, "grad_norm": 0.14453, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:06.371490+00:00", "epoch": 0, "step": 7868, "train_loss": 3.6219611167907715, "perplexity": 37.4108629999471, "lr": 0.0026291804804649314, "grad_norm": 0.157936, "tokens_per_sec": 108223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:06.676031+00:00", "epoch": 0, "step": 7869, "train_loss": 3.5073182582855225, "perplexity": 33.35868833710735, "lr": 0.0026291804804649314, "grad_norm": 0.176694, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:06.981087+00:00", "epoch": 0, "step": 7870, "train_loss": 3.6393308639526367, "perplexity": 38.066356630111606, "lr": 0.0026291804804649314, "grad_norm": 0.165889, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:07.286694+00:00", "epoch": 0, "step": 7871, "train_loss": 3.6357674598693848, "perplexity": 37.93095221306634, "lr": 0.0026291804804649314, "grad_norm": 0.181644, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:07.591902+00:00", "epoch": 0, "step": 7872, "train_loss": 3.661770820617676, "perplexity": 38.930220305310826, "lr": 0.0026291804804649314, "grad_norm": 0.163047, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:07.897316+00:00", "epoch": 0, "step": 7873, "train_loss": 3.611384630203247, "perplexity": 37.01727258007851, "lr": 0.0026291804804649314, "grad_norm": 0.168775, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:08.202108+00:00", "epoch": 0, "step": 7874, "train_loss": 3.5117194652557373, "perplexity": 33.50583039253447, "lr": 0.0026291804804649314, "grad_norm": 0.158315, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:08.506664+00:00", "epoch": 0, "step": 7875, "train_loss": 3.6703431606292725, "perplexity": 39.265377880420914, "lr": 0.0026291804804649314, "grad_norm": 0.196747, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:08.811533+00:00", "epoch": 0, "step": 7876, "train_loss": 3.552093029022217, "perplexity": 34.886259079269, "lr": 0.0026291804804649314, "grad_norm": 0.209005, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:09.117561+00:00", "epoch": 0, "step": 7877, "train_loss": 3.691411256790161, "perplexity": 40.10140041580612, "lr": 0.0026291804804649314, "grad_norm": 0.20523, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:09.421736+00:00", "epoch": 0, "step": 7878, "train_loss": 3.593879461288452, "perplexity": 36.37491763970168, "lr": 0.0026291804804649314, "grad_norm": 0.172027, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:09.725913+00:00", "epoch": 0, "step": 7879, "train_loss": 3.5073602199554443, "perplexity": 33.360088152745476, "lr": 0.0026291804804649314, "grad_norm": 0.184262, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:10.030457+00:00", "epoch": 0, "step": 7880, "train_loss": 3.510490894317627, "perplexity": 33.46469137932953, "lr": 0.0026291804804649314, "grad_norm": 0.192074, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:10.334920+00:00", "epoch": 0, "step": 7881, "train_loss": 3.5112531185150146, "perplexity": 33.49020870058139, "lr": 0.0026291804804649314, "grad_norm": 0.19346, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:10.640218+00:00", "epoch": 0, "step": 7882, "train_loss": 3.4553821086883545, "perplexity": 31.670387900899843, "lr": 0.0026291804804649314, "grad_norm": 0.162885, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:10.945085+00:00", "epoch": 0, "step": 7883, "train_loss": 3.6540329456329346, "perplexity": 38.63014559499616, "lr": 0.0026291804804649314, "grad_norm": 0.182466, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:11.249196+00:00", "epoch": 0, "step": 7884, "train_loss": 3.459869861602783, "perplexity": 31.812836173754295, "lr": 0.0026291804804649314, "grad_norm": 0.172439, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:11.554278+00:00", "epoch": 0, "step": 7885, "train_loss": 3.576221466064453, "perplexity": 35.73824722232921, "lr": 0.0026291804804649314, "grad_norm": 0.159706, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:11.858449+00:00", "epoch": 0, "step": 7886, "train_loss": 3.5537304878234863, "perplexity": 34.943430686543095, "lr": 0.0026291804804649314, "grad_norm": 0.187229, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:12.164720+00:00", "epoch": 0, "step": 7887, "train_loss": 3.728703260421753, "perplexity": 41.62509624215392, "lr": 0.0026291804804649314, "grad_norm": 0.165612, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:12.470222+00:00", "epoch": 0, "step": 7888, "train_loss": 3.666903018951416, "perplexity": 39.130531495799055, "lr": 0.0026291804804649314, "grad_norm": 0.160205, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:12.773652+00:00", "epoch": 0, "step": 7889, "train_loss": 3.559373140335083, "perplexity": 35.14116166242604, "lr": 0.0026291804804649314, "grad_norm": 0.161542, "tokens_per_sec": 107992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:13.077996+00:00", "epoch": 0, "step": 7890, "train_loss": 3.5293962955474854, "perplexity": 34.10337304071139, "lr": 0.0026291804804649314, "grad_norm": 0.146196, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:13.382146+00:00", "epoch": 0, "step": 7891, "train_loss": 3.474365472793579, "perplexity": 32.27734118627386, "lr": 0.0026291804804649314, "grad_norm": 0.174518, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:13.686753+00:00", "epoch": 0, "step": 7892, "train_loss": 3.5935652256011963, "perplexity": 36.36348913817385, "lr": 0.0026291804804649314, "grad_norm": 0.154899, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:13.991581+00:00", "epoch": 0, "step": 7893, "train_loss": 3.6686317920684814, "perplexity": 39.19823781427452, "lr": 0.0026291804804649314, "grad_norm": 0.164115, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:14.295664+00:00", "epoch": 0, "step": 7894, "train_loss": 3.5463171005249023, "perplexity": 34.68533934886575, "lr": 0.0026291804804649314, "grad_norm": 0.144948, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:14.599300+00:00", "epoch": 0, "step": 7895, "train_loss": 3.5001754760742188, "perplexity": 33.12126343807088, "lr": 0.0026291804804649314, "grad_norm": 0.155869, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:14.903627+00:00", "epoch": 0, "step": 7896, "train_loss": 3.628222942352295, "perplexity": 37.645858280029074, "lr": 0.0026291804804649314, "grad_norm": 0.167429, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:15.208315+00:00", "epoch": 0, "step": 7897, "train_loss": 3.631298065185547, "perplexity": 37.76180209777664, "lr": 0.0026291804804649314, "grad_norm": 0.167663, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:15.514683+00:00", "epoch": 0, "step": 7898, "train_loss": 3.555657386779785, "perplexity": 35.01082805977303, "lr": 0.0026291804804649314, "grad_norm": 0.173201, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:15.819987+00:00", "epoch": 0, "step": 7899, "train_loss": 3.539283514022827, "perplexity": 34.44223297130192, "lr": 0.0026291804804649314, "grad_norm": 0.175426, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:16.124733+00:00", "epoch": 0, "step": 7900, "train_loss": 3.5989105701446533, "perplexity": 36.558384944980176, "lr": 0.0026291804804649314, "grad_norm": 0.184631, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:16.428330+00:00", "epoch": 0, "step": 7901, "train_loss": 3.496213674545288, "perplexity": 32.990303156558184, "lr": 0.0026291804804649314, "grad_norm": 0.159963, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:16.733111+00:00", "epoch": 0, "step": 7902, "train_loss": 3.5052907466888428, "perplexity": 33.291121728840835, "lr": 0.0026291804804649314, "grad_norm": 0.166253, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:17.038326+00:00", "epoch": 0, "step": 7903, "train_loss": 3.5462560653686523, "perplexity": 34.68322238836418, "lr": 0.0026291804804649314, "grad_norm": 0.187371, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:17.343579+00:00", "epoch": 0, "step": 7904, "train_loss": 3.6051204204559326, "perplexity": 36.786113391082125, "lr": 0.0026291804804649314, "grad_norm": 0.160414, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:17.647011+00:00", "epoch": 0, "step": 7905, "train_loss": 3.6394596099853516, "perplexity": 38.07125783800641, "lr": 0.0026291804804649314, "grad_norm": 0.168742, "tokens_per_sec": 107991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:17.951005+00:00", "epoch": 0, "step": 7906, "train_loss": 3.5333118438720703, "perplexity": 34.23716821568001, "lr": 0.0026291804804649314, "grad_norm": 0.186126, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:18.254678+00:00", "epoch": 0, "step": 7907, "train_loss": 3.4893863201141357, "perplexity": 32.76583380196606, "lr": 0.0026291804804649314, "grad_norm": 0.166582, "tokens_per_sec": 107906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:18.558978+00:00", "epoch": 0, "step": 7908, "train_loss": 3.575317144393921, "perplexity": 35.70594295982307, "lr": 0.0026291804804649314, "grad_norm": 0.162486, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:18.864590+00:00", "epoch": 0, "step": 7909, "train_loss": 3.555436134338379, "perplexity": 35.00308268546224, "lr": 0.0026291804804649314, "grad_norm": 0.150931, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:19.170307+00:00", "epoch": 0, "step": 7910, "train_loss": 3.482908248901367, "perplexity": 32.55426043083002, "lr": 0.0026291804804649314, "grad_norm": 0.164, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:19.474118+00:00", "epoch": 0, "step": 7911, "train_loss": 3.438959836959839, "perplexity": 31.154535512435167, "lr": 0.0026291804804649314, "grad_norm": 0.157225, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:19.779090+00:00", "epoch": 0, "step": 7912, "train_loss": 3.5436174869537354, "perplexity": 34.59182861425463, "lr": 0.0026291804804649314, "grad_norm": 0.141511, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:20.083864+00:00", "epoch": 0, "step": 7913, "train_loss": 3.511359453201294, "perplexity": 33.4937700607617, "lr": 0.0026291804804649314, "grad_norm": 0.174788, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:20.389182+00:00", "epoch": 0, "step": 7914, "train_loss": 3.4227254390716553, "perplexity": 30.652843737445412, "lr": 0.0026291804804649314, "grad_norm": 0.168478, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:20.693500+00:00", "epoch": 0, "step": 7915, "train_loss": 3.566826581954956, "perplexity": 35.40406280138182, "lr": 0.0026291804804649314, "grad_norm": 0.16281, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:20.997545+00:00", "epoch": 0, "step": 7916, "train_loss": 3.620405912399292, "perplexity": 37.352726680168125, "lr": 0.0026291804804649314, "grad_norm": 0.163112, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:21.302294+00:00", "epoch": 0, "step": 7917, "train_loss": 3.6352789402008057, "perplexity": 37.912426722264065, "lr": 0.0026291804804649314, "grad_norm": 0.167434, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:21.606541+00:00", "epoch": 0, "step": 7918, "train_loss": 3.6461472511291504, "perplexity": 38.326718009184795, "lr": 0.0026291804804649314, "grad_norm": 0.178186, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:21.912061+00:00", "epoch": 0, "step": 7919, "train_loss": 3.5091357231140137, "perplexity": 33.419371708126896, "lr": 0.0026291804804649314, "grad_norm": 0.170714, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:22.218433+00:00", "epoch": 0, "step": 7920, "train_loss": 3.543992757797241, "perplexity": 34.60481235501441, "lr": 0.0026291804804649314, "grad_norm": 0.158058, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:22.522675+00:00", "epoch": 0, "step": 7921, "train_loss": 3.520242929458618, "perplexity": 33.79263669373606, "lr": 0.0026291804804649314, "grad_norm": 0.160027, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:22.826870+00:00", "epoch": 0, "step": 7922, "train_loss": 3.5403637886047363, "perplexity": 34.47946014428906, "lr": 0.0026291804804649314, "grad_norm": 0.176083, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:23.131507+00:00", "epoch": 0, "step": 7923, "train_loss": 3.454803943634033, "perplexity": 31.652082481645525, "lr": 0.0026291804804649314, "grad_norm": 0.152517, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:23.435848+00:00", "epoch": 0, "step": 7924, "train_loss": 3.550626516342163, "perplexity": 34.83513543386919, "lr": 0.0026291804804649314, "grad_norm": 0.154787, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:23.741045+00:00", "epoch": 0, "step": 7925, "train_loss": 3.4975922107696533, "perplexity": 33.03581284568273, "lr": 0.0026291804804649314, "grad_norm": 0.153787, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:24.045984+00:00", "epoch": 0, "step": 7926, "train_loss": 3.5149214267730713, "perplexity": 33.613286715755066, "lr": 0.0026291804804649314, "grad_norm": 0.161618, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:24.350074+00:00", "epoch": 0, "step": 7927, "train_loss": 3.5883092880249023, "perplexity": 36.17286629890682, "lr": 0.0026291804804649314, "grad_norm": 0.155238, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:24.654172+00:00", "epoch": 0, "step": 7928, "train_loss": 3.553781270980835, "perplexity": 34.945205269341024, "lr": 0.0026291804804649314, "grad_norm": 0.155957, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:24.959106+00:00", "epoch": 0, "step": 7929, "train_loss": 3.661701202392578, "perplexity": 38.927510146809816, "lr": 0.0026291804804649314, "grad_norm": 0.152072, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:25.264374+00:00", "epoch": 0, "step": 7930, "train_loss": 3.562732696533203, "perplexity": 35.259418904596636, "lr": 0.0026291804804649314, "grad_norm": 0.153228, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:25.569298+00:00", "epoch": 0, "step": 7931, "train_loss": 3.704058885574341, "perplexity": 40.611808967098725, "lr": 0.0026291804804649314, "grad_norm": 0.165692, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:25.875004+00:00", "epoch": 0, "step": 7932, "train_loss": 3.5823731422424316, "perplexity": 35.95877495856596, "lr": 0.0026291804804649314, "grad_norm": 0.168152, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:26.179542+00:00", "epoch": 0, "step": 7933, "train_loss": 3.5410377979278564, "perplexity": 34.502707455454605, "lr": 0.0026291804804649314, "grad_norm": 0.154099, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:26.483806+00:00", "epoch": 0, "step": 7934, "train_loss": 3.544036626815796, "perplexity": 34.6063304674685, "lr": 0.0026291804804649314, "grad_norm": 0.160376, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:26.788307+00:00", "epoch": 0, "step": 7935, "train_loss": 3.436607599258423, "perplexity": 31.081338761261513, "lr": 0.0026291804804649314, "grad_norm": 0.155795, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:27.093897+00:00", "epoch": 0, "step": 7936, "train_loss": 3.509892463684082, "perplexity": 33.444671073845534, "lr": 0.0026291804804649314, "grad_norm": 0.141904, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:27.397768+00:00", "epoch": 0, "step": 7937, "train_loss": 3.5132851600646973, "perplexity": 33.55833138679196, "lr": 0.0026291804804649314, "grad_norm": 0.198569, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:27.701583+00:00", "epoch": 0, "step": 7938, "train_loss": 3.528210163116455, "perplexity": 34.06294590464864, "lr": 0.0026291804804649314, "grad_norm": 0.177655, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:28.007739+00:00", "epoch": 0, "step": 7939, "train_loss": 3.441150665283203, "perplexity": 31.22286457252707, "lr": 0.0026291804804649314, "grad_norm": 0.148582, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:28.312340+00:00", "epoch": 0, "step": 7940, "train_loss": 3.510298252105713, "perplexity": 33.458245288075766, "lr": 0.0026291804804649314, "grad_norm": 0.168705, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:28.617250+00:00", "epoch": 0, "step": 7941, "train_loss": 3.610203742980957, "perplexity": 36.973585155921974, "lr": 0.0026291804804649314, "grad_norm": 0.171086, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:28.923687+00:00", "epoch": 0, "step": 7942, "train_loss": 3.5021796226501465, "perplexity": 33.18770986673119, "lr": 0.0026291804804649314, "grad_norm": 0.176646, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:29.227905+00:00", "epoch": 0, "step": 7943, "train_loss": 3.554175853729248, "perplexity": 34.9589967652443, "lr": 0.0026291804804649314, "grad_norm": 0.150879, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:29.532219+00:00", "epoch": 0, "step": 7944, "train_loss": 3.5080158710479736, "perplexity": 33.38196790294696, "lr": 0.0026291804804649314, "grad_norm": 0.166359, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:29.837208+00:00", "epoch": 0, "step": 7945, "train_loss": 3.60394024848938, "perplexity": 36.74272505918199, "lr": 0.0026291804804649314, "grad_norm": 0.147007, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:30.142846+00:00", "epoch": 0, "step": 7946, "train_loss": 3.5036637783050537, "perplexity": 33.23700216367586, "lr": 0.0026291804804649314, "grad_norm": 0.161612, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:30.447822+00:00", "epoch": 0, "step": 7947, "train_loss": 3.6746888160705566, "perplexity": 39.43638297896431, "lr": 0.0026291804804649314, "grad_norm": 0.149728, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:30.752114+00:00", "epoch": 0, "step": 7948, "train_loss": 3.6085550785064697, "perplexity": 36.912678340827995, "lr": 0.0026291804804649314, "grad_norm": 0.154765, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:31.057733+00:00", "epoch": 0, "step": 7949, "train_loss": 3.5662708282470703, "perplexity": 35.38439232868109, "lr": 0.0026291804804649314, "grad_norm": 0.175004, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:31.362643+00:00", "epoch": 0, "step": 7950, "train_loss": 3.501528739929199, "perplexity": 33.16611558826485, "lr": 0.0026291804804649314, "grad_norm": 0.155923, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:31.666871+00:00", "epoch": 0, "step": 7951, "train_loss": 3.5279552936553955, "perplexity": 34.05426540622779, "lr": 0.0026291804804649314, "grad_norm": 0.159782, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:31.973179+00:00", "epoch": 0, "step": 7952, "train_loss": 3.5429296493530273, "perplexity": 34.568043235033365, "lr": 0.0026291804804649314, "grad_norm": 0.170137, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:32.277449+00:00", "epoch": 0, "step": 7953, "train_loss": 3.5686745643615723, "perplexity": 35.46954937694466, "lr": 0.0026291804804649314, "grad_norm": 0.16912, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:32.581414+00:00", "epoch": 0, "step": 7954, "train_loss": 3.5586469173431396, "perplexity": 35.11565060734178, "lr": 0.0026291804804649314, "grad_norm": 0.144244, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:32.885421+00:00", "epoch": 0, "step": 7955, "train_loss": 3.5885708332061768, "perplexity": 36.18232837510646, "lr": 0.0026291804804649314, "grad_norm": 0.154206, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:33.189472+00:00", "epoch": 0, "step": 7956, "train_loss": 3.542438268661499, "perplexity": 34.55106133867344, "lr": 0.0026291804804649314, "grad_norm": 0.171444, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:33.494997+00:00", "epoch": 0, "step": 7957, "train_loss": 3.38254976272583, "perplexity": 29.445755165849697, "lr": 0.0026291804804649314, "grad_norm": 0.16352, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:33.799628+00:00", "epoch": 0, "step": 7958, "train_loss": 3.5403449535369873, "perplexity": 34.478810727437214, "lr": 0.0026291804804649314, "grad_norm": 0.144596, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:34.103959+00:00", "epoch": 0, "step": 7959, "train_loss": 3.683082103729248, "perplexity": 39.768776872957766, "lr": 0.0026291804804649314, "grad_norm": 0.157689, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:34.408571+00:00", "epoch": 0, "step": 7960, "train_loss": 3.6070356369018555, "perplexity": 36.85663427024789, "lr": 0.0026291804804649314, "grad_norm": 0.177617, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:34.712475+00:00", "epoch": 0, "step": 7961, "train_loss": 3.5884759426116943, "perplexity": 36.17889517534894, "lr": 0.0026291804804649314, "grad_norm": 0.180449, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:35.017990+00:00", "epoch": 0, "step": 7962, "train_loss": 3.512232780456543, "perplexity": 33.523033859613406, "lr": 0.0026291804804649314, "grad_norm": 0.18523, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:35.323468+00:00", "epoch": 0, "step": 7963, "train_loss": 3.6158370971679688, "perplexity": 37.182458232302224, "lr": 0.0026291804804649314, "grad_norm": 0.186318, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:35.627424+00:00", "epoch": 0, "step": 7964, "train_loss": 3.5685482025146484, "perplexity": 35.465067662340665, "lr": 0.0026291804804649314, "grad_norm": 0.151442, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:35.930905+00:00", "epoch": 0, "step": 7965, "train_loss": 3.502092123031616, "perplexity": 33.18480608181985, "lr": 0.0026291804804649314, "grad_norm": 0.17223, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:36.236001+00:00", "epoch": 0, "step": 7966, "train_loss": 3.5746657848358154, "perplexity": 35.68269312541796, "lr": 0.0026291804804649314, "grad_norm": 0.202422, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:36.540412+00:00", "epoch": 0, "step": 7967, "train_loss": 3.579341173171997, "perplexity": 35.849914179733005, "lr": 0.0026291804804649314, "grad_norm": 0.146098, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:36.844578+00:00", "epoch": 0, "step": 7968, "train_loss": 3.534637212753296, "perplexity": 34.28257517685419, "lr": 0.0026291804804649314, "grad_norm": 0.219144, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:37.150031+00:00", "epoch": 0, "step": 7969, "train_loss": 3.7329447269439697, "perplexity": 41.802022642809064, "lr": 0.0026291804804649314, "grad_norm": 0.172613, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:37.454706+00:00", "epoch": 0, "step": 7970, "train_loss": 3.6292476654052734, "perplexity": 37.6844546307656, "lr": 0.0026291804804649314, "grad_norm": 0.15113, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:37.759986+00:00", "epoch": 0, "step": 7971, "train_loss": 3.514061212539673, "perplexity": 33.584384520912785, "lr": 0.0026291804804649314, "grad_norm": 0.173024, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:38.066938+00:00", "epoch": 0, "step": 7972, "train_loss": 3.5454938411712646, "perplexity": 34.6567960696856, "lr": 0.0026291804804649314, "grad_norm": 0.181914, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:38.371972+00:00", "epoch": 0, "step": 7973, "train_loss": 3.5085866451263428, "perplexity": 33.40102690358525, "lr": 0.0026291804804649314, "grad_norm": 0.166422, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:38.677224+00:00", "epoch": 0, "step": 7974, "train_loss": 3.462772846221924, "perplexity": 31.90532252613617, "lr": 0.0026291804804649314, "grad_norm": 0.158667, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:38.981387+00:00", "epoch": 0, "step": 7975, "train_loss": 3.684265613555908, "perplexity": 39.81587147414571, "lr": 0.0026291804804649314, "grad_norm": 0.163519, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:39.286496+00:00", "epoch": 0, "step": 7976, "train_loss": 3.5182387828826904, "perplexity": 33.72497911711113, "lr": 0.0026291804804649314, "grad_norm": 0.185633, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:39.590590+00:00", "epoch": 0, "step": 7977, "train_loss": 3.5320351123809814, "perplexity": 34.19348443701515, "lr": 0.0026291804804649314, "grad_norm": 0.132291, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:39.897373+00:00", "epoch": 0, "step": 7978, "train_loss": 3.5678420066833496, "perplexity": 35.44003122075663, "lr": 0.0026291804804649314, "grad_norm": 0.182714, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:40.203211+00:00", "epoch": 0, "step": 7979, "train_loss": 3.6172821521759033, "perplexity": 37.23622777039155, "lr": 0.0026291804804649314, "grad_norm": 0.190582, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:40.508167+00:00", "epoch": 0, "step": 7980, "train_loss": 3.685070276260376, "perplexity": 39.84792271445631, "lr": 0.0026291804804649314, "grad_norm": 0.181896, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:40.812758+00:00", "epoch": 0, "step": 7981, "train_loss": 3.5483615398406982, "perplexity": 34.75632395744988, "lr": 0.0026291804804649314, "grad_norm": 0.174395, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:41.117844+00:00", "epoch": 0, "step": 7982, "train_loss": 3.6142232418060303, "perplexity": 37.12249951807388, "lr": 0.0026291804804649314, "grad_norm": 0.161284, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:41.422606+00:00", "epoch": 0, "step": 7983, "train_loss": 3.4874095916748047, "perplexity": 32.70112861980757, "lr": 0.0026291804804649314, "grad_norm": 0.157543, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:41.728576+00:00", "epoch": 0, "step": 7984, "train_loss": 3.5425992012023926, "perplexity": 34.5566221762131, "lr": 0.0026291804804649314, "grad_norm": 0.15637, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:42.033210+00:00", "epoch": 0, "step": 7985, "train_loss": 3.5639002323150635, "perplexity": 35.30060957892938, "lr": 0.0026291804804649314, "grad_norm": 0.168385, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:42.338955+00:00", "epoch": 0, "step": 7986, "train_loss": 3.552680730819702, "perplexity": 34.90676782235972, "lr": 0.0026291804804649314, "grad_norm": 0.162556, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:42.644463+00:00", "epoch": 0, "step": 7987, "train_loss": 3.5592854022979736, "perplexity": 35.13807858113376, "lr": 0.0026291804804649314, "grad_norm": 0.149323, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:42.949030+00:00", "epoch": 0, "step": 7988, "train_loss": 3.473118782043457, "perplexity": 32.23712639649023, "lr": 0.0026291804804649314, "grad_norm": 0.175306, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:43.255647+00:00", "epoch": 0, "step": 7989, "train_loss": 3.5246543884277344, "perplexity": 33.942040826774374, "lr": 0.0026291804804649314, "grad_norm": 0.177338, "tokens_per_sec": 106869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:43.561530+00:00", "epoch": 0, "step": 7990, "train_loss": 3.47351336479187, "perplexity": 32.2498491203433, "lr": 0.0026291804804649314, "grad_norm": 0.141041, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:43.866909+00:00", "epoch": 0, "step": 7991, "train_loss": 3.6632165908813477, "perplexity": 38.98654516678025, "lr": 0.0026291804804649314, "grad_norm": 0.160416, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:44.171998+00:00", "epoch": 0, "step": 7992, "train_loss": 3.456989049911499, "perplexity": 31.721321265271936, "lr": 0.0026291804804649314, "grad_norm": 0.160597, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:44.476674+00:00", "epoch": 0, "step": 7993, "train_loss": 3.5060946941375732, "perplexity": 33.31789680265412, "lr": 0.0026291804804649314, "grad_norm": 0.171086, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:44.781049+00:00", "epoch": 0, "step": 7994, "train_loss": 3.543806314468384, "perplexity": 34.59836112001801, "lr": 0.0026291804804649314, "grad_norm": 0.166823, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:45.086277+00:00", "epoch": 0, "step": 7995, "train_loss": 3.7084217071533203, "perplexity": 40.78937811321568, "lr": 0.0026291804804649314, "grad_norm": 0.155708, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:45.391412+00:00", "epoch": 0, "step": 7996, "train_loss": 3.605581521987915, "perplexity": 36.80307943555638, "lr": 0.0026291804804649314, "grad_norm": 0.160526, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:45.694969+00:00", "epoch": 0, "step": 7997, "train_loss": 3.6460211277008057, "perplexity": 38.321884416933294, "lr": 0.0026291804804649314, "grad_norm": 0.178837, "tokens_per_sec": 107937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:46.000555+00:00", "epoch": 0, "step": 7998, "train_loss": 3.5372684001922607, "perplexity": 34.372897833845414, "lr": 0.0026291804804649314, "grad_norm": 0.155855, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:46.307558+00:00", "epoch": 0, "step": 7999, "train_loss": 3.601830244064331, "perplexity": 36.66527948070848, "lr": 0.0026291804804649314, "grad_norm": 0.151832, "tokens_per_sec": 106735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:46.612918+00:00", "epoch": 0, "step": 8000, "train_loss": 3.593937873840332, "perplexity": 36.377042453522726, "lr": 0.0026291804804649314, "grad_norm": 0.16468, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:39:49.685851+00:00", "step": 8000, "epoch": 0, "val_loss": 3.5143845796585085, "val_ppl": 33.59524636265977, "eval_train_loss": 3.593937873840332, "eval_train_ppl": 36.377042453522726} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:39:50.663141+00:00", "step": 8000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5144_epoch_0000_step_0008000.pt", "val_loss": 3.5143845796585085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:51.970579+00:00", "epoch": 0, "step": 8001, "train_loss": 3.4747238159179688, "perplexity": 32.28890962217248, "lr": 0.0026291804804649314, "grad_norm": 0.152486, "tokens_per_sec": 6116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:52.274813+00:00", "epoch": 0, "step": 8002, "train_loss": 3.4954605102539062, "perplexity": 32.965465392891254, "lr": 0.0026291804804649314, "grad_norm": 0.161097, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:52.578550+00:00", "epoch": 0, "step": 8003, "train_loss": 3.419318437576294, "perplexity": 30.548587155008878, "lr": 0.0026291804804649314, "grad_norm": 0.15107, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:52.883034+00:00", "epoch": 0, "step": 8004, "train_loss": 3.4335217475891113, "perplexity": 30.985574193778856, "lr": 0.0026291804804649314, "grad_norm": 0.169541, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:53.186110+00:00", "epoch": 0, "step": 8005, "train_loss": 3.518681526184082, "perplexity": 33.739913931605386, "lr": 0.0026291804804649314, "grad_norm": 0.166994, "tokens_per_sec": 108118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:53.489415+00:00", "epoch": 0, "step": 8006, "train_loss": 3.5552515983581543, "perplexity": 34.99662395324056, "lr": 0.0026291804804649314, "grad_norm": 0.178573, "tokens_per_sec": 108036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:53.792674+00:00", "epoch": 0, "step": 8007, "train_loss": 3.583705186843872, "perplexity": 36.006705566390515, "lr": 0.0026291804804649314, "grad_norm": 0.190603, "tokens_per_sec": 108053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:54.095440+00:00", "epoch": 0, "step": 8008, "train_loss": 3.542290687561035, "perplexity": 34.545962631264565, "lr": 0.0026291804804649314, "grad_norm": 0.182586, "tokens_per_sec": 108230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:54.398168+00:00", "epoch": 0, "step": 8009, "train_loss": 3.5522358417510986, "perplexity": 34.89124163690645, "lr": 0.0026291804804649314, "grad_norm": 0.140729, "tokens_per_sec": 108243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:54.701834+00:00", "epoch": 0, "step": 8010, "train_loss": 3.5304508209228516, "perplexity": 34.1393548815161, "lr": 0.0026291804804649314, "grad_norm": 0.173876, "tokens_per_sec": 107967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:55.005186+00:00", "epoch": 0, "step": 8011, "train_loss": 3.5267019271850586, "perplexity": 34.01160966901515, "lr": 0.0026291804804649314, "grad_norm": 0.157146, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:55.309855+00:00", "epoch": 0, "step": 8012, "train_loss": 3.559950590133667, "perplexity": 35.161459779157695, "lr": 0.0026291804804649314, "grad_norm": 0.154318, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:55.614081+00:00", "epoch": 0, "step": 8013, "train_loss": 3.553199052810669, "perplexity": 34.9248654575516, "lr": 0.0026291804804649314, "grad_norm": 0.187143, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:55.917900+00:00", "epoch": 0, "step": 8014, "train_loss": 3.6215291023254395, "perplexity": 37.39470445658412, "lr": 0.0026291804804649314, "grad_norm": 0.16037, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:56.221756+00:00", "epoch": 0, "step": 8015, "train_loss": 3.553722858428955, "perplexity": 34.94316409034109, "lr": 0.0026291804804649314, "grad_norm": 0.16824, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:56.526300+00:00", "epoch": 0, "step": 8016, "train_loss": 3.6218674182891846, "perplexity": 37.40735782235844, "lr": 0.0026291804804649314, "grad_norm": 0.175051, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:56.830270+00:00", "epoch": 0, "step": 8017, "train_loss": 3.5336592197418213, "perplexity": 34.24906344770541, "lr": 0.0026291804804649314, "grad_norm": 0.189128, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:57.134313+00:00", "epoch": 0, "step": 8018, "train_loss": 3.660536050796509, "perplexity": 38.88218010954215, "lr": 0.0026291804804649314, "grad_norm": 0.167695, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:57.438158+00:00", "epoch": 0, "step": 8019, "train_loss": 3.6011462211608887, "perplexity": 36.64020816543703, "lr": 0.0026291804804649314, "grad_norm": 0.165328, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:57.741711+00:00", "epoch": 0, "step": 8020, "train_loss": 3.587928056716919, "perplexity": 36.159078698073714, "lr": 0.0026291804804649314, "grad_norm": 0.167953, "tokens_per_sec": 108014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:58.046746+00:00", "epoch": 0, "step": 8021, "train_loss": 3.6409013271331787, "perplexity": 38.12618540876701, "lr": 0.0026291804804649314, "grad_norm": 0.14756, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:58.352265+00:00", "epoch": 0, "step": 8022, "train_loss": 3.6001808643341064, "perplexity": 36.60485435761523, "lr": 0.0026291804804649314, "grad_norm": 0.164386, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:58.655975+00:00", "epoch": 0, "step": 8023, "train_loss": 3.5032691955566406, "perplexity": 33.22389000309941, "lr": 0.0026291804804649314, "grad_norm": 0.175798, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:58.959642+00:00", "epoch": 0, "step": 8024, "train_loss": 3.449306011199951, "perplexity": 31.478538973150677, "lr": 0.0026291804804649314, "grad_norm": 0.152929, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:59.262366+00:00", "epoch": 0, "step": 8025, "train_loss": 3.5128285884857178, "perplexity": 33.54301310365524, "lr": 0.0026291804804649314, "grad_norm": 0.171765, "tokens_per_sec": 108243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:59.565734+00:00", "epoch": 0, "step": 8026, "train_loss": 3.5502865314483643, "perplexity": 34.82329402711201, "lr": 0.0026291804804649314, "grad_norm": 0.143086, "tokens_per_sec": 108015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:39:59.869550+00:00", "epoch": 0, "step": 8027, "train_loss": 3.58370041847229, "perplexity": 36.00653387344828, "lr": 0.0026291804804649314, "grad_norm": 0.16223, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:00.173910+00:00", "epoch": 0, "step": 8028, "train_loss": 3.529083728790283, "perplexity": 34.09271512573213, "lr": 0.0026291804804649314, "grad_norm": 0.139899, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:00.476715+00:00", "epoch": 0, "step": 8029, "train_loss": 3.4637303352355957, "perplexity": 31.935886151765462, "lr": 0.0026291804804649314, "grad_norm": 0.159627, "tokens_per_sec": 108210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:00.780372+00:00", "epoch": 0, "step": 8030, "train_loss": 3.6354763507843018, "perplexity": 37.9199117753351, "lr": 0.0026291804804649314, "grad_norm": 0.16004, "tokens_per_sec": 108005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:01.084353+00:00", "epoch": 0, "step": 8031, "train_loss": 3.480616331100464, "perplexity": 32.479734178498695, "lr": 0.0026291804804649314, "grad_norm": 0.160526, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:01.388804+00:00", "epoch": 0, "step": 8032, "train_loss": 3.6371920108795166, "perplexity": 37.98502529514416, "lr": 0.0026291804804649314, "grad_norm": 0.154688, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:01.693620+00:00", "epoch": 0, "step": 8033, "train_loss": 3.527308225631714, "perplexity": 34.03223710768585, "lr": 0.0026291804804649314, "grad_norm": 0.14291, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:01.997625+00:00", "epoch": 0, "step": 8034, "train_loss": 3.6331255435943604, "perplexity": 37.83087407033429, "lr": 0.0026291804804649314, "grad_norm": 0.158355, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:02.301106+00:00", "epoch": 0, "step": 8035, "train_loss": 3.5155954360961914, "perplexity": 33.635950021163794, "lr": 0.0026291804804649314, "grad_norm": 0.159364, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:02.603696+00:00", "epoch": 0, "step": 8036, "train_loss": 3.5411486625671387, "perplexity": 34.50653279771459, "lr": 0.0026291804804649314, "grad_norm": 0.15201, "tokens_per_sec": 108291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:02.907396+00:00", "epoch": 0, "step": 8037, "train_loss": 3.456954002380371, "perplexity": 31.72020953075936, "lr": 0.0026291804804649314, "grad_norm": 0.150252, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:03.210532+00:00", "epoch": 0, "step": 8038, "train_loss": 3.5836448669433594, "perplexity": 36.00453371099667, "lr": 0.0026291804804649314, "grad_norm": 0.152677, "tokens_per_sec": 108097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:03.513719+00:00", "epoch": 0, "step": 8039, "train_loss": 3.4889793395996094, "perplexity": 32.75250145925232, "lr": 0.0026291804804649314, "grad_norm": 0.157352, "tokens_per_sec": 108078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:03.817355+00:00", "epoch": 0, "step": 8040, "train_loss": 3.566312074661255, "perplexity": 35.3858518380823, "lr": 0.0026291804804649314, "grad_norm": 0.161575, "tokens_per_sec": 107976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:04.121882+00:00", "epoch": 0, "step": 8041, "train_loss": 3.5695548057556152, "perplexity": 35.500784887912324, "lr": 0.0026291804804649314, "grad_norm": 0.181702, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:04.425586+00:00", "epoch": 0, "step": 8042, "train_loss": 3.5770950317382812, "perplexity": 35.769480568548765, "lr": 0.0026291804804649314, "grad_norm": 0.167881, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:04.730614+00:00", "epoch": 0, "step": 8043, "train_loss": 3.5569915771484375, "perplexity": 35.05757034398794, "lr": 0.0026291804804649314, "grad_norm": 0.154839, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:05.034744+00:00", "epoch": 0, "step": 8044, "train_loss": 3.6023712158203125, "perplexity": 36.685119727355385, "lr": 0.0026291804804649314, "grad_norm": 0.1729, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:05.338269+00:00", "epoch": 0, "step": 8045, "train_loss": 3.5414881706237793, "perplexity": 34.51825003254141, "lr": 0.0026291804804649314, "grad_norm": 0.204411, "tokens_per_sec": 107958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:05.642002+00:00", "epoch": 0, "step": 8046, "train_loss": 3.606360912322998, "perplexity": 36.83177458087663, "lr": 0.0026291804804649314, "grad_norm": 0.180799, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:05.946283+00:00", "epoch": 0, "step": 8047, "train_loss": 3.6015210151672363, "perplexity": 36.65394326960535, "lr": 0.0026291804804649314, "grad_norm": 0.171591, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:06.251328+00:00", "epoch": 0, "step": 8048, "train_loss": 3.5116662979125977, "perplexity": 33.50404902390855, "lr": 0.0026291804804649314, "grad_norm": 0.171363, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:06.555212+00:00", "epoch": 0, "step": 8049, "train_loss": 3.526447057723999, "perplexity": 34.00294225296579, "lr": 0.0026291804804649314, "grad_norm": 0.155049, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:06.858850+00:00", "epoch": 0, "step": 8050, "train_loss": 3.593940496444702, "perplexity": 36.37713785623834, "lr": 0.0026291804804649314, "grad_norm": 0.1748, "tokens_per_sec": 107981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:07.163714+00:00", "epoch": 0, "step": 8051, "train_loss": 3.5458176136016846, "perplexity": 34.668018801487186, "lr": 0.0026291804804649314, "grad_norm": 0.178409, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:07.468867+00:00", "epoch": 0, "step": 8052, "train_loss": 3.5261738300323486, "perplexity": 33.993652976646246, "lr": 0.0026291804804649314, "grad_norm": 0.157146, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:07.772634+00:00", "epoch": 0, "step": 8053, "train_loss": 3.597219705581665, "perplexity": 36.49662189858501, "lr": 0.0026291804804649314, "grad_norm": 0.135597, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:08.078206+00:00", "epoch": 0, "step": 8054, "train_loss": 3.5922327041625977, "perplexity": 36.31506627872917, "lr": 0.0026291804804649314, "grad_norm": 0.152114, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:08.381741+00:00", "epoch": 0, "step": 8055, "train_loss": 3.518388271331787, "perplexity": 33.730020988776566, "lr": 0.0026291804804649314, "grad_norm": 0.153683, "tokens_per_sec": 107955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:08.685264+00:00", "epoch": 0, "step": 8056, "train_loss": 3.489670515060425, "perplexity": 32.775147009664124, "lr": 0.0026291804804649314, "grad_norm": 0.147964, "tokens_per_sec": 107959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:08.988918+00:00", "epoch": 0, "step": 8057, "train_loss": 3.5337750911712646, "perplexity": 34.253032165570254, "lr": 0.0026291804804649314, "grad_norm": 0.148106, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:09.292631+00:00", "epoch": 0, "step": 8058, "train_loss": 3.522378921508789, "perplexity": 33.86489464079558, "lr": 0.0026291804804649314, "grad_norm": 0.137229, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:09.597182+00:00", "epoch": 0, "step": 8059, "train_loss": 3.5083370208740234, "perplexity": 33.39269023777802, "lr": 0.0026291804804649314, "grad_norm": 0.156075, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:09.901484+00:00", "epoch": 0, "step": 8060, "train_loss": 3.5710947513580322, "perplexity": 35.55549628095534, "lr": 0.0026291804804649314, "grad_norm": 0.195996, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:10.205511+00:00", "epoch": 0, "step": 8061, "train_loss": 3.543623208999634, "perplexity": 34.592026550851976, "lr": 0.0026291804804649314, "grad_norm": 0.198777, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:10.509380+00:00", "epoch": 0, "step": 8062, "train_loss": 3.531067132949829, "perplexity": 34.160401861610104, "lr": 0.0026291804804649314, "grad_norm": 0.18203, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:10.813946+00:00", "epoch": 0, "step": 8063, "train_loss": 3.54939603805542, "perplexity": 34.79229791682429, "lr": 0.0026291804804649314, "grad_norm": 0.23844, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:11.118842+00:00", "epoch": 0, "step": 8064, "train_loss": 3.4979403018951416, "perplexity": 33.04731432062214, "lr": 0.0026291804804649314, "grad_norm": 0.203465, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:11.422778+00:00", "epoch": 0, "step": 8065, "train_loss": 3.5186455249786377, "perplexity": 33.73869927589692, "lr": 0.0026291804804649314, "grad_norm": 0.170398, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:11.726092+00:00", "epoch": 0, "step": 8066, "train_loss": 3.5347840785980225, "perplexity": 34.28761048596535, "lr": 0.0026291804804649314, "grad_norm": 0.19736, "tokens_per_sec": 108031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:12.029870+00:00", "epoch": 0, "step": 8067, "train_loss": 3.6063597202301025, "perplexity": 36.831730674005996, "lr": 0.0026291804804649314, "grad_norm": 0.187713, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:12.333845+00:00", "epoch": 0, "step": 8068, "train_loss": 3.5344996452331543, "perplexity": 34.27785933238399, "lr": 0.0026291804804649314, "grad_norm": 0.175277, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:12.637691+00:00", "epoch": 0, "step": 8069, "train_loss": 3.575834035873413, "perplexity": 35.724403828228, "lr": 0.0026291804804649314, "grad_norm": 0.14673, "tokens_per_sec": 107846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:12.943068+00:00", "epoch": 0, "step": 8070, "train_loss": 3.606598138809204, "perplexity": 36.84051308980309, "lr": 0.0026291804804649314, "grad_norm": 0.164887, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:13.247554+00:00", "epoch": 0, "step": 8071, "train_loss": 3.5377819538116455, "perplexity": 34.39055469342073, "lr": 0.0026291804804649314, "grad_norm": 0.154243, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:13.551437+00:00", "epoch": 0, "step": 8072, "train_loss": 3.6318182945251465, "perplexity": 37.781452005930255, "lr": 0.0026291804804649314, "grad_norm": 0.163581, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:13.855737+00:00", "epoch": 0, "step": 8073, "train_loss": 3.5971784591674805, "perplexity": 36.49511657484666, "lr": 0.0026291804804649314, "grad_norm": 0.145694, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:14.161360+00:00", "epoch": 0, "step": 8074, "train_loss": 3.4564931392669678, "perplexity": 31.705594224323708, "lr": 0.0026291804804649314, "grad_norm": 0.181911, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:14.465561+00:00", "epoch": 0, "step": 8075, "train_loss": 3.576174736022949, "perplexity": 35.736577211573376, "lr": 0.0026291804804649314, "grad_norm": 0.184929, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:14.769507+00:00", "epoch": 0, "step": 8076, "train_loss": 3.43774151802063, "perplexity": 31.11660246374593, "lr": 0.0026291804804649314, "grad_norm": 0.13623, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:15.073227+00:00", "epoch": 0, "step": 8077, "train_loss": 3.5054566860198975, "perplexity": 33.29664649368579, "lr": 0.0026291804804649314, "grad_norm": 0.17011, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:15.376840+00:00", "epoch": 0, "step": 8078, "train_loss": 3.475965976715088, "perplexity": 32.3290425605079, "lr": 0.0026291804804649314, "grad_norm": 0.146644, "tokens_per_sec": 107932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:15.681297+00:00", "epoch": 0, "step": 8079, "train_loss": 3.714050054550171, "perplexity": 41.01960218611053, "lr": 0.0026291804804649314, "grad_norm": 0.147768, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:15.986550+00:00", "epoch": 0, "step": 8080, "train_loss": 3.5026443004608154, "perplexity": 33.20313504268412, "lr": 0.0026291804804649314, "grad_norm": 0.148526, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:16.291148+00:00", "epoch": 0, "step": 8081, "train_loss": 3.554579257965088, "perplexity": 34.97310221752843, "lr": 0.0026291804804649314, "grad_norm": 0.15255, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:16.595393+00:00", "epoch": 0, "step": 8082, "train_loss": 3.6216254234313965, "perplexity": 37.39830652934935, "lr": 0.0026291804804649314, "grad_norm": 0.13918, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:16.899679+00:00", "epoch": 0, "step": 8083, "train_loss": 3.678884983062744, "perplexity": 39.60221230796594, "lr": 0.0026291804804649314, "grad_norm": 0.148969, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:17.204498+00:00", "epoch": 0, "step": 8084, "train_loss": 3.5734705924987793, "perplexity": 35.640070919970086, "lr": 0.0026291804804649314, "grad_norm": 0.161011, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:17.509458+00:00", "epoch": 0, "step": 8085, "train_loss": 3.5451385974884033, "perplexity": 34.64448664836517, "lr": 0.0026291804804649314, "grad_norm": 0.156539, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:17.814511+00:00", "epoch": 0, "step": 8086, "train_loss": 3.6129984855651855, "perplexity": 37.07706133614169, "lr": 0.0026291804804649314, "grad_norm": 0.153027, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:18.117941+00:00", "epoch": 0, "step": 8087, "train_loss": 3.5585317611694336, "perplexity": 35.11160705620504, "lr": 0.0026291804804649314, "grad_norm": 0.146788, "tokens_per_sec": 107992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:18.421731+00:00", "epoch": 0, "step": 8088, "train_loss": 3.5264477729797363, "perplexity": 34.00296657377402, "lr": 0.0026291804804649314, "grad_norm": 0.164969, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:18.725452+00:00", "epoch": 0, "step": 8089, "train_loss": 3.5193800926208496, "perplexity": 33.763491737430506, "lr": 0.0026291804804649314, "grad_norm": 0.164904, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:19.031001+00:00", "epoch": 0, "step": 8090, "train_loss": 3.602811813354492, "perplexity": 36.70128666194257, "lr": 0.0026291804804649314, "grad_norm": 0.168426, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:19.334976+00:00", "epoch": 0, "step": 8091, "train_loss": 3.5863749980926514, "perplexity": 36.10296511417759, "lr": 0.0026291804804649314, "grad_norm": 0.153435, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:19.639033+00:00", "epoch": 0, "step": 8092, "train_loss": 3.727106809616089, "perplexity": 41.558696839514646, "lr": 0.0026291804804649314, "grad_norm": 0.171026, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:19.942484+00:00", "epoch": 0, "step": 8093, "train_loss": 3.6198179721832275, "perplexity": 37.33077196463572, "lr": 0.0026291804804649314, "grad_norm": 0.162247, "tokens_per_sec": 107985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:20.247000+00:00", "epoch": 0, "step": 8094, "train_loss": 3.542518377304077, "perplexity": 34.553829288163776, "lr": 0.0026291804804649314, "grad_norm": 0.180973, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:20.551674+00:00", "epoch": 0, "step": 8095, "train_loss": 3.5232911109924316, "perplexity": 33.89579993515083, "lr": 0.0026291804804649314, "grad_norm": 0.186829, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:20.856280+00:00", "epoch": 0, "step": 8096, "train_loss": 3.5650341510772705, "perplexity": 35.34066030529196, "lr": 0.0026291804804649314, "grad_norm": 0.162713, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:21.160447+00:00", "epoch": 0, "step": 8097, "train_loss": 3.5369038581848145, "perplexity": 34.360369752312046, "lr": 0.0026291804804649314, "grad_norm": 0.177632, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:21.464914+00:00", "epoch": 0, "step": 8098, "train_loss": 3.610347270965576, "perplexity": 36.978892280934915, "lr": 0.0026291804804649314, "grad_norm": 0.168155, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:21.770240+00:00", "epoch": 0, "step": 8099, "train_loss": 3.536553382873535, "perplexity": 34.348329401073606, "lr": 0.0026291804804649314, "grad_norm": 0.170064, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:22.076120+00:00", "epoch": 0, "step": 8100, "train_loss": 3.6027541160583496, "perplexity": 36.69916915802493, "lr": 0.0026291804804649314, "grad_norm": 0.165135, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:22.380148+00:00", "epoch": 0, "step": 8101, "train_loss": 3.563434600830078, "perplexity": 35.28417632988624, "lr": 0.0026291804804649314, "grad_norm": 0.174027, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:22.685412+00:00", "epoch": 0, "step": 8102, "train_loss": 3.549802541732788, "perplexity": 34.80644398889193, "lr": 0.0026291804804649314, "grad_norm": 0.155217, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:22.990292+00:00", "epoch": 0, "step": 8103, "train_loss": 3.5996596813201904, "perplexity": 36.585781499953626, "lr": 0.0026291804804649314, "grad_norm": 0.16447, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:23.295563+00:00", "epoch": 0, "step": 8104, "train_loss": 3.5465846061706543, "perplexity": 34.694619114105855, "lr": 0.0026291804804649314, "grad_norm": 0.175216, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:23.600336+00:00", "epoch": 0, "step": 8105, "train_loss": 3.605187177658081, "perplexity": 36.78856921106095, "lr": 0.0026291804804649314, "grad_norm": 0.183222, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:23.905784+00:00", "epoch": 0, "step": 8106, "train_loss": 3.5407023429870605, "perplexity": 34.49113529284592, "lr": 0.0026291804804649314, "grad_norm": 0.161425, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:24.211037+00:00", "epoch": 0, "step": 8107, "train_loss": 3.513051986694336, "perplexity": 33.550507389768136, "lr": 0.0026291804804649314, "grad_norm": 0.158127, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:24.515385+00:00", "epoch": 0, "step": 8108, "train_loss": 3.481931447982788, "perplexity": 32.522476924943554, "lr": 0.0026291804804649314, "grad_norm": 0.150984, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:24.819196+00:00", "epoch": 0, "step": 8109, "train_loss": 3.5300369262695312, "perplexity": 34.125227708845124, "lr": 0.0026291804804649314, "grad_norm": 0.17792, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:25.124341+00:00", "epoch": 0, "step": 8110, "train_loss": 3.5833022594451904, "perplexity": 35.99220040064215, "lr": 0.0026291804804649314, "grad_norm": 0.170343, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:25.430276+00:00", "epoch": 0, "step": 8111, "train_loss": 3.5424747467041016, "perplexity": 34.552321716748835, "lr": 0.0026291804804649314, "grad_norm": 0.199975, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:25.735867+00:00", "epoch": 0, "step": 8112, "train_loss": 3.5850324630737305, "perplexity": 36.05452814066674, "lr": 0.0026291804804649314, "grad_norm": 0.144346, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:26.040802+00:00", "epoch": 0, "step": 8113, "train_loss": 3.5490682125091553, "perplexity": 34.7808939821066, "lr": 0.0026291804804649314, "grad_norm": 0.208516, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:26.344752+00:00", "epoch": 0, "step": 8114, "train_loss": 3.5861382484436035, "perplexity": 36.09441876157013, "lr": 0.0026291804804649314, "grad_norm": 0.193456, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:26.648505+00:00", "epoch": 0, "step": 8115, "train_loss": 3.5769641399383545, "perplexity": 35.76479894325457, "lr": 0.0026291804804649314, "grad_norm": 0.17741, "tokens_per_sec": 107874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:26.952250+00:00", "epoch": 0, "step": 8116, "train_loss": 3.6104085445404053, "perplexity": 36.98115817927732, "lr": 0.0026291804804649314, "grad_norm": 0.146361, "tokens_per_sec": 107880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:27.256590+00:00", "epoch": 0, "step": 8117, "train_loss": 3.5451884269714355, "perplexity": 34.646213008236224, "lr": 0.0026291804804649314, "grad_norm": 0.163918, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:27.561294+00:00", "epoch": 0, "step": 8118, "train_loss": 3.57651948928833, "perplexity": 35.74889963723679, "lr": 0.0026291804804649314, "grad_norm": 0.185475, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:27.865626+00:00", "epoch": 0, "step": 8119, "train_loss": 3.5823121070861816, "perplexity": 35.956580276094876, "lr": 0.0026291804804649314, "grad_norm": 0.154072, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:28.170765+00:00", "epoch": 0, "step": 8120, "train_loss": 3.435753583908081, "perplexity": 31.054806152090592, "lr": 0.0026291804804649314, "grad_norm": 0.153694, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:28.474744+00:00", "epoch": 0, "step": 8121, "train_loss": 3.505631446838379, "perplexity": 33.3024659513715, "lr": 0.0026291804804649314, "grad_norm": 0.162623, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:28.779699+00:00", "epoch": 0, "step": 8122, "train_loss": 3.5044612884521484, "perplexity": 33.26351958271031, "lr": 0.0026291804804649314, "grad_norm": 0.161436, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:29.085423+00:00", "epoch": 0, "step": 8123, "train_loss": 3.5282065868377686, "perplexity": 34.06282408627903, "lr": 0.0026291804804649314, "grad_norm": 0.182387, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:29.389525+00:00", "epoch": 0, "step": 8124, "train_loss": 3.604633092880249, "perplexity": 36.76819087104824, "lr": 0.0026291804804649314, "grad_norm": 0.157461, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:29.693720+00:00", "epoch": 0, "step": 8125, "train_loss": 3.5411853790283203, "perplexity": 34.507799778745955, "lr": 0.0026291804804649314, "grad_norm": 0.17305, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:29.998080+00:00", "epoch": 0, "step": 8126, "train_loss": 3.536149501800537, "perplexity": 34.334459562010686, "lr": 0.0026291804804649314, "grad_norm": 0.180295, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:30.303051+00:00", "epoch": 0, "step": 8127, "train_loss": 3.4790561199188232, "perplexity": 32.429098445546764, "lr": 0.0026291804804649314, "grad_norm": 0.165318, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:30.608161+00:00", "epoch": 0, "step": 8128, "train_loss": 3.456193208694458, "perplexity": 31.69608617324512, "lr": 0.0026291804804649314, "grad_norm": 0.163127, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:30.913265+00:00", "epoch": 0, "step": 8129, "train_loss": 3.621480941772461, "perplexity": 37.39290355030569, "lr": 0.0026291804804649314, "grad_norm": 0.158467, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:31.218852+00:00", "epoch": 0, "step": 8130, "train_loss": 3.5500376224517822, "perplexity": 34.814627274599125, "lr": 0.0026291804804649314, "grad_norm": 0.144378, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:31.522743+00:00", "epoch": 0, "step": 8131, "train_loss": 3.5867300033569336, "perplexity": 36.11578413212412, "lr": 0.0026291804804649314, "grad_norm": 0.15175, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:31.828191+00:00", "epoch": 0, "step": 8132, "train_loss": 3.5276525020599365, "perplexity": 34.043955621810056, "lr": 0.0026291804804649314, "grad_norm": 0.177011, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:32.133957+00:00", "epoch": 0, "step": 8133, "train_loss": 3.549171209335327, "perplexity": 34.78447648828839, "lr": 0.0026291804804649314, "grad_norm": 0.149037, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:32.438547+00:00", "epoch": 0, "step": 8134, "train_loss": 3.5331215858459473, "perplexity": 34.23065493925645, "lr": 0.0026291804804649314, "grad_norm": 0.165636, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:32.742231+00:00", "epoch": 0, "step": 8135, "train_loss": 3.516996145248413, "perplexity": 33.68309721624742, "lr": 0.0026291804804649314, "grad_norm": 0.158648, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:33.045885+00:00", "epoch": 0, "step": 8136, "train_loss": 3.6190402507781982, "perplexity": 37.30175031105597, "lr": 0.0026291804804649314, "grad_norm": 0.154059, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:33.349781+00:00", "epoch": 0, "step": 8137, "train_loss": 3.570852518081665, "perplexity": 35.54688459965834, "lr": 0.0026291804804649314, "grad_norm": 0.166525, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:33.654750+00:00", "epoch": 0, "step": 8138, "train_loss": 3.5191993713378906, "perplexity": 33.757390507214225, "lr": 0.0026291804804649314, "grad_norm": 0.136915, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:33.958417+00:00", "epoch": 0, "step": 8139, "train_loss": 3.641603469848633, "perplexity": 38.15296483250828, "lr": 0.0026291804804649314, "grad_norm": 0.16075, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:34.263201+00:00", "epoch": 0, "step": 8140, "train_loss": 3.537687063217163, "perplexity": 34.38729150806654, "lr": 0.0026291804804649314, "grad_norm": 0.179733, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:34.566786+00:00", "epoch": 0, "step": 8141, "train_loss": 3.5510528087615967, "perplexity": 34.84998855369585, "lr": 0.0026291804804649314, "grad_norm": 0.215002, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:34.872087+00:00", "epoch": 0, "step": 8142, "train_loss": 3.544955015182495, "perplexity": 34.63812711738579, "lr": 0.0026291804804649314, "grad_norm": 0.184303, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:35.176558+00:00", "epoch": 0, "step": 8143, "train_loss": 3.5004968643188477, "perplexity": 33.13190993352474, "lr": 0.0026291804804649314, "grad_norm": 0.163602, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:35.481917+00:00", "epoch": 0, "step": 8144, "train_loss": 3.650351047515869, "perplexity": 38.488174855950355, "lr": 0.0026291804804649314, "grad_norm": 0.200413, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:35.786325+00:00", "epoch": 0, "step": 8145, "train_loss": 3.602047920227051, "perplexity": 36.673261506767766, "lr": 0.0026291804804649314, "grad_norm": 0.183566, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:36.090010+00:00", "epoch": 0, "step": 8146, "train_loss": 3.5276834964752197, "perplexity": 34.045010810660884, "lr": 0.0026291804804649314, "grad_norm": 0.186379, "tokens_per_sec": 107901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:36.393336+00:00", "epoch": 0, "step": 8147, "train_loss": 3.5094335079193115, "perplexity": 33.42932497111587, "lr": 0.0026291804804649314, "grad_norm": 0.151782, "tokens_per_sec": 108029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:36.697382+00:00", "epoch": 0, "step": 8148, "train_loss": 3.4846129417419434, "perplexity": 32.60980277338316, "lr": 0.0026291804804649314, "grad_norm": 0.181759, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:37.002985+00:00", "epoch": 0, "step": 8149, "train_loss": 3.6072611808776855, "perplexity": 36.864947999597454, "lr": 0.0026291804804649314, "grad_norm": 0.157226, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:37.308168+00:00", "epoch": 0, "step": 8150, "train_loss": 3.5681819915771484, "perplexity": 35.45208234449137, "lr": 0.0026291804804649314, "grad_norm": 0.173577, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:37.612102+00:00", "epoch": 0, "step": 8151, "train_loss": 3.479738712310791, "perplexity": 32.451241858026734, "lr": 0.0026291804804649314, "grad_norm": 0.146608, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:37.917311+00:00", "epoch": 0, "step": 8152, "train_loss": 3.5099165439605713, "perplexity": 33.44547644046877, "lr": 0.0026291804804649314, "grad_norm": 0.189923, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:38.221816+00:00", "epoch": 0, "step": 8153, "train_loss": 3.6086254119873047, "perplexity": 36.915274629284575, "lr": 0.0026291804804649314, "grad_norm": 0.177627, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:38.526111+00:00", "epoch": 0, "step": 8154, "train_loss": 3.5676119327545166, "perplexity": 35.4318783314553, "lr": 0.0026291804804649314, "grad_norm": 0.147411, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:38.829845+00:00", "epoch": 0, "step": 8155, "train_loss": 3.5643720626831055, "perplexity": 35.31726940853681, "lr": 0.0026291804804649314, "grad_norm": 0.154907, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:39.134460+00:00", "epoch": 0, "step": 8156, "train_loss": 3.569444417953491, "perplexity": 35.49686625058373, "lr": 0.0026291804804649314, "grad_norm": 0.164765, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:39.438178+00:00", "epoch": 0, "step": 8157, "train_loss": 3.4444665908813477, "perplexity": 31.32656911165608, "lr": 0.0026291804804649314, "grad_norm": 0.176849, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:39.742819+00:00", "epoch": 0, "step": 8158, "train_loss": 3.5715293884277344, "perplexity": 35.570953376540736, "lr": 0.0026291804804649314, "grad_norm": 0.144105, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:40.047450+00:00", "epoch": 0, "step": 8159, "train_loss": 3.55314040184021, "perplexity": 34.92281714036785, "lr": 0.0026291804804649314, "grad_norm": 0.175351, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:40.352808+00:00", "epoch": 0, "step": 8160, "train_loss": 3.493718385696411, "perplexity": 32.90808544210407, "lr": 0.0026291804804649314, "grad_norm": 0.165295, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:40.656561+00:00", "epoch": 0, "step": 8161, "train_loss": 3.487286329269409, "perplexity": 32.69709804844881, "lr": 0.0026291804804649314, "grad_norm": 0.153693, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:40.961570+00:00", "epoch": 0, "step": 8162, "train_loss": 3.525243043899536, "perplexity": 33.96202687670288, "lr": 0.0026291804804649314, "grad_norm": 0.169663, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:41.265633+00:00", "epoch": 0, "step": 8163, "train_loss": 3.6315739154815674, "perplexity": 37.77222013890738, "lr": 0.0026291804804649314, "grad_norm": 0.177235, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:41.571079+00:00", "epoch": 0, "step": 8164, "train_loss": 3.5707836151123047, "perplexity": 35.544435398137516, "lr": 0.0026291804804649314, "grad_norm": 0.147618, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:41.875042+00:00", "epoch": 0, "step": 8165, "train_loss": 3.5887391567230225, "perplexity": 36.18841922446843, "lr": 0.0026291804804649314, "grad_norm": 0.175785, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:42.178347+00:00", "epoch": 0, "step": 8166, "train_loss": 3.5296220779418945, "perplexity": 34.11107385125496, "lr": 0.0026291804804649314, "grad_norm": 0.149755, "tokens_per_sec": 108038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:42.482823+00:00", "epoch": 0, "step": 8167, "train_loss": 3.580411195755005, "perplexity": 35.8882949279759, "lr": 0.0026291804804649314, "grad_norm": 0.159059, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:42.786704+00:00", "epoch": 0, "step": 8168, "train_loss": 3.615968704223633, "perplexity": 37.1873520281745, "lr": 0.0026291804804649314, "grad_norm": 0.173233, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:43.091651+00:00", "epoch": 0, "step": 8169, "train_loss": 3.667560577392578, "perplexity": 39.15627056863616, "lr": 0.0026291804804649314, "grad_norm": 0.177509, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:43.397176+00:00", "epoch": 0, "step": 8170, "train_loss": 3.6304428577423096, "perplexity": 37.72952172872594, "lr": 0.0026291804804649314, "grad_norm": 0.151333, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:43.701352+00:00", "epoch": 0, "step": 8171, "train_loss": 3.5586133003234863, "perplexity": 35.11447014366711, "lr": 0.0026291804804649314, "grad_norm": 0.167846, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:44.005722+00:00", "epoch": 0, "step": 8172, "train_loss": 3.6800098419189453, "perplexity": 39.64678427108968, "lr": 0.0026291804804649314, "grad_norm": 0.149333, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:44.309419+00:00", "epoch": 0, "step": 8173, "train_loss": 3.5200841426849365, "perplexity": 33.78727129596963, "lr": 0.0026291804804649314, "grad_norm": 0.171548, "tokens_per_sec": 107899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:44.613811+00:00", "epoch": 0, "step": 8174, "train_loss": 3.5115625858306885, "perplexity": 33.500574429413724, "lr": 0.0026291804804649314, "grad_norm": 0.174312, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:44.917960+00:00", "epoch": 0, "step": 8175, "train_loss": 3.7452499866485596, "perplexity": 42.31958522788175, "lr": 0.0026291804804649314, "grad_norm": 0.154379, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:45.221296+00:00", "epoch": 0, "step": 8176, "train_loss": 3.5954339504241943, "perplexity": 36.431506025606474, "lr": 0.0026291804804649314, "grad_norm": 0.18253, "tokens_per_sec": 108025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:45.524446+00:00", "epoch": 0, "step": 8177, "train_loss": 3.5899813175201416, "perplexity": 36.23339899040622, "lr": 0.0026291804804649314, "grad_norm": 0.188318, "tokens_per_sec": 108092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:45.827821+00:00", "epoch": 0, "step": 8178, "train_loss": 3.4645941257476807, "perplexity": 31.963483984875605, "lr": 0.0026291804804649314, "grad_norm": 0.191192, "tokens_per_sec": 108012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:46.132437+00:00", "epoch": 0, "step": 8179, "train_loss": 3.487433671951294, "perplexity": 32.701916081507356, "lr": 0.0026291804804649314, "grad_norm": 0.141968, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:46.438887+00:00", "epoch": 0, "step": 8180, "train_loss": 3.566372871398926, "perplexity": 35.38800324783245, "lr": 0.0026291804804649314, "grad_norm": 0.162946, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:46.744033+00:00", "epoch": 0, "step": 8181, "train_loss": 3.4604756832122803, "perplexity": 31.832114916517593, "lr": 0.0026291804804649314, "grad_norm": 0.15107, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:47.048741+00:00", "epoch": 0, "step": 8182, "train_loss": 3.519481897354126, "perplexity": 33.76692919567311, "lr": 0.0026291804804649314, "grad_norm": 0.173711, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:47.351873+00:00", "epoch": 0, "step": 8183, "train_loss": 3.5161728858947754, "perplexity": 33.65537870273295, "lr": 0.0026291804804649314, "grad_norm": 0.145529, "tokens_per_sec": 108098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:47.655977+00:00", "epoch": 0, "step": 8184, "train_loss": 3.5273678302764893, "perplexity": 34.03426564754417, "lr": 0.0026291804804649314, "grad_norm": 0.184588, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:47.960273+00:00", "epoch": 0, "step": 8185, "train_loss": 3.3543238639831543, "perplexity": 28.626242412324448, "lr": 0.0026291804804649314, "grad_norm": 0.166243, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:48.265459+00:00", "epoch": 0, "step": 8186, "train_loss": 3.5490660667419434, "perplexity": 34.78081935048476, "lr": 0.0026291804804649314, "grad_norm": 0.209281, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:48.569169+00:00", "epoch": 0, "step": 8187, "train_loss": 3.5281760692596436, "perplexity": 34.0617845872454, "lr": 0.0026291804804649314, "grad_norm": 0.184662, "tokens_per_sec": 107888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:48.872801+00:00", "epoch": 0, "step": 8188, "train_loss": 3.6146271228790283, "perplexity": 37.137495621128856, "lr": 0.0026291804804649314, "grad_norm": 0.151743, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:49.178479+00:00", "epoch": 0, "step": 8189, "train_loss": 3.528115749359131, "perplexity": 34.059730045753305, "lr": 0.0026291804804649314, "grad_norm": 0.149614, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:49.483691+00:00", "epoch": 0, "step": 8190, "train_loss": 3.5560338497161865, "perplexity": 35.024010820164904, "lr": 0.0026291804804649314, "grad_norm": 0.170306, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:49.788716+00:00", "epoch": 0, "step": 8191, "train_loss": 3.621446371078491, "perplexity": 37.3916108740249, "lr": 0.0026291804804649314, "grad_norm": 0.162355, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:50.093598+00:00", "epoch": 0, "step": 8192, "train_loss": 3.551163911819458, "perplexity": 34.85386070909077, "lr": 0.0026291804804649314, "grad_norm": 0.130252, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:50.397574+00:00", "epoch": 0, "step": 8193, "train_loss": 3.527015447616577, "perplexity": 34.02227467531655, "lr": 0.0026291804804649314, "grad_norm": 0.163563, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:50.701514+00:00", "epoch": 0, "step": 8194, "train_loss": 3.533698797225952, "perplexity": 34.25041896629434, "lr": 0.0026291804804649314, "grad_norm": 0.153162, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:51.005770+00:00", "epoch": 0, "step": 8195, "train_loss": 3.58459734916687, "perplexity": 36.038843726568246, "lr": 0.0026291804804649314, "grad_norm": 0.162253, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:51.310143+00:00", "epoch": 0, "step": 8196, "train_loss": 3.429899215698242, "perplexity": 30.873531025605992, "lr": 0.0026291804804649314, "grad_norm": 0.199674, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:51.614231+00:00", "epoch": 0, "step": 8197, "train_loss": 3.6762490272521973, "perplexity": 39.49796008880799, "lr": 0.0026291804804649314, "grad_norm": 0.149749, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:51.918730+00:00", "epoch": 0, "step": 8198, "train_loss": 3.5491342544555664, "perplexity": 34.78319105589387, "lr": 0.0026291804804649314, "grad_norm": 0.194005, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:52.224023+00:00", "epoch": 0, "step": 8199, "train_loss": 3.511953353881836, "perplexity": 33.513667941692354, "lr": 0.0026291804804649314, "grad_norm": 0.177802, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:52.529216+00:00", "epoch": 0, "step": 8200, "train_loss": 3.4865832328796387, "perplexity": 32.674116916777045, "lr": 0.0026291804804649314, "grad_norm": 0.147085, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:52.834754+00:00", "epoch": 0, "step": 8201, "train_loss": 3.5354065895080566, "perplexity": 34.30896154251519, "lr": 0.0026291804804649314, "grad_norm": 0.173387, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:53.139175+00:00", "epoch": 0, "step": 8202, "train_loss": 3.4811007976531982, "perplexity": 32.49547333558963, "lr": 0.0026291804804649314, "grad_norm": 0.166941, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:53.442654+00:00", "epoch": 0, "step": 8203, "train_loss": 3.5618813037872314, "perplexity": 35.22941206672871, "lr": 0.0026291804804649314, "grad_norm": 0.170473, "tokens_per_sec": 107976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:53.746748+00:00", "epoch": 0, "step": 8204, "train_loss": 3.663006544113159, "perplexity": 38.97835702894119, "lr": 0.0026291804804649314, "grad_norm": 0.16409, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:54.051202+00:00", "epoch": 0, "step": 8205, "train_loss": 3.5504326820373535, "perplexity": 34.82838384397542, "lr": 0.0026291804804649314, "grad_norm": 0.16723, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:54.355466+00:00", "epoch": 0, "step": 8206, "train_loss": 3.6225266456604004, "perplexity": 37.4320259065646, "lr": 0.0026291804804649314, "grad_norm": 0.158092, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:54.659297+00:00", "epoch": 0, "step": 8207, "train_loss": 3.565610647201538, "perplexity": 35.36103993281073, "lr": 0.0026291804804649314, "grad_norm": 0.174755, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:54.964069+00:00", "epoch": 0, "step": 8208, "train_loss": 3.514963388442993, "perplexity": 33.614697214990464, "lr": 0.0026291804804649314, "grad_norm": 0.20281, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:55.269308+00:00", "epoch": 0, "step": 8209, "train_loss": 3.4859862327575684, "perplexity": 32.6546162865077, "lr": 0.0026291804804649314, "grad_norm": 0.188417, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:55.575891+00:00", "epoch": 0, "step": 8210, "train_loss": 3.572964906692505, "perplexity": 35.62205279810945, "lr": 0.0026291804804649314, "grad_norm": 0.153787, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:55.880740+00:00", "epoch": 0, "step": 8211, "train_loss": 3.5685319900512695, "perplexity": 35.46449269089082, "lr": 0.0026291804804649314, "grad_norm": 0.189054, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:56.185151+00:00", "epoch": 0, "step": 8212, "train_loss": 3.576974868774414, "perplexity": 35.76518265997755, "lr": 0.0026291804804649314, "grad_norm": 0.188025, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:56.488757+00:00", "epoch": 0, "step": 8213, "train_loss": 3.726970672607422, "perplexity": 41.55303954793493, "lr": 0.0026291804804649314, "grad_norm": 0.150731, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:56.793445+00:00", "epoch": 0, "step": 8214, "train_loss": 3.570631265640259, "perplexity": 35.53902063464909, "lr": 0.0026291804804649314, "grad_norm": 0.163774, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:57.097808+00:00", "epoch": 0, "step": 8215, "train_loss": 3.5662806034088135, "perplexity": 35.38473821852985, "lr": 0.0026291804804649314, "grad_norm": 0.154147, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:57.401624+00:00", "epoch": 0, "step": 8216, "train_loss": 3.5692081451416016, "perplexity": 35.48848029690687, "lr": 0.0026291804804649314, "grad_norm": 0.173734, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:57.704728+00:00", "epoch": 0, "step": 8217, "train_loss": 3.5839757919311523, "perplexity": 36.01645048254546, "lr": 0.0026291804804649314, "grad_norm": 0.152582, "tokens_per_sec": 108107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:58.010151+00:00", "epoch": 0, "step": 8218, "train_loss": 3.6006977558135986, "perplexity": 36.62377998576733, "lr": 0.0026291804804649314, "grad_norm": 0.169805, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:58.315073+00:00", "epoch": 0, "step": 8219, "train_loss": 3.4431192874908447, "perplexity": 31.28439113852244, "lr": 0.0026291804804649314, "grad_norm": 0.173575, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:58.619836+00:00", "epoch": 0, "step": 8220, "train_loss": 3.66815185546875, "perplexity": 39.17942965902453, "lr": 0.0026291804804649314, "grad_norm": 0.173054, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:58.923993+00:00", "epoch": 0, "step": 8221, "train_loss": 3.5282740592956543, "perplexity": 34.065122466280464, "lr": 0.0026291804804649314, "grad_norm": 0.18128, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:59.228439+00:00", "epoch": 0, "step": 8222, "train_loss": 3.566851854324341, "perplexity": 35.404957557240905, "lr": 0.0026291804804649314, "grad_norm": 0.15552, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:59.533785+00:00", "epoch": 0, "step": 8223, "train_loss": 3.5186355113983154, "perplexity": 33.73836143241327, "lr": 0.0026291804804649314, "grad_norm": 0.20055, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:40:59.839068+00:00", "epoch": 0, "step": 8224, "train_loss": 3.5192553997039795, "perplexity": 33.759281931633986, "lr": 0.0026291804804649314, "grad_norm": 0.174705, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:00.142971+00:00", "epoch": 0, "step": 8225, "train_loss": 3.4683847427368164, "perplexity": 32.084875238821795, "lr": 0.0026291804804649314, "grad_norm": 0.150329, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:00.446554+00:00", "epoch": 0, "step": 8226, "train_loss": 3.5903615951538086, "perplexity": 36.24718036184159, "lr": 0.0026291804804649314, "grad_norm": 0.159143, "tokens_per_sec": 107939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:00.750654+00:00", "epoch": 0, "step": 8227, "train_loss": 3.513890027999878, "perplexity": 33.57863588555706, "lr": 0.0026291804804649314, "grad_norm": 0.151453, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:01.055709+00:00", "epoch": 0, "step": 8228, "train_loss": 3.5976736545562744, "perplexity": 36.5131932636651, "lr": 0.0026291804804649314, "grad_norm": 0.172316, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:01.361041+00:00", "epoch": 0, "step": 8229, "train_loss": 3.571061611175537, "perplexity": 35.55431798484447, "lr": 0.0026291804804649314, "grad_norm": 0.15136, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:01.666746+00:00", "epoch": 0, "step": 8230, "train_loss": 3.5499086380004883, "perplexity": 34.81013701859593, "lr": 0.0026291804804649314, "grad_norm": 0.192825, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:01.971623+00:00", "epoch": 0, "step": 8231, "train_loss": 3.5558278560638428, "perplexity": 35.016796839298976, "lr": 0.0026291804804649314, "grad_norm": 0.16832, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:02.275595+00:00", "epoch": 0, "step": 8232, "train_loss": 3.390566349029541, "perplexity": 29.682758312241123, "lr": 0.0026291804804649314, "grad_norm": 0.197181, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:02.579886+00:00", "epoch": 0, "step": 8233, "train_loss": 3.518343210220337, "perplexity": 33.72850111078553, "lr": 0.0026291804804649314, "grad_norm": 0.156209, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:02.884832+00:00", "epoch": 0, "step": 8234, "train_loss": 3.6024372577667236, "perplexity": 36.68754256407005, "lr": 0.0026291804804649314, "grad_norm": 0.182626, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:03.189344+00:00", "epoch": 0, "step": 8235, "train_loss": 3.566429376602173, "perplexity": 35.390002910643624, "lr": 0.0026291804804649314, "grad_norm": 0.160351, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:03.494369+00:00", "epoch": 0, "step": 8236, "train_loss": 3.536224365234375, "perplexity": 34.33703005376919, "lr": 0.0026291804804649314, "grad_norm": 0.161567, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:03.798478+00:00", "epoch": 0, "step": 8237, "train_loss": 3.6277763843536377, "perplexity": 37.62905097389561, "lr": 0.0026291804804649314, "grad_norm": 0.15331, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:04.103785+00:00", "epoch": 0, "step": 8238, "train_loss": 3.572615385055542, "perplexity": 35.609604295540706, "lr": 0.0026291804804649314, "grad_norm": 0.141061, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:04.408573+00:00", "epoch": 0, "step": 8239, "train_loss": 3.6152634620666504, "perplexity": 37.16113518551563, "lr": 0.0026291804804649314, "grad_norm": 0.147152, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:04.712526+00:00", "epoch": 0, "step": 8240, "train_loss": 3.668468952178955, "perplexity": 39.1918552972375, "lr": 0.0026291804804649314, "grad_norm": 0.154322, "tokens_per_sec": 107896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:05.016845+00:00", "epoch": 0, "step": 8241, "train_loss": 3.6146609783172607, "perplexity": 37.138752948601535, "lr": 0.0026291804804649314, "grad_norm": 0.17232, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:05.320183+00:00", "epoch": 0, "step": 8242, "train_loss": 3.550372362136841, "perplexity": 34.82628306268703, "lr": 0.0026291804804649314, "grad_norm": 0.182881, "tokens_per_sec": 108025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:05.624611+00:00", "epoch": 0, "step": 8243, "train_loss": 3.6914050579071045, "perplexity": 40.101151832685005, "lr": 0.0026291804804649314, "grad_norm": 0.163527, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:05.929982+00:00", "epoch": 0, "step": 8244, "train_loss": 3.6159861087799072, "perplexity": 37.187999263167974, "lr": 0.0026291804804649314, "grad_norm": 0.138591, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:06.234664+00:00", "epoch": 0, "step": 8245, "train_loss": 3.528914451599121, "perplexity": 34.086944495108405, "lr": 0.0026291804804649314, "grad_norm": 0.172864, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:06.538575+00:00", "epoch": 0, "step": 8246, "train_loss": 3.523327350616455, "perplexity": 33.89702832845458, "lr": 0.0026291804804649314, "grad_norm": 0.171192, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:06.843459+00:00", "epoch": 0, "step": 8247, "train_loss": 3.533531904220581, "perplexity": 34.24470328790447, "lr": 0.0026291804804649314, "grad_norm": 0.160992, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:07.147667+00:00", "epoch": 0, "step": 8248, "train_loss": 3.505053758621216, "perplexity": 33.2832330650297, "lr": 0.0026291804804649314, "grad_norm": 0.157216, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:07.452789+00:00", "epoch": 0, "step": 8249, "train_loss": 3.5712597370147705, "perplexity": 35.561362911801574, "lr": 0.0026291804804649314, "grad_norm": 0.161557, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:07.757514+00:00", "epoch": 0, "step": 8250, "train_loss": 3.5022225379943848, "perplexity": 33.18913415928639, "lr": 0.0026291804804649314, "grad_norm": 0.167932, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:08.061799+00:00", "epoch": 0, "step": 8251, "train_loss": 3.519296407699585, "perplexity": 33.76066636050521, "lr": 0.0026291804804649314, "grad_norm": 0.182565, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:08.365380+00:00", "epoch": 0, "step": 8252, "train_loss": 3.6415412425994873, "perplexity": 38.15059075232701, "lr": 0.0026291804804649314, "grad_norm": 0.153322, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:08.669067+00:00", "epoch": 0, "step": 8253, "train_loss": 3.5265250205993652, "perplexity": 34.00559332345594, "lr": 0.0026291804804649314, "grad_norm": 0.167976, "tokens_per_sec": 107901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:08.972814+00:00", "epoch": 0, "step": 8254, "train_loss": 3.536252498626709, "perplexity": 34.33799608449607, "lr": 0.0026291804804649314, "grad_norm": 0.197944, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:09.278440+00:00", "epoch": 0, "step": 8255, "train_loss": 3.5513949394226074, "perplexity": 34.86191384320321, "lr": 0.0026291804804649314, "grad_norm": 0.155478, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:09.582086+00:00", "epoch": 0, "step": 8256, "train_loss": 3.500356912612915, "perplexity": 33.127273390662346, "lr": 0.0026291804804649314, "grad_norm": 0.163194, "tokens_per_sec": 107915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:09.886319+00:00", "epoch": 0, "step": 8257, "train_loss": 3.5073819160461426, "perplexity": 33.36081194409544, "lr": 0.0026291804804649314, "grad_norm": 0.171024, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:10.190627+00:00", "epoch": 0, "step": 8258, "train_loss": 3.572075843811035, "perplexity": 35.59039662745322, "lr": 0.0026291804804649314, "grad_norm": 0.138841, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:10.495663+00:00", "epoch": 0, "step": 8259, "train_loss": 3.551511764526367, "perplexity": 34.865986827814, "lr": 0.0026291804804649314, "grad_norm": 0.158943, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:10.800982+00:00", "epoch": 0, "step": 8260, "train_loss": 3.610748291015625, "perplexity": 36.99372453198442, "lr": 0.0026291804804649314, "grad_norm": 0.174346, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:11.105495+00:00", "epoch": 0, "step": 8261, "train_loss": 3.508204460144043, "perplexity": 33.38826397176509, "lr": 0.0026291804804649314, "grad_norm": 0.17217, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:11.409066+00:00", "epoch": 0, "step": 8262, "train_loss": 3.5136945247650146, "perplexity": 33.57207179529049, "lr": 0.0026291804804649314, "grad_norm": 0.144556, "tokens_per_sec": 107943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:11.713213+00:00", "epoch": 0, "step": 8263, "train_loss": 3.506298303604126, "perplexity": 33.32468133252259, "lr": 0.0026291804804649314, "grad_norm": 0.179031, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:12.017021+00:00", "epoch": 0, "step": 8264, "train_loss": 3.413013458251953, "perplexity": 30.356584865891666, "lr": 0.0026291804804649314, "grad_norm": 0.159972, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:12.321218+00:00", "epoch": 0, "step": 8265, "train_loss": 3.5488903522491455, "perplexity": 34.77470839336109, "lr": 0.0026291804804649314, "grad_norm": 0.184443, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:12.625509+00:00", "epoch": 0, "step": 8266, "train_loss": 3.4617085456848145, "perplexity": 31.871383737973833, "lr": 0.0026291804804649314, "grad_norm": 0.176931, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:12.929720+00:00", "epoch": 0, "step": 8267, "train_loss": 3.5343117713928223, "perplexity": 34.271420024220575, "lr": 0.0026291804804649314, "grad_norm": 0.168231, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:13.235220+00:00", "epoch": 0, "step": 8268, "train_loss": 3.57035756111145, "perplexity": 35.529294774818624, "lr": 0.0026291804804649314, "grad_norm": 0.165283, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:13.539216+00:00", "epoch": 0, "step": 8269, "train_loss": 3.6802453994750977, "perplexity": 39.65612447073599, "lr": 0.0026291804804649314, "grad_norm": 0.169742, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:13.845164+00:00", "epoch": 0, "step": 8270, "train_loss": 3.441157341003418, "perplexity": 31.22307300833099, "lr": 0.0026291804804649314, "grad_norm": 0.168087, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:14.149512+00:00", "epoch": 0, "step": 8271, "train_loss": 3.6158981323242188, "perplexity": 37.18472773870933, "lr": 0.0026291804804649314, "grad_norm": 0.170401, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:14.453156+00:00", "epoch": 0, "step": 8272, "train_loss": 3.5051698684692383, "perplexity": 33.28709780052516, "lr": 0.0026291804804649314, "grad_norm": 0.155954, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:14.757112+00:00", "epoch": 0, "step": 8273, "train_loss": 3.5864028930664062, "perplexity": 36.10397221948845, "lr": 0.0026291804804649314, "grad_norm": 0.15431, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:15.061587+00:00", "epoch": 0, "step": 8274, "train_loss": 3.5601966381073, "perplexity": 35.17011224950407, "lr": 0.0026291804804649314, "grad_norm": 0.149737, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:15.365184+00:00", "epoch": 0, "step": 8275, "train_loss": 3.47784686088562, "perplexity": 32.38990696641377, "lr": 0.0026291804804649314, "grad_norm": 0.163698, "tokens_per_sec": 107934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:15.669058+00:00", "epoch": 0, "step": 8276, "train_loss": 3.4977893829345703, "perplexity": 33.04232723062789, "lr": 0.0026291804804649314, "grad_norm": 0.150185, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:15.973321+00:00", "epoch": 0, "step": 8277, "train_loss": 3.5875463485717773, "perplexity": 36.145279117087625, "lr": 0.0026291804804649314, "grad_norm": 0.175307, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:16.277284+00:00", "epoch": 0, "step": 8278, "train_loss": 3.519601583480835, "perplexity": 33.770970870500776, "lr": 0.0026291804804649314, "grad_norm": 0.159603, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:16.581587+00:00", "epoch": 0, "step": 8279, "train_loss": 3.486450433731079, "perplexity": 32.669778109971865, "lr": 0.0026291804804649314, "grad_norm": 0.182999, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:16.886031+00:00", "epoch": 0, "step": 8280, "train_loss": 3.60624361038208, "perplexity": 36.82745439561886, "lr": 0.0026291804804649314, "grad_norm": 0.172093, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:17.190751+00:00", "epoch": 0, "step": 8281, "train_loss": 3.557375431060791, "perplexity": 35.07102991261086, "lr": 0.0026291804804649314, "grad_norm": 0.160977, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:17.495689+00:00", "epoch": 0, "step": 8282, "train_loss": 3.6138713359832764, "perplexity": 37.10943819265132, "lr": 0.0026291804804649314, "grad_norm": 0.164009, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:17.799460+00:00", "epoch": 0, "step": 8283, "train_loss": 3.6557133197784424, "perplexity": 38.695113262598454, "lr": 0.0026291804804649314, "grad_norm": 0.183255, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:18.103171+00:00", "epoch": 0, "step": 8284, "train_loss": 3.566537380218506, "perplexity": 35.39382536535576, "lr": 0.0026291804804649314, "grad_norm": 0.213428, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:18.406406+00:00", "epoch": 0, "step": 8285, "train_loss": 3.597808837890625, "perplexity": 36.5181295725241, "lr": 0.0026291804804649314, "grad_norm": 0.168391, "tokens_per_sec": 108062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:18.710913+00:00", "epoch": 0, "step": 8286, "train_loss": 3.5896713733673096, "perplexity": 36.22217040045971, "lr": 0.0026291804804649314, "grad_norm": 0.154254, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:19.015177+00:00", "epoch": 0, "step": 8287, "train_loss": 3.6159236431121826, "perplexity": 37.18567636251418, "lr": 0.0026291804804649314, "grad_norm": 0.157285, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:19.319256+00:00", "epoch": 0, "step": 8288, "train_loss": 3.538844585418701, "perplexity": 34.427118607366836, "lr": 0.0026291804804649314, "grad_norm": 0.164193, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:19.623458+00:00", "epoch": 0, "step": 8289, "train_loss": 3.600501537322998, "perplexity": 36.61659442793115, "lr": 0.0026291804804649314, "grad_norm": 0.144338, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:19.928093+00:00", "epoch": 0, "step": 8290, "train_loss": 3.53920578956604, "perplexity": 34.43955607148533, "lr": 0.0026291804804649314, "grad_norm": 0.165962, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:20.232894+00:00", "epoch": 0, "step": 8291, "train_loss": 3.6018357276916504, "perplexity": 36.665480539987975, "lr": 0.0026291804804649314, "grad_norm": 0.158575, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:20.538412+00:00", "epoch": 0, "step": 8292, "train_loss": 3.5079545974731445, "perplexity": 33.37992253310291, "lr": 0.0026291804804649314, "grad_norm": 0.1519, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:20.842806+00:00", "epoch": 0, "step": 8293, "train_loss": 3.4298603534698486, "perplexity": 30.872331234705385, "lr": 0.0026291804804649314, "grad_norm": 0.15333, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:21.146697+00:00", "epoch": 0, "step": 8294, "train_loss": 3.4822752475738525, "perplexity": 32.53366006147948, "lr": 0.0026291804804649314, "grad_norm": 0.157484, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:21.449399+00:00", "epoch": 0, "step": 8295, "train_loss": 3.6456804275512695, "perplexity": 38.30883036906624, "lr": 0.0026291804804649314, "grad_norm": 0.156548, "tokens_per_sec": 108251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:21.754358+00:00", "epoch": 0, "step": 8296, "train_loss": 3.4840893745422363, "perplexity": 32.59273381902433, "lr": 0.0026291804804649314, "grad_norm": 0.153244, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:22.059136+00:00", "epoch": 0, "step": 8297, "train_loss": 3.57716965675354, "perplexity": 35.772149966182674, "lr": 0.0026291804804649314, "grad_norm": 0.159127, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:22.364939+00:00", "epoch": 0, "step": 8298, "train_loss": 3.504509449005127, "perplexity": 33.26512161078439, "lr": 0.0026291804804649314, "grad_norm": 0.156822, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:22.670271+00:00", "epoch": 0, "step": 8299, "train_loss": 3.5297908782958984, "perplexity": 34.116832298596805, "lr": 0.0026291804804649314, "grad_norm": 0.147044, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:22.975970+00:00", "epoch": 0, "step": 8300, "train_loss": 3.585064172744751, "perplexity": 36.055671436019544, "lr": 0.0026291804804649314, "grad_norm": 0.148407, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:23.280493+00:00", "epoch": 0, "step": 8301, "train_loss": 3.5398006439208984, "perplexity": 34.46004868585001, "lr": 0.0026291804804649314, "grad_norm": 0.176672, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:23.585724+00:00", "epoch": 0, "step": 8302, "train_loss": 3.5483670234680176, "perplexity": 34.75651454870002, "lr": 0.0026291804804649314, "grad_norm": 0.185714, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:23.890068+00:00", "epoch": 0, "step": 8303, "train_loss": 3.520512819290161, "perplexity": 33.80175821360808, "lr": 0.0026291804804649314, "grad_norm": 0.153272, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:24.193805+00:00", "epoch": 0, "step": 8304, "train_loss": 3.507920742034912, "perplexity": 33.378792460327055, "lr": 0.0026291804804649314, "grad_norm": 0.212216, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:24.497013+00:00", "epoch": 0, "step": 8305, "train_loss": 3.612595796585083, "perplexity": 37.06213381790231, "lr": 0.0026291804804649314, "grad_norm": 0.17507, "tokens_per_sec": 108071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:24.801918+00:00", "epoch": 0, "step": 8306, "train_loss": 3.4815824031829834, "perplexity": 32.51112710440928, "lr": 0.0026291804804649314, "grad_norm": 0.172051, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:25.107451+00:00", "epoch": 0, "step": 8307, "train_loss": 3.499643325805664, "perplexity": 33.103642637710095, "lr": 0.0026291804804649314, "grad_norm": 0.169726, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:25.412526+00:00", "epoch": 0, "step": 8308, "train_loss": 3.4973597526550293, "perplexity": 33.02813429541878, "lr": 0.0026291804804649314, "grad_norm": 0.176604, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:25.717150+00:00", "epoch": 0, "step": 8309, "train_loss": 3.534769296646118, "perplexity": 34.28710365190224, "lr": 0.0026291804804649314, "grad_norm": 0.170502, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:26.021449+00:00", "epoch": 0, "step": 8310, "train_loss": 3.525995969772339, "perplexity": 33.98760739433949, "lr": 0.0026291804804649314, "grad_norm": 0.171402, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:26.325155+00:00", "epoch": 0, "step": 8311, "train_loss": 3.5356273651123047, "perplexity": 34.31653696043383, "lr": 0.0026291804804649314, "grad_norm": 0.140658, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:26.629935+00:00", "epoch": 0, "step": 8312, "train_loss": 3.502908229827881, "perplexity": 33.21189948164606, "lr": 0.0026291804804649314, "grad_norm": 0.169252, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:26.934502+00:00", "epoch": 0, "step": 8313, "train_loss": 3.591014862060547, "perplexity": 36.27086718129981, "lr": 0.0026291804804649314, "grad_norm": 0.158954, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:27.239612+00:00", "epoch": 0, "step": 8314, "train_loss": 3.6355221271514893, "perplexity": 37.92164765087098, "lr": 0.0026291804804649314, "grad_norm": 0.161684, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:27.543743+00:00", "epoch": 0, "step": 8315, "train_loss": 3.5649120807647705, "perplexity": 35.33634652314227, "lr": 0.0026291804804649314, "grad_norm": 0.156635, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:27.847773+00:00", "epoch": 0, "step": 8316, "train_loss": 3.6112418174743652, "perplexity": 37.011986419840056, "lr": 0.0026291804804649314, "grad_norm": 0.163367, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:28.152992+00:00", "epoch": 0, "step": 8317, "train_loss": 3.4836390018463135, "perplexity": 32.578058246613374, "lr": 0.0026291804804649314, "grad_norm": 0.163358, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:28.458517+00:00", "epoch": 0, "step": 8318, "train_loss": 3.5591862201690674, "perplexity": 35.134593684517, "lr": 0.0026291804804649314, "grad_norm": 0.159209, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:28.763569+00:00", "epoch": 0, "step": 8319, "train_loss": 3.6733999252319336, "perplexity": 39.38558652880601, "lr": 0.0026291804804649314, "grad_norm": 0.196982, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:29.068926+00:00", "epoch": 0, "step": 8320, "train_loss": 3.5035877227783203, "perplexity": 33.2344744020956, "lr": 0.0026291804804649314, "grad_norm": 0.180284, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:29.372357+00:00", "epoch": 0, "step": 8321, "train_loss": 3.5794363021850586, "perplexity": 35.85332470890483, "lr": 0.0026291804804649314, "grad_norm": 0.155656, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:29.676462+00:00", "epoch": 0, "step": 8322, "train_loss": 3.4987423419952393, "perplexity": 33.07383022389527, "lr": 0.0026291804804649314, "grad_norm": 0.173192, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:29.980344+00:00", "epoch": 0, "step": 8323, "train_loss": 3.6415257453918457, "perplexity": 38.14999952928164, "lr": 0.0026291804804649314, "grad_norm": 0.175997, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:30.284628+00:00", "epoch": 0, "step": 8324, "train_loss": 3.6604931354522705, "perplexity": 38.880511503202676, "lr": 0.0026291804804649314, "grad_norm": 0.143205, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:30.588272+00:00", "epoch": 0, "step": 8325, "train_loss": 3.4447104930877686, "perplexity": 31.334210662839727, "lr": 0.0026291804804649314, "grad_norm": 0.171218, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:30.892232+00:00", "epoch": 0, "step": 8326, "train_loss": 3.501197338104248, "perplexity": 33.15512609810149, "lr": 0.0026291804804649314, "grad_norm": 0.164004, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:31.196716+00:00", "epoch": 0, "step": 8327, "train_loss": 3.603003978729248, "perplexity": 36.70834005613534, "lr": 0.0026291804804649314, "grad_norm": 0.148217, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:31.501507+00:00", "epoch": 0, "step": 8328, "train_loss": 3.512145519256592, "perplexity": 33.52010872707996, "lr": 0.0026291804804649314, "grad_norm": 0.154086, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:31.806438+00:00", "epoch": 0, "step": 8329, "train_loss": 3.617504119873047, "perplexity": 37.24449392749605, "lr": 0.0026291804804649314, "grad_norm": 0.149161, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:32.111562+00:00", "epoch": 0, "step": 8330, "train_loss": 3.546912908554077, "perplexity": 34.706011310193205, "lr": 0.0026291804804649314, "grad_norm": 0.152888, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:32.415769+00:00", "epoch": 0, "step": 8331, "train_loss": 3.4961516857147217, "perplexity": 32.98825818962869, "lr": 0.0026291804804649314, "grad_norm": 0.146458, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:32.719085+00:00", "epoch": 0, "step": 8332, "train_loss": 3.542782783508301, "perplexity": 34.56296674295388, "lr": 0.0026291804804649314, "grad_norm": 0.162257, "tokens_per_sec": 108034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:33.022869+00:00", "epoch": 0, "step": 8333, "train_loss": 3.531214952468872, "perplexity": 34.165451809014804, "lr": 0.0026291804804649314, "grad_norm": 0.17184, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:33.327845+00:00", "epoch": 0, "step": 8334, "train_loss": 3.4659781455993652, "perplexity": 32.00775270856944, "lr": 0.0026291804804649314, "grad_norm": 0.160864, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:33.632594+00:00", "epoch": 0, "step": 8335, "train_loss": 3.5604193210601807, "perplexity": 35.177944906020116, "lr": 0.0026291804804649314, "grad_norm": 0.189211, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:33.937155+00:00", "epoch": 0, "step": 8336, "train_loss": 3.4551756381988525, "perplexity": 31.663849575416087, "lr": 0.0026291804804649314, "grad_norm": 0.157916, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:34.242073+00:00", "epoch": 0, "step": 8337, "train_loss": 3.5686159133911133, "perplexity": 35.467469114457266, "lr": 0.0026291804804649314, "grad_norm": 0.177576, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:34.546817+00:00", "epoch": 0, "step": 8338, "train_loss": 3.640547037124634, "perplexity": 38.11268007475647, "lr": 0.0026291804804649314, "grad_norm": 0.16969, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:34.852367+00:00", "epoch": 0, "step": 8339, "train_loss": 3.5665459632873535, "perplexity": 35.39412915429937, "lr": 0.0026291804804649314, "grad_norm": 0.153327, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:35.158406+00:00", "epoch": 0, "step": 8340, "train_loss": 3.539276599884033, "perplexity": 34.44199483374604, "lr": 0.0026291804804649314, "grad_norm": 0.162511, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:35.462343+00:00", "epoch": 0, "step": 8341, "train_loss": 3.4066543579101562, "perplexity": 30.164156779901546, "lr": 0.0026291804804649314, "grad_norm": 0.20102, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:35.766530+00:00", "epoch": 0, "step": 8342, "train_loss": 3.5727365016937256, "perplexity": 35.61391747229349, "lr": 0.0026291804804649314, "grad_norm": 0.164797, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:36.070568+00:00", "epoch": 0, "step": 8343, "train_loss": 3.4426681995391846, "perplexity": 31.270282309005573, "lr": 0.0026291804804649314, "grad_norm": 0.17485, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:36.374932+00:00", "epoch": 0, "step": 8344, "train_loss": 3.4954984188079834, "perplexity": 32.96671508970573, "lr": 0.0026291804804649314, "grad_norm": 0.189114, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:36.680305+00:00", "epoch": 0, "step": 8345, "train_loss": 3.658928394317627, "perplexity": 38.81972114051151, "lr": 0.0026291804804649314, "grad_norm": 0.209186, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:36.985313+00:00", "epoch": 0, "step": 8346, "train_loss": 3.599334716796875, "perplexity": 36.57389435046392, "lr": 0.0026291804804649314, "grad_norm": 0.16569, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:37.289682+00:00", "epoch": 0, "step": 8347, "train_loss": 3.61419677734375, "perplexity": 37.12151710408522, "lr": 0.0026291804804649314, "grad_norm": 0.173165, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:37.594496+00:00", "epoch": 0, "step": 8348, "train_loss": 3.5056955814361572, "perplexity": 33.30460186012241, "lr": 0.0026291804804649314, "grad_norm": 0.164188, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:37.898635+00:00", "epoch": 0, "step": 8349, "train_loss": 3.489926815032959, "perplexity": 32.78354835552887, "lr": 0.0026291804804649314, "grad_norm": 0.159789, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:38.204204+00:00", "epoch": 0, "step": 8350, "train_loss": 3.5449628829956055, "perplexity": 34.63839964476854, "lr": 0.0026291804804649314, "grad_norm": 0.16469, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:38.508455+00:00", "epoch": 0, "step": 8351, "train_loss": 3.616894483566284, "perplexity": 37.22179525144237, "lr": 0.0026291804804649314, "grad_norm": 0.163762, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:38.812308+00:00", "epoch": 0, "step": 8352, "train_loss": 3.6095597743988037, "perplexity": 36.94978299345695, "lr": 0.0026291804804649314, "grad_norm": 0.151294, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:39.116616+00:00", "epoch": 0, "step": 8353, "train_loss": 3.5474812984466553, "perplexity": 34.72574346348063, "lr": 0.0026291804804649314, "grad_norm": 0.157372, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:39.420498+00:00", "epoch": 0, "step": 8354, "train_loss": 3.5927655696868896, "perplexity": 36.33442248223024, "lr": 0.0026291804804649314, "grad_norm": 0.131717, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:39.724198+00:00", "epoch": 0, "step": 8355, "train_loss": 3.4780523777008057, "perplexity": 32.396564321013386, "lr": 0.0026291804804649314, "grad_norm": 0.154162, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:40.030282+00:00", "epoch": 0, "step": 8356, "train_loss": 3.557356834411621, "perplexity": 35.07037771503591, "lr": 0.0026291804804649314, "grad_norm": 0.16655, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:40.334337+00:00", "epoch": 0, "step": 8357, "train_loss": 3.5405685901641846, "perplexity": 34.486522314642606, "lr": 0.0026291804804649314, "grad_norm": 0.17138, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:40.638405+00:00", "epoch": 0, "step": 8358, "train_loss": 3.6311280727386475, "perplexity": 37.75538342221734, "lr": 0.0026291804804649314, "grad_norm": 0.153356, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:40.943391+00:00", "epoch": 0, "step": 8359, "train_loss": 3.5145440101623535, "perplexity": 33.60060289670028, "lr": 0.0026291804804649314, "grad_norm": 0.150976, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:41.247499+00:00", "epoch": 0, "step": 8360, "train_loss": 3.6494150161743164, "perplexity": 38.45216557354363, "lr": 0.0026291804804649314, "grad_norm": 0.163579, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:41.550877+00:00", "epoch": 0, "step": 8361, "train_loss": 3.5340421199798584, "perplexity": 34.262179933244546, "lr": 0.0026291804804649314, "grad_norm": 0.161266, "tokens_per_sec": 107943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:41.855422+00:00", "epoch": 0, "step": 8362, "train_loss": 3.537672519683838, "perplexity": 34.38679139898321, "lr": 0.0026291804804649314, "grad_norm": 0.163191, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:42.159472+00:00", "epoch": 0, "step": 8363, "train_loss": 3.609280824661255, "perplexity": 36.939477298640696, "lr": 0.0026291804804649314, "grad_norm": 0.171177, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:42.464036+00:00", "epoch": 0, "step": 8364, "train_loss": 3.655675172805786, "perplexity": 38.69363718932494, "lr": 0.0026291804804649314, "grad_norm": 0.197501, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:42.768149+00:00", "epoch": 0, "step": 8365, "train_loss": 3.5833706855773926, "perplexity": 35.994663291967115, "lr": 0.0026291804804649314, "grad_norm": 0.195707, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:43.073052+00:00", "epoch": 0, "step": 8366, "train_loss": 3.6235296726226807, "perplexity": 37.4695900735885, "lr": 0.0026291804804649314, "grad_norm": 0.173334, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:43.377340+00:00", "epoch": 0, "step": 8367, "train_loss": 3.5594043731689453, "perplexity": 35.14225923763007, "lr": 0.0026291804804649314, "grad_norm": 0.164933, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:43.682204+00:00", "epoch": 0, "step": 8368, "train_loss": 3.5875253677368164, "perplexity": 36.14452076690729, "lr": 0.0026291804804649314, "grad_norm": 0.170466, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:43.986430+00:00", "epoch": 0, "step": 8369, "train_loss": 3.459089517593384, "perplexity": 31.788020901108023, "lr": 0.0026291804804649314, "grad_norm": 0.165938, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:44.290994+00:00", "epoch": 0, "step": 8370, "train_loss": 3.5657763481140137, "perplexity": 35.366899774870866, "lr": 0.0026291804804649314, "grad_norm": 0.175111, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:44.595172+00:00", "epoch": 0, "step": 8371, "train_loss": 3.7429769039154053, "perplexity": 42.22349855727448, "lr": 0.0026291804804649314, "grad_norm": 0.181932, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:44.900871+00:00", "epoch": 0, "step": 8372, "train_loss": 3.5298731327056885, "perplexity": 34.11963867391822, "lr": 0.0026291804804649314, "grad_norm": 0.153764, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:45.205578+00:00", "epoch": 0, "step": 8373, "train_loss": 3.5122082233428955, "perplexity": 33.52221064076909, "lr": 0.0026291804804649314, "grad_norm": 0.166928, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:45.509735+00:00", "epoch": 0, "step": 8374, "train_loss": 3.4620323181152344, "perplexity": 31.881704484043883, "lr": 0.0026291804804649314, "grad_norm": 0.152536, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:45.813444+00:00", "epoch": 0, "step": 8375, "train_loss": 3.529508113861084, "perplexity": 34.10718663558372, "lr": 0.0026291804804649314, "grad_norm": 0.155864, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:46.118085+00:00", "epoch": 0, "step": 8376, "train_loss": 3.4482624530792236, "perplexity": 31.44570642249363, "lr": 0.0026291804804649314, "grad_norm": 0.158737, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:46.423019+00:00", "epoch": 0, "step": 8377, "train_loss": 3.5567266941070557, "perplexity": 35.048285417895784, "lr": 0.0026291804804649314, "grad_norm": 0.159163, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:46.727531+00:00", "epoch": 0, "step": 8378, "train_loss": 3.602431297302246, "perplexity": 36.68732388992753, "lr": 0.0026291804804649314, "grad_norm": 0.163047, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:47.033202+00:00", "epoch": 0, "step": 8379, "train_loss": 3.533963918685913, "perplexity": 34.25950069120184, "lr": 0.0026291804804649314, "grad_norm": 0.173128, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:47.337267+00:00", "epoch": 0, "step": 8380, "train_loss": 3.586162567138672, "perplexity": 36.09529654140685, "lr": 0.0026291804804649314, "grad_norm": 0.176201, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:47.641287+00:00", "epoch": 0, "step": 8381, "train_loss": 3.524864435195923, "perplexity": 33.949170991563584, "lr": 0.0026291804804649314, "grad_norm": 0.141694, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:47.946479+00:00", "epoch": 0, "step": 8382, "train_loss": 3.574592113494873, "perplexity": 35.68006443039791, "lr": 0.0026291804804649314, "grad_norm": 0.17037, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:48.250850+00:00", "epoch": 0, "step": 8383, "train_loss": 3.5641441345214844, "perplexity": 35.309220525565436, "lr": 0.0026291804804649314, "grad_norm": 0.144941, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:48.554294+00:00", "epoch": 0, "step": 8384, "train_loss": 3.4388034343719482, "perplexity": 31.14966324348416, "lr": 0.0026291804804649314, "grad_norm": 0.155663, "tokens_per_sec": 107987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:48.858005+00:00", "epoch": 0, "step": 8385, "train_loss": 3.5074214935302734, "perplexity": 33.3621323072289, "lr": 0.0026291804804649314, "grad_norm": 0.16566, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:49.163348+00:00", "epoch": 0, "step": 8386, "train_loss": 3.5056841373443604, "perplexity": 33.304220721382364, "lr": 0.0026291804804649314, "grad_norm": 0.166994, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:49.469123+00:00", "epoch": 0, "step": 8387, "train_loss": 3.569329023361206, "perplexity": 35.49277034050282, "lr": 0.0026291804804649314, "grad_norm": 0.160668, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:49.774434+00:00", "epoch": 0, "step": 8388, "train_loss": 3.455770254135132, "perplexity": 31.682683003750984, "lr": 0.0026291804804649314, "grad_norm": 0.158502, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:50.079190+00:00", "epoch": 0, "step": 8389, "train_loss": 3.4346375465393066, "perplexity": 31.020167160744545, "lr": 0.0026291804804649314, "grad_norm": 0.172918, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:50.382662+00:00", "epoch": 0, "step": 8390, "train_loss": 3.6024844646453857, "perplexity": 36.68927450931982, "lr": 0.0026291804804649314, "grad_norm": 0.163118, "tokens_per_sec": 108037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:50.686117+00:00", "epoch": 0, "step": 8391, "train_loss": 3.5177249908447266, "perplexity": 33.70765594199607, "lr": 0.0026291804804649314, "grad_norm": 0.170409, "tokens_per_sec": 107926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:50.990276+00:00", "epoch": 0, "step": 8392, "train_loss": 3.586132526397705, "perplexity": 36.094212228240195, "lr": 0.0026291804804649314, "grad_norm": 0.150702, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:51.295327+00:00", "epoch": 0, "step": 8393, "train_loss": 3.5801520347595215, "perplexity": 35.87899528684034, "lr": 0.0026291804804649314, "grad_norm": 0.1887, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:51.601013+00:00", "epoch": 0, "step": 8394, "train_loss": 3.4276535511016846, "perplexity": 30.804277219525275, "lr": 0.0026291804804649314, "grad_norm": 0.159899, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:51.905520+00:00", "epoch": 0, "step": 8395, "train_loss": 3.551126003265381, "perplexity": 34.852539474670486, "lr": 0.0026291804804649314, "grad_norm": 0.162913, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:52.209854+00:00", "epoch": 0, "step": 8396, "train_loss": 3.5392985343933105, "perplexity": 34.442750310286726, "lr": 0.0026291804804649314, "grad_norm": 0.162459, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:52.514325+00:00", "epoch": 0, "step": 8397, "train_loss": 3.516106128692627, "perplexity": 33.65313203880484, "lr": 0.0026291804804649314, "grad_norm": 0.176196, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:52.818547+00:00", "epoch": 0, "step": 8398, "train_loss": 3.5952351093292236, "perplexity": 36.42426266521963, "lr": 0.0026291804804649314, "grad_norm": 0.152765, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:53.124048+00:00", "epoch": 0, "step": 8399, "train_loss": 3.4817183017730713, "perplexity": 32.51554562097323, "lr": 0.0026291804804649314, "grad_norm": 0.172311, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:53.428283+00:00", "epoch": 0, "step": 8400, "train_loss": 3.5449881553649902, "perplexity": 34.639275050261, "lr": 0.0026291804804649314, "grad_norm": 0.150211, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:53.732223+00:00", "epoch": 0, "step": 8401, "train_loss": 3.545470952987671, "perplexity": 34.656002847652125, "lr": 0.0026291804804649314, "grad_norm": 0.161607, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:54.035435+00:00", "epoch": 0, "step": 8402, "train_loss": 3.4992191791534424, "perplexity": 33.08960481576701, "lr": 0.0026291804804649314, "grad_norm": 0.145757, "tokens_per_sec": 108070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:54.340012+00:00", "epoch": 0, "step": 8403, "train_loss": 3.505448818206787, "perplexity": 33.29638452292455, "lr": 0.0026291804804649314, "grad_norm": 0.155212, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:54.644004+00:00", "epoch": 0, "step": 8404, "train_loss": 3.5518081188201904, "perplexity": 34.876321043937835, "lr": 0.0026291804804649314, "grad_norm": 0.182203, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:54.948160+00:00", "epoch": 0, "step": 8405, "train_loss": 3.5050809383392334, "perplexity": 33.28413770621301, "lr": 0.0026291804804649314, "grad_norm": 0.198296, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:55.253038+00:00", "epoch": 0, "step": 8406, "train_loss": 3.626487970352173, "perplexity": 37.580600396662426, "lr": 0.0026291804804649314, "grad_norm": 0.175012, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:55.557046+00:00", "epoch": 0, "step": 8407, "train_loss": 3.6076393127441406, "perplexity": 36.87889044706703, "lr": 0.0026291804804649314, "grad_norm": 0.152728, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:55.862384+00:00", "epoch": 0, "step": 8408, "train_loss": 3.542177438735962, "perplexity": 34.54205056310829, "lr": 0.0026291804804649314, "grad_norm": 0.186215, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:56.167392+00:00", "epoch": 0, "step": 8409, "train_loss": 3.5289173126220703, "perplexity": 34.087042018778384, "lr": 0.0026291804804649314, "grad_norm": 0.168219, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:56.471815+00:00", "epoch": 0, "step": 8410, "train_loss": 3.455580949783325, "perplexity": 31.676685901638137, "lr": 0.0026291804804649314, "grad_norm": 0.151238, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:56.775797+00:00", "epoch": 0, "step": 8411, "train_loss": 3.496150493621826, "perplexity": 32.98821886458391, "lr": 0.0026291804804649314, "grad_norm": 0.166703, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:57.079669+00:00", "epoch": 0, "step": 8412, "train_loss": 3.5868992805480957, "perplexity": 36.12189822809252, "lr": 0.0026291804804649314, "grad_norm": 0.173792, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:57.385041+00:00", "epoch": 0, "step": 8413, "train_loss": 3.552029848098755, "perplexity": 34.88405500283277, "lr": 0.0026291804804649314, "grad_norm": 0.135767, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:57.689083+00:00", "epoch": 0, "step": 8414, "train_loss": 3.5483694076538086, "perplexity": 34.75659741478694, "lr": 0.0026291804804649314, "grad_norm": 0.182315, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:57.994428+00:00", "epoch": 0, "step": 8415, "train_loss": 3.5089640617370605, "perplexity": 33.41363538512921, "lr": 0.0026291804804649314, "grad_norm": 0.188141, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:58.299475+00:00", "epoch": 0, "step": 8416, "train_loss": 3.5732431411743164, "perplexity": 35.63196546046908, "lr": 0.0026291804804649314, "grad_norm": 0.157851, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:58.604046+00:00", "epoch": 0, "step": 8417, "train_loss": 3.619654417037964, "perplexity": 37.324666824081426, "lr": 0.0026291804804649314, "grad_norm": 0.213634, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:58.910589+00:00", "epoch": 0, "step": 8418, "train_loss": 3.610579490661621, "perplexity": 36.98748050519929, "lr": 0.0026291804804649314, "grad_norm": 0.184738, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:59.215374+00:00", "epoch": 0, "step": 8419, "train_loss": 3.5081913471221924, "perplexity": 33.38782615360064, "lr": 0.0026291804804649314, "grad_norm": 0.152412, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:59.519264+00:00", "epoch": 0, "step": 8420, "train_loss": 3.5682735443115234, "perplexity": 35.45532822815179, "lr": 0.0026291804804649314, "grad_norm": 0.158546, "tokens_per_sec": 107887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:41:59.823467+00:00", "epoch": 0, "step": 8421, "train_loss": 3.611409902572632, "perplexity": 37.0182081060862, "lr": 0.0026291804804649314, "grad_norm": 0.159451, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:00.127436+00:00", "epoch": 0, "step": 8422, "train_loss": 3.5741817951202393, "perplexity": 35.66542724751191, "lr": 0.0026291804804649314, "grad_norm": 0.157642, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:00.431876+00:00", "epoch": 0, "step": 8423, "train_loss": 3.5525271892547607, "perplexity": 34.90140859404391, "lr": 0.0026291804804649314, "grad_norm": 0.174152, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:00.735966+00:00", "epoch": 0, "step": 8424, "train_loss": 3.4914393424987793, "perplexity": 32.83317189186401, "lr": 0.0026291804804649314, "grad_norm": 0.14139, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:01.040182+00:00", "epoch": 0, "step": 8425, "train_loss": 3.566288471221924, "perplexity": 35.38501662013231, "lr": 0.0026291804804649314, "grad_norm": 0.172518, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:01.344635+00:00", "epoch": 0, "step": 8426, "train_loss": 3.563617706298828, "perplexity": 35.290637647065786, "lr": 0.0026291804804649314, "grad_norm": 0.17449, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:01.650966+00:00", "epoch": 0, "step": 8427, "train_loss": 3.5228757858276367, "perplexity": 33.88172507948011, "lr": 0.0026291804804649314, "grad_norm": 0.164673, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:01.955848+00:00", "epoch": 0, "step": 8428, "train_loss": 3.519906997680664, "perplexity": 33.781286579748496, "lr": 0.0026291804804649314, "grad_norm": 0.176163, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:02.260725+00:00", "epoch": 0, "step": 8429, "train_loss": 3.48153018951416, "perplexity": 32.50942962350182, "lr": 0.0026291804804649314, "grad_norm": 0.164112, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:02.565064+00:00", "epoch": 0, "step": 8430, "train_loss": 3.6708028316497803, "perplexity": 39.283431185714576, "lr": 0.0026291804804649314, "grad_norm": 0.159198, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:02.868718+00:00", "epoch": 0, "step": 8431, "train_loss": 3.569485902786255, "perplexity": 35.49833886268909, "lr": 0.0026291804804649314, "grad_norm": 0.149827, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:03.172633+00:00", "epoch": 0, "step": 8432, "train_loss": 3.523179054260254, "perplexity": 33.89200189537749, "lr": 0.0026291804804649314, "grad_norm": 0.150903, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:03.478462+00:00", "epoch": 0, "step": 8433, "train_loss": 3.5225017070770264, "perplexity": 33.86905301641639, "lr": 0.0026291804804649314, "grad_norm": 0.130804, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:03.783653+00:00", "epoch": 0, "step": 8434, "train_loss": 3.55914568901062, "perplexity": 35.133169667592114, "lr": 0.0026291804804649314, "grad_norm": 0.143657, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:04.088946+00:00", "epoch": 0, "step": 8435, "train_loss": 3.632462978363037, "perplexity": 37.80581695040961, "lr": 0.0026291804804649314, "grad_norm": 0.178672, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:04.392894+00:00", "epoch": 0, "step": 8436, "train_loss": 3.4562876224517822, "perplexity": 31.69907886110644, "lr": 0.0026291804804649314, "grad_norm": 0.161683, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:04.698106+00:00", "epoch": 0, "step": 8437, "train_loss": 3.5759658813476562, "perplexity": 35.729114239709176, "lr": 0.0026291804804649314, "grad_norm": 0.138523, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:05.004152+00:00", "epoch": 0, "step": 8438, "train_loss": 3.56882381439209, "perplexity": 35.474843603343515, "lr": 0.0026291804804649314, "grad_norm": 0.177067, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:05.307990+00:00", "epoch": 0, "step": 8439, "train_loss": 3.5814876556396484, "perplexity": 35.92694803832298, "lr": 0.0026291804804649314, "grad_norm": 0.158407, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:05.612172+00:00", "epoch": 0, "step": 8440, "train_loss": 3.519998788833618, "perplexity": 33.784387545310345, "lr": 0.0026291804804649314, "grad_norm": 0.17025, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:05.916157+00:00", "epoch": 0, "step": 8441, "train_loss": 3.604261875152588, "perplexity": 36.75454439984561, "lr": 0.0026291804804649314, "grad_norm": 0.151924, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:06.221863+00:00", "epoch": 0, "step": 8442, "train_loss": 3.5780277252197266, "perplexity": 35.80285809298983, "lr": 0.0026291804804649314, "grad_norm": 0.158133, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:06.525495+00:00", "epoch": 0, "step": 8443, "train_loss": 3.5307226181030273, "perplexity": 34.14863512302052, "lr": 0.0026291804804649314, "grad_norm": 0.155914, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:06.829713+00:00", "epoch": 0, "step": 8444, "train_loss": 3.5480501651763916, "perplexity": 34.74550340345571, "lr": 0.0026291804804649314, "grad_norm": 0.16523, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:07.133933+00:00", "epoch": 0, "step": 8445, "train_loss": 3.594327211380005, "perplexity": 36.39120815917371, "lr": 0.0026291804804649314, "grad_norm": 0.16616, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:07.439153+00:00", "epoch": 0, "step": 8446, "train_loss": 3.5038952827453613, "perplexity": 33.244697567985035, "lr": 0.0026291804804649314, "grad_norm": 0.170196, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:07.745062+00:00", "epoch": 0, "step": 8447, "train_loss": 3.56577205657959, "perplexity": 35.3667479969287, "lr": 0.0026291804804649314, "grad_norm": 0.208068, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:08.049922+00:00", "epoch": 0, "step": 8448, "train_loss": 3.602491617202759, "perplexity": 36.68953693239923, "lr": 0.0026291804804649314, "grad_norm": 0.150375, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:08.354273+00:00", "epoch": 0, "step": 8449, "train_loss": 3.5423035621643066, "perplexity": 34.54640739969118, "lr": 0.0026291804804649314, "grad_norm": 0.170956, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:08.659070+00:00", "epoch": 0, "step": 8450, "train_loss": 3.526632308959961, "perplexity": 34.00924192353736, "lr": 0.0026291804804649314, "grad_norm": 0.191061, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:08.965047+00:00", "epoch": 0, "step": 8451, "train_loss": 3.5460402965545654, "perplexity": 34.675739637901856, "lr": 0.0026291804804649314, "grad_norm": 0.186999, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:09.269562+00:00", "epoch": 0, "step": 8452, "train_loss": 3.4725899696350098, "perplexity": 32.22008351068099, "lr": 0.0026291804804649314, "grad_norm": 0.184295, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:09.573222+00:00", "epoch": 0, "step": 8453, "train_loss": 3.6118340492248535, "perplexity": 37.03391258539136, "lr": 0.0026291804804649314, "grad_norm": 0.161727, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:09.877472+00:00", "epoch": 0, "step": 8454, "train_loss": 3.5615084171295166, "perplexity": 35.21627793793277, "lr": 0.0026291804804649314, "grad_norm": 0.194011, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:10.183799+00:00", "epoch": 0, "step": 8455, "train_loss": 3.469681978225708, "perplexity": 32.126523885849146, "lr": 0.0026291804804649314, "grad_norm": 0.15492, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:10.489162+00:00", "epoch": 0, "step": 8456, "train_loss": 3.4819085597991943, "perplexity": 32.52173255303946, "lr": 0.0026291804804649314, "grad_norm": 0.174365, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:10.795675+00:00", "epoch": 0, "step": 8457, "train_loss": 3.5224571228027344, "perplexity": 33.867543022927926, "lr": 0.0026291804804649314, "grad_norm": 0.162418, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:11.101997+00:00", "epoch": 0, "step": 8458, "train_loss": 3.4739179611206055, "perplexity": 32.26289993087664, "lr": 0.0026291804804649314, "grad_norm": 0.191026, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:11.406131+00:00", "epoch": 0, "step": 8459, "train_loss": 3.5039455890655518, "perplexity": 33.24637002845284, "lr": 0.0026291804804649314, "grad_norm": 0.166943, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:11.710630+00:00", "epoch": 0, "step": 8460, "train_loss": 3.6244008541107178, "perplexity": 37.50224710986224, "lr": 0.0026291804804649314, "grad_norm": 0.157458, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:12.015394+00:00", "epoch": 0, "step": 8461, "train_loss": 3.5204455852508545, "perplexity": 33.79948566126501, "lr": 0.0026291804804649314, "grad_norm": 0.170753, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:12.320143+00:00", "epoch": 0, "step": 8462, "train_loss": 3.547023057937622, "perplexity": 34.70983436649411, "lr": 0.0026291804804649314, "grad_norm": 0.15836, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:12.624713+00:00", "epoch": 0, "step": 8463, "train_loss": 3.5012333393096924, "perplexity": 33.1563197440939, "lr": 0.0026291804804649314, "grad_norm": 0.166764, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:12.928692+00:00", "epoch": 0, "step": 8464, "train_loss": 3.5127220153808594, "perplexity": 33.539438511083574, "lr": 0.0026291804804649314, "grad_norm": 0.176916, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:13.233328+00:00", "epoch": 0, "step": 8465, "train_loss": 3.559335708618164, "perplexity": 35.13984629302891, "lr": 0.0026291804804649314, "grad_norm": 0.153004, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:13.538545+00:00", "epoch": 0, "step": 8466, "train_loss": 3.567992925643921, "perplexity": 35.4453801970519, "lr": 0.0026291804804649314, "grad_norm": 0.142143, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:13.844166+00:00", "epoch": 0, "step": 8467, "train_loss": 3.5052127838134766, "perplexity": 33.28852635843927, "lr": 0.0026291804804649314, "grad_norm": 0.156405, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:14.149319+00:00", "epoch": 0, "step": 8468, "train_loss": 3.525956630706787, "perplexity": 33.98627037992288, "lr": 0.0026291804804649314, "grad_norm": 0.144831, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:14.453686+00:00", "epoch": 0, "step": 8469, "train_loss": 3.536273241043091, "perplexity": 34.33870834489555, "lr": 0.0026291804804649314, "grad_norm": 0.153414, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:14.757648+00:00", "epoch": 0, "step": 8470, "train_loss": 3.516700506210327, "perplexity": 33.673140649634064, "lr": 0.0026291804804649314, "grad_norm": 0.15604, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:15.061295+00:00", "epoch": 0, "step": 8471, "train_loss": 3.598806858062744, "perplexity": 36.55459359537407, "lr": 0.0026291804804649314, "grad_norm": 0.168222, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:15.366259+00:00", "epoch": 0, "step": 8472, "train_loss": 3.605679512023926, "perplexity": 36.8066859473338, "lr": 0.0026291804804649314, "grad_norm": 0.164867, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:15.671097+00:00", "epoch": 0, "step": 8473, "train_loss": 3.563781261444092, "perplexity": 35.29641008447565, "lr": 0.0026291804804649314, "grad_norm": 0.177671, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:15.976484+00:00", "epoch": 0, "step": 8474, "train_loss": 3.6141443252563477, "perplexity": 37.119570054089436, "lr": 0.0026291804804649314, "grad_norm": 0.194999, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:16.282156+00:00", "epoch": 0, "step": 8475, "train_loss": 3.5031790733337402, "perplexity": 33.22089592719735, "lr": 0.0026291804804649314, "grad_norm": 0.152539, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:16.586449+00:00", "epoch": 0, "step": 8476, "train_loss": 3.518272638320923, "perplexity": 33.72612091038637, "lr": 0.0026291804804649314, "grad_norm": 0.170059, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:16.892043+00:00", "epoch": 0, "step": 8477, "train_loss": 3.4786767959594727, "perplexity": 32.416799644288695, "lr": 0.0026291804804649314, "grad_norm": 0.147719, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:17.197135+00:00", "epoch": 0, "step": 8478, "train_loss": 3.556959867477417, "perplexity": 35.05645869759062, "lr": 0.0026291804804649314, "grad_norm": 0.140576, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:17.503541+00:00", "epoch": 0, "step": 8479, "train_loss": 3.5444509983062744, "perplexity": 34.62067331563276, "lr": 0.0026291804804649314, "grad_norm": 0.169102, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:17.807548+00:00", "epoch": 0, "step": 8480, "train_loss": 3.5099220275878906, "perplexity": 33.44565984349994, "lr": 0.0026291804804649314, "grad_norm": 0.184751, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:18.111784+00:00", "epoch": 0, "step": 8481, "train_loss": 3.4759809970855713, "perplexity": 32.32952815835145, "lr": 0.0026291804804649314, "grad_norm": 0.190876, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:18.416206+00:00", "epoch": 0, "step": 8482, "train_loss": 3.5505530834198, "perplexity": 34.832577481993454, "lr": 0.0026291804804649314, "grad_norm": 0.164472, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:18.721286+00:00", "epoch": 0, "step": 8483, "train_loss": 3.479198455810547, "perplexity": 32.433714598706146, "lr": 0.0026291804804649314, "grad_norm": 0.157067, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:19.026737+00:00", "epoch": 0, "step": 8484, "train_loss": 3.5060646533966064, "perplexity": 33.316895923380365, "lr": 0.0026291804804649314, "grad_norm": 0.163142, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:19.331721+00:00", "epoch": 0, "step": 8485, "train_loss": 3.5334725379943848, "perplexity": 34.242670369447005, "lr": 0.0026291804804649314, "grad_norm": 0.171709, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:19.635841+00:00", "epoch": 0, "step": 8486, "train_loss": 3.551206350326538, "perplexity": 34.85533988629206, "lr": 0.0026291804804649314, "grad_norm": 0.156915, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:19.941187+00:00", "epoch": 0, "step": 8487, "train_loss": 3.5664918422698975, "perplexity": 35.39221363985283, "lr": 0.0026291804804649314, "grad_norm": 0.167089, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:20.245905+00:00", "epoch": 0, "step": 8488, "train_loss": 3.562656879425049, "perplexity": 35.25674573875722, "lr": 0.0026291804804649314, "grad_norm": 0.160916, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:20.550925+00:00", "epoch": 0, "step": 8489, "train_loss": 3.5044448375701904, "perplexity": 33.2629723729772, "lr": 0.0026291804804649314, "grad_norm": 0.157361, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:20.856544+00:00", "epoch": 0, "step": 8490, "train_loss": 3.586853265762329, "perplexity": 36.12023612492502, "lr": 0.0026291804804649314, "grad_norm": 0.169903, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:21.160239+00:00", "epoch": 0, "step": 8491, "train_loss": 3.5233724117279053, "perplexity": 33.89855580064046, "lr": 0.0026291804804649314, "grad_norm": 0.175023, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:21.464616+00:00", "epoch": 0, "step": 8492, "train_loss": 3.633334159851074, "perplexity": 37.83876702894217, "lr": 0.0026291804804649314, "grad_norm": 0.157061, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:21.768693+00:00", "epoch": 0, "step": 8493, "train_loss": 3.6235804557800293, "perplexity": 37.471492945993525, "lr": 0.0026291804804649314, "grad_norm": 0.166988, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:22.074160+00:00", "epoch": 0, "step": 8494, "train_loss": 3.4678955078125, "perplexity": 32.06918203645698, "lr": 0.0026291804804649314, "grad_norm": 0.162718, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:22.379171+00:00", "epoch": 0, "step": 8495, "train_loss": 3.487138271331787, "perplexity": 32.69225734190689, "lr": 0.0026291804804649314, "grad_norm": 0.155799, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:22.683814+00:00", "epoch": 0, "step": 8496, "train_loss": 3.592519760131836, "perplexity": 36.32549223162218, "lr": 0.0026291804804649314, "grad_norm": 0.158998, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:22.989753+00:00", "epoch": 0, "step": 8497, "train_loss": 3.588261365890503, "perplexity": 36.171132859481816, "lr": 0.0026291804804649314, "grad_norm": 0.152618, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:23.294755+00:00", "epoch": 0, "step": 8498, "train_loss": 3.6131479740142822, "perplexity": 37.0826043428353, "lr": 0.0026291804804649314, "grad_norm": 0.165283, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:23.599502+00:00", "epoch": 0, "step": 8499, "train_loss": 3.6611273288726807, "perplexity": 38.905177088329374, "lr": 0.0026291804804649314, "grad_norm": 0.159898, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:23.904540+00:00", "epoch": 0, "step": 8500, "train_loss": 3.49049973487854, "perplexity": 32.80233608240721, "lr": 0.0026291804804649314, "grad_norm": 0.153486, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:42:26.978249+00:00", "step": 8500, "epoch": 0, "val_loss": 3.509123992919922, "val_ppl": 33.41897969470954, "eval_train_loss": 3.49049973487854, "eval_train_ppl": 32.80233608240721} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:42:27.921415+00:00", "step": 8500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5091_epoch_0000_step_0008500.pt", "val_loss": 3.509123992919922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:29.025382+00:00", "epoch": 0, "step": 8501, "train_loss": 3.5720441341400146, "perplexity": 35.589268085577615, "lr": 0.0026291804804649314, "grad_norm": 0.162994, "tokens_per_sec": 6399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:29.328479+00:00", "epoch": 0, "step": 8502, "train_loss": 3.594905138015747, "perplexity": 36.41224568616379, "lr": 0.0026291804804649314, "grad_norm": 0.155682, "tokens_per_sec": 108104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:29.632128+00:00", "epoch": 0, "step": 8503, "train_loss": 3.5353713035583496, "perplexity": 34.30775093958246, "lr": 0.0026291804804649314, "grad_norm": 0.158606, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:29.936275+00:00", "epoch": 0, "step": 8504, "train_loss": 3.5696322917938232, "perplexity": 35.5035358096642, "lr": 0.0026291804804649314, "grad_norm": 0.149801, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:30.239009+00:00", "epoch": 0, "step": 8505, "train_loss": 3.529801845550537, "perplexity": 34.11720646863589, "lr": 0.0026291804804649314, "grad_norm": 0.16606, "tokens_per_sec": 108240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:30.541568+00:00", "epoch": 0, "step": 8506, "train_loss": 3.4508397579193115, "perplexity": 31.526856122694117, "lr": 0.0026291804804649314, "grad_norm": 0.148299, "tokens_per_sec": 108303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:30.845557+00:00", "epoch": 0, "step": 8507, "train_loss": 3.650200366973877, "perplexity": 38.48237587381063, "lr": 0.0026291804804649314, "grad_norm": 0.147531, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:31.150092+00:00", "epoch": 0, "step": 8508, "train_loss": 3.5906217098236084, "perplexity": 36.25661001153453, "lr": 0.0026291804804649314, "grad_norm": 0.152903, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:31.455213+00:00", "epoch": 0, "step": 8509, "train_loss": 3.5811612606048584, "perplexity": 35.91522357437561, "lr": 0.0026291804804649314, "grad_norm": 0.173323, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:31.759007+00:00", "epoch": 0, "step": 8510, "train_loss": 3.6068356037139893, "perplexity": 36.84926245752897, "lr": 0.0026291804804649314, "grad_norm": 0.177413, "tokens_per_sec": 107919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:32.063375+00:00", "epoch": 0, "step": 8511, "train_loss": 3.6035406589508057, "perplexity": 36.72804598362747, "lr": 0.0026291804804649314, "grad_norm": 0.178633, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:32.367060+00:00", "epoch": 0, "step": 8512, "train_loss": 3.5260720252990723, "perplexity": 33.99019243802446, "lr": 0.0026291804804649314, "grad_norm": 0.165361, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:32.671851+00:00", "epoch": 0, "step": 8513, "train_loss": 3.5267374515533447, "perplexity": 34.012817931424294, "lr": 0.0026291804804649314, "grad_norm": 0.170244, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:32.977291+00:00", "epoch": 0, "step": 8514, "train_loss": 3.5273189544677734, "perplexity": 34.03260223593722, "lr": 0.0026291804804649314, "grad_norm": 0.149467, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:33.450515+00:00", "epoch": 0, "step": 8515, "train_loss": 3.550145387649536, "perplexity": 34.818379281956375, "lr": 0.0026291804804649314, "grad_norm": 0.18181, "tokens_per_sec": 69244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:33.754395+00:00", "epoch": 0, "step": 8516, "train_loss": 3.606688976287842, "perplexity": 36.84385974112227, "lr": 0.0026291804804649314, "grad_norm": 0.141743, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:34.059851+00:00", "epoch": 0, "step": 8517, "train_loss": 3.553147554397583, "perplexity": 34.923066928714384, "lr": 0.0026291804804649314, "grad_norm": 0.182896, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:34.364986+00:00", "epoch": 0, "step": 8518, "train_loss": 3.5322117805480957, "perplexity": 34.199525870886696, "lr": 0.0026291804804649314, "grad_norm": 0.171474, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:34.670471+00:00", "epoch": 0, "step": 8519, "train_loss": 3.5622153282165527, "perplexity": 35.241181516522076, "lr": 0.0026291804804649314, "grad_norm": 0.157044, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:34.976136+00:00", "epoch": 0, "step": 8520, "train_loss": 3.468048572540283, "perplexity": 32.07409107276619, "lr": 0.0026291804804649314, "grad_norm": 0.151954, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:35.280463+00:00", "epoch": 0, "step": 8521, "train_loss": 3.4827425479888916, "perplexity": 32.54886660706476, "lr": 0.0026291804804649314, "grad_norm": 0.15293, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:35.584693+00:00", "epoch": 0, "step": 8522, "train_loss": 3.527233600616455, "perplexity": 34.029697546230985, "lr": 0.0026291804804649314, "grad_norm": 0.158577, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:35.889098+00:00", "epoch": 0, "step": 8523, "train_loss": 3.6343986988067627, "perplexity": 37.879069318352215, "lr": 0.0026291804804649314, "grad_norm": 0.168025, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:36.193358+00:00", "epoch": 0, "step": 8524, "train_loss": 3.5658085346221924, "perplexity": 35.36803813019947, "lr": 0.0026291804804649314, "grad_norm": 0.16983, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:36.498407+00:00", "epoch": 0, "step": 8525, "train_loss": 3.6297881603240967, "perplexity": 37.7048283924735, "lr": 0.0026291804804649314, "grad_norm": 0.158548, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:36.803660+00:00", "epoch": 0, "step": 8526, "train_loss": 3.5823895931243896, "perplexity": 35.959366516993974, "lr": 0.0026291804804649314, "grad_norm": 0.157398, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:37.108826+00:00", "epoch": 0, "step": 8527, "train_loss": 3.6801557540893555, "perplexity": 39.65256964150017, "lr": 0.0026291804804649314, "grad_norm": 0.132011, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:37.413567+00:00", "epoch": 0, "step": 8528, "train_loss": 3.560743808746338, "perplexity": 35.1893615681494, "lr": 0.0026291804804649314, "grad_norm": 0.169573, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:37.718885+00:00", "epoch": 0, "step": 8529, "train_loss": 3.5843112468719482, "perplexity": 36.02853440550228, "lr": 0.0026291804804649314, "grad_norm": 0.180346, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:38.024220+00:00", "epoch": 0, "step": 8530, "train_loss": 3.545881509780884, "perplexity": 34.670234026200454, "lr": 0.0026291804804649314, "grad_norm": 0.170316, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:38.327774+00:00", "epoch": 0, "step": 8531, "train_loss": 3.537968635559082, "perplexity": 34.3969753815602, "lr": 0.0026291804804649314, "grad_norm": 0.187684, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:38.632597+00:00", "epoch": 0, "step": 8532, "train_loss": 3.5432944297790527, "perplexity": 34.58065528074536, "lr": 0.0026291804804649314, "grad_norm": 0.169171, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:38.936032+00:00", "epoch": 0, "step": 8533, "train_loss": 3.5336520671844482, "perplexity": 34.248818480190195, "lr": 0.0026291804804649314, "grad_norm": 0.158768, "tokens_per_sec": 107991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:39.240823+00:00", "epoch": 0, "step": 8534, "train_loss": 3.5274033546447754, "perplexity": 34.035474714806924, "lr": 0.0026291804804649314, "grad_norm": 0.181993, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:39.546352+00:00", "epoch": 0, "step": 8535, "train_loss": 3.5098540782928467, "perplexity": 33.44338731170064, "lr": 0.0026291804804649314, "grad_norm": 0.18099, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:39.850905+00:00", "epoch": 0, "step": 8536, "train_loss": 3.6194489002227783, "perplexity": 37.316996765617866, "lr": 0.0026291804804649314, "grad_norm": 0.145214, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:40.156258+00:00", "epoch": 0, "step": 8537, "train_loss": 3.3837742805480957, "perplexity": 29.481834102982916, "lr": 0.0026291804804649314, "grad_norm": 0.177756, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:40.460827+00:00", "epoch": 0, "step": 8538, "train_loss": 3.531363010406494, "perplexity": 34.1705106498391, "lr": 0.0026291804804649314, "grad_norm": 0.179277, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:40.766447+00:00", "epoch": 0, "step": 8539, "train_loss": 3.5591514110565186, "perplexity": 35.13337070177668, "lr": 0.0026291804804649314, "grad_norm": 0.163383, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:41.072558+00:00", "epoch": 0, "step": 8540, "train_loss": 3.522726535797119, "perplexity": 33.87666860832663, "lr": 0.0026291804804649314, "grad_norm": 0.149267, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:41.377139+00:00", "epoch": 0, "step": 8541, "train_loss": 3.56650972366333, "perplexity": 35.392846507607636, "lr": 0.0026291804804649314, "grad_norm": 0.151141, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:41.681245+00:00", "epoch": 0, "step": 8542, "train_loss": 3.5856547355651855, "perplexity": 36.07697086374147, "lr": 0.0026291804804649314, "grad_norm": 0.162515, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:41.985394+00:00", "epoch": 0, "step": 8543, "train_loss": 3.595458745956421, "perplexity": 36.43240937538767, "lr": 0.0026291804804649314, "grad_norm": 0.15252, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:42.289899+00:00", "epoch": 0, "step": 8544, "train_loss": 3.521524667739868, "perplexity": 33.83597777984181, "lr": 0.0026291804804649314, "grad_norm": 0.152102, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:42.595201+00:00", "epoch": 0, "step": 8545, "train_loss": 3.6161303520202637, "perplexity": 37.19336376757151, "lr": 0.0026291804804649314, "grad_norm": 0.165654, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:42.900647+00:00", "epoch": 0, "step": 8546, "train_loss": 3.547311782836914, "perplexity": 34.71985740680704, "lr": 0.0026291804804649314, "grad_norm": 0.173605, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:43.205438+00:00", "epoch": 0, "step": 8547, "train_loss": 3.5235166549682617, "perplexity": 33.903445790838084, "lr": 0.0026291804804649314, "grad_norm": 0.146793, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:43.509598+00:00", "epoch": 0, "step": 8548, "train_loss": 3.5950145721435547, "perplexity": 36.41623064655333, "lr": 0.0026291804804649314, "grad_norm": 0.162266, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:43.813955+00:00", "epoch": 0, "step": 8549, "train_loss": 3.5530381202697754, "perplexity": 34.919245362453374, "lr": 0.0026291804804649314, "grad_norm": 0.172006, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:44.119972+00:00", "epoch": 0, "step": 8550, "train_loss": 3.518136978149414, "perplexity": 33.721545929367714, "lr": 0.0026291804804649314, "grad_norm": 0.154623, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:44.424806+00:00", "epoch": 0, "step": 8551, "train_loss": 3.5364856719970703, "perplexity": 34.34600372432245, "lr": 0.0026291804804649314, "grad_norm": 0.163936, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:44.729455+00:00", "epoch": 0, "step": 8552, "train_loss": 3.5291709899902344, "perplexity": 34.09569022676715, "lr": 0.0026291804804649314, "grad_norm": 0.179843, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:45.035000+00:00", "epoch": 0, "step": 8553, "train_loss": 3.5112409591674805, "perplexity": 33.48980148397055, "lr": 0.0026291804804649314, "grad_norm": 0.156478, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:45.339389+00:00", "epoch": 0, "step": 8554, "train_loss": 3.6386055946350098, "perplexity": 38.03875827894237, "lr": 0.0026291804804649314, "grad_norm": 0.166558, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:45.643777+00:00", "epoch": 0, "step": 8555, "train_loss": 3.503689765930176, "perplexity": 33.23786592565179, "lr": 0.0026291804804649314, "grad_norm": 0.172271, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:45.950584+00:00", "epoch": 0, "step": 8556, "train_loss": 3.5147738456726074, "perplexity": 33.60832639594508, "lr": 0.0026291804804649314, "grad_norm": 0.166063, "tokens_per_sec": 106804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:46.255701+00:00", "epoch": 0, "step": 8557, "train_loss": 3.5003247261047363, "perplexity": 33.12620715656569, "lr": 0.0026291804804649314, "grad_norm": 0.154992, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:46.560911+00:00", "epoch": 0, "step": 8558, "train_loss": 3.558838367462158, "perplexity": 35.122374146420604, "lr": 0.0026291804804649314, "grad_norm": 0.179548, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:46.866137+00:00", "epoch": 0, "step": 8559, "train_loss": 3.591857671737671, "perplexity": 36.30144950488674, "lr": 0.0026291804804649314, "grad_norm": 0.171961, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:47.171187+00:00", "epoch": 0, "step": 8560, "train_loss": 3.4798331260681152, "perplexity": 32.45430584633944, "lr": 0.0026291804804649314, "grad_norm": 0.17763, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:47.475464+00:00", "epoch": 0, "step": 8561, "train_loss": 3.594724416732788, "perplexity": 36.405665812987436, "lr": 0.0026291804804649314, "grad_norm": 0.150977, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:47.780920+00:00", "epoch": 0, "step": 8562, "train_loss": 3.6414201259613037, "perplexity": 38.145970360839165, "lr": 0.0026291804804649314, "grad_norm": 0.154685, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:48.085129+00:00", "epoch": 0, "step": 8563, "train_loss": 3.5783886909484863, "perplexity": 35.81578403052293, "lr": 0.0026291804804649314, "grad_norm": 0.138823, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:48.390553+00:00", "epoch": 0, "step": 8564, "train_loss": 3.4973549842834473, "perplexity": 33.027976805377286, "lr": 0.0026291804804649314, "grad_norm": 0.173747, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:48.694056+00:00", "epoch": 0, "step": 8565, "train_loss": 3.4604272842407227, "perplexity": 31.830574312175262, "lr": 0.0026291804804649314, "grad_norm": 0.149423, "tokens_per_sec": 107965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:48.999228+00:00", "epoch": 0, "step": 8566, "train_loss": 3.6287992000579834, "perplexity": 37.66755824773651, "lr": 0.0026291804804649314, "grad_norm": 0.164472, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:49.305607+00:00", "epoch": 0, "step": 8567, "train_loss": 3.6261868476867676, "perplexity": 37.56928572974028, "lr": 0.0026291804804649314, "grad_norm": 0.168695, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:49.610998+00:00", "epoch": 0, "step": 8568, "train_loss": 3.545619249343872, "perplexity": 34.66114258768787, "lr": 0.0026291804804649314, "grad_norm": 0.164224, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:49.915713+00:00", "epoch": 0, "step": 8569, "train_loss": 3.6768202781677246, "perplexity": 39.52052978053496, "lr": 0.0026291804804649314, "grad_norm": 0.157849, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:50.220117+00:00", "epoch": 0, "step": 8570, "train_loss": 3.516310453414917, "perplexity": 33.660008908196154, "lr": 0.0026291804804649314, "grad_norm": 0.168285, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:50.524090+00:00", "epoch": 0, "step": 8571, "train_loss": 3.578817367553711, "perplexity": 35.83114071052415, "lr": 0.0026291804804649314, "grad_norm": 0.145826, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:50.829519+00:00", "epoch": 0, "step": 8572, "train_loss": 3.670179843902588, "perplexity": 39.258965711054955, "lr": 0.0026291804804649314, "grad_norm": 0.161665, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:51.134917+00:00", "epoch": 0, "step": 8573, "train_loss": 3.587448835372925, "perplexity": 36.14175464714146, "lr": 0.0026291804804649314, "grad_norm": 0.163334, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:51.439501+00:00", "epoch": 0, "step": 8574, "train_loss": 3.6047933101654053, "perplexity": 36.77408224270689, "lr": 0.0026291804804649314, "grad_norm": 0.158644, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:51.744650+00:00", "epoch": 0, "step": 8575, "train_loss": 3.55613374710083, "perplexity": 35.027509802012226, "lr": 0.0026291804804649314, "grad_norm": 0.166991, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:52.048983+00:00", "epoch": 0, "step": 8576, "train_loss": 3.606851816177368, "perplexity": 36.84985987968993, "lr": 0.0026291804804649314, "grad_norm": 0.164879, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:52.354078+00:00", "epoch": 0, "step": 8577, "train_loss": 3.5197501182556152, "perplexity": 33.77598740660894, "lr": 0.0026291804804649314, "grad_norm": 0.14501, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:52.659760+00:00", "epoch": 0, "step": 8578, "train_loss": 3.575258493423462, "perplexity": 35.703848833029205, "lr": 0.0026291804804649314, "grad_norm": 0.168076, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:52.966076+00:00", "epoch": 0, "step": 8579, "train_loss": 3.5545361042022705, "perplexity": 34.971593029134155, "lr": 0.0026291804804649314, "grad_norm": 0.162296, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:53.271285+00:00", "epoch": 0, "step": 8580, "train_loss": 3.56182599067688, "perplexity": 35.22746347226335, "lr": 0.0026291804804649314, "grad_norm": 0.179674, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:53.575341+00:00", "epoch": 0, "step": 8581, "train_loss": 3.495781898498535, "perplexity": 32.976061808637766, "lr": 0.0026291804804649314, "grad_norm": 0.220687, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:53.880559+00:00", "epoch": 0, "step": 8582, "train_loss": 3.391878604888916, "perplexity": 29.721735254025653, "lr": 0.0026291804804649314, "grad_norm": 0.220234, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:54.186152+00:00", "epoch": 0, "step": 8583, "train_loss": 3.4109675884246826, "perplexity": 30.294542731561094, "lr": 0.0026291804804649314, "grad_norm": 0.17611, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:54.490966+00:00", "epoch": 0, "step": 8584, "train_loss": 3.501149892807007, "perplexity": 33.15355308060513, "lr": 0.0026291804804649314, "grad_norm": 0.212815, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:54.795305+00:00", "epoch": 0, "step": 8585, "train_loss": 3.4544031620025635, "perplexity": 31.639399450121918, "lr": 0.0026291804804649314, "grad_norm": 0.142644, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:55.100392+00:00", "epoch": 0, "step": 8586, "train_loss": 3.550245523452759, "perplexity": 34.82186602290357, "lr": 0.0026291804804649314, "grad_norm": 0.179124, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:55.406329+00:00", "epoch": 0, "step": 8587, "train_loss": 3.4868714809417725, "perplexity": 32.68353652518695, "lr": 0.0026291804804649314, "grad_norm": 0.167662, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:55.712856+00:00", "epoch": 0, "step": 8588, "train_loss": 3.565424680709839, "perplexity": 35.354464575688596, "lr": 0.0026291804804649314, "grad_norm": 0.158804, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:56.019014+00:00", "epoch": 0, "step": 8589, "train_loss": 3.52364182472229, "perplexity": 33.907689742410085, "lr": 0.0026291804804649314, "grad_norm": 0.153227, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:56.324202+00:00", "epoch": 0, "step": 8590, "train_loss": 3.582169771194458, "perplexity": 35.951462728391945, "lr": 0.0026291804804649314, "grad_norm": 0.155877, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:56.628489+00:00", "epoch": 0, "step": 8591, "train_loss": 3.6605095863342285, "perplexity": 38.881151127169055, "lr": 0.0026291804804649314, "grad_norm": 0.141536, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:56.932232+00:00", "epoch": 0, "step": 8592, "train_loss": 3.6249499320983887, "perplexity": 37.52284442248646, "lr": 0.0026291804804649314, "grad_norm": 0.153588, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:57.237111+00:00", "epoch": 0, "step": 8593, "train_loss": 3.7154626846313477, "perplexity": 41.0775886571504, "lr": 0.0026291804804649314, "grad_norm": 0.168492, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:57.542577+00:00", "epoch": 0, "step": 8594, "train_loss": 3.516691207885742, "perplexity": 33.67282754729817, "lr": 0.0026291804804649314, "grad_norm": 0.138766, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:57.848194+00:00", "epoch": 0, "step": 8595, "train_loss": 3.4783034324645996, "perplexity": 32.40469865385552, "lr": 0.0026291804804649314, "grad_norm": 0.160801, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:58.154134+00:00", "epoch": 0, "step": 8596, "train_loss": 3.579880952835083, "perplexity": 35.869270457923285, "lr": 0.0026291804804649314, "grad_norm": 0.200558, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:58.459078+00:00", "epoch": 0, "step": 8597, "train_loss": 3.6309168338775635, "perplexity": 37.747408860321784, "lr": 0.0026291804804649314, "grad_norm": 0.165559, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:58.765040+00:00", "epoch": 0, "step": 8598, "train_loss": 3.5167462825775146, "perplexity": 33.67468211896596, "lr": 0.0026291804804649314, "grad_norm": 0.182674, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:59.072482+00:00", "epoch": 0, "step": 8599, "train_loss": 3.551090717315674, "perplexity": 34.85130969141258, "lr": 0.0026291804804649314, "grad_norm": 0.169229, "tokens_per_sec": 106583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:59.378332+00:00", "epoch": 0, "step": 8600, "train_loss": 3.5608749389648438, "perplexity": 35.19397625937704, "lr": 0.0026291804804649314, "grad_norm": 0.172503, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:59.682389+00:00", "epoch": 0, "step": 8601, "train_loss": 3.5566699504852295, "perplexity": 35.04629670766622, "lr": 0.0026291804804649314, "grad_norm": 0.164049, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:42:59.987200+00:00", "epoch": 0, "step": 8602, "train_loss": 3.576420545578003, "perplexity": 35.74536268344912, "lr": 0.0026291804804649314, "grad_norm": 0.159286, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:00.291074+00:00", "epoch": 0, "step": 8603, "train_loss": 3.519843816757202, "perplexity": 33.779152314289796, "lr": 0.0026291804804649314, "grad_norm": 0.160414, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:00.595798+00:00", "epoch": 0, "step": 8604, "train_loss": 3.5309839248657227, "perplexity": 34.15755955827184, "lr": 0.0026291804804649314, "grad_norm": 0.15169, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:00.901989+00:00", "epoch": 0, "step": 8605, "train_loss": 3.6080331802368164, "perplexity": 36.89341870409635, "lr": 0.0026291804804649314, "grad_norm": 0.176245, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:01.208406+00:00", "epoch": 0, "step": 8606, "train_loss": 3.5706229209899902, "perplexity": 35.53872407518835, "lr": 0.0026291804804649314, "grad_norm": 0.161039, "tokens_per_sec": 106940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:01.512469+00:00", "epoch": 0, "step": 8607, "train_loss": 3.4964263439178467, "perplexity": 32.997319929730935, "lr": 0.0026291804804649314, "grad_norm": 0.173131, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:01.817550+00:00", "epoch": 0, "step": 8608, "train_loss": 3.647944927215576, "perplexity": 38.395678999811544, "lr": 0.0026291804804649314, "grad_norm": 0.168536, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:02.123370+00:00", "epoch": 0, "step": 8609, "train_loss": 3.5559091567993164, "perplexity": 35.01964384636626, "lr": 0.0026291804804649314, "grad_norm": 0.141254, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:02.428604+00:00", "epoch": 0, "step": 8610, "train_loss": 3.51442551612854, "perplexity": 33.59662166160545, "lr": 0.0026291804804649314, "grad_norm": 0.165021, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:02.733796+00:00", "epoch": 0, "step": 8611, "train_loss": 3.573284149169922, "perplexity": 35.63342668591286, "lr": 0.0026291804804649314, "grad_norm": 0.169367, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:03.038363+00:00", "epoch": 0, "step": 8612, "train_loss": 3.560152053833008, "perplexity": 35.16854425052697, "lr": 0.0026291804804649314, "grad_norm": 0.157046, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:03.342221+00:00", "epoch": 0, "step": 8613, "train_loss": 3.626556634902954, "perplexity": 37.58318094030167, "lr": 0.0026291804804649314, "grad_norm": 0.180478, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:03.646613+00:00", "epoch": 0, "step": 8614, "train_loss": 3.600482940673828, "perplexity": 36.61591348830239, "lr": 0.0026291804804649314, "grad_norm": 0.159989, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:03.952649+00:00", "epoch": 0, "step": 8615, "train_loss": 3.5079116821289062, "perplexity": 33.37849005297467, "lr": 0.0026291804804649314, "grad_norm": 0.151621, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:04.258476+00:00", "epoch": 0, "step": 8616, "train_loss": 3.616641044616699, "perplexity": 37.21236299405337, "lr": 0.0026291804804649314, "grad_norm": 0.149725, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:04.564336+00:00", "epoch": 0, "step": 8617, "train_loss": 3.494809865951538, "perplexity": 32.944023576912585, "lr": 0.0026291804804649314, "grad_norm": 0.153971, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:04.870576+00:00", "epoch": 0, "step": 8618, "train_loss": 3.548741579055786, "perplexity": 34.76953523376854, "lr": 0.0026291804804649314, "grad_norm": 0.175614, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:05.174382+00:00", "epoch": 0, "step": 8619, "train_loss": 3.619976282119751, "perplexity": 37.33668226459307, "lr": 0.0026291804804649314, "grad_norm": 0.159089, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:05.480528+00:00", "epoch": 0, "step": 8620, "train_loss": 3.5158371925354004, "perplexity": 33.64408271169586, "lr": 0.0026291804804649314, "grad_norm": 0.156207, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:05.786721+00:00", "epoch": 0, "step": 8621, "train_loss": 3.620955467224121, "perplexity": 37.373259692827055, "lr": 0.0026291804804649314, "grad_norm": 0.156436, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:06.091698+00:00", "epoch": 0, "step": 8622, "train_loss": 3.4641976356506348, "perplexity": 31.950813292076617, "lr": 0.0026291804804649314, "grad_norm": 0.149894, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:06.397538+00:00", "epoch": 0, "step": 8623, "train_loss": 3.5658278465270996, "perplexity": 35.36872116098389, "lr": 0.0026291804804649314, "grad_norm": 0.172221, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:06.703396+00:00", "epoch": 0, "step": 8624, "train_loss": 3.4843881130218506, "perplexity": 32.60247197728077, "lr": 0.0026291804804649314, "grad_norm": 0.14248, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:07.009008+00:00", "epoch": 0, "step": 8625, "train_loss": 3.6053285598754883, "perplexity": 36.79377082825067, "lr": 0.0026291804804649314, "grad_norm": 0.140305, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:07.313638+00:00", "epoch": 0, "step": 8626, "train_loss": 3.352419853210449, "perplexity": 28.57178959421591, "lr": 0.0026291804804649314, "grad_norm": 0.151664, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:07.618854+00:00", "epoch": 0, "step": 8627, "train_loss": 3.5083491802215576, "perplexity": 33.39309627357228, "lr": 0.0026291804804649314, "grad_norm": 0.153293, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:07.924839+00:00", "epoch": 0, "step": 8628, "train_loss": 3.5054142475128174, "perplexity": 33.29523346370148, "lr": 0.0026291804804649314, "grad_norm": 0.159083, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:08.230714+00:00", "epoch": 0, "step": 8629, "train_loss": 3.468794345855713, "perplexity": 32.09801999567194, "lr": 0.0026291804804649314, "grad_norm": 0.149269, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:08.536096+00:00", "epoch": 0, "step": 8630, "train_loss": 3.55208158493042, "perplexity": 34.88585984000212, "lr": 0.0026291804804649314, "grad_norm": 0.182134, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:08.840731+00:00", "epoch": 0, "step": 8631, "train_loss": 3.5576722621917725, "perplexity": 35.08144163126917, "lr": 0.0026291804804649314, "grad_norm": 0.157822, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:09.144249+00:00", "epoch": 0, "step": 8632, "train_loss": 3.5484323501586914, "perplexity": 34.7587851509395, "lr": 0.0026291804804649314, "grad_norm": 0.181647, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:09.449171+00:00", "epoch": 0, "step": 8633, "train_loss": 3.4567363262176514, "perplexity": 31.71330554871146, "lr": 0.0026291804804649314, "grad_norm": 0.178909, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:09.753799+00:00", "epoch": 0, "step": 8634, "train_loss": 3.6309897899627686, "perplexity": 37.75016286395831, "lr": 0.0026291804804649314, "grad_norm": 0.182685, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:10.060451+00:00", "epoch": 0, "step": 8635, "train_loss": 3.5730204582214355, "perplexity": 35.62403171257139, "lr": 0.0026291804804649314, "grad_norm": 0.162648, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:10.369673+00:00", "epoch": 0, "step": 8636, "train_loss": 3.4725944995880127, "perplexity": 32.22022946647563, "lr": 0.0026291804804649314, "grad_norm": 0.172434, "tokens_per_sec": 105971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:10.673722+00:00", "epoch": 0, "step": 8637, "train_loss": 3.4810709953308105, "perplexity": 32.49450490944789, "lr": 0.0026291804804649314, "grad_norm": 0.131718, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:10.978855+00:00", "epoch": 0, "step": 8638, "train_loss": 3.534203290939331, "perplexity": 34.26770244668045, "lr": 0.0026291804804649314, "grad_norm": 0.166745, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:11.284224+00:00", "epoch": 0, "step": 8639, "train_loss": 3.553506851196289, "perplexity": 34.93561692931537, "lr": 0.0026291804804649314, "grad_norm": 0.175289, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:11.589122+00:00", "epoch": 0, "step": 8640, "train_loss": 3.457848072052002, "perplexity": 31.74858228980026, "lr": 0.0026291804804649314, "grad_norm": 0.149622, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:11.893469+00:00", "epoch": 0, "step": 8641, "train_loss": 3.554342031478882, "perplexity": 34.964806655380066, "lr": 0.0026291804804649314, "grad_norm": 0.158874, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:12.198786+00:00", "epoch": 0, "step": 8642, "train_loss": 3.5267748832702637, "perplexity": 34.01409111342526, "lr": 0.0026291804804649314, "grad_norm": 0.172587, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:12.503693+00:00", "epoch": 0, "step": 8643, "train_loss": 3.500425338745117, "perplexity": 33.129540239405834, "lr": 0.0026291804804649314, "grad_norm": 0.1683, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:12.809765+00:00", "epoch": 0, "step": 8644, "train_loss": 3.470942735671997, "perplexity": 32.167053183500045, "lr": 0.0026291804804649314, "grad_norm": 0.157198, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:13.116885+00:00", "epoch": 0, "step": 8645, "train_loss": 3.567537307739258, "perplexity": 35.42923432564989, "lr": 0.0026291804804649314, "grad_norm": 0.174826, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:13.421933+00:00", "epoch": 0, "step": 8646, "train_loss": 3.544978618621826, "perplexity": 34.638944705966665, "lr": 0.0026291804804649314, "grad_norm": 0.177915, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:13.727562+00:00", "epoch": 0, "step": 8647, "train_loss": 3.554542064666748, "perplexity": 34.97180147669335, "lr": 0.0026291804804649314, "grad_norm": 0.143561, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:14.033804+00:00", "epoch": 0, "step": 8648, "train_loss": 3.642963409423828, "perplexity": 38.20488585602032, "lr": 0.0026291804804649314, "grad_norm": 0.160367, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:14.338547+00:00", "epoch": 0, "step": 8649, "train_loss": 3.613786458969116, "perplexity": 37.10628858800678, "lr": 0.0026291804804649314, "grad_norm": 0.148194, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:14.643210+00:00", "epoch": 0, "step": 8650, "train_loss": 3.3924570083618164, "perplexity": 29.738931381587744, "lr": 0.0026291804804649314, "grad_norm": 0.183968, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:14.947947+00:00", "epoch": 0, "step": 8651, "train_loss": 3.655195951461792, "perplexity": 38.675098814854195, "lr": 0.0026291804804649314, "grad_norm": 0.201876, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:15.252414+00:00", "epoch": 0, "step": 8652, "train_loss": 3.5435078144073486, "perplexity": 34.58803504835515, "lr": 0.0026291804804649314, "grad_norm": 0.196423, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:15.556969+00:00", "epoch": 0, "step": 8653, "train_loss": 3.4939613342285156, "perplexity": 32.91608138441849, "lr": 0.0026291804804649314, "grad_norm": 0.147499, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:15.863120+00:00", "epoch": 0, "step": 8654, "train_loss": 3.5969114303588867, "perplexity": 36.4853726283633, "lr": 0.0026291804804649314, "grad_norm": 0.162615, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:16.167824+00:00", "epoch": 0, "step": 8655, "train_loss": 3.5181102752685547, "perplexity": 33.72064547896673, "lr": 0.0026291804804649314, "grad_norm": 0.172842, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:16.471989+00:00", "epoch": 0, "step": 8656, "train_loss": 3.675436019897461, "perplexity": 39.46586100692161, "lr": 0.0026291804804649314, "grad_norm": 0.152457, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:16.777235+00:00", "epoch": 0, "step": 8657, "train_loss": 3.5725061893463135, "perplexity": 35.60571609183575, "lr": 0.0026291804804649314, "grad_norm": 0.154932, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:17.083107+00:00", "epoch": 0, "step": 8658, "train_loss": 3.5944743156433105, "perplexity": 36.39656185480684, "lr": 0.0026291804804649314, "grad_norm": 0.16053, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:17.388757+00:00", "epoch": 0, "step": 8659, "train_loss": 3.443557024002075, "perplexity": 31.298088456444752, "lr": 0.0026291804804649314, "grad_norm": 0.182261, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:17.693893+00:00", "epoch": 0, "step": 8660, "train_loss": 3.499833583831787, "perplexity": 33.109941470598585, "lr": 0.0026291804804649314, "grad_norm": 0.155603, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:17.998389+00:00", "epoch": 0, "step": 8661, "train_loss": 3.598898410797119, "perplexity": 36.557940421574905, "lr": 0.0026291804804649314, "grad_norm": 0.163671, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:18.303114+00:00", "epoch": 0, "step": 8662, "train_loss": 3.498814105987549, "perplexity": 33.07620381916142, "lr": 0.0026291804804649314, "grad_norm": 0.158322, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:18.607352+00:00", "epoch": 0, "step": 8663, "train_loss": 3.6366446018218994, "perplexity": 37.96423763843866, "lr": 0.0026291804804649314, "grad_norm": 0.145811, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:18.913678+00:00", "epoch": 0, "step": 8664, "train_loss": 3.4891788959503174, "perplexity": 32.759038081110596, "lr": 0.0026291804804649314, "grad_norm": 0.167777, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:19.220846+00:00", "epoch": 0, "step": 8665, "train_loss": 3.555781364440918, "perplexity": 35.0151688894276, "lr": 0.0026291804804649314, "grad_norm": 0.147788, "tokens_per_sec": 106678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:19.526518+00:00", "epoch": 0, "step": 8666, "train_loss": 3.5127501487731934, "perplexity": 33.54038210253902, "lr": 0.0026291804804649314, "grad_norm": 0.162586, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:19.832610+00:00", "epoch": 0, "step": 8667, "train_loss": 3.553454875946045, "perplexity": 34.93380118907022, "lr": 0.0026291804804649314, "grad_norm": 0.15734, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:20.138910+00:00", "epoch": 0, "step": 8668, "train_loss": 3.451479911804199, "perplexity": 31.54704462330347, "lr": 0.0026291804804649314, "grad_norm": 0.160084, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:20.445226+00:00", "epoch": 0, "step": 8669, "train_loss": 3.474393844604492, "perplexity": 32.27825696588588, "lr": 0.0026291804804649314, "grad_norm": 0.159502, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:20.751927+00:00", "epoch": 0, "step": 8670, "train_loss": 3.4598922729492188, "perplexity": 31.813549150236234, "lr": 0.0026291804804649314, "grad_norm": 0.157817, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:21.056793+00:00", "epoch": 0, "step": 8671, "train_loss": 3.606905698776245, "perplexity": 36.8518454994032, "lr": 0.0026291804804649314, "grad_norm": 0.156826, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:21.362480+00:00", "epoch": 0, "step": 8672, "train_loss": 3.557156562805176, "perplexity": 35.063354817419246, "lr": 0.0026291804804649314, "grad_norm": 0.181181, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:21.668251+00:00", "epoch": 0, "step": 8673, "train_loss": 3.5712082386016846, "perplexity": 35.55953160519954, "lr": 0.0026291804804649314, "grad_norm": 0.219583, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:21.974051+00:00", "epoch": 0, "step": 8674, "train_loss": 3.5374038219451904, "perplexity": 34.37755298712074, "lr": 0.0026291804804649314, "grad_norm": 0.17281, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:22.280586+00:00", "epoch": 0, "step": 8675, "train_loss": 3.4904823303222656, "perplexity": 32.801765177271115, "lr": 0.0026291804804649314, "grad_norm": 0.157119, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:22.587505+00:00", "epoch": 0, "step": 8676, "train_loss": 3.494617462158203, "perplexity": 32.937685631550586, "lr": 0.0026291804804649314, "grad_norm": 0.175713, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:22.892149+00:00", "epoch": 0, "step": 8677, "train_loss": 3.4805636405944824, "perplexity": 32.47802284995647, "lr": 0.0026291804804649314, "grad_norm": 0.17606, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:23.197475+00:00", "epoch": 0, "step": 8678, "train_loss": 3.6501739025115967, "perplexity": 38.48135747190165, "lr": 0.0026291804804649314, "grad_norm": 0.157216, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:23.502409+00:00", "epoch": 0, "step": 8679, "train_loss": 3.4757559299468994, "perplexity": 32.32225266272244, "lr": 0.0026291804804649314, "grad_norm": 0.155126, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:23.808051+00:00", "epoch": 0, "step": 8680, "train_loss": 3.674116373062134, "perplexity": 39.413814357491596, "lr": 0.0026291804804649314, "grad_norm": 0.146406, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:24.113149+00:00", "epoch": 0, "step": 8681, "train_loss": 3.6417829990386963, "perplexity": 38.15981501826893, "lr": 0.0026291804804649314, "grad_norm": 0.151878, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:24.418286+00:00", "epoch": 0, "step": 8682, "train_loss": 3.48237681388855, "perplexity": 32.53696455324523, "lr": 0.0026291804804649314, "grad_norm": 0.167767, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:24.722392+00:00", "epoch": 0, "step": 8683, "train_loss": 3.5965404510498047, "perplexity": 36.471839820385114, "lr": 0.0026291804804649314, "grad_norm": 0.160172, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:25.027357+00:00", "epoch": 0, "step": 8684, "train_loss": 3.5642013549804688, "perplexity": 35.31124099317579, "lr": 0.0026291804804649314, "grad_norm": 0.164571, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:25.332571+00:00", "epoch": 0, "step": 8685, "train_loss": 3.4947428703308105, "perplexity": 32.941816545535325, "lr": 0.0026291804804649314, "grad_norm": 0.15321, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:25.638229+00:00", "epoch": 0, "step": 8686, "train_loss": 3.6224372386932373, "perplexity": 37.42867937225749, "lr": 0.0026291804804649314, "grad_norm": 0.154173, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:25.944244+00:00", "epoch": 0, "step": 8687, "train_loss": 3.5561256408691406, "perplexity": 35.02722586205311, "lr": 0.0026291804804649314, "grad_norm": 0.169001, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:26.248946+00:00", "epoch": 0, "step": 8688, "train_loss": 3.482097625732422, "perplexity": 32.527881886051155, "lr": 0.0026291804804649314, "grad_norm": 0.179497, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:26.553268+00:00", "epoch": 0, "step": 8689, "train_loss": 3.5248498916625977, "perplexity": 33.94867725425426, "lr": 0.0026291804804649314, "grad_norm": 0.166142, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:26.858337+00:00", "epoch": 0, "step": 8690, "train_loss": 3.552379846572876, "perplexity": 34.89626650573395, "lr": 0.0026291804804649314, "grad_norm": 0.166061, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:27.163417+00:00", "epoch": 0, "step": 8691, "train_loss": 3.568620443344116, "perplexity": 35.467629780789395, "lr": 0.0026291804804649314, "grad_norm": 0.167528, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:27.469559+00:00", "epoch": 0, "step": 8692, "train_loss": 3.5810108184814453, "perplexity": 35.90982081828946, "lr": 0.0026291804804649314, "grad_norm": 0.154293, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:27.774162+00:00", "epoch": 0, "step": 8693, "train_loss": 3.437670946121216, "perplexity": 31.114406583491384, "lr": 0.0026291804804649314, "grad_norm": 0.162092, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:28.080041+00:00", "epoch": 0, "step": 8694, "train_loss": 3.56564998626709, "perplexity": 35.36243103044069, "lr": 0.0026291804804649314, "grad_norm": 0.197883, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:28.385173+00:00", "epoch": 0, "step": 8695, "train_loss": 3.544679641723633, "perplexity": 34.6285900097064, "lr": 0.0026291804804649314, "grad_norm": 0.169514, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:28.691084+00:00", "epoch": 0, "step": 8696, "train_loss": 3.505999803543091, "perplexity": 33.314735397615785, "lr": 0.0026291804804649314, "grad_norm": 0.17352, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:28.996633+00:00", "epoch": 0, "step": 8697, "train_loss": 3.5331342220306396, "perplexity": 34.23108748686728, "lr": 0.0026291804804649314, "grad_norm": 0.165436, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:29.302315+00:00", "epoch": 0, "step": 8698, "train_loss": 3.5374257564544678, "perplexity": 34.37830705014564, "lr": 0.0026291804804649314, "grad_norm": 0.170346, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:29.607481+00:00", "epoch": 0, "step": 8699, "train_loss": 3.4549663066864014, "perplexity": 31.657222027595946, "lr": 0.0026291804804649314, "grad_norm": 0.18378, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:29.913086+00:00", "epoch": 0, "step": 8700, "train_loss": 3.5675432682037354, "perplexity": 35.4294455009719, "lr": 0.0026291804804649314, "grad_norm": 0.159724, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:30.218181+00:00", "epoch": 0, "step": 8701, "train_loss": 3.5854287147521973, "perplexity": 36.06881763889066, "lr": 0.0026291804804649314, "grad_norm": 0.153191, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:30.523698+00:00", "epoch": 0, "step": 8702, "train_loss": 3.601525068283081, "perplexity": 36.654091832584655, "lr": 0.0026291804804649314, "grad_norm": 0.151026, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:30.828856+00:00", "epoch": 0, "step": 8703, "train_loss": 3.6006157398223877, "perplexity": 36.620776373323714, "lr": 0.0026291804804649314, "grad_norm": 0.1639, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:31.133993+00:00", "epoch": 0, "step": 8704, "train_loss": 3.6231701374053955, "perplexity": 37.456120857853605, "lr": 0.0026291804804649314, "grad_norm": 0.164228, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:31.438773+00:00", "epoch": 0, "step": 8705, "train_loss": 3.6087560653686523, "perplexity": 36.92009804982951, "lr": 0.0026291804804649314, "grad_norm": 0.154357, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:31.743840+00:00", "epoch": 0, "step": 8706, "train_loss": 3.556574583053589, "perplexity": 35.042954591727735, "lr": 0.0026291804804649314, "grad_norm": 0.164388, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:32.050576+00:00", "epoch": 0, "step": 8707, "train_loss": 3.5501205921173096, "perplexity": 34.81751595241421, "lr": 0.0026291804804649314, "grad_norm": 0.187526, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:32.356658+00:00", "epoch": 0, "step": 8708, "train_loss": 3.5575478076934814, "perplexity": 35.077075859727216, "lr": 0.0026291804804649314, "grad_norm": 0.182082, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:32.661275+00:00", "epoch": 0, "step": 8709, "train_loss": 3.574414014816284, "perplexity": 35.67371042390777, "lr": 0.0026291804804649314, "grad_norm": 0.191204, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:32.965169+00:00", "epoch": 0, "step": 8710, "train_loss": 3.5076355934143066, "perplexity": 33.369275900580945, "lr": 0.0026291804804649314, "grad_norm": 0.167475, "tokens_per_sec": 107896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:33.269301+00:00", "epoch": 0, "step": 8711, "train_loss": 3.585301399230957, "perplexity": 36.06422581088414, "lr": 0.0026291804804649314, "grad_norm": 0.14999, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:33.574287+00:00", "epoch": 0, "step": 8712, "train_loss": 3.631641387939453, "perplexity": 37.774768809421516, "lr": 0.0026291804804649314, "grad_norm": 0.173431, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:33.878878+00:00", "epoch": 0, "step": 8713, "train_loss": 3.4509551525115967, "perplexity": 31.530494361314922, "lr": 0.0026291804804649314, "grad_norm": 0.16241, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:34.185267+00:00", "epoch": 0, "step": 8714, "train_loss": 3.6307337284088135, "perplexity": 37.74049773607995, "lr": 0.0026291804804649314, "grad_norm": 0.156466, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:34.489874+00:00", "epoch": 0, "step": 8715, "train_loss": 3.6283581256866455, "perplexity": 37.65094771667162, "lr": 0.0026291804804649314, "grad_norm": 0.169902, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:34.794962+00:00", "epoch": 0, "step": 8716, "train_loss": 3.548698663711548, "perplexity": 34.768043119212514, "lr": 0.0026291804804649314, "grad_norm": 0.151956, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:35.100001+00:00", "epoch": 0, "step": 8717, "train_loss": 3.5469846725463867, "perplexity": 34.70850204149331, "lr": 0.0026291804804649314, "grad_norm": 0.173884, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:35.405144+00:00", "epoch": 0, "step": 8718, "train_loss": 3.5214807987213135, "perplexity": 33.83449346126281, "lr": 0.0026291804804649314, "grad_norm": 0.154989, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:35.710331+00:00", "epoch": 0, "step": 8719, "train_loss": 3.586838722229004, "perplexity": 36.119710812887185, "lr": 0.0026291804804649314, "grad_norm": 0.171391, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:36.015494+00:00", "epoch": 0, "step": 8720, "train_loss": 3.5264008045196533, "perplexity": 34.0013695443011, "lr": 0.0026291804804649314, "grad_norm": 0.162682, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:36.320230+00:00", "epoch": 0, "step": 8721, "train_loss": 3.404665946960449, "perplexity": 30.104237631958895, "lr": 0.0026291804804649314, "grad_norm": 0.161091, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:36.624862+00:00", "epoch": 0, "step": 8722, "train_loss": 3.608947515487671, "perplexity": 36.92716708365742, "lr": 0.0026291804804649314, "grad_norm": 0.155326, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:36.931429+00:00", "epoch": 0, "step": 8723, "train_loss": 3.465005397796631, "perplexity": 31.976632376023204, "lr": 0.0026291804804649314, "grad_norm": 0.13648, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:37.237929+00:00", "epoch": 0, "step": 8724, "train_loss": 3.4445877075195312, "perplexity": 31.330363510170468, "lr": 0.0026291804804649314, "grad_norm": 0.151177, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:37.544650+00:00", "epoch": 0, "step": 8725, "train_loss": 3.511155128479004, "perplexity": 33.48692715460684, "lr": 0.0026291804804649314, "grad_norm": 0.143589, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:37.850098+00:00", "epoch": 0, "step": 8726, "train_loss": 3.598764657974243, "perplexity": 36.55305102083785, "lr": 0.0026291804804649314, "grad_norm": 0.149039, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:38.155363+00:00", "epoch": 0, "step": 8727, "train_loss": 3.4350523948669434, "perplexity": 31.03303849485531, "lr": 0.0026291804804649314, "grad_norm": 0.153756, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:38.459919+00:00", "epoch": 0, "step": 8728, "train_loss": 3.571511745452881, "perplexity": 35.57032580464093, "lr": 0.0026291804804649314, "grad_norm": 0.158896, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:38.765848+00:00", "epoch": 0, "step": 8729, "train_loss": 3.50565767288208, "perplexity": 33.303339354751806, "lr": 0.0026291804804649314, "grad_norm": 0.171894, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:39.071590+00:00", "epoch": 0, "step": 8730, "train_loss": 3.50146222114563, "perplexity": 33.16390949197444, "lr": 0.0026291804804649314, "grad_norm": 0.188747, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:39.377391+00:00", "epoch": 0, "step": 8731, "train_loss": 3.5608880519866943, "perplexity": 35.194437761782574, "lr": 0.0026291804804649314, "grad_norm": 0.169047, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:39.682349+00:00", "epoch": 0, "step": 8732, "train_loss": 3.5604240894317627, "perplexity": 35.17811264793285, "lr": 0.0026291804804649314, "grad_norm": 0.183238, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:39.987072+00:00", "epoch": 0, "step": 8733, "train_loss": 3.4467058181762695, "perplexity": 31.396795016805434, "lr": 0.0026291804804649314, "grad_norm": 0.204535, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:40.293369+00:00", "epoch": 0, "step": 8734, "train_loss": 3.6051924228668213, "perplexity": 36.78876217529179, "lr": 0.0026291804804649314, "grad_norm": 0.200575, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:40.599010+00:00", "epoch": 0, "step": 8735, "train_loss": 3.6208064556121826, "perplexity": 37.36769105806276, "lr": 0.0026291804804649314, "grad_norm": 0.174298, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:40.905683+00:00", "epoch": 0, "step": 8736, "train_loss": 3.5701091289520264, "perplexity": 35.52046925171218, "lr": 0.0026291804804649314, "grad_norm": 0.19054, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:41.211061+00:00", "epoch": 0, "step": 8737, "train_loss": 3.534895420074463, "perplexity": 34.291428331678816, "lr": 0.0026291804804649314, "grad_norm": 0.154915, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:41.516790+00:00", "epoch": 0, "step": 8738, "train_loss": 3.5065176486968994, "perplexity": 33.33199173956276, "lr": 0.0026291804804649314, "grad_norm": 0.183749, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:41.821815+00:00", "epoch": 0, "step": 8739, "train_loss": 3.5450706481933594, "perplexity": 34.64213265989709, "lr": 0.0026291804804649314, "grad_norm": 0.163521, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:42.127331+00:00", "epoch": 0, "step": 8740, "train_loss": 3.568476676940918, "perplexity": 35.46253109374457, "lr": 0.0026291804804649314, "grad_norm": 0.159651, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:42.431736+00:00", "epoch": 0, "step": 8741, "train_loss": 3.5679450035095215, "perplexity": 35.44368161947831, "lr": 0.0026291804804649314, "grad_norm": 0.161277, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:42.737120+00:00", "epoch": 0, "step": 8742, "train_loss": 3.5025100708007812, "perplexity": 33.19867849626232, "lr": 0.0026291804804649314, "grad_norm": 0.176743, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:43.043379+00:00", "epoch": 0, "step": 8743, "train_loss": 3.6460213661193848, "perplexity": 38.321893553583614, "lr": 0.0026291804804649314, "grad_norm": 0.18917, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:43.349260+00:00", "epoch": 0, "step": 8744, "train_loss": 3.536898612976074, "perplexity": 34.36018952547297, "lr": 0.0026291804804649314, "grad_norm": 0.176056, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:43.653908+00:00", "epoch": 0, "step": 8745, "train_loss": 3.6224863529205322, "perplexity": 37.43051769806714, "lr": 0.0026291804804649314, "grad_norm": 0.149276, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:43.961148+00:00", "epoch": 0, "step": 8746, "train_loss": 3.5063579082489014, "perplexity": 33.326667697513365, "lr": 0.0026291804804649314, "grad_norm": 0.156505, "tokens_per_sec": 106654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:44.267411+00:00", "epoch": 0, "step": 8747, "train_loss": 3.633727550506592, "perplexity": 37.85365537458368, "lr": 0.0026291804804649314, "grad_norm": 0.16176, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:44.572190+00:00", "epoch": 0, "step": 8748, "train_loss": 3.5596563816070557, "perplexity": 35.151116499497704, "lr": 0.0026291804804649314, "grad_norm": 0.148896, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:44.877614+00:00", "epoch": 0, "step": 8749, "train_loss": 3.6311326026916504, "perplexity": 37.755554452717234, "lr": 0.0026291804804649314, "grad_norm": 0.162147, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:45.183206+00:00", "epoch": 0, "step": 8750, "train_loss": 3.6137142181396484, "perplexity": 37.10360809576236, "lr": 0.0026291804804649314, "grad_norm": 0.178676, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:45.489097+00:00", "epoch": 0, "step": 8751, "train_loss": 3.57643723487854, "perplexity": 35.74595925352791, "lr": 0.0026291804804649314, "grad_norm": 0.164995, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:45.795436+00:00", "epoch": 0, "step": 8752, "train_loss": 3.4863946437835693, "perplexity": 32.667955515607645, "lr": 0.0026291804804649314, "grad_norm": 0.148392, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:46.101652+00:00", "epoch": 0, "step": 8753, "train_loss": 3.506134271621704, "perplexity": 33.31921546728064, "lr": 0.0026291804804649314, "grad_norm": 0.154312, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:46.407123+00:00", "epoch": 0, "step": 8754, "train_loss": 3.553051471710205, "perplexity": 34.91971158779006, "lr": 0.0026291804804649314, "grad_norm": 0.163307, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:46.711550+00:00", "epoch": 0, "step": 8755, "train_loss": 3.441340923309326, "perplexity": 31.22880553825083, "lr": 0.0026291804804649314, "grad_norm": 0.156261, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:47.016443+00:00", "epoch": 0, "step": 8756, "train_loss": 3.5274977684020996, "perplexity": 34.03868828355723, "lr": 0.0026291804804649314, "grad_norm": 0.133286, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:47.322011+00:00", "epoch": 0, "step": 8757, "train_loss": 3.506721019744873, "perplexity": 33.338771191002586, "lr": 0.0026291804804649314, "grad_norm": 0.130347, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:47.626811+00:00", "epoch": 0, "step": 8758, "train_loss": 3.509610176086426, "perplexity": 33.43523139040916, "lr": 0.0026291804804649314, "grad_norm": 0.149244, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:47.931427+00:00", "epoch": 0, "step": 8759, "train_loss": 3.5735392570495605, "perplexity": 35.64251821344981, "lr": 0.0026291804804649314, "grad_norm": 0.145485, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:48.236099+00:00", "epoch": 0, "step": 8760, "train_loss": 3.462991714477539, "perplexity": 31.912306352663396, "lr": 0.0026291804804649314, "grad_norm": 0.160945, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:48.541238+00:00", "epoch": 0, "step": 8761, "train_loss": 3.485964059829712, "perplexity": 32.65389224608368, "lr": 0.0026291804804649314, "grad_norm": 0.149653, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:48.846927+00:00", "epoch": 0, "step": 8762, "train_loss": 3.434138774871826, "perplexity": 31.00469903809559, "lr": 0.0026291804804649314, "grad_norm": 0.1494, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:49.153609+00:00", "epoch": 0, "step": 8763, "train_loss": 3.5785064697265625, "perplexity": 35.82000261822706, "lr": 0.0026291804804649314, "grad_norm": 0.154993, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:49.458118+00:00", "epoch": 0, "step": 8764, "train_loss": 3.5208213329315186, "perplexity": 33.81218812592123, "lr": 0.0026291804804649314, "grad_norm": 0.158296, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:49.764390+00:00", "epoch": 0, "step": 8765, "train_loss": 3.5681493282318115, "perplexity": 35.450924379794444, "lr": 0.0026291804804649314, "grad_norm": 0.160548, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:50.069746+00:00", "epoch": 0, "step": 8766, "train_loss": 3.529646158218384, "perplexity": 34.11189526523454, "lr": 0.0026291804804649314, "grad_norm": 0.167452, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:50.375392+00:00", "epoch": 0, "step": 8767, "train_loss": 3.5372543334960938, "perplexity": 34.3724143241359, "lr": 0.0026291804804649314, "grad_norm": 0.182172, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:50.681094+00:00", "epoch": 0, "step": 8768, "train_loss": 3.628220319747925, "perplexity": 37.6457595499661, "lr": 0.0026291804804649314, "grad_norm": 0.180175, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:50.986872+00:00", "epoch": 0, "step": 8769, "train_loss": 3.532801628112793, "perplexity": 34.219704328457425, "lr": 0.0026291804804649314, "grad_norm": 0.139994, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:51.292563+00:00", "epoch": 0, "step": 8770, "train_loss": 3.559328079223633, "perplexity": 35.13957819830048, "lr": 0.0026291804804649314, "grad_norm": 0.157062, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:51.597636+00:00", "epoch": 0, "step": 8771, "train_loss": 3.5354042053222656, "perplexity": 34.30887974367409, "lr": 0.0026291804804649314, "grad_norm": 0.138856, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:51.903041+00:00", "epoch": 0, "step": 8772, "train_loss": 3.553461790084839, "perplexity": 34.934042727055264, "lr": 0.0026291804804649314, "grad_norm": 0.157537, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:52.209746+00:00", "epoch": 0, "step": 8773, "train_loss": 3.541958808898926, "perplexity": 34.53449946570046, "lr": 0.0026291804804649314, "grad_norm": 0.149576, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:52.515483+00:00", "epoch": 0, "step": 8774, "train_loss": 3.5100231170654297, "perplexity": 33.44904101867721, "lr": 0.0026291804804649314, "grad_norm": 0.161928, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:52.821500+00:00", "epoch": 0, "step": 8775, "train_loss": 3.452892780303955, "perplexity": 31.591647950810582, "lr": 0.0026291804804649314, "grad_norm": 0.167399, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:53.127290+00:00", "epoch": 0, "step": 8776, "train_loss": 3.5375285148620605, "perplexity": 34.38183989174532, "lr": 0.0026291804804649314, "grad_norm": 0.16836, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:53.431957+00:00", "epoch": 0, "step": 8777, "train_loss": 3.5609281063079834, "perplexity": 35.19584747933273, "lr": 0.0026291804804649314, "grad_norm": 0.178467, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:53.738396+00:00", "epoch": 0, "step": 8778, "train_loss": 3.373541831970215, "perplexity": 29.18170091896113, "lr": 0.0026291804804649314, "grad_norm": 0.212029, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:54.045354+00:00", "epoch": 0, "step": 8779, "train_loss": 3.554805040359497, "perplexity": 34.98099941977809, "lr": 0.0026291804804649314, "grad_norm": 0.204621, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:54.350969+00:00", "epoch": 0, "step": 8780, "train_loss": 3.5670812129974365, "perplexity": 35.41307892264444, "lr": 0.0026291804804649314, "grad_norm": 0.152982, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:54.655412+00:00", "epoch": 0, "step": 8781, "train_loss": 3.4793834686279297, "perplexity": 32.43971580675533, "lr": 0.0026291804804649314, "grad_norm": 0.159354, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:54.961350+00:00", "epoch": 0, "step": 8782, "train_loss": 3.4772865772247314, "perplexity": 32.3717645136973, "lr": 0.0026291804804649314, "grad_norm": 0.151468, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:55.267210+00:00", "epoch": 0, "step": 8783, "train_loss": 3.494856119155884, "perplexity": 32.94554737880714, "lr": 0.0026291804804649314, "grad_norm": 0.153797, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:55.572084+00:00", "epoch": 0, "step": 8784, "train_loss": 3.6008989810943604, "perplexity": 36.63115035770523, "lr": 0.0026291804804649314, "grad_norm": 0.156415, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:55.878586+00:00", "epoch": 0, "step": 8785, "train_loss": 3.5368924140930176, "perplexity": 34.359976531336464, "lr": 0.0026291804804649314, "grad_norm": 0.165967, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:56.184018+00:00", "epoch": 0, "step": 8786, "train_loss": 3.4871573448181152, "perplexity": 32.692880903177056, "lr": 0.0026291804804649314, "grad_norm": 0.173881, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:56.488531+00:00", "epoch": 0, "step": 8787, "train_loss": 3.5568840503692627, "perplexity": 35.05380091902385, "lr": 0.0026291804804649314, "grad_norm": 0.180752, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:56.793442+00:00", "epoch": 0, "step": 8788, "train_loss": 3.520012140274048, "perplexity": 33.78483861855935, "lr": 0.0026291804804649314, "grad_norm": 0.173018, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:57.098394+00:00", "epoch": 0, "step": 8789, "train_loss": 3.5453343391418457, "perplexity": 34.651268681205885, "lr": 0.0026291804804649314, "grad_norm": 0.169501, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:57.402655+00:00", "epoch": 0, "step": 8790, "train_loss": 3.5772342681884766, "perplexity": 35.77446133079224, "lr": 0.0026291804804649314, "grad_norm": 0.170159, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:57.708065+00:00", "epoch": 0, "step": 8791, "train_loss": 3.699340581893921, "perplexity": 40.42064146719359, "lr": 0.0026291804804649314, "grad_norm": 0.169247, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:58.013231+00:00", "epoch": 0, "step": 8792, "train_loss": 3.671982765197754, "perplexity": 39.329810380853324, "lr": 0.0026291804804649314, "grad_norm": 0.151825, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:58.318579+00:00", "epoch": 0, "step": 8793, "train_loss": 3.514747381210327, "perplexity": 33.60743698142786, "lr": 0.0026291804804649314, "grad_norm": 0.165834, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:58.623465+00:00", "epoch": 0, "step": 8794, "train_loss": 3.600560188293457, "perplexity": 36.618742089709855, "lr": 0.0026291804804649314, "grad_norm": 0.150349, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:58.929800+00:00", "epoch": 0, "step": 8795, "train_loss": 3.513120651245117, "perplexity": 33.552811199380656, "lr": 0.0026291804804649314, "grad_norm": 0.169674, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:59.235703+00:00", "epoch": 0, "step": 8796, "train_loss": 3.495368719100952, "perplexity": 32.962439593688096, "lr": 0.0026291804804649314, "grad_norm": 0.188876, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:59.540770+00:00", "epoch": 0, "step": 8797, "train_loss": 3.5144336223602295, "perplexity": 33.59689400470845, "lr": 0.0026291804804649314, "grad_norm": 0.161153, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:43:59.845588+00:00", "epoch": 0, "step": 8798, "train_loss": 3.5346145629882812, "perplexity": 34.281798693375954, "lr": 0.0026291804804649314, "grad_norm": 0.175539, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:00.150699+00:00", "epoch": 0, "step": 8799, "train_loss": 3.5356764793395996, "perplexity": 34.31822243202006, "lr": 0.0026291804804649314, "grad_norm": 0.180586, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:00.455455+00:00", "epoch": 0, "step": 8800, "train_loss": 3.5172479152679443, "perplexity": 33.69157867793552, "lr": 0.0026291804804649314, "grad_norm": 0.152417, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:00.760966+00:00", "epoch": 0, "step": 8801, "train_loss": 3.5574440956115723, "perplexity": 35.07343813180395, "lr": 0.0026291804804649314, "grad_norm": 0.169352, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:01.067127+00:00", "epoch": 0, "step": 8802, "train_loss": 3.6078734397888184, "perplexity": 36.88752580354419, "lr": 0.0026291804804649314, "grad_norm": 0.18043, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:01.372960+00:00", "epoch": 0, "step": 8803, "train_loss": 3.707110643386841, "perplexity": 40.73593567838929, "lr": 0.0026291804804649314, "grad_norm": 0.187547, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:01.677506+00:00", "epoch": 0, "step": 8804, "train_loss": 3.563197374343872, "perplexity": 35.27580698147164, "lr": 0.0026291804804649314, "grad_norm": 0.18582, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:01.982578+00:00", "epoch": 0, "step": 8805, "train_loss": 3.5094571113586426, "perplexity": 33.43011402747189, "lr": 0.0026291804804649314, "grad_norm": 0.236439, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:02.288251+00:00", "epoch": 0, "step": 8806, "train_loss": 3.641322374343872, "perplexity": 38.14224171278155, "lr": 0.0026291804804649314, "grad_norm": 0.169815, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:02.593762+00:00", "epoch": 0, "step": 8807, "train_loss": 3.6788504123687744, "perplexity": 39.6008432556684, "lr": 0.0026291804804649314, "grad_norm": 0.167058, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:02.898965+00:00", "epoch": 0, "step": 8808, "train_loss": 3.607839345932007, "perplexity": 36.88626818695994, "lr": 0.0026291804804649314, "grad_norm": 0.180621, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:03.204187+00:00", "epoch": 0, "step": 8809, "train_loss": 3.6098721027374268, "perplexity": 36.96132526018644, "lr": 0.0026291804804649314, "grad_norm": 0.16467, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:03.508855+00:00", "epoch": 0, "step": 8810, "train_loss": 3.5168471336364746, "perplexity": 33.67807841757507, "lr": 0.0026291804804649314, "grad_norm": 0.159918, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:03.813723+00:00", "epoch": 0, "step": 8811, "train_loss": 3.5917470455169678, "perplexity": 36.297433834845265, "lr": 0.0026291804804649314, "grad_norm": 0.179522, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:04.120074+00:00", "epoch": 0, "step": 8812, "train_loss": 3.5971407890319824, "perplexity": 36.49374182475395, "lr": 0.0026291804804649314, "grad_norm": 0.18676, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:04.425681+00:00", "epoch": 0, "step": 8813, "train_loss": 3.615147590637207, "perplexity": 37.15682952111851, "lr": 0.0026291804804649314, "grad_norm": 0.202705, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:04.731761+00:00", "epoch": 0, "step": 8814, "train_loss": 3.5877842903137207, "perplexity": 36.15388061105042, "lr": 0.0026291804804649314, "grad_norm": 0.144225, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:05.038186+00:00", "epoch": 0, "step": 8815, "train_loss": 3.709116220474243, "perplexity": 40.81771671929809, "lr": 0.0026291804804649314, "grad_norm": 0.150459, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:05.343547+00:00", "epoch": 0, "step": 8816, "train_loss": 3.463103771209717, "perplexity": 31.91588254179382, "lr": 0.0026291804804649314, "grad_norm": 0.148247, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:05.649306+00:00", "epoch": 0, "step": 8817, "train_loss": 3.5039501190185547, "perplexity": 33.246520633287695, "lr": 0.0026291804804649314, "grad_norm": 0.145984, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:05.954678+00:00", "epoch": 0, "step": 8818, "train_loss": 3.5268847942352295, "perplexity": 34.01782984046176, "lr": 0.0026291804804649314, "grad_norm": 0.13521, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:06.258660+00:00", "epoch": 0, "step": 8819, "train_loss": 3.527892589569092, "perplexity": 34.05213013157667, "lr": 0.0026291804804649314, "grad_norm": 0.152149, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:06.563375+00:00", "epoch": 0, "step": 8820, "train_loss": 3.447432279586792, "perplexity": 31.419611863575113, "lr": 0.0026291804804649314, "grad_norm": 0.149121, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:06.868249+00:00", "epoch": 0, "step": 8821, "train_loss": 3.560180425643921, "perplexity": 35.169542059969295, "lr": 0.0026291804804649314, "grad_norm": 0.152819, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:07.174227+00:00", "epoch": 0, "step": 8822, "train_loss": 3.5015084743499756, "perplexity": 33.16544346453237, "lr": 0.0026291804804649314, "grad_norm": 0.172031, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:07.479970+00:00", "epoch": 0, "step": 8823, "train_loss": 3.346407651901245, "perplexity": 28.400525596111525, "lr": 0.0026291804804649314, "grad_norm": 0.159649, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:07.784439+00:00", "epoch": 0, "step": 8824, "train_loss": 3.60851788520813, "perplexity": 36.91130546210104, "lr": 0.0026291804804649314, "grad_norm": 0.15942, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:08.090560+00:00", "epoch": 0, "step": 8825, "train_loss": 3.6067471504211426, "perplexity": 36.846003163075416, "lr": 0.0026291804804649314, "grad_norm": 0.175605, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:08.395049+00:00", "epoch": 0, "step": 8826, "train_loss": 3.48539662361145, "perplexity": 32.635368500975, "lr": 0.0026291804804649314, "grad_norm": 0.153252, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:08.699562+00:00", "epoch": 0, "step": 8827, "train_loss": 3.599591016769409, "perplexity": 36.58326943994767, "lr": 0.0026291804804649314, "grad_norm": 0.173755, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:09.005646+00:00", "epoch": 0, "step": 8828, "train_loss": 3.5978353023529053, "perplexity": 36.519096017974896, "lr": 0.0026291804804649314, "grad_norm": 0.154312, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:09.311643+00:00", "epoch": 0, "step": 8829, "train_loss": 3.6750385761260986, "perplexity": 39.45017866291436, "lr": 0.0026291804804649314, "grad_norm": 0.160812, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:09.616465+00:00", "epoch": 0, "step": 8830, "train_loss": 3.5483899116516113, "perplexity": 34.757310071290085, "lr": 0.0026291804804649314, "grad_norm": 0.153123, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:09.922024+00:00", "epoch": 0, "step": 8831, "train_loss": 3.5589303970336914, "perplexity": 35.12560659220255, "lr": 0.0026291804804649314, "grad_norm": 0.148162, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:10.229301+00:00", "epoch": 0, "step": 8832, "train_loss": 3.512272357940674, "perplexity": 33.524360643209214, "lr": 0.0026291804804649314, "grad_norm": 0.157346, "tokens_per_sec": 106640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:10.534944+00:00", "epoch": 0, "step": 8833, "train_loss": 3.5453922748565674, "perplexity": 34.653276285378375, "lr": 0.0026291804804649314, "grad_norm": 0.152833, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:10.840772+00:00", "epoch": 0, "step": 8834, "train_loss": 3.5320487022399902, "perplexity": 34.19394912480518, "lr": 0.0026291804804649314, "grad_norm": 0.156593, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:11.146406+00:00", "epoch": 0, "step": 8835, "train_loss": 3.4739112854003906, "perplexity": 32.26268455350228, "lr": 0.0026291804804649314, "grad_norm": 0.153392, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:11.450510+00:00", "epoch": 0, "step": 8836, "train_loss": 3.514894962310791, "perplexity": 33.61239716996739, "lr": 0.0026291804804649314, "grad_norm": 0.165876, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:11.755614+00:00", "epoch": 0, "step": 8837, "train_loss": 3.555183172225952, "perplexity": 34.994229351550906, "lr": 0.0026291804804649314, "grad_norm": 0.155304, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:12.060192+00:00", "epoch": 0, "step": 8838, "train_loss": 3.511026620864868, "perplexity": 33.482624105986645, "lr": 0.0026291804804649314, "grad_norm": 0.155289, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:12.365582+00:00", "epoch": 0, "step": 8839, "train_loss": 3.4245684146881104, "perplexity": 30.709388270121277, "lr": 0.0026291804804649314, "grad_norm": 0.152288, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:12.671211+00:00", "epoch": 0, "step": 8840, "train_loss": 3.4914116859436035, "perplexity": 32.83226385199067, "lr": 0.0026291804804649314, "grad_norm": 0.148661, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:12.976504+00:00", "epoch": 0, "step": 8841, "train_loss": 3.4692766666412354, "perplexity": 32.113505272025094, "lr": 0.0026291804804649314, "grad_norm": 0.142772, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:13.281871+00:00", "epoch": 0, "step": 8842, "train_loss": 3.5340049266815186, "perplexity": 34.26090563346237, "lr": 0.0026291804804649314, "grad_norm": 0.16762, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:13.587972+00:00", "epoch": 0, "step": 8843, "train_loss": 3.584582805633545, "perplexity": 36.03831959825486, "lr": 0.0026291804804649314, "grad_norm": 0.163658, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:13.895069+00:00", "epoch": 0, "step": 8844, "train_loss": 3.4877045154571533, "perplexity": 32.710774382659714, "lr": 0.0026291804804649314, "grad_norm": 0.143253, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:14.200571+00:00", "epoch": 0, "step": 8845, "train_loss": 3.5643224716186523, "perplexity": 35.31551803097996, "lr": 0.0026291804804649314, "grad_norm": 0.143642, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:14.505864+00:00", "epoch": 0, "step": 8846, "train_loss": 3.4857442378997803, "perplexity": 32.64671499335804, "lr": 0.0026291804804649314, "grad_norm": 0.145459, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:14.811669+00:00", "epoch": 0, "step": 8847, "train_loss": 3.579383134841919, "perplexity": 35.85141853356092, "lr": 0.0026291804804649314, "grad_norm": 0.144762, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:15.119308+00:00", "epoch": 0, "step": 8848, "train_loss": 3.5370309352874756, "perplexity": 34.364736445994055, "lr": 0.0026291804804649314, "grad_norm": 0.181105, "tokens_per_sec": 106515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:15.425445+00:00", "epoch": 0, "step": 8849, "train_loss": 3.5302491188049316, "perplexity": 34.13246959574378, "lr": 0.0026291804804649314, "grad_norm": 0.208776, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:15.731276+00:00", "epoch": 0, "step": 8850, "train_loss": 3.5379598140716553, "perplexity": 34.39667195041272, "lr": 0.0026291804804649314, "grad_norm": 0.190449, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:16.036253+00:00", "epoch": 0, "step": 8851, "train_loss": 3.4022350311279297, "perplexity": 30.03114564030716, "lr": 0.0026291804804649314, "grad_norm": 0.154572, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:16.342459+00:00", "epoch": 0, "step": 8852, "train_loss": 3.594531536102295, "perplexity": 36.39864454236723, "lr": 0.0026291804804649314, "grad_norm": 0.181843, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:16.648665+00:00", "epoch": 0, "step": 8853, "train_loss": 3.5218122005462646, "perplexity": 33.84570813232076, "lr": 0.0026291804804649314, "grad_norm": 0.146511, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:16.955853+00:00", "epoch": 0, "step": 8854, "train_loss": 3.5299878120422363, "perplexity": 34.12355171581283, "lr": 0.0026291804804649314, "grad_norm": 0.158701, "tokens_per_sec": 106670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:17.262521+00:00", "epoch": 0, "step": 8855, "train_loss": 3.5719101428985596, "perplexity": 35.58449975482876, "lr": 0.0026291804804649314, "grad_norm": 0.166091, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:17.569280+00:00", "epoch": 0, "step": 8856, "train_loss": 3.5064916610717773, "perplexity": 33.33112553151225, "lr": 0.0026291804804649314, "grad_norm": 0.153131, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:17.875183+00:00", "epoch": 0, "step": 8857, "train_loss": 3.583813190460205, "perplexity": 36.010594630816655, "lr": 0.0026291804804649314, "grad_norm": 0.161073, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:18.180817+00:00", "epoch": 0, "step": 8858, "train_loss": 3.5058205127716064, "perplexity": 33.30876290842665, "lr": 0.0026291804804649314, "grad_norm": 0.16005, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:18.486458+00:00", "epoch": 0, "step": 8859, "train_loss": 3.5645296573638916, "perplexity": 35.32283566092985, "lr": 0.0026291804804649314, "grad_norm": 0.153313, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:18.791162+00:00", "epoch": 0, "step": 8860, "train_loss": 3.478142261505127, "perplexity": 32.399476378332935, "lr": 0.0026291804804649314, "grad_norm": 0.171833, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:19.096828+00:00", "epoch": 0, "step": 8861, "train_loss": 3.561936140060425, "perplexity": 35.23134396936194, "lr": 0.0026291804804649314, "grad_norm": 0.173861, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:19.402450+00:00", "epoch": 0, "step": 8862, "train_loss": 3.4722113609313965, "perplexity": 32.20788701562798, "lr": 0.0026291804804649314, "grad_norm": 0.154177, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:19.709886+00:00", "epoch": 0, "step": 8863, "train_loss": 3.532949447631836, "perplexity": 34.224763042572185, "lr": 0.0026291804804649314, "grad_norm": 0.14832, "tokens_per_sec": 106580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:20.016911+00:00", "epoch": 0, "step": 8864, "train_loss": 3.5736567974090576, "perplexity": 35.64670789407746, "lr": 0.0026291804804649314, "grad_norm": 0.160333, "tokens_per_sec": 106728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:20.321996+00:00", "epoch": 0, "step": 8865, "train_loss": 3.459381341934204, "perplexity": 31.79729877304206, "lr": 0.0026291804804649314, "grad_norm": 0.148073, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:20.626250+00:00", "epoch": 0, "step": 8866, "train_loss": 3.5974299907684326, "perplexity": 36.504297404531656, "lr": 0.0026291804804649314, "grad_norm": 0.1453, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:20.930563+00:00", "epoch": 0, "step": 8867, "train_loss": 3.381434679031372, "perplexity": 29.41293898418813, "lr": 0.0026291804804649314, "grad_norm": 0.162946, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:21.236677+00:00", "epoch": 0, "step": 8868, "train_loss": 3.460188388824463, "perplexity": 31.822971042104513, "lr": 0.0026291804804649314, "grad_norm": 0.176174, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:21.541589+00:00", "epoch": 0, "step": 8869, "train_loss": 3.484038829803467, "perplexity": 32.59108646944044, "lr": 0.0026291804804649314, "grad_norm": 0.149271, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:21.847646+00:00", "epoch": 0, "step": 8870, "train_loss": 3.610050678253174, "perplexity": 36.96792623727638, "lr": 0.0026291804804649314, "grad_norm": 0.179266, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:22.153077+00:00", "epoch": 0, "step": 8871, "train_loss": 3.46938157081604, "perplexity": 32.11687428950465, "lr": 0.0026291804804649314, "grad_norm": 0.174927, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:22.459236+00:00", "epoch": 0, "step": 8872, "train_loss": 3.559891939163208, "perplexity": 35.159397585894304, "lr": 0.0026291804804649314, "grad_norm": 0.157809, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:22.764887+00:00", "epoch": 0, "step": 8873, "train_loss": 3.5231032371520996, "perplexity": 33.88943239921134, "lr": 0.0026291804804649314, "grad_norm": 0.196097, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:23.071177+00:00", "epoch": 0, "step": 8874, "train_loss": 3.6147494316101074, "perplexity": 37.14203813888287, "lr": 0.0026291804804649314, "grad_norm": 0.160656, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:23.375326+00:00", "epoch": 0, "step": 8875, "train_loss": 3.4864964485168457, "perplexity": 32.67128143740002, "lr": 0.0026291804804649314, "grad_norm": 0.164027, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:23.679611+00:00", "epoch": 0, "step": 8876, "train_loss": 3.5098319053649902, "perplexity": 33.44264578210747, "lr": 0.0026291804804649314, "grad_norm": 0.150701, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:23.984536+00:00", "epoch": 0, "step": 8877, "train_loss": 3.6415653228759766, "perplexity": 38.15150944016164, "lr": 0.0026291804804649314, "grad_norm": 0.167814, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:24.289964+00:00", "epoch": 0, "step": 8878, "train_loss": 3.592905044555664, "perplexity": 36.33949057446548, "lr": 0.0026291804804649314, "grad_norm": 0.160171, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:24.595390+00:00", "epoch": 0, "step": 8879, "train_loss": 3.5167644023895264, "perplexity": 33.67529230340371, "lr": 0.0026291804804649314, "grad_norm": 0.17649, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:24.901090+00:00", "epoch": 0, "step": 8880, "train_loss": 3.5358853340148926, "perplexity": 34.32539070175999, "lr": 0.0026291804804649314, "grad_norm": 0.15159, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:25.205638+00:00", "epoch": 0, "step": 8881, "train_loss": 3.511760950088501, "perplexity": 33.507220405136934, "lr": 0.0026291804804649314, "grad_norm": 0.184642, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:25.510374+00:00", "epoch": 0, "step": 8882, "train_loss": 3.3835113048553467, "perplexity": 29.47408211657285, "lr": 0.0026291804804649314, "grad_norm": 0.151298, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:25.816431+00:00", "epoch": 0, "step": 8883, "train_loss": 3.568902015686035, "perplexity": 35.47761789049081, "lr": 0.0026291804804649314, "grad_norm": 0.185488, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:26.123189+00:00", "epoch": 0, "step": 8884, "train_loss": 3.4842209815979004, "perplexity": 32.59702353503055, "lr": 0.0026291804804649314, "grad_norm": 0.159208, "tokens_per_sec": 106821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:26.427790+00:00", "epoch": 0, "step": 8885, "train_loss": 3.5159389972686768, "perplexity": 33.64750801291563, "lr": 0.0026291804804649314, "grad_norm": 0.172479, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:26.733204+00:00", "epoch": 0, "step": 8886, "train_loss": 3.5712482929229736, "perplexity": 35.56095594662867, "lr": 0.0026291804804649314, "grad_norm": 0.150063, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:27.037846+00:00", "epoch": 0, "step": 8887, "train_loss": 3.6009023189544678, "perplexity": 36.63127262756476, "lr": 0.0026291804804649314, "grad_norm": 0.163167, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:27.342446+00:00", "epoch": 0, "step": 8888, "train_loss": 3.623522996902466, "perplexity": 37.46933993792352, "lr": 0.0026291804804649314, "grad_norm": 0.20811, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:27.646844+00:00", "epoch": 0, "step": 8889, "train_loss": 3.510329008102417, "perplexity": 33.45927434558234, "lr": 0.0026291804804649314, "grad_norm": 0.197971, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:27.952487+00:00", "epoch": 0, "step": 8890, "train_loss": 3.584794282913208, "perplexity": 36.04594168996833, "lr": 0.0026291804804649314, "grad_norm": 0.167593, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:28.257976+00:00", "epoch": 0, "step": 8891, "train_loss": 3.5025389194488525, "perplexity": 33.19963624706953, "lr": 0.0026291804804649314, "grad_norm": 0.150062, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:28.564252+00:00", "epoch": 0, "step": 8892, "train_loss": 3.5359201431274414, "perplexity": 34.32658555894405, "lr": 0.0026291804804649314, "grad_norm": 0.138753, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:28.870223+00:00", "epoch": 0, "step": 8893, "train_loss": 3.576866865158081, "perplexity": 35.76132009950046, "lr": 0.0026291804804649314, "grad_norm": 0.161173, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:29.176763+00:00", "epoch": 0, "step": 8894, "train_loss": 3.5186846256256104, "perplexity": 33.74001850665785, "lr": 0.0026291804804649314, "grad_norm": 0.168687, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:29.483467+00:00", "epoch": 0, "step": 8895, "train_loss": 3.4894649982452393, "perplexity": 32.76841185795063, "lr": 0.0026291804804649314, "grad_norm": 0.190957, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:29.788448+00:00", "epoch": 0, "step": 8896, "train_loss": 3.586958169937134, "perplexity": 36.124025487245945, "lr": 0.0026291804804649314, "grad_norm": 0.160466, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:30.093959+00:00", "epoch": 0, "step": 8897, "train_loss": 3.488008499145508, "perplexity": 32.720719435995896, "lr": 0.0026291804804649314, "grad_norm": 0.157503, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:30.398600+00:00", "epoch": 0, "step": 8898, "train_loss": 3.573317527770996, "perplexity": 35.63461609969749, "lr": 0.0026291804804649314, "grad_norm": 0.15053, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:30.704667+00:00", "epoch": 0, "step": 8899, "train_loss": 3.6112613677978516, "perplexity": 37.01271002322075, "lr": 0.0026291804804649314, "grad_norm": 0.171558, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:31.009984+00:00", "epoch": 0, "step": 8900, "train_loss": 3.6661810874938965, "perplexity": 39.10229212883295, "lr": 0.0026291804804649314, "grad_norm": 0.175718, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:31.315916+00:00", "epoch": 0, "step": 8901, "train_loss": 3.5233492851257324, "perplexity": 33.8977718512913, "lr": 0.0026291804804649314, "grad_norm": 0.174144, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:31.621210+00:00", "epoch": 0, "step": 8902, "train_loss": 3.513592004776001, "perplexity": 33.56863016327993, "lr": 0.0026291804804649314, "grad_norm": 0.166812, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:31.926193+00:00", "epoch": 0, "step": 8903, "train_loss": 3.5661962032318115, "perplexity": 35.381751866387134, "lr": 0.0026291804804649314, "grad_norm": 0.178518, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:32.230935+00:00", "epoch": 0, "step": 8904, "train_loss": 3.5482230186462402, "perplexity": 34.75150980337921, "lr": 0.0026291804804649314, "grad_norm": 0.166651, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:32.536859+00:00", "epoch": 0, "step": 8905, "train_loss": 3.5839483737945557, "perplexity": 36.015462992124036, "lr": 0.0026291804804649314, "grad_norm": 0.166586, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:32.843500+00:00", "epoch": 0, "step": 8906, "train_loss": 3.6348230838775635, "perplexity": 37.895148041410444, "lr": 0.0026291804804649314, "grad_norm": 0.164417, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:33.149885+00:00", "epoch": 0, "step": 8907, "train_loss": 3.572361469268799, "perplexity": 35.60056360268628, "lr": 0.0026291804804649314, "grad_norm": 0.145021, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:33.455039+00:00", "epoch": 0, "step": 8908, "train_loss": 3.5310025215148926, "perplexity": 34.15819478032993, "lr": 0.0026291804804649314, "grad_norm": 0.146481, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:33.760239+00:00", "epoch": 0, "step": 8909, "train_loss": 3.5589077472686768, "perplexity": 35.124811014477096, "lr": 0.0026291804804649314, "grad_norm": 0.161851, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:34.065891+00:00", "epoch": 0, "step": 8910, "train_loss": 3.5453178882598877, "perplexity": 34.650698641963956, "lr": 0.0026291804804649314, "grad_norm": 0.160218, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:34.372753+00:00", "epoch": 0, "step": 8911, "train_loss": 3.6752538681030273, "perplexity": 39.45867288420493, "lr": 0.0026291804804649314, "grad_norm": 0.167213, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:34.678697+00:00", "epoch": 0, "step": 8912, "train_loss": 3.54557204246521, "perplexity": 34.65950638195583, "lr": 0.0026291804804649314, "grad_norm": 0.184012, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:34.984299+00:00", "epoch": 0, "step": 8913, "train_loss": 3.476574182510376, "perplexity": 32.34871125224915, "lr": 0.0026291804804649314, "grad_norm": 0.145521, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:35.289417+00:00", "epoch": 0, "step": 8914, "train_loss": 3.5814342498779297, "perplexity": 35.925029383530834, "lr": 0.0026291804804649314, "grad_norm": 0.162695, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:35.594810+00:00", "epoch": 0, "step": 8915, "train_loss": 3.610985517501831, "perplexity": 37.00250146428606, "lr": 0.0026291804804649314, "grad_norm": 0.151826, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:35.901225+00:00", "epoch": 0, "step": 8916, "train_loss": 3.5386977195739746, "perplexity": 34.422062810782066, "lr": 0.0026291804804649314, "grad_norm": 0.161302, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:36.207949+00:00", "epoch": 0, "step": 8917, "train_loss": 3.45497465133667, "perplexity": 31.65748619714444, "lr": 0.0026291804804649314, "grad_norm": 0.148554, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:36.512348+00:00", "epoch": 0, "step": 8918, "train_loss": 3.585428476333618, "perplexity": 36.068809039415434, "lr": 0.0026291804804649314, "grad_norm": 0.16811, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:36.817418+00:00", "epoch": 0, "step": 8919, "train_loss": 3.4296114444732666, "perplexity": 30.86464778999453, "lr": 0.0026291804804649314, "grad_norm": 0.181766, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:37.122829+00:00", "epoch": 0, "step": 8920, "train_loss": 3.524923086166382, "perplexity": 33.95116220178115, "lr": 0.0026291804804649314, "grad_norm": 0.173078, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:37.427127+00:00", "epoch": 0, "step": 8921, "train_loss": 3.5698611736297607, "perplexity": 35.51166285415354, "lr": 0.0026291804804649314, "grad_norm": 0.202028, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:37.732376+00:00", "epoch": 0, "step": 8922, "train_loss": 3.5216593742370605, "perplexity": 33.84053601289252, "lr": 0.0026291804804649314, "grad_norm": 0.150926, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:38.038939+00:00", "epoch": 0, "step": 8923, "train_loss": 3.475548267364502, "perplexity": 32.3155412371457, "lr": 0.0026291804804649314, "grad_norm": 0.151608, "tokens_per_sec": 106888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:38.344474+00:00", "epoch": 0, "step": 8924, "train_loss": 3.5162765979766846, "perplexity": 33.65886935313383, "lr": 0.0026291804804649314, "grad_norm": 0.140048, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:38.650026+00:00", "epoch": 0, "step": 8925, "train_loss": 3.5364556312561035, "perplexity": 34.344971960418874, "lr": 0.0026291804804649314, "grad_norm": 0.143352, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:38.955145+00:00", "epoch": 0, "step": 8926, "train_loss": 3.5186221599578857, "perplexity": 33.73791097969751, "lr": 0.0026291804804649314, "grad_norm": 0.155458, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:39.261138+00:00", "epoch": 0, "step": 8927, "train_loss": 3.644712209701538, "perplexity": 38.27175702613295, "lr": 0.0026291804804649314, "grad_norm": 0.140945, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:39.567247+00:00", "epoch": 0, "step": 8928, "train_loss": 3.552922248840332, "perplexity": 34.91519945398527, "lr": 0.0026291804804649314, "grad_norm": 0.145802, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:39.873421+00:00", "epoch": 0, "step": 8929, "train_loss": 3.5426034927368164, "perplexity": 34.55677047746496, "lr": 0.0026291804804649314, "grad_norm": 0.148221, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:40.178541+00:00", "epoch": 0, "step": 8930, "train_loss": 3.5938656330108643, "perplexity": 36.374414640721135, "lr": 0.0026291804804649314, "grad_norm": 0.192896, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:40.483330+00:00", "epoch": 0, "step": 8931, "train_loss": 3.5662648677825928, "perplexity": 35.38418142189611, "lr": 0.0026291804804649314, "grad_norm": 0.172186, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:40.788337+00:00", "epoch": 0, "step": 8932, "train_loss": 3.5270285606384277, "perplexity": 34.02272081307288, "lr": 0.0026291804804649314, "grad_norm": 0.147483, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:41.094133+00:00", "epoch": 0, "step": 8933, "train_loss": 3.552725076675415, "perplexity": 34.908315827172515, "lr": 0.0026291804804649314, "grad_norm": 0.152635, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:41.399512+00:00", "epoch": 0, "step": 8934, "train_loss": 3.575619697570801, "perplexity": 35.716747540696765, "lr": 0.0026291804804649314, "grad_norm": 0.164392, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:41.704331+00:00", "epoch": 0, "step": 8935, "train_loss": 3.561849355697632, "perplexity": 35.22828657229426, "lr": 0.0026291804804649314, "grad_norm": 0.170763, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:42.009133+00:00", "epoch": 0, "step": 8936, "train_loss": 3.5226943492889404, "perplexity": 33.87557825420284, "lr": 0.0026291804804649314, "grad_norm": 0.16281, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:42.314032+00:00", "epoch": 0, "step": 8937, "train_loss": 3.5112252235412598, "perplexity": 33.48927450511838, "lr": 0.0026291804804649314, "grad_norm": 0.151094, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:42.618603+00:00", "epoch": 0, "step": 8938, "train_loss": 3.624581813812256, "perplexity": 37.509034119375336, "lr": 0.0026291804804649314, "grad_norm": 0.160252, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:42.924280+00:00", "epoch": 0, "step": 8939, "train_loss": 3.521085739135742, "perplexity": 33.821129460260146, "lr": 0.0026291804804649314, "grad_norm": 0.161729, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:43.230292+00:00", "epoch": 0, "step": 8940, "train_loss": 3.5014567375183105, "perplexity": 33.16372763395295, "lr": 0.0026291804804649314, "grad_norm": 0.157214, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:43.534364+00:00", "epoch": 0, "step": 8941, "train_loss": 3.59356951713562, "perplexity": 36.36364519367412, "lr": 0.0026291804804649314, "grad_norm": 0.182844, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:43.839663+00:00", "epoch": 0, "step": 8942, "train_loss": 3.694795846939087, "perplexity": 40.23735716976953, "lr": 0.0026291804804649314, "grad_norm": 0.162729, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:44.145258+00:00", "epoch": 0, "step": 8943, "train_loss": 3.545498847961426, "perplexity": 34.656969589425564, "lr": 0.0026291804804649314, "grad_norm": 0.156727, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:44.451315+00:00", "epoch": 0, "step": 8944, "train_loss": 3.545783281326294, "perplexity": 34.6668285899499, "lr": 0.0026291804804649314, "grad_norm": 0.175757, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:44.755821+00:00", "epoch": 0, "step": 8945, "train_loss": 3.5535800457000732, "perplexity": 34.93817411804585, "lr": 0.0026291804804649314, "grad_norm": 0.198065, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:45.060469+00:00", "epoch": 0, "step": 8946, "train_loss": 3.57462215423584, "perplexity": 35.681136302070975, "lr": 0.0026291804804649314, "grad_norm": 0.185264, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:45.365720+00:00", "epoch": 0, "step": 8947, "train_loss": 3.601396322250366, "perplexity": 36.649373067446106, "lr": 0.0026291804804649314, "grad_norm": 0.157315, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:45.671927+00:00", "epoch": 0, "step": 8948, "train_loss": 3.548156499862671, "perplexity": 34.749198252101536, "lr": 0.0026291804804649314, "grad_norm": 0.167494, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:45.978392+00:00", "epoch": 0, "step": 8949, "train_loss": 3.586782693862915, "perplexity": 36.11768714119877, "lr": 0.0026291804804649314, "grad_norm": 0.170072, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:46.286281+00:00", "epoch": 0, "step": 8950, "train_loss": 3.4588756561279297, "perplexity": 31.781223395262415, "lr": 0.0026291804804649314, "grad_norm": 0.161976, "tokens_per_sec": 106491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:46.590675+00:00", "epoch": 0, "step": 8951, "train_loss": 3.4760518074035645, "perplexity": 32.33181750357459, "lr": 0.0026291804804649314, "grad_norm": 0.171917, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:46.895046+00:00", "epoch": 0, "step": 8952, "train_loss": 3.513979196548462, "perplexity": 33.581630177278754, "lr": 0.0026291804804649314, "grad_norm": 0.159536, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:47.199784+00:00", "epoch": 0, "step": 8953, "train_loss": 3.612332344055176, "perplexity": 37.0523709910614, "lr": 0.0026291804804649314, "grad_norm": 0.175032, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:47.505372+00:00", "epoch": 0, "step": 8954, "train_loss": 3.4822213649749756, "perplexity": 32.53190711055159, "lr": 0.0026291804804649314, "grad_norm": 0.166733, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:47.810907+00:00", "epoch": 0, "step": 8955, "train_loss": 3.5249216556549072, "perplexity": 33.95111363428878, "lr": 0.0026291804804649314, "grad_norm": 0.152465, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:48.116206+00:00", "epoch": 0, "step": 8956, "train_loss": 3.556546449661255, "perplexity": 35.04196872840557, "lr": 0.0026291804804649314, "grad_norm": 0.150954, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:48.420903+00:00", "epoch": 0, "step": 8957, "train_loss": 3.6099588871002197, "perplexity": 36.96453306443875, "lr": 0.0026291804804649314, "grad_norm": 0.168961, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:48.725606+00:00", "epoch": 0, "step": 8958, "train_loss": 3.5174612998962402, "perplexity": 33.698768710022286, "lr": 0.0026291804804649314, "grad_norm": 0.175415, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:49.031913+00:00", "epoch": 0, "step": 8959, "train_loss": 3.5851998329162598, "perplexity": 36.06056308638398, "lr": 0.0026291804804649314, "grad_norm": 0.173515, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:49.338649+00:00", "epoch": 0, "step": 8960, "train_loss": 3.6014816761016846, "perplexity": 36.652501366090085, "lr": 0.0026291804804649314, "grad_norm": 0.176178, "tokens_per_sec": 106886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:49.644313+00:00", "epoch": 0, "step": 8961, "train_loss": 3.549309015274048, "perplexity": 34.789270326025836, "lr": 0.0026291804804649314, "grad_norm": 0.180745, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:49.949214+00:00", "epoch": 0, "step": 8962, "train_loss": 3.5231072902679443, "perplexity": 33.88956975728513, "lr": 0.0026291804804649314, "grad_norm": 0.155998, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:50.253945+00:00", "epoch": 0, "step": 8963, "train_loss": 3.588846445083618, "perplexity": 36.19230202892572, "lr": 0.0026291804804649314, "grad_norm": 0.162403, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:50.558729+00:00", "epoch": 0, "step": 8964, "train_loss": 3.601529121398926, "perplexity": 36.65424039616611, "lr": 0.0026291804804649314, "grad_norm": 0.163568, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:50.865331+00:00", "epoch": 0, "step": 8965, "train_loss": 3.512929677963257, "perplexity": 33.54640412072016, "lr": 0.0026291804804649314, "grad_norm": 0.159857, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:51.171652+00:00", "epoch": 0, "step": 8966, "train_loss": 3.5812644958496094, "perplexity": 35.91893148266171, "lr": 0.0026291804804649314, "grad_norm": 0.150036, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:51.476390+00:00", "epoch": 0, "step": 8967, "train_loss": 3.5329127311706543, "perplexity": 34.223506453457375, "lr": 0.0026291804804649314, "grad_norm": 0.165703, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:51.781366+00:00", "epoch": 0, "step": 8968, "train_loss": 3.4820916652679443, "perplexity": 32.52768800534445, "lr": 0.0026291804804649314, "grad_norm": 0.14354, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:52.086439+00:00", "epoch": 0, "step": 8969, "train_loss": 3.5590598583221436, "perplexity": 35.130154292858876, "lr": 0.0026291804804649314, "grad_norm": 0.153402, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:52.392654+00:00", "epoch": 0, "step": 8970, "train_loss": 3.5957586765289307, "perplexity": 36.44333820765307, "lr": 0.0026291804804649314, "grad_norm": 0.158018, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:52.697743+00:00", "epoch": 0, "step": 8971, "train_loss": 3.531104803085327, "perplexity": 34.16168871281468, "lr": 0.0026291804804649314, "grad_norm": 0.155944, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:53.003344+00:00", "epoch": 0, "step": 8972, "train_loss": 3.5610365867614746, "perplexity": 35.1996657479283, "lr": 0.0026291804804649314, "grad_norm": 0.183861, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:53.308579+00:00", "epoch": 0, "step": 8973, "train_loss": 3.4444804191589355, "perplexity": 31.327002307144802, "lr": 0.0026291804804649314, "grad_norm": 0.179969, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:53.613000+00:00", "epoch": 0, "step": 8974, "train_loss": 3.6080548763275146, "perplexity": 36.89421915573802, "lr": 0.0026291804804649314, "grad_norm": 0.163828, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:53.918500+00:00", "epoch": 0, "step": 8975, "train_loss": 3.5989649295806885, "perplexity": 36.56037229218319, "lr": 0.0026291804804649314, "grad_norm": 0.146375, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:54.224467+00:00", "epoch": 0, "step": 8976, "train_loss": 3.5237855911254883, "perplexity": 33.91256487943717, "lr": 0.0026291804804649314, "grad_norm": 0.149396, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:54.531404+00:00", "epoch": 0, "step": 8977, "train_loss": 3.490903377532959, "perplexity": 32.81557917697377, "lr": 0.0026291804804649314, "grad_norm": 0.152788, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:54.837356+00:00", "epoch": 0, "step": 8978, "train_loss": 3.5701465606689453, "perplexity": 35.52179886874681, "lr": 0.0026291804804649314, "grad_norm": 0.138511, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:55.143444+00:00", "epoch": 0, "step": 8979, "train_loss": 3.6434988975524902, "perplexity": 38.225349597409405, "lr": 0.0026291804804649314, "grad_norm": 0.161488, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:55.448343+00:00", "epoch": 0, "step": 8980, "train_loss": 3.550438165664673, "perplexity": 34.82857483037621, "lr": 0.0026291804804649314, "grad_norm": 0.160376, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:55.753876+00:00", "epoch": 0, "step": 8981, "train_loss": 3.462109327316284, "perplexity": 31.884159763172633, "lr": 0.0026291804804649314, "grad_norm": 0.151762, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:56.059377+00:00", "epoch": 0, "step": 8982, "train_loss": 3.598836660385132, "perplexity": 36.55568302339081, "lr": 0.0026291804804649314, "grad_norm": 0.201489, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:56.363738+00:00", "epoch": 0, "step": 8983, "train_loss": 3.528688669204712, "perplexity": 34.07924913193366, "lr": 0.0026291804804649314, "grad_norm": 0.180627, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:56.668166+00:00", "epoch": 0, "step": 8984, "train_loss": 3.593585252761841, "perplexity": 36.36421740290493, "lr": 0.0026291804804649314, "grad_norm": 0.144848, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:56.974821+00:00", "epoch": 0, "step": 8985, "train_loss": 3.6355011463165283, "perplexity": 37.92085203138658, "lr": 0.0026291804804649314, "grad_norm": 0.174402, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:57.281339+00:00", "epoch": 0, "step": 8986, "train_loss": 3.599193811416626, "perplexity": 36.568741255031235, "lr": 0.0026291804804649314, "grad_norm": 0.161175, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:57.586928+00:00", "epoch": 0, "step": 8987, "train_loss": 3.6058967113494873, "perplexity": 36.81468120294834, "lr": 0.0026291804804649314, "grad_norm": 0.159609, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:57.893115+00:00", "epoch": 0, "step": 8988, "train_loss": 3.7396764755249023, "perplexity": 42.08437263774656, "lr": 0.0026291804804649314, "grad_norm": 0.178239, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:58.198453+00:00", "epoch": 0, "step": 8989, "train_loss": 3.675278425216675, "perplexity": 39.45964188721724, "lr": 0.0026291804804649314, "grad_norm": 0.190982, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:58.503654+00:00", "epoch": 0, "step": 8990, "train_loss": 3.4696483612060547, "perplexity": 32.12544390601724, "lr": 0.0026291804804649314, "grad_norm": 0.14352, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:58.808909+00:00", "epoch": 0, "step": 8991, "train_loss": 3.586165189743042, "perplexity": 36.09539120521343, "lr": 0.0026291804804649314, "grad_norm": 0.165213, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:59.114604+00:00", "epoch": 0, "step": 8992, "train_loss": 3.5446338653564453, "perplexity": 34.62700487493606, "lr": 0.0026291804804649314, "grad_norm": 0.196945, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:59.419995+00:00", "epoch": 0, "step": 8993, "train_loss": 3.505136013031006, "perplexity": 33.2859708703181, "lr": 0.0026291804804649314, "grad_norm": 0.162503, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:44:59.724292+00:00", "epoch": 0, "step": 8994, "train_loss": 3.454329490661621, "perplexity": 31.637068618996597, "lr": 0.0026291804804649314, "grad_norm": 0.140976, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:00.029056+00:00", "epoch": 0, "step": 8995, "train_loss": 3.5656657218933105, "perplexity": 35.36298748481571, "lr": 0.0026291804804649314, "grad_norm": 0.159606, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:00.334847+00:00", "epoch": 0, "step": 8996, "train_loss": 3.4865453243255615, "perplexity": 32.672878311725995, "lr": 0.0026291804804649314, "grad_norm": 0.14558, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:00.641165+00:00", "epoch": 0, "step": 8997, "train_loss": 3.517210006713867, "perplexity": 33.690301503111336, "lr": 0.0026291804804649314, "grad_norm": 0.158692, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:00.947287+00:00", "epoch": 0, "step": 8998, "train_loss": 3.551532506942749, "perplexity": 34.866710040130904, "lr": 0.0026291804804649314, "grad_norm": 0.159101, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:01.255276+00:00", "epoch": 0, "step": 8999, "train_loss": 3.5037453174591064, "perplexity": 33.23971239120887, "lr": 0.0026291804804649314, "grad_norm": 0.153482, "tokens_per_sec": 106395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:01.560679+00:00", "epoch": 0, "step": 9000, "train_loss": 3.5705881118774414, "perplexity": 35.537487025272604, "lr": 0.0026291804804649314, "grad_norm": 0.159679, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:45:04.633529+00:00", "step": 9000, "epoch": 0, "val_loss": 3.5061901450157165, "val_ppl": 33.321077176944165, "eval_train_loss": 3.5705881118774414, "eval_train_ppl": 35.537487025272604} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:45:05.650966+00:00", "step": 9000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5062_epoch_0000_step_0009000.pt", "val_loss": 3.5061901450157165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:06.770469+00:00", "epoch": 0, "step": 9001, "train_loss": 3.538938045501709, "perplexity": 34.43033631909106, "lr": 0.0026291804804649314, "grad_norm": 0.150378, "tokens_per_sec": 6290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:07.074125+00:00", "epoch": 0, "step": 9002, "train_loss": 3.523102045059204, "perplexity": 33.88939199988382, "lr": 0.0026291804804649314, "grad_norm": 0.180521, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:07.378251+00:00", "epoch": 0, "step": 9003, "train_loss": 3.5426366329193115, "perplexity": 34.55791571412159, "lr": 0.0026291804804649314, "grad_norm": 0.162483, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:07.682736+00:00", "epoch": 0, "step": 9004, "train_loss": 3.5928735733032227, "perplexity": 36.338346943179864, "lr": 0.0026291804804649314, "grad_norm": 0.17949, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:07.986481+00:00", "epoch": 0, "step": 9005, "train_loss": 3.609951972961426, "perplexity": 36.96427748741023, "lr": 0.0026291804804649314, "grad_norm": 0.189528, "tokens_per_sec": 107877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:08.290544+00:00", "epoch": 0, "step": 9006, "train_loss": 3.5493834018707275, "perplexity": 34.791858277699625, "lr": 0.0026291804804649314, "grad_norm": 0.191232, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:08.594522+00:00", "epoch": 0, "step": 9007, "train_loss": 3.480919361114502, "perplexity": 32.489578004214756, "lr": 0.0026291804804649314, "grad_norm": 0.193162, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:08.898498+00:00", "epoch": 0, "step": 9008, "train_loss": 3.5255932807922363, "perplexity": 33.97392371470015, "lr": 0.0026291804804649314, "grad_norm": 0.182434, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:09.204242+00:00", "epoch": 0, "step": 9009, "train_loss": 3.5107977390289307, "perplexity": 33.47496141846771, "lr": 0.0026291804804649314, "grad_norm": 0.221742, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:09.509366+00:00", "epoch": 0, "step": 9010, "train_loss": 3.539240598678589, "perplexity": 34.44075490273376, "lr": 0.0026291804804649314, "grad_norm": 0.162445, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:09.814953+00:00", "epoch": 0, "step": 9011, "train_loss": 3.5202560424804688, "perplexity": 33.79307982022477, "lr": 0.0026291804804649314, "grad_norm": 0.173365, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:10.119994+00:00", "epoch": 0, "step": 9012, "train_loss": 3.455561876296997, "perplexity": 31.676081722564593, "lr": 0.0026291804804649314, "grad_norm": 0.14201, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:10.425765+00:00", "epoch": 0, "step": 9013, "train_loss": 3.5671186447143555, "perplexity": 35.41440451979943, "lr": 0.0026291804804649314, "grad_norm": 0.172535, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:10.730101+00:00", "epoch": 0, "step": 9014, "train_loss": 3.600684642791748, "perplexity": 36.62329974048887, "lr": 0.0026291804804649314, "grad_norm": 0.151635, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:11.035929+00:00", "epoch": 0, "step": 9015, "train_loss": 3.554567337036133, "perplexity": 34.97268530814653, "lr": 0.0026291804804649314, "grad_norm": 0.15904, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:11.340804+00:00", "epoch": 0, "step": 9016, "train_loss": 3.6015546321868896, "perplexity": 36.65517548664823, "lr": 0.0026291804804649314, "grad_norm": 0.171102, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:11.644911+00:00", "epoch": 0, "step": 9017, "train_loss": 3.5550856590270996, "perplexity": 34.990817118677036, "lr": 0.0026291804804649314, "grad_norm": 0.153694, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:11.950571+00:00", "epoch": 0, "step": 9018, "train_loss": 3.5636343955993652, "perplexity": 35.29122662803845, "lr": 0.0026291804804649314, "grad_norm": 0.166443, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:12.255949+00:00", "epoch": 0, "step": 9019, "train_loss": 3.5946085453033447, "perplexity": 36.40144768083507, "lr": 0.0026291804804649314, "grad_norm": 0.166591, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:12.561432+00:00", "epoch": 0, "step": 9020, "train_loss": 3.4659550189971924, "perplexity": 32.00701248656554, "lr": 0.0026291804804649314, "grad_norm": 0.155247, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:12.866080+00:00", "epoch": 0, "step": 9021, "train_loss": 3.5082337856292725, "perplexity": 33.389243113163865, "lr": 0.0026291804804649314, "grad_norm": 0.162614, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:13.172218+00:00", "epoch": 0, "step": 9022, "train_loss": 3.5051093101501465, "perplexity": 33.28508205087074, "lr": 0.0026291804804649314, "grad_norm": 0.185641, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:13.477193+00:00", "epoch": 0, "step": 9023, "train_loss": 3.4759840965270996, "perplexity": 32.3296283619889, "lr": 0.0026291804804649314, "grad_norm": 0.169904, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:13.781502+00:00", "epoch": 0, "step": 9024, "train_loss": 3.5648162364959717, "perplexity": 35.332959899144576, "lr": 0.0026291804804649314, "grad_norm": 0.17229, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:14.086213+00:00", "epoch": 0, "step": 9025, "train_loss": 3.542501211166382, "perplexity": 34.55323613746338, "lr": 0.0026291804804649314, "grad_norm": 0.187855, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:14.392422+00:00", "epoch": 0, "step": 9026, "train_loss": 3.5629961490631104, "perplexity": 35.26870931144697, "lr": 0.0026291804804649314, "grad_norm": 0.161201, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:14.696735+00:00", "epoch": 0, "step": 9027, "train_loss": 3.5018301010131836, "perplexity": 33.17611207100987, "lr": 0.0026291804804649314, "grad_norm": 0.159835, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:15.001002+00:00", "epoch": 0, "step": 9028, "train_loss": 3.460364818572998, "perplexity": 31.82858605619631, "lr": 0.0026291804804649314, "grad_norm": 0.154783, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:15.305009+00:00", "epoch": 0, "step": 9029, "train_loss": 3.5593693256378174, "perplexity": 35.14102760978842, "lr": 0.0026291804804649314, "grad_norm": 0.155921, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:15.610203+00:00", "epoch": 0, "step": 9030, "train_loss": 3.424567699432373, "perplexity": 30.709366305062982, "lr": 0.0026291804804649314, "grad_norm": 0.169924, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:15.915091+00:00", "epoch": 0, "step": 9031, "train_loss": 3.583512783050537, "perplexity": 35.999778406081376, "lr": 0.0026291804804649314, "grad_norm": 0.156907, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:16.219729+00:00", "epoch": 0, "step": 9032, "train_loss": 3.5954558849334717, "perplexity": 36.43230514157746, "lr": 0.0026291804804649314, "grad_norm": 0.15888, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:16.524390+00:00", "epoch": 0, "step": 9033, "train_loss": 3.54725980758667, "perplexity": 34.718052880425645, "lr": 0.0026291804804649314, "grad_norm": 0.153711, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:16.828546+00:00", "epoch": 0, "step": 9034, "train_loss": 3.5455567836761475, "perplexity": 34.65897752389382, "lr": 0.0026291804804649314, "grad_norm": 0.147983, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:17.133385+00:00", "epoch": 0, "step": 9035, "train_loss": 3.4809820652008057, "perplexity": 32.49161529739054, "lr": 0.0026291804804649314, "grad_norm": 0.159389, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:17.439301+00:00", "epoch": 0, "step": 9036, "train_loss": 3.32889723777771, "perplexity": 27.90754933690242, "lr": 0.0026291804804649314, "grad_norm": 0.161537, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:17.745733+00:00", "epoch": 0, "step": 9037, "train_loss": 3.521045446395874, "perplexity": 33.819766741742775, "lr": 0.0026291804804649314, "grad_norm": 0.170167, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:18.051051+00:00", "epoch": 0, "step": 9038, "train_loss": 3.5755841732025146, "perplexity": 35.715478748339805, "lr": 0.0026291804804649314, "grad_norm": 0.149034, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:18.356415+00:00", "epoch": 0, "step": 9039, "train_loss": 3.5522100925445557, "perplexity": 34.89034322668572, "lr": 0.0026291804804649314, "grad_norm": 0.207777, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:18.661536+00:00", "epoch": 0, "step": 9040, "train_loss": 3.489635944366455, "perplexity": 32.77401396967214, "lr": 0.0026291804804649314, "grad_norm": 0.192342, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:18.967161+00:00", "epoch": 0, "step": 9041, "train_loss": 3.5851542949676514, "perplexity": 36.05892099970428, "lr": 0.0026291804804649314, "grad_norm": 0.167983, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:19.272717+00:00", "epoch": 0, "step": 9042, "train_loss": 3.561692237854004, "perplexity": 35.22275201467358, "lr": 0.0026291804804649314, "grad_norm": 0.155257, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:19.578165+00:00", "epoch": 0, "step": 9043, "train_loss": 3.5813612937927246, "perplexity": 35.922408529630914, "lr": 0.0026291804804649314, "grad_norm": 0.150589, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:19.883295+00:00", "epoch": 0, "step": 9044, "train_loss": 3.5771381855010986, "perplexity": 35.77102418953561, "lr": 0.0026291804804649314, "grad_norm": 0.186171, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:20.188383+00:00", "epoch": 0, "step": 9045, "train_loss": 3.6553869247436523, "perplexity": 38.682485430701824, "lr": 0.0026291804804649314, "grad_norm": 0.162786, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:20.494690+00:00", "epoch": 0, "step": 9046, "train_loss": 3.543241500854492, "perplexity": 34.57882501228825, "lr": 0.0026291804804649314, "grad_norm": 0.201495, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:20.801919+00:00", "epoch": 0, "step": 9047, "train_loss": 3.5418641567230225, "perplexity": 34.531230854875304, "lr": 0.0026291804804649314, "grad_norm": 0.157378, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:21.107503+00:00", "epoch": 0, "step": 9048, "train_loss": 3.6184604167938232, "perplexity": 37.280127757900324, "lr": 0.0026291804804649314, "grad_norm": 0.160132, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:21.413873+00:00", "epoch": 0, "step": 9049, "train_loss": 3.588615894317627, "perplexity": 36.18395882777263, "lr": 0.0026291804804649314, "grad_norm": 0.17737, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:21.720267+00:00", "epoch": 0, "step": 9050, "train_loss": 3.515963554382324, "perplexity": 33.648334308739535, "lr": 0.0026291804804649314, "grad_norm": 0.154406, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:22.027131+00:00", "epoch": 0, "step": 9051, "train_loss": 3.512259006500244, "perplexity": 33.52391304769317, "lr": 0.0026291804804649314, "grad_norm": 0.158629, "tokens_per_sec": 106689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:22.335070+00:00", "epoch": 0, "step": 9052, "train_loss": 3.5104265213012695, "perplexity": 33.46253722553941, "lr": 0.0026291804804649314, "grad_norm": 0.15696, "tokens_per_sec": 106412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:22.641769+00:00", "epoch": 0, "step": 9053, "train_loss": 3.601229190826416, "perplexity": 36.64324831737182, "lr": 0.0026291804804649314, "grad_norm": 0.165973, "tokens_per_sec": 106840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:22.947805+00:00", "epoch": 0, "step": 9054, "train_loss": 3.6649646759033203, "perplexity": 39.05475656473627, "lr": 0.0026291804804649314, "grad_norm": 0.153977, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:23.253710+00:00", "epoch": 0, "step": 9055, "train_loss": 3.491483211517334, "perplexity": 32.834612282484976, "lr": 0.0026291804804649314, "grad_norm": 0.172822, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:23.559641+00:00", "epoch": 0, "step": 9056, "train_loss": 3.540832281112671, "perplexity": 34.495617297501376, "lr": 0.0026291804804649314, "grad_norm": 0.169969, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:23.866404+00:00", "epoch": 0, "step": 9057, "train_loss": 3.47756290435791, "perplexity": 32.38071094659546, "lr": 0.0026291804804649314, "grad_norm": 0.173041, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:24.173713+00:00", "epoch": 0, "step": 9058, "train_loss": 3.4948558807373047, "perplexity": 32.94553952397749, "lr": 0.0026291804804649314, "grad_norm": 0.152579, "tokens_per_sec": 106630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:24.479131+00:00", "epoch": 0, "step": 9059, "train_loss": 3.5046443939208984, "perplexity": 33.26961087271301, "lr": 0.0026291804804649314, "grad_norm": 0.161166, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:24.785571+00:00", "epoch": 0, "step": 9060, "train_loss": 3.5055103302001953, "perplexity": 33.29843271290332, "lr": 0.0026291804804649314, "grad_norm": 0.152092, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:25.092359+00:00", "epoch": 0, "step": 9061, "train_loss": 3.5703258514404297, "perplexity": 35.52816817043195, "lr": 0.0026291804804649314, "grad_norm": 0.155034, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:25.400222+00:00", "epoch": 0, "step": 9062, "train_loss": 3.557389497756958, "perplexity": 35.071523249602706, "lr": 0.0026291804804649314, "grad_norm": 0.158374, "tokens_per_sec": 106437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:25.705559+00:00", "epoch": 0, "step": 9063, "train_loss": 3.596754312515259, "perplexity": 36.47964057560765, "lr": 0.0026291804804649314, "grad_norm": 0.158218, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:26.011075+00:00", "epoch": 0, "step": 9064, "train_loss": 3.592860698699951, "perplexity": 36.33787910439106, "lr": 0.0026291804804649314, "grad_norm": 0.147113, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:26.317182+00:00", "epoch": 0, "step": 9065, "train_loss": 3.551161766052246, "perplexity": 34.85378592089949, "lr": 0.0026291804804649314, "grad_norm": 0.188811, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:26.623651+00:00", "epoch": 0, "step": 9066, "train_loss": 3.55580472946167, "perplexity": 35.01598702913322, "lr": 0.0026291804804649314, "grad_norm": 0.158163, "tokens_per_sec": 106922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:26.929496+00:00", "epoch": 0, "step": 9067, "train_loss": 3.5097148418426514, "perplexity": 33.43873109733375, "lr": 0.0026291804804649314, "grad_norm": 0.163394, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:27.235631+00:00", "epoch": 0, "step": 9068, "train_loss": 3.598741292953491, "perplexity": 36.55219696801972, "lr": 0.0026291804804649314, "grad_norm": 0.144253, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:27.541192+00:00", "epoch": 0, "step": 9069, "train_loss": 3.5152089595794678, "perplexity": 33.62295302804101, "lr": 0.0026291804804649314, "grad_norm": 0.166875, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:27.848525+00:00", "epoch": 0, "step": 9070, "train_loss": 3.5145747661590576, "perplexity": 33.601636332624324, "lr": 0.0026291804804649314, "grad_norm": 0.149044, "tokens_per_sec": 106683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:28.155728+00:00", "epoch": 0, "step": 9071, "train_loss": 3.540034294128418, "perplexity": 34.46810122407857, "lr": 0.0026291804804649314, "grad_norm": 0.155624, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:28.461467+00:00", "epoch": 0, "step": 9072, "train_loss": 3.5124917030334473, "perplexity": 33.53171485373017, "lr": 0.0026291804804649314, "grad_norm": 0.137768, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:28.767440+00:00", "epoch": 0, "step": 9073, "train_loss": 3.6277124881744385, "perplexity": 37.62664669812432, "lr": 0.0026291804804649314, "grad_norm": 0.148814, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:29.072937+00:00", "epoch": 0, "step": 9074, "train_loss": 3.557854175567627, "perplexity": 35.08782399524724, "lr": 0.0026291804804649314, "grad_norm": 0.147408, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:29.378580+00:00", "epoch": 0, "step": 9075, "train_loss": 3.5159130096435547, "perplexity": 33.64663360545303, "lr": 0.0026291804804649314, "grad_norm": 0.16441, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:29.683920+00:00", "epoch": 0, "step": 9076, "train_loss": 3.599896192550659, "perplexity": 36.59443547149423, "lr": 0.0026291804804649314, "grad_norm": 0.147952, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:29.990228+00:00", "epoch": 0, "step": 9077, "train_loss": 3.480381727218628, "perplexity": 32.47211520053527, "lr": 0.0026291804804649314, "grad_norm": 0.15545, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:30.296180+00:00", "epoch": 0, "step": 9078, "train_loss": 3.5624911785125732, "perplexity": 35.25090414780923, "lr": 0.0026291804804649314, "grad_norm": 0.156061, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:30.602275+00:00", "epoch": 0, "step": 9079, "train_loss": 3.5953800678253174, "perplexity": 36.42954305426629, "lr": 0.0026291804804649314, "grad_norm": 0.170153, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:30.908653+00:00", "epoch": 0, "step": 9080, "train_loss": 3.523446798324585, "perplexity": 33.90107749262771, "lr": 0.0026291804804649314, "grad_norm": 0.15152, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:31.214780+00:00", "epoch": 0, "step": 9081, "train_loss": 3.5132367610931396, "perplexity": 33.55670723736955, "lr": 0.0026291804804649314, "grad_norm": 0.147647, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:31.520035+00:00", "epoch": 0, "step": 9082, "train_loss": 3.616427421569824, "perplexity": 37.20441442471819, "lr": 0.0026291804804649314, "grad_norm": 0.15853, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:31.824428+00:00", "epoch": 0, "step": 9083, "train_loss": 3.6053853034973145, "perplexity": 36.79585869930423, "lr": 0.0026291804804649314, "grad_norm": 0.166146, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:32.129779+00:00", "epoch": 0, "step": 9084, "train_loss": 3.5466694831848145, "perplexity": 34.6975640147591, "lr": 0.0026291804804649314, "grad_norm": 0.157944, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:32.435057+00:00", "epoch": 0, "step": 9085, "train_loss": 3.5255842208862305, "perplexity": 33.973615915538964, "lr": 0.0026291804804649314, "grad_norm": 0.175441, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:32.741022+00:00", "epoch": 0, "step": 9086, "train_loss": 3.5345346927642822, "perplexity": 34.27906070777844, "lr": 0.0026291804804649314, "grad_norm": 0.16678, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:33.047411+00:00", "epoch": 0, "step": 9087, "train_loss": 3.5715205669403076, "perplexity": 35.570639589206806, "lr": 0.0026291804804649314, "grad_norm": 0.164088, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:33.352850+00:00", "epoch": 0, "step": 9088, "train_loss": 3.6120522022247314, "perplexity": 37.041992525818664, "lr": 0.0026291804804649314, "grad_norm": 0.17552, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:33.658526+00:00", "epoch": 0, "step": 9089, "train_loss": 3.613619804382324, "perplexity": 37.100105170076546, "lr": 0.0026291804804649314, "grad_norm": 0.189285, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:33.965090+00:00", "epoch": 0, "step": 9090, "train_loss": 3.551197052001953, "perplexity": 34.855015791535045, "lr": 0.0026291804804649314, "grad_norm": 0.165919, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:34.270643+00:00", "epoch": 0, "step": 9091, "train_loss": 3.4469194412231445, "perplexity": 31.403502812263376, "lr": 0.0026291804804649314, "grad_norm": 0.162329, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:34.577624+00:00", "epoch": 0, "step": 9092, "train_loss": 3.6411144733428955, "perplexity": 38.134312726800594, "lr": 0.0026291804804649314, "grad_norm": 0.149218, "tokens_per_sec": 106739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:34.882591+00:00", "epoch": 0, "step": 9093, "train_loss": 3.539135456085205, "perplexity": 34.43713390280919, "lr": 0.0026291804804649314, "grad_norm": 0.171366, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:35.187069+00:00", "epoch": 0, "step": 9094, "train_loss": 3.4425790309906006, "perplexity": 31.26749410763005, "lr": 0.0026291804804649314, "grad_norm": 0.160127, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:35.493223+00:00", "epoch": 0, "step": 9095, "train_loss": 3.54630970954895, "perplexity": 34.6850829913041, "lr": 0.0026291804804649314, "grad_norm": 0.183361, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:35.798890+00:00", "epoch": 0, "step": 9096, "train_loss": 3.4541378021240234, "perplexity": 31.631004736786068, "lr": 0.0026291804804649314, "grad_norm": 0.15461, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:36.105719+00:00", "epoch": 0, "step": 9097, "train_loss": 3.511704921722412, "perplexity": 33.50534310291704, "lr": 0.0026291804804649314, "grad_norm": 0.176856, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:36.412345+00:00", "epoch": 0, "step": 9098, "train_loss": 3.618560314178467, "perplexity": 37.283852131187, "lr": 0.0026291804804649314, "grad_norm": 0.156491, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:36.718589+00:00", "epoch": 0, "step": 9099, "train_loss": 3.6059958934783936, "perplexity": 36.818332742485794, "lr": 0.0026291804804649314, "grad_norm": 0.169054, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:37.024990+00:00", "epoch": 0, "step": 9100, "train_loss": 3.5034852027893066, "perplexity": 33.231067378792005, "lr": 0.0026291804804649314, "grad_norm": 0.169536, "tokens_per_sec": 107025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:37.330824+00:00", "epoch": 0, "step": 9101, "train_loss": 3.5676591396331787, "perplexity": 35.43355099931687, "lr": 0.0026291804804649314, "grad_norm": 0.189758, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:37.636270+00:00", "epoch": 0, "step": 9102, "train_loss": 3.495522975921631, "perplexity": 32.96752466701517, "lr": 0.0026291804804649314, "grad_norm": 0.193826, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:37.942042+00:00", "epoch": 0, "step": 9103, "train_loss": 3.5162599086761475, "perplexity": 33.65830761483498, "lr": 0.0026291804804649314, "grad_norm": 0.159361, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:38.246990+00:00", "epoch": 0, "step": 9104, "train_loss": 3.5572285652160645, "perplexity": 35.065879554392424, "lr": 0.0026291804804649314, "grad_norm": 0.153872, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:38.552249+00:00", "epoch": 0, "step": 9105, "train_loss": 3.5744054317474365, "perplexity": 35.67340423530917, "lr": 0.0026291804804649314, "grad_norm": 0.16894, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:38.857865+00:00", "epoch": 0, "step": 9106, "train_loss": 3.5298941135406494, "perplexity": 34.12035453993587, "lr": 0.0026291804804649314, "grad_norm": 0.159209, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:39.163982+00:00", "epoch": 0, "step": 9107, "train_loss": 3.6311049461364746, "perplexity": 37.75451027858151, "lr": 0.0026291804804649314, "grad_norm": 0.157954, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:39.469458+00:00", "epoch": 0, "step": 9108, "train_loss": 3.477321147918701, "perplexity": 32.37288364740606, "lr": 0.0026291804804649314, "grad_norm": 0.141199, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:39.776506+00:00", "epoch": 0, "step": 9109, "train_loss": 3.572528600692749, "perplexity": 35.60651407281605, "lr": 0.0026291804804649314, "grad_norm": 0.14712, "tokens_per_sec": 106719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:40.082162+00:00", "epoch": 0, "step": 9110, "train_loss": 3.5600059032440186, "perplexity": 35.16340472265252, "lr": 0.0026291804804649314, "grad_norm": 0.151813, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:40.387034+00:00", "epoch": 0, "step": 9111, "train_loss": 3.6569466590881348, "perplexity": 38.74286690904962, "lr": 0.0026291804804649314, "grad_norm": 0.157952, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:40.691905+00:00", "epoch": 0, "step": 9112, "train_loss": 3.5528225898742676, "perplexity": 34.91172001468929, "lr": 0.0026291804804649314, "grad_norm": 0.184839, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:40.997396+00:00", "epoch": 0, "step": 9113, "train_loss": 3.600611925125122, "perplexity": 36.62063667641467, "lr": 0.0026291804804649314, "grad_norm": 0.151515, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:41.303714+00:00", "epoch": 0, "step": 9114, "train_loss": 3.595839500427246, "perplexity": 36.446283819350946, "lr": 0.0026291804804649314, "grad_norm": 0.175376, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:41.608636+00:00", "epoch": 0, "step": 9115, "train_loss": 3.5973076820373535, "perplexity": 36.49983288326771, "lr": 0.0026291804804649314, "grad_norm": 0.178747, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:41.914476+00:00", "epoch": 0, "step": 9116, "train_loss": 3.567260265350342, "perplexity": 35.41942028545019, "lr": 0.0026291804804649314, "grad_norm": 0.1716, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:42.220612+00:00", "epoch": 0, "step": 9117, "train_loss": 3.4867208003997803, "perplexity": 32.67861212320422, "lr": 0.0026291804804649314, "grad_norm": 0.191843, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:42.526290+00:00", "epoch": 0, "step": 9118, "train_loss": 3.4937379360198975, "perplexity": 32.9087288121088, "lr": 0.0026291804804649314, "grad_norm": 0.191624, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:42.832463+00:00", "epoch": 0, "step": 9119, "train_loss": 3.537895441055298, "perplexity": 34.394457804153014, "lr": 0.0026291804804649314, "grad_norm": 0.185433, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:43.139967+00:00", "epoch": 0, "step": 9120, "train_loss": 3.488797664642334, "perplexity": 32.74655169042741, "lr": 0.0026291804804649314, "grad_norm": 0.16642, "tokens_per_sec": 106629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:43.446051+00:00", "epoch": 0, "step": 9121, "train_loss": 3.6570963859558105, "perplexity": 38.748668191449774, "lr": 0.0026291804804649314, "grad_norm": 0.172902, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:43.750793+00:00", "epoch": 0, "step": 9122, "train_loss": 3.6184444427490234, "perplexity": 37.27953224822574, "lr": 0.0026291804804649314, "grad_norm": 0.163718, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:44.055900+00:00", "epoch": 0, "step": 9123, "train_loss": 3.582839250564575, "perplexity": 35.97553954958241, "lr": 0.0026291804804649314, "grad_norm": 0.17911, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:44.361089+00:00", "epoch": 0, "step": 9124, "train_loss": 3.553924322128296, "perplexity": 34.950204578623, "lr": 0.0026291804804649314, "grad_norm": 0.172989, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:44.666486+00:00", "epoch": 0, "step": 9125, "train_loss": 3.576448440551758, "perplexity": 35.74635981331043, "lr": 0.0026291804804649314, "grad_norm": 0.157196, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:44.972974+00:00", "epoch": 0, "step": 9126, "train_loss": 3.4703433513641357, "perplexity": 32.1477785336301, "lr": 0.0026291804804649314, "grad_norm": 0.192474, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:45.278683+00:00", "epoch": 0, "step": 9127, "train_loss": 3.453610897064209, "perplexity": 31.614342590413347, "lr": 0.0026291804804649314, "grad_norm": 0.165968, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:45.583660+00:00", "epoch": 0, "step": 9128, "train_loss": 3.5493340492248535, "perplexity": 34.79014124980903, "lr": 0.0026291804804649314, "grad_norm": 0.152553, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:45.889226+00:00", "epoch": 0, "step": 9129, "train_loss": 3.631093978881836, "perplexity": 37.75409621752409, "lr": 0.0026291804804649314, "grad_norm": 0.158432, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:46.194887+00:00", "epoch": 0, "step": 9130, "train_loss": 3.6029810905456543, "perplexity": 36.70749987852383, "lr": 0.0026291804804649314, "grad_norm": 0.17823, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:46.500649+00:00", "epoch": 0, "step": 9131, "train_loss": 3.5618395805358887, "perplexity": 35.22794221177818, "lr": 0.0026291804804649314, "grad_norm": 0.151683, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:46.806790+00:00", "epoch": 0, "step": 9132, "train_loss": 3.4873788356781006, "perplexity": 32.70012287946987, "lr": 0.0026291804804649314, "grad_norm": 0.185585, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:47.112043+00:00", "epoch": 0, "step": 9133, "train_loss": 3.527279853820801, "perplexity": 34.03127156518684, "lr": 0.0026291804804649314, "grad_norm": 0.173813, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:47.417465+00:00", "epoch": 0, "step": 9134, "train_loss": 3.4834096431732178, "perplexity": 32.5705870432273, "lr": 0.0026291804804649314, "grad_norm": 0.214753, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:47.722628+00:00", "epoch": 0, "step": 9135, "train_loss": 3.5822577476501465, "perplexity": 35.95462574979327, "lr": 0.0026291804804649314, "grad_norm": 0.214626, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:48.028825+00:00", "epoch": 0, "step": 9136, "train_loss": 3.4100868701934814, "perplexity": 30.267873521225248, "lr": 0.0026291804804649314, "grad_norm": 0.181549, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:48.335829+00:00", "epoch": 0, "step": 9137, "train_loss": 3.5154693126678467, "perplexity": 33.63170800734577, "lr": 0.0026291804804649314, "grad_norm": 0.171685, "tokens_per_sec": 106733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:48.641545+00:00", "epoch": 0, "step": 9138, "train_loss": 3.50162935256958, "perplexity": 33.16945268659969, "lr": 0.0026291804804649314, "grad_norm": 0.190235, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:48.946695+00:00", "epoch": 0, "step": 9139, "train_loss": 3.5886080265045166, "perplexity": 36.18367414026691, "lr": 0.0026291804804649314, "grad_norm": 0.140684, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:49.252248+00:00", "epoch": 0, "step": 9140, "train_loss": 3.5702314376831055, "perplexity": 35.52481398092743, "lr": 0.0026291804804649314, "grad_norm": 0.168111, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:49.557810+00:00", "epoch": 0, "step": 9141, "train_loss": 3.4226882457733154, "perplexity": 30.651703678284733, "lr": 0.0026291804804649314, "grad_norm": 0.170234, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:49.864868+00:00", "epoch": 0, "step": 9142, "train_loss": 3.4878334999084473, "perplexity": 32.71499383606095, "lr": 0.0026291804804649314, "grad_norm": 0.165972, "tokens_per_sec": 106724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:50.170451+00:00", "epoch": 0, "step": 9143, "train_loss": 3.5587081909179688, "perplexity": 35.11780233470827, "lr": 0.0026291804804649314, "grad_norm": 0.139562, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:50.475294+00:00", "epoch": 0, "step": 9144, "train_loss": 3.5442519187927246, "perplexity": 34.61378173483936, "lr": 0.0026291804804649314, "grad_norm": 0.144357, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:50.780121+00:00", "epoch": 0, "step": 9145, "train_loss": 3.560992956161499, "perplexity": 35.19812999889585, "lr": 0.0026291804804649314, "grad_norm": 0.135603, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:51.086109+00:00", "epoch": 0, "step": 9146, "train_loss": 3.513828754425049, "perplexity": 33.57657846553186, "lr": 0.0026291804804649314, "grad_norm": 0.139527, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:51.393057+00:00", "epoch": 0, "step": 9147, "train_loss": 3.533007860183716, "perplexity": 34.226762256708014, "lr": 0.0026291804804649314, "grad_norm": 0.148605, "tokens_per_sec": 106754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:51.699391+00:00", "epoch": 0, "step": 9148, "train_loss": 3.5406153202056885, "perplexity": 34.48813390891633, "lr": 0.0026291804804649314, "grad_norm": 0.150468, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:52.004550+00:00", "epoch": 0, "step": 9149, "train_loss": 3.5422518253326416, "perplexity": 34.544620124261286, "lr": 0.0026291804804649314, "grad_norm": 0.160825, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:52.309637+00:00", "epoch": 0, "step": 9150, "train_loss": 3.404527425765991, "perplexity": 30.100067845812415, "lr": 0.0026291804804649314, "grad_norm": 0.162939, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:52.614059+00:00", "epoch": 0, "step": 9151, "train_loss": 3.546391725540161, "perplexity": 34.68792783942579, "lr": 0.0026291804804649314, "grad_norm": 0.142623, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:52.919608+00:00", "epoch": 0, "step": 9152, "train_loss": 3.624755620956421, "perplexity": 37.5155540240626, "lr": 0.0026291804804649314, "grad_norm": 0.162263, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:53.226706+00:00", "epoch": 0, "step": 9153, "train_loss": 3.541524648666382, "perplexity": 34.519509213702754, "lr": 0.0026291804804649314, "grad_norm": 0.155266, "tokens_per_sec": 106707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:53.531540+00:00", "epoch": 0, "step": 9154, "train_loss": 3.6537318229675293, "perplexity": 38.61851493380533, "lr": 0.0026291804804649314, "grad_norm": 0.142309, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:53.836169+00:00", "epoch": 0, "step": 9155, "train_loss": 3.6357362270355225, "perplexity": 37.92976754043807, "lr": 0.0026291804804649314, "grad_norm": 0.155852, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:54.141765+00:00", "epoch": 0, "step": 9156, "train_loss": 3.6542060375213623, "perplexity": 38.63683273857589, "lr": 0.0026291804804649314, "grad_norm": 0.200521, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:54.447678+00:00", "epoch": 0, "step": 9157, "train_loss": 3.6756749153137207, "perplexity": 39.475290346483064, "lr": 0.0026291804804649314, "grad_norm": 0.170694, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:54.754682+00:00", "epoch": 0, "step": 9158, "train_loss": 3.4386212825775146, "perplexity": 31.14398979315862, "lr": 0.0026291804804649314, "grad_norm": 0.150301, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:55.060834+00:00", "epoch": 0, "step": 9159, "train_loss": 3.5276525020599365, "perplexity": 34.043955621810056, "lr": 0.0026291804804649314, "grad_norm": 0.186221, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:55.365748+00:00", "epoch": 0, "step": 9160, "train_loss": 3.489344596862793, "perplexity": 32.764466733366355, "lr": 0.0026291804804649314, "grad_norm": 0.159342, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:55.670291+00:00", "epoch": 0, "step": 9161, "train_loss": 3.4571359157562256, "perplexity": 31.72598038603988, "lr": 0.0026291804804649314, "grad_norm": 0.165343, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:55.975514+00:00", "epoch": 0, "step": 9162, "train_loss": 3.552496910095215, "perplexity": 34.900351824723835, "lr": 0.0026291804804649314, "grad_norm": 0.159971, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:56.281258+00:00", "epoch": 0, "step": 9163, "train_loss": 3.6300601959228516, "perplexity": 37.71508684330985, "lr": 0.0026291804804649314, "grad_norm": 0.15315, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:56.587458+00:00", "epoch": 0, "step": 9164, "train_loss": 3.5158538818359375, "perplexity": 33.64464421258904, "lr": 0.0026291804804649314, "grad_norm": 0.143813, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:56.892369+00:00", "epoch": 0, "step": 9165, "train_loss": 3.6470203399658203, "perplexity": 38.360195251001265, "lr": 0.0026291804804649314, "grad_norm": 0.147314, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:57.199665+00:00", "epoch": 0, "step": 9166, "train_loss": 3.469118356704712, "perplexity": 32.108421787437614, "lr": 0.0026291804804649314, "grad_norm": 0.154086, "tokens_per_sec": 106633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:57.504601+00:00", "epoch": 0, "step": 9167, "train_loss": 3.4920647144317627, "perplexity": 32.85371125773808, "lr": 0.0026291804804649314, "grad_norm": 0.148904, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:57.810192+00:00", "epoch": 0, "step": 9168, "train_loss": 3.593367576599121, "perplexity": 36.3563026410591, "lr": 0.0026291804804649314, "grad_norm": 0.182343, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:58.117737+00:00", "epoch": 0, "step": 9169, "train_loss": 3.54410982131958, "perplexity": 34.608863553357715, "lr": 0.0026291804804649314, "grad_norm": 0.148484, "tokens_per_sec": 106547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:58.424329+00:00", "epoch": 0, "step": 9170, "train_loss": 3.586737632751465, "perplexity": 36.11605967474118, "lr": 0.0026291804804649314, "grad_norm": 0.145525, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:58.730556+00:00", "epoch": 0, "step": 9171, "train_loss": 3.614406108856201, "perplexity": 37.129288620788394, "lr": 0.0026291804804649314, "grad_norm": 0.151111, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:59.036552+00:00", "epoch": 0, "step": 9172, "train_loss": 3.6057331562042236, "perplexity": 36.80866046479113, "lr": 0.0026291804804649314, "grad_norm": 0.163105, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:59.342150+00:00", "epoch": 0, "step": 9173, "train_loss": 3.6089401245117188, "perplexity": 36.926894156862126, "lr": 0.0026291804804649314, "grad_norm": 0.179217, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:59.647231+00:00", "epoch": 0, "step": 9174, "train_loss": 3.547581434249878, "perplexity": 34.729220927801286, "lr": 0.0026291804804649314, "grad_norm": 0.152426, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:45:59.953747+00:00", "epoch": 0, "step": 9175, "train_loss": 3.536226272583008, "perplexity": 34.337095546518974, "lr": 0.0026291804804649314, "grad_norm": 0.16038, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:00.259087+00:00", "epoch": 0, "step": 9176, "train_loss": 3.664858818054199, "perplexity": 39.0506225310222, "lr": 0.0026291804804649314, "grad_norm": 0.159779, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:00.564679+00:00", "epoch": 0, "step": 9177, "train_loss": 3.5208613872528076, "perplexity": 33.81354247729155, "lr": 0.0026291804804649314, "grad_norm": 0.141885, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:00.869483+00:00", "epoch": 0, "step": 9178, "train_loss": 3.5174407958984375, "perplexity": 33.69807775762637, "lr": 0.0026291804804649314, "grad_norm": 0.154874, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:01.174985+00:00", "epoch": 0, "step": 9179, "train_loss": 3.459883689880371, "perplexity": 31.813276093525424, "lr": 0.0026291804804649314, "grad_norm": 0.162704, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:01.480976+00:00", "epoch": 0, "step": 9180, "train_loss": 3.6747515201568604, "perplexity": 39.43885587885578, "lr": 0.0026291804804649314, "grad_norm": 0.150995, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:01.786466+00:00", "epoch": 0, "step": 9181, "train_loss": 3.6101160049438477, "perplexity": 36.97034130844198, "lr": 0.0026291804804649314, "grad_norm": 0.175291, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:02.092140+00:00", "epoch": 0, "step": 9182, "train_loss": 3.557497262954712, "perplexity": 35.075302942897274, "lr": 0.0026291804804649314, "grad_norm": 0.181155, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:02.397495+00:00", "epoch": 0, "step": 9183, "train_loss": 3.4834651947021484, "perplexity": 32.57239643939261, "lr": 0.0026291804804649314, "grad_norm": 0.170068, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:02.703071+00:00", "epoch": 0, "step": 9184, "train_loss": 3.514676094055176, "perplexity": 33.605041288245616, "lr": 0.0026291804804649314, "grad_norm": 0.179545, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:03.008973+00:00", "epoch": 0, "step": 9185, "train_loss": 3.58974289894104, "perplexity": 36.22476130463621, "lr": 0.0026291804804649314, "grad_norm": 0.168655, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:03.314989+00:00", "epoch": 0, "step": 9186, "train_loss": 3.597506046295166, "perplexity": 36.507073863679985, "lr": 0.0026291804804649314, "grad_norm": 0.145083, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:03.620161+00:00", "epoch": 0, "step": 9187, "train_loss": 3.4598288536071777, "perplexity": 31.81153161985704, "lr": 0.0026291804804649314, "grad_norm": 0.179392, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:03.925863+00:00", "epoch": 0, "step": 9188, "train_loss": 3.534754753112793, "perplexity": 34.28660499989375, "lr": 0.0026291804804649314, "grad_norm": 0.17252, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:04.231220+00:00", "epoch": 0, "step": 9189, "train_loss": 3.565312623977661, "perplexity": 35.35050309187996, "lr": 0.0026291804804649314, "grad_norm": 0.168481, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:04.536502+00:00", "epoch": 0, "step": 9190, "train_loss": 3.4990012645721436, "perplexity": 33.08239489399031, "lr": 0.0026291804804649314, "grad_norm": 0.173821, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:04.841792+00:00", "epoch": 0, "step": 9191, "train_loss": 3.6195755004882812, "perplexity": 37.32172140638024, "lr": 0.0026291804804649314, "grad_norm": 0.175297, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:05.147083+00:00", "epoch": 0, "step": 9192, "train_loss": 3.443370819091797, "perplexity": 31.292261141246026, "lr": 0.0026291804804649314, "grad_norm": 0.169834, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:05.452359+00:00", "epoch": 0, "step": 9193, "train_loss": 3.6089844703674316, "perplexity": 36.92853174789225, "lr": 0.0026291804804649314, "grad_norm": 0.17474, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:05.756569+00:00", "epoch": 0, "step": 9194, "train_loss": 3.5762088298797607, "perplexity": 35.737795630089934, "lr": 0.0026291804804649314, "grad_norm": 0.183693, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:06.062730+00:00", "epoch": 0, "step": 9195, "train_loss": 3.545903444290161, "perplexity": 34.67099450911073, "lr": 0.0026291804804649314, "grad_norm": 0.176519, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:06.368795+00:00", "epoch": 0, "step": 9196, "train_loss": 3.443998098373413, "perplexity": 31.311896286050782, "lr": 0.0026291804804649314, "grad_norm": 0.162547, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:06.674688+00:00", "epoch": 0, "step": 9197, "train_loss": 3.4419517517089844, "perplexity": 31.247886806658034, "lr": 0.0026291804804649314, "grad_norm": 0.150477, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:06.980555+00:00", "epoch": 0, "step": 9198, "train_loss": 3.5238993167877197, "perplexity": 33.91642182764889, "lr": 0.0026291804804649314, "grad_norm": 0.15499, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:07.286178+00:00", "epoch": 0, "step": 9199, "train_loss": 3.5016937255859375, "perplexity": 33.17158797304673, "lr": 0.0026291804804649314, "grad_norm": 0.152299, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:07.591790+00:00", "epoch": 0, "step": 9200, "train_loss": 3.5785422325134277, "perplexity": 35.82128366425296, "lr": 0.0026291804804649314, "grad_norm": 0.177714, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:07.897291+00:00", "epoch": 0, "step": 9201, "train_loss": 3.4710705280303955, "perplexity": 32.17116414975853, "lr": 0.0026291804804649314, "grad_norm": 0.206392, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:08.203536+00:00", "epoch": 0, "step": 9202, "train_loss": 3.4869511127471924, "perplexity": 32.68613927783753, "lr": 0.0026291804804649314, "grad_norm": 0.159797, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:08.509140+00:00", "epoch": 0, "step": 9203, "train_loss": 3.4808642864227295, "perplexity": 32.487788699993494, "lr": 0.0026291804804649314, "grad_norm": 0.163196, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:08.814304+00:00", "epoch": 0, "step": 9204, "train_loss": 3.5068554878234863, "perplexity": 33.34325449293171, "lr": 0.0026291804804649314, "grad_norm": 0.170927, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:09.119865+00:00", "epoch": 0, "step": 9205, "train_loss": 3.5699520111083984, "perplexity": 35.51488879058519, "lr": 0.0026291804804649314, "grad_norm": 0.162669, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:09.425368+00:00", "epoch": 0, "step": 9206, "train_loss": 3.4241786003112793, "perplexity": 30.697419641993115, "lr": 0.0026291804804649314, "grad_norm": 0.175034, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:09.731647+00:00", "epoch": 0, "step": 9207, "train_loss": 3.522124767303467, "perplexity": 33.856288829057775, "lr": 0.0026291804804649314, "grad_norm": 0.166588, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:10.038647+00:00", "epoch": 0, "step": 9208, "train_loss": 3.5780394077301025, "perplexity": 35.803276362694206, "lr": 0.0026291804804649314, "grad_norm": 0.143733, "tokens_per_sec": 106742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:10.343983+00:00", "epoch": 0, "step": 9209, "train_loss": 3.5467605590820312, "perplexity": 34.700724270442365, "lr": 0.0026291804804649314, "grad_norm": 0.153359, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:10.649441+00:00", "epoch": 0, "step": 9210, "train_loss": 3.5723204612731934, "perplexity": 35.599103724864044, "lr": 0.0026291804804649314, "grad_norm": 0.159351, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:10.955087+00:00", "epoch": 0, "step": 9211, "train_loss": 3.622551679611206, "perplexity": 37.4329629897891, "lr": 0.0026291804804649314, "grad_norm": 0.174443, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:11.261203+00:00", "epoch": 0, "step": 9212, "train_loss": 3.6298651695251465, "perplexity": 37.70773212298928, "lr": 0.0026291804804649314, "grad_norm": 0.157026, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:11.566752+00:00", "epoch": 0, "step": 9213, "train_loss": 3.52130389213562, "perplexity": 33.82850844595583, "lr": 0.0026291804804649314, "grad_norm": 0.16507, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:11.871492+00:00", "epoch": 0, "step": 9214, "train_loss": 3.5761542320251465, "perplexity": 35.73584447638478, "lr": 0.0026291804804649314, "grad_norm": 0.16213, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:12.177308+00:00", "epoch": 0, "step": 9215, "train_loss": 3.5804977416992188, "perplexity": 35.8914010487558, "lr": 0.0026291804804649314, "grad_norm": 0.166567, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:12.482756+00:00", "epoch": 0, "step": 9216, "train_loss": 3.582393169403076, "perplexity": 35.95949511793999, "lr": 0.0026291804804649314, "grad_norm": 0.162364, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:12.787575+00:00", "epoch": 0, "step": 9217, "train_loss": 3.5795986652374268, "perplexity": 35.85914643674607, "lr": 0.0026291804804649314, "grad_norm": 0.141461, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:13.094425+00:00", "epoch": 0, "step": 9218, "train_loss": 3.5241494178771973, "perplexity": 33.9249054225331, "lr": 0.0026291804804649314, "grad_norm": 0.159772, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:13.399612+00:00", "epoch": 0, "step": 9219, "train_loss": 3.4113378524780273, "perplexity": 30.305761788625706, "lr": 0.0026291804804649314, "grad_norm": 0.151356, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:13.704559+00:00", "epoch": 0, "step": 9220, "train_loss": 3.5914297103881836, "perplexity": 36.285917211415985, "lr": 0.0026291804804649314, "grad_norm": 0.139434, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:14.008863+00:00", "epoch": 0, "step": 9221, "train_loss": 3.55067777633667, "perplexity": 34.836921128487134, "lr": 0.0026291804804649314, "grad_norm": 0.159887, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:14.313597+00:00", "epoch": 0, "step": 9222, "train_loss": 3.518645763397217, "perplexity": 33.738707319830624, "lr": 0.0026291804804649314, "grad_norm": 0.163289, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:14.619652+00:00", "epoch": 0, "step": 9223, "train_loss": 3.594710111618042, "perplexity": 36.40514502948552, "lr": 0.0026291804804649314, "grad_norm": 0.149969, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:14.924588+00:00", "epoch": 0, "step": 9224, "train_loss": 3.4748947620391846, "perplexity": 32.29442975784013, "lr": 0.0026291804804649314, "grad_norm": 0.1597, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:15.230223+00:00", "epoch": 0, "step": 9225, "train_loss": 3.4188826084136963, "perplexity": 30.535276090736264, "lr": 0.0026291804804649314, "grad_norm": 0.166706, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:15.534666+00:00", "epoch": 0, "step": 9226, "train_loss": 3.5492544174194336, "perplexity": 34.78737095835361, "lr": 0.0026291804804649314, "grad_norm": 0.150396, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:15.839675+00:00", "epoch": 0, "step": 9227, "train_loss": 3.6159920692443848, "perplexity": 37.18822092157716, "lr": 0.0026291804804649314, "grad_norm": 0.157903, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:16.146262+00:00", "epoch": 0, "step": 9228, "train_loss": 3.559579372406006, "perplexity": 35.148409644327785, "lr": 0.0026291804804649314, "grad_norm": 0.155215, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:16.452202+00:00", "epoch": 0, "step": 9229, "train_loss": 3.5617198944091797, "perplexity": 35.22372616812892, "lr": 0.0026291804804649314, "grad_norm": 0.15356, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:16.757720+00:00", "epoch": 0, "step": 9230, "train_loss": 3.642549991607666, "perplexity": 38.18909453997362, "lr": 0.0026291804804649314, "grad_norm": 0.167555, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:17.061988+00:00", "epoch": 0, "step": 9231, "train_loss": 3.683913230895996, "perplexity": 39.80184352319743, "lr": 0.0026291804804649314, "grad_norm": 0.14645, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:17.366429+00:00", "epoch": 0, "step": 9232, "train_loss": 3.5264837741851807, "perplexity": 34.00419074359502, "lr": 0.0026291804804649314, "grad_norm": 0.152611, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:17.670708+00:00", "epoch": 0, "step": 9233, "train_loss": 3.5535125732421875, "perplexity": 34.93581683309086, "lr": 0.0026291804804649314, "grad_norm": 0.151377, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:17.976781+00:00", "epoch": 0, "step": 9234, "train_loss": 3.6199190616607666, "perplexity": 37.3345459036193, "lr": 0.0026291804804649314, "grad_norm": 0.156875, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:18.283052+00:00", "epoch": 0, "step": 9235, "train_loss": 3.50626802444458, "perplexity": 33.32367230445604, "lr": 0.0026291804804649314, "grad_norm": 0.188633, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:18.588345+00:00", "epoch": 0, "step": 9236, "train_loss": 3.538672685623169, "perplexity": 34.421201101341076, "lr": 0.0026291804804649314, "grad_norm": 0.168738, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:18.893522+00:00", "epoch": 0, "step": 9237, "train_loss": 3.6525988578796387, "perplexity": 38.57478628083527, "lr": 0.0026291804804649314, "grad_norm": 0.166216, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:19.198894+00:00", "epoch": 0, "step": 9238, "train_loss": 3.6113827228546143, "perplexity": 37.017201975301596, "lr": 0.0026291804804649314, "grad_norm": 0.154043, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:19.503178+00:00", "epoch": 0, "step": 9239, "train_loss": 3.6510655879974365, "perplexity": 38.51568604270437, "lr": 0.0026291804804649314, "grad_norm": 0.180296, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:19.808792+00:00", "epoch": 0, "step": 9240, "train_loss": 3.5870628356933594, "perplexity": 36.1278066335663, "lr": 0.0026291804804649314, "grad_norm": 0.195374, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:20.114283+00:00", "epoch": 0, "step": 9241, "train_loss": 3.5844619274139404, "perplexity": 36.03396361362142, "lr": 0.0026291804804649314, "grad_norm": 0.155983, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:20.419054+00:00", "epoch": 0, "step": 9242, "train_loss": 3.4429304599761963, "perplexity": 31.27848434239683, "lr": 0.0026291804804649314, "grad_norm": 0.154303, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:20.724762+00:00", "epoch": 0, "step": 9243, "train_loss": 3.565073013305664, "perplexity": 35.34203374879169, "lr": 0.0026291804804649314, "grad_norm": 0.151837, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:21.029051+00:00", "epoch": 0, "step": 9244, "train_loss": 3.5409317016601562, "perplexity": 34.49904704114961, "lr": 0.0026291804804649314, "grad_norm": 0.159774, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:21.335069+00:00", "epoch": 0, "step": 9245, "train_loss": 3.583794116973877, "perplexity": 36.009907789782545, "lr": 0.0026291804804649314, "grad_norm": 0.142719, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:21.640585+00:00", "epoch": 0, "step": 9246, "train_loss": 3.550124406814575, "perplexity": 34.81764877095044, "lr": 0.0026291804804649314, "grad_norm": 0.145873, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:21.947269+00:00", "epoch": 0, "step": 9247, "train_loss": 3.500614643096924, "perplexity": 33.13581239920135, "lr": 0.0026291804804649314, "grad_norm": 0.160611, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:22.252524+00:00", "epoch": 0, "step": 9248, "train_loss": 3.5246846675872803, "perplexity": 33.94306857880354, "lr": 0.0026291804804649314, "grad_norm": 0.154758, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:22.557010+00:00", "epoch": 0, "step": 9249, "train_loss": 3.5062484741210938, "perplexity": 33.323020822251095, "lr": 0.0026291804804649314, "grad_norm": 0.150714, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:22.863042+00:00", "epoch": 0, "step": 9250, "train_loss": 3.6873626708984375, "perplexity": 39.9393746607516, "lr": 0.0026291804804649314, "grad_norm": 0.146636, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:23.169644+00:00", "epoch": 0, "step": 9251, "train_loss": 3.6167588233947754, "perplexity": 37.21674607880825, "lr": 0.0026291804804649314, "grad_norm": 0.156982, "tokens_per_sec": 106807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:23.474896+00:00", "epoch": 0, "step": 9252, "train_loss": 3.59173321723938, "perplexity": 36.296931907324876, "lr": 0.0026291804804649314, "grad_norm": 0.221642, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:23.779485+00:00", "epoch": 0, "step": 9253, "train_loss": 3.576793670654297, "perplexity": 35.758702663213256, "lr": 0.0026291804804649314, "grad_norm": 0.201281, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:24.085639+00:00", "epoch": 0, "step": 9254, "train_loss": 3.5235230922698975, "perplexity": 33.90366403824759, "lr": 0.0026291804804649314, "grad_norm": 0.163532, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:24.391124+00:00", "epoch": 0, "step": 9255, "train_loss": 3.6854922771453857, "perplexity": 39.864742121760386, "lr": 0.0026291804804649314, "grad_norm": 0.171234, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:24.698137+00:00", "epoch": 0, "step": 9256, "train_loss": 3.611144542694092, "perplexity": 37.008386262098696, "lr": 0.0026291804804649314, "grad_norm": 0.174059, "tokens_per_sec": 106731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:25.004046+00:00", "epoch": 0, "step": 9257, "train_loss": 3.465456485748291, "perplexity": 31.99105990341994, "lr": 0.0026291804804649314, "grad_norm": 0.171478, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:25.308640+00:00", "epoch": 0, "step": 9258, "train_loss": 3.5049636363983154, "perplexity": 33.28023364123997, "lr": 0.0026291804804649314, "grad_norm": 0.187769, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:25.612723+00:00", "epoch": 0, "step": 9259, "train_loss": 3.5796761512756348, "perplexity": 35.86192512759046, "lr": 0.0026291804804649314, "grad_norm": 0.159816, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:25.919402+00:00", "epoch": 0, "step": 9260, "train_loss": 3.55033540725708, "perplexity": 34.824996085364134, "lr": 0.0026291804804649314, "grad_norm": 0.175376, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:26.225834+00:00", "epoch": 0, "step": 9261, "train_loss": 3.5720057487487793, "perplexity": 35.58790200381733, "lr": 0.0026291804804649314, "grad_norm": 0.157334, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:26.531553+00:00", "epoch": 0, "step": 9262, "train_loss": 3.5438413619995117, "perplexity": 34.59957372840568, "lr": 0.0026291804804649314, "grad_norm": 0.177353, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:26.836868+00:00", "epoch": 0, "step": 9263, "train_loss": 3.5260825157165527, "perplexity": 33.990549011203676, "lr": 0.0026291804804649314, "grad_norm": 0.15175, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:27.142343+00:00", "epoch": 0, "step": 9264, "train_loss": 3.5649490356445312, "perplexity": 35.33765239770828, "lr": 0.0026291804804649314, "grad_norm": 0.184808, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:27.448031+00:00", "epoch": 0, "step": 9265, "train_loss": 3.5344460010528564, "perplexity": 34.276020574037524, "lr": 0.0026291804804649314, "grad_norm": 0.165699, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:27.753436+00:00", "epoch": 0, "step": 9266, "train_loss": 3.578035831451416, "perplexity": 35.803148320429, "lr": 0.0026291804804649314, "grad_norm": 0.170994, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:28.059355+00:00", "epoch": 0, "step": 9267, "train_loss": 3.555116653442383, "perplexity": 34.991901655401115, "lr": 0.0026291804804649314, "grad_norm": 0.182343, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:28.364816+00:00", "epoch": 0, "step": 9268, "train_loss": 3.5136454105377197, "perplexity": 33.57042296941631, "lr": 0.0026291804804649314, "grad_norm": 0.187017, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:28.669364+00:00", "epoch": 0, "step": 9269, "train_loss": 3.5241973400115967, "perplexity": 33.92653121536567, "lr": 0.0026291804804649314, "grad_norm": 0.149088, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:28.975225+00:00", "epoch": 0, "step": 9270, "train_loss": 3.4197304248809814, "perplexity": 30.56117537800886, "lr": 0.0026291804804649314, "grad_norm": 0.182139, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:29.281865+00:00", "epoch": 0, "step": 9271, "train_loss": 3.611043930053711, "perplexity": 37.00466293795051, "lr": 0.0026291804804649314, "grad_norm": 0.193929, "tokens_per_sec": 106805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:29.587626+00:00", "epoch": 0, "step": 9272, "train_loss": 3.501063108444214, "perplexity": 33.15067599547151, "lr": 0.0026291804804649314, "grad_norm": 0.132206, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:29.892564+00:00", "epoch": 0, "step": 9273, "train_loss": 3.536417245864868, "perplexity": 34.343653640535486, "lr": 0.0026291804804649314, "grad_norm": 0.150904, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:30.196750+00:00", "epoch": 0, "step": 9274, "train_loss": 3.538933277130127, "perplexity": 34.43017214284522, "lr": 0.0026291804804649314, "grad_norm": 0.130376, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:30.500916+00:00", "epoch": 0, "step": 9275, "train_loss": 3.4446794986724854, "perplexity": 31.33323949235238, "lr": 0.0026291804804649314, "grad_norm": 0.151503, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:30.807036+00:00", "epoch": 0, "step": 9276, "train_loss": 3.574357748031616, "perplexity": 35.67170323539459, "lr": 0.0026291804804649314, "grad_norm": 0.136625, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:31.113322+00:00", "epoch": 0, "step": 9277, "train_loss": 3.564229726791382, "perplexity": 35.31224285124055, "lr": 0.0026291804804649314, "grad_norm": 0.152139, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:31.418808+00:00", "epoch": 0, "step": 9278, "train_loss": 3.5623605251312256, "perplexity": 35.246298798845515, "lr": 0.0026291804804649314, "grad_norm": 0.147651, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:31.723911+00:00", "epoch": 0, "step": 9279, "train_loss": 3.5917129516601562, "perplexity": 36.29619633642914, "lr": 0.0026291804804649314, "grad_norm": 0.16604, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:32.028753+00:00", "epoch": 0, "step": 9280, "train_loss": 3.501424551010132, "perplexity": 33.162660226540375, "lr": 0.0026291804804649314, "grad_norm": 0.153339, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:32.334672+00:00", "epoch": 0, "step": 9281, "train_loss": 3.5241682529449463, "perplexity": 33.925544406442754, "lr": 0.0026291804804649314, "grad_norm": 0.1586, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:32.640212+00:00", "epoch": 0, "step": 9282, "train_loss": 3.544137716293335, "perplexity": 34.609828980163435, "lr": 0.0026291804804649314, "grad_norm": 0.174812, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:32.945312+00:00", "epoch": 0, "step": 9283, "train_loss": 3.5943689346313477, "perplexity": 36.39272655037429, "lr": 0.0026291804804649314, "grad_norm": 0.174197, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:33.250400+00:00", "epoch": 0, "step": 9284, "train_loss": 3.50685715675354, "perplexity": 33.34331014053766, "lr": 0.0026291804804649314, "grad_norm": 0.151665, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:33.556670+00:00", "epoch": 0, "step": 9285, "train_loss": 3.631784200668335, "perplexity": 37.7801639124736, "lr": 0.0026291804804649314, "grad_norm": 0.160124, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:33.861961+00:00", "epoch": 0, "step": 9286, "train_loss": 3.6023614406585693, "perplexity": 36.684761126129175, "lr": 0.0026291804804649314, "grad_norm": 0.163897, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:34.168675+00:00", "epoch": 0, "step": 9287, "train_loss": 3.5522139072418213, "perplexity": 34.89047632303649, "lr": 0.0026291804804649314, "grad_norm": 0.146102, "tokens_per_sec": 106836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:34.474244+00:00", "epoch": 0, "step": 9288, "train_loss": 3.5219781398773193, "perplexity": 33.85132493249845, "lr": 0.0026291804804649314, "grad_norm": 0.189795, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:34.779637+00:00", "epoch": 0, "step": 9289, "train_loss": 3.5983622074127197, "perplexity": 36.538343184716084, "lr": 0.0026291804804649314, "grad_norm": 0.175357, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:35.085183+00:00", "epoch": 0, "step": 9290, "train_loss": 3.6274845600128174, "perplexity": 37.618071503015805, "lr": 0.0026291804804649314, "grad_norm": 0.162936, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:35.391697+00:00", "epoch": 0, "step": 9291, "train_loss": 3.5251569747924805, "perplexity": 33.95910392116569, "lr": 0.0026291804804649314, "grad_norm": 0.161361, "tokens_per_sec": 106847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:35.696659+00:00", "epoch": 0, "step": 9292, "train_loss": 3.2939486503601074, "perplexity": 26.949066279662873, "lr": 0.0026291804804649314, "grad_norm": 0.172994, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:36.000768+00:00", "epoch": 0, "step": 9293, "train_loss": 3.513385534286499, "perplexity": 33.56169994724537, "lr": 0.0026291804804649314, "grad_norm": 0.161607, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:36.305115+00:00", "epoch": 0, "step": 9294, "train_loss": 3.5276541709899902, "perplexity": 34.04401243883816, "lr": 0.0026291804804649314, "grad_norm": 0.180856, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:36.610470+00:00", "epoch": 0, "step": 9295, "train_loss": 3.4943032264709473, "perplexity": 32.92733706129973, "lr": 0.0026291804804649314, "grad_norm": 0.157472, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:36.915552+00:00", "epoch": 0, "step": 9296, "train_loss": 3.512956142425537, "perplexity": 33.54729192001417, "lr": 0.0026291804804649314, "grad_norm": 0.200924, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:37.220331+00:00", "epoch": 0, "step": 9297, "train_loss": 3.495147705078125, "perplexity": 32.9551552373134, "lr": 0.0026291804804649314, "grad_norm": 0.15882, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:37.524719+00:00", "epoch": 0, "step": 9298, "train_loss": 3.5946879386901855, "perplexity": 36.404337829780204, "lr": 0.0026291804804649314, "grad_norm": 0.172329, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:37.830122+00:00", "epoch": 0, "step": 9299, "train_loss": 3.5308921337127686, "perplexity": 34.15442434039272, "lr": 0.0026291804804649314, "grad_norm": 0.163513, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:38.135764+00:00", "epoch": 0, "step": 9300, "train_loss": 3.5148472785949707, "perplexity": 33.61079444418496, "lr": 0.0026291804804649314, "grad_norm": 0.156953, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:38.441352+00:00", "epoch": 0, "step": 9301, "train_loss": 3.524946928024292, "perplexity": 33.951971670215826, "lr": 0.0026291804804649314, "grad_norm": 0.159255, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:38.746581+00:00", "epoch": 0, "step": 9302, "train_loss": 3.5861809253692627, "perplexity": 36.095959193266545, "lr": 0.0026291804804649314, "grad_norm": 0.159072, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:39.050948+00:00", "epoch": 0, "step": 9303, "train_loss": 3.599048614501953, "perplexity": 36.56343197208262, "lr": 0.0026291804804649314, "grad_norm": 0.14973, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:39.356458+00:00", "epoch": 0, "step": 9304, "train_loss": 3.5652923583984375, "perplexity": 35.34978670071802, "lr": 0.0026291804804649314, "grad_norm": 0.148831, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:39.660940+00:00", "epoch": 0, "step": 9305, "train_loss": 3.5140461921691895, "perplexity": 33.58388007480332, "lr": 0.0026291804804649314, "grad_norm": 0.1483, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:39.966117+00:00", "epoch": 0, "step": 9306, "train_loss": 3.498114824295044, "perplexity": 33.05308232053566, "lr": 0.0026291804804649314, "grad_norm": 0.158652, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:40.271327+00:00", "epoch": 0, "step": 9307, "train_loss": 3.6069698333740234, "perplexity": 36.85420905348367, "lr": 0.0026291804804649314, "grad_norm": 0.147703, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:40.576330+00:00", "epoch": 0, "step": 9308, "train_loss": 3.6319327354431152, "perplexity": 37.785775997396065, "lr": 0.0026291804804649314, "grad_norm": 0.146792, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:40.881119+00:00", "epoch": 0, "step": 9309, "train_loss": 3.520437240600586, "perplexity": 33.79920361755469, "lr": 0.0026291804804649314, "grad_norm": 0.1414, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:41.185439+00:00", "epoch": 0, "step": 9310, "train_loss": 3.4639675617218018, "perplexity": 31.94346308851063, "lr": 0.0026291804804649314, "grad_norm": 0.157195, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:41.491168+00:00", "epoch": 0, "step": 9311, "train_loss": 3.421111822128296, "perplexity": 30.60342167429136, "lr": 0.0026291804804649314, "grad_norm": 0.147405, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:41.796376+00:00", "epoch": 0, "step": 9312, "train_loss": 3.569782018661499, "perplexity": 35.50885204085378, "lr": 0.0026291804804649314, "grad_norm": 0.170526, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:42.100864+00:00", "epoch": 0, "step": 9313, "train_loss": 3.535435914993286, "perplexity": 34.30996768421287, "lr": 0.0026291804804649314, "grad_norm": 0.178043, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:42.405329+00:00", "epoch": 0, "step": 9314, "train_loss": 3.581993579864502, "perplexity": 35.945128950354736, "lr": 0.0026291804804649314, "grad_norm": 0.212647, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:42.710315+00:00", "epoch": 0, "step": 9315, "train_loss": 3.5709781646728516, "perplexity": 35.55135122513784, "lr": 0.0026291804804649314, "grad_norm": 0.168451, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:43.016112+00:00", "epoch": 0, "step": 9316, "train_loss": 3.493943691253662, "perplexity": 32.91550065194529, "lr": 0.0026291804804649314, "grad_norm": 0.165574, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:43.321621+00:00", "epoch": 0, "step": 9317, "train_loss": 3.544632911682129, "perplexity": 34.6269718520666, "lr": 0.0026291804804649314, "grad_norm": 0.203635, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:43.626040+00:00", "epoch": 0, "step": 9318, "train_loss": 3.403090000152588, "perplexity": 30.05683231869712, "lr": 0.0026291804804649314, "grad_norm": 0.18945, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:43.930527+00:00", "epoch": 0, "step": 9319, "train_loss": 3.5103423595428467, "perplexity": 33.45972107807284, "lr": 0.0026291804804649314, "grad_norm": 0.166768, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:44.235420+00:00", "epoch": 0, "step": 9320, "train_loss": 3.554373264312744, "perplexity": 34.96589872243145, "lr": 0.0026291804804649314, "grad_norm": 0.135506, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:44.541196+00:00", "epoch": 0, "step": 9321, "train_loss": 3.4746317863464355, "perplexity": 32.285938224385006, "lr": 0.0026291804804649314, "grad_norm": 0.157587, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:44.845884+00:00", "epoch": 0, "step": 9322, "train_loss": 3.4581196308135986, "perplexity": 31.757205066231585, "lr": 0.0026291804804649314, "grad_norm": 0.159158, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:45.151431+00:00", "epoch": 0, "step": 9323, "train_loss": 3.556647300720215, "perplexity": 35.04550292627067, "lr": 0.0026291804804649314, "grad_norm": 0.18172, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:45.457522+00:00", "epoch": 0, "step": 9324, "train_loss": 3.5107510089874268, "perplexity": 33.47339716868029, "lr": 0.0026291804804649314, "grad_norm": 0.199253, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:45.762977+00:00", "epoch": 0, "step": 9325, "train_loss": 3.5678601264953613, "perplexity": 35.44067339327804, "lr": 0.0026291804804649314, "grad_norm": 0.165513, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:46.068576+00:00", "epoch": 0, "step": 9326, "train_loss": 3.497433662414551, "perplexity": 33.03057548709483, "lr": 0.0026291804804649314, "grad_norm": 0.158287, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:46.374688+00:00", "epoch": 0, "step": 9327, "train_loss": 3.4291129112243652, "perplexity": 30.849264571697095, "lr": 0.0026291804804649314, "grad_norm": 0.1692, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:46.679958+00:00", "epoch": 0, "step": 9328, "train_loss": 3.5428009033203125, "perplexity": 34.563593023087854, "lr": 0.0026291804804649314, "grad_norm": 0.161725, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:46.985646+00:00", "epoch": 0, "step": 9329, "train_loss": 3.426115036010742, "perplexity": 30.75692081277185, "lr": 0.0026291804804649314, "grad_norm": 0.147963, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:47.291517+00:00", "epoch": 0, "step": 9330, "train_loss": 3.5197713375091553, "perplexity": 33.77670411545327, "lr": 0.0026291804804649314, "grad_norm": 0.156836, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:47.596243+00:00", "epoch": 0, "step": 9331, "train_loss": 3.6039459705352783, "perplexity": 36.742935303342726, "lr": 0.0026291804804649314, "grad_norm": 0.157626, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:47.900476+00:00", "epoch": 0, "step": 9332, "train_loss": 3.476349353790283, "perplexity": 32.34143915042097, "lr": 0.0026291804804649314, "grad_norm": 0.145868, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:48.206123+00:00", "epoch": 0, "step": 9333, "train_loss": 3.4916188716888428, "perplexity": 32.839066933771186, "lr": 0.0026291804804649314, "grad_norm": 0.154868, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:48.511325+00:00", "epoch": 0, "step": 9334, "train_loss": 3.4210128784179688, "perplexity": 30.60039380799883, "lr": 0.0026291804804649314, "grad_norm": 0.140192, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:48.815768+00:00", "epoch": 0, "step": 9335, "train_loss": 3.564314365386963, "perplexity": 35.315231756368874, "lr": 0.0026291804804649314, "grad_norm": 0.154373, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:49.120336+00:00", "epoch": 0, "step": 9336, "train_loss": 3.562220335006714, "perplexity": 35.24135796216467, "lr": 0.0026291804804649314, "grad_norm": 0.185847, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:49.425309+00:00", "epoch": 0, "step": 9337, "train_loss": 3.5381367206573486, "perplexity": 34.40275748647651, "lr": 0.0026291804804649314, "grad_norm": 0.196424, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:49.730986+00:00", "epoch": 0, "step": 9338, "train_loss": 3.5816266536712646, "perplexity": 35.93194216046082, "lr": 0.0026291804804649314, "grad_norm": 0.19151, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:50.036498+00:00", "epoch": 0, "step": 9339, "train_loss": 3.6018216609954834, "perplexity": 36.66496478144092, "lr": 0.0026291804804649314, "grad_norm": 0.189705, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:50.340889+00:00", "epoch": 0, "step": 9340, "train_loss": 3.4945199489593506, "perplexity": 32.93447392905609, "lr": 0.0026291804804649314, "grad_norm": 0.187605, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:50.644879+00:00", "epoch": 0, "step": 9341, "train_loss": 3.7007501125335693, "perplexity": 40.4776557720811, "lr": 0.0026291804804649314, "grad_norm": 0.190613, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:50.950228+00:00", "epoch": 0, "step": 9342, "train_loss": 3.547792434692383, "perplexity": 34.736549581932344, "lr": 0.0026291804804649314, "grad_norm": 0.160163, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:51.257142+00:00", "epoch": 0, "step": 9343, "train_loss": 3.4226560592651367, "perplexity": 30.650717122850573, "lr": 0.0026291804804649314, "grad_norm": 0.167847, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:51.561865+00:00", "epoch": 0, "step": 9344, "train_loss": 3.417994499206543, "perplexity": 30.50816946949873, "lr": 0.0026291804804649314, "grad_norm": 0.146591, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:51.866509+00:00", "epoch": 0, "step": 9345, "train_loss": 3.5810229778289795, "perplexity": 35.91025746093533, "lr": 0.0026291804804649314, "grad_norm": 0.16149, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:52.170754+00:00", "epoch": 0, "step": 9346, "train_loss": 3.6151440143585205, "perplexity": 37.15669663817865, "lr": 0.0026291804804649314, "grad_norm": 0.161981, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:52.475034+00:00", "epoch": 0, "step": 9347, "train_loss": 3.522441864013672, "perplexity": 33.86702624917555, "lr": 0.0026291804804649314, "grad_norm": 0.157749, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:52.779317+00:00", "epoch": 0, "step": 9348, "train_loss": 3.5435750484466553, "perplexity": 34.590360619841036, "lr": 0.0026291804804649314, "grad_norm": 0.183843, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:53.084614+00:00", "epoch": 0, "step": 9349, "train_loss": 3.5182271003723145, "perplexity": 33.72458512699407, "lr": 0.0026291804804649314, "grad_norm": 0.181966, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:53.390117+00:00", "epoch": 0, "step": 9350, "train_loss": 3.499776601791382, "perplexity": 33.10805485232808, "lr": 0.0026291804804649314, "grad_norm": 0.165888, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:53.695321+00:00", "epoch": 0, "step": 9351, "train_loss": 3.6014606952667236, "perplexity": 36.651732374075095, "lr": 0.0026291804804649314, "grad_norm": 0.188439, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:53.999555+00:00", "epoch": 0, "step": 9352, "train_loss": 3.5873730182647705, "perplexity": 36.139014587693495, "lr": 0.0026291804804649314, "grad_norm": 0.162751, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:54.304013+00:00", "epoch": 0, "step": 9353, "train_loss": 3.615408182144165, "perplexity": 37.16651353704855, "lr": 0.0026291804804649314, "grad_norm": 0.164568, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:54.608913+00:00", "epoch": 0, "step": 9354, "train_loss": 3.506152868270874, "perplexity": 33.31983509880283, "lr": 0.0026291804804649314, "grad_norm": 0.158006, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:54.914223+00:00", "epoch": 0, "step": 9355, "train_loss": 3.548982858657837, "perplexity": 34.7779254255436, "lr": 0.0026291804804649314, "grad_norm": 0.179336, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:55.218768+00:00", "epoch": 0, "step": 9356, "train_loss": 3.6097114086151123, "perplexity": 36.955386269657396, "lr": 0.0026291804804649314, "grad_norm": 0.180953, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:55.522633+00:00", "epoch": 0, "step": 9357, "train_loss": 3.5541281700134277, "perplexity": 34.95732983012033, "lr": 0.0026291804804649314, "grad_norm": 0.160339, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:55.827721+00:00", "epoch": 0, "step": 9358, "train_loss": 3.540398359298706, "perplexity": 34.48065214375795, "lr": 0.0026291804804649314, "grad_norm": 0.172713, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:56.133095+00:00", "epoch": 0, "step": 9359, "train_loss": 3.508686065673828, "perplexity": 33.40434781704736, "lr": 0.0026291804804649314, "grad_norm": 0.178957, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:56.438990+00:00", "epoch": 0, "step": 9360, "train_loss": 3.4763894081115723, "perplexity": 32.34273459075947, "lr": 0.0026291804804649314, "grad_norm": 0.162872, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:56.743454+00:00", "epoch": 0, "step": 9361, "train_loss": 3.4824230670928955, "perplexity": 32.53846952692015, "lr": 0.0026291804804649314, "grad_norm": 0.169105, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:57.047529+00:00", "epoch": 0, "step": 9362, "train_loss": 3.462712526321411, "perplexity": 31.90339805829801, "lr": 0.0026291804804649314, "grad_norm": 0.162718, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:57.351417+00:00", "epoch": 0, "step": 9363, "train_loss": 3.5749871730804443, "perplexity": 35.6941629665625, "lr": 0.0026291804804649314, "grad_norm": 0.174968, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:57.655812+00:00", "epoch": 0, "step": 9364, "train_loss": 3.64451265335083, "perplexity": 38.264120415958004, "lr": 0.0026291804804649314, "grad_norm": 0.166262, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:57.960940+00:00", "epoch": 0, "step": 9365, "train_loss": 3.4601988792419434, "perplexity": 31.82330488010726, "lr": 0.0026291804804649314, "grad_norm": 0.152474, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:58.265903+00:00", "epoch": 0, "step": 9366, "train_loss": 3.6429619789123535, "perplexity": 38.20483120353181, "lr": 0.0026291804804649314, "grad_norm": 0.161067, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:58.570027+00:00", "epoch": 0, "step": 9367, "train_loss": 3.6256678104400635, "perplexity": 37.54979093081551, "lr": 0.0026291804804649314, "grad_norm": 0.154161, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:58.873566+00:00", "epoch": 0, "step": 9368, "train_loss": 3.608827590942383, "perplexity": 36.92273887546687, "lr": 0.0026291804804649314, "grad_norm": 0.146871, "tokens_per_sec": 107953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:59.178015+00:00", "epoch": 0, "step": 9369, "train_loss": 3.444430112838745, "perplexity": 31.325426400575495, "lr": 0.0026291804804649314, "grad_norm": 0.138387, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:59.483674+00:00", "epoch": 0, "step": 9370, "train_loss": 3.552953004837036, "perplexity": 34.916273322258455, "lr": 0.0026291804804649314, "grad_norm": 0.180491, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:46:59.789304+00:00", "epoch": 0, "step": 9371, "train_loss": 3.5516881942749023, "perplexity": 34.87213876777911, "lr": 0.0026291804804649314, "grad_norm": 0.180827, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:00.093380+00:00", "epoch": 0, "step": 9372, "train_loss": 3.6116204261779785, "perplexity": 37.02600213310471, "lr": 0.0026291804804649314, "grad_norm": 0.16305, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:00.397751+00:00", "epoch": 0, "step": 9373, "train_loss": 3.55137038230896, "perplexity": 34.86105774573468, "lr": 0.0026291804804649314, "grad_norm": 0.197086, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:00.702341+00:00", "epoch": 0, "step": 9374, "train_loss": 3.5458459854125977, "perplexity": 34.66900240991467, "lr": 0.0026291804804649314, "grad_norm": 0.206246, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:01.007596+00:00", "epoch": 0, "step": 9375, "train_loss": 3.4938793182373047, "perplexity": 32.913381850080974, "lr": 0.0026291804804649314, "grad_norm": 0.167853, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:01.311712+00:00", "epoch": 0, "step": 9376, "train_loss": 3.439636468887329, "perplexity": 31.17562279921022, "lr": 0.0026291804804649314, "grad_norm": 0.165146, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:01.616487+00:00", "epoch": 0, "step": 9377, "train_loss": 3.542893409729004, "perplexity": 34.56679052484231, "lr": 0.0026291804804649314, "grad_norm": 0.14703, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:01.922102+00:00", "epoch": 0, "step": 9378, "train_loss": 3.533435106277466, "perplexity": 34.241388631492164, "lr": 0.0026291804804649314, "grad_norm": 0.164863, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:02.226626+00:00", "epoch": 0, "step": 9379, "train_loss": 3.611333131790161, "perplexity": 37.01536629836953, "lr": 0.0026291804804649314, "grad_norm": 0.160313, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:02.532107+00:00", "epoch": 0, "step": 9380, "train_loss": 3.5712101459503174, "perplexity": 35.55959942968821, "lr": 0.0026291804804649314, "grad_norm": 0.158388, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:02.837852+00:00", "epoch": 0, "step": 9381, "train_loss": 3.5408003330230713, "perplexity": 34.49451524603347, "lr": 0.0026291804804649314, "grad_norm": 0.151901, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:03.143294+00:00", "epoch": 0, "step": 9382, "train_loss": 3.6003096103668213, "perplexity": 36.60956739077753, "lr": 0.0026291804804649314, "grad_norm": 0.155325, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:03.448067+00:00", "epoch": 0, "step": 9383, "train_loss": 3.675396680831909, "perplexity": 39.46430848736593, "lr": 0.0026291804804649314, "grad_norm": 0.154508, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:03.753611+00:00", "epoch": 0, "step": 9384, "train_loss": 3.5327067375183105, "perplexity": 34.21645735442677, "lr": 0.0026291804804649314, "grad_norm": 0.162812, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:04.058365+00:00", "epoch": 0, "step": 9385, "train_loss": 3.5955257415771484, "perplexity": 36.434850269032054, "lr": 0.0026291804804649314, "grad_norm": 0.164035, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:04.363951+00:00", "epoch": 0, "step": 9386, "train_loss": 3.6147725582122803, "perplexity": 37.142897117955386, "lr": 0.0026291804804649314, "grad_norm": 0.18419, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:04.668785+00:00", "epoch": 0, "step": 9387, "train_loss": 3.49310302734375, "perplexity": 32.88784140616429, "lr": 0.0026291804804649314, "grad_norm": 0.175208, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:04.972745+00:00", "epoch": 0, "step": 9388, "train_loss": 3.4776484966278076, "perplexity": 32.38348260376079, "lr": 0.0026291804804649314, "grad_norm": 0.156138, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:05.277204+00:00", "epoch": 0, "step": 9389, "train_loss": 3.519632339477539, "perplexity": 33.77200954634223, "lr": 0.0026291804804649314, "grad_norm": 0.164302, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:05.582396+00:00", "epoch": 0, "step": 9390, "train_loss": 3.508481025695801, "perplexity": 33.39749929243959, "lr": 0.0026291804804649314, "grad_norm": 0.169426, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:05.887369+00:00", "epoch": 0, "step": 9391, "train_loss": 3.5218851566314697, "perplexity": 33.84817747276269, "lr": 0.0026291804804649314, "grad_norm": 0.174649, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:06.192727+00:00", "epoch": 0, "step": 9392, "train_loss": 3.6341497898101807, "perplexity": 37.86964205053128, "lr": 0.0026291804804649314, "grad_norm": 0.171878, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:06.498015+00:00", "epoch": 0, "step": 9393, "train_loss": 3.617271661758423, "perplexity": 37.23583714886574, "lr": 0.0026291804804649314, "grad_norm": 0.150216, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:06.803188+00:00", "epoch": 0, "step": 9394, "train_loss": 3.6173617839813232, "perplexity": 37.23919307650071, "lr": 0.0026291804804649314, "grad_norm": 0.143352, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:07.108298+00:00", "epoch": 0, "step": 9395, "train_loss": 3.5967366695404053, "perplexity": 36.47899697190386, "lr": 0.0026291804804649314, "grad_norm": 0.158188, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:07.413012+00:00", "epoch": 0, "step": 9396, "train_loss": 3.6101748943328857, "perplexity": 36.97251853336126, "lr": 0.0026291804804649314, "grad_norm": 0.142003, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:07.718887+00:00", "epoch": 0, "step": 9397, "train_loss": 3.4824466705322266, "perplexity": 32.539237555775586, "lr": 0.0026291804804649314, "grad_norm": 0.148945, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:08.024573+00:00", "epoch": 0, "step": 9398, "train_loss": 3.4793410301208496, "perplexity": 32.43833914285838, "lr": 0.0026291804804649314, "grad_norm": 0.165868, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:08.330057+00:00", "epoch": 0, "step": 9399, "train_loss": 3.5861892700195312, "perplexity": 36.09626040267886, "lr": 0.0026291804804649314, "grad_norm": 0.19207, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:08.635722+00:00", "epoch": 0, "step": 9400, "train_loss": 3.487659215927124, "perplexity": 32.709292633514806, "lr": 0.0026291804804649314, "grad_norm": 0.192996, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:08.940296+00:00", "epoch": 0, "step": 9401, "train_loss": 3.5152387619018555, "perplexity": 33.62395508505852, "lr": 0.0026291804804649314, "grad_norm": 0.172608, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:09.246098+00:00", "epoch": 0, "step": 9402, "train_loss": 3.520817995071411, "perplexity": 33.812075265755695, "lr": 0.0026291804804649314, "grad_norm": 0.184446, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:09.550380+00:00", "epoch": 0, "step": 9403, "train_loss": 3.621408224105835, "perplexity": 37.39018452447295, "lr": 0.0026291804804649314, "grad_norm": 0.175434, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:09.856205+00:00", "epoch": 0, "step": 9404, "train_loss": 3.5286190509796143, "perplexity": 34.07687667768042, "lr": 0.0026291804804649314, "grad_norm": 0.162452, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:10.162199+00:00", "epoch": 0, "step": 9405, "train_loss": 3.459362268447876, "perplexity": 31.796692293482497, "lr": 0.0026291804804649314, "grad_norm": 0.157554, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:10.467513+00:00", "epoch": 0, "step": 9406, "train_loss": 3.5156168937683105, "perplexity": 33.636671778094346, "lr": 0.0026291804804649314, "grad_norm": 0.15141, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:10.772695+00:00", "epoch": 0, "step": 9407, "train_loss": 3.5605671405792236, "perplexity": 35.183145277265766, "lr": 0.0026291804804649314, "grad_norm": 0.157495, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:11.078425+00:00", "epoch": 0, "step": 9408, "train_loss": 3.596889019012451, "perplexity": 36.48455495120015, "lr": 0.0026291804804649314, "grad_norm": 0.161574, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:11.383838+00:00", "epoch": 0, "step": 9409, "train_loss": 3.583512783050537, "perplexity": 35.999778406081376, "lr": 0.0026291804804649314, "grad_norm": 0.155752, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:11.688555+00:00", "epoch": 0, "step": 9410, "train_loss": 3.6212263107299805, "perplexity": 37.383383368411664, "lr": 0.0026291804804649314, "grad_norm": 0.158972, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:11.993044+00:00", "epoch": 0, "step": 9411, "train_loss": 3.60163950920105, "perplexity": 36.658286800534704, "lr": 0.0026291804804649314, "grad_norm": 0.157595, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:12.298239+00:00", "epoch": 0, "step": 9412, "train_loss": 3.633327007293701, "perplexity": 37.83849638595797, "lr": 0.0026291804804649314, "grad_norm": 0.153023, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:12.603783+00:00", "epoch": 0, "step": 9413, "train_loss": 3.5701136589050293, "perplexity": 35.52063015813298, "lr": 0.0026291804804649314, "grad_norm": 0.17336, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:12.909822+00:00", "epoch": 0, "step": 9414, "train_loss": 3.5586330890655518, "perplexity": 35.11516502173491, "lr": 0.0026291804804649314, "grad_norm": 0.154303, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:13.213938+00:00", "epoch": 0, "step": 9415, "train_loss": 3.5746655464172363, "perplexity": 35.682684618001986, "lr": 0.0026291804804649314, "grad_norm": 0.197158, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:13.517978+00:00", "epoch": 0, "step": 9416, "train_loss": 3.509615182876587, "perplexity": 33.435398794015796, "lr": 0.0026291804804649314, "grad_norm": 0.20976, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:13.822177+00:00", "epoch": 0, "step": 9417, "train_loss": 3.4077985286712646, "perplexity": 30.19868947800703, "lr": 0.0026291804804649314, "grad_norm": 0.17838, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:14.127483+00:00", "epoch": 0, "step": 9418, "train_loss": 3.6216835975646973, "perplexity": 37.400482206702065, "lr": 0.0026291804804649314, "grad_norm": 0.180178, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:14.432976+00:00", "epoch": 0, "step": 9419, "train_loss": 3.5806260108947754, "perplexity": 35.896005105168726, "lr": 0.0026291804804649314, "grad_norm": 0.168579, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:14.738809+00:00", "epoch": 0, "step": 9420, "train_loss": 3.6138734817504883, "perplexity": 37.109517820952476, "lr": 0.0026291804804649314, "grad_norm": 0.164086, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:15.043012+00:00", "epoch": 0, "step": 9421, "train_loss": 3.5110628604888916, "perplexity": 33.48383752568237, "lr": 0.0026291804804649314, "grad_norm": 0.15637, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:15.347644+00:00", "epoch": 0, "step": 9422, "train_loss": 3.503898859024048, "perplexity": 33.24481646050098, "lr": 0.0026291804804649314, "grad_norm": 0.146942, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:15.652852+00:00", "epoch": 0, "step": 9423, "train_loss": 3.5583813190460205, "perplexity": 35.106325188800696, "lr": 0.0026291804804649314, "grad_norm": 0.148445, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:15.958751+00:00", "epoch": 0, "step": 9424, "train_loss": 3.5016560554504395, "perplexity": 33.170338418368694, "lr": 0.0026291804804649314, "grad_norm": 0.165342, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:16.264714+00:00", "epoch": 0, "step": 9425, "train_loss": 3.652398109436035, "perplexity": 38.56704322975578, "lr": 0.0026291804804649314, "grad_norm": 0.169457, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:16.569720+00:00", "epoch": 0, "step": 9426, "train_loss": 3.392699718475342, "perplexity": 29.74615019700381, "lr": 0.0026291804804649314, "grad_norm": 0.175837, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:16.874992+00:00", "epoch": 0, "step": 9427, "train_loss": 3.56133770942688, "perplexity": 35.21026676112226, "lr": 0.0026291804804649314, "grad_norm": 0.173084, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:17.180263+00:00", "epoch": 0, "step": 9428, "train_loss": 3.5747482776641846, "perplexity": 35.6856368131126, "lr": 0.0026291804804649314, "grad_norm": 0.148042, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:17.485277+00:00", "epoch": 0, "step": 9429, "train_loss": 3.5657761096954346, "perplexity": 35.36689134274588, "lr": 0.0026291804804649314, "grad_norm": 0.148995, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:17.790367+00:00", "epoch": 0, "step": 9430, "train_loss": 3.5640194416046143, "perplexity": 35.30481799035388, "lr": 0.0026291804804649314, "grad_norm": 0.156449, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:18.095779+00:00", "epoch": 0, "step": 9431, "train_loss": 3.651215076446533, "perplexity": 38.521444123249324, "lr": 0.0026291804804649314, "grad_norm": 0.155482, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:18.400863+00:00", "epoch": 0, "step": 9432, "train_loss": 3.6381092071533203, "perplexity": 38.0198810011235, "lr": 0.0026291804804649314, "grad_norm": 0.174747, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:18.705578+00:00", "epoch": 0, "step": 9433, "train_loss": 3.5672881603240967, "perplexity": 35.42040832303004, "lr": 0.0026291804804649314, "grad_norm": 0.158196, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:19.010361+00:00", "epoch": 0, "step": 9434, "train_loss": 3.6568236351013184, "perplexity": 38.738100900274475, "lr": 0.0026291804804649314, "grad_norm": 0.160102, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:19.315526+00:00", "epoch": 0, "step": 9435, "train_loss": 3.5567686557769775, "perplexity": 35.049756133336444, "lr": 0.0026291804804649314, "grad_norm": 0.166014, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:19.621869+00:00", "epoch": 0, "step": 9436, "train_loss": 3.507647752761841, "perplexity": 33.36968165167041, "lr": 0.0026291804804649314, "grad_norm": 0.167355, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:19.927381+00:00", "epoch": 0, "step": 9437, "train_loss": 3.494103193283081, "perplexity": 32.92075115982081, "lr": 0.0026291804804649314, "grad_norm": 0.155322, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:20.231871+00:00", "epoch": 0, "step": 9438, "train_loss": 3.5375428199768066, "perplexity": 34.38233173142806, "lr": 0.0026291804804649314, "grad_norm": 0.158721, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:20.537320+00:00", "epoch": 0, "step": 9439, "train_loss": 3.5221123695373535, "perplexity": 33.85586908930934, "lr": 0.0026291804804649314, "grad_norm": 0.153799, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:20.843378+00:00", "epoch": 0, "step": 9440, "train_loss": 3.523862361907959, "perplexity": 33.915168473517255, "lr": 0.0026291804804649314, "grad_norm": 0.153478, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:21.149177+00:00", "epoch": 0, "step": 9441, "train_loss": 3.642380475997925, "perplexity": 38.18262144098836, "lr": 0.0026291804804649314, "grad_norm": 0.14858, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:21.454011+00:00", "epoch": 0, "step": 9442, "train_loss": 3.608320474624634, "perplexity": 36.90401949893895, "lr": 0.0026291804804649314, "grad_norm": 0.152292, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:21.758846+00:00", "epoch": 0, "step": 9443, "train_loss": 3.5700390338897705, "perplexity": 35.51797952946826, "lr": 0.0026291804804649314, "grad_norm": 0.177842, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:22.063552+00:00", "epoch": 0, "step": 9444, "train_loss": 3.6015233993530273, "perplexity": 36.65403065952025, "lr": 0.0026291804804649314, "grad_norm": 0.177098, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:22.368632+00:00", "epoch": 0, "step": 9445, "train_loss": 3.6219282150268555, "perplexity": 37.40963213681368, "lr": 0.0026291804804649314, "grad_norm": 0.159985, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:22.675198+00:00", "epoch": 0, "step": 9446, "train_loss": 3.4650533199310303, "perplexity": 31.97816480121582, "lr": 0.0026291804804649314, "grad_norm": 0.168676, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:22.979482+00:00", "epoch": 0, "step": 9447, "train_loss": 3.5397515296936035, "perplexity": 34.45835624874797, "lr": 0.0026291804804649314, "grad_norm": 0.171569, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:23.283665+00:00", "epoch": 0, "step": 9448, "train_loss": 3.5959041118621826, "perplexity": 36.448638742123265, "lr": 0.0026291804804649314, "grad_norm": 0.173088, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:23.587395+00:00", "epoch": 0, "step": 9449, "train_loss": 3.5116922855377197, "perplexity": 33.50491972588835, "lr": 0.0026291804804649314, "grad_norm": 0.160153, "tokens_per_sec": 107884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:23.891950+00:00", "epoch": 0, "step": 9450, "train_loss": 3.5478434562683105, "perplexity": 34.73832194064819, "lr": 0.0026291804804649314, "grad_norm": 0.161796, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:24.196657+00:00", "epoch": 0, "step": 9451, "train_loss": 3.537208080291748, "perplexity": 34.370824526599215, "lr": 0.0026291804804649314, "grad_norm": 0.135888, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:24.502386+00:00", "epoch": 0, "step": 9452, "train_loss": 3.5324339866638184, "perplexity": 34.207126059062894, "lr": 0.0026291804804649314, "grad_norm": 0.162963, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:24.807214+00:00", "epoch": 0, "step": 9453, "train_loss": 3.465139865875244, "perplexity": 31.980932501447654, "lr": 0.0026291804804649314, "grad_norm": 0.142162, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:25.111303+00:00", "epoch": 0, "step": 9454, "train_loss": 3.5181214809417725, "perplexity": 33.72102334361777, "lr": 0.0026291804804649314, "grad_norm": 0.149287, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:25.415739+00:00", "epoch": 0, "step": 9455, "train_loss": 3.584035634994507, "perplexity": 36.018605881765694, "lr": 0.0026291804804649314, "grad_norm": 0.139695, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:25.719895+00:00", "epoch": 0, "step": 9456, "train_loss": 3.615300416946411, "perplexity": 37.162508496173295, "lr": 0.0026291804804649314, "grad_norm": 0.153119, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:26.024930+00:00", "epoch": 0, "step": 9457, "train_loss": 3.5123887062072754, "perplexity": 33.528261371376054, "lr": 0.0026291804804649314, "grad_norm": 0.16076, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:26.330352+00:00", "epoch": 0, "step": 9458, "train_loss": 3.5937352180480957, "perplexity": 36.36967118210575, "lr": 0.0026291804804649314, "grad_norm": 0.164225, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:26.634920+00:00", "epoch": 0, "step": 9459, "train_loss": 3.5375397205352783, "perplexity": 34.382225165566396, "lr": 0.0026291804804649314, "grad_norm": 0.17159, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:26.939824+00:00", "epoch": 0, "step": 9460, "train_loss": 3.605438709259033, "perplexity": 36.797823862641216, "lr": 0.0026291804804649314, "grad_norm": 0.171362, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:27.244966+00:00", "epoch": 0, "step": 9461, "train_loss": 3.5068798065185547, "perplexity": 33.34406536722997, "lr": 0.0026291804804649314, "grad_norm": 0.147939, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:27.550380+00:00", "epoch": 0, "step": 9462, "train_loss": 3.481107473373413, "perplexity": 32.495690267001955, "lr": 0.0026291804804649314, "grad_norm": 0.185406, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:27.854978+00:00", "epoch": 0, "step": 9463, "train_loss": 3.4410645961761475, "perplexity": 31.220177364098063, "lr": 0.0026291804804649314, "grad_norm": 0.180176, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:28.160021+00:00", "epoch": 0, "step": 9464, "train_loss": 3.5439908504486084, "perplexity": 34.604746351635825, "lr": 0.0026291804804649314, "grad_norm": 0.159132, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:28.464334+00:00", "epoch": 0, "step": 9465, "train_loss": 3.5801079273223877, "perplexity": 35.87741279121147, "lr": 0.0026291804804649314, "grad_norm": 0.160884, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:28.768896+00:00", "epoch": 0, "step": 9466, "train_loss": 3.551044225692749, "perplexity": 34.84968943512843, "lr": 0.0026291804804649314, "grad_norm": 0.155943, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:29.074035+00:00", "epoch": 0, "step": 9467, "train_loss": 3.6047441959381104, "perplexity": 36.77227615642569, "lr": 0.0026291804804649314, "grad_norm": 0.175157, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:29.379470+00:00", "epoch": 0, "step": 9468, "train_loss": 3.537736415863037, "perplexity": 34.38898865376588, "lr": 0.0026291804804649314, "grad_norm": 0.163639, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:29.684313+00:00", "epoch": 0, "step": 9469, "train_loss": 3.621330499649048, "perplexity": 37.38727850562744, "lr": 0.0026291804804649314, "grad_norm": 0.166476, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:29.988940+00:00", "epoch": 0, "step": 9470, "train_loss": 3.610572576522827, "perplexity": 36.98722476950954, "lr": 0.0026291804804649314, "grad_norm": 0.187684, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:30.293514+00:00", "epoch": 0, "step": 9471, "train_loss": 3.604421377182007, "perplexity": 36.76040729182691, "lr": 0.0026291804804649314, "grad_norm": 0.17316, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:30.598824+00:00", "epoch": 0, "step": 9472, "train_loss": 3.600937604904175, "perplexity": 36.63256521961343, "lr": 0.0026291804804649314, "grad_norm": 0.198219, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:30.903781+00:00", "epoch": 0, "step": 9473, "train_loss": 3.520197629928589, "perplexity": 33.7911059378469, "lr": 0.0026291804804649314, "grad_norm": 0.246036, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:31.208734+00:00", "epoch": 0, "step": 9474, "train_loss": 3.5000035762786865, "perplexity": 33.11557038898912, "lr": 0.0026291804804649314, "grad_norm": 0.235827, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:31.514627+00:00", "epoch": 0, "step": 9475, "train_loss": 3.659113883972168, "perplexity": 38.8269224650402, "lr": 0.0026291804804649314, "grad_norm": 0.155634, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:31.819312+00:00", "epoch": 0, "step": 9476, "train_loss": 3.5394797325134277, "perplexity": 34.44899183735402, "lr": 0.0026291804804649314, "grad_norm": 0.19654, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:32.124752+00:00", "epoch": 0, "step": 9477, "train_loss": 3.574610948562622, "perplexity": 35.680736473157715, "lr": 0.0026291804804649314, "grad_norm": 0.161688, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:32.430114+00:00", "epoch": 0, "step": 9478, "train_loss": 3.4927542209625244, "perplexity": 32.87637191764877, "lr": 0.0026291804804649314, "grad_norm": 0.177956, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:32.734211+00:00", "epoch": 0, "step": 9479, "train_loss": 3.4852559566497803, "perplexity": 32.63077810571104, "lr": 0.0026291804804649314, "grad_norm": 0.164091, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:33.038782+00:00", "epoch": 0, "step": 9480, "train_loss": 3.509402275085449, "perplexity": 33.42828089486773, "lr": 0.0026291804804649314, "grad_norm": 0.169508, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:33.343675+00:00", "epoch": 0, "step": 9481, "train_loss": 3.464975595474243, "perplexity": 31.975679412316598, "lr": 0.0026291804804649314, "grad_norm": 0.174489, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:33.648819+00:00", "epoch": 0, "step": 9482, "train_loss": 3.4358723163604736, "perplexity": 31.058493584288208, "lr": 0.0026291804804649314, "grad_norm": 0.185657, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:33.954718+00:00", "epoch": 0, "step": 9483, "train_loss": 3.5430386066436768, "perplexity": 34.571809880565326, "lr": 0.0026291804804649314, "grad_norm": 0.160717, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:34.261055+00:00", "epoch": 0, "step": 9484, "train_loss": 3.4786107540130615, "perplexity": 32.41465884643577, "lr": 0.0026291804804649314, "grad_norm": 0.175001, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:34.566230+00:00", "epoch": 0, "step": 9485, "train_loss": 3.358215093612671, "perplexity": 28.73785070089276, "lr": 0.0026291804804649314, "grad_norm": 0.173485, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:34.870324+00:00", "epoch": 0, "step": 9486, "train_loss": 3.535217523574829, "perplexity": 34.30247549984726, "lr": 0.0026291804804649314, "grad_norm": 0.170817, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:35.175994+00:00", "epoch": 0, "step": 9487, "train_loss": 3.528823137283325, "perplexity": 34.0838320112036, "lr": 0.0026291804804649314, "grad_norm": 0.18618, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:35.481623+00:00", "epoch": 0, "step": 9488, "train_loss": 3.510843276977539, "perplexity": 33.476485834249566, "lr": 0.0026291804804649314, "grad_norm": 0.170261, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:35.787018+00:00", "epoch": 0, "step": 9489, "train_loss": 3.5540332794189453, "perplexity": 34.954012865688085, "lr": 0.0026291804804649314, "grad_norm": 0.179141, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:36.092178+00:00", "epoch": 0, "step": 9490, "train_loss": 3.616251230239868, "perplexity": 37.197859906902025, "lr": 0.0026291804804649314, "grad_norm": 0.164839, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:36.396620+00:00", "epoch": 0, "step": 9491, "train_loss": 3.5704972743988037, "perplexity": 35.534259036167526, "lr": 0.0026291804804649314, "grad_norm": 0.167161, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:36.701643+00:00", "epoch": 0, "step": 9492, "train_loss": 3.4814565181732178, "perplexity": 32.50703469844794, "lr": 0.0026291804804649314, "grad_norm": 0.138522, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:37.006429+00:00", "epoch": 0, "step": 9493, "train_loss": 3.58563494682312, "perplexity": 36.07625695293427, "lr": 0.0026291804804649314, "grad_norm": 0.156539, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:37.311672+00:00", "epoch": 0, "step": 9494, "train_loss": 3.556428909301758, "perplexity": 35.03785012485956, "lr": 0.0026291804804649314, "grad_norm": 0.142941, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:37.616151+00:00", "epoch": 0, "step": 9495, "train_loss": 3.6009035110473633, "perplexity": 36.63131629547064, "lr": 0.0026291804804649314, "grad_norm": 0.146362, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:37.920690+00:00", "epoch": 0, "step": 9496, "train_loss": 3.5725114345550537, "perplexity": 35.60590285173879, "lr": 0.0026291804804649314, "grad_norm": 0.169286, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:38.226160+00:00", "epoch": 0, "step": 9497, "train_loss": 3.4974279403686523, "perplexity": 33.03038648516658, "lr": 0.0026291804804649314, "grad_norm": 0.147996, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:38.531921+00:00", "epoch": 0, "step": 9498, "train_loss": 3.5185258388519287, "perplexity": 33.734661463300284, "lr": 0.0026291804804649314, "grad_norm": 0.155836, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:38.837738+00:00", "epoch": 0, "step": 9499, "train_loss": 3.5685834884643555, "perplexity": 35.46631910301356, "lr": 0.0026291804804649314, "grad_norm": 0.139294, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:39.142794+00:00", "epoch": 0, "step": 9500, "train_loss": 3.6975269317626953, "perplexity": 40.34739900367071, "lr": 0.0026291804804649314, "grad_norm": 0.157484, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:47:42.217151+00:00", "step": 9500, "epoch": 0, "val_loss": 3.5020875930786133, "val_ppl": 33.18465575654837, "eval_train_loss": 3.6975269317626953, "eval_train_ppl": 40.34739900367071} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:47:43.223297+00:00", "step": 9500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p5021_epoch_0000_step_0009500.pt", "val_loss": 3.5020875930786133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:44.343770+00:00", "epoch": 0, "step": 9501, "train_loss": 3.4591379165649414, "perplexity": 31.78955944585918, "lr": 0.0026291804804649314, "grad_norm": 0.169579, "tokens_per_sec": 6300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:44.647127+00:00", "epoch": 0, "step": 9502, "train_loss": 3.4345591068267822, "perplexity": 31.017734043177764, "lr": 0.0026291804804649314, "grad_norm": 0.149998, "tokens_per_sec": 108009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:44.949389+00:00", "epoch": 0, "step": 9503, "train_loss": 3.551727533340454, "perplexity": 34.87351063211578, "lr": 0.0026291804804649314, "grad_norm": 0.179829, "tokens_per_sec": 108409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:45.252018+00:00", "epoch": 0, "step": 9504, "train_loss": 3.5165491104125977, "perplexity": 33.668043063529645, "lr": 0.0026291804804649314, "grad_norm": 0.179783, "tokens_per_sec": 108278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:45.556298+00:00", "epoch": 0, "step": 9505, "train_loss": 3.625223398208618, "perplexity": 37.533107051972145, "lr": 0.0026291804804649314, "grad_norm": 0.170469, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:45.861043+00:00", "epoch": 0, "step": 9506, "train_loss": 3.5659072399139404, "perplexity": 35.371529315018044, "lr": 0.0026291804804649314, "grad_norm": 0.178418, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:46.164601+00:00", "epoch": 0, "step": 9507, "train_loss": 3.5812785625457764, "perplexity": 35.91943674691122, "lr": 0.0026291804804649314, "grad_norm": 0.165168, "tokens_per_sec": 107947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:46.468517+00:00", "epoch": 0, "step": 9508, "train_loss": 3.6875405311584473, "perplexity": 39.94647892007736, "lr": 0.0026291804804649314, "grad_norm": 0.201946, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:46.772211+00:00", "epoch": 0, "step": 9509, "train_loss": 3.730189085006714, "perplexity": 41.68698980360277, "lr": 0.0026291804804649314, "grad_norm": 0.152874, "tokens_per_sec": 107901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:47.076089+00:00", "epoch": 0, "step": 9510, "train_loss": 3.6267199516296387, "perplexity": 37.58931940363446, "lr": 0.0026291804804649314, "grad_norm": 0.179329, "tokens_per_sec": 107888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:47.381102+00:00", "epoch": 0, "step": 9511, "train_loss": 3.510989189147949, "perplexity": 33.48137081733592, "lr": 0.0026291804804649314, "grad_norm": 0.142805, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:47.684724+00:00", "epoch": 0, "step": 9512, "train_loss": 3.6184470653533936, "perplexity": 37.279630017818135, "lr": 0.0026291804804649314, "grad_norm": 0.199345, "tokens_per_sec": 107925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:47.990008+00:00", "epoch": 0, "step": 9513, "train_loss": 3.6265017986297607, "perplexity": 37.58112007522975, "lr": 0.0026291804804649314, "grad_norm": 0.162979, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:48.294929+00:00", "epoch": 0, "step": 9514, "train_loss": 3.4689137935638428, "perplexity": 32.10185425958839, "lr": 0.0026291804804649314, "grad_norm": 0.186375, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:48.600612+00:00", "epoch": 0, "step": 9515, "train_loss": 3.4881420135498047, "perplexity": 32.725088415013886, "lr": 0.0026291804804649314, "grad_norm": 0.195792, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:48.904616+00:00", "epoch": 0, "step": 9516, "train_loss": 3.5598716735839844, "perplexity": 35.15868506755689, "lr": 0.0026291804804649314, "grad_norm": 0.15511, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:49.208438+00:00", "epoch": 0, "step": 9517, "train_loss": 3.557581663131714, "perplexity": 35.078263429605094, "lr": 0.0026291804804649314, "grad_norm": 0.174388, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:49.512847+00:00", "epoch": 0, "step": 9518, "train_loss": 3.5223405361175537, "perplexity": 33.86359474851425, "lr": 0.0026291804804649314, "grad_norm": 0.196597, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:49.817296+00:00", "epoch": 0, "step": 9519, "train_loss": 3.5668981075286865, "perplexity": 35.40659518785019, "lr": 0.0026291804804649314, "grad_norm": 0.169538, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:50.122374+00:00", "epoch": 0, "step": 9520, "train_loss": 3.5716488361358643, "perplexity": 35.57520249916649, "lr": 0.0026291804804649314, "grad_norm": 0.187654, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:50.428034+00:00", "epoch": 0, "step": 9521, "train_loss": 3.438021659851074, "perplexity": 31.12532074683821, "lr": 0.0026291804804649314, "grad_norm": 0.181776, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:50.733544+00:00", "epoch": 0, "step": 9522, "train_loss": 3.50327205657959, "perplexity": 33.223985057547154, "lr": 0.0026291804804649314, "grad_norm": 0.188493, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:51.037153+00:00", "epoch": 0, "step": 9523, "train_loss": 3.6665537357330322, "perplexity": 39.11686624448154, "lr": 0.0026291804804649314, "grad_norm": 0.160228, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:51.341529+00:00", "epoch": 0, "step": 9524, "train_loss": 3.6155993938446045, "perplexity": 37.17362088878413, "lr": 0.0026291804804649314, "grad_norm": 0.187488, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:51.646254+00:00", "epoch": 0, "step": 9525, "train_loss": 3.4285836219787598, "perplexity": 30.832940708128266, "lr": 0.0026291804804649314, "grad_norm": 0.156581, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:51.951627+00:00", "epoch": 0, "step": 9526, "train_loss": 3.573671817779541, "perplexity": 35.64724332485772, "lr": 0.0026291804804649314, "grad_norm": 0.174046, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:52.256617+00:00", "epoch": 0, "step": 9527, "train_loss": 3.518524646759033, "perplexity": 33.73462124847399, "lr": 0.0026291804804649314, "grad_norm": 0.164912, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:52.560618+00:00", "epoch": 0, "step": 9528, "train_loss": 3.6012730598449707, "perplexity": 36.64485585597247, "lr": 0.0026291804804649314, "grad_norm": 0.145543, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:52.865037+00:00", "epoch": 0, "step": 9529, "train_loss": 3.7242729663848877, "perplexity": 41.4410927222975, "lr": 0.0026291804804649314, "grad_norm": 0.171831, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:53.169594+00:00", "epoch": 0, "step": 9530, "train_loss": 3.627185106277466, "perplexity": 37.60680831747316, "lr": 0.0026291804804649314, "grad_norm": 0.166404, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:53.474827+00:00", "epoch": 0, "step": 9531, "train_loss": 3.531968832015991, "perplexity": 34.19121815549219, "lr": 0.0026291804804649314, "grad_norm": 0.158131, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:53.779403+00:00", "epoch": 0, "step": 9532, "train_loss": 3.537876844406128, "perplexity": 34.39381818843521, "lr": 0.0026291804804649314, "grad_norm": 0.17189, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:54.084556+00:00", "epoch": 0, "step": 9533, "train_loss": 3.4818074703216553, "perplexity": 32.51844511425254, "lr": 0.0026291804804649314, "grad_norm": 0.175631, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:54.388672+00:00", "epoch": 0, "step": 9534, "train_loss": 3.521190881729126, "perplexity": 33.82468568847502, "lr": 0.0026291804804649314, "grad_norm": 0.138305, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:54.693097+00:00", "epoch": 0, "step": 9535, "train_loss": 3.518434762954712, "perplexity": 33.73158918864742, "lr": 0.0026291804804649314, "grad_norm": 0.167107, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:54.998255+00:00", "epoch": 0, "step": 9536, "train_loss": 3.4683592319488525, "perplexity": 32.08405673881306, "lr": 0.0026291804804649314, "grad_norm": 0.169007, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:55.303670+00:00", "epoch": 0, "step": 9537, "train_loss": 3.5310633182525635, "perplexity": 34.160271550267076, "lr": 0.0026291804804649314, "grad_norm": 0.152572, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:55.607544+00:00", "epoch": 0, "step": 9538, "train_loss": 3.590790271759033, "perplexity": 36.26272201100077, "lr": 0.0026291804804649314, "grad_norm": 0.170543, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:55.911618+00:00", "epoch": 0, "step": 9539, "train_loss": 3.5084340572357178, "perplexity": 33.395930700164676, "lr": 0.0026291804804649314, "grad_norm": 0.184903, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:56.215914+00:00", "epoch": 0, "step": 9540, "train_loss": 3.5114071369171143, "perplexity": 33.49536720625364, "lr": 0.0026291804804649314, "grad_norm": 0.138256, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:56.520248+00:00", "epoch": 0, "step": 9541, "train_loss": 3.604872941970825, "perplexity": 36.77701074586799, "lr": 0.0026291804804649314, "grad_norm": 0.189697, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:56.825108+00:00", "epoch": 0, "step": 9542, "train_loss": 3.6545228958129883, "perplexity": 38.64907707914925, "lr": 0.0026291804804649314, "grad_norm": 0.144473, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:57.129609+00:00", "epoch": 0, "step": 9543, "train_loss": 3.4709393978118896, "perplexity": 32.16694581455564, "lr": 0.0026291804804649314, "grad_norm": 0.173334, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:57.433405+00:00", "epoch": 0, "step": 9544, "train_loss": 3.66206431388855, "perplexity": 38.94164773985962, "lr": 0.0026291804804649314, "grad_norm": 0.177706, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:57.738044+00:00", "epoch": 0, "step": 9545, "train_loss": 3.504693031311035, "perplexity": 33.271229059108585, "lr": 0.0026291804804649314, "grad_norm": 0.144525, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:58.041713+00:00", "epoch": 0, "step": 9546, "train_loss": 3.5590169429779053, "perplexity": 35.12864670254387, "lr": 0.0026291804804649314, "grad_norm": 0.172105, "tokens_per_sec": 107898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:58.345837+00:00", "epoch": 0, "step": 9547, "train_loss": 3.584629535675049, "perplexity": 36.04000370977441, "lr": 0.0026291804804649314, "grad_norm": 0.157798, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:58.650066+00:00", "epoch": 0, "step": 9548, "train_loss": 3.633744955062866, "perplexity": 37.85431420639216, "lr": 0.0026291804804649314, "grad_norm": 0.16441, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:58.953938+00:00", "epoch": 0, "step": 9549, "train_loss": 3.6012356281280518, "perplexity": 36.64348420177338, "lr": 0.0026291804804649314, "grad_norm": 0.146013, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:59.257939+00:00", "epoch": 0, "step": 9550, "train_loss": 3.5469934940338135, "perplexity": 34.708808223458156, "lr": 0.0026291804804649314, "grad_norm": 0.195813, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:59.562712+00:00", "epoch": 0, "step": 9551, "train_loss": 3.6195878982543945, "perplexity": 37.322184115221454, "lr": 0.0026291804804649314, "grad_norm": 0.195133, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:47:59.866922+00:00", "epoch": 0, "step": 9552, "train_loss": 3.524068593978882, "perplexity": 33.92216359023115, "lr": 0.0026291804804649314, "grad_norm": 0.15501, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:00.171214+00:00", "epoch": 0, "step": 9553, "train_loss": 3.4925553798675537, "perplexity": 32.86983539374432, "lr": 0.0026291804804649314, "grad_norm": 0.169898, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:00.475051+00:00", "epoch": 0, "step": 9554, "train_loss": 3.5747880935668945, "perplexity": 35.6870576972428, "lr": 0.0026291804804649314, "grad_norm": 0.177722, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:00.778515+00:00", "epoch": 0, "step": 9555, "train_loss": 3.5046966075897217, "perplexity": 33.27134804650871, "lr": 0.0026291804804649314, "grad_norm": 0.149128, "tokens_per_sec": 107979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:01.082455+00:00", "epoch": 0, "step": 9556, "train_loss": 3.5773651599884033, "perplexity": 35.779144220896505, "lr": 0.0026291804804649314, "grad_norm": 0.163033, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:01.387086+00:00", "epoch": 0, "step": 9557, "train_loss": 3.5053088665008545, "perplexity": 33.29172496317345, "lr": 0.0026291804804649314, "grad_norm": 0.147172, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:01.692016+00:00", "epoch": 0, "step": 9558, "train_loss": 3.5985770225524902, "perplexity": 36.54619301711553, "lr": 0.0026291804804649314, "grad_norm": 0.161966, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:01.997025+00:00", "epoch": 0, "step": 9559, "train_loss": 3.5392696857452393, "perplexity": 34.441756697836674, "lr": 0.0026291804804649314, "grad_norm": 0.158912, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:02.301157+00:00", "epoch": 0, "step": 9560, "train_loss": 3.6179723739624023, "perplexity": 37.26193789787007, "lr": 0.0026291804804649314, "grad_norm": 0.138241, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:02.604924+00:00", "epoch": 0, "step": 9561, "train_loss": 3.4928762912750244, "perplexity": 32.88038539160065, "lr": 0.0026291804804649314, "grad_norm": 0.161173, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:02.909741+00:00", "epoch": 0, "step": 9562, "train_loss": 3.4792213439941406, "perplexity": 32.434456956016085, "lr": 0.0026291804804649314, "grad_norm": 0.145618, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:03.215023+00:00", "epoch": 0, "step": 9563, "train_loss": 3.4798455238342285, "perplexity": 32.45470820972689, "lr": 0.0026291804804649314, "grad_norm": 0.146175, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:03.520193+00:00", "epoch": 0, "step": 9564, "train_loss": 3.408722400665283, "perplexity": 30.226602093328236, "lr": 0.0026291804804649314, "grad_norm": 0.163882, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:03.823913+00:00", "epoch": 0, "step": 9565, "train_loss": 3.5092275142669678, "perplexity": 33.42243945158069, "lr": 0.0026291804804649314, "grad_norm": 0.192292, "tokens_per_sec": 107888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:04.127746+00:00", "epoch": 0, "step": 9566, "train_loss": 3.522749185562134, "perplexity": 33.87743591559972, "lr": 0.0026291804804649314, "grad_norm": 0.182406, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:04.433088+00:00", "epoch": 0, "step": 9567, "train_loss": 3.507579803466797, "perplexity": 33.367414282360286, "lr": 0.0026291804804649314, "grad_norm": 0.166158, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:04.738416+00:00", "epoch": 0, "step": 9568, "train_loss": 3.5532360076904297, "perplexity": 34.926156125603335, "lr": 0.0026291804804649314, "grad_norm": 0.178287, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:05.044176+00:00", "epoch": 0, "step": 9569, "train_loss": 3.652167320251465, "perplexity": 38.558143400329186, "lr": 0.0026291804804649314, "grad_norm": 0.147047, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:05.348249+00:00", "epoch": 0, "step": 9570, "train_loss": 3.5354857444763184, "perplexity": 34.31167737476157, "lr": 0.0026291804804649314, "grad_norm": 0.140799, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:05.652676+00:00", "epoch": 0, "step": 9571, "train_loss": 3.591597318649292, "perplexity": 36.29199954061258, "lr": 0.0026291804804649314, "grad_norm": 0.170553, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:05.957197+00:00", "epoch": 0, "step": 9572, "train_loss": 3.4934144020080566, "perplexity": 32.89808344121427, "lr": 0.0026291804804649314, "grad_norm": 0.166984, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:06.261486+00:00", "epoch": 0, "step": 9573, "train_loss": 3.5914082527160645, "perplexity": 36.285138608455455, "lr": 0.0026291804804649314, "grad_norm": 0.155424, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:06.566187+00:00", "epoch": 0, "step": 9574, "train_loss": 3.506042242050171, "perplexity": 33.31614925525062, "lr": 0.0026291804804649314, "grad_norm": 0.161314, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:06.870810+00:00", "epoch": 0, "step": 9575, "train_loss": 3.5671546459198, "perplexity": 35.41567950400259, "lr": 0.0026291804804649314, "grad_norm": 0.18156, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:07.175776+00:00", "epoch": 0, "step": 9576, "train_loss": 3.58711576461792, "perplexity": 36.12971889012454, "lr": 0.0026291804804649314, "grad_norm": 0.155227, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:07.480157+00:00", "epoch": 0, "step": 9577, "train_loss": 3.4812655448913574, "perplexity": 32.50082731608898, "lr": 0.0026291804804649314, "grad_norm": 0.147814, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:07.784310+00:00", "epoch": 0, "step": 9578, "train_loss": 3.6381545066833496, "perplexity": 38.02160332287451, "lr": 0.0026291804804649314, "grad_norm": 0.148594, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:08.089882+00:00", "epoch": 0, "step": 9579, "train_loss": 3.5337581634521484, "perplexity": 34.25245234477041, "lr": 0.0026291804804649314, "grad_norm": 0.157499, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:08.394253+00:00", "epoch": 0, "step": 9580, "train_loss": 3.50185489654541, "perplexity": 33.1769347005646, "lr": 0.0026291804804649314, "grad_norm": 0.175709, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:08.699212+00:00", "epoch": 0, "step": 9581, "train_loss": 3.658637046813965, "perplexity": 38.808412759078706, "lr": 0.0026291804804649314, "grad_norm": 0.169507, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:09.004272+00:00", "epoch": 0, "step": 9582, "train_loss": 3.5276389122009277, "perplexity": 34.043492972396734, "lr": 0.0026291804804649314, "grad_norm": 0.160593, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:09.308951+00:00", "epoch": 0, "step": 9583, "train_loss": 3.6203181743621826, "perplexity": 37.349449569014276, "lr": 0.0026291804804649314, "grad_norm": 0.182848, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:09.615256+00:00", "epoch": 0, "step": 9584, "train_loss": 3.511456251144409, "perplexity": 33.49701234573148, "lr": 0.0026291804804649314, "grad_norm": 0.187155, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:09.919635+00:00", "epoch": 0, "step": 9585, "train_loss": 3.495352268218994, "perplexity": 32.961897336945604, "lr": 0.0026291804804649314, "grad_norm": 0.141027, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:10.223496+00:00", "epoch": 0, "step": 9586, "train_loss": 3.6047065258026123, "perplexity": 36.77089096589063, "lr": 0.0026291804804649314, "grad_norm": 0.174587, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:10.527028+00:00", "epoch": 0, "step": 9587, "train_loss": 3.5048437118530273, "perplexity": 33.276242763660335, "lr": 0.0026291804804649314, "grad_norm": 0.154473, "tokens_per_sec": 107953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:10.831344+00:00", "epoch": 0, "step": 9588, "train_loss": 3.4940640926361084, "perplexity": 32.919463962316925, "lr": 0.0026291804804649314, "grad_norm": 0.157368, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:11.136650+00:00", "epoch": 0, "step": 9589, "train_loss": 3.545483112335205, "perplexity": 34.65642424459685, "lr": 0.0026291804804649314, "grad_norm": 0.170809, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:11.442300+00:00", "epoch": 0, "step": 9590, "train_loss": 3.5497140884399414, "perplexity": 34.80336538046736, "lr": 0.0026291804804649314, "grad_norm": 0.165015, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:11.747323+00:00", "epoch": 0, "step": 9591, "train_loss": 3.5589659214019775, "perplexity": 35.12685442935159, "lr": 0.0026291804804649314, "grad_norm": 0.164203, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:12.052184+00:00", "epoch": 0, "step": 9592, "train_loss": 3.5712344646453857, "perplexity": 35.56046420325854, "lr": 0.0026291804804649314, "grad_norm": 0.152906, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:12.355868+00:00", "epoch": 0, "step": 9593, "train_loss": 3.5110957622528076, "perplexity": 33.4849392211234, "lr": 0.0026291804804649314, "grad_norm": 0.166945, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:12.659910+00:00", "epoch": 0, "step": 9594, "train_loss": 3.5216333866119385, "perplexity": 33.83965658915581, "lr": 0.0026291804804649314, "grad_norm": 0.145122, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:12.964757+00:00", "epoch": 0, "step": 9595, "train_loss": 3.4957387447357178, "perplexity": 32.974638798192174, "lr": 0.0026291804804649314, "grad_norm": 0.213958, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:13.269419+00:00", "epoch": 0, "step": 9596, "train_loss": 3.5643558502197266, "perplexity": 35.31669683324132, "lr": 0.0026291804804649314, "grad_norm": 0.188498, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:13.572922+00:00", "epoch": 0, "step": 9597, "train_loss": 3.607452869415283, "perplexity": 36.87201526490335, "lr": 0.0026291804804649314, "grad_norm": 0.157169, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:13.877021+00:00", "epoch": 0, "step": 9598, "train_loss": 3.5482614040374756, "perplexity": 34.75284377928146, "lr": 0.0026291804804649314, "grad_norm": 0.162734, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:14.182863+00:00", "epoch": 0, "step": 9599, "train_loss": 3.5019776821136475, "perplexity": 33.18100859944706, "lr": 0.0026291804804649314, "grad_norm": 0.193502, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:14.488538+00:00", "epoch": 0, "step": 9600, "train_loss": 3.5687029361724854, "perplexity": 35.47055572656867, "lr": 0.0026291804804649314, "grad_norm": 0.182659, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:14.794101+00:00", "epoch": 0, "step": 9601, "train_loss": 3.573136329650879, "perplexity": 35.62815975920517, "lr": 0.0026291804804649314, "grad_norm": 0.200156, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:15.099003+00:00", "epoch": 0, "step": 9602, "train_loss": 3.546311378479004, "perplexity": 34.685140878329825, "lr": 0.0026291804804649314, "grad_norm": 0.17081, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:15.403534+00:00", "epoch": 0, "step": 9603, "train_loss": 3.6073079109191895, "perplexity": 36.86667074039908, "lr": 0.0026291804804649314, "grad_norm": 0.197708, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:15.708564+00:00", "epoch": 0, "step": 9604, "train_loss": 3.546698808670044, "perplexity": 34.69858155257994, "lr": 0.0026291804804649314, "grad_norm": 0.179081, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:16.013536+00:00", "epoch": 0, "step": 9605, "train_loss": 3.5902035236358643, "perplexity": 36.241451167843714, "lr": 0.0026291804804649314, "grad_norm": 0.161485, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:16.318762+00:00", "epoch": 0, "step": 9606, "train_loss": 3.598484992980957, "perplexity": 36.54282984138918, "lr": 0.0026291804804649314, "grad_norm": 0.173177, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:16.623049+00:00", "epoch": 0, "step": 9607, "train_loss": 3.5024008750915527, "perplexity": 33.195053540936854, "lr": 0.0026291804804649314, "grad_norm": 0.171791, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:16.927037+00:00", "epoch": 0, "step": 9608, "train_loss": 3.5290653705596924, "perplexity": 34.09208924955139, "lr": 0.0026291804804649314, "grad_norm": 0.145718, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:17.230750+00:00", "epoch": 0, "step": 9609, "train_loss": 3.4774136543273926, "perplexity": 32.37587848512998, "lr": 0.0026291804804649314, "grad_norm": 0.170571, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:17.536093+00:00", "epoch": 0, "step": 9610, "train_loss": 3.576280117034912, "perplexity": 35.74034336668115, "lr": 0.0026291804804649314, "grad_norm": 0.1669, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:17.840615+00:00", "epoch": 0, "step": 9611, "train_loss": 3.5990262031555176, "perplexity": 36.562612545524075, "lr": 0.0026291804804649314, "grad_norm": 0.159568, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:18.146285+00:00", "epoch": 0, "step": 9612, "train_loss": 3.5106918811798096, "perplexity": 33.47141801860419, "lr": 0.0026291804804649314, "grad_norm": 0.165909, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:18.450933+00:00", "epoch": 0, "step": 9613, "train_loss": 3.5943055152893066, "perplexity": 36.39041862078585, "lr": 0.0026291804804649314, "grad_norm": 0.141435, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:18.755327+00:00", "epoch": 0, "step": 9614, "train_loss": 3.5527970790863037, "perplexity": 34.91082940056272, "lr": 0.0026291804804649314, "grad_norm": 0.179526, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:19.059692+00:00", "epoch": 0, "step": 9615, "train_loss": 3.6682016849517822, "perplexity": 39.181381998391565, "lr": 0.0026291804804649314, "grad_norm": 0.143805, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:19.364538+00:00", "epoch": 0, "step": 9616, "train_loss": 3.6120543479919434, "perplexity": 37.04207200939697, "lr": 0.0026291804804649314, "grad_norm": 0.160904, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:19.669879+00:00", "epoch": 0, "step": 9617, "train_loss": 3.608041763305664, "perplexity": 36.89373536420806, "lr": 0.0026291804804649314, "grad_norm": 0.149642, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:19.974650+00:00", "epoch": 0, "step": 9618, "train_loss": 3.5607824325561523, "perplexity": 35.190720741606164, "lr": 0.0026291804804649314, "grad_norm": 0.172299, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:20.280623+00:00", "epoch": 0, "step": 9619, "train_loss": 3.46785044670105, "perplexity": 32.06773699602893, "lr": 0.0026291804804649314, "grad_norm": 0.152559, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:20.584892+00:00", "epoch": 0, "step": 9620, "train_loss": 3.4995927810668945, "perplexity": 33.101969465026045, "lr": 0.0026291804804649314, "grad_norm": 0.166, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:20.890117+00:00", "epoch": 0, "step": 9621, "train_loss": 3.5699660778045654, "perplexity": 35.515388371248925, "lr": 0.0026291804804649314, "grad_norm": 0.177534, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:21.195067+00:00", "epoch": 0, "step": 9622, "train_loss": 3.4878811836242676, "perplexity": 32.716553845723325, "lr": 0.0026291804804649314, "grad_norm": 0.165992, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:21.499727+00:00", "epoch": 0, "step": 9623, "train_loss": 3.540731191635132, "perplexity": 34.492130329822224, "lr": 0.0026291804804649314, "grad_norm": 0.175749, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:21.803981+00:00", "epoch": 0, "step": 9624, "train_loss": 3.497154712677002, "perplexity": 33.0213629017155, "lr": 0.0026291804804649314, "grad_norm": 0.154257, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:22.109199+00:00", "epoch": 0, "step": 9625, "train_loss": 3.6152291297912598, "perplexity": 37.159859381089376, "lr": 0.0026291804804649314, "grad_norm": 0.171621, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:22.413537+00:00", "epoch": 0, "step": 9626, "train_loss": 3.59831166267395, "perplexity": 36.5364964103775, "lr": 0.0026291804804649314, "grad_norm": 0.163921, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:22.718013+00:00", "epoch": 0, "step": 9627, "train_loss": 3.5425422191619873, "perplexity": 34.55465312547278, "lr": 0.0026291804804649314, "grad_norm": 0.186527, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:23.023367+00:00", "epoch": 0, "step": 9628, "train_loss": 3.488345146179199, "perplexity": 32.73173662348197, "lr": 0.0026291804804649314, "grad_norm": 0.164834, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:23.328536+00:00", "epoch": 0, "step": 9629, "train_loss": 3.5912108421325684, "perplexity": 36.27797624505695, "lr": 0.0026291804804649314, "grad_norm": 0.154508, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:23.633290+00:00", "epoch": 0, "step": 9630, "train_loss": 3.5018601417541504, "perplexity": 33.17710872096885, "lr": 0.0026291804804649314, "grad_norm": 0.151288, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:23.938110+00:00", "epoch": 0, "step": 9631, "train_loss": 3.4404218196868896, "perplexity": 31.200116216205778, "lr": 0.0026291804804649314, "grad_norm": 0.140184, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:24.243200+00:00", "epoch": 0, "step": 9632, "train_loss": 3.519976854324341, "perplexity": 33.78364650947547, "lr": 0.0026291804804649314, "grad_norm": 0.145538, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:24.548828+00:00", "epoch": 0, "step": 9633, "train_loss": 3.5385701656341553, "perplexity": 34.41767242106555, "lr": 0.0026291804804649314, "grad_norm": 0.149507, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:24.854094+00:00", "epoch": 0, "step": 9634, "train_loss": 3.4945497512817383, "perplexity": 32.935455467491785, "lr": 0.0026291804804649314, "grad_norm": 0.159351, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:25.160008+00:00", "epoch": 0, "step": 9635, "train_loss": 3.6176300048828125, "perplexity": 37.249182746097624, "lr": 0.0026291804804649314, "grad_norm": 0.215922, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:25.465114+00:00", "epoch": 0, "step": 9636, "train_loss": 3.6139674186706543, "perplexity": 37.113003938500434, "lr": 0.0026291804804649314, "grad_norm": 0.174933, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:25.771093+00:00", "epoch": 0, "step": 9637, "train_loss": 3.6353936195373535, "perplexity": 37.916774743516825, "lr": 0.0026291804804649314, "grad_norm": 0.156597, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:26.076872+00:00", "epoch": 0, "step": 9638, "train_loss": 3.623337984085083, "perplexity": 37.46240827101952, "lr": 0.0026291804804649314, "grad_norm": 0.160718, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:26.383241+00:00", "epoch": 0, "step": 9639, "train_loss": 3.592280626296997, "perplexity": 36.31680661591611, "lr": 0.0026291804804649314, "grad_norm": 0.164562, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:26.687334+00:00", "epoch": 0, "step": 9640, "train_loss": 3.5899693965911865, "perplexity": 36.232967057205585, "lr": 0.0026291804804649314, "grad_norm": 0.170199, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:26.990855+00:00", "epoch": 0, "step": 9641, "train_loss": 3.4366202354431152, "perplexity": 31.081731513280022, "lr": 0.0026291804804649314, "grad_norm": 0.175503, "tokens_per_sec": 107904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:27.296327+00:00", "epoch": 0, "step": 9642, "train_loss": 3.5380454063415527, "perplexity": 34.399616165641106, "lr": 0.0026291804804649314, "grad_norm": 0.152885, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:27.602112+00:00", "epoch": 0, "step": 9643, "train_loss": 3.511477470397949, "perplexity": 33.497723134870455, "lr": 0.0026291804804649314, "grad_norm": 0.160601, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:27.906944+00:00", "epoch": 0, "step": 9644, "train_loss": 3.482866048812866, "perplexity": 32.55288666714543, "lr": 0.0026291804804649314, "grad_norm": 0.175306, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:28.213055+00:00", "epoch": 0, "step": 9645, "train_loss": 3.541330099105835, "perplexity": 34.512794111586196, "lr": 0.0026291804804649314, "grad_norm": 0.169526, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:28.518378+00:00", "epoch": 0, "step": 9646, "train_loss": 3.4750161170959473, "perplexity": 32.298349088006965, "lr": 0.0026291804804649314, "grad_norm": 0.167995, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:28.822842+00:00", "epoch": 0, "step": 9647, "train_loss": 3.5450596809387207, "perplexity": 34.64175273289036, "lr": 0.0026291804804649314, "grad_norm": 0.178398, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:29.127164+00:00", "epoch": 0, "step": 9648, "train_loss": 3.548506259918213, "perplexity": 34.761354259331185, "lr": 0.0026291804804649314, "grad_norm": 0.189475, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:29.432557+00:00", "epoch": 0, "step": 9649, "train_loss": 3.511704683303833, "perplexity": 33.505335114621694, "lr": 0.0026291804804649314, "grad_norm": 0.162396, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:29.739618+00:00", "epoch": 0, "step": 9650, "train_loss": 3.644986152648926, "perplexity": 38.28224274023253, "lr": 0.0026291804804649314, "grad_norm": 0.141384, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:30.043738+00:00", "epoch": 0, "step": 9651, "train_loss": 3.4791696071624756, "perplexity": 32.43277894338431, "lr": 0.0026291804804649314, "grad_norm": 0.150538, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:30.348871+00:00", "epoch": 0, "step": 9652, "train_loss": 3.5057969093322754, "perplexity": 33.307976716340605, "lr": 0.0026291804804649314, "grad_norm": 0.155176, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:30.654347+00:00", "epoch": 0, "step": 9653, "train_loss": 3.551999807357788, "perplexity": 34.88300707571289, "lr": 0.0026291804804649314, "grad_norm": 0.164247, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:30.958690+00:00", "epoch": 0, "step": 9654, "train_loss": 3.5409016609191895, "perplexity": 34.49801067978046, "lr": 0.0026291804804649314, "grad_norm": 0.150542, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:31.264277+00:00", "epoch": 0, "step": 9655, "train_loss": 3.5323808193206787, "perplexity": 34.205307405400816, "lr": 0.0026291804804649314, "grad_norm": 0.168307, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:31.568756+00:00", "epoch": 0, "step": 9656, "train_loss": 3.6108310222625732, "perplexity": 36.99678519554872, "lr": 0.0026291804804649314, "grad_norm": 0.158719, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:31.873033+00:00", "epoch": 0, "step": 9657, "train_loss": 3.5496761798858643, "perplexity": 34.80204606021569, "lr": 0.0026291804804649314, "grad_norm": 0.151254, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:32.177377+00:00", "epoch": 0, "step": 9658, "train_loss": 3.5843796730041504, "perplexity": 36.03099978310772, "lr": 0.0026291804804649314, "grad_norm": 0.159927, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:32.482491+00:00", "epoch": 0, "step": 9659, "train_loss": 3.5172905921936035, "perplexity": 33.693016561616105, "lr": 0.0026291804804649314, "grad_norm": 0.148412, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:32.787965+00:00", "epoch": 0, "step": 9660, "train_loss": 3.544055700302124, "perplexity": 34.606990537134436, "lr": 0.0026291804804649314, "grad_norm": 0.153024, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:33.093057+00:00", "epoch": 0, "step": 9661, "train_loss": 3.6185781955718994, "perplexity": 37.28451882437632, "lr": 0.0026291804804649314, "grad_norm": 0.148494, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:33.397416+00:00", "epoch": 0, "step": 9662, "train_loss": 3.582703113555908, "perplexity": 35.97064228060029, "lr": 0.0026291804804649314, "grad_norm": 0.142999, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:33.701618+00:00", "epoch": 0, "step": 9663, "train_loss": 3.558826208114624, "perplexity": 35.12194708386354, "lr": 0.0026291804804649314, "grad_norm": 0.160946, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:34.005811+00:00", "epoch": 0, "step": 9664, "train_loss": 3.538501262664795, "perplexity": 34.41530102293639, "lr": 0.0026291804804649314, "grad_norm": 0.164281, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:34.311952+00:00", "epoch": 0, "step": 9665, "train_loss": 3.5520241260528564, "perplexity": 34.88385539524, "lr": 0.0026291804804649314, "grad_norm": 0.175726, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:34.616791+00:00", "epoch": 0, "step": 9666, "train_loss": 3.634235382080078, "perplexity": 37.872883537875744, "lr": 0.0026291804804649314, "grad_norm": 0.184895, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:34.922191+00:00", "epoch": 0, "step": 9667, "train_loss": 3.498887777328491, "perplexity": 33.07864067721226, "lr": 0.0026291804804649314, "grad_norm": 0.181499, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:35.226822+00:00", "epoch": 0, "step": 9668, "train_loss": 3.474801778793335, "perplexity": 32.291427056541, "lr": 0.0026291804804649314, "grad_norm": 0.165415, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:35.532096+00:00", "epoch": 0, "step": 9669, "train_loss": 3.510366439819336, "perplexity": 33.4605268071087, "lr": 0.0026291804804649314, "grad_norm": 0.154998, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:35.837693+00:00", "epoch": 0, "step": 9670, "train_loss": 3.5461177825927734, "perplexity": 34.67842662768908, "lr": 0.0026291804804649314, "grad_norm": 0.174154, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:36.143201+00:00", "epoch": 0, "step": 9671, "train_loss": 3.497875213623047, "perplexity": 33.04516339803632, "lr": 0.0026291804804649314, "grad_norm": 0.168901, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:36.448178+00:00", "epoch": 0, "step": 9672, "train_loss": 3.565701961517334, "perplexity": 35.36426904940807, "lr": 0.0026291804804649314, "grad_norm": 0.17522, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:36.752434+00:00", "epoch": 0, "step": 9673, "train_loss": 3.588585138320923, "perplexity": 36.18284597116777, "lr": 0.0026291804804649314, "grad_norm": 0.154315, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:37.057495+00:00", "epoch": 0, "step": 9674, "train_loss": 3.658050298690796, "perplexity": 38.785648674774116, "lr": 0.0026291804804649314, "grad_norm": 0.162137, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:37.362009+00:00", "epoch": 0, "step": 9675, "train_loss": 3.5321478843688965, "perplexity": 34.197340721665206, "lr": 0.0026291804804649314, "grad_norm": 0.164315, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:37.667473+00:00", "epoch": 0, "step": 9676, "train_loss": 3.553337335586548, "perplexity": 34.9296952988285, "lr": 0.0026291804804649314, "grad_norm": 0.154944, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:37.973417+00:00", "epoch": 0, "step": 9677, "train_loss": 3.5267999172210693, "perplexity": 34.01494263116728, "lr": 0.0026291804804649314, "grad_norm": 0.161821, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:38.277916+00:00", "epoch": 0, "step": 9678, "train_loss": 3.501868486404419, "perplexity": 33.17738557349317, "lr": 0.0026291804804649314, "grad_norm": 0.163913, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:38.582926+00:00", "epoch": 0, "step": 9679, "train_loss": 3.4888079166412354, "perplexity": 32.74688740976026, "lr": 0.0026291804804649314, "grad_norm": 0.174065, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:38.889134+00:00", "epoch": 0, "step": 9680, "train_loss": 3.4960100650787354, "perplexity": 32.9835867023211, "lr": 0.0026291804804649314, "grad_norm": 0.194424, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:39.194815+00:00", "epoch": 0, "step": 9681, "train_loss": 3.5984103679656982, "perplexity": 36.54010293390369, "lr": 0.0026291804804649314, "grad_norm": 0.16937, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:39.500563+00:00", "epoch": 0, "step": 9682, "train_loss": 3.5188822746276855, "perplexity": 33.74668784671818, "lr": 0.0026291804804649314, "grad_norm": 0.16126, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:39.805572+00:00", "epoch": 0, "step": 9683, "train_loss": 3.484102487564087, "perplexity": 32.593161211057264, "lr": 0.0026291804804649314, "grad_norm": 0.186074, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:40.110810+00:00", "epoch": 0, "step": 9684, "train_loss": 3.5670156478881836, "perplexity": 35.41075713637081, "lr": 0.0026291804804649314, "grad_norm": 0.193795, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:40.415181+00:00", "epoch": 0, "step": 9685, "train_loss": 3.5752665996551514, "perplexity": 35.70413825787312, "lr": 0.0026291804804649314, "grad_norm": 0.164837, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:40.720526+00:00", "epoch": 0, "step": 9686, "train_loss": 3.513896942138672, "perplexity": 33.57886805370871, "lr": 0.0026291804804649314, "grad_norm": 0.218648, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:41.025242+00:00", "epoch": 0, "step": 9687, "train_loss": 3.4508914947509766, "perplexity": 31.528487264536956, "lr": 0.0026291804804649314, "grad_norm": 0.175608, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:41.329544+00:00", "epoch": 0, "step": 9688, "train_loss": 3.5604000091552734, "perplexity": 35.17726555945302, "lr": 0.0026291804804649314, "grad_norm": 0.16031, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:41.634608+00:00", "epoch": 0, "step": 9689, "train_loss": 3.66499924659729, "perplexity": 39.056106738111616, "lr": 0.0026291804804649314, "grad_norm": 0.136675, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:41.939161+00:00", "epoch": 0, "step": 9690, "train_loss": 3.4351165294647217, "perplexity": 31.035028850121652, "lr": 0.0026291804804649314, "grad_norm": 0.156949, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:42.244877+00:00", "epoch": 0, "step": 9691, "train_loss": 3.61961030960083, "perplexity": 37.32302056499234, "lr": 0.0026291804804649314, "grad_norm": 0.153487, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:42.550496+00:00", "epoch": 0, "step": 9692, "train_loss": 3.5780768394470215, "perplexity": 35.80461656588268, "lr": 0.0026291804804649314, "grad_norm": 0.146365, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:42.854476+00:00", "epoch": 0, "step": 9693, "train_loss": 3.610522747039795, "perplexity": 36.98538176113893, "lr": 0.0026291804804649314, "grad_norm": 0.170302, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:43.158613+00:00", "epoch": 0, "step": 9694, "train_loss": 3.4459502696990967, "perplexity": 31.373082175371042, "lr": 0.0026291804804649314, "grad_norm": 0.16834, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:43.463093+00:00", "epoch": 0, "step": 9695, "train_loss": 3.532139778137207, "perplexity": 34.19706351122172, "lr": 0.0026291804804649314, "grad_norm": 0.151526, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:43.768144+00:00", "epoch": 0, "step": 9696, "train_loss": 3.549825668334961, "perplexity": 34.807248952983116, "lr": 0.0026291804804649314, "grad_norm": 0.150182, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:44.072841+00:00", "epoch": 0, "step": 9697, "train_loss": 3.5812339782714844, "perplexity": 35.91783534058991, "lr": 0.0026291804804649314, "grad_norm": 0.145692, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:44.378352+00:00", "epoch": 0, "step": 9698, "train_loss": 3.5782225131988525, "perplexity": 35.80983273863121, "lr": 0.0026291804804649314, "grad_norm": 0.152889, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:44.683628+00:00", "epoch": 0, "step": 9699, "train_loss": 3.5090060234069824, "perplexity": 33.415037506485604, "lr": 0.0026291804804649314, "grad_norm": 0.178636, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:44.989016+00:00", "epoch": 0, "step": 9700, "train_loss": 3.489898681640625, "perplexity": 32.78262605607465, "lr": 0.0026291804804649314, "grad_norm": 0.186673, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:45.294216+00:00", "epoch": 0, "step": 9701, "train_loss": 3.5379505157470703, "perplexity": 34.39635212047922, "lr": 0.0026291804804649314, "grad_norm": 0.18393, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:45.598923+00:00", "epoch": 0, "step": 9702, "train_loss": 3.509533405303955, "perplexity": 33.4326646400602, "lr": 0.0026291804804649314, "grad_norm": 0.16404, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:45.904643+00:00", "epoch": 0, "step": 9703, "train_loss": 3.5488290786743164, "perplexity": 34.77257768794282, "lr": 0.0026291804804649314, "grad_norm": 0.162836, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:46.210195+00:00", "epoch": 0, "step": 9704, "train_loss": 3.519308090209961, "perplexity": 33.76106077214413, "lr": 0.0026291804804649314, "grad_norm": 0.160978, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:46.514038+00:00", "epoch": 0, "step": 9705, "train_loss": 3.5255866050720215, "perplexity": 33.97369691504786, "lr": 0.0026291804804649314, "grad_norm": 0.172985, "tokens_per_sec": 107846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:46.819508+00:00", "epoch": 0, "step": 9706, "train_loss": 3.5212278366088867, "perplexity": 33.82593569876443, "lr": 0.0026291804804649314, "grad_norm": 0.169659, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:47.124000+00:00", "epoch": 0, "step": 9707, "train_loss": 3.621569871902466, "perplexity": 37.39622905394622, "lr": 0.0026291804804649314, "grad_norm": 0.150384, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:47.428272+00:00", "epoch": 0, "step": 9708, "train_loss": 3.626648187637329, "perplexity": 37.586621940797365, "lr": 0.0026291804804649314, "grad_norm": 0.152229, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:47.732945+00:00", "epoch": 0, "step": 9709, "train_loss": 3.4938063621520996, "perplexity": 32.910980706180425, "lr": 0.0026291804804649314, "grad_norm": 0.162823, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:48.038951+00:00", "epoch": 0, "step": 9710, "train_loss": 3.5432684421539307, "perplexity": 34.579756623316484, "lr": 0.0026291804804649314, "grad_norm": 0.188699, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:48.344180+00:00", "epoch": 0, "step": 9711, "train_loss": 3.4739108085632324, "perplexity": 32.26266916945913, "lr": 0.0026291804804649314, "grad_norm": 0.188696, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:48.649342+00:00", "epoch": 0, "step": 9712, "train_loss": 3.602989912033081, "perplexity": 36.707823694700735, "lr": 0.0026291804804649314, "grad_norm": 0.165028, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:48.954390+00:00", "epoch": 0, "step": 9713, "train_loss": 3.489780902862549, "perplexity": 32.778765185804396, "lr": 0.0026291804804649314, "grad_norm": 0.16986, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:49.259008+00:00", "epoch": 0, "step": 9714, "train_loss": 3.535047769546509, "perplexity": 34.296653010659284, "lr": 0.0026291804804649314, "grad_norm": 0.171024, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:49.564801+00:00", "epoch": 0, "step": 9715, "train_loss": 3.5083630084991455, "perplexity": 33.39355804576982, "lr": 0.0026291804804649314, "grad_norm": 0.178334, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:49.870541+00:00", "epoch": 0, "step": 9716, "train_loss": 3.589339017868042, "perplexity": 36.21013376326379, "lr": 0.0026291804804649314, "grad_norm": 0.157191, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:50.175848+00:00", "epoch": 0, "step": 9717, "train_loss": 3.5418460369110107, "perplexity": 34.53060516113242, "lr": 0.0026291804804649314, "grad_norm": 0.182158, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:50.480838+00:00", "epoch": 0, "step": 9718, "train_loss": 3.57922101020813, "perplexity": 35.84560660660134, "lr": 0.0026291804804649314, "grad_norm": 0.14631, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:50.785083+00:00", "epoch": 0, "step": 9719, "train_loss": 3.5916149616241455, "perplexity": 36.29263984509628, "lr": 0.0026291804804649314, "grad_norm": 0.169372, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:51.091636+00:00", "epoch": 0, "step": 9720, "train_loss": 3.51711106300354, "perplexity": 33.68696822458477, "lr": 0.0026291804804649314, "grad_norm": 0.173863, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:51.397409+00:00", "epoch": 0, "step": 9721, "train_loss": 3.542276382446289, "perplexity": 34.54546845083977, "lr": 0.0026291804804649314, "grad_norm": 0.172398, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:51.703472+00:00", "epoch": 0, "step": 9722, "train_loss": 3.5600805282592773, "perplexity": 35.166028890179554, "lr": 0.0026291804804649314, "grad_norm": 0.176538, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:52.008973+00:00", "epoch": 0, "step": 9723, "train_loss": 3.549607992172241, "perplexity": 34.79967306917077, "lr": 0.0026291804804649314, "grad_norm": 0.13952, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:52.314040+00:00", "epoch": 0, "step": 9724, "train_loss": 3.5867574214935303, "perplexity": 36.11677437320197, "lr": 0.0026291804804649314, "grad_norm": 0.167309, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:52.621680+00:00", "epoch": 0, "step": 9725, "train_loss": 3.610097646713257, "perplexity": 36.969662604621135, "lr": 0.0026291804804649314, "grad_norm": 0.156994, "tokens_per_sec": 106514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:52.927622+00:00", "epoch": 0, "step": 9726, "train_loss": 3.5421829223632812, "perplexity": 34.54223997935977, "lr": 0.0026291804804649314, "grad_norm": 0.161699, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:53.232169+00:00", "epoch": 0, "step": 9727, "train_loss": 3.616609811782837, "perplexity": 37.211200764652304, "lr": 0.0026291804804649314, "grad_norm": 0.180118, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:53.536332+00:00", "epoch": 0, "step": 9728, "train_loss": 3.5740416049957275, "perplexity": 35.660427657280096, "lr": 0.0026291804804649314, "grad_norm": 0.180284, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:53.841361+00:00", "epoch": 0, "step": 9729, "train_loss": 3.4715473651885986, "perplexity": 32.186508214267114, "lr": 0.0026291804804649314, "grad_norm": 0.14463, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:54.146947+00:00", "epoch": 0, "step": 9730, "train_loss": 3.524085283279419, "perplexity": 33.92272973213842, "lr": 0.0026291804804649314, "grad_norm": 0.16263, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:54.451165+00:00", "epoch": 0, "step": 9731, "train_loss": 3.6138386726379395, "perplexity": 37.10822609405208, "lr": 0.0026291804804649314, "grad_norm": 0.194064, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:54.755628+00:00", "epoch": 0, "step": 9732, "train_loss": 3.4572181701660156, "perplexity": 31.728590095160115, "lr": 0.0026291804804649314, "grad_norm": 0.174008, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:55.060316+00:00", "epoch": 0, "step": 9733, "train_loss": 3.6636605262756348, "perplexity": 39.003856516356855, "lr": 0.0026291804804649314, "grad_norm": 0.165522, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:55.364705+00:00", "epoch": 0, "step": 9734, "train_loss": 3.651801109313965, "perplexity": 38.544025571695705, "lr": 0.0026291804804649314, "grad_norm": 0.157812, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:55.670076+00:00", "epoch": 0, "step": 9735, "train_loss": 3.576382637023926, "perplexity": 35.74400765411859, "lr": 0.0026291804804649314, "grad_norm": 0.157287, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:55.976328+00:00", "epoch": 0, "step": 9736, "train_loss": 3.560199499130249, "perplexity": 35.17021287214629, "lr": 0.0026291804804649314, "grad_norm": 0.155099, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:56.280683+00:00", "epoch": 0, "step": 9737, "train_loss": 3.5346925258636475, "perplexity": 34.28447150516351, "lr": 0.0026291804804649314, "grad_norm": 0.181169, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:56.585569+00:00", "epoch": 0, "step": 9738, "train_loss": 3.4857735633850098, "perplexity": 32.647672388154334, "lr": 0.0026291804804649314, "grad_norm": 0.151716, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:56.891077+00:00", "epoch": 0, "step": 9739, "train_loss": 3.63557767868042, "perplexity": 37.92375431489122, "lr": 0.0026291804804649314, "grad_norm": 0.164762, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:57.196015+00:00", "epoch": 0, "step": 9740, "train_loss": 3.5141353607177734, "perplexity": 33.58687483416259, "lr": 0.0026291804804649314, "grad_norm": 0.149539, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:57.501498+00:00", "epoch": 0, "step": 9741, "train_loss": 3.5654098987579346, "perplexity": 35.353941971556196, "lr": 0.0026291804804649314, "grad_norm": 0.141677, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:57.807025+00:00", "epoch": 0, "step": 9742, "train_loss": 3.589364767074585, "perplexity": 36.21106615748116, "lr": 0.0026291804804649314, "grad_norm": 0.187713, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:58.111657+00:00", "epoch": 0, "step": 9743, "train_loss": 3.5609235763549805, "perplexity": 35.195688044158864, "lr": 0.0026291804804649314, "grad_norm": 0.181557, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:58.415268+00:00", "epoch": 0, "step": 9744, "train_loss": 3.584036350250244, "perplexity": 36.018631644289414, "lr": 0.0026291804804649314, "grad_norm": 0.161037, "tokens_per_sec": 107927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:58.719115+00:00", "epoch": 0, "step": 9745, "train_loss": 3.4995055198669434, "perplexity": 33.09908107347387, "lr": 0.0026291804804649314, "grad_norm": 0.158858, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:59.025240+00:00", "epoch": 0, "step": 9746, "train_loss": 3.5084428787231445, "perplexity": 33.396225303246865, "lr": 0.0026291804804649314, "grad_norm": 0.153103, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:59.330593+00:00", "epoch": 0, "step": 9747, "train_loss": 3.5199434757232666, "perplexity": 33.78251887743529, "lr": 0.0026291804804649314, "grad_norm": 0.151947, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:59.636071+00:00", "epoch": 0, "step": 9748, "train_loss": 3.491541624069214, "perplexity": 32.83653029199575, "lr": 0.0026291804804649314, "grad_norm": 0.138167, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:48:59.941030+00:00", "epoch": 0, "step": 9749, "train_loss": 3.497798204421997, "perplexity": 33.042618714387764, "lr": 0.0026291804804649314, "grad_norm": 0.158368, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:00.246611+00:00", "epoch": 0, "step": 9750, "train_loss": 3.4819583892822266, "perplexity": 32.523353134535924, "lr": 0.0026291804804649314, "grad_norm": 0.158561, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:00.551657+00:00", "epoch": 0, "step": 9751, "train_loss": 3.5302536487579346, "perplexity": 34.132624214577135, "lr": 0.0026291804804649314, "grad_norm": 0.163178, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:00.857019+00:00", "epoch": 0, "step": 9752, "train_loss": 3.5776379108428955, "perplexity": 35.78890434403621, "lr": 0.0026291804804649314, "grad_norm": 0.175008, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:01.162942+00:00", "epoch": 0, "step": 9753, "train_loss": 3.4764621257781982, "perplexity": 32.3450865644652, "lr": 0.0026291804804649314, "grad_norm": 0.174541, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:01.467111+00:00", "epoch": 0, "step": 9754, "train_loss": 3.5984835624694824, "perplexity": 36.542777566489164, "lr": 0.0026291804804649314, "grad_norm": 0.149542, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:01.771859+00:00", "epoch": 0, "step": 9755, "train_loss": 3.446578025817871, "perplexity": 31.392783002681917, "lr": 0.0026291804804649314, "grad_norm": 0.161014, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:02.077499+00:00", "epoch": 0, "step": 9756, "train_loss": 3.5578415393829346, "perplexity": 35.08738062182407, "lr": 0.0026291804804649314, "grad_norm": 0.177781, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:02.383188+00:00", "epoch": 0, "step": 9757, "train_loss": 3.4951651096343994, "perplexity": 32.955728812158654, "lr": 0.0026291804804649314, "grad_norm": 0.164397, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:02.688881+00:00", "epoch": 0, "step": 9758, "train_loss": 3.476726770401001, "perplexity": 32.353647650471295, "lr": 0.0026291804804649314, "grad_norm": 0.150143, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:02.993946+00:00", "epoch": 0, "step": 9759, "train_loss": 3.566753625869751, "perplexity": 35.401479953777844, "lr": 0.0026291804804649314, "grad_norm": 0.167713, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:03.299421+00:00", "epoch": 0, "step": 9760, "train_loss": 3.5591049194335938, "perplexity": 35.131737332323226, "lr": 0.0026291804804649314, "grad_norm": 0.14698, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:03.603846+00:00", "epoch": 0, "step": 9761, "train_loss": 3.511756420135498, "perplexity": 33.50706861934703, "lr": 0.0026291804804649314, "grad_norm": 0.153444, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:03.907906+00:00", "epoch": 0, "step": 9762, "train_loss": 3.4991750717163086, "perplexity": 33.08814535028969, "lr": 0.0026291804804649314, "grad_norm": 0.186162, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:04.212460+00:00", "epoch": 0, "step": 9763, "train_loss": 3.5861597061157227, "perplexity": 36.09519327208282, "lr": 0.0026291804804649314, "grad_norm": 0.158589, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:04.517860+00:00", "epoch": 0, "step": 9764, "train_loss": 3.49531888961792, "perplexity": 32.960797133285475, "lr": 0.0026291804804649314, "grad_norm": 0.155362, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:04.822915+00:00", "epoch": 0, "step": 9765, "train_loss": 3.5894927978515625, "perplexity": 36.215702585212874, "lr": 0.0026291804804649314, "grad_norm": 0.155856, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:05.128867+00:00", "epoch": 0, "step": 9766, "train_loss": 3.551445484161377, "perplexity": 34.86367597406427, "lr": 0.0026291804804649314, "grad_norm": 0.197874, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:05.433907+00:00", "epoch": 0, "step": 9767, "train_loss": 3.586538791656494, "perplexity": 36.108879031816535, "lr": 0.0026291804804649314, "grad_norm": 0.180107, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:05.739508+00:00", "epoch": 0, "step": 9768, "train_loss": 3.3858985900878906, "perplexity": 29.544529212751456, "lr": 0.0026291804804649314, "grad_norm": 0.172562, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:06.045524+00:00", "epoch": 0, "step": 9769, "train_loss": 3.5799341201782227, "perplexity": 35.87117758243181, "lr": 0.0026291804804649314, "grad_norm": 0.156556, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:06.351261+00:00", "epoch": 0, "step": 9770, "train_loss": 3.5201754570007324, "perplexity": 33.7903566983992, "lr": 0.0026291804804649314, "grad_norm": 0.167278, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:06.655780+00:00", "epoch": 0, "step": 9771, "train_loss": 3.527740240097046, "perplexity": 34.046942702690146, "lr": 0.0026291804804649314, "grad_norm": 0.16955, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:06.960850+00:00", "epoch": 0, "step": 9772, "train_loss": 3.57358455657959, "perplexity": 35.644132839344564, "lr": 0.0026291804804649314, "grad_norm": 0.157521, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:07.266135+00:00", "epoch": 0, "step": 9773, "train_loss": 3.5364277362823486, "perplexity": 34.34401392168972, "lr": 0.0026291804804649314, "grad_norm": 0.160137, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:07.571714+00:00", "epoch": 0, "step": 9774, "train_loss": 3.578329086303711, "perplexity": 35.81364930705879, "lr": 0.0026291804804649314, "grad_norm": 0.147561, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:07.876313+00:00", "epoch": 0, "step": 9775, "train_loss": 3.6100215911865234, "perplexity": 36.966850964380335, "lr": 0.0026291804804649314, "grad_norm": 0.150178, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:08.182140+00:00", "epoch": 0, "step": 9776, "train_loss": 3.504772901535034, "perplexity": 33.27388654575188, "lr": 0.0026291804804649314, "grad_norm": 0.141881, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:08.486711+00:00", "epoch": 0, "step": 9777, "train_loss": 3.6148316860198975, "perplexity": 37.1450933609594, "lr": 0.0026291804804649314, "grad_norm": 0.151893, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:08.792123+00:00", "epoch": 0, "step": 9778, "train_loss": 3.5695371627807617, "perplexity": 35.50015855398247, "lr": 0.0026291804804649314, "grad_norm": 0.15769, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:09.098374+00:00", "epoch": 0, "step": 9779, "train_loss": 3.5138678550720215, "perplexity": 33.577891357140274, "lr": 0.0026291804804649314, "grad_norm": 0.173102, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:09.404788+00:00", "epoch": 0, "step": 9780, "train_loss": 3.5681025981903076, "perplexity": 35.449267795333256, "lr": 0.0026291804804649314, "grad_norm": 0.180208, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:09.709436+00:00", "epoch": 0, "step": 9781, "train_loss": 3.58040189743042, "perplexity": 35.88796122851228, "lr": 0.0026291804804649314, "grad_norm": 0.184274, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:10.014557+00:00", "epoch": 0, "step": 9782, "train_loss": 3.5615572929382324, "perplexity": 35.217999204060746, "lr": 0.0026291804804649314, "grad_norm": 0.163493, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:10.319966+00:00", "epoch": 0, "step": 9783, "train_loss": 3.462217330932617, "perplexity": 31.887603553698362, "lr": 0.0026291804804649314, "grad_norm": 0.176529, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:10.625588+00:00", "epoch": 0, "step": 9784, "train_loss": 3.435248613357544, "perplexity": 31.039128348278904, "lr": 0.0026291804804649314, "grad_norm": 0.152125, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:10.931315+00:00", "epoch": 0, "step": 9785, "train_loss": 3.5821468830108643, "perplexity": 35.95063987412941, "lr": 0.0026291804804649314, "grad_norm": 0.181078, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:11.238170+00:00", "epoch": 0, "step": 9786, "train_loss": 3.538571357727051, "perplexity": 34.417713450152775, "lr": 0.0026291804804649314, "grad_norm": 0.197726, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:11.543017+00:00", "epoch": 0, "step": 9787, "train_loss": 3.492915153503418, "perplexity": 32.8816632214769, "lr": 0.0026291804804649314, "grad_norm": 0.206071, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:11.847714+00:00", "epoch": 0, "step": 9788, "train_loss": 3.5025787353515625, "perplexity": 33.200958146872495, "lr": 0.0026291804804649314, "grad_norm": 0.185633, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:12.153289+00:00", "epoch": 0, "step": 9789, "train_loss": 3.552776336669922, "perplexity": 34.91010527311316, "lr": 0.0026291804804649314, "grad_norm": 0.190394, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:12.458298+00:00", "epoch": 0, "step": 9790, "train_loss": 3.562321901321411, "perplexity": 35.244937478793894, "lr": 0.0026291804804649314, "grad_norm": 0.155428, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:12.762303+00:00", "epoch": 0, "step": 9791, "train_loss": 3.6703667640686035, "perplexity": 39.26630468932342, "lr": 0.0026291804804649314, "grad_norm": 0.166678, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:13.066839+00:00", "epoch": 0, "step": 9792, "train_loss": 3.69563627243042, "perplexity": 40.27118788454516, "lr": 0.0026291804804649314, "grad_norm": 0.198778, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:13.371142+00:00", "epoch": 0, "step": 9793, "train_loss": 3.5249173641204834, "perplexity": 33.95096793222854, "lr": 0.0026291804804649314, "grad_norm": 0.202962, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:13.675784+00:00", "epoch": 0, "step": 9794, "train_loss": 3.5576512813568115, "perplexity": 35.0807056010534, "lr": 0.0026291804804649314, "grad_norm": 0.177383, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:13.982034+00:00", "epoch": 0, "step": 9795, "train_loss": 3.5459201335906982, "perplexity": 34.67157314858654, "lr": 0.0026291804804649314, "grad_norm": 0.185903, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:14.288774+00:00", "epoch": 0, "step": 9796, "train_loss": 3.5294477939605713, "perplexity": 34.10512935552718, "lr": 0.0026291804804649314, "grad_norm": 0.175272, "tokens_per_sec": 106830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:14.594073+00:00", "epoch": 0, "step": 9797, "train_loss": 3.5752508640289307, "perplexity": 35.70357643531929, "lr": 0.0026291804804649314, "grad_norm": 0.148413, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:14.899790+00:00", "epoch": 0, "step": 9798, "train_loss": 3.5301735401153564, "perplexity": 34.129890005902006, "lr": 0.0026291804804649314, "grad_norm": 0.175859, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:15.205441+00:00", "epoch": 0, "step": 9799, "train_loss": 3.548380136489868, "perplexity": 34.756970314622976, "lr": 0.0026291804804649314, "grad_norm": 0.154418, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:15.511506+00:00", "epoch": 0, "step": 9800, "train_loss": 3.527599811553955, "perplexity": 34.04216187581984, "lr": 0.0026291804804649314, "grad_norm": 0.147408, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:15.816127+00:00", "epoch": 0, "step": 9801, "train_loss": 3.4777400493621826, "perplexity": 32.3864475358635, "lr": 0.0026291804804649314, "grad_norm": 0.156468, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:16.121382+00:00", "epoch": 0, "step": 9802, "train_loss": 3.5312137603759766, "perplexity": 34.165411080646706, "lr": 0.0026291804804649314, "grad_norm": 0.159238, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:16.425712+00:00", "epoch": 0, "step": 9803, "train_loss": 3.511187791824341, "perplexity": 33.488020967536464, "lr": 0.0026291804804649314, "grad_norm": 0.153129, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:16.729724+00:00", "epoch": 0, "step": 9804, "train_loss": 3.44599986076355, "perplexity": 31.37463803848943, "lr": 0.0026291804804649314, "grad_norm": 0.15551, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:17.034739+00:00", "epoch": 0, "step": 9805, "train_loss": 3.614837646484375, "perplexity": 37.145314763628726, "lr": 0.0026291804804649314, "grad_norm": 0.148784, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:17.340715+00:00", "epoch": 0, "step": 9806, "train_loss": 3.4578609466552734, "perplexity": 31.748991042832937, "lr": 0.0026291804804649314, "grad_norm": 0.170198, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:17.645691+00:00", "epoch": 0, "step": 9807, "train_loss": 3.445096254348755, "perplexity": 31.346300519206228, "lr": 0.0026291804804649314, "grad_norm": 0.181391, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:17.950731+00:00", "epoch": 0, "step": 9808, "train_loss": 3.4641051292419434, "perplexity": 31.947857773788495, "lr": 0.0026291804804649314, "grad_norm": 0.142519, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:18.255582+00:00", "epoch": 0, "step": 9809, "train_loss": 3.5258917808532715, "perplexity": 33.984066446730345, "lr": 0.0026291804804649314, "grad_norm": 0.180338, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:18.560693+00:00", "epoch": 0, "step": 9810, "train_loss": 3.5661137104034424, "perplexity": 35.3788332459873, "lr": 0.0026291804804649314, "grad_norm": 0.175962, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:18.866848+00:00", "epoch": 0, "step": 9811, "train_loss": 3.488452196121216, "perplexity": 32.73524074154409, "lr": 0.0026291804804649314, "grad_norm": 0.140231, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:19.171171+00:00", "epoch": 0, "step": 9812, "train_loss": 3.4540815353393555, "perplexity": 31.62922501192388, "lr": 0.0026291804804649314, "grad_norm": 0.181371, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:19.477809+00:00", "epoch": 0, "step": 9813, "train_loss": 3.583353042602539, "perplexity": 35.99402824462987, "lr": 0.0026291804804649314, "grad_norm": 0.161369, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:19.782713+00:00", "epoch": 0, "step": 9814, "train_loss": 3.513350009918213, "perplexity": 33.560507710233, "lr": 0.0026291804804649314, "grad_norm": 0.152273, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:20.087791+00:00", "epoch": 0, "step": 9815, "train_loss": 3.5419998168945312, "perplexity": 34.535915685340754, "lr": 0.0026291804804649314, "grad_norm": 0.203262, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:20.393347+00:00", "epoch": 0, "step": 9816, "train_loss": 3.5177063941955566, "perplexity": 33.70702909837279, "lr": 0.0026291804804649314, "grad_norm": 0.169199, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:20.698891+00:00", "epoch": 0, "step": 9817, "train_loss": 3.6239099502563477, "perplexity": 37.48384163023853, "lr": 0.0026291804804649314, "grad_norm": 0.199295, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:21.004513+00:00", "epoch": 0, "step": 9818, "train_loss": 3.5498011112213135, "perplexity": 34.80639419791002, "lr": 0.0026291804804649314, "grad_norm": 0.214765, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:21.308589+00:00", "epoch": 0, "step": 9819, "train_loss": 3.5385775566101074, "perplexity": 34.417926802194806, "lr": 0.0026291804804649314, "grad_norm": 0.135596, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:21.613470+00:00", "epoch": 0, "step": 9820, "train_loss": 3.459177255630493, "perplexity": 31.790810042020563, "lr": 0.0026291804804649314, "grad_norm": 0.168304, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:21.919135+00:00", "epoch": 0, "step": 9821, "train_loss": 3.56846022605896, "perplexity": 35.46194770863023, "lr": 0.0026291804804649314, "grad_norm": 0.174008, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:22.225031+00:00", "epoch": 0, "step": 9822, "train_loss": 3.501030206680298, "perplexity": 33.14958529769929, "lr": 0.0026291804804649314, "grad_norm": 0.159086, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:22.530984+00:00", "epoch": 0, "step": 9823, "train_loss": 3.5643608570098877, "perplexity": 35.31687365697421, "lr": 0.0026291804804649314, "grad_norm": 0.156107, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:22.836084+00:00", "epoch": 0, "step": 9824, "train_loss": 3.583888292312622, "perplexity": 36.01329919473768, "lr": 0.0026291804804649314, "grad_norm": 0.168662, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:23.141920+00:00", "epoch": 0, "step": 9825, "train_loss": 3.58465313911438, "perplexity": 36.040854387854885, "lr": 0.0026291804804649314, "grad_norm": 0.167197, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:23.446586+00:00", "epoch": 0, "step": 9826, "train_loss": 3.6118106842041016, "perplexity": 37.03304729736406, "lr": 0.0026291804804649314, "grad_norm": 0.164222, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:23.751804+00:00", "epoch": 0, "step": 9827, "train_loss": 3.720094680786133, "perplexity": 41.268301238846455, "lr": 0.0026291804804649314, "grad_norm": 0.186449, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:24.057750+00:00", "epoch": 0, "step": 9828, "train_loss": 3.594905138015747, "perplexity": 36.41224568616379, "lr": 0.0026291804804649314, "grad_norm": 0.153129, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:24.363306+00:00", "epoch": 0, "step": 9829, "train_loss": 3.5740528106689453, "perplexity": 35.66082725861813, "lr": 0.0026291804804649314, "grad_norm": 0.152223, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:24.669140+00:00", "epoch": 0, "step": 9830, "train_loss": 3.491910696029663, "perplexity": 32.84865157127953, "lr": 0.0026291804804649314, "grad_norm": 0.142416, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:24.974425+00:00", "epoch": 0, "step": 9831, "train_loss": 3.519620418548584, "perplexity": 33.7716069550154, "lr": 0.0026291804804649314, "grad_norm": 0.160828, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:25.279299+00:00", "epoch": 0, "step": 9832, "train_loss": 3.5955753326416016, "perplexity": 36.43665715684246, "lr": 0.0026291804804649314, "grad_norm": 0.154085, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:25.585947+00:00", "epoch": 0, "step": 9833, "train_loss": 3.5926461219787598, "perplexity": 36.33008267793351, "lr": 0.0026291804804649314, "grad_norm": 0.169032, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:25.891397+00:00", "epoch": 0, "step": 9834, "train_loss": 3.642916202545166, "perplexity": 38.20308236517834, "lr": 0.0026291804804649314, "grad_norm": 0.166151, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:26.195972+00:00", "epoch": 0, "step": 9835, "train_loss": 3.5618515014648438, "perplexity": 35.22836216407762, "lr": 0.0026291804804649314, "grad_norm": 0.196432, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:26.500842+00:00", "epoch": 0, "step": 9836, "train_loss": 3.5115253925323486, "perplexity": 33.4993284557255, "lr": 0.0026291804804649314, "grad_norm": 0.17025, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:26.806457+00:00", "epoch": 0, "step": 9837, "train_loss": 3.5843868255615234, "perplexity": 36.03125749782254, "lr": 0.0026291804804649314, "grad_norm": 0.163179, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:27.111958+00:00", "epoch": 0, "step": 9838, "train_loss": 3.5452160835266113, "perplexity": 34.64717121638822, "lr": 0.0026291804804649314, "grad_norm": 0.185973, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:27.416528+00:00", "epoch": 0, "step": 9839, "train_loss": 3.4648382663726807, "perplexity": 31.97128852249643, "lr": 0.0026291804804649314, "grad_norm": 0.170657, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:27.721247+00:00", "epoch": 0, "step": 9840, "train_loss": 3.5875766277313232, "perplexity": 36.146373582330504, "lr": 0.0026291804804649314, "grad_norm": 0.15654, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:28.026585+00:00", "epoch": 0, "step": 9841, "train_loss": 3.5520105361938477, "perplexity": 34.88338133178473, "lr": 0.0026291804804649314, "grad_norm": 0.18618, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:28.331126+00:00", "epoch": 0, "step": 9842, "train_loss": 3.489253044128418, "perplexity": 32.76146719415668, "lr": 0.0026291804804649314, "grad_norm": 0.151067, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:28.635609+00:00", "epoch": 0, "step": 9843, "train_loss": 3.529705047607422, "perplexity": 34.11390415305615, "lr": 0.0026291804804649314, "grad_norm": 0.180001, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:28.941931+00:00", "epoch": 0, "step": 9844, "train_loss": 3.507737636566162, "perplexity": 33.37268118040876, "lr": 0.0026291804804649314, "grad_norm": 0.163623, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:29.247318+00:00", "epoch": 0, "step": 9845, "train_loss": 3.515927314758301, "perplexity": 33.64711492785026, "lr": 0.0026291804804649314, "grad_norm": 0.153748, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:29.551405+00:00", "epoch": 0, "step": 9846, "train_loss": 3.358678102493286, "perplexity": 28.75115966182286, "lr": 0.0026291804804649314, "grad_norm": 0.175639, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:29.856307+00:00", "epoch": 0, "step": 9847, "train_loss": 3.5986409187316895, "perplexity": 36.54852825381916, "lr": 0.0026291804804649314, "grad_norm": 0.168563, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:30.161683+00:00", "epoch": 0, "step": 9848, "train_loss": 3.5543437004089355, "perplexity": 34.964865009245415, "lr": 0.0026291804804649314, "grad_norm": 0.148134, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:30.466978+00:00", "epoch": 0, "step": 9849, "train_loss": 3.5787901878356934, "perplexity": 35.830166843458166, "lr": 0.0026291804804649314, "grad_norm": 0.159915, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:30.772470+00:00", "epoch": 0, "step": 9850, "train_loss": 3.5609514713287354, "perplexity": 35.19666984064667, "lr": 0.0026291804804649314, "grad_norm": 0.152329, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:31.076806+00:00", "epoch": 0, "step": 9851, "train_loss": 3.551940441131592, "perplexity": 34.88093626469315, "lr": 0.0026291804804649314, "grad_norm": 0.177895, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:31.381087+00:00", "epoch": 0, "step": 9852, "train_loss": 3.560732841491699, "perplexity": 35.1889756395768, "lr": 0.0026291804804649314, "grad_norm": 0.176931, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:31.685380+00:00", "epoch": 0, "step": 9853, "train_loss": 3.5756354331970215, "perplexity": 35.71730957050782, "lr": 0.0026291804804649314, "grad_norm": 0.167194, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:31.990956+00:00", "epoch": 0, "step": 9854, "train_loss": 3.630268096923828, "perplexity": 37.7229286627493, "lr": 0.0026291804804649314, "grad_norm": 0.171599, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:32.295354+00:00", "epoch": 0, "step": 9855, "train_loss": 3.612860918045044, "perplexity": 37.071961087581855, "lr": 0.0026291804804649314, "grad_norm": 0.161224, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:32.600382+00:00", "epoch": 0, "step": 9856, "train_loss": 3.5497920513153076, "perplexity": 34.806078856678674, "lr": 0.0026291804804649314, "grad_norm": 0.149183, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:32.905166+00:00", "epoch": 0, "step": 9857, "train_loss": 3.5165436267852783, "perplexity": 33.667858441035115, "lr": 0.0026291804804649314, "grad_norm": 0.154554, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:33.210445+00:00", "epoch": 0, "step": 9858, "train_loss": 3.634883165359497, "perplexity": 37.89742490646089, "lr": 0.0026291804804649314, "grad_norm": 0.178467, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:33.514910+00:00", "epoch": 0, "step": 9859, "train_loss": 3.558213949203491, "perplexity": 35.10044994036548, "lr": 0.0026291804804649314, "grad_norm": 0.157582, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:33.821021+00:00", "epoch": 0, "step": 9860, "train_loss": 3.5330309867858887, "perplexity": 34.22755381457538, "lr": 0.0026291804804649314, "grad_norm": 0.152962, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:34.127140+00:00", "epoch": 0, "step": 9861, "train_loss": 3.501018762588501, "perplexity": 33.14920593297286, "lr": 0.0026291804804649314, "grad_norm": 0.159519, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:34.432934+00:00", "epoch": 0, "step": 9862, "train_loss": 3.5352962017059326, "perplexity": 34.30517446068502, "lr": 0.0026291804804649314, "grad_norm": 0.154267, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:34.737423+00:00", "epoch": 0, "step": 9863, "train_loss": 3.522050380706787, "perplexity": 33.85377046862287, "lr": 0.0026291804804649314, "grad_norm": 0.160207, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:35.042330+00:00", "epoch": 0, "step": 9864, "train_loss": 3.6839990615844727, "perplexity": 39.80525988944209, "lr": 0.0026291804804649314, "grad_norm": 0.161911, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:35.348133+00:00", "epoch": 0, "step": 9865, "train_loss": 3.4886953830718994, "perplexity": 32.74320249297771, "lr": 0.0026291804804649314, "grad_norm": 0.148647, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:35.653428+00:00", "epoch": 0, "step": 9866, "train_loss": 3.4214682579040527, "perplexity": 30.614331772895873, "lr": 0.0026291804804649314, "grad_norm": 0.16832, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:35.958254+00:00", "epoch": 0, "step": 9867, "train_loss": 3.617053985595703, "perplexity": 37.227732676826705, "lr": 0.0026291804804649314, "grad_norm": 0.149374, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:36.263375+00:00", "epoch": 0, "step": 9868, "train_loss": 3.5768990516662598, "perplexity": 35.76247115004637, "lr": 0.0026291804804649314, "grad_norm": 0.181203, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:36.568039+00:00", "epoch": 0, "step": 9869, "train_loss": 3.612286329269409, "perplexity": 37.05066607337412, "lr": 0.0026291804804649314, "grad_norm": 0.164868, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:36.872509+00:00", "epoch": 0, "step": 9870, "train_loss": 3.607422351837158, "perplexity": 36.87089003746657, "lr": 0.0026291804804649314, "grad_norm": 0.168645, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:37.177926+00:00", "epoch": 0, "step": 9871, "train_loss": 3.537275552749634, "perplexity": 34.373143688848494, "lr": 0.0026291804804649314, "grad_norm": 0.17074, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:37.484323+00:00", "epoch": 0, "step": 9872, "train_loss": 3.434291124343872, "perplexity": 31.0094229474584, "lr": 0.0026291804804649314, "grad_norm": 0.145082, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:37.788937+00:00", "epoch": 0, "step": 9873, "train_loss": 3.554918050765991, "perplexity": 34.98495286012783, "lr": 0.0026291804804649314, "grad_norm": 0.161273, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:38.093427+00:00", "epoch": 0, "step": 9874, "train_loss": 3.485476493835449, "perplexity": 32.637975199264865, "lr": 0.0026291804804649314, "grad_norm": 0.147041, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:38.398408+00:00", "epoch": 0, "step": 9875, "train_loss": 3.596794843673706, "perplexity": 36.48111916766424, "lr": 0.0026291804804649314, "grad_norm": 0.147609, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:38.703777+00:00", "epoch": 0, "step": 9876, "train_loss": 3.5478780269622803, "perplexity": 34.739522889303714, "lr": 0.0026291804804649314, "grad_norm": 0.160247, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:39.009659+00:00", "epoch": 0, "step": 9877, "train_loss": 3.5979554653167725, "perplexity": 36.523484524452854, "lr": 0.0026291804804649314, "grad_norm": 0.148073, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:39.314657+00:00", "epoch": 0, "step": 9878, "train_loss": 3.5027692317962646, "perplexity": 33.20728341381148, "lr": 0.0026291804804649314, "grad_norm": 0.141822, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:39.619163+00:00", "epoch": 0, "step": 9879, "train_loss": 3.546746253967285, "perplexity": 34.7002278761504, "lr": 0.0026291804804649314, "grad_norm": 0.166193, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:39.924157+00:00", "epoch": 0, "step": 9880, "train_loss": 3.5739519596099854, "perplexity": 35.65723100777161, "lr": 0.0026291804804649314, "grad_norm": 0.153676, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:40.227650+00:00", "epoch": 0, "step": 9881, "train_loss": 3.5288963317871094, "perplexity": 34.08632685167791, "lr": 0.0026291804804649314, "grad_norm": 0.157145, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:40.533241+00:00", "epoch": 0, "step": 9882, "train_loss": 3.457479476928711, "perplexity": 31.736882073650612, "lr": 0.0026291804804649314, "grad_norm": 0.213062, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:40.838211+00:00", "epoch": 0, "step": 9883, "train_loss": 3.519326686859131, "perplexity": 33.76168862058483, "lr": 0.0026291804804649314, "grad_norm": 0.200471, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:41.142677+00:00", "epoch": 0, "step": 9884, "train_loss": 3.46636700630188, "perplexity": 32.02020168607581, "lr": 0.0026291804804649314, "grad_norm": 0.181813, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:41.446865+00:00", "epoch": 0, "step": 9885, "train_loss": 3.547868251800537, "perplexity": 34.73918330650834, "lr": 0.0026291804804649314, "grad_norm": 0.172361, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:41.751255+00:00", "epoch": 0, "step": 9886, "train_loss": 3.6523189544677734, "perplexity": 38.56399057749087, "lr": 0.0026291804804649314, "grad_norm": 0.173226, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:42.057285+00:00", "epoch": 0, "step": 9887, "train_loss": 3.451794147491455, "perplexity": 31.55695938825647, "lr": 0.0026291804804649314, "grad_norm": 0.183857, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:42.363114+00:00", "epoch": 0, "step": 9888, "train_loss": 3.547201633453369, "perplexity": 34.716033246535225, "lr": 0.0026291804804649314, "grad_norm": 0.192126, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:42.667696+00:00", "epoch": 0, "step": 9889, "train_loss": 3.5186688899993896, "perplexity": 33.739487590515104, "lr": 0.0026291804804649314, "grad_norm": 0.177661, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:42.972701+00:00", "epoch": 0, "step": 9890, "train_loss": 3.6514902114868164, "perplexity": 38.532044180486665, "lr": 0.0026291804804649314, "grad_norm": 0.196487, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:43.277617+00:00", "epoch": 0, "step": 9891, "train_loss": 3.5785653591156006, "perplexity": 35.82211209840898, "lr": 0.0026291804804649314, "grad_norm": 0.182982, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:43.583355+00:00", "epoch": 0, "step": 9892, "train_loss": 3.612597703933716, "perplexity": 37.062204508379985, "lr": 0.0026291804804649314, "grad_norm": 0.156711, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:43.887864+00:00", "epoch": 0, "step": 9893, "train_loss": 3.566565990447998, "perplexity": 35.394838005307925, "lr": 0.0026291804804649314, "grad_norm": 0.16348, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:44.193994+00:00", "epoch": 0, "step": 9894, "train_loss": 3.5598223209381104, "perplexity": 35.156949936240366, "lr": 0.0026291804804649314, "grad_norm": 0.166886, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:44.499583+00:00", "epoch": 0, "step": 9895, "train_loss": 3.563239574432373, "perplexity": 35.277295655059056, "lr": 0.0026291804804649314, "grad_norm": 0.177849, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:44.804143+00:00", "epoch": 0, "step": 9896, "train_loss": 3.4794516563415527, "perplexity": 32.441927872223765, "lr": 0.0026291804804649314, "grad_norm": 0.161145, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:45.108664+00:00", "epoch": 0, "step": 9897, "train_loss": 3.665432929992676, "perplexity": 39.07304839648478, "lr": 0.0026291804804649314, "grad_norm": 0.174858, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:45.413483+00:00", "epoch": 0, "step": 9898, "train_loss": 3.5057101249694824, "perplexity": 33.30508623023166, "lr": 0.0026291804804649314, "grad_norm": 0.164877, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:45.718593+00:00", "epoch": 0, "step": 9899, "train_loss": 3.598052501678467, "perplexity": 36.527028802466646, "lr": 0.0026291804804649314, "grad_norm": 0.167927, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:46.022348+00:00", "epoch": 0, "step": 9900, "train_loss": 3.487117052078247, "perplexity": 32.69156364396945, "lr": 0.0026291804804649314, "grad_norm": 0.146549, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:46.326934+00:00", "epoch": 0, "step": 9901, "train_loss": 3.5099079608917236, "perplexity": 33.445189376873785, "lr": 0.0026291804804649314, "grad_norm": 0.161705, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:46.631133+00:00", "epoch": 0, "step": 9902, "train_loss": 3.624190092086792, "perplexity": 37.49434389323782, "lr": 0.0026291804804649314, "grad_norm": 0.161505, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:46.936272+00:00", "epoch": 0, "step": 9903, "train_loss": 3.520966053009033, "perplexity": 33.817081782504395, "lr": 0.0026291804804649314, "grad_norm": 0.158214, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:47.242348+00:00", "epoch": 0, "step": 9904, "train_loss": 3.5760812759399414, "perplexity": 35.7332374241716, "lr": 0.0026291804804649314, "grad_norm": 0.145749, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:47.545866+00:00", "epoch": 0, "step": 9905, "train_loss": 3.6233465671539307, "perplexity": 37.46272981482882, "lr": 0.0026291804804649314, "grad_norm": 0.183142, "tokens_per_sec": 107953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:47.850303+00:00", "epoch": 0, "step": 9906, "train_loss": 3.5939297676086426, "perplexity": 36.37674757398361, "lr": 0.0026291804804649314, "grad_norm": 0.17137, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:48.153999+00:00", "epoch": 0, "step": 9907, "train_loss": 3.603109121322632, "perplexity": 36.71219986911948, "lr": 0.0026291804804649314, "grad_norm": 0.173668, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:48.459096+00:00", "epoch": 0, "step": 9908, "train_loss": 3.5607125759124756, "perplexity": 35.18826252182908, "lr": 0.0026291804804649314, "grad_norm": 0.171434, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:48.763463+00:00", "epoch": 0, "step": 9909, "train_loss": 3.4550230503082275, "perplexity": 31.659018423997285, "lr": 0.0026291804804649314, "grad_norm": 0.168626, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:49.069494+00:00", "epoch": 0, "step": 9910, "train_loss": 3.5178020000457764, "perplexity": 33.71025184160222, "lr": 0.0026291804804649314, "grad_norm": 0.166326, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:49.374999+00:00", "epoch": 0, "step": 9911, "train_loss": 3.5157570838928223, "perplexity": 33.64138763784983, "lr": 0.0026291804804649314, "grad_norm": 0.166589, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:49.679729+00:00", "epoch": 0, "step": 9912, "train_loss": 3.4673287868499756, "perplexity": 32.05101290764527, "lr": 0.0026291804804649314, "grad_norm": 0.183085, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:49.985271+00:00", "epoch": 0, "step": 9913, "train_loss": 3.5160505771636963, "perplexity": 33.65126260779214, "lr": 0.0026291804804649314, "grad_norm": 0.152415, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:50.291168+00:00", "epoch": 0, "step": 9914, "train_loss": 3.47615647315979, "perplexity": 32.335201714806175, "lr": 0.0026291804804649314, "grad_norm": 0.167128, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:50.597805+00:00", "epoch": 0, "step": 9915, "train_loss": 3.5479276180267334, "perplexity": 34.741245701940095, "lr": 0.0026291804804649314, "grad_norm": 0.174191, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:50.903484+00:00", "epoch": 0, "step": 9916, "train_loss": 3.4359676837921143, "perplexity": 31.061455694294136, "lr": 0.0026291804804649314, "grad_norm": 0.145675, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:51.208189+00:00", "epoch": 0, "step": 9917, "train_loss": 3.5021581649780273, "perplexity": 33.18699774337477, "lr": 0.0026291804804649314, "grad_norm": 0.152646, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:51.513430+00:00", "epoch": 0, "step": 9918, "train_loss": 3.526073694229126, "perplexity": 33.99024916532549, "lr": 0.0026291804804649314, "grad_norm": 0.152763, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:51.818641+00:00", "epoch": 0, "step": 9919, "train_loss": 3.4358580112457275, "perplexity": 31.058049292151477, "lr": 0.0026291804804649314, "grad_norm": 0.151176, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:52.123343+00:00", "epoch": 0, "step": 9920, "train_loss": 3.527080774307251, "perplexity": 34.0244973105282, "lr": 0.0026291804804649314, "grad_norm": 0.16434, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:52.427883+00:00", "epoch": 0, "step": 9921, "train_loss": 3.6137285232543945, "perplexity": 37.10413887093006, "lr": 0.0026291804804649314, "grad_norm": 0.165457, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:52.732819+00:00", "epoch": 0, "step": 9922, "train_loss": 3.5256991386413574, "perplexity": 33.97752031155144, "lr": 0.0026291804804649314, "grad_norm": 0.173276, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:53.038025+00:00", "epoch": 0, "step": 9923, "train_loss": 3.669113874435425, "perplexity": 39.2171391491741, "lr": 0.0026291804804649314, "grad_norm": 0.186176, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:53.342658+00:00", "epoch": 0, "step": 9924, "train_loss": 3.564305067062378, "perplexity": 35.31490338540786, "lr": 0.0026291804804649314, "grad_norm": 0.193995, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:53.647895+00:00", "epoch": 0, "step": 9925, "train_loss": 3.567648410797119, "perplexity": 35.433170840596524, "lr": 0.0026291804804649314, "grad_norm": 0.153672, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:53.954000+00:00", "epoch": 0, "step": 9926, "train_loss": 3.4960970878601074, "perplexity": 32.98645715067095, "lr": 0.0026291804804649314, "grad_norm": 0.197935, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:54.259900+00:00", "epoch": 0, "step": 9927, "train_loss": 3.4951605796813965, "perplexity": 32.95557952459409, "lr": 0.0026291804804649314, "grad_norm": 0.148243, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:54.566098+00:00", "epoch": 0, "step": 9928, "train_loss": 3.5262680053710938, "perplexity": 33.996854491179995, "lr": 0.0026291804804649314, "grad_norm": 0.173394, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:54.871369+00:00", "epoch": 0, "step": 9929, "train_loss": 3.6122918128967285, "perplexity": 37.05086924597585, "lr": 0.0026291804804649314, "grad_norm": 0.181579, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:55.175725+00:00", "epoch": 0, "step": 9930, "train_loss": 3.5645956993103027, "perplexity": 35.32516852678231, "lr": 0.0026291804804649314, "grad_norm": 0.158426, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:55.479256+00:00", "epoch": 0, "step": 9931, "train_loss": 3.543367624282837, "perplexity": 34.583186487283236, "lr": 0.0026291804804649314, "grad_norm": 0.203617, "tokens_per_sec": 107898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:55.784029+00:00", "epoch": 0, "step": 9932, "train_loss": 3.5003838539123535, "perplexity": 33.1281658944769, "lr": 0.0026291804804649314, "grad_norm": 0.179201, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:56.089620+00:00", "epoch": 0, "step": 9933, "train_loss": 3.553485870361328, "perplexity": 34.93488395859152, "lr": 0.0026291804804649314, "grad_norm": 0.166927, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:56.395842+00:00", "epoch": 0, "step": 9934, "train_loss": 3.523512363433838, "perplexity": 33.90330029334559, "lr": 0.0026291804804649314, "grad_norm": 0.173459, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:56.700931+00:00", "epoch": 0, "step": 9935, "train_loss": 3.648308038711548, "perplexity": 38.40962344379275, "lr": 0.0026291804804649314, "grad_norm": 0.148127, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:57.005414+00:00", "epoch": 0, "step": 9936, "train_loss": 3.5533504486083984, "perplexity": 34.930153335689305, "lr": 0.0026291804804649314, "grad_norm": 0.155444, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:57.311670+00:00", "epoch": 0, "step": 9937, "train_loss": 3.393681049346924, "perplexity": 29.77535534011338, "lr": 0.0026291804804649314, "grad_norm": 0.161036, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:57.616617+00:00", "epoch": 0, "step": 9938, "train_loss": 3.52642560005188, "perplexity": 34.0022126368078, "lr": 0.0026291804804649314, "grad_norm": 0.158454, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:57.921513+00:00", "epoch": 0, "step": 9939, "train_loss": 3.5290634632110596, "perplexity": 34.09202422411359, "lr": 0.0026291804804649314, "grad_norm": 0.136901, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:58.227543+00:00", "epoch": 0, "step": 9940, "train_loss": 3.4861795902252197, "perplexity": 32.6609309108902, "lr": 0.0026291804804649314, "grad_norm": 0.151212, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:58.533522+00:00", "epoch": 0, "step": 9941, "train_loss": 3.494158983230591, "perplexity": 32.92258785803417, "lr": 0.0026291804804649314, "grad_norm": 0.156155, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:58.838234+00:00", "epoch": 0, "step": 9942, "train_loss": 3.515502691268921, "perplexity": 33.63283060544607, "lr": 0.0026291804804649314, "grad_norm": 0.167104, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:59.142961+00:00", "epoch": 0, "step": 9943, "train_loss": 3.6209068298339844, "perplexity": 37.371441999218995, "lr": 0.0026291804804649314, "grad_norm": 0.15452, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:59.447057+00:00", "epoch": 0, "step": 9944, "train_loss": 3.52547287940979, "perplexity": 33.96983345355927, "lr": 0.0026291804804649314, "grad_norm": 0.160479, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:49:59.751641+00:00", "epoch": 0, "step": 9945, "train_loss": 3.50128436088562, "perplexity": 33.158011474936224, "lr": 0.0026291804804649314, "grad_norm": 0.149933, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:00.056626+00:00", "epoch": 0, "step": 9946, "train_loss": 3.4484641551971436, "perplexity": 31.452049727786083, "lr": 0.0026291804804649314, "grad_norm": 0.159552, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:00.361487+00:00", "epoch": 0, "step": 9947, "train_loss": 3.5196499824523926, "perplexity": 33.77260539031363, "lr": 0.0026291804804649314, "grad_norm": 0.149557, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:00.667832+00:00", "epoch": 0, "step": 9948, "train_loss": 3.5456087589263916, "perplexity": 34.66077897973898, "lr": 0.0026291804804649314, "grad_norm": 0.150324, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:00.972076+00:00", "epoch": 0, "step": 9949, "train_loss": 3.4833297729492188, "perplexity": 32.567985727029615, "lr": 0.0026291804804649314, "grad_norm": 0.147952, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:01.276691+00:00", "epoch": 0, "step": 9950, "train_loss": 3.6009013652801514, "perplexity": 36.631237693277534, "lr": 0.0026291804804649314, "grad_norm": 0.14712, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:01.580196+00:00", "epoch": 0, "step": 9951, "train_loss": 3.5161707401275635, "perplexity": 33.655306486202306, "lr": 0.0026291804804649314, "grad_norm": 0.146694, "tokens_per_sec": 107907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:01.885793+00:00", "epoch": 0, "step": 9952, "train_loss": 3.5338730812072754, "perplexity": 34.256388785880624, "lr": 0.0026291804804649314, "grad_norm": 0.175336, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:02.190177+00:00", "epoch": 0, "step": 9953, "train_loss": 3.5402817726135254, "perplexity": 34.476632393150894, "lr": 0.0026291804804649314, "grad_norm": 0.157315, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:02.494188+00:00", "epoch": 0, "step": 9954, "train_loss": 3.4279086589813232, "perplexity": 30.812136635827443, "lr": 0.0026291804804649314, "grad_norm": 0.166074, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:02.798348+00:00", "epoch": 0, "step": 9955, "train_loss": 3.4337804317474365, "perplexity": 30.993590707787284, "lr": 0.0026291804804649314, "grad_norm": 0.136078, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:03.103095+00:00", "epoch": 0, "step": 9956, "train_loss": 3.5162370204925537, "perplexity": 33.65753724612704, "lr": 0.0026291804804649314, "grad_norm": 0.163686, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:03.408646+00:00", "epoch": 0, "step": 9957, "train_loss": 3.429896593093872, "perplexity": 30.87345005665478, "lr": 0.0026291804804649314, "grad_norm": 0.153808, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:03.714137+00:00", "epoch": 0, "step": 9958, "train_loss": 3.4407057762145996, "perplexity": 31.20897695084286, "lr": 0.0026291804804649314, "grad_norm": 0.161061, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:04.018419+00:00", "epoch": 0, "step": 9959, "train_loss": 3.496617317199707, "perplexity": 33.00362213796805, "lr": 0.0026291804804649314, "grad_norm": 0.155682, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:04.322943+00:00", "epoch": 0, "step": 9960, "train_loss": 3.523623466491699, "perplexity": 33.90706726293681, "lr": 0.0026291804804649314, "grad_norm": 0.151377, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:04.628330+00:00", "epoch": 0, "step": 9961, "train_loss": 3.533745050430298, "perplexity": 34.252003194559244, "lr": 0.0026291804804649314, "grad_norm": 0.145521, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:04.933122+00:00", "epoch": 0, "step": 9962, "train_loss": 3.4557418823242188, "perplexity": 31.681784121411102, "lr": 0.0026291804804649314, "grad_norm": 0.159056, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:05.239610+00:00", "epoch": 0, "step": 9963, "train_loss": 3.5687272548675537, "perplexity": 35.471418334686, "lr": 0.0026291804804649314, "grad_norm": 0.159545, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:05.544896+00:00", "epoch": 0, "step": 9964, "train_loss": 3.491309642791748, "perplexity": 32.82891371523683, "lr": 0.0026291804804649314, "grad_norm": 0.16276, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:05.849988+00:00", "epoch": 0, "step": 9965, "train_loss": 3.607961654663086, "perplexity": 36.89077997552607, "lr": 0.0026291804804649314, "grad_norm": 0.165415, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:06.154870+00:00", "epoch": 0, "step": 9966, "train_loss": 3.443037271499634, "perplexity": 31.281825423390057, "lr": 0.0026291804804649314, "grad_norm": 0.174062, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:06.459351+00:00", "epoch": 0, "step": 9967, "train_loss": 3.6144697666168213, "perplexity": 37.13165226338671, "lr": 0.0026291804804649314, "grad_norm": 0.204361, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:06.765355+00:00", "epoch": 0, "step": 9968, "train_loss": 3.5370943546295166, "perplexity": 34.36691590407804, "lr": 0.0026291804804649314, "grad_norm": 0.195264, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:07.070060+00:00", "epoch": 0, "step": 9969, "train_loss": 3.461996555328369, "perplexity": 31.88056432582926, "lr": 0.0026291804804649314, "grad_norm": 0.161494, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:07.374830+00:00", "epoch": 0, "step": 9970, "train_loss": 3.582080364227295, "perplexity": 35.948248560830955, "lr": 0.0026291804804649314, "grad_norm": 0.160618, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:07.678191+00:00", "epoch": 0, "step": 9971, "train_loss": 3.600200891494751, "perplexity": 36.60558745625474, "lr": 0.0026291804804649314, "grad_norm": 0.172377, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:07.983142+00:00", "epoch": 0, "step": 9972, "train_loss": 3.5388355255126953, "perplexity": 34.426806702321116, "lr": 0.0026291804804649314, "grad_norm": 0.158437, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:08.288640+00:00", "epoch": 0, "step": 9973, "train_loss": 3.5121805667877197, "perplexity": 33.52128354472109, "lr": 0.0026291804804649314, "grad_norm": 0.153899, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:08.594564+00:00", "epoch": 0, "step": 9974, "train_loss": 3.4054243564605713, "perplexity": 30.127077631714535, "lr": 0.0026291804804649314, "grad_norm": 0.169255, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:08.899095+00:00", "epoch": 0, "step": 9975, "train_loss": 3.5074849128723145, "perplexity": 33.36424817880179, "lr": 0.0026291804804649314, "grad_norm": 0.14934, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:09.203186+00:00", "epoch": 0, "step": 9976, "train_loss": 3.5863382816314697, "perplexity": 36.10163956539532, "lr": 0.0026291804804649314, "grad_norm": 0.148443, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:09.508044+00:00", "epoch": 0, "step": 9977, "train_loss": 3.600170135498047, "perplexity": 36.60446163224059, "lr": 0.0026291804804649314, "grad_norm": 0.142894, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:09.812853+00:00", "epoch": 0, "step": 9978, "train_loss": 3.5466549396514893, "perplexity": 34.69705939325005, "lr": 0.0026291804804649314, "grad_norm": 0.155346, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:10.118361+00:00", "epoch": 0, "step": 9979, "train_loss": 3.6117656230926514, "perplexity": 37.031378584689755, "lr": 0.0026291804804649314, "grad_norm": 0.156663, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:10.424059+00:00", "epoch": 0, "step": 9980, "train_loss": 3.565385580062866, "perplexity": 35.35308222027598, "lr": 0.0026291804804649314, "grad_norm": 0.160418, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:10.728998+00:00", "epoch": 0, "step": 9981, "train_loss": 3.5202252864837646, "perplexity": 33.792040496355995, "lr": 0.0026291804804649314, "grad_norm": 0.172655, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:11.033808+00:00", "epoch": 0, "step": 9982, "train_loss": 3.536195755004883, "perplexity": 34.33604767751234, "lr": 0.0026291804804649314, "grad_norm": 0.159754, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:11.338502+00:00", "epoch": 0, "step": 9983, "train_loss": 3.466120719909668, "perplexity": 32.01231651716941, "lr": 0.0026291804804649314, "grad_norm": 0.169096, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:11.644871+00:00", "epoch": 0, "step": 9984, "train_loss": 3.5541176795959473, "perplexity": 34.956963115059914, "lr": 0.0026291804804649314, "grad_norm": 0.176372, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:11.949766+00:00", "epoch": 0, "step": 9985, "train_loss": 3.4894895553588867, "perplexity": 32.76921656544528, "lr": 0.0026291804804649314, "grad_norm": 0.152504, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:12.253949+00:00", "epoch": 0, "step": 9986, "train_loss": 3.5860161781311035, "perplexity": 36.090012973505836, "lr": 0.0026291804804649314, "grad_norm": 0.15301, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:12.558827+00:00", "epoch": 0, "step": 9987, "train_loss": 3.6023237705230713, "perplexity": 36.68337923223507, "lr": 0.0026291804804649314, "grad_norm": 0.174569, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:12.864300+00:00", "epoch": 0, "step": 9988, "train_loss": 3.598682165145874, "perplexity": 36.55003578064318, "lr": 0.0026291804804649314, "grad_norm": 0.163707, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:13.169976+00:00", "epoch": 0, "step": 9989, "train_loss": 3.4930038452148438, "perplexity": 32.884579681793554, "lr": 0.0026291804804649314, "grad_norm": 0.16043, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:13.474792+00:00", "epoch": 0, "step": 9990, "train_loss": 3.540332555770874, "perplexity": 34.47838326985571, "lr": 0.0026291804804649314, "grad_norm": 0.141292, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:13.779759+00:00", "epoch": 0, "step": 9991, "train_loss": 3.5306413173675537, "perplexity": 34.14585892672453, "lr": 0.0026291804804649314, "grad_norm": 0.166638, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:14.084728+00:00", "epoch": 0, "step": 9992, "train_loss": 3.6479008197784424, "perplexity": 38.393985502162046, "lr": 0.0026291804804649314, "grad_norm": 0.181607, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:14.388771+00:00", "epoch": 0, "step": 9993, "train_loss": 3.6295125484466553, "perplexity": 37.69443792586543, "lr": 0.0026291804804649314, "grad_norm": 0.156407, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:14.692802+00:00", "epoch": 0, "step": 9994, "train_loss": 3.5358388423919678, "perplexity": 34.323794895734835, "lr": 0.0026291804804649314, "grad_norm": 0.1586, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:14.998187+00:00", "epoch": 0, "step": 9995, "train_loss": 3.549778938293457, "perplexity": 34.80562244679855, "lr": 0.0026291804804649314, "grad_norm": 0.16418, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:15.305431+00:00", "epoch": 0, "step": 9996, "train_loss": 3.4888968467712402, "perplexity": 32.7497997242092, "lr": 0.0026291804804649314, "grad_norm": 0.192203, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:15.610517+00:00", "epoch": 0, "step": 9997, "train_loss": 3.487935781478882, "perplexity": 32.71834014813737, "lr": 0.0026291804804649314, "grad_norm": 0.18968, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:15.914637+00:00", "epoch": 0, "step": 9998, "train_loss": 3.6174871921539307, "perplexity": 37.24386346850035, "lr": 0.0026291804804649314, "grad_norm": 0.161551, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:16.219462+00:00", "epoch": 0, "step": 9999, "train_loss": 3.4928925037384033, "perplexity": 32.88091846796592, "lr": 0.0026291804804649314, "grad_norm": 0.155555, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:16.525002+00:00", "epoch": 0, "step": 10000, "train_loss": 3.486008405685425, "perplexity": 32.655340342985994, "lr": 0.0026291804804649314, "grad_norm": 0.15706, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:50:19.597770+00:00", "step": 10000, "epoch": 0, "val_loss": 3.4971402883529663, "val_ppl": 33.02088659431212, "eval_train_loss": 3.486008405685425, "eval_train_ppl": 32.655340342985994} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:50:20.610825+00:00", "step": 10000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4971_epoch_0000_step_0010000.pt", "val_loss": 3.4971402883529663} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T21:50:22.520594+00:00", "step": 10000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0010000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:23.666012+00:00", "epoch": 0, "step": 10001, "train_loss": 3.4975175857543945, "perplexity": 33.03334763962919, "lr": 0.0026291804804649314, "grad_norm": 0.150228, "tokens_per_sec": 4589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:23.970187+00:00", "epoch": 0, "step": 10002, "train_loss": 3.38490891456604, "perplexity": 29.515304179420443, "lr": 0.0026291804804649314, "grad_norm": 0.164252, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:24.273143+00:00", "epoch": 0, "step": 10003, "train_loss": 3.621372699737549, "perplexity": 37.38885628538018, "lr": 0.0026291804804649314, "grad_norm": 0.21552, "tokens_per_sec": 108158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:24.578202+00:00", "epoch": 0, "step": 10004, "train_loss": 3.565614938735962, "perplexity": 35.3611916862565, "lr": 0.0026291804804649314, "grad_norm": 0.183251, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:24.883213+00:00", "epoch": 0, "step": 10005, "train_loss": 3.5348708629608154, "perplexity": 34.29058624351581, "lr": 0.0026291804804649314, "grad_norm": 0.164789, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:25.188944+00:00", "epoch": 0, "step": 10006, "train_loss": 3.5739803314208984, "perplexity": 35.658242682338894, "lr": 0.0026291804804649314, "grad_norm": 0.18512, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:25.493124+00:00", "epoch": 0, "step": 10007, "train_loss": 3.4920623302459717, "perplexity": 32.85363292847989, "lr": 0.0026291804804649314, "grad_norm": 0.177202, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:25.797443+00:00", "epoch": 0, "step": 10008, "train_loss": 3.368745803833008, "perplexity": 29.042079741439302, "lr": 0.0026291804804649314, "grad_norm": 0.149008, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:26.102589+00:00", "epoch": 0, "step": 10009, "train_loss": 3.5512819290161133, "perplexity": 34.85797430675713, "lr": 0.0026291804804649314, "grad_norm": 0.170528, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:26.844496+00:00", "epoch": 0, "step": 10010, "train_loss": 3.6273138523101807, "perplexity": 37.611650356537076, "lr": 0.0026291804804649314, "grad_norm": 0.164493, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:27.147328+00:00", "epoch": 0, "step": 10011, "train_loss": 3.548530340194702, "perplexity": 34.76219133243133, "lr": 0.0026291804804649314, "grad_norm": 0.155721, "tokens_per_sec": 44311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:27.452281+00:00", "epoch": 0, "step": 10012, "train_loss": 3.5926566123962402, "perplexity": 36.33046379766695, "lr": 0.0026291804804649314, "grad_norm": 0.16645, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:27.757374+00:00", "epoch": 0, "step": 10013, "train_loss": 3.6273751258850098, "perplexity": 37.61395502741663, "lr": 0.0026291804804649314, "grad_norm": 0.151781, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:28.062797+00:00", "epoch": 0, "step": 10014, "train_loss": 3.5182273387908936, "perplexity": 33.724593167562695, "lr": 0.0026291804804649314, "grad_norm": 0.160984, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:28.367406+00:00", "epoch": 0, "step": 10015, "train_loss": 3.500261068344116, "perplexity": 33.12409848351768, "lr": 0.0026291804804649314, "grad_norm": 0.160869, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:28.671051+00:00", "epoch": 0, "step": 10016, "train_loss": 3.4153571128845215, "perplexity": 30.427813651875397, "lr": 0.0026291804804649314, "grad_norm": 0.154867, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:28.975564+00:00", "epoch": 0, "step": 10017, "train_loss": 3.5493414402008057, "perplexity": 34.79039838385661, "lr": 0.0026291804804649314, "grad_norm": 0.166346, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:29.280306+00:00", "epoch": 0, "step": 10018, "train_loss": 3.601172685623169, "perplexity": 36.641177841674896, "lr": 0.0026291804804649314, "grad_norm": 0.165311, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:29.586262+00:00", "epoch": 0, "step": 10019, "train_loss": 3.5404601097106934, "perplexity": 34.48278140397409, "lr": 0.0026291804804649314, "grad_norm": 0.165161, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:29.890796+00:00", "epoch": 0, "step": 10020, "train_loss": 3.5618860721588135, "perplexity": 35.229580054056576, "lr": 0.0026291804804649314, "grad_norm": 0.151124, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:30.194754+00:00", "epoch": 0, "step": 10021, "train_loss": 3.5583553314208984, "perplexity": 35.10541287063682, "lr": 0.0026291804804649314, "grad_norm": 0.153763, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:30.498665+00:00", "epoch": 0, "step": 10022, "train_loss": 3.59985089302063, "perplexity": 36.59277779831188, "lr": 0.0026291804804649314, "grad_norm": 0.145249, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:30.802940+00:00", "epoch": 0, "step": 10023, "train_loss": 3.5913357734680176, "perplexity": 36.282508784198825, "lr": 0.0026291804804649314, "grad_norm": 0.150249, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:31.108687+00:00", "epoch": 0, "step": 10024, "train_loss": 3.674595594406128, "perplexity": 39.43270682505522, "lr": 0.0026291804804649314, "grad_norm": 0.174999, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:31.413097+00:00", "epoch": 0, "step": 10025, "train_loss": 3.6030218601226807, "perplexity": 36.70899645827482, "lr": 0.0026291804804649314, "grad_norm": 0.153545, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:31.717425+00:00", "epoch": 0, "step": 10026, "train_loss": 3.4154913425445557, "perplexity": 30.431898241087847, "lr": 0.0026291804804649314, "grad_norm": 0.162845, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:32.021433+00:00", "epoch": 0, "step": 10027, "train_loss": 3.48262882232666, "perplexity": 32.54516517613282, "lr": 0.0026291804804649314, "grad_norm": 0.165505, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:32.325542+00:00", "epoch": 0, "step": 10028, "train_loss": 3.4770851135253906, "perplexity": 32.36524343516654, "lr": 0.0026291804804649314, "grad_norm": 0.193818, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:32.630048+00:00", "epoch": 0, "step": 10029, "train_loss": 3.5176990032196045, "perplexity": 33.70677997145196, "lr": 0.0026291804804649314, "grad_norm": 0.178181, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:32.936253+00:00", "epoch": 0, "step": 10030, "train_loss": 3.5550074577331543, "perplexity": 34.988080898491525, "lr": 0.0026291804804649314, "grad_norm": 0.224721, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:33.240550+00:00", "epoch": 0, "step": 10031, "train_loss": 3.580169677734375, "perplexity": 35.8796283046361, "lr": 0.0026291804804649314, "grad_norm": 0.164099, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:33.544780+00:00", "epoch": 0, "step": 10032, "train_loss": 3.6130526065826416, "perplexity": 37.079068038727385, "lr": 0.0026291804804649314, "grad_norm": 0.181692, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:33.849320+00:00", "epoch": 0, "step": 10033, "train_loss": 3.4792726039886475, "perplexity": 32.43611958871439, "lr": 0.0026291804804649314, "grad_norm": 0.181639, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:34.153498+00:00", "epoch": 0, "step": 10034, "train_loss": 3.492659330368042, "perplexity": 32.87325240718152, "lr": 0.0026291804804649314, "grad_norm": 0.151784, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:34.460453+00:00", "epoch": 0, "step": 10035, "train_loss": 3.5896761417388916, "perplexity": 36.22234312163948, "lr": 0.0026291804804649314, "grad_norm": 0.168497, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:34.766587+00:00", "epoch": 0, "step": 10036, "train_loss": 3.5233538150787354, "perplexity": 33.89792540695249, "lr": 0.0026291804804649314, "grad_norm": 0.167129, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:35.071238+00:00", "epoch": 0, "step": 10037, "train_loss": 3.579134702682495, "perplexity": 35.84251299449318, "lr": 0.0026291804804649314, "grad_norm": 0.164866, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:35.376155+00:00", "epoch": 0, "step": 10038, "train_loss": 3.5633533000946045, "perplexity": 35.28130781700773, "lr": 0.0026291804804649314, "grad_norm": 0.178666, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:35.681737+00:00", "epoch": 0, "step": 10039, "train_loss": 3.5534753799438477, "perplexity": 34.934517478996426, "lr": 0.0026291804804649314, "grad_norm": 0.165255, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:35.988201+00:00", "epoch": 0, "step": 10040, "train_loss": 3.42814302444458, "perplexity": 30.81935878278188, "lr": 0.0026291804804649314, "grad_norm": 0.184116, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:36.293678+00:00", "epoch": 0, "step": 10041, "train_loss": 3.6052329540252686, "perplexity": 36.790253296658825, "lr": 0.0026291804804649314, "grad_norm": 0.164504, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:36.600085+00:00", "epoch": 0, "step": 10042, "train_loss": 3.5230607986450195, "perplexity": 33.88799421281198, "lr": 0.0026291804804649314, "grad_norm": 0.176352, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:36.904358+00:00", "epoch": 0, "step": 10043, "train_loss": 3.574155807495117, "perplexity": 35.66450039980212, "lr": 0.0026291804804649314, "grad_norm": 0.165179, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:37.210032+00:00", "epoch": 0, "step": 10044, "train_loss": 3.6370205879211426, "perplexity": 37.978514347813025, "lr": 0.0026291804804649314, "grad_norm": 0.201347, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:37.516123+00:00", "epoch": 0, "step": 10045, "train_loss": 3.563767433166504, "perplexity": 35.295921999293846, "lr": 0.0026291804804649314, "grad_norm": 0.169659, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:37.820472+00:00", "epoch": 0, "step": 10046, "train_loss": 3.4659245014190674, "perplexity": 32.00603572496571, "lr": 0.0026291804804649314, "grad_norm": 0.163797, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:38.125533+00:00", "epoch": 0, "step": 10047, "train_loss": 3.5787205696105957, "perplexity": 35.827672497664544, "lr": 0.0026291804804649314, "grad_norm": 0.160592, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:38.430726+00:00", "epoch": 0, "step": 10048, "train_loss": 3.5382039546966553, "perplexity": 34.40507060058474, "lr": 0.0026291804804649314, "grad_norm": 0.137694, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:38.735965+00:00", "epoch": 0, "step": 10049, "train_loss": 3.537684440612793, "perplexity": 34.38720132392381, "lr": 0.0026291804804649314, "grad_norm": 0.142965, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:39.042731+00:00", "epoch": 0, "step": 10050, "train_loss": 3.547682762145996, "perplexity": 34.73274014498615, "lr": 0.0026291804804649314, "grad_norm": 0.154462, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:39.349155+00:00", "epoch": 0, "step": 10051, "train_loss": 3.604128360748291, "perplexity": 36.749637466325304, "lr": 0.0026291804804649314, "grad_norm": 0.160928, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:39.654818+00:00", "epoch": 0, "step": 10052, "train_loss": 3.578421115875244, "perplexity": 35.81694537352502, "lr": 0.0026291804804649314, "grad_norm": 0.150221, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:39.959835+00:00", "epoch": 0, "step": 10053, "train_loss": 3.5537192821502686, "perplexity": 34.943039124071575, "lr": 0.0026291804804649314, "grad_norm": 0.161244, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:40.265061+00:00", "epoch": 0, "step": 10054, "train_loss": 3.5740184783935547, "perplexity": 35.65960296229258, "lr": 0.0026291804804649314, "grad_norm": 0.148263, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:40.571289+00:00", "epoch": 0, "step": 10055, "train_loss": 3.5697546005249023, "perplexity": 35.50787846764498, "lr": 0.0026291804804649314, "grad_norm": 0.163448, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:40.876638+00:00", "epoch": 0, "step": 10056, "train_loss": 3.560882568359375, "perplexity": 35.19424476913133, "lr": 0.0026291804804649314, "grad_norm": 0.160106, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:41.182003+00:00", "epoch": 0, "step": 10057, "train_loss": 3.4689810276031494, "perplexity": 32.10401266947799, "lr": 0.0026291804804649314, "grad_norm": 0.165811, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:41.487549+00:00", "epoch": 0, "step": 10058, "train_loss": 3.481370687484741, "perplexity": 32.50424471701417, "lr": 0.0026291804804649314, "grad_norm": 0.156486, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:41.791958+00:00", "epoch": 0, "step": 10059, "train_loss": 3.5200388431549072, "perplexity": 33.78574078312497, "lr": 0.0026291804804649314, "grad_norm": 0.149257, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:42.097344+00:00", "epoch": 0, "step": 10060, "train_loss": 3.658468723297119, "perplexity": 38.80188094030468, "lr": 0.0026291804804649314, "grad_norm": 0.194199, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:42.402329+00:00", "epoch": 0, "step": 10061, "train_loss": 3.6336140632629395, "perplexity": 37.84935971132916, "lr": 0.0026291804804649314, "grad_norm": 0.157137, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:42.709924+00:00", "epoch": 0, "step": 10062, "train_loss": 3.5358245372772217, "perplexity": 34.32330389342226, "lr": 0.0026291804804649314, "grad_norm": 0.193536, "tokens_per_sec": 106529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:43.015072+00:00", "epoch": 0, "step": 10063, "train_loss": 3.559281587600708, "perplexity": 35.137944540257145, "lr": 0.0026291804804649314, "grad_norm": 0.210378, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:43.320146+00:00", "epoch": 0, "step": 10064, "train_loss": 3.5755746364593506, "perplexity": 35.71513814061615, "lr": 0.0026291804804649314, "grad_norm": 0.182556, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:43.624881+00:00", "epoch": 0, "step": 10065, "train_loss": 3.5175540447235107, "perplexity": 33.70189424144172, "lr": 0.0026291804804649314, "grad_norm": 0.149266, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:43.930960+00:00", "epoch": 0, "step": 10066, "train_loss": 3.619556427001953, "perplexity": 37.321009557825995, "lr": 0.0026291804804649314, "grad_norm": 0.151874, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:44.236816+00:00", "epoch": 0, "step": 10067, "train_loss": 3.5370309352874756, "perplexity": 34.364736445994055, "lr": 0.0026291804804649314, "grad_norm": 0.162245, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:44.542251+00:00", "epoch": 0, "step": 10068, "train_loss": 3.5277557373046875, "perplexity": 34.04747033931921, "lr": 0.0026291804804649314, "grad_norm": 0.158402, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:44.847725+00:00", "epoch": 0, "step": 10069, "train_loss": 3.4466922283172607, "perplexity": 31.39636834168706, "lr": 0.0026291804804649314, "grad_norm": 0.171099, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:45.153114+00:00", "epoch": 0, "step": 10070, "train_loss": 3.4565320014953613, "perplexity": 31.706826398310163, "lr": 0.0026291804804649314, "grad_norm": 0.147199, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:45.458824+00:00", "epoch": 0, "step": 10071, "train_loss": 3.584097385406494, "perplexity": 36.02083011419103, "lr": 0.0026291804804649314, "grad_norm": 0.16036, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:45.764360+00:00", "epoch": 0, "step": 10072, "train_loss": 3.5128121376037598, "perplexity": 33.54246129604503, "lr": 0.0026291804804649314, "grad_norm": 0.156718, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:46.071841+00:00", "epoch": 0, "step": 10073, "train_loss": 3.5059762001037598, "perplexity": 33.31394906456012, "lr": 0.0026291804804649314, "grad_norm": 0.163818, "tokens_per_sec": 106569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:46.376770+00:00", "epoch": 0, "step": 10074, "train_loss": 3.5161099433898926, "perplexity": 33.65326041556047, "lr": 0.0026291804804649314, "grad_norm": 0.177479, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:46.682090+00:00", "epoch": 0, "step": 10075, "train_loss": 3.4361233711242676, "perplexity": 31.066291945925766, "lr": 0.0026291804804649314, "grad_norm": 0.162776, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:46.987323+00:00", "epoch": 0, "step": 10076, "train_loss": 3.519437313079834, "perplexity": 33.765423755199585, "lr": 0.0026291804804649314, "grad_norm": 0.173015, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:47.293824+00:00", "epoch": 0, "step": 10077, "train_loss": 3.495744228363037, "perplexity": 32.97481961931811, "lr": 0.0026291804804649314, "grad_norm": 0.168979, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:47.600094+00:00", "epoch": 0, "step": 10078, "train_loss": 3.372030258178711, "perplexity": 29.13762394585166, "lr": 0.0026291804804649314, "grad_norm": 0.176628, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:47.904998+00:00", "epoch": 0, "step": 10079, "train_loss": 3.5464999675750732, "perplexity": 34.691682734537096, "lr": 0.0026291804804649314, "grad_norm": 0.151533, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:48.209808+00:00", "epoch": 0, "step": 10080, "train_loss": 3.496326208114624, "perplexity": 32.994015882025074, "lr": 0.0026291804804649314, "grad_norm": 0.179033, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:48.513815+00:00", "epoch": 0, "step": 10081, "train_loss": 3.6145706176757812, "perplexity": 37.13539721867659, "lr": 0.0026291804804649314, "grad_norm": 0.173978, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:48.819359+00:00", "epoch": 0, "step": 10082, "train_loss": 3.5178072452545166, "perplexity": 33.71042865937354, "lr": 0.0026291804804649314, "grad_norm": 0.155815, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:49.125034+00:00", "epoch": 0, "step": 10083, "train_loss": 3.5793800354003906, "perplexity": 35.85130741435767, "lr": 0.0026291804804649314, "grad_norm": 0.19048, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:49.430595+00:00", "epoch": 0, "step": 10084, "train_loss": 3.5318901538848877, "perplexity": 34.188528160170854, "lr": 0.0026291804804649314, "grad_norm": 0.173582, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:49.735733+00:00", "epoch": 0, "step": 10085, "train_loss": 3.606224536895752, "perplexity": 36.82675197436977, "lr": 0.0026291804804649314, "grad_norm": 0.158816, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:50.040883+00:00", "epoch": 0, "step": 10086, "train_loss": 3.5697519779205322, "perplexity": 35.50778534464985, "lr": 0.0026291804804649314, "grad_norm": 0.146978, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:50.347652+00:00", "epoch": 0, "step": 10087, "train_loss": 3.484996795654297, "perplexity": 32.62232257649277, "lr": 0.0026291804804649314, "grad_norm": 0.14741, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:50.653755+00:00", "epoch": 0, "step": 10088, "train_loss": 3.5501954555511475, "perplexity": 34.82012260878653, "lr": 0.0026291804804649314, "grad_norm": 0.161413, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:50.960857+00:00", "epoch": 0, "step": 10089, "train_loss": 3.5462164878845215, "perplexity": 34.68184974084364, "lr": 0.0026291804804649314, "grad_norm": 0.177765, "tokens_per_sec": 106701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:51.266786+00:00", "epoch": 0, "step": 10090, "train_loss": 3.5238864421844482, "perplexity": 33.915985169984374, "lr": 0.0026291804804649314, "grad_norm": 0.151026, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:51.572716+00:00", "epoch": 0, "step": 10091, "train_loss": 3.541551113128662, "perplexity": 34.52042276604056, "lr": 0.0026291804804649314, "grad_norm": 0.159176, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:51.877475+00:00", "epoch": 0, "step": 10092, "train_loss": 3.6527748107910156, "perplexity": 38.581574223948785, "lr": 0.0026291804804649314, "grad_norm": 0.176062, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:52.183600+00:00", "epoch": 0, "step": 10093, "train_loss": 3.4687812328338623, "perplexity": 32.09759909639401, "lr": 0.0026291804804649314, "grad_norm": 0.161457, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:52.489864+00:00", "epoch": 0, "step": 10094, "train_loss": 3.5948574542999268, "perplexity": 36.41050945638339, "lr": 0.0026291804804649314, "grad_norm": 0.183723, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:52.794766+00:00", "epoch": 0, "step": 10095, "train_loss": 3.5375232696533203, "perplexity": 34.381659552291175, "lr": 0.0026291804804649314, "grad_norm": 0.159222, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:53.099843+00:00", "epoch": 0, "step": 10096, "train_loss": 3.541253089904785, "perplexity": 34.51013641122067, "lr": 0.0026291804804649314, "grad_norm": 0.153501, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:53.405329+00:00", "epoch": 0, "step": 10097, "train_loss": 3.5406739711761475, "perplexity": 34.49015673075906, "lr": 0.0026291804804649314, "grad_norm": 0.17042, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:53.711115+00:00", "epoch": 0, "step": 10098, "train_loss": 3.543588399887085, "perplexity": 34.590822454063364, "lr": 0.0026291804804649314, "grad_norm": 0.175527, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:54.017077+00:00", "epoch": 0, "step": 10099, "train_loss": 3.4649765491485596, "perplexity": 31.975709906715345, "lr": 0.0026291804804649314, "grad_norm": 0.18222, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:54.323131+00:00", "epoch": 0, "step": 10100, "train_loss": 3.6103007793426514, "perplexity": 36.97717311218259, "lr": 0.0026291804804649314, "grad_norm": 0.163713, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:54.628633+00:00", "epoch": 0, "step": 10101, "train_loss": 3.6520419120788574, "perplexity": 38.55330819721955, "lr": 0.0026291804804649314, "grad_norm": 0.156892, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:54.934472+00:00", "epoch": 0, "step": 10102, "train_loss": 3.539710760116577, "perplexity": 34.456951424775916, "lr": 0.0026291804804649314, "grad_norm": 0.178591, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:55.239413+00:00", "epoch": 0, "step": 10103, "train_loss": 3.469663619995117, "perplexity": 32.125934105129254, "lr": 0.0026291804804649314, "grad_norm": 0.181469, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:55.545820+00:00", "epoch": 0, "step": 10104, "train_loss": 3.512664556503296, "perplexity": 33.53751142795717, "lr": 0.0026291804804649314, "grad_norm": 0.169751, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:55.853303+00:00", "epoch": 0, "step": 10105, "train_loss": 3.4443352222442627, "perplexity": 31.32245405326806, "lr": 0.0026291804804649314, "grad_norm": 0.210185, "tokens_per_sec": 106570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:56.159740+00:00", "epoch": 0, "step": 10106, "train_loss": 3.5963969230651855, "perplexity": 36.46660546636747, "lr": 0.0026291804804649314, "grad_norm": 0.159298, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:56.464766+00:00", "epoch": 0, "step": 10107, "train_loss": 3.5829107761383057, "perplexity": 35.97811281271492, "lr": 0.0026291804804649314, "grad_norm": 0.168061, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:56.770455+00:00", "epoch": 0, "step": 10108, "train_loss": 3.496187448501587, "perplexity": 32.98943796277124, "lr": 0.0026291804804649314, "grad_norm": 0.202198, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:57.076403+00:00", "epoch": 0, "step": 10109, "train_loss": 3.4506261348724365, "perplexity": 31.52012197894056, "lr": 0.0026291804804649314, "grad_norm": 0.18053, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:57.383441+00:00", "epoch": 0, "step": 10110, "train_loss": 3.5031075477600098, "perplexity": 33.21851986853181, "lr": 0.0026291804804649314, "grad_norm": 0.163176, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:57.688457+00:00", "epoch": 0, "step": 10111, "train_loss": 3.507176399230957, "perplexity": 33.353956440757415, "lr": 0.0026291804804649314, "grad_norm": 0.200382, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:57.993555+00:00", "epoch": 0, "step": 10112, "train_loss": 3.6195480823516846, "perplexity": 37.32069812835295, "lr": 0.0026291804804649314, "grad_norm": 0.154928, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:58.298411+00:00", "epoch": 0, "step": 10113, "train_loss": 3.563638925552368, "perplexity": 35.291386495998594, "lr": 0.0026291804804649314, "grad_norm": 0.145883, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:58.604967+00:00", "epoch": 0, "step": 10114, "train_loss": 3.631978988647461, "perplexity": 37.78752375103392, "lr": 0.0026291804804649314, "grad_norm": 0.165676, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:58.911604+00:00", "epoch": 0, "step": 10115, "train_loss": 3.551694631576538, "perplexity": 34.87236325097758, "lr": 0.0026291804804649314, "grad_norm": 0.172431, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:59.217839+00:00", "epoch": 0, "step": 10116, "train_loss": 3.5410659313201904, "perplexity": 34.503678147314396, "lr": 0.0026291804804649314, "grad_norm": 0.166135, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:59.523475+00:00", "epoch": 0, "step": 10117, "train_loss": 3.5734283924102783, "perplexity": 35.6385669375574, "lr": 0.0026291804804649314, "grad_norm": 0.170711, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:50:59.829665+00:00", "epoch": 0, "step": 10118, "train_loss": 3.5402727127075195, "perplexity": 34.47632003951696, "lr": 0.0026291804804649314, "grad_norm": 0.154375, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:00.135172+00:00", "epoch": 0, "step": 10119, "train_loss": 3.540742874145508, "perplexity": 34.49253328684646, "lr": 0.0026291804804649314, "grad_norm": 0.161898, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:00.441314+00:00", "epoch": 0, "step": 10120, "train_loss": 3.5163767337799072, "perplexity": 33.66223997980946, "lr": 0.0026291804804649314, "grad_norm": 0.150705, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:00.747510+00:00", "epoch": 0, "step": 10121, "train_loss": 3.6706011295318604, "perplexity": 39.27550843349007, "lr": 0.0026291804804649314, "grad_norm": 0.148832, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:01.053019+00:00", "epoch": 0, "step": 10122, "train_loss": 3.471597194671631, "perplexity": 32.18811209129189, "lr": 0.0026291804804649314, "grad_norm": 0.15115, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:01.357962+00:00", "epoch": 0, "step": 10123, "train_loss": 3.6417133808135986, "perplexity": 38.15715849214969, "lr": 0.0026291804804649314, "grad_norm": 0.160819, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:01.664441+00:00", "epoch": 0, "step": 10124, "train_loss": 3.6141867637634277, "perplexity": 37.12114538665313, "lr": 0.0026291804804649314, "grad_norm": 0.166845, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:01.970641+00:00", "epoch": 0, "step": 10125, "train_loss": 3.5493035316467285, "perplexity": 34.789079555155716, "lr": 0.0026291804804649314, "grad_norm": 0.152516, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:02.275777+00:00", "epoch": 0, "step": 10126, "train_loss": 3.59698224067688, "perplexity": 36.487956260674046, "lr": 0.0026291804804649314, "grad_norm": 0.155288, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:02.580840+00:00", "epoch": 0, "step": 10127, "train_loss": 3.4874107837677, "perplexity": 32.701167602613914, "lr": 0.0026291804804649314, "grad_norm": 0.170725, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:02.886727+00:00", "epoch": 0, "step": 10128, "train_loss": 3.528985023498535, "perplexity": 34.08935016041183, "lr": 0.0026291804804649314, "grad_norm": 0.159559, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:03.192462+00:00", "epoch": 0, "step": 10129, "train_loss": 3.5394070148468018, "perplexity": 34.44648687812849, "lr": 0.0026291804804649314, "grad_norm": 0.152361, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:03.498901+00:00", "epoch": 0, "step": 10130, "train_loss": 3.541416645050049, "perplexity": 34.515781183197646, "lr": 0.0026291804804649314, "grad_norm": 0.165062, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:03.802933+00:00", "epoch": 0, "step": 10131, "train_loss": 3.60491681098938, "perplexity": 36.77862415262381, "lr": 0.0026291804804649314, "grad_norm": 0.163889, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:04.107394+00:00", "epoch": 0, "step": 10132, "train_loss": 3.5094220638275146, "perplexity": 33.428942405041255, "lr": 0.0026291804804649314, "grad_norm": 0.159256, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:04.413040+00:00", "epoch": 0, "step": 10133, "train_loss": 3.4939005374908447, "perplexity": 32.91408025488509, "lr": 0.0026291804804649314, "grad_norm": 0.160974, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:04.719061+00:00", "epoch": 0, "step": 10134, "train_loss": 3.5626537799835205, "perplexity": 35.25663646270467, "lr": 0.0026291804804649314, "grad_norm": 0.171209, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:05.025216+00:00", "epoch": 0, "step": 10135, "train_loss": 3.622448205947876, "perplexity": 37.42908986436594, "lr": 0.0026291804804649314, "grad_norm": 0.167306, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:05.329959+00:00", "epoch": 0, "step": 10136, "train_loss": 3.5172863006591797, "perplexity": 33.69287196718595, "lr": 0.0026291804804649314, "grad_norm": 0.162667, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:05.635025+00:00", "epoch": 0, "step": 10137, "train_loss": 3.5810463428497314, "perplexity": 35.91109651464832, "lr": 0.0026291804804649314, "grad_norm": 0.158984, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:05.941475+00:00", "epoch": 0, "step": 10138, "train_loss": 3.4140703678131104, "perplexity": 30.388685991685662, "lr": 0.0026291804804649314, "grad_norm": 0.153825, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:06.247471+00:00", "epoch": 0, "step": 10139, "train_loss": 3.553842067718506, "perplexity": 34.94732988840294, "lr": 0.0026291804804649314, "grad_norm": 0.136694, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:06.554171+00:00", "epoch": 0, "step": 10140, "train_loss": 3.557560920715332, "perplexity": 35.0775358292052, "lr": 0.0026291804804649314, "grad_norm": 0.172979, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:06.859150+00:00", "epoch": 0, "step": 10141, "train_loss": 3.5001718997955322, "perplexity": 33.12114498741419, "lr": 0.0026291804804649314, "grad_norm": 0.166093, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:07.163290+00:00", "epoch": 0, "step": 10142, "train_loss": 3.547992706298828, "perplexity": 34.7435070231852, "lr": 0.0026291804804649314, "grad_norm": 0.148992, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:07.468875+00:00", "epoch": 0, "step": 10143, "train_loss": 3.5347096920013428, "perplexity": 34.28506004217361, "lr": 0.0026291804804649314, "grad_norm": 0.150101, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:07.773886+00:00", "epoch": 0, "step": 10144, "train_loss": 3.438624143600464, "perplexity": 31.144078896955612, "lr": 0.0026291804804649314, "grad_norm": 0.156455, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:08.079957+00:00", "epoch": 0, "step": 10145, "train_loss": 3.4308953285217285, "perplexity": 30.904299867841885, "lr": 0.0026291804804649314, "grad_norm": 0.158829, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:08.384531+00:00", "epoch": 0, "step": 10146, "train_loss": 3.5891475677490234, "perplexity": 36.203201992410484, "lr": 0.0026291804804649314, "grad_norm": 0.193873, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:08.689223+00:00", "epoch": 0, "step": 10147, "train_loss": 3.5649633407592773, "perplexity": 35.33815791049639, "lr": 0.0026291804804649314, "grad_norm": 0.173922, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:08.994327+00:00", "epoch": 0, "step": 10148, "train_loss": 3.5304481983184814, "perplexity": 34.1392653476122, "lr": 0.0026291804804649314, "grad_norm": 0.165348, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:09.300241+00:00", "epoch": 0, "step": 10149, "train_loss": 3.5472171306610107, "perplexity": 34.71657125227973, "lr": 0.0026291804804649314, "grad_norm": 0.17137, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:09.607011+00:00", "epoch": 0, "step": 10150, "train_loss": 3.538618803024292, "perplexity": 34.4193464475365, "lr": 0.0026291804804649314, "grad_norm": 0.168791, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:09.912665+00:00", "epoch": 0, "step": 10151, "train_loss": 3.5700836181640625, "perplexity": 35.519563108110994, "lr": 0.0026291804804649314, "grad_norm": 0.179179, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:10.217878+00:00", "epoch": 0, "step": 10152, "train_loss": 3.5619537830352783, "perplexity": 35.23196556056099, "lr": 0.0026291804804649314, "grad_norm": 0.16256, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:10.522616+00:00", "epoch": 0, "step": 10153, "train_loss": 3.6942081451416016, "perplexity": 40.21371655013257, "lr": 0.0026291804804649314, "grad_norm": 0.165848, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:10.827682+00:00", "epoch": 0, "step": 10154, "train_loss": 3.442728281021118, "perplexity": 31.2721611303478, "lr": 0.0026291804804649314, "grad_norm": 0.188366, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:11.133964+00:00", "epoch": 0, "step": 10155, "train_loss": 3.6124579906463623, "perplexity": 37.05702678765817, "lr": 0.0026291804804649314, "grad_norm": 0.166406, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:11.439769+00:00", "epoch": 0, "step": 10156, "train_loss": 3.523806571960449, "perplexity": 33.91327640082814, "lr": 0.0026291804804649314, "grad_norm": 0.179809, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:11.745515+00:00", "epoch": 0, "step": 10157, "train_loss": 3.6237947940826416, "perplexity": 37.47952538298662, "lr": 0.0026291804804649314, "grad_norm": 0.185958, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:12.050031+00:00", "epoch": 0, "step": 10158, "train_loss": 3.55899977684021, "perplexity": 35.12804368453328, "lr": 0.0026291804804649314, "grad_norm": 0.204547, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:12.354223+00:00", "epoch": 0, "step": 10159, "train_loss": 3.5385355949401855, "perplexity": 34.41648259881174, "lr": 0.0026291804804649314, "grad_norm": 0.194526, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:12.660102+00:00", "epoch": 0, "step": 10160, "train_loss": 3.4104762077331543, "perplexity": 30.279660234989308, "lr": 0.0026291804804649314, "grad_norm": 0.186381, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:12.966412+00:00", "epoch": 0, "step": 10161, "train_loss": 3.4082705974578857, "perplexity": 30.212948702108967, "lr": 0.0026291804804649314, "grad_norm": 0.158961, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:13.271236+00:00", "epoch": 0, "step": 10162, "train_loss": 3.4409306049346924, "perplexity": 31.215994414020198, "lr": 0.0026291804804649314, "grad_norm": 0.216654, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:13.575947+00:00", "epoch": 0, "step": 10163, "train_loss": 3.580366611480713, "perplexity": 35.88669491005904, "lr": 0.0026291804804649314, "grad_norm": 0.195537, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:13.881002+00:00", "epoch": 0, "step": 10164, "train_loss": 3.51261568069458, "perplexity": 33.535872295021115, "lr": 0.0026291804804649314, "grad_norm": 0.188155, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:14.186198+00:00", "epoch": 0, "step": 10165, "train_loss": 3.5051429271698, "perplexity": 33.28620101493622, "lr": 0.0026291804804649314, "grad_norm": 0.162991, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:14.491338+00:00", "epoch": 0, "step": 10166, "train_loss": 3.6216816902160645, "perplexity": 37.400410871011495, "lr": 0.0026291804804649314, "grad_norm": 0.171055, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:14.796844+00:00", "epoch": 0, "step": 10167, "train_loss": 3.5573976039886475, "perplexity": 35.07180754864817, "lr": 0.0026291804804649314, "grad_norm": 0.159346, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:15.103501+00:00", "epoch": 0, "step": 10168, "train_loss": 3.538569688796997, "perplexity": 34.41765600944435, "lr": 0.0026291804804649314, "grad_norm": 0.150946, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:15.408348+00:00", "epoch": 0, "step": 10169, "train_loss": 3.434326410293579, "perplexity": 31.010517163702083, "lr": 0.0026291804804649314, "grad_norm": 0.149277, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:15.713954+00:00", "epoch": 0, "step": 10170, "train_loss": 3.5993120670318604, "perplexity": 36.57306596973255, "lr": 0.0026291804804649314, "grad_norm": 0.13615, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:16.019342+00:00", "epoch": 0, "step": 10171, "train_loss": 3.4673478603363037, "perplexity": 32.051624238031856, "lr": 0.0026291804804649314, "grad_norm": 0.150738, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:16.323472+00:00", "epoch": 0, "step": 10172, "train_loss": 3.4307785034179688, "perplexity": 30.900689680687552, "lr": 0.0026291804804649314, "grad_norm": 0.157302, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:16.627798+00:00", "epoch": 0, "step": 10173, "train_loss": 3.5089616775512695, "perplexity": 33.41355572090947, "lr": 0.0026291804804649314, "grad_norm": 0.184766, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:16.933158+00:00", "epoch": 0, "step": 10174, "train_loss": 3.5457587242126465, "perplexity": 34.66597728315328, "lr": 0.0026291804804649314, "grad_norm": 0.150804, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:17.238833+00:00", "epoch": 0, "step": 10175, "train_loss": 3.5896220207214355, "perplexity": 36.2203827846233, "lr": 0.0026291804804649314, "grad_norm": 0.164569, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:17.545490+00:00", "epoch": 0, "step": 10176, "train_loss": 3.4765191078186035, "perplexity": 32.34692970600721, "lr": 0.0026291804804649314, "grad_norm": 0.146706, "tokens_per_sec": 106865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:17.851052+00:00", "epoch": 0, "step": 10177, "train_loss": 3.4942235946655273, "perplexity": 32.92471510239892, "lr": 0.0026291804804649314, "grad_norm": 0.163525, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:18.156373+00:00", "epoch": 0, "step": 10178, "train_loss": 3.519592761993408, "perplexity": 33.77067296161985, "lr": 0.0026291804804649314, "grad_norm": 0.166324, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:18.461927+00:00", "epoch": 0, "step": 10179, "train_loss": 3.4776790142059326, "perplexity": 32.384470884300995, "lr": 0.0026291804804649314, "grad_norm": 0.159801, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:18.767461+00:00", "epoch": 0, "step": 10180, "train_loss": 3.535360097885132, "perplexity": 34.307366500290556, "lr": 0.0026291804804649314, "grad_norm": 0.157939, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:19.072530+00:00", "epoch": 0, "step": 10181, "train_loss": 3.5984978675842285, "perplexity": 36.543300318854406, "lr": 0.0026291804804649314, "grad_norm": 0.160852, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:19.378908+00:00", "epoch": 0, "step": 10182, "train_loss": 3.561257839202881, "perplexity": 35.20745462153359, "lr": 0.0026291804804649314, "grad_norm": 0.167511, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:19.684918+00:00", "epoch": 0, "step": 10183, "train_loss": 3.508112907409668, "perplexity": 33.38520732482677, "lr": 0.0026291804804649314, "grad_norm": 0.204698, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:19.990415+00:00", "epoch": 0, "step": 10184, "train_loss": 3.5307445526123047, "perplexity": 34.14938416478934, "lr": 0.0026291804804649314, "grad_norm": 0.195487, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:20.295349+00:00", "epoch": 0, "step": 10185, "train_loss": 3.5525403022766113, "perplexity": 34.9018662599781, "lr": 0.0026291804804649314, "grad_norm": 0.184307, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:20.600659+00:00", "epoch": 0, "step": 10186, "train_loss": 3.4710311889648438, "perplexity": 32.16989859111628, "lr": 0.0026291804804649314, "grad_norm": 0.167867, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:20.907160+00:00", "epoch": 0, "step": 10187, "train_loss": 3.4885501861572266, "perplexity": 32.73844862613097, "lr": 0.0026291804804649314, "grad_norm": 0.208954, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:21.213000+00:00", "epoch": 0, "step": 10188, "train_loss": 3.6061971187591553, "perplexity": 36.82574226729594, "lr": 0.0026291804804649314, "grad_norm": 0.163221, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:21.518260+00:00", "epoch": 0, "step": 10189, "train_loss": 3.613708734512329, "perplexity": 37.10340463396122, "lr": 0.0026291804804649314, "grad_norm": 0.183618, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:21.823162+00:00", "epoch": 0, "step": 10190, "train_loss": 3.4920899868011475, "perplexity": 32.85454155935644, "lr": 0.0026291804804649314, "grad_norm": 0.224989, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:22.128433+00:00", "epoch": 0, "step": 10191, "train_loss": 3.543498992919922, "perplexity": 34.58772993178465, "lr": 0.0026291804804649314, "grad_norm": 0.170379, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:22.434163+00:00", "epoch": 0, "step": 10192, "train_loss": 3.478321075439453, "perplexity": 32.40527037418241, "lr": 0.0026291804804649314, "grad_norm": 0.171717, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:22.740779+00:00", "epoch": 0, "step": 10193, "train_loss": 3.518458604812622, "perplexity": 33.73239342199111, "lr": 0.0026291804804649314, "grad_norm": 0.146785, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:23.046729+00:00", "epoch": 0, "step": 10194, "train_loss": 3.5625412464141846, "perplexity": 35.25266913079394, "lr": 0.0026291804804649314, "grad_norm": 0.153442, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:23.351510+00:00", "epoch": 0, "step": 10195, "train_loss": 3.5710229873657227, "perplexity": 35.55294476814814, "lr": 0.0026291804804649314, "grad_norm": 0.169896, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:23.655966+00:00", "epoch": 0, "step": 10196, "train_loss": 3.488414764404297, "perplexity": 32.73401542821231, "lr": 0.0026291804804649314, "grad_norm": 0.168582, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:23.961300+00:00", "epoch": 0, "step": 10197, "train_loss": 3.5656850337982178, "perplexity": 35.3636704180616, "lr": 0.0026291804804649314, "grad_norm": 0.167154, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:24.267022+00:00", "epoch": 0, "step": 10198, "train_loss": 3.487781524658203, "perplexity": 32.713293510257394, "lr": 0.0026291804804649314, "grad_norm": 0.158429, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:24.573332+00:00", "epoch": 0, "step": 10199, "train_loss": 3.5983870029449463, "perplexity": 36.539249183614345, "lr": 0.0026291804804649314, "grad_norm": 0.165992, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:24.878279+00:00", "epoch": 0, "step": 10200, "train_loss": 3.5255722999572754, "perplexity": 33.97321092089125, "lr": 0.0026291804804649314, "grad_norm": 0.155612, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:25.182959+00:00", "epoch": 0, "step": 10201, "train_loss": 3.5376944541931152, "perplexity": 34.38754566465037, "lr": 0.0026291804804649314, "grad_norm": 0.198014, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:25.487295+00:00", "epoch": 0, "step": 10202, "train_loss": 3.479163408279419, "perplexity": 32.43257789700357, "lr": 0.0026291804804649314, "grad_norm": 0.164286, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:25.791857+00:00", "epoch": 0, "step": 10203, "train_loss": 3.622920274734497, "perplexity": 37.446763140575364, "lr": 0.0026291804804649314, "grad_norm": 0.161879, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:26.098431+00:00", "epoch": 0, "step": 10204, "train_loss": 3.4894540309906006, "perplexity": 32.76805248040437, "lr": 0.0026291804804649314, "grad_norm": 0.152555, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:26.403777+00:00", "epoch": 0, "step": 10205, "train_loss": 3.5407650470733643, "perplexity": 34.49329809577761, "lr": 0.0026291804804649314, "grad_norm": 0.182213, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:26.708818+00:00", "epoch": 0, "step": 10206, "train_loss": 3.4706060886383057, "perplexity": 32.15622606302448, "lr": 0.0026291804804649314, "grad_norm": 0.166695, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:27.014224+00:00", "epoch": 0, "step": 10207, "train_loss": 3.4994983673095703, "perplexity": 33.09884433124415, "lr": 0.0026291804804649314, "grad_norm": 0.189298, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:27.319288+00:00", "epoch": 0, "step": 10208, "train_loss": 3.4700870513916016, "perplexity": 32.139540114673274, "lr": 0.0026291804804649314, "grad_norm": 0.157748, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:27.624787+00:00", "epoch": 0, "step": 10209, "train_loss": 3.6225852966308594, "perplexity": 37.43422139559342, "lr": 0.0026291804804649314, "grad_norm": 0.186787, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:27.931082+00:00", "epoch": 0, "step": 10210, "train_loss": 3.611048698425293, "perplexity": 37.00483939035436, "lr": 0.0026291804804649314, "grad_norm": 0.226239, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:28.235811+00:00", "epoch": 0, "step": 10211, "train_loss": 3.485062837600708, "perplexity": 32.624477089315505, "lr": 0.0026291804804649314, "grad_norm": 0.179805, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:28.539145+00:00", "epoch": 0, "step": 10212, "train_loss": 3.602703094482422, "perplexity": 36.6972967563461, "lr": 0.0026291804804649314, "grad_norm": 0.199986, "tokens_per_sec": 108023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:28.842947+00:00", "epoch": 0, "step": 10213, "train_loss": 3.535616159439087, "perplexity": 34.31615242268919, "lr": 0.0026291804804649314, "grad_norm": 0.201002, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:29.147472+00:00", "epoch": 0, "step": 10214, "train_loss": 3.530350923538208, "perplexity": 34.13594461959097, "lr": 0.0026291804804649314, "grad_norm": 0.170387, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:29.453217+00:00", "epoch": 0, "step": 10215, "train_loss": 3.623676300048828, "perplexity": 37.47508454595012, "lr": 0.0026291804804649314, "grad_norm": 0.187534, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:29.759202+00:00", "epoch": 0, "step": 10216, "train_loss": 3.5467188358306885, "perplexity": 34.69927647360546, "lr": 0.0026291804804649314, "grad_norm": 0.157148, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:30.064469+00:00", "epoch": 0, "step": 10217, "train_loss": 3.4821484088897705, "perplexity": 32.529533796539454, "lr": 0.0026291804804649314, "grad_norm": 0.1695, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:30.369075+00:00", "epoch": 0, "step": 10218, "train_loss": 3.5353782176971436, "perplexity": 34.307988148954216, "lr": 0.0026291804804649314, "grad_norm": 0.145293, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:30.674086+00:00", "epoch": 0, "step": 10219, "train_loss": 3.4917409420013428, "perplexity": 32.84307585361407, "lr": 0.0026291804804649314, "grad_norm": 0.166459, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:30.979393+00:00", "epoch": 0, "step": 10220, "train_loss": 3.658912420272827, "perplexity": 38.81910103749969, "lr": 0.0026291804804649314, "grad_norm": 0.154138, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:31.285153+00:00", "epoch": 0, "step": 10221, "train_loss": 3.518127679824829, "perplexity": 33.72123237694591, "lr": 0.0026291804804649314, "grad_norm": 0.141535, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:31.590436+00:00", "epoch": 0, "step": 10222, "train_loss": 3.485384941101074, "perplexity": 32.63498724017091, "lr": 0.0026291804804649314, "grad_norm": 0.161234, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:31.894993+00:00", "epoch": 0, "step": 10223, "train_loss": 3.522519826889038, "perplexity": 33.86966672285016, "lr": 0.0026291804804649314, "grad_norm": 0.151629, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:32.199711+00:00", "epoch": 0, "step": 10224, "train_loss": 3.4815683364868164, "perplexity": 32.51066978347876, "lr": 0.0026291804804649314, "grad_norm": 0.164848, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:32.504694+00:00", "epoch": 0, "step": 10225, "train_loss": 3.6313092708587646, "perplexity": 37.76222524656189, "lr": 0.0026291804804649314, "grad_norm": 0.153855, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:32.810029+00:00", "epoch": 0, "step": 10226, "train_loss": 3.4984090328216553, "perplexity": 33.06280824984066, "lr": 0.0026291804804649314, "grad_norm": 0.157551, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:33.114526+00:00", "epoch": 0, "step": 10227, "train_loss": 3.4809060096740723, "perplexity": 32.48914422444524, "lr": 0.0026291804804649314, "grad_norm": 0.135148, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:33.419176+00:00", "epoch": 0, "step": 10228, "train_loss": 3.687640905380249, "perplexity": 39.95048871804949, "lr": 0.0026291804804649314, "grad_norm": 0.158746, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:33.723888+00:00", "epoch": 0, "step": 10229, "train_loss": 3.561600923538208, "perplexity": 35.219535820017455, "lr": 0.0026291804804649314, "grad_norm": 0.175582, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:34.029023+00:00", "epoch": 0, "step": 10230, "train_loss": 3.535609006881714, "perplexity": 34.31590697531795, "lr": 0.0026291804804649314, "grad_norm": 0.161666, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:34.333667+00:00", "epoch": 0, "step": 10231, "train_loss": 3.5713961124420166, "perplexity": 35.56621293856684, "lr": 0.0026291804804649314, "grad_norm": 0.151329, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:34.639055+00:00", "epoch": 0, "step": 10232, "train_loss": 3.4756596088409424, "perplexity": 32.319139497533115, "lr": 0.0026291804804649314, "grad_norm": 0.172206, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:34.944308+00:00", "epoch": 0, "step": 10233, "train_loss": 3.5996792316436768, "perplexity": 36.586496770808814, "lr": 0.0026291804804649314, "grad_norm": 0.171519, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:35.249493+00:00", "epoch": 0, "step": 10234, "train_loss": 3.508371353149414, "perplexity": 33.39383670449559, "lr": 0.0026291804804649314, "grad_norm": 0.148089, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:35.554953+00:00", "epoch": 0, "step": 10235, "train_loss": 3.531280279159546, "perplexity": 34.16768379782029, "lr": 0.0026291804804649314, "grad_norm": 0.198674, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:35.860060+00:00", "epoch": 0, "step": 10236, "train_loss": 3.5501253604888916, "perplexity": 34.81768197566367, "lr": 0.0026291804804649314, "grad_norm": 0.172786, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:36.165007+00:00", "epoch": 0, "step": 10237, "train_loss": 3.527470111846924, "perplexity": 34.03774690371432, "lr": 0.0026291804804649314, "grad_norm": 0.167522, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:36.469389+00:00", "epoch": 0, "step": 10238, "train_loss": 3.5294177532196045, "perplexity": 34.104104827559446, "lr": 0.0026291804804649314, "grad_norm": 0.16855, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:36.773171+00:00", "epoch": 0, "step": 10239, "train_loss": 3.5885088443756104, "perplexity": 36.18008554439926, "lr": 0.0026291804804649314, "grad_norm": 0.160112, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:37.077834+00:00", "epoch": 0, "step": 10240, "train_loss": 3.496922016143799, "perplexity": 33.013679838992324, "lr": 0.0026291804804649314, "grad_norm": 0.176907, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:37.382841+00:00", "epoch": 0, "step": 10241, "train_loss": 3.4661924839019775, "perplexity": 32.014613931240575, "lr": 0.0026291804804649314, "grad_norm": 0.154954, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:37.689156+00:00", "epoch": 0, "step": 10242, "train_loss": 3.548604965209961, "perplexity": 34.76478555828581, "lr": 0.0026291804804649314, "grad_norm": 0.172885, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:37.994617+00:00", "epoch": 0, "step": 10243, "train_loss": 3.6117868423461914, "perplexity": 37.032164371237755, "lr": 0.0026291804804649314, "grad_norm": 0.189743, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:38.299921+00:00", "epoch": 0, "step": 10244, "train_loss": 3.5800039768218994, "perplexity": 35.87368351002928, "lr": 0.0026291804804649314, "grad_norm": 0.142874, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:38.604879+00:00", "epoch": 0, "step": 10245, "train_loss": 3.565598726272583, "perplexity": 35.36061839887846, "lr": 0.0026291804804649314, "grad_norm": 0.179538, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:38.910947+00:00", "epoch": 0, "step": 10246, "train_loss": 3.612091302871704, "perplexity": 37.04344092000797, "lr": 0.0026291804804649314, "grad_norm": 0.167582, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:39.217804+00:00", "epoch": 0, "step": 10247, "train_loss": 3.5742597579956055, "perplexity": 35.668207935165086, "lr": 0.0026291804804649314, "grad_norm": 0.145824, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:39.523055+00:00", "epoch": 0, "step": 10248, "train_loss": 3.4989395141601562, "perplexity": 33.08035210554824, "lr": 0.0026291804804649314, "grad_norm": 0.156494, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:39.827618+00:00", "epoch": 0, "step": 10249, "train_loss": 3.4715659618377686, "perplexity": 32.187106781034025, "lr": 0.0026291804804649314, "grad_norm": 0.151835, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:40.132062+00:00", "epoch": 0, "step": 10250, "train_loss": 3.510113477706909, "perplexity": 33.45206363204112, "lr": 0.0026291804804649314, "grad_norm": 0.151568, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:40.436446+00:00", "epoch": 0, "step": 10251, "train_loss": 3.4507012367248535, "perplexity": 31.5224892873831, "lr": 0.0026291804804649314, "grad_norm": 0.15934, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:40.742256+00:00", "epoch": 0, "step": 10252, "train_loss": 3.5828399658203125, "perplexity": 35.97556528130268, "lr": 0.0026291804804649314, "grad_norm": 0.144071, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:41.047606+00:00", "epoch": 0, "step": 10253, "train_loss": 3.646984100341797, "perplexity": 38.358805117137024, "lr": 0.0026291804804649314, "grad_norm": 0.151634, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:41.353519+00:00", "epoch": 0, "step": 10254, "train_loss": 3.58575439453125, "perplexity": 36.08056643651905, "lr": 0.0026291804804649314, "grad_norm": 0.167927, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:41.658869+00:00", "epoch": 0, "step": 10255, "train_loss": 3.571451187133789, "perplexity": 35.56817179072304, "lr": 0.0026291804804649314, "grad_norm": 0.160679, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:41.963865+00:00", "epoch": 0, "step": 10256, "train_loss": 3.66469669342041, "perplexity": 39.044291976328424, "lr": 0.0026291804804649314, "grad_norm": 0.147674, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:42.269026+00:00", "epoch": 0, "step": 10257, "train_loss": 3.5642452239990234, "perplexity": 35.312790096640676, "lr": 0.0026291804804649314, "grad_norm": 0.151786, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:42.575039+00:00", "epoch": 0, "step": 10258, "train_loss": 3.5256872177124023, "perplexity": 33.977115270359974, "lr": 0.0026291804804649314, "grad_norm": 0.178335, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:42.879384+00:00", "epoch": 0, "step": 10259, "train_loss": 3.6456522941589355, "perplexity": 38.30775262687195, "lr": 0.0026291804804649314, "grad_norm": 0.164305, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:43.184645+00:00", "epoch": 0, "step": 10260, "train_loss": 3.6150143146514893, "perplexity": 37.15187773802215, "lr": 0.0026291804804649314, "grad_norm": 0.16308, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:43.489408+00:00", "epoch": 0, "step": 10261, "train_loss": 3.5610361099243164, "perplexity": 35.19964896342372, "lr": 0.0026291804804649314, "grad_norm": 0.158158, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:43.795027+00:00", "epoch": 0, "step": 10262, "train_loss": 3.5757155418395996, "perplexity": 35.72017095030315, "lr": 0.0026291804804649314, "grad_norm": 0.153373, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:44.101006+00:00", "epoch": 0, "step": 10263, "train_loss": 3.5367431640625, "perplexity": 34.354848686465594, "lr": 0.0026291804804649314, "grad_norm": 0.16545, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:44.406579+00:00", "epoch": 0, "step": 10264, "train_loss": 3.5856947898864746, "perplexity": 36.07841593126399, "lr": 0.0026291804804649314, "grad_norm": 0.171092, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:44.710330+00:00", "epoch": 0, "step": 10265, "train_loss": 3.576676845550537, "perplexity": 35.75452539307407, "lr": 0.0026291804804649314, "grad_norm": 0.155124, "tokens_per_sec": 107877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:45.014425+00:00", "epoch": 0, "step": 10266, "train_loss": 3.521836280822754, "perplexity": 33.84652315614352, "lr": 0.0026291804804649314, "grad_norm": 0.160016, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:45.317553+00:00", "epoch": 0, "step": 10267, "train_loss": 3.652722120285034, "perplexity": 38.579541394837236, "lr": 0.0026291804804649314, "grad_norm": 0.16562, "tokens_per_sec": 108104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:45.621961+00:00", "epoch": 0, "step": 10268, "train_loss": 3.441823720932007, "perplexity": 31.243886371526145, "lr": 0.0026291804804649314, "grad_norm": 0.170471, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:45.927089+00:00", "epoch": 0, "step": 10269, "train_loss": 3.537010908126831, "perplexity": 34.36404822478832, "lr": 0.0026291804804649314, "grad_norm": 0.168458, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:46.233671+00:00", "epoch": 0, "step": 10270, "train_loss": 3.518639087677002, "perplexity": 33.73848209041193, "lr": 0.0026291804804649314, "grad_norm": 0.184091, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:46.539009+00:00", "epoch": 0, "step": 10271, "train_loss": 3.465561866760254, "perplexity": 31.99443133132545, "lr": 0.0026291804804649314, "grad_norm": 0.173802, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:46.843900+00:00", "epoch": 0, "step": 10272, "train_loss": 3.5258989334106445, "perplexity": 33.984309520584674, "lr": 0.0026291804804649314, "grad_norm": 0.154618, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:47.148685+00:00", "epoch": 0, "step": 10273, "train_loss": 3.5324366092681885, "perplexity": 34.20721577093883, "lr": 0.0026291804804649314, "grad_norm": 0.17086, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:47.454396+00:00", "epoch": 0, "step": 10274, "train_loss": 3.5636985301971436, "perplexity": 35.29349008924566, "lr": 0.0026291804804649314, "grad_norm": 0.171096, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:47.761229+00:00", "epoch": 0, "step": 10275, "train_loss": 3.52667498588562, "perplexity": 34.01069336439795, "lr": 0.0026291804804649314, "grad_norm": 0.146153, "tokens_per_sec": 106794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:48.065598+00:00", "epoch": 0, "step": 10276, "train_loss": 3.4345216751098633, "perplexity": 31.016573017867316, "lr": 0.0026291804804649314, "grad_norm": 0.158391, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:48.369404+00:00", "epoch": 0, "step": 10277, "train_loss": 3.578533172607422, "perplexity": 35.820959128260085, "lr": 0.0026291804804649314, "grad_norm": 0.163057, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:48.674019+00:00", "epoch": 0, "step": 10278, "train_loss": 3.4696154594421387, "perplexity": 32.12438693963427, "lr": 0.0026291804804649314, "grad_norm": 0.167986, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:48.978974+00:00", "epoch": 0, "step": 10279, "train_loss": 3.5431835651397705, "perplexity": 34.57682172137853, "lr": 0.0026291804804649314, "grad_norm": 0.16115, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:49.284625+00:00", "epoch": 0, "step": 10280, "train_loss": 3.4882898330688477, "perplexity": 32.729926179393225, "lr": 0.0026291804804649314, "grad_norm": 0.167154, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:49.589631+00:00", "epoch": 0, "step": 10281, "train_loss": 3.4365181922912598, "perplexity": 31.07855999724978, "lr": 0.0026291804804649314, "grad_norm": 0.190183, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:49.893411+00:00", "epoch": 0, "step": 10282, "train_loss": 3.531270742416382, "perplexity": 34.16735795094916, "lr": 0.0026291804804649314, "grad_norm": 0.201028, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:50.197542+00:00", "epoch": 0, "step": 10283, "train_loss": 3.5181703567504883, "perplexity": 33.72267152618221, "lr": 0.0026291804804649314, "grad_norm": 0.172111, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:50.502748+00:00", "epoch": 0, "step": 10284, "train_loss": 3.622145652770996, "perplexity": 37.41776728724697, "lr": 0.0026291804804649314, "grad_norm": 0.166403, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:50.808986+00:00", "epoch": 0, "step": 10285, "train_loss": 3.4926645755767822, "perplexity": 32.87342483470457, "lr": 0.0026291804804649314, "grad_norm": 0.170786, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:51.114748+00:00", "epoch": 0, "step": 10286, "train_loss": 3.482697010040283, "perplexity": 32.54738443219781, "lr": 0.0026291804804649314, "grad_norm": 0.156842, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:51.419697+00:00", "epoch": 0, "step": 10287, "train_loss": 3.570600748062134, "perplexity": 35.537936086359366, "lr": 0.0026291804804649314, "grad_norm": 0.152455, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:51.724189+00:00", "epoch": 0, "step": 10288, "train_loss": 3.4871442317962646, "perplexity": 32.692452203526194, "lr": 0.0026291804804649314, "grad_norm": 0.162356, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:52.028947+00:00", "epoch": 0, "step": 10289, "train_loss": 3.5656960010528564, "perplexity": 35.36405826256682, "lr": 0.0026291804804649314, "grad_norm": 0.172449, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:52.334878+00:00", "epoch": 0, "step": 10290, "train_loss": 3.541597604751587, "perplexity": 34.52202771382703, "lr": 0.0026291804804649314, "grad_norm": 0.176366, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:52.639907+00:00", "epoch": 0, "step": 10291, "train_loss": 3.5282304286956787, "perplexity": 34.06363621697223, "lr": 0.0026291804804649314, "grad_norm": 0.186072, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:52.944516+00:00", "epoch": 0, "step": 10292, "train_loss": 3.543774127960205, "perplexity": 34.597247537506114, "lr": 0.0026291804804649314, "grad_norm": 0.147605, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:53.250374+00:00", "epoch": 0, "step": 10293, "train_loss": 3.571147918701172, "perplexity": 35.55738672248104, "lr": 0.0026291804804649314, "grad_norm": 0.141946, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:53.555710+00:00", "epoch": 0, "step": 10294, "train_loss": 3.520995616912842, "perplexity": 33.818081562235925, "lr": 0.0026291804804649314, "grad_norm": 0.149763, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:53.861063+00:00", "epoch": 0, "step": 10295, "train_loss": 3.4565680027008057, "perplexity": 31.707967902828962, "lr": 0.0026291804804649314, "grad_norm": 0.166126, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:54.166794+00:00", "epoch": 0, "step": 10296, "train_loss": 3.4084362983703613, "perplexity": 30.217955430075722, "lr": 0.0026291804804649314, "grad_norm": 0.159862, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:54.471862+00:00", "epoch": 0, "step": 10297, "train_loss": 3.585653066635132, "perplexity": 36.076910653850796, "lr": 0.0026291804804649314, "grad_norm": 0.209922, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:54.776938+00:00", "epoch": 0, "step": 10298, "train_loss": 3.60632586479187, "perplexity": 36.830483740731026, "lr": 0.0026291804804649314, "grad_norm": 0.200261, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:55.082883+00:00", "epoch": 0, "step": 10299, "train_loss": 3.4891417026519775, "perplexity": 32.75781968709211, "lr": 0.0026291804804649314, "grad_norm": 0.169774, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:55.388521+00:00", "epoch": 0, "step": 10300, "train_loss": 3.4791743755340576, "perplexity": 32.432933595294465, "lr": 0.0026291804804649314, "grad_norm": 0.156038, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:55.694961+00:00", "epoch": 0, "step": 10301, "train_loss": 3.588522434234619, "perplexity": 36.180577230001695, "lr": 0.0026291804804649314, "grad_norm": 0.148158, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:56.000202+00:00", "epoch": 0, "step": 10302, "train_loss": 3.558187484741211, "perplexity": 35.09952103812352, "lr": 0.0026291804804649314, "grad_norm": 0.186658, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:56.304797+00:00", "epoch": 0, "step": 10303, "train_loss": 3.6451337337493896, "perplexity": 38.287892892661965, "lr": 0.0026291804804649314, "grad_norm": 0.198752, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:56.609038+00:00", "epoch": 0, "step": 10304, "train_loss": 3.5217437744140625, "perplexity": 33.84339228065491, "lr": 0.0026291804804649314, "grad_norm": 0.182294, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:56.914202+00:00", "epoch": 0, "step": 10305, "train_loss": 3.5206425189971924, "perplexity": 33.806142576064644, "lr": 0.0026291804804649314, "grad_norm": 0.160784, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:57.219850+00:00", "epoch": 0, "step": 10306, "train_loss": 3.4478189945220947, "perplexity": 31.431764646424035, "lr": 0.0026291804804649314, "grad_norm": 0.158905, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:57.525237+00:00", "epoch": 0, "step": 10307, "train_loss": 3.553889513015747, "perplexity": 34.94898801419211, "lr": 0.0026291804804649314, "grad_norm": 0.174386, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:57.829618+00:00", "epoch": 0, "step": 10308, "train_loss": 3.5545153617858887, "perplexity": 34.970867641313184, "lr": 0.0026291804804649314, "grad_norm": 0.158604, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:58.133925+00:00", "epoch": 0, "step": 10309, "train_loss": 3.495018243789673, "perplexity": 32.950889096611235, "lr": 0.0026291804804649314, "grad_norm": 0.161948, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:58.439078+00:00", "epoch": 0, "step": 10310, "train_loss": 3.5744707584381104, "perplexity": 35.67573473687408, "lr": 0.0026291804804649314, "grad_norm": 0.164904, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:58.744643+00:00", "epoch": 0, "step": 10311, "train_loss": 3.5502095222473145, "perplexity": 34.820612416316735, "lr": 0.0026291804804649314, "grad_norm": 0.173341, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:59.049952+00:00", "epoch": 0, "step": 10312, "train_loss": 3.581547737121582, "perplexity": 35.929106647448044, "lr": 0.0026291804804649314, "grad_norm": 0.190404, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:59.355477+00:00", "epoch": 0, "step": 10313, "train_loss": 3.548997163772583, "perplexity": 34.77842293131587, "lr": 0.0026291804804649314, "grad_norm": 0.161457, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:59.659822+00:00", "epoch": 0, "step": 10314, "train_loss": 3.4874279499053955, "perplexity": 32.70172896017793, "lr": 0.0026291804804649314, "grad_norm": 0.152498, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:51:59.963539+00:00", "epoch": 0, "step": 10315, "train_loss": 3.4840848445892334, "perplexity": 32.592586175806304, "lr": 0.0026291804804649314, "grad_norm": 0.178534, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:00.268687+00:00", "epoch": 0, "step": 10316, "train_loss": 3.510331392288208, "perplexity": 33.45935411880391, "lr": 0.0026291804804649314, "grad_norm": 0.163081, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:00.574492+00:00", "epoch": 0, "step": 10317, "train_loss": 3.509631872177124, "perplexity": 33.4359568120913, "lr": 0.0026291804804649314, "grad_norm": 0.151264, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:00.880602+00:00", "epoch": 0, "step": 10318, "train_loss": 3.5970473289489746, "perplexity": 36.4903312759913, "lr": 0.0026291804804649314, "grad_norm": 0.155463, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:01.186079+00:00", "epoch": 0, "step": 10319, "train_loss": 3.465332269668579, "perplexity": 31.98708634616787, "lr": 0.0026291804804649314, "grad_norm": 0.154226, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:01.490459+00:00", "epoch": 0, "step": 10320, "train_loss": 3.567533254623413, "perplexity": 35.429090727149884, "lr": 0.0026291804804649314, "grad_norm": 0.135246, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:01.795212+00:00", "epoch": 0, "step": 10321, "train_loss": 3.4820237159729004, "perplexity": 32.52547784696528, "lr": 0.0026291804804649314, "grad_norm": 0.146978, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:02.099417+00:00", "epoch": 0, "step": 10322, "train_loss": 3.654557704925537, "perplexity": 38.65042244263852, "lr": 0.0026291804804649314, "grad_norm": 0.150834, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:02.404823+00:00", "epoch": 0, "step": 10323, "train_loss": 3.6553783416748047, "perplexity": 38.682153417691026, "lr": 0.0026291804804649314, "grad_norm": 0.147338, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:02.710848+00:00", "epoch": 0, "step": 10324, "train_loss": 3.560661792755127, "perplexity": 35.18647559612986, "lr": 0.0026291804804649314, "grad_norm": 0.169301, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:03.015555+00:00", "epoch": 0, "step": 10325, "train_loss": 3.5130674839019775, "perplexity": 33.55102733297645, "lr": 0.0026291804804649314, "grad_norm": 0.158033, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:03.320392+00:00", "epoch": 0, "step": 10326, "train_loss": 3.4721250534057617, "perplexity": 32.20510735254839, "lr": 0.0026291804804649314, "grad_norm": 0.140002, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:03.626675+00:00", "epoch": 0, "step": 10327, "train_loss": 3.5999526977539062, "perplexity": 36.59650330592941, "lr": 0.0026291804804649314, "grad_norm": 0.163807, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:03.932877+00:00", "epoch": 0, "step": 10328, "train_loss": 3.5945651531219482, "perplexity": 36.39986817688352, "lr": 0.0026291804804649314, "grad_norm": 0.150849, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:04.237971+00:00", "epoch": 0, "step": 10329, "train_loss": 3.593437433242798, "perplexity": 36.35884245904865, "lr": 0.0026291804804649314, "grad_norm": 0.159202, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:04.543237+00:00", "epoch": 0, "step": 10330, "train_loss": 3.558546304702759, "perplexity": 35.112117706745686, "lr": 0.0026291804804649314, "grad_norm": 0.185317, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:04.847941+00:00", "epoch": 0, "step": 10331, "train_loss": 3.5063090324401855, "perplexity": 33.325038869483315, "lr": 0.0026291804804649314, "grad_norm": 0.180965, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:05.153961+00:00", "epoch": 0, "step": 10332, "train_loss": 3.5293288230895996, "perplexity": 34.10107207993675, "lr": 0.0026291804804649314, "grad_norm": 0.173521, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:05.459885+00:00", "epoch": 0, "step": 10333, "train_loss": 3.5735092163085938, "perplexity": 35.64144750187532, "lr": 0.0026291804804649314, "grad_norm": 0.163667, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:05.764024+00:00", "epoch": 0, "step": 10334, "train_loss": 3.639404773712158, "perplexity": 38.069170209350204, "lr": 0.0026291804804649314, "grad_norm": 0.163911, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:06.069874+00:00", "epoch": 0, "step": 10335, "train_loss": 3.550507068634033, "perplexity": 34.83097470527891, "lr": 0.0026291804804649314, "grad_norm": 0.14897, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:06.374448+00:00", "epoch": 0, "step": 10336, "train_loss": 3.5592169761657715, "perplexity": 35.13567430058219, "lr": 0.0026291804804649314, "grad_norm": 0.155887, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:06.679992+00:00", "epoch": 0, "step": 10337, "train_loss": 3.568567991256714, "perplexity": 35.465769478360976, "lr": 0.0026291804804649314, "grad_norm": 0.170035, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:06.985258+00:00", "epoch": 0, "step": 10338, "train_loss": 3.4788520336151123, "perplexity": 32.42248078602214, "lr": 0.0026291804804649314, "grad_norm": 0.159936, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:07.289543+00:00", "epoch": 0, "step": 10339, "train_loss": 3.6299045085906982, "perplexity": 37.70921553911298, "lr": 0.0026291804804649314, "grad_norm": 0.175829, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:07.593612+00:00", "epoch": 0, "step": 10340, "train_loss": 3.5530900955200195, "perplexity": 34.92106034613613, "lr": 0.0026291804804649314, "grad_norm": 0.188015, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:07.898927+00:00", "epoch": 0, "step": 10341, "train_loss": 3.6642518043518066, "perplexity": 39.02692546100964, "lr": 0.0026291804804649314, "grad_norm": 0.17578, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:08.205199+00:00", "epoch": 0, "step": 10342, "train_loss": 3.5111196041107178, "perplexity": 33.48573757380351, "lr": 0.0026291804804649314, "grad_norm": 0.219003, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:08.510393+00:00", "epoch": 0, "step": 10343, "train_loss": 3.5385541915893555, "perplexity": 34.41712263601555, "lr": 0.0026291804804649314, "grad_norm": 0.183496, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:08.814888+00:00", "epoch": 0, "step": 10344, "train_loss": 3.559746503829956, "perplexity": 35.15428453900731, "lr": 0.0026291804804649314, "grad_norm": 0.187467, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:09.119384+00:00", "epoch": 0, "step": 10345, "train_loss": 3.496580123901367, "perplexity": 33.002394647230936, "lr": 0.0026291804804649314, "grad_norm": 0.165555, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:09.424109+00:00", "epoch": 0, "step": 10346, "train_loss": 3.5776009559631348, "perplexity": 35.787581793816905, "lr": 0.0026291804804649314, "grad_norm": 0.184759, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:09.730014+00:00", "epoch": 0, "step": 10347, "train_loss": 3.4956307411193848, "perplexity": 32.97107761026874, "lr": 0.0026291804804649314, "grad_norm": 0.153152, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:10.035789+00:00", "epoch": 0, "step": 10348, "train_loss": 3.6376476287841797, "perplexity": 38.0023358959881, "lr": 0.0026291804804649314, "grad_norm": 0.175819, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:10.340605+00:00", "epoch": 0, "step": 10349, "train_loss": 3.5570921897888184, "perplexity": 35.061097756153735, "lr": 0.0026291804804649314, "grad_norm": 0.18789, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:10.646393+00:00", "epoch": 0, "step": 10350, "train_loss": 3.5475103855133057, "perplexity": 34.726753548185364, "lr": 0.0026291804804649314, "grad_norm": 0.15066, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:10.951381+00:00", "epoch": 0, "step": 10351, "train_loss": 3.669976234436035, "perplexity": 39.250973027709726, "lr": 0.0026291804804649314, "grad_norm": 0.139932, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:11.256596+00:00", "epoch": 0, "step": 10352, "train_loss": 3.518371105194092, "perplexity": 33.7294419795615, "lr": 0.0026291804804649314, "grad_norm": 0.148662, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:11.562504+00:00", "epoch": 0, "step": 10353, "train_loss": 3.5626816749572754, "perplexity": 35.25761995937073, "lr": 0.0026291804804649314, "grad_norm": 0.165493, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:11.867113+00:00", "epoch": 0, "step": 10354, "train_loss": 3.4694697856903076, "perplexity": 32.11970760050022, "lr": 0.0026291804804649314, "grad_norm": 0.164846, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:12.172048+00:00", "epoch": 0, "step": 10355, "train_loss": 3.5169503688812256, "perplexity": 33.68155536171176, "lr": 0.0026291804804649314, "grad_norm": 0.144029, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:12.477882+00:00", "epoch": 0, "step": 10356, "train_loss": 3.5574722290039062, "perplexity": 35.07442488047963, "lr": 0.0026291804804649314, "grad_norm": 0.156716, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:12.783688+00:00", "epoch": 0, "step": 10357, "train_loss": 3.6211307048797607, "perplexity": 37.3798094691062, "lr": 0.0026291804804649314, "grad_norm": 0.158147, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:13.088192+00:00", "epoch": 0, "step": 10358, "train_loss": 3.534545421600342, "perplexity": 34.279428484173955, "lr": 0.0026291804804649314, "grad_norm": 0.154713, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:13.392907+00:00", "epoch": 0, "step": 10359, "train_loss": 3.681344747543335, "perplexity": 39.69974432687584, "lr": 0.0026291804804649314, "grad_norm": 0.160094, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:13.697611+00:00", "epoch": 0, "step": 10360, "train_loss": 3.4876868724823, "perplexity": 32.710197272380825, "lr": 0.0026291804804649314, "grad_norm": 0.167303, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:14.002045+00:00", "epoch": 0, "step": 10361, "train_loss": 3.542625904083252, "perplexity": 34.55754494989827, "lr": 0.0026291804804649314, "grad_norm": 0.164367, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:14.305927+00:00", "epoch": 0, "step": 10362, "train_loss": 3.5115303993225098, "perplexity": 33.499496180253495, "lr": 0.0026291804804649314, "grad_norm": 0.151745, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:14.610373+00:00", "epoch": 0, "step": 10363, "train_loss": 3.4680469036102295, "perplexity": 32.074037543396315, "lr": 0.0026291804804649314, "grad_norm": 0.146944, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:14.915044+00:00", "epoch": 0, "step": 10364, "train_loss": 3.529550313949585, "perplexity": 34.10862599224854, "lr": 0.0026291804804649314, "grad_norm": 0.163531, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:15.220868+00:00", "epoch": 0, "step": 10365, "train_loss": 3.5087265968322754, "perplexity": 33.40570176139984, "lr": 0.0026291804804649314, "grad_norm": 0.154713, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:15.526450+00:00", "epoch": 0, "step": 10366, "train_loss": 3.6052048206329346, "perplexity": 36.789218276588144, "lr": 0.0026291804804649314, "grad_norm": 0.177937, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:15.832133+00:00", "epoch": 0, "step": 10367, "train_loss": 3.536102294921875, "perplexity": 34.33283877760061, "lr": 0.0026291804804649314, "grad_norm": 0.193321, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:16.136790+00:00", "epoch": 0, "step": 10368, "train_loss": 3.7281980514526367, "perplexity": 41.60407218141126, "lr": 0.0026291804804649314, "grad_norm": 0.169489, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:16.441611+00:00", "epoch": 0, "step": 10369, "train_loss": 3.4906299114227295, "perplexity": 32.806606455104905, "lr": 0.0026291804804649314, "grad_norm": 0.177945, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:16.745932+00:00", "epoch": 0, "step": 10370, "train_loss": 3.539367914199829, "perplexity": 34.44514002453722, "lr": 0.0026291804804649314, "grad_norm": 0.149407, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:17.050062+00:00", "epoch": 0, "step": 10371, "train_loss": 3.5993282794952393, "perplexity": 36.573658914031775, "lr": 0.0026291804804649314, "grad_norm": 0.171712, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:17.355331+00:00", "epoch": 0, "step": 10372, "train_loss": 3.568441867828369, "perplexity": 35.46129669599273, "lr": 0.0026291804804649314, "grad_norm": 0.164394, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:17.660879+00:00", "epoch": 0, "step": 10373, "train_loss": 3.5395147800445557, "perplexity": 34.45019921062537, "lr": 0.0026291804804649314, "grad_norm": 0.170487, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:17.965987+00:00", "epoch": 0, "step": 10374, "train_loss": 3.634160041809082, "perplexity": 37.8700302920501, "lr": 0.0026291804804649314, "grad_norm": 0.153957, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:18.270345+00:00", "epoch": 0, "step": 10375, "train_loss": 3.556236505508423, "perplexity": 35.03110935808058, "lr": 0.0026291804804649314, "grad_norm": 0.181087, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:18.575460+00:00", "epoch": 0, "step": 10376, "train_loss": 3.5286049842834473, "perplexity": 34.07639733198129, "lr": 0.0026291804804649314, "grad_norm": 0.183323, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:18.881120+00:00", "epoch": 0, "step": 10377, "train_loss": 3.4935731887817383, "perplexity": 32.90330763649979, "lr": 0.0026291804804649314, "grad_norm": 0.190952, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:19.185603+00:00", "epoch": 0, "step": 10378, "train_loss": 3.5949137210845947, "perplexity": 36.41255821631665, "lr": 0.0026291804804649314, "grad_norm": 0.174659, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:19.490320+00:00", "epoch": 0, "step": 10379, "train_loss": 3.4011270999908447, "perplexity": 29.997891623951645, "lr": 0.0026291804804649314, "grad_norm": 0.160941, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:19.795413+00:00", "epoch": 0, "step": 10380, "train_loss": 3.581444501876831, "perplexity": 35.925397688780535, "lr": 0.0026291804804649314, "grad_norm": 0.180563, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:20.099483+00:00", "epoch": 0, "step": 10381, "train_loss": 3.4501893520355225, "perplexity": 31.50635753689757, "lr": 0.0026291804804649314, "grad_norm": 0.141655, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:20.405521+00:00", "epoch": 0, "step": 10382, "train_loss": 3.4649388790130615, "perplexity": 31.974505400077618, "lr": 0.0026291804804649314, "grad_norm": 0.183398, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:20.712400+00:00", "epoch": 0, "step": 10383, "train_loss": 3.545858144760132, "perplexity": 34.669423964926544, "lr": 0.0026291804804649314, "grad_norm": 0.142937, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:21.018220+00:00", "epoch": 0, "step": 10384, "train_loss": 3.531015396118164, "perplexity": 34.15863455636716, "lr": 0.0026291804804649314, "grad_norm": 0.150927, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:21.324131+00:00", "epoch": 0, "step": 10385, "train_loss": 3.5415241718292236, "perplexity": 34.519492753522, "lr": 0.0026291804804649314, "grad_norm": 0.173243, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:21.629093+00:00", "epoch": 0, "step": 10386, "train_loss": 3.5764803886413574, "perplexity": 35.747501859459604, "lr": 0.0026291804804649314, "grad_norm": 0.18842, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:21.934542+00:00", "epoch": 0, "step": 10387, "train_loss": 3.5340890884399414, "perplexity": 34.26378921286749, "lr": 0.0026291804804649314, "grad_norm": 0.156905, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:22.241240+00:00", "epoch": 0, "step": 10388, "train_loss": 3.4516851902008057, "perplexity": 31.553521214770825, "lr": 0.0026291804804649314, "grad_norm": 0.159843, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:22.547199+00:00", "epoch": 0, "step": 10389, "train_loss": 3.5802483558654785, "perplexity": 35.882451357790615, "lr": 0.0026291804804649314, "grad_norm": 0.158194, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:22.852160+00:00", "epoch": 0, "step": 10390, "train_loss": 3.588620901107788, "perplexity": 36.184139993715206, "lr": 0.0026291804804649314, "grad_norm": 0.181269, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:23.155807+00:00", "epoch": 0, "step": 10391, "train_loss": 3.6302034854888916, "perplexity": 37.72049140893647, "lr": 0.0026291804804649314, "grad_norm": 0.150181, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:23.461719+00:00", "epoch": 0, "step": 10392, "train_loss": 3.4552674293518066, "perplexity": 31.666756170073377, "lr": 0.0026291804804649314, "grad_norm": 0.210803, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:23.766787+00:00", "epoch": 0, "step": 10393, "train_loss": 3.611065149307251, "perplexity": 37.00544815760641, "lr": 0.0026291804804649314, "grad_norm": 0.193301, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:24.072239+00:00", "epoch": 0, "step": 10394, "train_loss": 3.4818108081817627, "perplexity": 32.518553656454394, "lr": 0.0026291804804649314, "grad_norm": 0.1895, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:24.378290+00:00", "epoch": 0, "step": 10395, "train_loss": 3.522709608078003, "perplexity": 33.87609515844945, "lr": 0.0026291804804649314, "grad_norm": 0.160985, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:24.683810+00:00", "epoch": 0, "step": 10396, "train_loss": 3.586229085922241, "perplexity": 36.09769763648344, "lr": 0.0026291804804649314, "grad_norm": 0.155759, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:24.989116+00:00", "epoch": 0, "step": 10397, "train_loss": 3.565481185913086, "perplexity": 35.35646234333674, "lr": 0.0026291804804649314, "grad_norm": 0.181599, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:25.294101+00:00", "epoch": 0, "step": 10398, "train_loss": 3.5418787002563477, "perplexity": 34.53173306463395, "lr": 0.0026291804804649314, "grad_norm": 0.17735, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:25.600733+00:00", "epoch": 0, "step": 10399, "train_loss": 3.5252344608306885, "perplexity": 33.961735379538965, "lr": 0.0026291804804649314, "grad_norm": 0.165647, "tokens_per_sec": 106863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:25.906196+00:00", "epoch": 0, "step": 10400, "train_loss": 3.6190640926361084, "perplexity": 37.30263966468856, "lr": 0.0026291804804649314, "grad_norm": 0.153078, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:26.211884+00:00", "epoch": 0, "step": 10401, "train_loss": 3.4232566356658936, "perplexity": 30.669130749062035, "lr": 0.0026291804804649314, "grad_norm": 0.163989, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:26.516467+00:00", "epoch": 0, "step": 10402, "train_loss": 3.5526554584503174, "perplexity": 34.90588565677653, "lr": 0.0026291804804649314, "grad_norm": 0.159704, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:26.821004+00:00", "epoch": 0, "step": 10403, "train_loss": 3.595430850982666, "perplexity": 36.43139310845875, "lr": 0.0026291804804649314, "grad_norm": 0.183571, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:27.126295+00:00", "epoch": 0, "step": 10404, "train_loss": 3.5192296504974365, "perplexity": 33.75841266810226, "lr": 0.0026291804804649314, "grad_norm": 0.162875, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:27.432305+00:00", "epoch": 0, "step": 10405, "train_loss": 3.4836275577545166, "perplexity": 32.577685422457556, "lr": 0.0026291804804649314, "grad_norm": 0.160779, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:27.737117+00:00", "epoch": 0, "step": 10406, "train_loss": 3.6234686374664307, "perplexity": 37.46730318109487, "lr": 0.0026291804804649314, "grad_norm": 0.153915, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:28.042319+00:00", "epoch": 0, "step": 10407, "train_loss": 3.4977076053619385, "perplexity": 33.03962521979631, "lr": 0.0026291804804649314, "grad_norm": 0.164951, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:28.347192+00:00", "epoch": 0, "step": 10408, "train_loss": 3.6098976135253906, "perplexity": 36.96226818474534, "lr": 0.0026291804804649314, "grad_norm": 0.13833, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:28.652953+00:00", "epoch": 0, "step": 10409, "train_loss": 3.481215238571167, "perplexity": 32.49919236018822, "lr": 0.0026291804804649314, "grad_norm": 0.156844, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:28.957554+00:00", "epoch": 0, "step": 10410, "train_loss": 3.604288339614868, "perplexity": 36.75551710197047, "lr": 0.0026291804804649314, "grad_norm": 0.136356, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:29.264462+00:00", "epoch": 0, "step": 10411, "train_loss": 3.6136584281921387, "perplexity": 37.101538145156034, "lr": 0.0026291804804649314, "grad_norm": 0.156937, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:29.568784+00:00", "epoch": 0, "step": 10412, "train_loss": 3.5224335193634033, "perplexity": 33.866743641865, "lr": 0.0026291804804649314, "grad_norm": 0.174747, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:29.872962+00:00", "epoch": 0, "step": 10413, "train_loss": 3.5242276191711426, "perplexity": 33.92755849776972, "lr": 0.0026291804804649314, "grad_norm": 0.15252, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:30.178142+00:00", "epoch": 0, "step": 10414, "train_loss": 3.4786906242370605, "perplexity": 32.41724791589209, "lr": 0.0026291804804649314, "grad_norm": 0.15797, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:30.484406+00:00", "epoch": 0, "step": 10415, "train_loss": 3.53273344039917, "perplexity": 34.21737104460996, "lr": 0.0026291804804649314, "grad_norm": 0.154893, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:30.788750+00:00", "epoch": 0, "step": 10416, "train_loss": 3.4939920902252197, "perplexity": 32.91709376687739, "lr": 0.0026291804804649314, "grad_norm": 0.172406, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:31.092781+00:00", "epoch": 0, "step": 10417, "train_loss": 3.5462589263916016, "perplexity": 34.68332161800133, "lr": 0.0026291804804649314, "grad_norm": 0.169309, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:31.397679+00:00", "epoch": 0, "step": 10418, "train_loss": 3.5378735065460205, "perplexity": 34.393703386873135, "lr": 0.0026291804804649314, "grad_norm": 0.164494, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:31.702573+00:00", "epoch": 0, "step": 10419, "train_loss": 3.519798994064331, "perplexity": 33.777638275652066, "lr": 0.0026291804804649314, "grad_norm": 0.155371, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:32.008948+00:00", "epoch": 0, "step": 10420, "train_loss": 3.452996015548706, "perplexity": 31.594909490668908, "lr": 0.0026291804804649314, "grad_norm": 0.144378, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:32.314576+00:00", "epoch": 0, "step": 10421, "train_loss": 3.5929152965545654, "perplexity": 36.33986312879263, "lr": 0.0026291804804649314, "grad_norm": 0.146833, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:32.620359+00:00", "epoch": 0, "step": 10422, "train_loss": 3.6414191722869873, "perplexity": 38.1459339820243, "lr": 0.0026291804804649314, "grad_norm": 0.166298, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:32.925739+00:00", "epoch": 0, "step": 10423, "train_loss": 3.474337100982666, "perplexity": 32.276425432643805, "lr": 0.0026291804804649314, "grad_norm": 0.157071, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:33.229989+00:00", "epoch": 0, "step": 10424, "train_loss": 3.415449380874634, "perplexity": 30.43062129461035, "lr": 0.0026291804804649314, "grad_norm": 0.166646, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:33.535187+00:00", "epoch": 0, "step": 10425, "train_loss": 3.5411839485168457, "perplexity": 34.50775041497772, "lr": 0.0026291804804649314, "grad_norm": 0.191647, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:33.839812+00:00", "epoch": 0, "step": 10426, "train_loss": 3.531768798828125, "perplexity": 34.18437946113323, "lr": 0.0026291804804649314, "grad_norm": 0.167726, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:34.144983+00:00", "epoch": 0, "step": 10427, "train_loss": 3.569753408432007, "perplexity": 35.50783613898055, "lr": 0.0026291804804649314, "grad_norm": 0.145566, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:34.450272+00:00", "epoch": 0, "step": 10428, "train_loss": 3.5285146236419678, "perplexity": 34.073318305972535, "lr": 0.0026291804804649314, "grad_norm": 0.165494, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:34.754729+00:00", "epoch": 0, "step": 10429, "train_loss": 3.5288901329040527, "perplexity": 34.08611555517883, "lr": 0.0026291804804649314, "grad_norm": 0.144196, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:35.059478+00:00", "epoch": 0, "step": 10430, "train_loss": 3.490849733352661, "perplexity": 32.81381885934364, "lr": 0.0026291804804649314, "grad_norm": 0.151047, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:35.364642+00:00", "epoch": 0, "step": 10431, "train_loss": 3.3881521224975586, "perplexity": 29.611183842837065, "lr": 0.0026291804804649314, "grad_norm": 0.150236, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:35.670143+00:00", "epoch": 0, "step": 10432, "train_loss": 3.5210986137390137, "perplexity": 33.82156489668717, "lr": 0.0026291804804649314, "grad_norm": 0.153507, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:35.975125+00:00", "epoch": 0, "step": 10433, "train_loss": 3.4443204402923584, "perplexity": 31.321991049680776, "lr": 0.0026291804804649314, "grad_norm": 0.151376, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:36.280419+00:00", "epoch": 0, "step": 10434, "train_loss": 3.5886874198913574, "perplexity": 36.18654699874673, "lr": 0.0026291804804649314, "grad_norm": 0.138271, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:36.584431+00:00", "epoch": 0, "step": 10435, "train_loss": 3.594083547592163, "perplexity": 36.382342019771926, "lr": 0.0026291804804649314, "grad_norm": 0.154405, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:36.889937+00:00", "epoch": 0, "step": 10436, "train_loss": 3.5094501972198486, "perplexity": 33.42988288782267, "lr": 0.0026291804804649314, "grad_norm": 0.157648, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:37.194298+00:00", "epoch": 0, "step": 10437, "train_loss": 3.5776851177215576, "perplexity": 35.790593866379254, "lr": 0.0026291804804649314, "grad_norm": 0.149565, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:37.499051+00:00", "epoch": 0, "step": 10438, "train_loss": 3.5247247219085693, "perplexity": 33.94442817260655, "lr": 0.0026291804804649314, "grad_norm": 0.143989, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:37.803602+00:00", "epoch": 0, "step": 10439, "train_loss": 3.5757970809936523, "perplexity": 35.723083661573455, "lr": 0.0026291804804649314, "grad_norm": 0.16585, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:38.109143+00:00", "epoch": 0, "step": 10440, "train_loss": 3.468120813369751, "perplexity": 32.076408215404854, "lr": 0.0026291804804649314, "grad_norm": 0.169214, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:38.413937+00:00", "epoch": 0, "step": 10441, "train_loss": 3.6041576862335205, "perplexity": 36.75071518307821, "lr": 0.0026291804804649314, "grad_norm": 0.145618, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:38.719100+00:00", "epoch": 0, "step": 10442, "train_loss": 3.56998348236084, "perplexity": 35.51600650620361, "lr": 0.0026291804804649314, "grad_norm": 0.162333, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:39.024427+00:00", "epoch": 0, "step": 10443, "train_loss": 3.4700732231140137, "perplexity": 32.13909568326389, "lr": 0.0026291804804649314, "grad_norm": 0.153044, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:39.329900+00:00", "epoch": 0, "step": 10444, "train_loss": 3.5145905017852783, "perplexity": 33.60216507957413, "lr": 0.0026291804804649314, "grad_norm": 0.183295, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:39.634686+00:00", "epoch": 0, "step": 10445, "train_loss": 3.452456474304199, "perplexity": 31.577867331769465, "lr": 0.0026291804804649314, "grad_norm": 0.170271, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:39.939805+00:00", "epoch": 0, "step": 10446, "train_loss": 3.5380873680114746, "perplexity": 34.40105966126562, "lr": 0.0026291804804649314, "grad_norm": 0.181828, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:40.245946+00:00", "epoch": 0, "step": 10447, "train_loss": 3.5450291633605957, "perplexity": 34.64069556662611, "lr": 0.0026291804804649314, "grad_norm": 0.155534, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:40.551697+00:00", "epoch": 0, "step": 10448, "train_loss": 3.6329495906829834, "perplexity": 37.824218203478424, "lr": 0.0026291804804649314, "grad_norm": 0.176972, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:40.856418+00:00", "epoch": 0, "step": 10449, "train_loss": 3.4781343936920166, "perplexity": 32.39922146631072, "lr": 0.0026291804804649314, "grad_norm": 0.174139, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:41.161032+00:00", "epoch": 0, "step": 10450, "train_loss": 3.4289064407348633, "perplexity": 30.842895766447995, "lr": 0.0026291804804649314, "grad_norm": 0.1921, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:41.465660+00:00", "epoch": 0, "step": 10451, "train_loss": 3.53019642829895, "perplexity": 34.13067118603036, "lr": 0.0026291804804649314, "grad_norm": 0.193356, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:41.771797+00:00", "epoch": 0, "step": 10452, "train_loss": 3.613663911819458, "perplexity": 37.10174159672202, "lr": 0.0026291804804649314, "grad_norm": 0.1776, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:42.077257+00:00", "epoch": 0, "step": 10453, "train_loss": 3.5183351039886475, "perplexity": 33.72822770084915, "lr": 0.0026291804804649314, "grad_norm": 0.161069, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:42.382332+00:00", "epoch": 0, "step": 10454, "train_loss": 3.5625429153442383, "perplexity": 35.25272796508202, "lr": 0.0026291804804649314, "grad_norm": 0.16011, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:42.687432+00:00", "epoch": 0, "step": 10455, "train_loss": 3.468217372894287, "perplexity": 32.079505647671844, "lr": 0.0026291804804649314, "grad_norm": 0.17723, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:42.991425+00:00", "epoch": 0, "step": 10456, "train_loss": 3.570589542388916, "perplexity": 35.53753786209194, "lr": 0.0026291804804649314, "grad_norm": 0.161212, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:43.296815+00:00", "epoch": 0, "step": 10457, "train_loss": 3.525559425354004, "perplexity": 33.97277353209439, "lr": 0.0026291804804649314, "grad_norm": 0.160155, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:43.602264+00:00", "epoch": 0, "step": 10458, "train_loss": 3.5109095573425293, "perplexity": 33.47870474148343, "lr": 0.0026291804804649314, "grad_norm": 0.151697, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:43.907758+00:00", "epoch": 0, "step": 10459, "train_loss": 3.595036029815674, "perplexity": 36.41701206247401, "lr": 0.0026291804804649314, "grad_norm": 0.153173, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:44.212824+00:00", "epoch": 0, "step": 10460, "train_loss": 3.526792287826538, "perplexity": 34.014683118739946, "lr": 0.0026291804804649314, "grad_norm": 0.163932, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:44.517632+00:00", "epoch": 0, "step": 10461, "train_loss": 3.6073174476623535, "perplexity": 36.86702233004575, "lr": 0.0026291804804649314, "grad_norm": 0.151137, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:44.821835+00:00", "epoch": 0, "step": 10462, "train_loss": 3.5559771060943604, "perplexity": 35.02202348732486, "lr": 0.0026291804804649314, "grad_norm": 0.1695, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:45.127198+00:00", "epoch": 0, "step": 10463, "train_loss": 3.641228199005127, "perplexity": 38.13864982338411, "lr": 0.0026291804804649314, "grad_norm": 0.141835, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:45.432547+00:00", "epoch": 0, "step": 10464, "train_loss": 3.585564136505127, "perplexity": 36.073702472150295, "lr": 0.0026291804804649314, "grad_norm": 0.164623, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:45.738081+00:00", "epoch": 0, "step": 10465, "train_loss": 3.561729669570923, "perplexity": 35.22407048743229, "lr": 0.0026291804804649314, "grad_norm": 0.168682, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:46.042831+00:00", "epoch": 0, "step": 10466, "train_loss": 3.6290230751037598, "perplexity": 37.67599201808364, "lr": 0.0026291804804649314, "grad_norm": 0.136676, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:46.347723+00:00", "epoch": 0, "step": 10467, "train_loss": 3.517263412475586, "perplexity": 33.69210080737182, "lr": 0.0026291804804649314, "grad_norm": 0.148885, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:46.653038+00:00", "epoch": 0, "step": 10468, "train_loss": 3.4885196685791016, "perplexity": 32.7374495432122, "lr": 0.0026291804804649314, "grad_norm": 0.144004, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:46.958137+00:00", "epoch": 0, "step": 10469, "train_loss": 3.5360145568847656, "perplexity": 34.32982661386046, "lr": 0.0026291804804649314, "grad_norm": 0.150836, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:47.263993+00:00", "epoch": 0, "step": 10470, "train_loss": 3.482051134109497, "perplexity": 32.526369647185454, "lr": 0.0026291804804649314, "grad_norm": 0.146924, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:47.567916+00:00", "epoch": 0, "step": 10471, "train_loss": 3.483773708343506, "perplexity": 32.5824470183165, "lr": 0.0026291804804649314, "grad_norm": 0.138196, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:47.872380+00:00", "epoch": 0, "step": 10472, "train_loss": 3.5168988704681396, "perplexity": 33.679820859722795, "lr": 0.0026291804804649314, "grad_norm": 0.135559, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:48.176684+00:00", "epoch": 0, "step": 10473, "train_loss": 3.541891574859619, "perplexity": 34.53217764985935, "lr": 0.0026291804804649314, "grad_norm": 0.133466, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:48.482557+00:00", "epoch": 0, "step": 10474, "train_loss": 3.5708131790161133, "perplexity": 35.54548624594006, "lr": 0.0026291804804649314, "grad_norm": 0.145161, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:48.788883+00:00", "epoch": 0, "step": 10475, "train_loss": 3.5686354637145996, "perplexity": 35.46816252172984, "lr": 0.0026291804804649314, "grad_norm": 0.150429, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:49.093662+00:00", "epoch": 0, "step": 10476, "train_loss": 3.511629581451416, "perplexity": 33.50281889637624, "lr": 0.0026291804804649314, "grad_norm": 0.161966, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:49.398516+00:00", "epoch": 0, "step": 10477, "train_loss": 3.6326348781585693, "perplexity": 37.81231632121756, "lr": 0.0026291804804649314, "grad_norm": 0.142507, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:49.703576+00:00", "epoch": 0, "step": 10478, "train_loss": 3.559666395187378, "perplexity": 35.15146848978853, "lr": 0.0026291804804649314, "grad_norm": 0.150257, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:50.009870+00:00", "epoch": 0, "step": 10479, "train_loss": 3.5506632328033447, "perplexity": 34.83641448024799, "lr": 0.0026291804804649314, "grad_norm": 0.151874, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:50.315132+00:00", "epoch": 0, "step": 10480, "train_loss": 3.4405035972595215, "perplexity": 31.202667790305117, "lr": 0.0026291804804649314, "grad_norm": 0.157399, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:50.620291+00:00", "epoch": 0, "step": 10481, "train_loss": 3.6261932849884033, "perplexity": 37.56952757534317, "lr": 0.0026291804804649314, "grad_norm": 0.189201, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:50.925219+00:00", "epoch": 0, "step": 10482, "train_loss": 3.520402193069458, "perplexity": 33.79801905967184, "lr": 0.0026291804804649314, "grad_norm": 0.188839, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:51.229797+00:00", "epoch": 0, "step": 10483, "train_loss": 3.703235149383545, "perplexity": 40.578369324891426, "lr": 0.0026291804804649314, "grad_norm": 0.189007, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:51.535453+00:00", "epoch": 0, "step": 10484, "train_loss": 3.4969418048858643, "perplexity": 33.01433314465132, "lr": 0.0026291804804649314, "grad_norm": 0.231013, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:51.840315+00:00", "epoch": 0, "step": 10485, "train_loss": 3.53535795211792, "perplexity": 34.30729288474737, "lr": 0.0026291804804649314, "grad_norm": 0.23138, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:52.144657+00:00", "epoch": 0, "step": 10486, "train_loss": 3.5500426292419434, "perplexity": 34.81480158456879, "lr": 0.0026291804804649314, "grad_norm": 0.184029, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:52.449383+00:00", "epoch": 0, "step": 10487, "train_loss": 3.4466190338134766, "perplexity": 31.39407038418562, "lr": 0.0026291804804649314, "grad_norm": 0.204909, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:52.754419+00:00", "epoch": 0, "step": 10488, "train_loss": 3.5297739505767822, "perplexity": 34.11625478333054, "lr": 0.0026291804804649314, "grad_norm": 0.187544, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:53.058341+00:00", "epoch": 0, "step": 10489, "train_loss": 3.5264670848846436, "perplexity": 34.00362324217179, "lr": 0.0026291804804649314, "grad_norm": 0.160691, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:53.363011+00:00", "epoch": 0, "step": 10490, "train_loss": 3.4923484325408936, "perplexity": 32.86303377299472, "lr": 0.0026291804804649314, "grad_norm": 0.194145, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:53.667346+00:00", "epoch": 0, "step": 10491, "train_loss": 3.475893974304199, "perplexity": 32.3267148753023, "lr": 0.0026291804804649314, "grad_norm": 0.194521, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:53.972593+00:00", "epoch": 0, "step": 10492, "train_loss": 3.4619386196136475, "perplexity": 31.878717356052583, "lr": 0.0026291804804649314, "grad_norm": 0.190852, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:54.277333+00:00", "epoch": 0, "step": 10493, "train_loss": 3.4036405086517334, "perplexity": 30.073383415687736, "lr": 0.0026291804804649314, "grad_norm": 0.16962, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:54.582107+00:00", "epoch": 0, "step": 10494, "train_loss": 3.540998935699463, "perplexity": 34.50136662941119, "lr": 0.0026291804804649314, "grad_norm": 0.163604, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:54.886828+00:00", "epoch": 0, "step": 10495, "train_loss": 3.5010266304016113, "perplexity": 33.14946674575591, "lr": 0.0026291804804649314, "grad_norm": 0.145729, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:55.191827+00:00", "epoch": 0, "step": 10496, "train_loss": 3.4859349727630615, "perplexity": 32.65294245395692, "lr": 0.0026291804804649314, "grad_norm": 0.147551, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:55.497134+00:00", "epoch": 0, "step": 10497, "train_loss": 3.504474639892578, "perplexity": 33.26396370157531, "lr": 0.0026291804804649314, "grad_norm": 0.148258, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:55.801557+00:00", "epoch": 0, "step": 10498, "train_loss": 3.580893039703369, "perplexity": 35.905591652530426, "lr": 0.0026291804804649314, "grad_norm": 0.147924, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:56.107404+00:00", "epoch": 0, "step": 10499, "train_loss": 3.653510808944702, "perplexity": 38.609980643597844, "lr": 0.0026291804804649314, "grad_norm": 0.144742, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:52:56.411602+00:00", "epoch": 0, "step": 10500, "train_loss": 3.5198094844818115, "perplexity": 33.77799261903768, "lr": 0.0026291804804649314, "grad_norm": 0.148365, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:52:59.483768+00:00", "step": 10500, "epoch": 0, "val_loss": 3.492388963699341, "val_ppl": 32.86436577681727, "eval_train_loss": 3.5198094844818115, "eval_train_ppl": 33.77799261903768} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:53:00.507146+00:00", "step": 10500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4924_epoch_0000_step_0010500.pt", "val_loss": 3.492388963699341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:01.647378+00:00", "epoch": 0, "step": 10501, "train_loss": 3.498218536376953, "perplexity": 33.0565105022855, "lr": 0.0026291804804649314, "grad_norm": 0.151322, "tokens_per_sec": 6258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:01.951149+00:00", "epoch": 0, "step": 10502, "train_loss": 3.43365740776062, "perplexity": 30.989777987225505, "lr": 0.0026291804804649314, "grad_norm": 0.185516, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:02.254754+00:00", "epoch": 0, "step": 10503, "train_loss": 3.5665950775146484, "perplexity": 35.39586755229325, "lr": 0.0026291804804649314, "grad_norm": 0.179518, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:02.559304+00:00", "epoch": 0, "step": 10504, "train_loss": 3.4662084579467773, "perplexity": 32.01512533820237, "lr": 0.0026291804804649314, "grad_norm": 0.161078, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:02.863735+00:00", "epoch": 0, "step": 10505, "train_loss": 3.52067232131958, "perplexity": 33.80715009263747, "lr": 0.0026291804804649314, "grad_norm": 0.163614, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:03.168254+00:00", "epoch": 0, "step": 10506, "train_loss": 3.5118701457977295, "perplexity": 33.510879449605696, "lr": 0.0026291804804649314, "grad_norm": 0.17153, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:03.471759+00:00", "epoch": 0, "step": 10507, "train_loss": 3.4728281497955322, "perplexity": 32.227758609335446, "lr": 0.0026291804804649314, "grad_norm": 0.148749, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:03.775797+00:00", "epoch": 0, "step": 10508, "train_loss": 3.557539701461792, "perplexity": 35.07679151797578, "lr": 0.0026291804804649314, "grad_norm": 0.159404, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:04.080515+00:00", "epoch": 0, "step": 10509, "train_loss": 3.5388472080230713, "perplexity": 34.42720889619694, "lr": 0.0026291804804649314, "grad_norm": 0.180432, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:04.384084+00:00", "epoch": 0, "step": 10510, "train_loss": 3.5555317401885986, "perplexity": 35.00642934492026, "lr": 0.0026291804804649314, "grad_norm": 0.154081, "tokens_per_sec": 108027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:04.688263+00:00", "epoch": 0, "step": 10511, "train_loss": 3.510363817214966, "perplexity": 33.46043905349994, "lr": 0.0026291804804649314, "grad_norm": 0.175899, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:04.992550+00:00", "epoch": 0, "step": 10512, "train_loss": 3.622701644897461, "perplexity": 37.438577055746215, "lr": 0.0026291804804649314, "grad_norm": 0.188914, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:05.296748+00:00", "epoch": 0, "step": 10513, "train_loss": 3.5590462684631348, "perplexity": 35.12967688225906, "lr": 0.0026291804804649314, "grad_norm": 0.180616, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:05.601760+00:00", "epoch": 0, "step": 10514, "train_loss": 3.5672059059143066, "perplexity": 35.417494958069106, "lr": 0.0026291804804649314, "grad_norm": 0.175002, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:05.906249+00:00", "epoch": 0, "step": 10515, "train_loss": 3.590327501296997, "perplexity": 36.24594457673097, "lr": 0.0026291804804649314, "grad_norm": 0.166845, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:06.209869+00:00", "epoch": 0, "step": 10516, "train_loss": 3.471510171890259, "perplexity": 32.18531111412677, "lr": 0.0026291804804649314, "grad_norm": 0.157238, "tokens_per_sec": 107924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:06.513551+00:00", "epoch": 0, "step": 10517, "train_loss": 3.560011148452759, "perplexity": 35.163589162534024, "lr": 0.0026291804804649314, "grad_norm": 0.154744, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:06.817677+00:00", "epoch": 0, "step": 10518, "train_loss": 3.535193681716919, "perplexity": 34.3016576748497, "lr": 0.0026291804804649314, "grad_norm": 0.169794, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:07.121821+00:00", "epoch": 0, "step": 10519, "train_loss": 3.4406468868255615, "perplexity": 31.207139127372404, "lr": 0.0026291804804649314, "grad_norm": 0.172212, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:07.426484+00:00", "epoch": 0, "step": 10520, "train_loss": 3.5787594318389893, "perplexity": 35.829064867911086, "lr": 0.0026291804804649314, "grad_norm": 0.174967, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:07.729225+00:00", "epoch": 0, "step": 10521, "train_loss": 3.584928512573242, "perplexity": 36.05078044921222, "lr": 0.0026291804804649314, "grad_norm": 0.145021, "tokens_per_sec": 108170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:08.033340+00:00", "epoch": 0, "step": 10522, "train_loss": 3.599680185317993, "perplexity": 36.58653166242775, "lr": 0.0026291804804649314, "grad_norm": 0.164572, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:08.338107+00:00", "epoch": 0, "step": 10523, "train_loss": 3.4416871070861816, "perplexity": 31.23961831559481, "lr": 0.0026291804804649314, "grad_norm": 0.164751, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:08.642696+00:00", "epoch": 0, "step": 10524, "train_loss": 3.6567888259887695, "perplexity": 38.73675248482902, "lr": 0.0026291804804649314, "grad_norm": 0.168843, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:08.947801+00:00", "epoch": 0, "step": 10525, "train_loss": 3.6460378170013428, "perplexity": 38.32252398771646, "lr": 0.0026291804804649314, "grad_norm": 0.185399, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:09.251989+00:00", "epoch": 0, "step": 10526, "train_loss": 3.5200998783111572, "perplexity": 33.78780296402482, "lr": 0.0026291804804649314, "grad_norm": 0.179147, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:09.556363+00:00", "epoch": 0, "step": 10527, "train_loss": 3.4415993690490723, "perplexity": 31.236877533040953, "lr": 0.0026291804804649314, "grad_norm": 0.153897, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:09.861189+00:00", "epoch": 0, "step": 10528, "train_loss": 3.5196423530578613, "perplexity": 33.772347726765666, "lr": 0.0026291804804649314, "grad_norm": 0.158432, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:10.165775+00:00", "epoch": 0, "step": 10529, "train_loss": 3.5087289810180664, "perplexity": 33.40578140689426, "lr": 0.0026291804804649314, "grad_norm": 0.174337, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:10.471366+00:00", "epoch": 0, "step": 10530, "train_loss": 3.599520683288574, "perplexity": 36.58069650175055, "lr": 0.0026291804804649314, "grad_norm": 0.171379, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:10.775342+00:00", "epoch": 0, "step": 10531, "train_loss": 3.468841552734375, "perplexity": 32.099535278772784, "lr": 0.0026291804804649314, "grad_norm": 0.201314, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:11.080987+00:00", "epoch": 0, "step": 10532, "train_loss": 3.465315818786621, "perplexity": 31.986560134714544, "lr": 0.0026291804804649314, "grad_norm": 0.17355, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:11.385389+00:00", "epoch": 0, "step": 10533, "train_loss": 3.5179691314697266, "perplexity": 33.71588635483322, "lr": 0.0026291804804649314, "grad_norm": 0.186211, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:11.690219+00:00", "epoch": 0, "step": 10534, "train_loss": 3.536877393722534, "perplexity": 34.35946043563514, "lr": 0.0026291804804649314, "grad_norm": 0.19285, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:11.994888+00:00", "epoch": 0, "step": 10535, "train_loss": 3.5151331424713135, "perplexity": 33.62040392960867, "lr": 0.0026291804804649314, "grad_norm": 0.207676, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:12.298468+00:00", "epoch": 0, "step": 10536, "train_loss": 3.620824098587036, "perplexity": 37.36835034111227, "lr": 0.0026291804804649314, "grad_norm": 0.186938, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:12.602562+00:00", "epoch": 0, "step": 10537, "train_loss": 3.546347141265869, "perplexity": 34.68638133781146, "lr": 0.0026291804804649314, "grad_norm": 0.192992, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:12.906570+00:00", "epoch": 0, "step": 10538, "train_loss": 3.511781692504883, "perplexity": 33.507915433062635, "lr": 0.0026291804804649314, "grad_norm": 0.18357, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:13.211189+00:00", "epoch": 0, "step": 10539, "train_loss": 3.643012523651123, "perplexity": 38.20676230554784, "lr": 0.0026291804804649314, "grad_norm": 0.165476, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:13.516245+00:00", "epoch": 0, "step": 10540, "train_loss": 3.515596389770508, "perplexity": 33.63598209892073, "lr": 0.0026291804804649314, "grad_norm": 0.16997, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:13.819510+00:00", "epoch": 0, "step": 10541, "train_loss": 3.6193222999572754, "perplexity": 37.312272722958404, "lr": 0.0026291804804649314, "grad_norm": 0.15562, "tokens_per_sec": 107981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:14.123828+00:00", "epoch": 0, "step": 10542, "train_loss": 3.6442317962646484, "perplexity": 38.253375175601754, "lr": 0.0026291804804649314, "grad_norm": 0.166734, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:14.428564+00:00", "epoch": 0, "step": 10543, "train_loss": 3.5417542457580566, "perplexity": 34.52743570253884, "lr": 0.0026291804804649314, "grad_norm": 0.150129, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:14.732780+00:00", "epoch": 0, "step": 10544, "train_loss": 3.54362416267395, "perplexity": 34.59205954039498, "lr": 0.0026291804804649314, "grad_norm": 0.161048, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:15.038052+00:00", "epoch": 0, "step": 10545, "train_loss": 3.6520793437957764, "perplexity": 38.55475134074778, "lr": 0.0026291804804649314, "grad_norm": 0.158026, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:15.342009+00:00", "epoch": 0, "step": 10546, "train_loss": 3.559784173965454, "perplexity": 35.155608830612195, "lr": 0.0026291804804649314, "grad_norm": 0.1703, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:15.646245+00:00", "epoch": 0, "step": 10547, "train_loss": 3.604123115539551, "perplexity": 36.74944470731119, "lr": 0.0026291804804649314, "grad_norm": 0.179832, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:15.950856+00:00", "epoch": 0, "step": 10548, "train_loss": 3.688673734664917, "perplexity": 39.99177206839102, "lr": 0.0026291804804649314, "grad_norm": 0.170067, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:16.255657+00:00", "epoch": 0, "step": 10549, "train_loss": 3.4255478382110596, "perplexity": 30.739480501482493, "lr": 0.0026291804804649314, "grad_norm": 0.17875, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:16.559936+00:00", "epoch": 0, "step": 10550, "train_loss": 3.482529640197754, "perplexity": 32.54193743743469, "lr": 0.0026291804804649314, "grad_norm": 0.186464, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:16.863807+00:00", "epoch": 0, "step": 10551, "train_loss": 3.6203718185424805, "perplexity": 37.35145320336216, "lr": 0.0026291804804649314, "grad_norm": 0.172647, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:17.168233+00:00", "epoch": 0, "step": 10552, "train_loss": 3.5915098190307617, "perplexity": 36.288824143422126, "lr": 0.0026291804804649314, "grad_norm": 0.157913, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:17.473504+00:00", "epoch": 0, "step": 10553, "train_loss": 3.587629795074463, "perplexity": 36.14829544006759, "lr": 0.0026291804804649314, "grad_norm": 0.161797, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:17.779213+00:00", "epoch": 0, "step": 10554, "train_loss": 3.4690799713134766, "perplexity": 32.10718931675994, "lr": 0.0026291804804649314, "grad_norm": 0.187097, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:18.083252+00:00", "epoch": 0, "step": 10555, "train_loss": 3.6043996810913086, "perplexity": 36.75960974334807, "lr": 0.0026291804804649314, "grad_norm": 0.156519, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:18.387914+00:00", "epoch": 0, "step": 10556, "train_loss": 3.428436040878296, "perplexity": 30.828390684565022, "lr": 0.0026291804804649314, "grad_norm": 0.149327, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:18.693442+00:00", "epoch": 0, "step": 10557, "train_loss": 3.449918270111084, "perplexity": 31.497817890388625, "lr": 0.0026291804804649314, "grad_norm": 0.148386, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:19.000472+00:00", "epoch": 0, "step": 10558, "train_loss": 3.6027920246124268, "perplexity": 36.7005603968333, "lr": 0.0026291804804649314, "grad_norm": 0.15819, "tokens_per_sec": 106730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:19.306312+00:00", "epoch": 0, "step": 10559, "train_loss": 3.5065884590148926, "perplexity": 33.33435207206414, "lr": 0.0026291804804649314, "grad_norm": 0.166313, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:19.611469+00:00", "epoch": 0, "step": 10560, "train_loss": 3.502413272857666, "perplexity": 33.1954650879979, "lr": 0.0026291804804649314, "grad_norm": 0.155284, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:19.916114+00:00", "epoch": 0, "step": 10561, "train_loss": 3.4979114532470703, "perplexity": 33.0463609640332, "lr": 0.0026291804804649314, "grad_norm": 0.160039, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:20.220984+00:00", "epoch": 0, "step": 10562, "train_loss": 3.62176775932312, "perplexity": 37.40363002951174, "lr": 0.0026291804804649314, "grad_norm": 0.142604, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:20.526084+00:00", "epoch": 0, "step": 10563, "train_loss": 3.4543113708496094, "perplexity": 31.63649536645425, "lr": 0.0026291804804649314, "grad_norm": 0.160488, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:20.831192+00:00", "epoch": 0, "step": 10564, "train_loss": 3.6610965728759766, "perplexity": 38.9039805392317, "lr": 0.0026291804804649314, "grad_norm": 0.144078, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:21.135904+00:00", "epoch": 0, "step": 10565, "train_loss": 3.543708324432373, "perplexity": 34.59497099146807, "lr": 0.0026291804804649314, "grad_norm": 0.150602, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:21.440278+00:00", "epoch": 0, "step": 10566, "train_loss": 3.4507224559783936, "perplexity": 31.523158178172157, "lr": 0.0026291804804649314, "grad_norm": 0.147936, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:21.746184+00:00", "epoch": 0, "step": 10567, "train_loss": 3.5609359741210938, "perplexity": 35.19612439477231, "lr": 0.0026291804804649314, "grad_norm": 0.159295, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:22.051900+00:00", "epoch": 0, "step": 10568, "train_loss": 3.6022298336029053, "perplexity": 36.679933470413374, "lr": 0.0026291804804649314, "grad_norm": 0.165153, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:22.358880+00:00", "epoch": 0, "step": 10569, "train_loss": 3.475978136062622, "perplexity": 32.32943566296177, "lr": 0.0026291804804649314, "grad_norm": 0.178516, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:22.664905+00:00", "epoch": 0, "step": 10570, "train_loss": 3.4692270755767822, "perplexity": 32.11191276860263, "lr": 0.0026291804804649314, "grad_norm": 0.188237, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:22.970694+00:00", "epoch": 0, "step": 10571, "train_loss": 3.5431571006774902, "perplexity": 34.57590667649245, "lr": 0.0026291804804649314, "grad_norm": 0.164929, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:23.275559+00:00", "epoch": 0, "step": 10572, "train_loss": 3.6373343467712402, "perplexity": 37.990432312388876, "lr": 0.0026291804804649314, "grad_norm": 0.170513, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:23.580289+00:00", "epoch": 0, "step": 10573, "train_loss": 3.5651023387908936, "perplexity": 35.343070186277316, "lr": 0.0026291804804649314, "grad_norm": 0.146715, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:23.884845+00:00", "epoch": 0, "step": 10574, "train_loss": 3.5731582641601562, "perplexity": 35.62894125397676, "lr": 0.0026291804804649314, "grad_norm": 0.154222, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:24.190255+00:00", "epoch": 0, "step": 10575, "train_loss": 3.5251567363739014, "perplexity": 33.95909582468535, "lr": 0.0026291804804649314, "grad_norm": 0.140064, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:24.494652+00:00", "epoch": 0, "step": 10576, "train_loss": 3.536543607711792, "perplexity": 34.34799364223915, "lr": 0.0026291804804649314, "grad_norm": 0.158094, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:24.799926+00:00", "epoch": 0, "step": 10577, "train_loss": 3.596402168273926, "perplexity": 36.466796741826826, "lr": 0.0026291804804649314, "grad_norm": 0.177906, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:25.105554+00:00", "epoch": 0, "step": 10578, "train_loss": 3.560702323913574, "perplexity": 35.187901773649564, "lr": 0.0026291804804649314, "grad_norm": 0.159195, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:25.411569+00:00", "epoch": 0, "step": 10579, "train_loss": 3.5538153648376465, "perplexity": 34.94639670647596, "lr": 0.0026291804804649314, "grad_norm": 0.15782, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:25.715691+00:00", "epoch": 0, "step": 10580, "train_loss": 3.5728726387023926, "perplexity": 35.6187661745215, "lr": 0.0026291804804649314, "grad_norm": 0.194182, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:26.020460+00:00", "epoch": 0, "step": 10581, "train_loss": 3.582336902618408, "perplexity": 35.95747184969335, "lr": 0.0026291804804649314, "grad_norm": 0.203522, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:26.324552+00:00", "epoch": 0, "step": 10582, "train_loss": 3.5602900981903076, "perplexity": 35.17339940472081, "lr": 0.0026291804804649314, "grad_norm": 0.165079, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:26.629650+00:00", "epoch": 0, "step": 10583, "train_loss": 3.596585750579834, "perplexity": 36.47349201500982, "lr": 0.0026291804804649314, "grad_norm": 0.167059, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:26.935033+00:00", "epoch": 0, "step": 10584, "train_loss": 3.6363415718078613, "perplexity": 37.9527350778727, "lr": 0.0026291804804649314, "grad_norm": 0.141878, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:27.239104+00:00", "epoch": 0, "step": 10585, "train_loss": 3.5647006034851074, "perplexity": 35.32887447881797, "lr": 0.0026291804804649314, "grad_norm": 0.161535, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:27.543859+00:00", "epoch": 0, "step": 10586, "train_loss": 3.47499418258667, "perplexity": 32.29764064733893, "lr": 0.0026291804804649314, "grad_norm": 0.164757, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:27.849491+00:00", "epoch": 0, "step": 10587, "train_loss": 3.5307624340057373, "perplexity": 34.14999480882263, "lr": 0.0026291804804649314, "grad_norm": 0.156648, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:28.154477+00:00", "epoch": 0, "step": 10588, "train_loss": 3.4659435749053955, "perplexity": 32.006646197472435, "lr": 0.0026291804804649314, "grad_norm": 0.158005, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:28.460992+00:00", "epoch": 0, "step": 10589, "train_loss": 3.5375237464904785, "perplexity": 34.381675946747926, "lr": 0.0026291804804649314, "grad_norm": 0.142296, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:28.766038+00:00", "epoch": 0, "step": 10590, "train_loss": 3.481837511062622, "perplexity": 32.519422007112084, "lr": 0.0026291804804649314, "grad_norm": 0.150055, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:29.070361+00:00", "epoch": 0, "step": 10591, "train_loss": 3.473517656326294, "perplexity": 32.24998752197794, "lr": 0.0026291804804649314, "grad_norm": 0.182349, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:29.374674+00:00", "epoch": 0, "step": 10592, "train_loss": 3.636590003967285, "perplexity": 37.9621649290948, "lr": 0.0026291804804649314, "grad_norm": 0.165625, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:29.679739+00:00", "epoch": 0, "step": 10593, "train_loss": 3.44773530960083, "perplexity": 31.429134391732255, "lr": 0.0026291804804649314, "grad_norm": 0.156631, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:29.984351+00:00", "epoch": 0, "step": 10594, "train_loss": 3.480130910873413, "perplexity": 32.46397168458542, "lr": 0.0026291804804649314, "grad_norm": 0.164233, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:30.288906+00:00", "epoch": 0, "step": 10595, "train_loss": 3.489928722381592, "perplexity": 32.783610885244634, "lr": 0.0026291804804649314, "grad_norm": 0.157156, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:30.593642+00:00", "epoch": 0, "step": 10596, "train_loss": 3.5145952701568604, "perplexity": 33.6023253075652, "lr": 0.0026291804804649314, "grad_norm": 0.163995, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:30.898526+00:00", "epoch": 0, "step": 10597, "train_loss": 3.5136544704437256, "perplexity": 33.570727115670756, "lr": 0.0026291804804649314, "grad_norm": 0.159633, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:31.202240+00:00", "epoch": 0, "step": 10598, "train_loss": 3.5224058628082275, "perplexity": 33.86580701735281, "lr": 0.0026291804804649314, "grad_norm": 0.157421, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:31.508125+00:00", "epoch": 0, "step": 10599, "train_loss": 3.5074124336242676, "perplexity": 33.36183005081525, "lr": 0.0026291804804649314, "grad_norm": 0.157945, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:31.814008+00:00", "epoch": 0, "step": 10600, "train_loss": 3.491542339324951, "perplexity": 32.83655377852083, "lr": 0.0026291804804649314, "grad_norm": 0.136316, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:32.119620+00:00", "epoch": 0, "step": 10601, "train_loss": 3.51902174949646, "perplexity": 33.75139498983362, "lr": 0.0026291804804649314, "grad_norm": 0.158162, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:32.424177+00:00", "epoch": 0, "step": 10602, "train_loss": 3.4996726512908936, "perplexity": 33.10461343232775, "lr": 0.0026291804804649314, "grad_norm": 0.144813, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:32.729532+00:00", "epoch": 0, "step": 10603, "train_loss": 3.580791711807251, "perplexity": 35.901953598790676, "lr": 0.0026291804804649314, "grad_norm": 0.142173, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:33.035308+00:00", "epoch": 0, "step": 10604, "train_loss": 3.602046251296997, "perplexity": 36.673200301710544, "lr": 0.0026291804804649314, "grad_norm": 0.154483, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:33.339167+00:00", "epoch": 0, "step": 10605, "train_loss": 3.5444962978363037, "perplexity": 34.62224165138542, "lr": 0.0026291804804649314, "grad_norm": 0.1724, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:33.644963+00:00", "epoch": 0, "step": 10606, "train_loss": 3.568567991256714, "perplexity": 35.465769478360976, "lr": 0.0026291804804649314, "grad_norm": 0.19105, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:33.948890+00:00", "epoch": 0, "step": 10607, "train_loss": 3.552816152572632, "perplexity": 34.91149527814028, "lr": 0.0026291804804649314, "grad_norm": 0.164898, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:34.253840+00:00", "epoch": 0, "step": 10608, "train_loss": 3.481205463409424, "perplexity": 32.49887467687908, "lr": 0.0026291804804649314, "grad_norm": 0.153249, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:34.559530+00:00", "epoch": 0, "step": 10609, "train_loss": 3.531744956970215, "perplexity": 34.18356445173109, "lr": 0.0026291804804649314, "grad_norm": 0.171312, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:34.865949+00:00", "epoch": 0, "step": 10610, "train_loss": 3.4997496604919434, "perplexity": 33.10716289032384, "lr": 0.0026291804804649314, "grad_norm": 0.145356, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:35.173169+00:00", "epoch": 0, "step": 10611, "train_loss": 3.4659922122955322, "perplexity": 32.00820295506851, "lr": 0.0026291804804649314, "grad_norm": 0.172732, "tokens_per_sec": 106572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:35.478159+00:00", "epoch": 0, "step": 10612, "train_loss": 3.645327091217041, "perplexity": 38.29529685845637, "lr": 0.0026291804804649314, "grad_norm": 0.175752, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:35.784213+00:00", "epoch": 0, "step": 10613, "train_loss": 3.455777883529663, "perplexity": 31.68292472436152, "lr": 0.0026291804804649314, "grad_norm": 0.162071, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:36.088912+00:00", "epoch": 0, "step": 10614, "train_loss": 3.444725513458252, "perplexity": 31.334681317827386, "lr": 0.0026291804804649314, "grad_norm": 0.175988, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:36.394259+00:00", "epoch": 0, "step": 10615, "train_loss": 3.5326666831970215, "perplexity": 34.21508686489772, "lr": 0.0026291804804649314, "grad_norm": 0.199276, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:36.699449+00:00", "epoch": 0, "step": 10616, "train_loss": 3.5805258750915527, "perplexity": 35.89241080982686, "lr": 0.0026291804804649314, "grad_norm": 0.217903, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:37.005779+00:00", "epoch": 0, "step": 10617, "train_loss": 3.5778326988220215, "perplexity": 35.7958762713903, "lr": 0.0026291804804649314, "grad_norm": 0.191662, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:37.310876+00:00", "epoch": 0, "step": 10618, "train_loss": 3.5360491275787354, "perplexity": 34.33101344030495, "lr": 0.0026291804804649314, "grad_norm": 0.142578, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:37.614996+00:00", "epoch": 0, "step": 10619, "train_loss": 3.432847738265991, "perplexity": 30.964696664505077, "lr": 0.0026291804804649314, "grad_norm": 0.220025, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:37.920335+00:00", "epoch": 0, "step": 10620, "train_loss": 3.4307456016540527, "perplexity": 30.899673010216052, "lr": 0.0026291804804649314, "grad_norm": 0.19488, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:38.225920+00:00", "epoch": 0, "step": 10621, "train_loss": 3.5056612491607666, "perplexity": 33.3034584569875, "lr": 0.0026291804804649314, "grad_norm": 0.167579, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:38.531395+00:00", "epoch": 0, "step": 10622, "train_loss": 3.526160717010498, "perplexity": 33.99320722005459, "lr": 0.0026291804804649314, "grad_norm": 0.172935, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:38.836955+00:00", "epoch": 0, "step": 10623, "train_loss": 3.647669792175293, "perplexity": 38.38511645625278, "lr": 0.0026291804804649314, "grad_norm": 0.180926, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:39.142274+00:00", "epoch": 0, "step": 10624, "train_loss": 3.545355796813965, "perplexity": 34.652012224745086, "lr": 0.0026291804804649314, "grad_norm": 0.146862, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:39.446998+00:00", "epoch": 0, "step": 10625, "train_loss": 3.462362051010132, "perplexity": 31.892218664097925, "lr": 0.0026291804804649314, "grad_norm": 0.171726, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:39.752647+00:00", "epoch": 0, "step": 10626, "train_loss": 3.4652721881866455, "perplexity": 31.985164572349543, "lr": 0.0026291804804649314, "grad_norm": 0.169036, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:40.057459+00:00", "epoch": 0, "step": 10627, "train_loss": 3.592122793197632, "perplexity": 36.31107507409427, "lr": 0.0026291804804649314, "grad_norm": 0.17783, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:40.363206+00:00", "epoch": 0, "step": 10628, "train_loss": 3.5104482173919678, "perplexity": 33.46326323965785, "lr": 0.0026291804804649314, "grad_norm": 0.161226, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:40.667246+00:00", "epoch": 0, "step": 10629, "train_loss": 3.4850170612335205, "perplexity": 32.62298369345434, "lr": 0.0026291804804649314, "grad_norm": 0.183426, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:40.971182+00:00", "epoch": 0, "step": 10630, "train_loss": 3.4535365104675293, "perplexity": 31.611990994526472, "lr": 0.0026291804804649314, "grad_norm": 0.176098, "tokens_per_sec": 107870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:41.275435+00:00", "epoch": 0, "step": 10631, "train_loss": 3.474419355392456, "perplexity": 32.279080420158614, "lr": 0.0026291804804649314, "grad_norm": 0.180458, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:41.580876+00:00", "epoch": 0, "step": 10632, "train_loss": 3.527597665786743, "perplexity": 34.04208882934344, "lr": 0.0026291804804649314, "grad_norm": 0.180947, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:41.886953+00:00", "epoch": 0, "step": 10633, "train_loss": 3.5565826892852783, "perplexity": 35.0432386591881, "lr": 0.0026291804804649314, "grad_norm": 0.18242, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:42.191471+00:00", "epoch": 0, "step": 10634, "train_loss": 3.5250437259674072, "perplexity": 33.955258310305645, "lr": 0.0026291804804649314, "grad_norm": 0.145939, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:42.496293+00:00", "epoch": 0, "step": 10635, "train_loss": 3.4504525661468506, "perplexity": 31.514651546300104, "lr": 0.0026291804804649314, "grad_norm": 0.157833, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:42.801175+00:00", "epoch": 0, "step": 10636, "train_loss": 3.5602118968963623, "perplexity": 35.17064890692255, "lr": 0.0026291804804649314, "grad_norm": 0.171565, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:43.105782+00:00", "epoch": 0, "step": 10637, "train_loss": 3.43857741355896, "perplexity": 31.14262356686024, "lr": 0.0026291804804649314, "grad_norm": 0.159829, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:43.412018+00:00", "epoch": 0, "step": 10638, "train_loss": 3.566678762435913, "perplexity": 35.39882977662748, "lr": 0.0026291804804649314, "grad_norm": 0.164163, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:43.717837+00:00", "epoch": 0, "step": 10639, "train_loss": 3.4344630241394043, "perplexity": 31.01475391910598, "lr": 0.0026291804804649314, "grad_norm": 0.202941, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:44.023665+00:00", "epoch": 0, "step": 10640, "train_loss": 3.49812650680542, "perplexity": 33.0534684657684, "lr": 0.0026291804804649314, "grad_norm": 0.193114, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:44.328190+00:00", "epoch": 0, "step": 10641, "train_loss": 3.5391180515289307, "perplexity": 34.43653454499004, "lr": 0.0026291804804649314, "grad_norm": 0.16792, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:44.632577+00:00", "epoch": 0, "step": 10642, "train_loss": 3.59051513671875, "perplexity": 36.25274623792479, "lr": 0.0026291804804649314, "grad_norm": 0.161743, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:44.937838+00:00", "epoch": 0, "step": 10643, "train_loss": 3.601133108139038, "perplexity": 36.639727704736906, "lr": 0.0026291804804649314, "grad_norm": 0.162387, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:45.243754+00:00", "epoch": 0, "step": 10644, "train_loss": 3.584407329559326, "perplexity": 36.03199629022117, "lr": 0.0026291804804649314, "grad_norm": 0.187445, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:45.547528+00:00", "epoch": 0, "step": 10645, "train_loss": 3.563082456588745, "perplexity": 35.27175339784186, "lr": 0.0026291804804649314, "grad_norm": 0.227211, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:45.851846+00:00", "epoch": 0, "step": 10646, "train_loss": 3.622727155685425, "perplexity": 37.439532155529776, "lr": 0.0026291804804649314, "grad_norm": 0.183372, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:46.156619+00:00", "epoch": 0, "step": 10647, "train_loss": 3.3755156993865967, "perplexity": 29.239358613161347, "lr": 0.0026291804804649314, "grad_norm": 0.169089, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:46.462462+00:00", "epoch": 0, "step": 10648, "train_loss": 3.531292676925659, "perplexity": 34.16810740339852, "lr": 0.0026291804804649314, "grad_norm": 0.163148, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:46.767735+00:00", "epoch": 0, "step": 10649, "train_loss": 3.5567078590393066, "perplexity": 35.04762528728227, "lr": 0.0026291804804649314, "grad_norm": 0.176164, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:47.073037+00:00", "epoch": 0, "step": 10650, "train_loss": 3.473310947418213, "perplexity": 32.24332185122214, "lr": 0.0026291804804649314, "grad_norm": 0.179177, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:47.377867+00:00", "epoch": 0, "step": 10651, "train_loss": 3.610621929168701, "perplexity": 36.989050231960775, "lr": 0.0026291804804649314, "grad_norm": 0.1401, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:47.683342+00:00", "epoch": 0, "step": 10652, "train_loss": 3.4699292182922363, "perplexity": 32.1344678317424, "lr": 0.0026291804804649314, "grad_norm": 0.162482, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:47.988608+00:00", "epoch": 0, "step": 10653, "train_loss": 3.5228660106658936, "perplexity": 33.881393881756075, "lr": 0.0026291804804649314, "grad_norm": 0.178888, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:48.293948+00:00", "epoch": 0, "step": 10654, "train_loss": 3.5499050617218018, "perplexity": 34.810012528067446, "lr": 0.0026291804804649314, "grad_norm": 0.159694, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:48.599255+00:00", "epoch": 0, "step": 10655, "train_loss": 3.5206053256988525, "perplexity": 33.80488523750052, "lr": 0.0026291804804649314, "grad_norm": 0.178946, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:48.904739+00:00", "epoch": 0, "step": 10656, "train_loss": 3.442819833755493, "perplexity": 31.27502431327321, "lr": 0.0026291804804649314, "grad_norm": 0.166862, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:49.209559+00:00", "epoch": 0, "step": 10657, "train_loss": 3.6548750400543213, "perplexity": 38.66268952570717, "lr": 0.0026291804804649314, "grad_norm": 0.158527, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:49.514078+00:00", "epoch": 0, "step": 10658, "train_loss": 3.523921012878418, "perplexity": 33.91715768939565, "lr": 0.0026291804804649314, "grad_norm": 0.159531, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:49.819221+00:00", "epoch": 0, "step": 10659, "train_loss": 3.5418381690979004, "perplexity": 34.53033348185319, "lr": 0.0026291804804649314, "grad_norm": 0.158302, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:50.124874+00:00", "epoch": 0, "step": 10660, "train_loss": 3.583503007888794, "perplexity": 35.999426504144694, "lr": 0.0026291804804649314, "grad_norm": 0.146848, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:50.430207+00:00", "epoch": 0, "step": 10661, "train_loss": 3.623976469039917, "perplexity": 37.48633509271741, "lr": 0.0026291804804649314, "grad_norm": 0.152414, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:50.734602+00:00", "epoch": 0, "step": 10662, "train_loss": 3.5262019634246826, "perplexity": 33.99460934687521, "lr": 0.0026291804804649314, "grad_norm": 0.156722, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:51.039115+00:00", "epoch": 0, "step": 10663, "train_loss": 3.562685012817383, "perplexity": 35.25773764457028, "lr": 0.0026291804804649314, "grad_norm": 0.150506, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:51.345725+00:00", "epoch": 0, "step": 10664, "train_loss": 3.484551429748535, "perplexity": 32.60779694110192, "lr": 0.0026291804804649314, "grad_norm": 0.158702, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:51.650944+00:00", "epoch": 0, "step": 10665, "train_loss": 3.5051136016845703, "perplexity": 33.28522489525267, "lr": 0.0026291804804649314, "grad_norm": 0.167783, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:51.958299+00:00", "epoch": 0, "step": 10666, "train_loss": 3.4666225910186768, "perplexity": 32.028386606181215, "lr": 0.0026291804804649314, "grad_norm": 0.155005, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:52.264722+00:00", "epoch": 0, "step": 10667, "train_loss": 3.524287462234497, "perplexity": 33.92958888755414, "lr": 0.0026291804804649314, "grad_norm": 0.191101, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:52.570070+00:00", "epoch": 0, "step": 10668, "train_loss": 3.547006607055664, "perplexity": 34.709263363802926, "lr": 0.0026291804804649314, "grad_norm": 0.161603, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:52.874776+00:00", "epoch": 0, "step": 10669, "train_loss": 3.5329673290252686, "perplexity": 34.225375034496906, "lr": 0.0026291804804649314, "grad_norm": 0.156248, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:53.180598+00:00", "epoch": 0, "step": 10670, "train_loss": 3.5853872299194336, "perplexity": 36.06732136105956, "lr": 0.0026291804804649314, "grad_norm": 0.150354, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:53.486837+00:00", "epoch": 0, "step": 10671, "train_loss": 3.5756828784942627, "perplexity": 35.71900422907852, "lr": 0.0026291804804649314, "grad_norm": 0.158831, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:53.792747+00:00", "epoch": 0, "step": 10672, "train_loss": 3.543008804321289, "perplexity": 34.57077957569456, "lr": 0.0026291804804649314, "grad_norm": 0.150818, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:54.097918+00:00", "epoch": 0, "step": 10673, "train_loss": 3.6032538414001465, "perplexity": 36.717513245997154, "lr": 0.0026291804804649314, "grad_norm": 0.151102, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:54.401648+00:00", "epoch": 0, "step": 10674, "train_loss": 3.5582690238952637, "perplexity": 35.10238314006172, "lr": 0.0026291804804649314, "grad_norm": 0.149506, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:54.706998+00:00", "epoch": 0, "step": 10675, "train_loss": 3.486044406890869, "perplexity": 32.65651599576487, "lr": 0.0026291804804649314, "grad_norm": 0.162584, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:55.012786+00:00", "epoch": 0, "step": 10676, "train_loss": 3.514425039291382, "perplexity": 33.59660564149167, "lr": 0.0026291804804649314, "grad_norm": 0.152081, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:55.318433+00:00", "epoch": 0, "step": 10677, "train_loss": 3.594500780105591, "perplexity": 36.39752508299078, "lr": 0.0026291804804649314, "grad_norm": 0.170729, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:55.623705+00:00", "epoch": 0, "step": 10678, "train_loss": 3.534881591796875, "perplexity": 34.290954143567575, "lr": 0.0026291804804649314, "grad_norm": 0.185869, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:55.928256+00:00", "epoch": 0, "step": 10679, "train_loss": 3.5702786445617676, "perplexity": 35.526491036094484, "lr": 0.0026291804804649314, "grad_norm": 0.192952, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:56.235626+00:00", "epoch": 0, "step": 10680, "train_loss": 3.5405309200286865, "perplexity": 34.48522322714271, "lr": 0.0026291804804649314, "grad_norm": 0.163593, "tokens_per_sec": 106674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:56.541113+00:00", "epoch": 0, "step": 10681, "train_loss": 3.465406656265259, "perplexity": 31.98946584515918, "lr": 0.0026291804804649314, "grad_norm": 0.157993, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:56.846383+00:00", "epoch": 0, "step": 10682, "train_loss": 3.5239779949188232, "perplexity": 33.91909041331029, "lr": 0.0026291804804649314, "grad_norm": 0.144985, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:57.151796+00:00", "epoch": 0, "step": 10683, "train_loss": 3.5630974769592285, "perplexity": 35.27228319662437, "lr": 0.0026291804804649314, "grad_norm": 0.153557, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:57.457757+00:00", "epoch": 0, "step": 10684, "train_loss": 3.426420211791992, "perplexity": 30.76630851248617, "lr": 0.0026291804804649314, "grad_norm": 0.141213, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:57.762911+00:00", "epoch": 0, "step": 10685, "train_loss": 3.564457416534424, "perplexity": 35.320284002150636, "lr": 0.0026291804804649314, "grad_norm": 0.15873, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:58.068642+00:00", "epoch": 0, "step": 10686, "train_loss": 3.490030288696289, "perplexity": 32.78694076488366, "lr": 0.0026291804804649314, "grad_norm": 0.168177, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:58.374419+00:00", "epoch": 0, "step": 10687, "train_loss": 3.5787785053253174, "perplexity": 35.82974825960731, "lr": 0.0026291804804649314, "grad_norm": 0.16821, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:58.680469+00:00", "epoch": 0, "step": 10688, "train_loss": 3.575533151626587, "perplexity": 35.71365653481555, "lr": 0.0026291804804649314, "grad_norm": 0.180435, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:58.985154+00:00", "epoch": 0, "step": 10689, "train_loss": 3.5145106315612793, "perplexity": 33.599481374297866, "lr": 0.0026291804804649314, "grad_norm": 0.161902, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:59.289206+00:00", "epoch": 0, "step": 10690, "train_loss": 3.587503671646118, "perplexity": 36.14373658061341, "lr": 0.0026291804804649314, "grad_norm": 0.193439, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:59.593770+00:00", "epoch": 0, "step": 10691, "train_loss": 3.476555824279785, "perplexity": 32.3481173925998, "lr": 0.0026291804804649314, "grad_norm": 0.15183, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:53:59.899739+00:00", "epoch": 0, "step": 10692, "train_loss": 3.5155932903289795, "perplexity": 33.63587784632253, "lr": 0.0026291804804649314, "grad_norm": 0.144439, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:00.204715+00:00", "epoch": 0, "step": 10693, "train_loss": 3.4711458683013916, "perplexity": 32.17358802529041, "lr": 0.0026291804804649314, "grad_norm": 0.152274, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:00.509563+00:00", "epoch": 0, "step": 10694, "train_loss": 3.500229597091675, "perplexity": 33.123056043055946, "lr": 0.0026291804804649314, "grad_norm": 0.168644, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:00.814365+00:00", "epoch": 0, "step": 10695, "train_loss": 3.3855433464050293, "perplexity": 29.534035569396124, "lr": 0.0026291804804649314, "grad_norm": 0.158151, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:01.118795+00:00", "epoch": 0, "step": 10696, "train_loss": 3.527780294418335, "perplexity": 34.04830645718402, "lr": 0.0026291804804649314, "grad_norm": 0.158803, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:01.423235+00:00", "epoch": 0, "step": 10697, "train_loss": 3.4840919971466064, "perplexity": 32.59281929698256, "lr": 0.0026291804804649314, "grad_norm": 0.151881, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:01.728962+00:00", "epoch": 0, "step": 10698, "train_loss": 3.4963014125823975, "perplexity": 32.99319778798357, "lr": 0.0026291804804649314, "grad_norm": 0.147315, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:02.035888+00:00", "epoch": 0, "step": 10699, "train_loss": 3.579591989517212, "perplexity": 35.858907051916354, "lr": 0.0026291804804649314, "grad_norm": 0.165655, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:02.340578+00:00", "epoch": 0, "step": 10700, "train_loss": 3.633262872695923, "perplexity": 37.83606970702959, "lr": 0.0026291804804649314, "grad_norm": 0.155184, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:02.644923+00:00", "epoch": 0, "step": 10701, "train_loss": 3.5740809440612793, "perplexity": 35.661830532775035, "lr": 0.0026291804804649314, "grad_norm": 0.176664, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:02.950241+00:00", "epoch": 0, "step": 10702, "train_loss": 3.53830885887146, "perplexity": 34.40868002544376, "lr": 0.0026291804804649314, "grad_norm": 0.180436, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:03.255897+00:00", "epoch": 0, "step": 10703, "train_loss": 3.510342836380005, "perplexity": 33.45973703291495, "lr": 0.0026291804804649314, "grad_norm": 0.154797, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:03.562007+00:00", "epoch": 0, "step": 10704, "train_loss": 3.5129857063293457, "perplexity": 33.54828372358625, "lr": 0.0026291804804649314, "grad_norm": 0.148054, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:03.866646+00:00", "epoch": 0, "step": 10705, "train_loss": 3.5396087169647217, "perplexity": 34.453435508239856, "lr": 0.0026291804804649314, "grad_norm": 0.166258, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:04.172782+00:00", "epoch": 0, "step": 10706, "train_loss": 3.4508635997772217, "perplexity": 31.52760779047869, "lr": 0.0026291804804649314, "grad_norm": 0.145949, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:04.477365+00:00", "epoch": 0, "step": 10707, "train_loss": 3.3890674114227295, "perplexity": 29.638299038695862, "lr": 0.0026291804804649314, "grad_norm": 0.175745, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:04.784089+00:00", "epoch": 0, "step": 10708, "train_loss": 3.513716220855713, "perplexity": 33.57280018590667, "lr": 0.0026291804804649314, "grad_norm": 0.158617, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:05.089087+00:00", "epoch": 0, "step": 10709, "train_loss": 3.551797866821289, "perplexity": 34.87596349376561, "lr": 0.0026291804804649314, "grad_norm": 0.145889, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:05.393243+00:00", "epoch": 0, "step": 10710, "train_loss": 3.4571168422698975, "perplexity": 31.725375266757627, "lr": 0.0026291804804649314, "grad_norm": 0.15742, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:05.697348+00:00", "epoch": 0, "step": 10711, "train_loss": 3.4958677291870117, "perplexity": 32.97889228819532, "lr": 0.0026291804804649314, "grad_norm": 0.171029, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:06.001501+00:00", "epoch": 0, "step": 10712, "train_loss": 3.467949390411377, "perplexity": 32.07091005388357, "lr": 0.0026291804804649314, "grad_norm": 0.181952, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:06.307146+00:00", "epoch": 0, "step": 10713, "train_loss": 3.498042345046997, "perplexity": 33.05068674479925, "lr": 0.0026291804804649314, "grad_norm": 0.169196, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:06.612786+00:00", "epoch": 0, "step": 10714, "train_loss": 3.5591108798980713, "perplexity": 35.13194673441969, "lr": 0.0026291804804649314, "grad_norm": 0.164676, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:06.918943+00:00", "epoch": 0, "step": 10715, "train_loss": 3.448544979095459, "perplexity": 31.45459190778816, "lr": 0.0026291804804649314, "grad_norm": 0.145343, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:07.224578+00:00", "epoch": 0, "step": 10716, "train_loss": 3.441087007522583, "perplexity": 31.220877058149263, "lr": 0.0026291804804649314, "grad_norm": 0.157386, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:07.528659+00:00", "epoch": 0, "step": 10717, "train_loss": 3.593853235244751, "perplexity": 36.37396368203136, "lr": 0.0026291804804649314, "grad_norm": 0.151926, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:07.835356+00:00", "epoch": 0, "step": 10718, "train_loss": 3.5105350017547607, "perplexity": 33.46616745365343, "lr": 0.0026291804804649314, "grad_norm": 0.157191, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:08.142100+00:00", "epoch": 0, "step": 10719, "train_loss": 3.5359630584716797, "perplexity": 34.32805872779039, "lr": 0.0026291804804649314, "grad_norm": 0.171033, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:08.448592+00:00", "epoch": 0, "step": 10720, "train_loss": 3.657052516937256, "perplexity": 38.74696836269109, "lr": 0.0026291804804649314, "grad_norm": 0.176145, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:08.754146+00:00", "epoch": 0, "step": 10721, "train_loss": 3.630964517593384, "perplexity": 37.749208839953354, "lr": 0.0026291804804649314, "grad_norm": 0.194035, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:09.059380+00:00", "epoch": 0, "step": 10722, "train_loss": 3.502451181411743, "perplexity": 33.19672350393352, "lr": 0.0026291804804649314, "grad_norm": 0.218289, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:09.363486+00:00", "epoch": 0, "step": 10723, "train_loss": 3.4959328174591064, "perplexity": 32.98103889716873, "lr": 0.0026291804804649314, "grad_norm": 0.182704, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:09.668690+00:00", "epoch": 0, "step": 10724, "train_loss": 3.5516738891601562, "perplexity": 34.871639921400636, "lr": 0.0026291804804649314, "grad_norm": 0.164729, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:09.973222+00:00", "epoch": 0, "step": 10725, "train_loss": 3.458582878112793, "perplexity": 31.771919913752015, "lr": 0.0026291804804649314, "grad_norm": 0.178498, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:10.279075+00:00", "epoch": 0, "step": 10726, "train_loss": 3.527111291885376, "perplexity": 34.02553567162709, "lr": 0.0026291804804649314, "grad_norm": 0.152328, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:10.584405+00:00", "epoch": 0, "step": 10727, "train_loss": 3.584725856781006, "perplexity": 36.04347528998097, "lr": 0.0026291804804649314, "grad_norm": 0.168416, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:10.888608+00:00", "epoch": 0, "step": 10728, "train_loss": 3.5590341091156006, "perplexity": 35.12924973090603, "lr": 0.0026291804804649314, "grad_norm": 0.183392, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:11.194329+00:00", "epoch": 0, "step": 10729, "train_loss": 3.5112462043762207, "perplexity": 33.4899771454307, "lr": 0.0026291804804649314, "grad_norm": 0.167897, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:11.501714+00:00", "epoch": 0, "step": 10730, "train_loss": 3.556835651397705, "perplexity": 35.052104392165596, "lr": 0.0026291804804649314, "grad_norm": 0.174891, "tokens_per_sec": 106677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:11.807328+00:00", "epoch": 0, "step": 10731, "train_loss": 3.587684392929077, "perplexity": 36.15026911332525, "lr": 0.0026291804804649314, "grad_norm": 0.19326, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:12.112088+00:00", "epoch": 0, "step": 10732, "train_loss": 3.550142765045166, "perplexity": 34.81828796724245, "lr": 0.0026291804804649314, "grad_norm": 0.173833, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:12.416624+00:00", "epoch": 0, "step": 10733, "train_loss": 3.623199701309204, "perplexity": 37.45722822337663, "lr": 0.0026291804804649314, "grad_norm": 0.156251, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:12.721400+00:00", "epoch": 0, "step": 10734, "train_loss": 3.6049952507019043, "perplexity": 36.7815091704779, "lr": 0.0026291804804649314, "grad_norm": 0.164834, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:13.027725+00:00", "epoch": 0, "step": 10735, "train_loss": 3.4416017532348633, "perplexity": 31.236952007649304, "lr": 0.0026291804804649314, "grad_norm": 0.163016, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:13.332878+00:00", "epoch": 0, "step": 10736, "train_loss": 3.3997418880462646, "perplexity": 29.956366953036838, "lr": 0.0026291804804649314, "grad_norm": 0.167297, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:13.636848+00:00", "epoch": 0, "step": 10737, "train_loss": 3.605074644088745, "perplexity": 36.784429494989745, "lr": 0.0026291804804649314, "grad_norm": 0.177207, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:13.941863+00:00", "epoch": 0, "step": 10738, "train_loss": 3.4714531898498535, "perplexity": 32.183477181679514, "lr": 0.0026291804804649314, "grad_norm": 0.164213, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:14.246780+00:00", "epoch": 0, "step": 10739, "train_loss": 3.5456173419952393, "perplexity": 34.661076476868, "lr": 0.0026291804804649314, "grad_norm": 0.19597, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:14.551219+00:00", "epoch": 0, "step": 10740, "train_loss": 3.5732295513153076, "perplexity": 35.631481230372586, "lr": 0.0026291804804649314, "grad_norm": 0.198374, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:14.857574+00:00", "epoch": 0, "step": 10741, "train_loss": 3.528132677078247, "perplexity": 34.06030660417659, "lr": 0.0026291804804649314, "grad_norm": 0.194757, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:15.163302+00:00", "epoch": 0, "step": 10742, "train_loss": 3.533066749572754, "perplexity": 34.228777909175754, "lr": 0.0026291804804649314, "grad_norm": 0.16994, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:15.467691+00:00", "epoch": 0, "step": 10743, "train_loss": 3.435009479522705, "perplexity": 31.031706729902325, "lr": 0.0026291804804649314, "grad_norm": 0.149447, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:15.772306+00:00", "epoch": 0, "step": 10744, "train_loss": 3.3923211097717285, "perplexity": 29.734890177345463, "lr": 0.0026291804804649314, "grad_norm": 0.17959, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:16.078473+00:00", "epoch": 0, "step": 10745, "train_loss": 3.4988741874694824, "perplexity": 33.07819114620379, "lr": 0.0026291804804649314, "grad_norm": 0.167837, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:16.384644+00:00", "epoch": 0, "step": 10746, "train_loss": 3.5335493087768555, "perplexity": 34.24529930695665, "lr": 0.0026291804804649314, "grad_norm": 0.159574, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:16.691518+00:00", "epoch": 0, "step": 10747, "train_loss": 3.469860792160034, "perplexity": 32.13226906962555, "lr": 0.0026291804804649314, "grad_norm": 0.171664, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:16.995907+00:00", "epoch": 0, "step": 10748, "train_loss": 3.4334185123443604, "perplexity": 30.982375555552153, "lr": 0.0026291804804649314, "grad_norm": 0.146873, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:17.300423+00:00", "epoch": 0, "step": 10749, "train_loss": 3.51703143119812, "perplexity": 33.6842857772914, "lr": 0.0026291804804649314, "grad_norm": 0.166323, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:17.606120+00:00", "epoch": 0, "step": 10750, "train_loss": 3.5429465770721436, "perplexity": 34.56862839811237, "lr": 0.0026291804804649314, "grad_norm": 0.170629, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:17.911754+00:00", "epoch": 0, "step": 10751, "train_loss": 3.6127400398254395, "perplexity": 37.06748016575673, "lr": 0.0026291804804649314, "grad_norm": 0.15353, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:18.217662+00:00", "epoch": 0, "step": 10752, "train_loss": 3.6221566200256348, "perplexity": 37.41817765967914, "lr": 0.0026291804804649314, "grad_norm": 0.199512, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:18.522807+00:00", "epoch": 0, "step": 10753, "train_loss": 3.4745328426361084, "perplexity": 32.28274389189786, "lr": 0.0026291804804649314, "grad_norm": 0.190552, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:18.828622+00:00", "epoch": 0, "step": 10754, "train_loss": 3.4561023712158203, "perplexity": 31.6932071114598, "lr": 0.0026291804804649314, "grad_norm": 0.166282, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:19.133371+00:00", "epoch": 0, "step": 10755, "train_loss": 3.498063564300537, "perplexity": 33.05138806314166, "lr": 0.0026291804804649314, "grad_norm": 0.190651, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:19.438595+00:00", "epoch": 0, "step": 10756, "train_loss": 3.518911123275757, "perplexity": 33.74766140708232, "lr": 0.0026291804804649314, "grad_norm": 0.160693, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:19.744138+00:00", "epoch": 0, "step": 10757, "train_loss": 3.5736565589904785, "perplexity": 35.64669939524103, "lr": 0.0026291804804649314, "grad_norm": 0.173429, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:20.049683+00:00", "epoch": 0, "step": 10758, "train_loss": 3.4484732151031494, "perplexity": 31.452334681691134, "lr": 0.0026291804804649314, "grad_norm": 0.171432, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:20.353979+00:00", "epoch": 0, "step": 10759, "train_loss": 3.4762635231018066, "perplexity": 32.33866338155715, "lr": 0.0026291804804649314, "grad_norm": 0.160287, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:20.658560+00:00", "epoch": 0, "step": 10760, "train_loss": 3.584603786468506, "perplexity": 36.03907572022263, "lr": 0.0026291804804649314, "grad_norm": 0.14557, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:20.963937+00:00", "epoch": 0, "step": 10761, "train_loss": 3.5113587379455566, "perplexity": 33.493746104159065, "lr": 0.0026291804804649314, "grad_norm": 0.141654, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:21.268842+00:00", "epoch": 0, "step": 10762, "train_loss": 3.6419036388397217, "perplexity": 38.16441888845931, "lr": 0.0026291804804649314, "grad_norm": 0.160665, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:21.877588+00:00", "epoch": 0, "step": 10763, "train_loss": 3.579442024230957, "perplexity": 35.853529863861375, "lr": 0.0026291804804649314, "grad_norm": 0.14732, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:22.181107+00:00", "epoch": 0, "step": 10764, "train_loss": 3.450833559036255, "perplexity": 31.526660692005596, "lr": 0.0026291804804649314, "grad_norm": 0.159523, "tokens_per_sec": 107959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:22.490178+00:00", "epoch": 0, "step": 10765, "train_loss": 3.484816789627075, "perplexity": 32.61645089029234, "lr": 0.0026291804804649314, "grad_norm": 0.183886, "tokens_per_sec": 106022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:22.795798+00:00", "epoch": 0, "step": 10766, "train_loss": 3.5123610496520996, "perplexity": 33.527334107988004, "lr": 0.0026291804804649314, "grad_norm": 0.191372, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:23.100080+00:00", "epoch": 0, "step": 10767, "train_loss": 3.5152981281280518, "perplexity": 33.62595127163416, "lr": 0.0026291804804649314, "grad_norm": 0.17719, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:23.404728+00:00", "epoch": 0, "step": 10768, "train_loss": 3.661308526992798, "perplexity": 38.91222727200133, "lr": 0.0026291804804649314, "grad_norm": 0.165547, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:23.708726+00:00", "epoch": 0, "step": 10769, "train_loss": 3.4721944332122803, "perplexity": 32.207341814177774, "lr": 0.0026291804804649314, "grad_norm": 0.158885, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:24.014813+00:00", "epoch": 0, "step": 10770, "train_loss": 3.5079538822174072, "perplexity": 33.379898657930354, "lr": 0.0026291804804649314, "grad_norm": 0.14669, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:24.320394+00:00", "epoch": 0, "step": 10771, "train_loss": 3.5259897708892822, "perplexity": 33.98739670978888, "lr": 0.0026291804804649314, "grad_norm": 0.156703, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:24.625288+00:00", "epoch": 0, "step": 10772, "train_loss": 3.5302364826202393, "perplexity": 34.13203829427897, "lr": 0.0026291804804649314, "grad_norm": 0.154863, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:24.929311+00:00", "epoch": 0, "step": 10773, "train_loss": 3.535196542739868, "perplexity": 34.30175581281989, "lr": 0.0026291804804649314, "grad_norm": 0.181912, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:25.234992+00:00", "epoch": 0, "step": 10774, "train_loss": 3.5180118083953857, "perplexity": 33.71732527591287, "lr": 0.0026291804804649314, "grad_norm": 0.213592, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:25.539630+00:00", "epoch": 0, "step": 10775, "train_loss": 3.631352663040161, "perplexity": 37.76386386744113, "lr": 0.0026291804804649314, "grad_norm": 0.169213, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:25.843145+00:00", "epoch": 0, "step": 10776, "train_loss": 3.6413121223449707, "perplexity": 38.141850680565845, "lr": 0.0026291804804649314, "grad_norm": 0.187215, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:26.148419+00:00", "epoch": 0, "step": 10777, "train_loss": 3.5180675983428955, "perplexity": 33.71920641619405, "lr": 0.0026291804804649314, "grad_norm": 0.216591, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:26.453325+00:00", "epoch": 0, "step": 10778, "train_loss": 3.5571765899658203, "perplexity": 35.064057043890685, "lr": 0.0026291804804649314, "grad_norm": 0.166961, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:26.758762+00:00", "epoch": 0, "step": 10779, "train_loss": 3.4883832931518555, "perplexity": 32.73298526395971, "lr": 0.0026291804804649314, "grad_norm": 0.188863, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:27.063850+00:00", "epoch": 0, "step": 10780, "train_loss": 3.503589153289795, "perplexity": 33.2345219444266, "lr": 0.0026291804804649314, "grad_norm": 0.175935, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:27.367562+00:00", "epoch": 0, "step": 10781, "train_loss": 3.563082218170166, "perplexity": 35.27174498840154, "lr": 0.0026291804804649314, "grad_norm": 0.173025, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:27.671528+00:00", "epoch": 0, "step": 10782, "train_loss": 3.476362705230713, "perplexity": 32.341870958101815, "lr": 0.0026291804804649314, "grad_norm": 0.201579, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:27.976827+00:00", "epoch": 0, "step": 10783, "train_loss": 3.570871353149414, "perplexity": 35.547554133943386, "lr": 0.0026291804804649314, "grad_norm": 0.184515, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:28.281409+00:00", "epoch": 0, "step": 10784, "train_loss": 3.610426664352417, "perplexity": 36.98182827698251, "lr": 0.0026291804804649314, "grad_norm": 0.168178, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:28.585382+00:00", "epoch": 0, "step": 10785, "train_loss": 3.4764199256896973, "perplexity": 32.34372162775003, "lr": 0.0026291804804649314, "grad_norm": 0.174488, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:28.890420+00:00", "epoch": 0, "step": 10786, "train_loss": 3.547776937484741, "perplexity": 34.736011266581926, "lr": 0.0026291804804649314, "grad_norm": 0.170716, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:29.194873+00:00", "epoch": 0, "step": 10787, "train_loss": 3.527003765106201, "perplexity": 34.021877212061334, "lr": 0.0026291804804649314, "grad_norm": 0.167596, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:29.500012+00:00", "epoch": 0, "step": 10788, "train_loss": 3.447101354598999, "perplexity": 31.409216049114956, "lr": 0.0026291804804649314, "grad_norm": 0.16767, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:29.804778+00:00", "epoch": 0, "step": 10789, "train_loss": 3.5998644828796387, "perplexity": 36.593275092381965, "lr": 0.0026291804804649314, "grad_norm": 0.166965, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:30.108979+00:00", "epoch": 0, "step": 10790, "train_loss": 3.60135555267334, "perplexity": 36.647878918465985, "lr": 0.0026291804804649314, "grad_norm": 0.163959, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:30.413565+00:00", "epoch": 0, "step": 10791, "train_loss": 3.592038154602051, "perplexity": 36.30800188575297, "lr": 0.0026291804804649314, "grad_norm": 0.172409, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:30.718053+00:00", "epoch": 0, "step": 10792, "train_loss": 3.488443613052368, "perplexity": 32.73495977392484, "lr": 0.0026291804804649314, "grad_norm": 0.178388, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:31.022134+00:00", "epoch": 0, "step": 10793, "train_loss": 3.5597503185272217, "perplexity": 35.1544186422162, "lr": 0.0026291804804649314, "grad_norm": 0.156061, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:31.328282+00:00", "epoch": 0, "step": 10794, "train_loss": 3.4782662391662598, "perplexity": 32.40349343864397, "lr": 0.0026291804804649314, "grad_norm": 0.152621, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:31.633360+00:00", "epoch": 0, "step": 10795, "train_loss": 3.5141220092773438, "perplexity": 33.58642640399763, "lr": 0.0026291804804649314, "grad_norm": 0.167288, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:31.937385+00:00", "epoch": 0, "step": 10796, "train_loss": 3.6574225425720215, "perplexity": 38.761308387179454, "lr": 0.0026291804804649314, "grad_norm": 0.151883, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:32.242542+00:00", "epoch": 0, "step": 10797, "train_loss": 3.581068754196167, "perplexity": 35.91190133969177, "lr": 0.0026291804804649314, "grad_norm": 0.155202, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:32.547659+00:00", "epoch": 0, "step": 10798, "train_loss": 3.461400032043457, "perplexity": 31.86155249793554, "lr": 0.0026291804804649314, "grad_norm": 0.166176, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:32.852647+00:00", "epoch": 0, "step": 10799, "train_loss": 3.5681874752044678, "perplexity": 35.45227675103167, "lr": 0.0026291804804649314, "grad_norm": 0.170414, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:33.158099+00:00", "epoch": 0, "step": 10800, "train_loss": 3.5259389877319336, "perplexity": 33.9856707662987, "lr": 0.0026291804804649314, "grad_norm": 0.17061, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:33.462954+00:00", "epoch": 0, "step": 10801, "train_loss": 3.544795036315918, "perplexity": 34.63258619229648, "lr": 0.0026291804804649314, "grad_norm": 0.166719, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:33.767838+00:00", "epoch": 0, "step": 10802, "train_loss": 3.551445245742798, "perplexity": 34.86366766191717, "lr": 0.0026291804804649314, "grad_norm": 0.159782, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:34.072465+00:00", "epoch": 0, "step": 10803, "train_loss": 3.5221283435821533, "perplexity": 33.856409908798426, "lr": 0.0026291804804649314, "grad_norm": 0.165056, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:34.377029+00:00", "epoch": 0, "step": 10804, "train_loss": 3.445983648300171, "perplexity": 31.3741293824425, "lr": 0.0026291804804649314, "grad_norm": 0.153686, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:34.681492+00:00", "epoch": 0, "step": 10805, "train_loss": 3.4481594562530518, "perplexity": 31.44246778132312, "lr": 0.0026291804804649314, "grad_norm": 0.188828, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:34.985407+00:00", "epoch": 0, "step": 10806, "train_loss": 3.4872512817382812, "perplexity": 32.69595211596833, "lr": 0.0026291804804649314, "grad_norm": 0.193331, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:35.290400+00:00", "epoch": 0, "step": 10807, "train_loss": 3.561483860015869, "perplexity": 35.21541313841175, "lr": 0.0026291804804649314, "grad_norm": 0.174221, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:35.594583+00:00", "epoch": 0, "step": 10808, "train_loss": 3.6674482822418213, "perplexity": 39.15187375620455, "lr": 0.0026291804804649314, "grad_norm": 0.176666, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:35.899560+00:00", "epoch": 0, "step": 10809, "train_loss": 3.6722805500030518, "perplexity": 39.3415239447541, "lr": 0.0026291804804649314, "grad_norm": 0.16446, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:36.205406+00:00", "epoch": 0, "step": 10810, "train_loss": 3.533921003341675, "perplexity": 34.258030464484115, "lr": 0.0026291804804649314, "grad_norm": 0.178824, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:36.510903+00:00", "epoch": 0, "step": 10811, "train_loss": 3.540294647216797, "perplexity": 34.477076268972446, "lr": 0.0026291804804649314, "grad_norm": 0.187079, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:36.815065+00:00", "epoch": 0, "step": 10812, "train_loss": 3.572437286376953, "perplexity": 35.60326283679007, "lr": 0.0026291804804649314, "grad_norm": 0.164436, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:37.119676+00:00", "epoch": 0, "step": 10813, "train_loss": 3.6446259021759033, "perplexity": 38.26845402802118, "lr": 0.0026291804804649314, "grad_norm": 0.162983, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:37.423968+00:00", "epoch": 0, "step": 10814, "train_loss": 3.5612294673919678, "perplexity": 35.206455736458494, "lr": 0.0026291804804649314, "grad_norm": 0.142786, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:37.729111+00:00", "epoch": 0, "step": 10815, "train_loss": 3.557936429977417, "perplexity": 35.090710242202434, "lr": 0.0026291804804649314, "grad_norm": 0.161659, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:38.033579+00:00", "epoch": 0, "step": 10816, "train_loss": 3.5449438095092773, "perplexity": 34.63773897602714, "lr": 0.0026291804804649314, "grad_norm": 0.166139, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:38.336931+00:00", "epoch": 0, "step": 10817, "train_loss": 3.5610477924346924, "perplexity": 35.20006018609001, "lr": 0.0026291804804649314, "grad_norm": 0.155633, "tokens_per_sec": 108019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:38.640903+00:00", "epoch": 0, "step": 10818, "train_loss": 3.5643911361694336, "perplexity": 35.31794303841623, "lr": 0.0026291804804649314, "grad_norm": 0.149841, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:38.945709+00:00", "epoch": 0, "step": 10819, "train_loss": 3.500606060028076, "perplexity": 33.135527993462745, "lr": 0.0026291804804649314, "grad_norm": 0.156029, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:39.251696+00:00", "epoch": 0, "step": 10820, "train_loss": 3.5461480617523193, "perplexity": 34.67947667719897, "lr": 0.0026291804804649314, "grad_norm": 0.155307, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:39.557150+00:00", "epoch": 0, "step": 10821, "train_loss": 3.4173741340637207, "perplexity": 30.489249133953663, "lr": 0.0026291804804649314, "grad_norm": 0.147231, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:39.861724+00:00", "epoch": 0, "step": 10822, "train_loss": 3.580703020095825, "perplexity": 35.898769534284625, "lr": 0.0026291804804649314, "grad_norm": 0.157725, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:40.165382+00:00", "epoch": 0, "step": 10823, "train_loss": 3.5671393871307373, "perplexity": 35.41513910774244, "lr": 0.0026291804804649314, "grad_norm": 0.154373, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:40.469427+00:00", "epoch": 0, "step": 10824, "train_loss": 3.5246083736419678, "perplexity": 33.94047902697047, "lr": 0.0026291804804649314, "grad_norm": 0.151444, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:40.774851+00:00", "epoch": 0, "step": 10825, "train_loss": 3.4932878017425537, "perplexity": 32.89391879874406, "lr": 0.0026291804804649314, "grad_norm": 0.146936, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:41.078980+00:00", "epoch": 0, "step": 10826, "train_loss": 3.465184450149536, "perplexity": 31.982358379900052, "lr": 0.0026291804804649314, "grad_norm": 0.160784, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:41.384302+00:00", "epoch": 0, "step": 10827, "train_loss": 3.527080535888672, "perplexity": 34.02448919845686, "lr": 0.0026291804804649314, "grad_norm": 0.155696, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:41.688435+00:00", "epoch": 0, "step": 10828, "train_loss": 3.568801164627075, "perplexity": 35.474040115571405, "lr": 0.0026291804804649314, "grad_norm": 0.153951, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:41.992820+00:00", "epoch": 0, "step": 10829, "train_loss": 3.5286929607391357, "perplexity": 34.07939538451827, "lr": 0.0026291804804649314, "grad_norm": 0.152435, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:42.298029+00:00", "epoch": 0, "step": 10830, "train_loss": 3.4640939235687256, "perplexity": 31.947499778540067, "lr": 0.0026291804804649314, "grad_norm": 0.146954, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:42.602712+00:00", "epoch": 0, "step": 10831, "train_loss": 3.5343425273895264, "perplexity": 34.27247409211125, "lr": 0.0026291804804649314, "grad_norm": 0.161451, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:42.907916+00:00", "epoch": 0, "step": 10832, "train_loss": 3.575942277908325, "perplexity": 35.728270919681535, "lr": 0.0026291804804649314, "grad_norm": 0.155182, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:43.212138+00:00", "epoch": 0, "step": 10833, "train_loss": 3.5475621223449707, "perplexity": 34.72855024686531, "lr": 0.0026291804804649314, "grad_norm": 0.163322, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:43.515971+00:00", "epoch": 0, "step": 10834, "train_loss": 3.4813389778137207, "perplexity": 32.50321403444881, "lr": 0.0026291804804649314, "grad_norm": 0.151465, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:43.820436+00:00", "epoch": 0, "step": 10835, "train_loss": 3.5409722328186035, "perplexity": 34.50044535582898, "lr": 0.0026291804804649314, "grad_norm": 0.165245, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:44.125267+00:00", "epoch": 0, "step": 10836, "train_loss": 3.5693695545196533, "perplexity": 35.49420893275494, "lr": 0.0026291804804649314, "grad_norm": 0.164671, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:44.430648+00:00", "epoch": 0, "step": 10837, "train_loss": 3.476205825805664, "perplexity": 32.336797581945476, "lr": 0.0026291804804649314, "grad_norm": 0.163173, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:44.734545+00:00", "epoch": 0, "step": 10838, "train_loss": 3.6285765171051025, "perplexity": 37.65917125849251, "lr": 0.0026291804804649314, "grad_norm": 0.175412, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:45.038671+00:00", "epoch": 0, "step": 10839, "train_loss": 3.587923526763916, "perplexity": 36.15891489951758, "lr": 0.0026291804804649314, "grad_norm": 0.155003, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:45.343912+00:00", "epoch": 0, "step": 10840, "train_loss": 3.613306760787964, "perplexity": 37.08849303744965, "lr": 0.0026291804804649314, "grad_norm": 0.15511, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:45.648545+00:00", "epoch": 0, "step": 10841, "train_loss": 3.5168495178222656, "perplexity": 33.678158712466825, "lr": 0.0026291804804649314, "grad_norm": 0.182493, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:45.954393+00:00", "epoch": 0, "step": 10842, "train_loss": 3.5132148265838623, "perplexity": 33.555971195535726, "lr": 0.0026291804804649314, "grad_norm": 0.158223, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:46.260393+00:00", "epoch": 0, "step": 10843, "train_loss": 3.705280065536499, "perplexity": 40.661433588566005, "lr": 0.0026291804804649314, "grad_norm": 0.169403, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:46.565170+00:00", "epoch": 0, "step": 10844, "train_loss": 3.567600727081299, "perplexity": 35.43148129562976, "lr": 0.0026291804804649314, "grad_norm": 0.160644, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:46.870003+00:00", "epoch": 0, "step": 10845, "train_loss": 3.4980671405792236, "perplexity": 33.051506264327706, "lr": 0.0026291804804649314, "grad_norm": 0.172912, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:47.174298+00:00", "epoch": 0, "step": 10846, "train_loss": 3.554990291595459, "perplexity": 34.98748029343217, "lr": 0.0026291804804649314, "grad_norm": 0.195588, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:47.478997+00:00", "epoch": 0, "step": 10847, "train_loss": 3.5953786373138428, "perplexity": 36.429490941424206, "lr": 0.0026291804804649314, "grad_norm": 0.159709, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:47.782839+00:00", "epoch": 0, "step": 10848, "train_loss": 3.496809959411621, "perplexity": 33.009980641176284, "lr": 0.0026291804804649314, "grad_norm": 0.204379, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:48.087419+00:00", "epoch": 0, "step": 10849, "train_loss": 3.6060168743133545, "perplexity": 36.81910522995229, "lr": 0.0026291804804649314, "grad_norm": 0.230025, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:48.392076+00:00", "epoch": 0, "step": 10850, "train_loss": 3.545290231704712, "perplexity": 34.64974033625686, "lr": 0.0026291804804649314, "grad_norm": 0.202607, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:48.696355+00:00", "epoch": 0, "step": 10851, "train_loss": 3.5771570205688477, "perplexity": 35.771697945544766, "lr": 0.0026291804804649314, "grad_norm": 0.207588, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:49.001898+00:00", "epoch": 0, "step": 10852, "train_loss": 3.564396619796753, "perplexity": 35.318136709384554, "lr": 0.0026291804804649314, "grad_norm": 0.217574, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:49.306957+00:00", "epoch": 0, "step": 10853, "train_loss": 3.5266315937042236, "perplexity": 34.00921759824065, "lr": 0.0026291804804649314, "grad_norm": 0.203133, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:49.611172+00:00", "epoch": 0, "step": 10854, "train_loss": 3.547297954559326, "perplexity": 34.71937729430058, "lr": 0.0026291804804649314, "grad_norm": 0.168453, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:49.915327+00:00", "epoch": 0, "step": 10855, "train_loss": 3.52795147895813, "perplexity": 34.05413549976244, "lr": 0.0026291804804649314, "grad_norm": 0.161528, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:50.219792+00:00", "epoch": 0, "step": 10856, "train_loss": 3.5485033988952637, "perplexity": 34.76125480644117, "lr": 0.0026291804804649314, "grad_norm": 0.17479, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:50.526793+00:00", "epoch": 0, "step": 10857, "train_loss": 3.5122759342193604, "perplexity": 33.524480535880045, "lr": 0.0026291804804649314, "grad_norm": 0.143022, "tokens_per_sec": 106737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:50.831926+00:00", "epoch": 0, "step": 10858, "train_loss": 3.584761142730713, "perplexity": 36.044747140676414, "lr": 0.0026291804804649314, "grad_norm": 0.173609, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:51.137239+00:00", "epoch": 0, "step": 10859, "train_loss": 3.5372347831726074, "perplexity": 34.371742338885646, "lr": 0.0026291804804649314, "grad_norm": 0.166627, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:51.442694+00:00", "epoch": 0, "step": 10860, "train_loss": 3.496800661087036, "perplexity": 33.00967370508874, "lr": 0.0026291804804649314, "grad_norm": 0.176465, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:51.747219+00:00", "epoch": 0, "step": 10861, "train_loss": 3.5576822757720947, "perplexity": 35.08179292386161, "lr": 0.0026291804804649314, "grad_norm": 0.168155, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:52.052516+00:00", "epoch": 0, "step": 10862, "train_loss": 3.510917901992798, "perplexity": 33.47898411073156, "lr": 0.0026291804804649314, "grad_norm": 0.167489, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:52.359340+00:00", "epoch": 0, "step": 10863, "train_loss": 3.521843910217285, "perplexity": 33.84678138560725, "lr": 0.0026291804804649314, "grad_norm": 0.171161, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:52.664840+00:00", "epoch": 0, "step": 10864, "train_loss": 3.5976057052612305, "perplexity": 36.510712302213776, "lr": 0.0026291804804649314, "grad_norm": 0.174822, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:52.969981+00:00", "epoch": 0, "step": 10865, "train_loss": 3.516453504562378, "perplexity": 33.66482435551343, "lr": 0.0026291804804649314, "grad_norm": 0.16591, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:53.274803+00:00", "epoch": 0, "step": 10866, "train_loss": 3.421231985092163, "perplexity": 30.607099293096585, "lr": 0.0026291804804649314, "grad_norm": 0.186367, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:53.579861+00:00", "epoch": 0, "step": 10867, "train_loss": 3.606071949005127, "perplexity": 36.82113308666545, "lr": 0.0026291804804649314, "grad_norm": 0.194262, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:53.883677+00:00", "epoch": 0, "step": 10868, "train_loss": 3.469285488128662, "perplexity": 32.1137885621576, "lr": 0.0026291804804649314, "grad_norm": 0.200857, "tokens_per_sec": 107856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:54.189099+00:00", "epoch": 0, "step": 10869, "train_loss": 3.5225343704223633, "perplexity": 33.870159311058835, "lr": 0.0026291804804649314, "grad_norm": 0.19168, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:54.493535+00:00", "epoch": 0, "step": 10870, "train_loss": 3.526148796081543, "perplexity": 33.99280199186172, "lr": 0.0026291804804649314, "grad_norm": 0.158863, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:54.797677+00:00", "epoch": 0, "step": 10871, "train_loss": 3.5408878326416016, "perplexity": 34.4975336350109, "lr": 0.0026291804804649314, "grad_norm": 0.153575, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:55.101937+00:00", "epoch": 0, "step": 10872, "train_loss": 3.5115559101104736, "perplexity": 33.500350789698274, "lr": 0.0026291804804649314, "grad_norm": 0.151033, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:55.405944+00:00", "epoch": 0, "step": 10873, "train_loss": 3.5184664726257324, "perplexity": 33.732658823202385, "lr": 0.0026291804804649314, "grad_norm": 0.164947, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:55.711809+00:00", "epoch": 0, "step": 10874, "train_loss": 3.6142709255218506, "perplexity": 37.124269698995505, "lr": 0.0026291804804649314, "grad_norm": 0.155258, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:56.016842+00:00", "epoch": 0, "step": 10875, "train_loss": 3.661764621734619, "perplexity": 38.929978982175754, "lr": 0.0026291804804649314, "grad_norm": 0.152371, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:56.320730+00:00", "epoch": 0, "step": 10876, "train_loss": 3.512669801712036, "perplexity": 33.53768733966658, "lr": 0.0026291804804649314, "grad_norm": 0.156327, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:56.624802+00:00", "epoch": 0, "step": 10877, "train_loss": 3.537766456604004, "perplexity": 34.390021739983396, "lr": 0.0026291804804649314, "grad_norm": 0.156852, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:56.928686+00:00", "epoch": 0, "step": 10878, "train_loss": 3.5297839641571045, "perplexity": 34.11659641089856, "lr": 0.0026291804804649314, "grad_norm": 0.160589, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:57.233282+00:00", "epoch": 0, "step": 10879, "train_loss": 3.519346237182617, "perplexity": 33.762348678970966, "lr": 0.0026291804804649314, "grad_norm": 0.146593, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:57.537457+00:00", "epoch": 0, "step": 10880, "train_loss": 3.516494035720825, "perplexity": 33.666188857495726, "lr": 0.0026291804804649314, "grad_norm": 0.158753, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:57.841487+00:00", "epoch": 0, "step": 10881, "train_loss": 3.4881608486175537, "perplexity": 32.72570480007609, "lr": 0.0026291804804649314, "grad_norm": 0.162377, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:58.146993+00:00", "epoch": 0, "step": 10882, "train_loss": 3.5535404682159424, "perplexity": 34.93679138037695, "lr": 0.0026291804804649314, "grad_norm": 0.136083, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:58.451833+00:00", "epoch": 0, "step": 10883, "train_loss": 3.462289810180664, "perplexity": 31.889914826984615, "lr": 0.0026291804804649314, "grad_norm": 0.186379, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:58.756845+00:00", "epoch": 0, "step": 10884, "train_loss": 3.4272217750549316, "perplexity": 30.790979541500363, "lr": 0.0026291804804649314, "grad_norm": 0.175317, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:59.062358+00:00", "epoch": 0, "step": 10885, "train_loss": 3.5061676502227783, "perplexity": 33.32032763464302, "lr": 0.0026291804804649314, "grad_norm": 0.14567, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:59.366101+00:00", "epoch": 0, "step": 10886, "train_loss": 3.6422297954559326, "perplexity": 38.17686849633425, "lr": 0.0026291804804649314, "grad_norm": 0.158026, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:59.670172+00:00", "epoch": 0, "step": 10887, "train_loss": 3.458327054977417, "perplexity": 31.763792961158334, "lr": 0.0026291804804649314, "grad_norm": 0.162464, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:54:59.974310+00:00", "epoch": 0, "step": 10888, "train_loss": 3.5801076889038086, "perplexity": 35.87740423737071, "lr": 0.0026291804804649314, "grad_norm": 0.166638, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:00.279963+00:00", "epoch": 0, "step": 10889, "train_loss": 3.441946268081665, "perplexity": 31.24771545536208, "lr": 0.0026291804804649314, "grad_norm": 0.160623, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:00.584529+00:00", "epoch": 0, "step": 10890, "train_loss": 3.576484441757202, "perplexity": 35.74764674851942, "lr": 0.0026291804804649314, "grad_norm": 0.150215, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:00.890322+00:00", "epoch": 0, "step": 10891, "train_loss": 3.5261003971099854, "perplexity": 33.99115681501771, "lr": 0.0026291804804649314, "grad_norm": 0.191559, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:01.194148+00:00", "epoch": 0, "step": 10892, "train_loss": 3.5829286575317383, "perplexity": 35.97875615725702, "lr": 0.0026291804804649314, "grad_norm": 0.149155, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:01.497503+00:00", "epoch": 0, "step": 10893, "train_loss": 3.585564374923706, "perplexity": 36.07371107279221, "lr": 0.0026291804804649314, "grad_norm": 0.158468, "tokens_per_sec": 108018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:01.801978+00:00", "epoch": 0, "step": 10894, "train_loss": 3.55051326751709, "perplexity": 34.831190619087074, "lr": 0.0026291804804649314, "grad_norm": 0.152112, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:02.106674+00:00", "epoch": 0, "step": 10895, "train_loss": 3.5475051403045654, "perplexity": 34.72657139959184, "lr": 0.0026291804804649314, "grad_norm": 0.166209, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:02.410750+00:00", "epoch": 0, "step": 10896, "train_loss": 3.5594351291656494, "perplexity": 35.143340089460615, "lr": 0.0026291804804649314, "grad_norm": 0.156788, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:02.714995+00:00", "epoch": 0, "step": 10897, "train_loss": 3.5030946731567383, "perplexity": 33.2180921960203, "lr": 0.0026291804804649314, "grad_norm": 0.155586, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:03.019811+00:00", "epoch": 0, "step": 10898, "train_loss": 3.4888253211975098, "perplexity": 32.747457359764844, "lr": 0.0026291804804649314, "grad_norm": 0.174504, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:03.323928+00:00", "epoch": 0, "step": 10899, "train_loss": 3.4882681369781494, "perplexity": 32.72921607564955, "lr": 0.0026291804804649314, "grad_norm": 0.152527, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:03.629674+00:00", "epoch": 0, "step": 10900, "train_loss": 3.4690935611724854, "perplexity": 32.10762565190078, "lr": 0.0026291804804649314, "grad_norm": 0.168413, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:03.934531+00:00", "epoch": 0, "step": 10901, "train_loss": 3.502319097518921, "perplexity": 33.192339041029015, "lr": 0.0026291804804649314, "grad_norm": 0.154113, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:04.238683+00:00", "epoch": 0, "step": 10902, "train_loss": 3.579066753387451, "perplexity": 35.84007760374507, "lr": 0.0026291804804649314, "grad_norm": 0.162336, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:04.544741+00:00", "epoch": 0, "step": 10903, "train_loss": 3.5877859592437744, "perplexity": 36.15394094939868, "lr": 0.0026291804804649314, "grad_norm": 0.186132, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:04.849173+00:00", "epoch": 0, "step": 10904, "train_loss": 3.555135726928711, "perplexity": 34.992569079323964, "lr": 0.0026291804804649314, "grad_norm": 0.161777, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:05.154665+00:00", "epoch": 0, "step": 10905, "train_loss": 3.582129955291748, "perplexity": 35.95003131694633, "lr": 0.0026291804804649314, "grad_norm": 0.168793, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:05.459420+00:00", "epoch": 0, "step": 10906, "train_loss": 3.4886393547058105, "perplexity": 32.74136799623391, "lr": 0.0026291804804649314, "grad_norm": 0.187547, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:05.764854+00:00", "epoch": 0, "step": 10907, "train_loss": 3.534357786178589, "perplexity": 34.27299705255394, "lr": 0.0026291804804649314, "grad_norm": 0.186603, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:06.070203+00:00", "epoch": 0, "step": 10908, "train_loss": 3.500502824783325, "perplexity": 33.13210741568552, "lr": 0.0026291804804649314, "grad_norm": 0.154415, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:06.374221+00:00", "epoch": 0, "step": 10909, "train_loss": 3.6176671981811523, "perplexity": 37.250568191828904, "lr": 0.0026291804804649314, "grad_norm": 0.178645, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:06.678463+00:00", "epoch": 0, "step": 10910, "train_loss": 3.5423061847686768, "perplexity": 34.546498001369, "lr": 0.0026291804804649314, "grad_norm": 0.162294, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:06.983436+00:00", "epoch": 0, "step": 10911, "train_loss": 3.597959518432617, "perplexity": 36.523632558666684, "lr": 0.0026291804804649314, "grad_norm": 0.144585, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:07.288487+00:00", "epoch": 0, "step": 10912, "train_loss": 3.479762077331543, "perplexity": 32.45200009082421, "lr": 0.0026291804804649314, "grad_norm": 0.172717, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:07.592869+00:00", "epoch": 0, "step": 10913, "train_loss": 3.500730037689209, "perplexity": 33.13963631338846, "lr": 0.0026291804804649314, "grad_norm": 0.180365, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:07.897362+00:00", "epoch": 0, "step": 10914, "train_loss": 3.460482597351074, "perplexity": 31.832335008939104, "lr": 0.0026291804804649314, "grad_norm": 0.174255, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:08.202534+00:00", "epoch": 0, "step": 10915, "train_loss": 3.420459032058716, "perplexity": 30.583450583701218, "lr": 0.0026291804804649314, "grad_norm": 0.141962, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:08.507911+00:00", "epoch": 0, "step": 10916, "train_loss": 3.624049425125122, "perplexity": 37.48907004873909, "lr": 0.0026291804804649314, "grad_norm": 0.18648, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:08.812883+00:00", "epoch": 0, "step": 10917, "train_loss": 3.5437660217285156, "perplexity": 34.59696708533846, "lr": 0.0026291804804649314, "grad_norm": 0.189899, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:09.116972+00:00", "epoch": 0, "step": 10918, "train_loss": 3.4341506958007812, "perplexity": 31.005068645113123, "lr": 0.0026291804804649314, "grad_norm": 0.172996, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:09.420676+00:00", "epoch": 0, "step": 10919, "train_loss": 3.4932076930999756, "perplexity": 32.891283817103826, "lr": 0.0026291804804649314, "grad_norm": 0.179592, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:09.724885+00:00", "epoch": 0, "step": 10920, "train_loss": 3.59598708152771, "perplexity": 36.45166299894768, "lr": 0.0026291804804649314, "grad_norm": 0.176614, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:10.030199+00:00", "epoch": 0, "step": 10921, "train_loss": 3.5661628246307373, "perplexity": 35.38057089271601, "lr": 0.0026291804804649314, "grad_norm": 0.194952, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:10.335432+00:00", "epoch": 0, "step": 10922, "train_loss": 3.4927785396575928, "perplexity": 32.877171437833994, "lr": 0.0026291804804649314, "grad_norm": 0.19674, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:10.640161+00:00", "epoch": 0, "step": 10923, "train_loss": 3.5492658615112305, "perplexity": 34.78776907049824, "lr": 0.0026291804804649314, "grad_norm": 0.191879, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:10.944641+00:00", "epoch": 0, "step": 10924, "train_loss": 3.4519646167755127, "perplexity": 31.562339339075574, "lr": 0.0026291804804649314, "grad_norm": 0.185753, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:11.248347+00:00", "epoch": 0, "step": 10925, "train_loss": 3.534735679626465, "perplexity": 34.2859510410387, "lr": 0.0026291804804649314, "grad_norm": 0.184065, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:11.552780+00:00", "epoch": 0, "step": 10926, "train_loss": 3.4777297973632812, "perplexity": 32.386115511740904, "lr": 0.0026291804804649314, "grad_norm": 0.16922, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:11.858947+00:00", "epoch": 0, "step": 10927, "train_loss": 3.4582602977752686, "perplexity": 31.761672569987102, "lr": 0.0026291804804649314, "grad_norm": 0.167204, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:12.163899+00:00", "epoch": 0, "step": 10928, "train_loss": 3.392746925354004, "perplexity": 29.747554453051837, "lr": 0.0026291804804649314, "grad_norm": 0.163529, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:12.468279+00:00", "epoch": 0, "step": 10929, "train_loss": 3.6174697875976562, "perplexity": 37.24321526122363, "lr": 0.0026291804804649314, "grad_norm": 0.149282, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:12.771958+00:00", "epoch": 0, "step": 10930, "train_loss": 3.480638027191162, "perplexity": 32.48043886940178, "lr": 0.0026291804804649314, "grad_norm": 0.164126, "tokens_per_sec": 107967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:13.076028+00:00", "epoch": 0, "step": 10931, "train_loss": 3.5846376419067383, "perplexity": 36.040295859578684, "lr": 0.0026291804804649314, "grad_norm": 0.173816, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:13.380561+00:00", "epoch": 0, "step": 10932, "train_loss": 3.533752679824829, "perplexity": 34.25226451760197, "lr": 0.0026291804804649314, "grad_norm": 0.166458, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:13.685760+00:00", "epoch": 0, "step": 10933, "train_loss": 3.542170763015747, "perplexity": 34.54181997081277, "lr": 0.0026291804804649314, "grad_norm": 0.149752, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:13.989581+00:00", "epoch": 0, "step": 10934, "train_loss": 3.4962947368621826, "perplexity": 32.99297753536131, "lr": 0.0026291804804649314, "grad_norm": 0.167267, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:14.293550+00:00", "epoch": 0, "step": 10935, "train_loss": 3.5973873138427734, "perplexity": 36.50273954658761, "lr": 0.0026291804804649314, "grad_norm": 0.146302, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:14.597428+00:00", "epoch": 0, "step": 10936, "train_loss": 3.5576279163360596, "perplexity": 35.07988594921467, "lr": 0.0026291804804649314, "grad_norm": 0.152738, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:14.902208+00:00", "epoch": 0, "step": 10937, "train_loss": 3.513232707977295, "perplexity": 33.55657122842339, "lr": 0.0026291804804649314, "grad_norm": 0.151978, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:15.206932+00:00", "epoch": 0, "step": 10938, "train_loss": 3.404107093811035, "perplexity": 30.08741848410218, "lr": 0.0026291804804649314, "grad_norm": 0.160083, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:15.512350+00:00", "epoch": 0, "step": 10939, "train_loss": 3.5463056564331055, "perplexity": 34.68494240892955, "lr": 0.0026291804804649314, "grad_norm": 0.167759, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:15.816831+00:00", "epoch": 0, "step": 10940, "train_loss": 3.5244569778442383, "perplexity": 33.935340970022786, "lr": 0.0026291804804649314, "grad_norm": 0.160006, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:16.120780+00:00", "epoch": 0, "step": 10941, "train_loss": 3.500680685043335, "perplexity": 33.13800082501126, "lr": 0.0026291804804649314, "grad_norm": 0.16942, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:16.425197+00:00", "epoch": 0, "step": 10942, "train_loss": 3.390838861465454, "perplexity": 29.690848335279288, "lr": 0.0026291804804649314, "grad_norm": 0.190146, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:16.729771+00:00", "epoch": 0, "step": 10943, "train_loss": 3.373356342315674, "perplexity": 29.176288517326526, "lr": 0.0026291804804649314, "grad_norm": 0.185888, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:17.035673+00:00", "epoch": 0, "step": 10944, "train_loss": 3.458850622177124, "perplexity": 31.780427795637937, "lr": 0.0026291804804649314, "grad_norm": 0.179297, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:17.340059+00:00", "epoch": 0, "step": 10945, "train_loss": 3.5358965396881104, "perplexity": 34.32577534302635, "lr": 0.0026291804804649314, "grad_norm": 0.173902, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:17.644734+00:00", "epoch": 0, "step": 10946, "train_loss": 3.5019006729125977, "perplexity": 33.178453454870876, "lr": 0.0026291804804649314, "grad_norm": 0.174578, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:17.949985+00:00", "epoch": 0, "step": 10947, "train_loss": 3.5996739864349365, "perplexity": 36.586304867499464, "lr": 0.0026291804804649314, "grad_norm": 0.15629, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:18.255045+00:00", "epoch": 0, "step": 10948, "train_loss": 3.5162720680236816, "perplexity": 33.658716880382876, "lr": 0.0026291804804649314, "grad_norm": 0.163516, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:18.560599+00:00", "epoch": 0, "step": 10949, "train_loss": 3.481072425842285, "perplexity": 32.49455139324327, "lr": 0.0026291804804649314, "grad_norm": 0.171542, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:18.865429+00:00", "epoch": 0, "step": 10950, "train_loss": 3.4391496181488037, "perplexity": 31.16044861830675, "lr": 0.0026291804804649314, "grad_norm": 0.1668, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:19.170386+00:00", "epoch": 0, "step": 10951, "train_loss": 3.6375269889831543, "perplexity": 37.997751578278255, "lr": 0.0026291804804649314, "grad_norm": 0.167137, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:19.474284+00:00", "epoch": 0, "step": 10952, "train_loss": 3.5541441440582275, "perplexity": 34.95788824453317, "lr": 0.0026291804804649314, "grad_norm": 0.17466, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:19.779053+00:00", "epoch": 0, "step": 10953, "train_loss": 3.547807216644287, "perplexity": 34.73706305973268, "lr": 0.0026291804804649314, "grad_norm": 0.177447, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:20.084125+00:00", "epoch": 0, "step": 10954, "train_loss": 3.5173492431640625, "perplexity": 33.69499274768718, "lr": 0.0026291804804649314, "grad_norm": 0.156627, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:20.389710+00:00", "epoch": 0, "step": 10955, "train_loss": 3.4830710887908936, "perplexity": 32.5595619946435, "lr": 0.0026291804804649314, "grad_norm": 0.192242, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:20.694665+00:00", "epoch": 0, "step": 10956, "train_loss": 3.5071349143981934, "perplexity": 33.352572786153004, "lr": 0.0026291804804649314, "grad_norm": 0.184698, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:20.999143+00:00", "epoch": 0, "step": 10957, "train_loss": 3.5287058353424072, "perplexity": 34.07983414603802, "lr": 0.0026291804804649314, "grad_norm": 0.169609, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:21.304626+00:00", "epoch": 0, "step": 10958, "train_loss": 3.5480923652648926, "perplexity": 34.746969697713, "lr": 0.0026291804804649314, "grad_norm": 0.159835, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:21.609403+00:00", "epoch": 0, "step": 10959, "train_loss": 3.5650980472564697, "perplexity": 35.34291851060043, "lr": 0.0026291804804649314, "grad_norm": 0.166513, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:21.915256+00:00", "epoch": 0, "step": 10960, "train_loss": 3.4774374961853027, "perplexity": 32.37665039542639, "lr": 0.0026291804804649314, "grad_norm": 0.182415, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:22.219160+00:00", "epoch": 0, "step": 10961, "train_loss": 3.6122405529022217, "perplexity": 37.04897006729819, "lr": 0.0026291804804649314, "grad_norm": 0.169544, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:22.523367+00:00", "epoch": 0, "step": 10962, "train_loss": 3.6298539638519287, "perplexity": 37.70730958483274, "lr": 0.0026291804804649314, "grad_norm": 0.155732, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:22.827616+00:00", "epoch": 0, "step": 10963, "train_loss": 3.560955762863159, "perplexity": 35.19682088869101, "lr": 0.0026291804804649314, "grad_norm": 0.166153, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:23.131873+00:00", "epoch": 0, "step": 10964, "train_loss": 3.5582163333892822, "perplexity": 35.100533626459246, "lr": 0.0026291804804649314, "grad_norm": 0.146248, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:23.437273+00:00", "epoch": 0, "step": 10965, "train_loss": 3.496926784515381, "perplexity": 33.013837260860406, "lr": 0.0026291804804649314, "grad_norm": 0.156721, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:23.741425+00:00", "epoch": 0, "step": 10966, "train_loss": 3.6055994033813477, "perplexity": 36.80373753178311, "lr": 0.0026291804804649314, "grad_norm": 0.166145, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:24.045721+00:00", "epoch": 0, "step": 10967, "train_loss": 3.545619010925293, "perplexity": 34.66113432382849, "lr": 0.0026291804804649314, "grad_norm": 0.166881, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:24.350037+00:00", "epoch": 0, "step": 10968, "train_loss": 3.544691801071167, "perplexity": 34.62901107332687, "lr": 0.0026291804804649314, "grad_norm": 0.15603, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:24.655707+00:00", "epoch": 0, "step": 10969, "train_loss": 3.584106683731079, "perplexity": 36.021165049118416, "lr": 0.0026291804804649314, "grad_norm": 0.164543, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:24.960861+00:00", "epoch": 0, "step": 10970, "train_loss": 3.515167713165283, "perplexity": 33.62156623039472, "lr": 0.0026291804804649314, "grad_norm": 0.168722, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:25.264537+00:00", "epoch": 0, "step": 10971, "train_loss": 3.516768217086792, "perplexity": 33.675420764694195, "lr": 0.0026291804804649314, "grad_norm": 0.154308, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:25.568902+00:00", "epoch": 0, "step": 10972, "train_loss": 3.584007501602173, "perplexity": 36.01759257044911, "lr": 0.0026291804804649314, "grad_norm": 0.15322, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:25.872791+00:00", "epoch": 0, "step": 10973, "train_loss": 3.5449070930480957, "perplexity": 34.63646722417586, "lr": 0.0026291804804649314, "grad_norm": 0.145938, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:26.178433+00:00", "epoch": 0, "step": 10974, "train_loss": 3.5827157497406006, "perplexity": 35.971096815151434, "lr": 0.0026291804804649314, "grad_norm": 0.183948, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:26.483466+00:00", "epoch": 0, "step": 10975, "train_loss": 3.4699854850769043, "perplexity": 32.13627598579334, "lr": 0.0026291804804649314, "grad_norm": 0.173131, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:26.788231+00:00", "epoch": 0, "step": 10976, "train_loss": 3.5030572414398193, "perplexity": 33.21684880906783, "lr": 0.0026291804804649314, "grad_norm": 0.167784, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:27.092939+00:00", "epoch": 0, "step": 10977, "train_loss": 3.4434216022491455, "perplexity": 31.2938503014183, "lr": 0.0026291804804649314, "grad_norm": 0.165406, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:27.397663+00:00", "epoch": 0, "step": 10978, "train_loss": 3.5258564949035645, "perplexity": 33.98286730782737, "lr": 0.0026291804804649314, "grad_norm": 0.193759, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:27.703018+00:00", "epoch": 0, "step": 10979, "train_loss": 3.545732259750366, "perplexity": 34.665059878844424, "lr": 0.0026291804804649314, "grad_norm": 0.185694, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:28.008324+00:00", "epoch": 0, "step": 10980, "train_loss": 3.477843761444092, "perplexity": 32.389806575946594, "lr": 0.0026291804804649314, "grad_norm": 0.207061, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:28.312300+00:00", "epoch": 0, "step": 10981, "train_loss": 3.5561277866363525, "perplexity": 35.027301022406526, "lr": 0.0026291804804649314, "grad_norm": 0.203769, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:28.616542+00:00", "epoch": 0, "step": 10982, "train_loss": 3.5822641849517822, "perplexity": 35.95485720130939, "lr": 0.0026291804804649314, "grad_norm": 0.173004, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:28.920538+00:00", "epoch": 0, "step": 10983, "train_loss": 3.461226463317871, "perplexity": 31.856022808777716, "lr": 0.0026291804804649314, "grad_norm": 0.166811, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:29.226308+00:00", "epoch": 0, "step": 10984, "train_loss": 3.5031955242156982, "perplexity": 33.22144244473012, "lr": 0.0026291804804649314, "grad_norm": 0.181542, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:29.530760+00:00", "epoch": 0, "step": 10985, "train_loss": 3.5187134742736816, "perplexity": 33.74099187461777, "lr": 0.0026291804804649314, "grad_norm": 0.150418, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:29.834742+00:00", "epoch": 0, "step": 10986, "train_loss": 3.463646173477173, "perplexity": 31.933198484531122, "lr": 0.0026291804804649314, "grad_norm": 0.147338, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:30.139235+00:00", "epoch": 0, "step": 10987, "train_loss": 3.4299652576446533, "perplexity": 30.875570041017042, "lr": 0.0026291804804649314, "grad_norm": 0.159497, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:30.444153+00:00", "epoch": 0, "step": 10988, "train_loss": 3.54707932472229, "perplexity": 34.71178743221612, "lr": 0.0026291804804649314, "grad_norm": 0.154306, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:30.749730+00:00", "epoch": 0, "step": 10989, "train_loss": 3.596467971801758, "perplexity": 36.46919646465542, "lr": 0.0026291804804649314, "grad_norm": 0.127643, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:31.055532+00:00", "epoch": 0, "step": 10990, "train_loss": 3.612874984741211, "perplexity": 37.072482571262555, "lr": 0.0026291804804649314, "grad_norm": 0.138893, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:31.360570+00:00", "epoch": 0, "step": 10991, "train_loss": 3.4851577281951904, "perplexity": 32.6275729922249, "lr": 0.0026291804804649314, "grad_norm": 0.151643, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:31.664741+00:00", "epoch": 0, "step": 10992, "train_loss": 3.481931209564209, "perplexity": 32.522469170981736, "lr": 0.0026291804804649314, "grad_norm": 0.137264, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:31.969432+00:00", "epoch": 0, "step": 10993, "train_loss": 3.510566234588623, "perplexity": 33.46721271322465, "lr": 0.0026291804804649314, "grad_norm": 0.153386, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:32.273869+00:00", "epoch": 0, "step": 10994, "train_loss": 3.544875144958496, "perplexity": 34.63536067289376, "lr": 0.0026291804804649314, "grad_norm": 0.176797, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:32.579966+00:00", "epoch": 0, "step": 10995, "train_loss": 3.5004396438598633, "perplexity": 33.130014164670214, "lr": 0.0026291804804649314, "grad_norm": 0.19436, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:32.885450+00:00", "epoch": 0, "step": 10996, "train_loss": 3.6029393672943115, "perplexity": 36.705968354230535, "lr": 0.0026291804804649314, "grad_norm": 0.167349, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:33.189292+00:00", "epoch": 0, "step": 10997, "train_loss": 3.532285690307617, "perplexity": 34.202053643031924, "lr": 0.0026291804804649314, "grad_norm": 0.158557, "tokens_per_sec": 107846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:33.493857+00:00", "epoch": 0, "step": 10998, "train_loss": 3.550666332244873, "perplexity": 34.836522453845056, "lr": 0.0026291804804649314, "grad_norm": 0.186281, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:33.798122+00:00", "epoch": 0, "step": 10999, "train_loss": 3.56683611869812, "perplexity": 35.40440044244571, "lr": 0.0026291804804649314, "grad_norm": 0.154925, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:34.102779+00:00", "epoch": 0, "step": 11000, "train_loss": 3.5541343688964844, "perplexity": 34.95754652719156, "lr": 0.0026291804804649314, "grad_norm": 0.159143, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:55:37.177050+00:00", "step": 11000, "epoch": 0, "val_loss": 3.4924290060997008, "val_ppl": 32.865681771256895, "eval_train_loss": 3.5541343688964844, "eval_train_ppl": 34.95754652719156} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:55:38.120636+00:00", "step": 11000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4924_epoch_0000_step_0011000.pt", "val_loss": 3.4924290060997008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:39.478019+00:00", "epoch": 0, "step": 11001, "train_loss": 3.561110258102417, "perplexity": 35.202259050029525, "lr": 0.0026291804804649314, "grad_norm": 0.170384, "tokens_per_sec": 6096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:39.781710+00:00", "epoch": 0, "step": 11002, "train_loss": 3.6380460262298584, "perplexity": 38.01747894581477, "lr": 0.0026291804804649314, "grad_norm": 0.143821, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:40.085678+00:00", "epoch": 0, "step": 11003, "train_loss": 3.4829602241516113, "perplexity": 32.55595249063466, "lr": 0.0026291804804649314, "grad_norm": 0.153086, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:40.389734+00:00", "epoch": 0, "step": 11004, "train_loss": 3.533590316772461, "perplexity": 34.24670366683494, "lr": 0.0026291804804649314, "grad_norm": 0.164857, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:40.693922+00:00", "epoch": 0, "step": 11005, "train_loss": 3.6230883598327637, "perplexity": 37.45305791245147, "lr": 0.0026291804804649314, "grad_norm": 0.161972, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:40.997551+00:00", "epoch": 0, "step": 11006, "train_loss": 3.483053207397461, "perplexity": 32.55897978951082, "lr": 0.0026291804804649314, "grad_norm": 0.192451, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:41.300775+00:00", "epoch": 0, "step": 11007, "train_loss": 3.5559887886047363, "perplexity": 35.02243263486756, "lr": 0.0026291804804649314, "grad_norm": 0.178907, "tokens_per_sec": 108065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:41.605171+00:00", "epoch": 0, "step": 11008, "train_loss": 3.5503387451171875, "perplexity": 34.8251123265233, "lr": 0.0026291804804649314, "grad_norm": 0.180874, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:41.910638+00:00", "epoch": 0, "step": 11009, "train_loss": 3.612117052078247, "perplexity": 37.04439477149969, "lr": 0.0026291804804649314, "grad_norm": 0.142212, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:42.328267+00:00", "epoch": 0, "step": 11010, "train_loss": 3.5078413486480713, "perplexity": 33.376142510140625, "lr": 0.0026291804804649314, "grad_norm": 0.165836, "tokens_per_sec": 78491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:42.632929+00:00", "epoch": 0, "step": 11011, "train_loss": 3.3770391941070557, "perplexity": 29.283938571681187, "lr": 0.0026291804804649314, "grad_norm": 0.160796, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:42.938061+00:00", "epoch": 0, "step": 11012, "train_loss": 3.659830331802368, "perplexity": 38.854749896654816, "lr": 0.0026291804804649314, "grad_norm": 0.157217, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:43.243223+00:00", "epoch": 0, "step": 11013, "train_loss": 3.5195462703704834, "perplexity": 33.7691029447232, "lr": 0.0026291804804649314, "grad_norm": 0.164807, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:43.547202+00:00", "epoch": 0, "step": 11014, "train_loss": 3.5297446250915527, "perplexity": 34.11525432227438, "lr": 0.0026291804804649314, "grad_norm": 0.165138, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:43.850397+00:00", "epoch": 0, "step": 11015, "train_loss": 3.4436099529266357, "perplexity": 31.29974507444867, "lr": 0.0026291804804649314, "grad_norm": 0.149021, "tokens_per_sec": 108076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:44.154007+00:00", "epoch": 0, "step": 11016, "train_loss": 3.558443784713745, "perplexity": 35.10851819733813, "lr": 0.0026291804804649314, "grad_norm": 0.166498, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:44.459159+00:00", "epoch": 0, "step": 11017, "train_loss": 3.4626548290252686, "perplexity": 31.901557371594116, "lr": 0.0026291804804649314, "grad_norm": 0.189692, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:44.763566+00:00", "epoch": 0, "step": 11018, "train_loss": 3.570652484893799, "perplexity": 35.5397747541394, "lr": 0.0026291804804649314, "grad_norm": 0.173003, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:45.066886+00:00", "epoch": 0, "step": 11019, "train_loss": 3.4879238605499268, "perplexity": 32.717950117453704, "lr": 0.0026291804804649314, "grad_norm": 0.154728, "tokens_per_sec": 108029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:45.370990+00:00", "epoch": 0, "step": 11020, "train_loss": 3.5488266944885254, "perplexity": 34.77249478375601, "lr": 0.0026291804804649314, "grad_norm": 0.177757, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:45.673726+00:00", "epoch": 0, "step": 11021, "train_loss": 3.515709161758423, "perplexity": 33.6397755093787, "lr": 0.0026291804804649314, "grad_norm": 0.164076, "tokens_per_sec": 108178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:45.978295+00:00", "epoch": 0, "step": 11022, "train_loss": 3.5198333263397217, "perplexity": 33.77879795873856, "lr": 0.0026291804804649314, "grad_norm": 0.141744, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:46.282634+00:00", "epoch": 0, "step": 11023, "train_loss": 3.5560312271118164, "perplexity": 35.02391896616152, "lr": 0.0026291804804649314, "grad_norm": 0.166503, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:46.585962+00:00", "epoch": 0, "step": 11024, "train_loss": 3.6306090354919434, "perplexity": 37.73579205672165, "lr": 0.0026291804804649314, "grad_norm": 0.151858, "tokens_per_sec": 108020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:46.889883+00:00", "epoch": 0, "step": 11025, "train_loss": 3.5744693279266357, "perplexity": 35.675683702362676, "lr": 0.0026291804804649314, "grad_norm": 0.189374, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:47.193203+00:00", "epoch": 0, "step": 11026, "train_loss": 3.5304205417633057, "perplexity": 34.13832118619264, "lr": 0.0026291804804649314, "grad_norm": 0.165999, "tokens_per_sec": 108032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:47.497421+00:00", "epoch": 0, "step": 11027, "train_loss": 3.509972095489502, "perplexity": 33.44733443942771, "lr": 0.0026291804804649314, "grad_norm": 0.196934, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:47.801518+00:00", "epoch": 0, "step": 11028, "train_loss": 3.4189584255218506, "perplexity": 30.53759127483033, "lr": 0.0026291804804649314, "grad_norm": 0.174147, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:48.105590+00:00", "epoch": 0, "step": 11029, "train_loss": 3.462372303009033, "perplexity": 31.892545624764626, "lr": 0.0026291804804649314, "grad_norm": 0.163031, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:48.409642+00:00", "epoch": 0, "step": 11030, "train_loss": 3.4419381618499756, "perplexity": 31.24746215516749, "lr": 0.0026291804804649314, "grad_norm": 0.16128, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:48.712590+00:00", "epoch": 0, "step": 11031, "train_loss": 3.566941499710083, "perplexity": 35.40813159058491, "lr": 0.0026291804804649314, "grad_norm": 0.16306, "tokens_per_sec": 108098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:49.017517+00:00", "epoch": 0, "step": 11032, "train_loss": 3.5356478691101074, "perplexity": 34.31724059384589, "lr": 0.0026291804804649314, "grad_norm": 0.162668, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:49.321932+00:00", "epoch": 0, "step": 11033, "train_loss": 3.4953501224517822, "perplexity": 32.96182660846294, "lr": 0.0026291804804649314, "grad_norm": 0.173728, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:49.626239+00:00", "epoch": 0, "step": 11034, "train_loss": 3.464975118637085, "perplexity": 31.97566416512813, "lr": 0.0026291804804649314, "grad_norm": 0.145432, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:49.930656+00:00", "epoch": 0, "step": 11035, "train_loss": 3.562645673751831, "perplexity": 35.25635066539928, "lr": 0.0026291804804649314, "grad_norm": 0.165461, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:50.234152+00:00", "epoch": 0, "step": 11036, "train_loss": 3.5111982822418213, "perplexity": 33.48837227269967, "lr": 0.0026291804804649314, "grad_norm": 0.173499, "tokens_per_sec": 107963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:50.538012+00:00", "epoch": 0, "step": 11037, "train_loss": 3.5052757263183594, "perplexity": 33.29062168761407, "lr": 0.0026291804804649314, "grad_norm": 0.163265, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:50.842560+00:00", "epoch": 0, "step": 11038, "train_loss": 3.488572597503662, "perplexity": 32.739182347066695, "lr": 0.0026291804804649314, "grad_norm": 0.175336, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:51.147460+00:00", "epoch": 0, "step": 11039, "train_loss": 3.620093822479248, "perplexity": 37.34107108957586, "lr": 0.0026291804804649314, "grad_norm": 0.209993, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:51.451739+00:00", "epoch": 0, "step": 11040, "train_loss": 3.548253297805786, "perplexity": 34.75256206581974, "lr": 0.0026291804804649314, "grad_norm": 0.166779, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:51.756009+00:00", "epoch": 0, "step": 11041, "train_loss": 3.589803695678711, "perplexity": 36.22696371889556, "lr": 0.0026291804804649314, "grad_norm": 0.160826, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:52.059682+00:00", "epoch": 0, "step": 11042, "train_loss": 3.4687142372131348, "perplexity": 32.09544876985072, "lr": 0.0026291804804649314, "grad_norm": 0.165642, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:52.364280+00:00", "epoch": 0, "step": 11043, "train_loss": 3.627722978591919, "perplexity": 37.627041419426966, "lr": 0.0026291804804649314, "grad_norm": 0.203726, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:52.668820+00:00", "epoch": 0, "step": 11044, "train_loss": 3.459958553314209, "perplexity": 31.815657833766934, "lr": 0.0026291804804649314, "grad_norm": 0.15399, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:52.972878+00:00", "epoch": 0, "step": 11045, "train_loss": 3.6567859649658203, "perplexity": 38.736641658249724, "lr": 0.0026291804804649314, "grad_norm": 0.151481, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:53.277058+00:00", "epoch": 0, "step": 11046, "train_loss": 3.577054977416992, "perplexity": 35.768047874974705, "lr": 0.0026291804804649314, "grad_norm": 0.148798, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:53.581550+00:00", "epoch": 0, "step": 11047, "train_loss": 3.494584798812866, "perplexity": 32.93660979412052, "lr": 0.0026291804804649314, "grad_norm": 0.161988, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:53.886405+00:00", "epoch": 0, "step": 11048, "train_loss": 3.460759162902832, "perplexity": 31.841139953750677, "lr": 0.0026291804804649314, "grad_norm": 0.177889, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:54.190678+00:00", "epoch": 0, "step": 11049, "train_loss": 3.6053106784820557, "perplexity": 36.793112910240886, "lr": 0.0026291804804649314, "grad_norm": 0.153131, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:54.495535+00:00", "epoch": 0, "step": 11050, "train_loss": 3.5886294841766357, "perplexity": 36.18445056601279, "lr": 0.0026291804804649314, "grad_norm": 0.144224, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:54.798894+00:00", "epoch": 0, "step": 11051, "train_loss": 3.4651594161987305, "perplexity": 31.98155774513529, "lr": 0.0026291804804649314, "grad_norm": 0.166997, "tokens_per_sec": 107946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:55.102390+00:00", "epoch": 0, "step": 11052, "train_loss": 3.503890037536621, "perplexity": 33.2445231930641, "lr": 0.0026291804804649314, "grad_norm": 0.148452, "tokens_per_sec": 107970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:55.406025+00:00", "epoch": 0, "step": 11053, "train_loss": 3.556490659713745, "perplexity": 35.04001379334294, "lr": 0.0026291804804649314, "grad_norm": 0.177843, "tokens_per_sec": 107919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:55.710042+00:00", "epoch": 0, "step": 11054, "train_loss": 3.460641622543335, "perplexity": 31.837397554659493, "lr": 0.0026291804804649314, "grad_norm": 0.175743, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:56.014060+00:00", "epoch": 0, "step": 11055, "train_loss": 3.4688916206359863, "perplexity": 32.10114247538103, "lr": 0.0026291804804649314, "grad_norm": 0.167853, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:56.318284+00:00", "epoch": 0, "step": 11056, "train_loss": 3.4390742778778076, "perplexity": 31.158101070097054, "lr": 0.0026291804804649314, "grad_norm": 0.153267, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:56.621599+00:00", "epoch": 0, "step": 11057, "train_loss": 3.6103646755218506, "perplexity": 36.97953588774741, "lr": 0.0026291804804649314, "grad_norm": 0.167108, "tokens_per_sec": 108032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:56.926709+00:00", "epoch": 0, "step": 11058, "train_loss": 3.553121328353882, "perplexity": 34.92215104684496, "lr": 0.0026291804804649314, "grad_norm": 0.17222, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:57.232257+00:00", "epoch": 0, "step": 11059, "train_loss": 3.4828414916992188, "perplexity": 32.55208727202345, "lr": 0.0026291804804649314, "grad_norm": 0.182282, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:57.538109+00:00", "epoch": 0, "step": 11060, "train_loss": 3.5896918773651123, "perplexity": 36.222913107376215, "lr": 0.0026291804804649314, "grad_norm": 0.170642, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:57.841441+00:00", "epoch": 0, "step": 11061, "train_loss": 3.565047264099121, "perplexity": 35.34112373118121, "lr": 0.0026291804804649314, "grad_norm": 0.160132, "tokens_per_sec": 107969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:58.144888+00:00", "epoch": 0, "step": 11062, "train_loss": 3.5100433826446533, "perplexity": 33.44971888973663, "lr": 0.0026291804804649314, "grad_norm": 0.159521, "tokens_per_sec": 107986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:58.448582+00:00", "epoch": 0, "step": 11063, "train_loss": 3.502443552017212, "perplexity": 33.19647023399891, "lr": 0.0026291804804649314, "grad_norm": 0.166258, "tokens_per_sec": 107898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:58.753296+00:00", "epoch": 0, "step": 11064, "train_loss": 3.5649755001068115, "perplexity": 35.33858760205202, "lr": 0.0026291804804649314, "grad_norm": 0.166774, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:59.058173+00:00", "epoch": 0, "step": 11065, "train_loss": 3.475163459777832, "perplexity": 32.303108363995676, "lr": 0.0026291804804649314, "grad_norm": 0.156925, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:59.363193+00:00", "epoch": 0, "step": 11066, "train_loss": 3.545431613922119, "perplexity": 34.65463953970015, "lr": 0.0026291804804649314, "grad_norm": 0.169799, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:59.667540+00:00", "epoch": 0, "step": 11067, "train_loss": 3.56958270072937, "perplexity": 35.50177519518728, "lr": 0.0026291804804649314, "grad_norm": 0.147904, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:55:59.971889+00:00", "epoch": 0, "step": 11068, "train_loss": 3.5475409030914307, "perplexity": 34.72781334077087, "lr": 0.0026291804804649314, "grad_norm": 0.157789, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:00.275867+00:00", "epoch": 0, "step": 11069, "train_loss": 3.5435855388641357, "perplexity": 34.59072348906806, "lr": 0.0026291804804649314, "grad_norm": 0.173463, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:00.580726+00:00", "epoch": 0, "step": 11070, "train_loss": 3.5259475708007812, "perplexity": 33.98596246890257, "lr": 0.0026291804804649314, "grad_norm": 0.180226, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:00.885944+00:00", "epoch": 0, "step": 11071, "train_loss": 3.5377824306488037, "perplexity": 34.39057109211901, "lr": 0.0026291804804649314, "grad_norm": 0.156154, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:01.190015+00:00", "epoch": 0, "step": 11072, "train_loss": 3.4582793712615967, "perplexity": 31.762278381592076, "lr": 0.0026291804804649314, "grad_norm": 0.171173, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:01.493487+00:00", "epoch": 0, "step": 11073, "train_loss": 3.5194449424743652, "perplexity": 33.76568136592164, "lr": 0.0026291804804649314, "grad_norm": 0.183762, "tokens_per_sec": 107978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:01.798858+00:00", "epoch": 0, "step": 11074, "train_loss": 3.4980130195617676, "perplexity": 33.04971753158468, "lr": 0.0026291804804649314, "grad_norm": 0.164982, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:02.104705+00:00", "epoch": 0, "step": 11075, "train_loss": 3.467665910720825, "perplexity": 32.06181989072498, "lr": 0.0026291804804649314, "grad_norm": 0.152601, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:02.410007+00:00", "epoch": 0, "step": 11076, "train_loss": 3.5023481845855713, "perplexity": 33.19330452284843, "lr": 0.0026291804804649314, "grad_norm": 0.184509, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:02.713228+00:00", "epoch": 0, "step": 11077, "train_loss": 3.5489730834960938, "perplexity": 34.77758546735905, "lr": 0.0026291804804649314, "grad_norm": 0.172285, "tokens_per_sec": 108067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:03.016643+00:00", "epoch": 0, "step": 11078, "train_loss": 3.561089515686035, "perplexity": 35.20152887768752, "lr": 0.0026291804804649314, "grad_norm": 0.176088, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:03.320166+00:00", "epoch": 0, "step": 11079, "train_loss": 3.5328385829925537, "perplexity": 34.22096893688292, "lr": 0.0026291804804649314, "grad_norm": 0.188735, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:03.624637+00:00", "epoch": 0, "step": 11080, "train_loss": 3.571934223175049, "perplexity": 35.58535664973868, "lr": 0.0026291804804649314, "grad_norm": 0.167807, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:03.929461+00:00", "epoch": 0, "step": 11081, "train_loss": 3.5518698692321777, "perplexity": 34.87847473762595, "lr": 0.0026291804804649314, "grad_norm": 0.170472, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:04.233241+00:00", "epoch": 0, "step": 11082, "train_loss": 3.5161473751068115, "perplexity": 33.654520138454394, "lr": 0.0026291804804649314, "grad_norm": 0.172382, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:04.537128+00:00", "epoch": 0, "step": 11083, "train_loss": 3.505457878112793, "perplexity": 33.29668618640518, "lr": 0.0026291804804649314, "grad_norm": 0.181277, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:04.841634+00:00", "epoch": 0, "step": 11084, "train_loss": 3.5781660079956055, "perplexity": 35.80780935392049, "lr": 0.0026291804804649314, "grad_norm": 0.177041, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:05.146101+00:00", "epoch": 0, "step": 11085, "train_loss": 3.503286361694336, "perplexity": 33.22446033386515, "lr": 0.0026291804804649314, "grad_norm": 0.148329, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:05.451476+00:00", "epoch": 0, "step": 11086, "train_loss": 3.556840658187866, "perplexity": 35.05227989113634, "lr": 0.0026291804804649314, "grad_norm": 0.163984, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:05.756599+00:00", "epoch": 0, "step": 11087, "train_loss": 3.484153985977173, "perplexity": 32.59483975035777, "lr": 0.0026291804804649314, "grad_norm": 0.187804, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:06.061247+00:00", "epoch": 0, "step": 11088, "train_loss": 3.525568962097168, "perplexity": 33.97309752325505, "lr": 0.0026291804804649314, "grad_norm": 0.175741, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:06.364679+00:00", "epoch": 0, "step": 11089, "train_loss": 3.5372748374938965, "perplexity": 34.37311910326905, "lr": 0.0026291804804649314, "grad_norm": 0.155734, "tokens_per_sec": 107991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:06.669340+00:00", "epoch": 0, "step": 11090, "train_loss": 3.5478551387786865, "perplexity": 34.738727773825275, "lr": 0.0026291804804649314, "grad_norm": 0.157129, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:06.974463+00:00", "epoch": 0, "step": 11091, "train_loss": 3.5151965618133545, "perplexity": 33.62253618111732, "lr": 0.0026291804804649314, "grad_norm": 0.149768, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:07.277882+00:00", "epoch": 0, "step": 11092, "train_loss": 3.5417604446411133, "perplexity": 34.52764973473838, "lr": 0.0026291804804649314, "grad_norm": 0.175771, "tokens_per_sec": 107995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:07.582518+00:00", "epoch": 0, "step": 11093, "train_loss": 3.5727479457855225, "perplexity": 35.614325043566424, "lr": 0.0026291804804649314, "grad_norm": 0.157516, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:07.886323+00:00", "epoch": 0, "step": 11094, "train_loss": 3.502021312713623, "perplexity": 33.182456338342675, "lr": 0.0026291804804649314, "grad_norm": 0.157574, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:08.191229+00:00", "epoch": 0, "step": 11095, "train_loss": 3.497542142868042, "perplexity": 33.03415885326183, "lr": 0.0026291804804649314, "grad_norm": 0.143125, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:08.496196+00:00", "epoch": 0, "step": 11096, "train_loss": 3.5638527870178223, "perplexity": 35.298934770746314, "lr": 0.0026291804804649314, "grad_norm": 0.139222, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:08.801068+00:00", "epoch": 0, "step": 11097, "train_loss": 3.4879887104034424, "perplexity": 32.720071940525365, "lr": 0.0026291804804649314, "grad_norm": 0.150303, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:09.106509+00:00", "epoch": 0, "step": 11098, "train_loss": 3.56331467628479, "perplexity": 35.279945144800564, "lr": 0.0026291804804649314, "grad_norm": 0.150225, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:09.410251+00:00", "epoch": 0, "step": 11099, "train_loss": 3.531826972961426, "perplexity": 34.18636816562583, "lr": 0.0026291804804649314, "grad_norm": 0.159532, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:09.714802+00:00", "epoch": 0, "step": 11100, "train_loss": 3.5316669940948486, "perplexity": 34.1808995066409, "lr": 0.0026291804804649314, "grad_norm": 0.168787, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:10.019631+00:00", "epoch": 0, "step": 11101, "train_loss": 3.460064172744751, "perplexity": 31.819018362895118, "lr": 0.0026291804804649314, "grad_norm": 0.190928, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:10.325220+00:00", "epoch": 0, "step": 11102, "train_loss": 3.444615602493286, "perplexity": 31.33123748202797, "lr": 0.0026291804804649314, "grad_norm": 0.169637, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:10.629082+00:00", "epoch": 0, "step": 11103, "train_loss": 3.545773983001709, "perplexity": 34.666506248023964, "lr": 0.0026291804804649314, "grad_norm": 0.156622, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:10.933689+00:00", "epoch": 0, "step": 11104, "train_loss": 3.450181484222412, "perplexity": 31.50610965173984, "lr": 0.0026291804804649314, "grad_norm": 0.148258, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:11.238075+00:00", "epoch": 0, "step": 11105, "train_loss": 3.538881778717041, "perplexity": 34.4283990892727, "lr": 0.0026291804804649314, "grad_norm": 0.183611, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:11.543497+00:00", "epoch": 0, "step": 11106, "train_loss": 3.532426595687866, "perplexity": 34.20687323595111, "lr": 0.0026291804804649314, "grad_norm": 0.185853, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:11.849376+00:00", "epoch": 0, "step": 11107, "train_loss": 3.6249704360961914, "perplexity": 37.52361379869367, "lr": 0.0026291804804649314, "grad_norm": 0.171001, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:12.153533+00:00", "epoch": 0, "step": 11108, "train_loss": 3.4812870025634766, "perplexity": 32.501524715667394, "lr": 0.0026291804804649314, "grad_norm": 0.18145, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:12.457354+00:00", "epoch": 0, "step": 11109, "train_loss": 3.542461395263672, "perplexity": 34.55186039656338, "lr": 0.0026291804804649314, "grad_norm": 0.176966, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:12.763314+00:00", "epoch": 0, "step": 11110, "train_loss": 3.5819246768951416, "perplexity": 35.94265230956095, "lr": 0.0026291804804649314, "grad_norm": 0.157565, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:13.067356+00:00", "epoch": 0, "step": 11111, "train_loss": 3.697382926940918, "perplexity": 40.34158920199778, "lr": 0.0026291804804649314, "grad_norm": 0.178772, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:13.372555+00:00", "epoch": 0, "step": 11112, "train_loss": 3.539238452911377, "perplexity": 34.44068100097042, "lr": 0.0026291804804649314, "grad_norm": 0.16267, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:13.676689+00:00", "epoch": 0, "step": 11113, "train_loss": 3.535444736480713, "perplexity": 34.31027035049639, "lr": 0.0026291804804649314, "grad_norm": 0.17255, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:13.981658+00:00", "epoch": 0, "step": 11114, "train_loss": 3.48022723197937, "perplexity": 32.46709880084307, "lr": 0.0026291804804649314, "grad_norm": 0.145, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:14.286339+00:00", "epoch": 0, "step": 11115, "train_loss": 3.54229998588562, "perplexity": 34.54628385233161, "lr": 0.0026291804804649314, "grad_norm": 0.168373, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:14.593485+00:00", "epoch": 0, "step": 11116, "train_loss": 3.470273017883301, "perplexity": 32.14551754797715, "lr": 0.0026291804804649314, "grad_norm": 0.138829, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:14.898552+00:00", "epoch": 0, "step": 11117, "train_loss": 3.589064598083496, "perplexity": 36.2001983494575, "lr": 0.0026291804804649314, "grad_norm": 0.155139, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:15.203353+00:00", "epoch": 0, "step": 11118, "train_loss": 3.542335271835327, "perplexity": 34.547502872273206, "lr": 0.0026291804804649314, "grad_norm": 0.178286, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:15.507466+00:00", "epoch": 0, "step": 11119, "train_loss": 3.5147011280059814, "perplexity": 33.605882565726255, "lr": 0.0026291804804649314, "grad_norm": 0.178939, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:15.811808+00:00", "epoch": 0, "step": 11120, "train_loss": 3.647559404373169, "perplexity": 38.380879441474576, "lr": 0.0026291804804649314, "grad_norm": 0.148493, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:16.116605+00:00", "epoch": 0, "step": 11121, "train_loss": 3.6280856132507324, "perplexity": 37.64068876310461, "lr": 0.0026291804804649314, "grad_norm": 0.162288, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:16.423042+00:00", "epoch": 0, "step": 11122, "train_loss": 3.5007619857788086, "perplexity": 33.140695078371365, "lr": 0.0026291804804649314, "grad_norm": 0.158262, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:16.727684+00:00", "epoch": 0, "step": 11123, "train_loss": 3.4298980236053467, "perplexity": 30.873494221510935, "lr": 0.0026291804804649314, "grad_norm": 0.167907, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:17.031441+00:00", "epoch": 0, "step": 11124, "train_loss": 3.4743611812591553, "perplexity": 32.27720266725028, "lr": 0.0026291804804649314, "grad_norm": 0.154832, "tokens_per_sec": 107875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:17.335880+00:00", "epoch": 0, "step": 11125, "train_loss": 3.535038948059082, "perplexity": 34.29635046450043, "lr": 0.0026291804804649314, "grad_norm": 0.165323, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:17.639536+00:00", "epoch": 0, "step": 11126, "train_loss": 3.5484447479248047, "perplexity": 34.759216084899485, "lr": 0.0026291804804649314, "grad_norm": 0.18327, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:17.944442+00:00", "epoch": 0, "step": 11127, "train_loss": 3.5220675468444824, "perplexity": 33.854351612096224, "lr": 0.0026291804804649314, "grad_norm": 0.183149, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:18.250420+00:00", "epoch": 0, "step": 11128, "train_loss": 3.4736039638519287, "perplexity": 32.25277105872107, "lr": 0.0026291804804649314, "grad_norm": 0.152326, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:18.554638+00:00", "epoch": 0, "step": 11129, "train_loss": 3.474494218826294, "perplexity": 32.281497033416855, "lr": 0.0026291804804649314, "grad_norm": 0.17895, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:18.858555+00:00", "epoch": 0, "step": 11130, "train_loss": 3.510923385620117, "perplexity": 33.47916769750682, "lr": 0.0026291804804649314, "grad_norm": 0.186353, "tokens_per_sec": 107874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:19.162579+00:00", "epoch": 0, "step": 11131, "train_loss": 3.5905003547668457, "perplexity": 36.2522103555342, "lr": 0.0026291804804649314, "grad_norm": 0.160731, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:19.467348+00:00", "epoch": 0, "step": 11132, "train_loss": 3.5569636821746826, "perplexity": 35.056592427622824, "lr": 0.0026291804804649314, "grad_norm": 0.173077, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:19.772456+00:00", "epoch": 0, "step": 11133, "train_loss": 3.521449327468872, "perplexity": 33.8334286641332, "lr": 0.0026291804804649314, "grad_norm": 0.160628, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:20.078062+00:00", "epoch": 0, "step": 11134, "train_loss": 3.5812020301818848, "perplexity": 35.916687852698345, "lr": 0.0026291804804649314, "grad_norm": 0.155248, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:20.382089+00:00", "epoch": 0, "step": 11135, "train_loss": 3.565490245819092, "perplexity": 35.356782671013335, "lr": 0.0026291804804649314, "grad_norm": 0.197378, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:20.686651+00:00", "epoch": 0, "step": 11136, "train_loss": 3.5357577800750732, "perplexity": 34.321012642165456, "lr": 0.0026291804804649314, "grad_norm": 0.185256, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:20.991366+00:00", "epoch": 0, "step": 11137, "train_loss": 3.4545915126800537, "perplexity": 31.645359313698282, "lr": 0.0026291804804649314, "grad_norm": 0.158856, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:21.297028+00:00", "epoch": 0, "step": 11138, "train_loss": 3.54207444190979, "perplexity": 34.53849302474155, "lr": 0.0026291804804649314, "grad_norm": 0.173519, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:21.602409+00:00", "epoch": 0, "step": 11139, "train_loss": 3.495795726776123, "perplexity": 32.97651781392709, "lr": 0.0026291804804649314, "grad_norm": 0.158414, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:21.906668+00:00", "epoch": 0, "step": 11140, "train_loss": 3.6072094440460205, "perplexity": 36.86304077332581, "lr": 0.0026291804804649314, "grad_norm": 0.14861, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:22.211301+00:00", "epoch": 0, "step": 11141, "train_loss": 3.3919031620025635, "perplexity": 29.72246514301803, "lr": 0.0026291804804649314, "grad_norm": 0.186129, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:22.515531+00:00", "epoch": 0, "step": 11142, "train_loss": 3.5148000717163086, "perplexity": 33.60920782093996, "lr": 0.0026291804804649314, "grad_norm": 0.192718, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:22.820736+00:00", "epoch": 0, "step": 11143, "train_loss": 3.521190643310547, "perplexity": 33.82467762404248, "lr": 0.0026291804804649314, "grad_norm": 0.147566, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:23.125870+00:00", "epoch": 0, "step": 11144, "train_loss": 3.57138729095459, "perplexity": 35.56589919305044, "lr": 0.0026291804804649314, "grad_norm": 0.148236, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:23.430957+00:00", "epoch": 0, "step": 11145, "train_loss": 3.500321388244629, "perplexity": 33.12609658610484, "lr": 0.0026291804804649314, "grad_norm": 0.143114, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:23.734658+00:00", "epoch": 0, "step": 11146, "train_loss": 3.636927843093872, "perplexity": 37.97499220039285, "lr": 0.0026291804804649314, "grad_norm": 0.164281, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:24.038924+00:00", "epoch": 0, "step": 11147, "train_loss": 3.506281614303589, "perplexity": 33.3241251715415, "lr": 0.0026291804804649314, "grad_norm": 0.162691, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:24.343681+00:00", "epoch": 0, "step": 11148, "train_loss": 3.4730536937713623, "perplexity": 32.23502820592032, "lr": 0.0026291804804649314, "grad_norm": 0.140775, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:24.649182+00:00", "epoch": 0, "step": 11149, "train_loss": 3.4649384021759033, "perplexity": 31.97449015344896, "lr": 0.0026291804804649314, "grad_norm": 0.165878, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:24.954100+00:00", "epoch": 0, "step": 11150, "train_loss": 3.573289394378662, "perplexity": 35.633613591164135, "lr": 0.0026291804804649314, "grad_norm": 0.186641, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:25.257795+00:00", "epoch": 0, "step": 11151, "train_loss": 3.5666420459747314, "perplexity": 35.39753008072837, "lr": 0.0026291804804649314, "grad_norm": 0.179829, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:25.562863+00:00", "epoch": 0, "step": 11152, "train_loss": 3.6793456077575684, "perplexity": 39.62045826687168, "lr": 0.0026291804804649314, "grad_norm": 0.167899, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:25.867586+00:00", "epoch": 0, "step": 11153, "train_loss": 3.614175319671631, "perplexity": 37.12072057128854, "lr": 0.0026291804804649314, "grad_norm": 0.168176, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:26.172261+00:00", "epoch": 0, "step": 11154, "train_loss": 3.4830000400543213, "perplexity": 32.557248761077574, "lr": 0.0026291804804649314, "grad_norm": 0.176111, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:26.477234+00:00", "epoch": 0, "step": 11155, "train_loss": 3.5919764041900635, "perplexity": 36.30575992089993, "lr": 0.0026291804804649314, "grad_norm": 0.175316, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:26.781649+00:00", "epoch": 0, "step": 11156, "train_loss": 3.6109588146209717, "perplexity": 37.00151340409005, "lr": 0.0026291804804649314, "grad_norm": 0.181798, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:27.086570+00:00", "epoch": 0, "step": 11157, "train_loss": 3.4911842346191406, "perplexity": 32.82479695930193, "lr": 0.0026291804804649314, "grad_norm": 0.186083, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:27.391364+00:00", "epoch": 0, "step": 11158, "train_loss": 3.5331759452819824, "perplexity": 34.23251574892989, "lr": 0.0026291804804649314, "grad_norm": 0.144109, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:27.697090+00:00", "epoch": 0, "step": 11159, "train_loss": 3.5786142349243164, "perplexity": 35.823862975895125, "lr": 0.0026291804804649314, "grad_norm": 0.193737, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:28.004243+00:00", "epoch": 0, "step": 11160, "train_loss": 3.571645975112915, "perplexity": 35.57510071784131, "lr": 0.0026291804804649314, "grad_norm": 0.175765, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:28.309212+00:00", "epoch": 0, "step": 11161, "train_loss": 3.450376510620117, "perplexity": 31.512254774021976, "lr": 0.0026291804804649314, "grad_norm": 0.181234, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:28.614372+00:00", "epoch": 0, "step": 11162, "train_loss": 3.478318691253662, "perplexity": 32.40519311408933, "lr": 0.0026291804804649314, "grad_norm": 0.175974, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:28.919466+00:00", "epoch": 0, "step": 11163, "train_loss": 3.5525124073028564, "perplexity": 34.90089268691374, "lr": 0.0026291804804649314, "grad_norm": 0.150221, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:29.223704+00:00", "epoch": 0, "step": 11164, "train_loss": 3.6540541648864746, "perplexity": 38.63096530654663, "lr": 0.0026291804804649314, "grad_norm": 0.180622, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:29.528253+00:00", "epoch": 0, "step": 11165, "train_loss": 3.515308380126953, "perplexity": 33.626296006616755, "lr": 0.0026291804804649314, "grad_norm": 0.184054, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:29.832347+00:00", "epoch": 0, "step": 11166, "train_loss": 3.437105178833008, "perplexity": 31.09680804886148, "lr": 0.0026291804804649314, "grad_norm": 0.146007, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:30.137087+00:00", "epoch": 0, "step": 11167, "train_loss": 3.479243516921997, "perplexity": 32.43517613086331, "lr": 0.0026291804804649314, "grad_norm": 0.189071, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:30.441178+00:00", "epoch": 0, "step": 11168, "train_loss": 3.5050528049468994, "perplexity": 33.283201323680295, "lr": 0.0026291804804649314, "grad_norm": 0.162815, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:30.746680+00:00", "epoch": 0, "step": 11169, "train_loss": 3.552691698074341, "perplexity": 34.90715065587035, "lr": 0.0026291804804649314, "grad_norm": 0.162722, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:31.052225+00:00", "epoch": 0, "step": 11170, "train_loss": 3.4686505794525146, "perplexity": 32.093405710484916, "lr": 0.0026291804804649314, "grad_norm": 0.192232, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:31.356824+00:00", "epoch": 0, "step": 11171, "train_loss": 3.5339198112487793, "perplexity": 34.25798962575373, "lr": 0.0026291804804649314, "grad_norm": 0.175042, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:31.661382+00:00", "epoch": 0, "step": 11172, "train_loss": 3.4522368907928467, "perplexity": 31.570934114017735, "lr": 0.0026291804804649314, "grad_norm": 0.175104, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:31.966200+00:00", "epoch": 0, "step": 11173, "train_loss": 3.541041374206543, "perplexity": 34.50283084697254, "lr": 0.0026291804804649314, "grad_norm": 0.165847, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:32.270222+00:00", "epoch": 0, "step": 11174, "train_loss": 3.493032455444336, "perplexity": 32.88552053062389, "lr": 0.0026291804804649314, "grad_norm": 0.172903, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:32.575506+00:00", "epoch": 0, "step": 11175, "train_loss": 3.619077444076538, "perplexity": 37.30313771198473, "lr": 0.0026291804804649314, "grad_norm": 0.141427, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:32.879294+00:00", "epoch": 0, "step": 11176, "train_loss": 3.588809013366699, "perplexity": 36.19094731427633, "lr": 0.0026291804804649314, "grad_norm": 0.167239, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:33.184599+00:00", "epoch": 0, "step": 11177, "train_loss": 3.4907805919647217, "perplexity": 32.81155014479605, "lr": 0.0026291804804649314, "grad_norm": 0.176292, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:33.489323+00:00", "epoch": 0, "step": 11178, "train_loss": 3.578368902206421, "perplexity": 35.81507528822346, "lr": 0.0026291804804649314, "grad_norm": 0.166684, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:33.793913+00:00", "epoch": 0, "step": 11179, "train_loss": 3.4633007049560547, "perplexity": 31.922168475046334, "lr": 0.0026291804804649314, "grad_norm": 0.161465, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:34.098453+00:00", "epoch": 0, "step": 11180, "train_loss": 3.520373821258545, "perplexity": 33.797060162268735, "lr": 0.0026291804804649314, "grad_norm": 0.16655, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:34.404013+00:00", "epoch": 0, "step": 11181, "train_loss": 3.534271478652954, "perplexity": 34.27003916262816, "lr": 0.0026291804804649314, "grad_norm": 0.152354, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:34.708843+00:00", "epoch": 0, "step": 11182, "train_loss": 3.5783114433288574, "perplexity": 35.81301745331856, "lr": 0.0026291804804649314, "grad_norm": 0.179884, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:35.014491+00:00", "epoch": 0, "step": 11183, "train_loss": 3.4880728721618652, "perplexity": 32.72282583520029, "lr": 0.0026291804804649314, "grad_norm": 0.137137, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:35.319334+00:00", "epoch": 0, "step": 11184, "train_loss": 3.6211256980895996, "perplexity": 37.37962231671244, "lr": 0.0026291804804649314, "grad_norm": 0.16784, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:35.625608+00:00", "epoch": 0, "step": 11185, "train_loss": 3.6017870903015137, "perplexity": 36.66369727007355, "lr": 0.0026291804804649314, "grad_norm": 0.184624, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:35.931421+00:00", "epoch": 0, "step": 11186, "train_loss": 3.496405601501465, "perplexity": 32.99663549267993, "lr": 0.0026291804804649314, "grad_norm": 0.154737, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:36.235872+00:00", "epoch": 0, "step": 11187, "train_loss": 3.4019999504089355, "perplexity": 30.02408672673756, "lr": 0.0026291804804649314, "grad_norm": 0.146766, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:36.541434+00:00", "epoch": 0, "step": 11188, "train_loss": 3.4233481884002686, "perplexity": 30.67193872037978, "lr": 0.0026291804804649314, "grad_norm": 0.158571, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:36.847851+00:00", "epoch": 0, "step": 11189, "train_loss": 3.539896249771118, "perplexity": 34.46334342559892, "lr": 0.0026291804804649314, "grad_norm": 0.167362, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:37.153176+00:00", "epoch": 0, "step": 11190, "train_loss": 3.5042381286621094, "perplexity": 33.25609733086915, "lr": 0.0026291804804649314, "grad_norm": 0.161817, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:37.459062+00:00", "epoch": 0, "step": 11191, "train_loss": 3.5255930423736572, "perplexity": 33.9739156146865, "lr": 0.0026291804804649314, "grad_norm": 0.152757, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:37.764704+00:00", "epoch": 0, "step": 11192, "train_loss": 3.494736671447754, "perplexity": 32.9416123436998, "lr": 0.0026291804804649314, "grad_norm": 0.158951, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:38.070382+00:00", "epoch": 0, "step": 11193, "train_loss": 3.5016095638275146, "perplexity": 33.16879631135046, "lr": 0.0026291804804649314, "grad_norm": 0.156871, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:38.375891+00:00", "epoch": 0, "step": 11194, "train_loss": 3.4928743839263916, "perplexity": 32.88032267730233, "lr": 0.0026291804804649314, "grad_norm": 0.156824, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:38.681848+00:00", "epoch": 0, "step": 11195, "train_loss": 3.4601190090179443, "perplexity": 31.820763247119846, "lr": 0.0026291804804649314, "grad_norm": 0.155773, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:38.988135+00:00", "epoch": 0, "step": 11196, "train_loss": 3.446445941925049, "perplexity": 31.38863679552602, "lr": 0.0026291804804649314, "grad_norm": 0.153294, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:39.293416+00:00", "epoch": 0, "step": 11197, "train_loss": 3.4484336376190186, "perplexity": 31.451089902047176, "lr": 0.0026291804804649314, "grad_norm": 0.158372, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:39.599857+00:00", "epoch": 0, "step": 11198, "train_loss": 3.5386765003204346, "perplexity": 34.42133240805324, "lr": 0.0026291804804649314, "grad_norm": 0.152753, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:39.905741+00:00", "epoch": 0, "step": 11199, "train_loss": 3.6242754459381104, "perplexity": 37.49754431647404, "lr": 0.0026291804804649314, "grad_norm": 0.16113, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:40.210869+00:00", "epoch": 0, "step": 11200, "train_loss": 3.5008938312530518, "perplexity": 33.14506481708953, "lr": 0.0026291804804649314, "grad_norm": 0.158604, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:40.515706+00:00", "epoch": 0, "step": 11201, "train_loss": 3.6323559284210205, "perplexity": 37.80177005651092, "lr": 0.0026291804804649314, "grad_norm": 0.175092, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:40.820626+00:00", "epoch": 0, "step": 11202, "train_loss": 3.5060126781463623, "perplexity": 33.31516431437819, "lr": 0.0026291804804649314, "grad_norm": 0.17855, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:41.124623+00:00", "epoch": 0, "step": 11203, "train_loss": 3.5134658813476562, "perplexity": 33.56439663953777, "lr": 0.0026291804804649314, "grad_norm": 0.172076, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:41.429809+00:00", "epoch": 0, "step": 11204, "train_loss": 3.4244604110717773, "perplexity": 30.70607172423543, "lr": 0.0026291804804649314, "grad_norm": 0.166177, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:41.735343+00:00", "epoch": 0, "step": 11205, "train_loss": 3.493784189224243, "perplexity": 32.91025098146964, "lr": 0.0026291804804649314, "grad_norm": 0.187123, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:42.039547+00:00", "epoch": 0, "step": 11206, "train_loss": 3.522886276245117, "perplexity": 33.88208051478548, "lr": 0.0026291804804649314, "grad_norm": 0.159788, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:42.344148+00:00", "epoch": 0, "step": 11207, "train_loss": 3.350407600402832, "perplexity": 28.51435373749886, "lr": 0.0026291804804649314, "grad_norm": 0.168451, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:42.648993+00:00", "epoch": 0, "step": 11208, "train_loss": 3.6207022666931152, "perplexity": 37.36379796153566, "lr": 0.0026291804804649314, "grad_norm": 0.159129, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:42.955019+00:00", "epoch": 0, "step": 11209, "train_loss": 3.5585458278656006, "perplexity": 35.11210096398725, "lr": 0.0026291804804649314, "grad_norm": 0.17086, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:43.259969+00:00", "epoch": 0, "step": 11210, "train_loss": 3.618103265762329, "perplexity": 37.26681549920233, "lr": 0.0026291804804649314, "grad_norm": 0.15703, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:43.563718+00:00", "epoch": 0, "step": 11211, "train_loss": 3.5573220252990723, "perplexity": 35.06915696755759, "lr": 0.0026291804804649314, "grad_norm": 0.195732, "tokens_per_sec": 107821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:43.868049+00:00", "epoch": 0, "step": 11212, "train_loss": 3.567976474761963, "perplexity": 35.44479709408261, "lr": 0.0026291804804649314, "grad_norm": 0.183004, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:44.172509+00:00", "epoch": 0, "step": 11213, "train_loss": 3.5939183235168457, "perplexity": 36.37633127752717, "lr": 0.0026291804804649314, "grad_norm": 0.167815, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:44.477159+00:00", "epoch": 0, "step": 11214, "train_loss": 3.6033856868743896, "perplexity": 36.7223546030926, "lr": 0.0026291804804649314, "grad_norm": 0.156557, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:44.781772+00:00", "epoch": 0, "step": 11215, "train_loss": 3.4869611263275146, "perplexity": 32.68646658475737, "lr": 0.0026291804804649314, "grad_norm": 0.1903, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:45.086342+00:00", "epoch": 0, "step": 11216, "train_loss": 3.4373300075531006, "perplexity": 31.10380029041298, "lr": 0.0026291804804649314, "grad_norm": 0.16308, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:45.390171+00:00", "epoch": 0, "step": 11217, "train_loss": 3.495209217071533, "perplexity": 32.95718243695303, "lr": 0.0026291804804649314, "grad_norm": 0.153447, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:45.694658+00:00", "epoch": 0, "step": 11218, "train_loss": 3.444520950317383, "perplexity": 31.32827205257094, "lr": 0.0026291804804649314, "grad_norm": 0.17285, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:45.999324+00:00", "epoch": 0, "step": 11219, "train_loss": 3.539074659347534, "perplexity": 34.435040301055885, "lr": 0.0026291804804649314, "grad_norm": 0.186603, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:46.304128+00:00", "epoch": 0, "step": 11220, "train_loss": 3.543837308883667, "perplexity": 34.59943349260938, "lr": 0.0026291804804649314, "grad_norm": 0.166061, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:46.608694+00:00", "epoch": 0, "step": 11221, "train_loss": 3.3704893589019775, "perplexity": 29.09276037618569, "lr": 0.0026291804804649314, "grad_norm": 0.189089, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:46.912865+00:00", "epoch": 0, "step": 11222, "train_loss": 3.507986307144165, "perplexity": 33.38098101624712, "lr": 0.0026291804804649314, "grad_norm": 0.1786, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:47.217509+00:00", "epoch": 0, "step": 11223, "train_loss": 3.515319347381592, "perplexity": 33.626664796789925, "lr": 0.0026291804804649314, "grad_norm": 0.141917, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:47.523474+00:00", "epoch": 0, "step": 11224, "train_loss": 3.599165916442871, "perplexity": 36.567721185181156, "lr": 0.0026291804804649314, "grad_norm": 0.160097, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:47.828586+00:00", "epoch": 0, "step": 11225, "train_loss": 3.576969861984253, "perplexity": 35.76500359166117, "lr": 0.0026291804804649314, "grad_norm": 0.166164, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:48.133282+00:00", "epoch": 0, "step": 11226, "train_loss": 3.6544315814971924, "perplexity": 38.645548026248605, "lr": 0.0026291804804649314, "grad_norm": 0.168739, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:48.438707+00:00", "epoch": 0, "step": 11227, "train_loss": 3.494016647338867, "perplexity": 32.91790212561541, "lr": 0.0026291804804649314, "grad_norm": 0.161731, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:48.742704+00:00", "epoch": 0, "step": 11228, "train_loss": 3.489891529083252, "perplexity": 32.78239157729951, "lr": 0.0026291804804649314, "grad_norm": 0.17034, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:49.048414+00:00", "epoch": 0, "step": 11229, "train_loss": 3.616128444671631, "perplexity": 37.19329282692764, "lr": 0.0026291804804649314, "grad_norm": 0.173056, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:49.354437+00:00", "epoch": 0, "step": 11230, "train_loss": 3.428807020187378, "perplexity": 30.839829501292268, "lr": 0.0026291804804649314, "grad_norm": 0.165742, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:49.659440+00:00", "epoch": 0, "step": 11231, "train_loss": 3.5835843086242676, "perplexity": 36.00235340297401, "lr": 0.0026291804804649314, "grad_norm": 0.173577, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:49.964449+00:00", "epoch": 0, "step": 11232, "train_loss": 3.5464060306549072, "perplexity": 34.68842405776306, "lr": 0.0026291804804649314, "grad_norm": 0.160373, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:50.269184+00:00", "epoch": 0, "step": 11233, "train_loss": 3.395336151123047, "perplexity": 29.8246773888675, "lr": 0.0026291804804649314, "grad_norm": 0.146095, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:50.573965+00:00", "epoch": 0, "step": 11234, "train_loss": 3.532904624938965, "perplexity": 34.223229030909266, "lr": 0.0026291804804649314, "grad_norm": 0.149214, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:50.878683+00:00", "epoch": 0, "step": 11235, "train_loss": 3.5528512001037598, "perplexity": 34.912718861299425, "lr": 0.0026291804804649314, "grad_norm": 0.147546, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:51.184734+00:00", "epoch": 0, "step": 11236, "train_loss": 3.5941433906555176, "perplexity": 36.38451931571778, "lr": 0.0026291804804649314, "grad_norm": 0.157881, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:51.490217+00:00", "epoch": 0, "step": 11237, "train_loss": 3.603219509124756, "perplexity": 36.71625267186005, "lr": 0.0026291804804649314, "grad_norm": 0.136505, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:51.795041+00:00", "epoch": 0, "step": 11238, "train_loss": 3.5991008281707764, "perplexity": 36.56534113285235, "lr": 0.0026291804804649314, "grad_norm": 0.172664, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:52.099509+00:00", "epoch": 0, "step": 11239, "train_loss": 3.558544158935547, "perplexity": 35.112042364395606, "lr": 0.0026291804804649314, "grad_norm": 0.198035, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:52.404396+00:00", "epoch": 0, "step": 11240, "train_loss": 3.527439594268799, "perplexity": 34.03670816996389, "lr": 0.0026291804804649314, "grad_norm": 0.17603, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:52.709425+00:00", "epoch": 0, "step": 11241, "train_loss": 3.5202107429504395, "perplexity": 33.791549044262645, "lr": 0.0026291804804649314, "grad_norm": 0.200272, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:53.014561+00:00", "epoch": 0, "step": 11242, "train_loss": 3.540032386779785, "perplexity": 34.46803548145552, "lr": 0.0026291804804649314, "grad_norm": 0.194139, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:53.319848+00:00", "epoch": 0, "step": 11243, "train_loss": 3.632767915725708, "perplexity": 37.81734711442364, "lr": 0.0026291804804649314, "grad_norm": 0.160152, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:53.624578+00:00", "epoch": 0, "step": 11244, "train_loss": 3.526747703552246, "perplexity": 34.0131666325838, "lr": 0.0026291804804649314, "grad_norm": 0.149989, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:53.930186+00:00", "epoch": 0, "step": 11245, "train_loss": 3.5460641384124756, "perplexity": 34.676566381814744, "lr": 0.0026291804804649314, "grad_norm": 0.163029, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:54.236300+00:00", "epoch": 0, "step": 11246, "train_loss": 3.496150255203247, "perplexity": 32.98821099958057, "lr": 0.0026291804804649314, "grad_norm": 0.191566, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:54.541854+00:00", "epoch": 0, "step": 11247, "train_loss": 3.4126110076904297, "perplexity": 30.344370299310906, "lr": 0.0026291804804649314, "grad_norm": 0.192826, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:54.847532+00:00", "epoch": 0, "step": 11248, "train_loss": 3.436985492706299, "perplexity": 31.093086415071507, "lr": 0.0026291804804649314, "grad_norm": 0.143324, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:55.152662+00:00", "epoch": 0, "step": 11249, "train_loss": 3.5874648094177246, "perplexity": 36.14233198176051, "lr": 0.0026291804804649314, "grad_norm": 0.195822, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:55.457682+00:00", "epoch": 0, "step": 11250, "train_loss": 3.557647466659546, "perplexity": 35.080571779036916, "lr": 0.0026291804804649314, "grad_norm": 0.177428, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:55.762932+00:00", "epoch": 0, "step": 11251, "train_loss": 3.531623363494873, "perplexity": 34.17940820602113, "lr": 0.0026291804804649314, "grad_norm": 0.183582, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:56.068417+00:00", "epoch": 0, "step": 11252, "train_loss": 3.5587446689605713, "perplexity": 35.119083386762945, "lr": 0.0026291804804649314, "grad_norm": 0.194021, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:56.373276+00:00", "epoch": 0, "step": 11253, "train_loss": 3.5442726612091064, "perplexity": 34.61449971575895, "lr": 0.0026291804804649314, "grad_norm": 0.178634, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:56.677322+00:00", "epoch": 0, "step": 11254, "train_loss": 3.569385051727295, "perplexity": 35.49475899814307, "lr": 0.0026291804804649314, "grad_norm": 0.214797, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:56.981260+00:00", "epoch": 0, "step": 11255, "train_loss": 3.5144691467285156, "perplexity": 33.59808753434392, "lr": 0.0026291804804649314, "grad_norm": 0.160419, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:57.286048+00:00", "epoch": 0, "step": 11256, "train_loss": 3.5237393379211426, "perplexity": 33.91099635091893, "lr": 0.0026291804804649314, "grad_norm": 0.153369, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:57.591630+00:00", "epoch": 0, "step": 11257, "train_loss": 3.4596829414367676, "perplexity": 31.806890268857348, "lr": 0.0026291804804649314, "grad_norm": 0.147064, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:57.896626+00:00", "epoch": 0, "step": 11258, "train_loss": 3.5289266109466553, "perplexity": 34.087358972632785, "lr": 0.0026291804804649314, "grad_norm": 0.141331, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:58.202834+00:00", "epoch": 0, "step": 11259, "train_loss": 3.5692903995513916, "perplexity": 35.491399500965095, "lr": 0.0026291804804649314, "grad_norm": 0.158964, "tokens_per_sec": 107013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:58.506903+00:00", "epoch": 0, "step": 11260, "train_loss": 3.4928059577941895, "perplexity": 32.87807288096928, "lr": 0.0026291804804649314, "grad_norm": 0.157021, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:58.810998+00:00", "epoch": 0, "step": 11261, "train_loss": 3.5295825004577637, "perplexity": 34.109723847485974, "lr": 0.0026291804804649314, "grad_norm": 0.166415, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:59.115558+00:00", "epoch": 0, "step": 11262, "train_loss": 3.535966157913208, "perplexity": 34.328165125766084, "lr": 0.0026291804804649314, "grad_norm": 0.1546, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:59.421739+00:00", "epoch": 0, "step": 11263, "train_loss": 3.510698080062866, "perplexity": 33.47162550465332, "lr": 0.0026291804804649314, "grad_norm": 0.172771, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:56:59.726082+00:00", "epoch": 0, "step": 11264, "train_loss": 3.5407416820526123, "perplexity": 34.49249216856709, "lr": 0.0026291804804649314, "grad_norm": 0.196575, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:00.032051+00:00", "epoch": 0, "step": 11265, "train_loss": 3.510741949081421, "perplexity": 33.473093904222026, "lr": 0.0026291804804649314, "grad_norm": 0.217193, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:00.337153+00:00", "epoch": 0, "step": 11266, "train_loss": 3.530569553375244, "perplexity": 34.1434085714918, "lr": 0.0026291804804649314, "grad_norm": 0.16821, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:00.641837+00:00", "epoch": 0, "step": 11267, "train_loss": 3.63346266746521, "perplexity": 37.84362991106689, "lr": 0.0026291804804649314, "grad_norm": 0.211905, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:00.947462+00:00", "epoch": 0, "step": 11268, "train_loss": 3.484396457672119, "perplexity": 32.60274403464242, "lr": 0.0026291804804649314, "grad_norm": 0.174599, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:01.253318+00:00", "epoch": 0, "step": 11269, "train_loss": 3.3923914432525635, "perplexity": 29.73698160922186, "lr": 0.0026291804804649314, "grad_norm": 0.173237, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:01.558142+00:00", "epoch": 0, "step": 11270, "train_loss": 3.4791388511657715, "perplexity": 32.431781456281456, "lr": 0.0026291804804649314, "grad_norm": 0.173571, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:01.863060+00:00", "epoch": 0, "step": 11271, "train_loss": 3.546246290206909, "perplexity": 34.68288335591261, "lr": 0.0026291804804649314, "grad_norm": 0.154594, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:02.167808+00:00", "epoch": 0, "step": 11272, "train_loss": 3.453970193862915, "perplexity": 31.625703563357664, "lr": 0.0026291804804649314, "grad_norm": 0.168933, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:02.473555+00:00", "epoch": 0, "step": 11273, "train_loss": 3.5182509422302246, "perplexity": 33.725389193345926, "lr": 0.0026291804804649314, "grad_norm": 0.170188, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:02.779296+00:00", "epoch": 0, "step": 11274, "train_loss": 3.4769375324249268, "perplexity": 32.36046728936669, "lr": 0.0026291804804649314, "grad_norm": 0.16876, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:03.083300+00:00", "epoch": 0, "step": 11275, "train_loss": 3.521247625350952, "perplexity": 33.82660507810416, "lr": 0.0026291804804649314, "grad_norm": 0.161928, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:03.388998+00:00", "epoch": 0, "step": 11276, "train_loss": 3.4945318698883057, "perplexity": 32.93486654092011, "lr": 0.0026291804804649314, "grad_norm": 0.155415, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:03.692475+00:00", "epoch": 0, "step": 11277, "train_loss": 3.520172595977783, "perplexity": 33.79026002355152, "lr": 0.0026291804804649314, "grad_norm": 0.14617, "tokens_per_sec": 107976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:03.997820+00:00", "epoch": 0, "step": 11278, "train_loss": 3.6019344329833984, "perplexity": 36.66909979555866, "lr": 0.0026291804804649314, "grad_norm": 0.162603, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:04.303508+00:00", "epoch": 0, "step": 11279, "train_loss": 3.5153768062591553, "perplexity": 33.628597002716006, "lr": 0.0026291804804649314, "grad_norm": 0.174969, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:04.607974+00:00", "epoch": 0, "step": 11280, "train_loss": 3.559264898300171, "perplexity": 35.13735811743396, "lr": 0.0026291804804649314, "grad_norm": 0.160833, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:04.911721+00:00", "epoch": 0, "step": 11281, "train_loss": 3.633942127227783, "perplexity": 37.86177875935253, "lr": 0.0026291804804649314, "grad_norm": 0.155241, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:05.215758+00:00", "epoch": 0, "step": 11282, "train_loss": 3.529451370239258, "perplexity": 34.1052513251925, "lr": 0.0026291804804649314, "grad_norm": 0.172175, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:05.520696+00:00", "epoch": 0, "step": 11283, "train_loss": 3.529435396194458, "perplexity": 34.10470653073122, "lr": 0.0026291804804649314, "grad_norm": 0.155769, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:05.825564+00:00", "epoch": 0, "step": 11284, "train_loss": 3.530197858810425, "perplexity": 34.13072001038205, "lr": 0.0026291804804649314, "grad_norm": 0.162133, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:06.131005+00:00", "epoch": 0, "step": 11285, "train_loss": 3.587204933166504, "perplexity": 36.13294066835723, "lr": 0.0026291804804649314, "grad_norm": 0.155083, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:06.434749+00:00", "epoch": 0, "step": 11286, "train_loss": 3.515321969985962, "perplexity": 33.62675298634361, "lr": 0.0026291804804649314, "grad_norm": 0.158283, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:06.738768+00:00", "epoch": 0, "step": 11287, "train_loss": 3.6494858264923096, "perplexity": 38.454888480019214, "lr": 0.0026291804804649314, "grad_norm": 0.151401, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:07.042712+00:00", "epoch": 0, "step": 11288, "train_loss": 3.5627496242523193, "perplexity": 35.26001577118783, "lr": 0.0026291804804649314, "grad_norm": 0.153378, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:07.346939+00:00", "epoch": 0, "step": 11289, "train_loss": 3.545825719833374, "perplexity": 34.668299829618846, "lr": 0.0026291804804649314, "grad_norm": 0.158923, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:07.651740+00:00", "epoch": 0, "step": 11290, "train_loss": 3.571075916290283, "perplexity": 35.55482659708083, "lr": 0.0026291804804649314, "grad_norm": 0.155399, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:07.955368+00:00", "epoch": 0, "step": 11291, "train_loss": 3.4462685585021973, "perplexity": 31.38306946548321, "lr": 0.0026291804804649314, "grad_norm": 0.190194, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:08.259488+00:00", "epoch": 0, "step": 11292, "train_loss": 3.5689685344696045, "perplexity": 35.479977896968336, "lr": 0.0026291804804649314, "grad_norm": 0.189057, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:08.563725+00:00", "epoch": 0, "step": 11293, "train_loss": 3.521505832672119, "perplexity": 33.83534048290976, "lr": 0.0026291804804649314, "grad_norm": 0.197306, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:08.869470+00:00", "epoch": 0, "step": 11294, "train_loss": 3.654191017150879, "perplexity": 38.63625240339229, "lr": 0.0026291804804649314, "grad_norm": 0.20777, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:09.173740+00:00", "epoch": 0, "step": 11295, "train_loss": 3.476597547531128, "perplexity": 32.3494670893889, "lr": 0.0026291804804649314, "grad_norm": 0.1808, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:09.478563+00:00", "epoch": 0, "step": 11296, "train_loss": 3.6340343952178955, "perplexity": 37.86527235075159, "lr": 0.0026291804804649314, "grad_norm": 0.195619, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:09.782480+00:00", "epoch": 0, "step": 11297, "train_loss": 3.5288331508636475, "perplexity": 34.08417331410196, "lr": 0.0026291804804649314, "grad_norm": 0.183042, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:10.087847+00:00", "epoch": 0, "step": 11298, "train_loss": 3.5147597789764404, "perplexity": 33.607853641154044, "lr": 0.0026291804804649314, "grad_norm": 0.181959, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:10.392316+00:00", "epoch": 0, "step": 11299, "train_loss": 3.492130756378174, "perplexity": 32.85588105242431, "lr": 0.0026291804804649314, "grad_norm": 0.179717, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:10.697211+00:00", "epoch": 0, "step": 11300, "train_loss": 3.4791760444641113, "perplexity": 32.43298772363724, "lr": 0.0026291804804649314, "grad_norm": 0.164236, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:11.001801+00:00", "epoch": 0, "step": 11301, "train_loss": 3.50069522857666, "perplexity": 33.13848277213519, "lr": 0.0026291804804649314, "grad_norm": 0.164073, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:11.306055+00:00", "epoch": 0, "step": 11302, "train_loss": 3.5394229888916016, "perplexity": 34.44703713224796, "lr": 0.0026291804804649314, "grad_norm": 0.154375, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:11.610140+00:00", "epoch": 0, "step": 11303, "train_loss": 3.4534926414489746, "perplexity": 31.61060423792503, "lr": 0.0026291804804649314, "grad_norm": 0.178335, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:11.914102+00:00", "epoch": 0, "step": 11304, "train_loss": 3.50225567817688, "perplexity": 33.190234071474826, "lr": 0.0026291804804649314, "grad_norm": 0.186682, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:12.219762+00:00", "epoch": 0, "step": 11305, "train_loss": 3.655724287033081, "perplexity": 38.69553764408602, "lr": 0.0026291804804649314, "grad_norm": 0.180809, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:12.525129+00:00", "epoch": 0, "step": 11306, "train_loss": 3.4819347858428955, "perplexity": 32.52258548060305, "lr": 0.0026291804804649314, "grad_norm": 0.176683, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:12.830835+00:00", "epoch": 0, "step": 11307, "train_loss": 3.5380985736846924, "perplexity": 34.40144515045836, "lr": 0.0026291804804649314, "grad_norm": 0.195236, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:13.134902+00:00", "epoch": 0, "step": 11308, "train_loss": 3.5480713844299316, "perplexity": 34.746240684924054, "lr": 0.0026291804804649314, "grad_norm": 0.171131, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:13.438726+00:00", "epoch": 0, "step": 11309, "train_loss": 3.6076626777648926, "perplexity": 36.87975213317425, "lr": 0.0026291804804649314, "grad_norm": 0.158017, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:13.743084+00:00", "epoch": 0, "step": 11310, "train_loss": 3.544377565383911, "perplexity": 34.61813111175888, "lr": 0.0026291804804649314, "grad_norm": 0.17606, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:14.047856+00:00", "epoch": 0, "step": 11311, "train_loss": 3.4111316204071045, "perplexity": 30.299512413044127, "lr": 0.0026291804804649314, "grad_norm": 0.170555, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:14.352127+00:00", "epoch": 0, "step": 11312, "train_loss": 3.506990671157837, "perplexity": 33.34776224993211, "lr": 0.0026291804804649314, "grad_norm": 0.158894, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:14.656125+00:00", "epoch": 0, "step": 11313, "train_loss": 3.546034097671509, "perplexity": 34.675524687713164, "lr": 0.0026291804804649314, "grad_norm": 0.154111, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:14.960388+00:00", "epoch": 0, "step": 11314, "train_loss": 3.5211281776428223, "perplexity": 33.822564808959015, "lr": 0.0026291804804649314, "grad_norm": 0.153825, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:15.266230+00:00", "epoch": 0, "step": 11315, "train_loss": 3.5363030433654785, "perplexity": 34.33973173340162, "lr": 0.0026291804804649314, "grad_norm": 0.188541, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:15.570394+00:00", "epoch": 0, "step": 11316, "train_loss": 3.4969849586486816, "perplexity": 33.01575786809428, "lr": 0.0026291804804649314, "grad_norm": 0.1862, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:15.875392+00:00", "epoch": 0, "step": 11317, "train_loss": 3.597956657409668, "perplexity": 36.52352806386523, "lr": 0.0026291804804649314, "grad_norm": 0.190566, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:16.180634+00:00", "epoch": 0, "step": 11318, "train_loss": 3.489190101623535, "perplexity": 32.759405170243, "lr": 0.0026291804804649314, "grad_norm": 0.149821, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:16.486376+00:00", "epoch": 0, "step": 11319, "train_loss": 3.5520377159118652, "perplexity": 34.88432946513777, "lr": 0.0026291804804649314, "grad_norm": 0.144457, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:16.791252+00:00", "epoch": 0, "step": 11320, "train_loss": 3.395188331604004, "perplexity": 29.82026904522791, "lr": 0.0026291804804649314, "grad_norm": 0.154479, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:17.096899+00:00", "epoch": 0, "step": 11321, "train_loss": 3.5383739471435547, "perplexity": 34.41091969985915, "lr": 0.0026291804804649314, "grad_norm": 0.139578, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:17.403813+00:00", "epoch": 0, "step": 11322, "train_loss": 3.4966843128204346, "perplexity": 33.005833310188045, "lr": 0.0026291804804649314, "grad_norm": 0.145692, "tokens_per_sec": 106766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:17.708189+00:00", "epoch": 0, "step": 11323, "train_loss": 3.521801233291626, "perplexity": 33.84533693985673, "lr": 0.0026291804804649314, "grad_norm": 0.139869, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:18.011769+00:00", "epoch": 0, "step": 11324, "train_loss": 3.5141146183013916, "perplexity": 33.58617816844511, "lr": 0.0026291804804649314, "grad_norm": 0.150648, "tokens_per_sec": 107939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:18.315083+00:00", "epoch": 0, "step": 11325, "train_loss": 3.62473726272583, "perplexity": 37.514865311192885, "lr": 0.0026291804804649314, "grad_norm": 0.16079, "tokens_per_sec": 108033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:18.619095+00:00", "epoch": 0, "step": 11326, "train_loss": 3.528848171234131, "perplexity": 34.08468527485767, "lr": 0.0026291804804649314, "grad_norm": 0.155459, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:18.923670+00:00", "epoch": 0, "step": 11327, "train_loss": 3.477679491043091, "perplexity": 32.38448632642375, "lr": 0.0026291804804649314, "grad_norm": 0.150682, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:19.229136+00:00", "epoch": 0, "step": 11328, "train_loss": 3.4870338439941406, "perplexity": 32.688843554760474, "lr": 0.0026291804804649314, "grad_norm": 0.175642, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:19.535313+00:00", "epoch": 0, "step": 11329, "train_loss": 3.4376697540283203, "perplexity": 31.114369492250457, "lr": 0.0026291804804649314, "grad_norm": 0.147172, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:19.839798+00:00", "epoch": 0, "step": 11330, "train_loss": 3.5131475925445557, "perplexity": 33.55371516789117, "lr": 0.0026291804804649314, "grad_norm": 0.165937, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:20.144123+00:00", "epoch": 0, "step": 11331, "train_loss": 3.6263279914855957, "perplexity": 37.57458877568465, "lr": 0.0026291804804649314, "grad_norm": 0.166611, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:20.448462+00:00", "epoch": 0, "step": 11332, "train_loss": 3.563758134841919, "perplexity": 35.29559380788039, "lr": 0.0026291804804649314, "grad_norm": 0.146729, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:20.752910+00:00", "epoch": 0, "step": 11333, "train_loss": 3.435154438018799, "perplexity": 31.03620536549096, "lr": 0.0026291804804649314, "grad_norm": 0.17456, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:21.057725+00:00", "epoch": 0, "step": 11334, "train_loss": 3.3985507488250732, "perplexity": 29.920705992235565, "lr": 0.0026291804804649314, "grad_norm": 0.189603, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:21.362549+00:00", "epoch": 0, "step": 11335, "train_loss": 3.5918307304382324, "perplexity": 36.300471509839866, "lr": 0.0026291804804649314, "grad_norm": 0.171817, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:21.668055+00:00", "epoch": 0, "step": 11336, "train_loss": 3.564558267593384, "perplexity": 35.32384626982123, "lr": 0.0026291804804649314, "grad_norm": 0.163958, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:21.973574+00:00", "epoch": 0, "step": 11337, "train_loss": 3.496814727783203, "perplexity": 33.01013804540518, "lr": 0.0026291804804649314, "grad_norm": 0.194288, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:22.278929+00:00", "epoch": 0, "step": 11338, "train_loss": 3.52656626701355, "perplexity": 34.00699596116944, "lr": 0.0026291804804649314, "grad_norm": 0.157193, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:22.583775+00:00", "epoch": 0, "step": 11339, "train_loss": 3.517267942428589, "perplexity": 33.69225343135074, "lr": 0.0026291804804649314, "grad_norm": 0.144382, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:22.888369+00:00", "epoch": 0, "step": 11340, "train_loss": 3.4461941719055176, "perplexity": 31.38073507257716, "lr": 0.0026291804804649314, "grad_norm": 0.148276, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:23.192917+00:00", "epoch": 0, "step": 11341, "train_loss": 3.369882345199585, "perplexity": 29.075106030753798, "lr": 0.0026291804804649314, "grad_norm": 0.164804, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:23.497852+00:00", "epoch": 0, "step": 11342, "train_loss": 3.4479753971099854, "perplexity": 31.436681040215042, "lr": 0.0026291804804649314, "grad_norm": 0.151604, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:23.803511+00:00", "epoch": 0, "step": 11343, "train_loss": 3.5330750942230225, "perplexity": 34.229063537548264, "lr": 0.0026291804804649314, "grad_norm": 0.144579, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:24.109421+00:00", "epoch": 0, "step": 11344, "train_loss": 3.4757726192474365, "perplexity": 32.322792103012596, "lr": 0.0026291804804649314, "grad_norm": 0.145713, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:24.414265+00:00", "epoch": 0, "step": 11345, "train_loss": 3.595418930053711, "perplexity": 36.43095881499836, "lr": 0.0026291804804649314, "grad_norm": 0.148208, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:24.719906+00:00", "epoch": 0, "step": 11346, "train_loss": 3.5765538215637207, "perplexity": 35.750126999372995, "lr": 0.0026291804804649314, "grad_norm": 0.148995, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:25.024274+00:00", "epoch": 0, "step": 11347, "train_loss": 3.495135545730591, "perplexity": 32.954754526564024, "lr": 0.0026291804804649314, "grad_norm": 0.163151, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:25.328475+00:00", "epoch": 0, "step": 11348, "train_loss": 3.5068306922912598, "perplexity": 33.34242773944033, "lr": 0.0026291804804649314, "grad_norm": 0.159337, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:25.634546+00:00", "epoch": 0, "step": 11349, "train_loss": 3.554252862930298, "perplexity": 34.96168903331807, "lr": 0.0026291804804649314, "grad_norm": 0.170163, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:25.939154+00:00", "epoch": 0, "step": 11350, "train_loss": 3.5653927326202393, "perplexity": 35.35333508612919, "lr": 0.0026291804804649314, "grad_norm": 0.150156, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:26.243479+00:00", "epoch": 0, "step": 11351, "train_loss": 3.4702515602111816, "perplexity": 32.14482778740186, "lr": 0.0026291804804649314, "grad_norm": 0.155692, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:26.546559+00:00", "epoch": 0, "step": 11352, "train_loss": 3.5262980461120605, "perplexity": 33.99787579721977, "lr": 0.0026291804804649314, "grad_norm": 0.160746, "tokens_per_sec": 108111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:26.850164+00:00", "epoch": 0, "step": 11353, "train_loss": 3.5927276611328125, "perplexity": 36.333045122917724, "lr": 0.0026291804804649314, "grad_norm": 0.140113, "tokens_per_sec": 107930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:27.154542+00:00", "epoch": 0, "step": 11354, "train_loss": 3.585012435913086, "perplexity": 36.05380607807016, "lr": 0.0026291804804649314, "grad_norm": 0.176826, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:27.458337+00:00", "epoch": 0, "step": 11355, "train_loss": 3.643056869506836, "perplexity": 38.2084566546847, "lr": 0.0026291804804649314, "grad_norm": 0.184044, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:27.762213+00:00", "epoch": 0, "step": 11356, "train_loss": 3.5432956218719482, "perplexity": 34.58069650412341, "lr": 0.0026291804804649314, "grad_norm": 0.20103, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:28.066960+00:00", "epoch": 0, "step": 11357, "train_loss": 3.504014730453491, "perplexity": 33.248668808090045, "lr": 0.0026291804804649314, "grad_norm": 0.216914, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:28.370685+00:00", "epoch": 0, "step": 11358, "train_loss": 3.4495484828948975, "perplexity": 31.486172553276084, "lr": 0.0026291804804649314, "grad_norm": 0.174432, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:28.674894+00:00", "epoch": 0, "step": 11359, "train_loss": 3.5496346950531006, "perplexity": 34.800602333101644, "lr": 0.0026291804804649314, "grad_norm": 0.172641, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:28.980564+00:00", "epoch": 0, "step": 11360, "train_loss": 3.428790807723999, "perplexity": 30.839329515738875, "lr": 0.0026291804804649314, "grad_norm": 0.192968, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:29.284351+00:00", "epoch": 0, "step": 11361, "train_loss": 3.6400792598724365, "perplexity": 38.09485599917079, "lr": 0.0026291804804649314, "grad_norm": 0.206911, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:29.588373+00:00", "epoch": 0, "step": 11362, "train_loss": 3.5997426509857178, "perplexity": 36.58881713593885, "lr": 0.0026291804804649314, "grad_norm": 0.163977, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:29.892619+00:00", "epoch": 0, "step": 11363, "train_loss": 3.4776179790496826, "perplexity": 32.382494353380054, "lr": 0.0026291804804649314, "grad_norm": 0.181433, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:30.197040+00:00", "epoch": 0, "step": 11364, "train_loss": 3.5119543075561523, "perplexity": 33.51369990283196, "lr": 0.0026291804804649314, "grad_norm": 0.170173, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:30.501542+00:00", "epoch": 0, "step": 11365, "train_loss": 3.629880666732788, "perplexity": 37.70831649207172, "lr": 0.0026291804804649314, "grad_norm": 0.177298, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:30.806486+00:00", "epoch": 0, "step": 11366, "train_loss": 3.529536724090576, "perplexity": 34.10816246397997, "lr": 0.0026291804804649314, "grad_norm": 0.161222, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:31.110745+00:00", "epoch": 0, "step": 11367, "train_loss": 3.528162717819214, "perplexity": 34.06132981639348, "lr": 0.0026291804804649314, "grad_norm": 0.148366, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:31.414631+00:00", "epoch": 0, "step": 11368, "train_loss": 3.4596450328826904, "perplexity": 31.805684538491455, "lr": 0.0026291804804649314, "grad_norm": 0.152768, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:31.720485+00:00", "epoch": 0, "step": 11369, "train_loss": 3.4574778079986572, "perplexity": 31.736829107058508, "lr": 0.0026291804804649314, "grad_norm": 0.147486, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:32.026065+00:00", "epoch": 0, "step": 11370, "train_loss": 3.5620224475860596, "perplexity": 35.23438483070744, "lr": 0.0026291804804649314, "grad_norm": 0.159951, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:32.331253+00:00", "epoch": 0, "step": 11371, "train_loss": 3.5703909397125244, "perplexity": 35.53048071276773, "lr": 0.0026291804804649314, "grad_norm": 0.172271, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:32.634869+00:00", "epoch": 0, "step": 11372, "train_loss": 3.589103937149048, "perplexity": 36.201622459444756, "lr": 0.0026291804804649314, "grad_norm": 0.168985, "tokens_per_sec": 107926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:32.938733+00:00", "epoch": 0, "step": 11373, "train_loss": 3.588759183883667, "perplexity": 36.189143983011114, "lr": 0.0026291804804649314, "grad_norm": 0.156538, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:33.244065+00:00", "epoch": 0, "step": 11374, "train_loss": 3.4708189964294434, "perplexity": 32.16307310295509, "lr": 0.0026291804804649314, "grad_norm": 0.155922, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:33.549886+00:00", "epoch": 0, "step": 11375, "train_loss": 3.5733509063720703, "perplexity": 35.63580555318375, "lr": 0.0026291804804649314, "grad_norm": 0.154708, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:33.854522+00:00", "epoch": 0, "step": 11376, "train_loss": 3.512443780899048, "perplexity": 33.53010798088701, "lr": 0.0026291804804649314, "grad_norm": 0.151548, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:34.158091+00:00", "epoch": 0, "step": 11377, "train_loss": 3.441209077835083, "perplexity": 31.224688432991403, "lr": 0.0026291804804649314, "grad_norm": 0.209125, "tokens_per_sec": 107943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:34.462908+00:00", "epoch": 0, "step": 11378, "train_loss": 3.6065661907196045, "perplexity": 36.839336124591, "lr": 0.0026291804804649314, "grad_norm": 0.187814, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:34.766607+00:00", "epoch": 0, "step": 11379, "train_loss": 3.5148937702178955, "perplexity": 33.6123571008914, "lr": 0.0026291804804649314, "grad_norm": 0.167205, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:35.072312+00:00", "epoch": 0, "step": 11380, "train_loss": 3.607142686843872, "perplexity": 36.860579981999784, "lr": 0.0026291804804649314, "grad_norm": 0.183442, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:35.377078+00:00", "epoch": 0, "step": 11381, "train_loss": 3.4651968479156494, "perplexity": 31.982754892156926, "lr": 0.0026291804804649314, "grad_norm": 0.159869, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:35.682381+00:00", "epoch": 0, "step": 11382, "train_loss": 3.4479823112487793, "perplexity": 31.436898398542397, "lr": 0.0026291804804649314, "grad_norm": 0.152368, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:35.987310+00:00", "epoch": 0, "step": 11383, "train_loss": 3.5951218605041504, "perplexity": 36.4201378938358, "lr": 0.0026291804804649314, "grad_norm": 0.149134, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:36.291592+00:00", "epoch": 0, "step": 11384, "train_loss": 3.521773099899292, "perplexity": 33.84438476910788, "lr": 0.0026291804804649314, "grad_norm": 0.17004, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:36.596756+00:00", "epoch": 0, "step": 11385, "train_loss": 3.732762098312378, "perplexity": 41.79438909368957, "lr": 0.0026291804804649314, "grad_norm": 0.160281, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:36.902317+00:00", "epoch": 0, "step": 11386, "train_loss": 3.5066988468170166, "perplexity": 33.3380319810294, "lr": 0.0026291804804649314, "grad_norm": 0.175667, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:37.208097+00:00", "epoch": 0, "step": 11387, "train_loss": 3.5449178218841553, "perplexity": 34.63683883514786, "lr": 0.0026291804804649314, "grad_norm": 0.169989, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:37.512451+00:00", "epoch": 0, "step": 11388, "train_loss": 3.5698564052581787, "perplexity": 35.51149352175327, "lr": 0.0026291804804649314, "grad_norm": 0.184664, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:37.816668+00:00", "epoch": 0, "step": 11389, "train_loss": 3.487455129623413, "perplexity": 32.70261779602885, "lr": 0.0026291804804649314, "grad_norm": 0.200079, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:38.121387+00:00", "epoch": 0, "step": 11390, "train_loss": 3.4952023029327393, "perplexity": 32.95695456720717, "lr": 0.0026291804804649314, "grad_norm": 0.19447, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:38.426670+00:00", "epoch": 0, "step": 11391, "train_loss": 3.556077241897583, "perplexity": 35.02553062136916, "lr": 0.0026291804804649314, "grad_norm": 0.166264, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:38.732385+00:00", "epoch": 0, "step": 11392, "train_loss": 3.642981767654419, "perplexity": 38.20558723656259, "lr": 0.0026291804804649314, "grad_norm": 0.148283, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:39.037881+00:00", "epoch": 0, "step": 11393, "train_loss": 3.452547073364258, "perplexity": 31.580728386470852, "lr": 0.0026291804804649314, "grad_norm": 0.164446, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:39.343258+00:00", "epoch": 0, "step": 11394, "train_loss": 3.49336576461792, "perplexity": 32.89648340320634, "lr": 0.0026291804804649314, "grad_norm": 0.168796, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:39.647050+00:00", "epoch": 0, "step": 11395, "train_loss": 3.455152750015259, "perplexity": 31.663124855707512, "lr": 0.0026291804804649314, "grad_norm": 0.189935, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:39.952213+00:00", "epoch": 0, "step": 11396, "train_loss": 3.5466113090515137, "perplexity": 34.695545572756025, "lr": 0.0026291804804649314, "grad_norm": 0.147305, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:40.260047+00:00", "epoch": 0, "step": 11397, "train_loss": 3.6038565635681152, "perplexity": 36.739650375782475, "lr": 0.0026291804804649314, "grad_norm": 0.168386, "tokens_per_sec": 106456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:40.565410+00:00", "epoch": 0, "step": 11398, "train_loss": 3.4216344356536865, "perplexity": 30.619419616387916, "lr": 0.0026291804804649314, "grad_norm": 0.158431, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:40.870269+00:00", "epoch": 0, "step": 11399, "train_loss": 3.5630557537078857, "perplexity": 35.27081155298822, "lr": 0.0026291804804649314, "grad_norm": 0.144772, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:41.174411+00:00", "epoch": 0, "step": 11400, "train_loss": 3.473642110824585, "perplexity": 32.25400142776402, "lr": 0.0026291804804649314, "grad_norm": 0.155149, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:41.479933+00:00", "epoch": 0, "step": 11401, "train_loss": 3.427340030670166, "perplexity": 30.794620963034742, "lr": 0.0026291804804649314, "grad_norm": 0.142041, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:41.785354+00:00", "epoch": 0, "step": 11402, "train_loss": 3.609241247177124, "perplexity": 36.9380153559943, "lr": 0.0026291804804649314, "grad_norm": 0.172093, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:42.092796+00:00", "epoch": 0, "step": 11403, "train_loss": 3.567997932434082, "perplexity": 35.445557665077, "lr": 0.0026291804804649314, "grad_norm": 0.165769, "tokens_per_sec": 106589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:42.397528+00:00", "epoch": 0, "step": 11404, "train_loss": 3.517836809158325, "perplexity": 33.711425285975785, "lr": 0.0026291804804649314, "grad_norm": 0.150608, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:42.702253+00:00", "epoch": 0, "step": 11405, "train_loss": 3.5290286540985107, "perplexity": 34.090837531659325, "lr": 0.0026291804804649314, "grad_norm": 0.171224, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:43.008264+00:00", "epoch": 0, "step": 11406, "train_loss": 3.530426263809204, "perplexity": 34.138516527792234, "lr": 0.0026291804804649314, "grad_norm": 0.180164, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:43.313172+00:00", "epoch": 0, "step": 11407, "train_loss": 3.4991238117218018, "perplexity": 33.08644929561105, "lr": 0.0026291804804649314, "grad_norm": 0.179144, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:43.619036+00:00", "epoch": 0, "step": 11408, "train_loss": 3.560421943664551, "perplexity": 35.17803716397314, "lr": 0.0026291804804649314, "grad_norm": 0.189177, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:43.924270+00:00", "epoch": 0, "step": 11409, "train_loss": 3.5641963481903076, "perplexity": 35.31106419764439, "lr": 0.0026291804804649314, "grad_norm": 0.179191, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:44.228795+00:00", "epoch": 0, "step": 11410, "train_loss": 3.5923101902008057, "perplexity": 36.31788029836458, "lr": 0.0026291804804649314, "grad_norm": 0.194215, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:44.533534+00:00", "epoch": 0, "step": 11411, "train_loss": 3.5962815284729004, "perplexity": 36.462397660081365, "lr": 0.0026291804804649314, "grad_norm": 0.175861, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:44.838531+00:00", "epoch": 0, "step": 11412, "train_loss": 3.5280601978302, "perplexity": 34.05783802822702, "lr": 0.0026291804804649314, "grad_norm": 0.17102, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:45.143573+00:00", "epoch": 0, "step": 11413, "train_loss": 3.473984956741333, "perplexity": 32.265061476290214, "lr": 0.0026291804804649314, "grad_norm": 0.15647, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:45.447825+00:00", "epoch": 0, "step": 11414, "train_loss": 3.477762460708618, "perplexity": 32.38717336789245, "lr": 0.0026291804804649314, "grad_norm": 0.171448, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:45.752263+00:00", "epoch": 0, "step": 11415, "train_loss": 3.5064806938171387, "perplexity": 33.33075998257569, "lr": 0.0026291804804649314, "grad_norm": 0.17973, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:46.057609+00:00", "epoch": 0, "step": 11416, "train_loss": 3.527773857116699, "perplexity": 34.04808727867063, "lr": 0.0026291804804649314, "grad_norm": 0.174278, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:46.362836+00:00", "epoch": 0, "step": 11417, "train_loss": 3.535515308380127, "perplexity": 34.312691776885764, "lr": 0.0026291804804649314, "grad_norm": 0.171541, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:46.667506+00:00", "epoch": 0, "step": 11418, "train_loss": 3.626227855682373, "perplexity": 37.570826402434115, "lr": 0.0026291804804649314, "grad_norm": 0.177932, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:46.970458+00:00", "epoch": 0, "step": 11419, "train_loss": 3.5234594345092773, "perplexity": 33.90150587561073, "lr": 0.0026291804804649314, "grad_norm": 0.158152, "tokens_per_sec": 108163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:47.275027+00:00", "epoch": 0, "step": 11420, "train_loss": 3.5103065967559814, "perplexity": 33.458524486596204, "lr": 0.0026291804804649314, "grad_norm": 0.173286, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:47.581114+00:00", "epoch": 0, "step": 11421, "train_loss": 3.471653461456299, "perplexity": 32.18992326381775, "lr": 0.0026291804804649314, "grad_norm": 0.177856, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:47.886310+00:00", "epoch": 0, "step": 11422, "train_loss": 3.5310580730438232, "perplexity": 34.160092372982085, "lr": 0.0026291804804649314, "grad_norm": 0.142978, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:48.190551+00:00", "epoch": 0, "step": 11423, "train_loss": 3.4550533294677734, "perplexity": 31.659977046980288, "lr": 0.0026291804804649314, "grad_norm": 0.164193, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:48.496135+00:00", "epoch": 0, "step": 11424, "train_loss": 3.523228883743286, "perplexity": 33.8936907583881, "lr": 0.0026291804804649314, "grad_norm": 0.165271, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:48.801860+00:00", "epoch": 0, "step": 11425, "train_loss": 3.4937243461608887, "perplexity": 32.908281590162936, "lr": 0.0026291804804649314, "grad_norm": 0.163081, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:49.105872+00:00", "epoch": 0, "step": 11426, "train_loss": 3.5711240768432617, "perplexity": 35.55653897842509, "lr": 0.0026291804804649314, "grad_norm": 0.178753, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:49.409942+00:00", "epoch": 0, "step": 11427, "train_loss": 3.511850357055664, "perplexity": 33.51021631801717, "lr": 0.0026291804804649314, "grad_norm": 0.174712, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:49.713839+00:00", "epoch": 0, "step": 11428, "train_loss": 3.485882043838501, "perplexity": 32.65121421456642, "lr": 0.0026291804804649314, "grad_norm": 0.200578, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:50.018883+00:00", "epoch": 0, "step": 11429, "train_loss": 3.6288950443267822, "perplexity": 37.67116864032915, "lr": 0.0026291804804649314, "grad_norm": 0.178548, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:50.325470+00:00", "epoch": 0, "step": 11430, "train_loss": 3.5569961071014404, "perplexity": 35.057729153493696, "lr": 0.0026291804804649314, "grad_norm": 0.195039, "tokens_per_sec": 106942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:50.629666+00:00", "epoch": 0, "step": 11431, "train_loss": 3.4808897972106934, "perplexity": 32.488617499654055, "lr": 0.0026291804804649314, "grad_norm": 0.173405, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:50.933752+00:00", "epoch": 0, "step": 11432, "train_loss": 3.606919050216675, "perplexity": 36.85233752790774, "lr": 0.0026291804804649314, "grad_norm": 0.176925, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:51.237366+00:00", "epoch": 0, "step": 11433, "train_loss": 3.5451412200927734, "perplexity": 34.64457750726639, "lr": 0.0026291804804649314, "grad_norm": 0.149154, "tokens_per_sec": 107927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:51.542227+00:00", "epoch": 0, "step": 11434, "train_loss": 3.5740573406219482, "perplexity": 35.66098880085554, "lr": 0.0026291804804649314, "grad_norm": 0.180274, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:51.847953+00:00", "epoch": 0, "step": 11435, "train_loss": 3.5677919387817383, "perplexity": 35.438256857180065, "lr": 0.0026291804804649314, "grad_norm": 0.181002, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:52.153656+00:00", "epoch": 0, "step": 11436, "train_loss": 3.543713331222534, "perplexity": 34.595144201662066, "lr": 0.0026291804804649314, "grad_norm": 0.17242, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:52.458459+00:00", "epoch": 0, "step": 11437, "train_loss": 3.5791690349578857, "perplexity": 35.843743570644115, "lr": 0.0026291804804649314, "grad_norm": 0.169615, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:52.762513+00:00", "epoch": 0, "step": 11438, "train_loss": 3.6042473316192627, "perplexity": 36.754009862791314, "lr": 0.0026291804804649314, "grad_norm": 0.172845, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:53.066676+00:00", "epoch": 0, "step": 11439, "train_loss": 3.5754342079162598, "perplexity": 35.71012306793869, "lr": 0.0026291804804649314, "grad_norm": 0.1703, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:53.371674+00:00", "epoch": 0, "step": 11440, "train_loss": 3.445826292037964, "perplexity": 31.36919285511981, "lr": 0.0026291804804649314, "grad_norm": 0.182152, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:53.675485+00:00", "epoch": 0, "step": 11441, "train_loss": 3.5937132835388184, "perplexity": 36.368873439964865, "lr": 0.0026291804804649314, "grad_norm": 0.185533, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:53.979101+00:00", "epoch": 0, "step": 11442, "train_loss": 3.443047285079956, "perplexity": 31.28213866802991, "lr": 0.0026291804804649314, "grad_norm": 0.14766, "tokens_per_sec": 107934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:54.283075+00:00", "epoch": 0, "step": 11443, "train_loss": 3.5455284118652344, "perplexity": 34.657994199886474, "lr": 0.0026291804804649314, "grad_norm": 0.179739, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:54.587439+00:00", "epoch": 0, "step": 11444, "train_loss": 3.5026729106903076, "perplexity": 33.20408500558681, "lr": 0.0026291804804649314, "grad_norm": 0.165573, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:54.893429+00:00", "epoch": 0, "step": 11445, "train_loss": 3.45404314994812, "perplexity": 31.628010935048884, "lr": 0.0026291804804649314, "grad_norm": 0.182376, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:55.197985+00:00", "epoch": 0, "step": 11446, "train_loss": 3.5575642585754395, "perplexity": 35.07765291330812, "lr": 0.0026291804804649314, "grad_norm": 0.18631, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:55.501470+00:00", "epoch": 0, "step": 11447, "train_loss": 3.529188871383667, "perplexity": 34.09629991066943, "lr": 0.0026291804804649314, "grad_norm": 0.176719, "tokens_per_sec": 107972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:55.806325+00:00", "epoch": 0, "step": 11448, "train_loss": 3.532421112060547, "perplexity": 34.20668565872082, "lr": 0.0026291804804649314, "grad_norm": 0.166321, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:56.110491+00:00", "epoch": 0, "step": 11449, "train_loss": 3.626075267791748, "perplexity": 37.56509398664407, "lr": 0.0026291804804649314, "grad_norm": 0.170061, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:56.414645+00:00", "epoch": 0, "step": 11450, "train_loss": 3.6128318309783936, "perplexity": 37.070882788661194, "lr": 0.0026291804804649314, "grad_norm": 0.183814, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:56.719454+00:00", "epoch": 0, "step": 11451, "train_loss": 3.541226863861084, "perplexity": 34.509231358743044, "lr": 0.0026291804804649314, "grad_norm": 0.161962, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:57.024613+00:00", "epoch": 0, "step": 11452, "train_loss": 3.5435497760772705, "perplexity": 34.58948645051651, "lr": 0.0026291804804649314, "grad_norm": 0.146739, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:57.330634+00:00", "epoch": 0, "step": 11453, "train_loss": 3.513627529144287, "perplexity": 33.569822688842436, "lr": 0.0026291804804649314, "grad_norm": 0.17683, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:57.634940+00:00", "epoch": 0, "step": 11454, "train_loss": 3.4892818927764893, "perplexity": 32.76241233182697, "lr": 0.0026291804804649314, "grad_norm": 0.179146, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:57.939095+00:00", "epoch": 0, "step": 11455, "train_loss": 3.5544800758361816, "perplexity": 34.969633682807206, "lr": 0.0026291804804649314, "grad_norm": 0.142187, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:58.244966+00:00", "epoch": 0, "step": 11456, "train_loss": 3.538602113723755, "perplexity": 34.41877201751278, "lr": 0.0026291804804649314, "grad_norm": 0.170683, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:58.549947+00:00", "epoch": 0, "step": 11457, "train_loss": 3.584381580352783, "perplexity": 36.03106850685143, "lr": 0.0026291804804649314, "grad_norm": 0.193456, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:58.854364+00:00", "epoch": 0, "step": 11458, "train_loss": 3.5293939113616943, "perplexity": 34.10329173203088, "lr": 0.0026291804804649314, "grad_norm": 0.182273, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:59.158520+00:00", "epoch": 0, "step": 11459, "train_loss": 3.491325855255127, "perplexity": 32.82944595711267, "lr": 0.0026291804804649314, "grad_norm": 0.159762, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:59.462839+00:00", "epoch": 0, "step": 11460, "train_loss": 3.5443480014801025, "perplexity": 34.61710767978906, "lr": 0.0026291804804649314, "grad_norm": 0.189652, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:57:59.768430+00:00", "epoch": 0, "step": 11461, "train_loss": 3.5784454345703125, "perplexity": 35.817816405488976, "lr": 0.0026291804804649314, "grad_norm": 0.176456, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:00.073048+00:00", "epoch": 0, "step": 11462, "train_loss": 3.5863120555877686, "perplexity": 36.10069277463373, "lr": 0.0026291804804649314, "grad_norm": 0.158902, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:00.376899+00:00", "epoch": 0, "step": 11463, "train_loss": 3.6118476390838623, "perplexity": 37.03441587446175, "lr": 0.0026291804804649314, "grad_norm": 0.194545, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:00.680573+00:00", "epoch": 0, "step": 11464, "train_loss": 3.4377951622009277, "perplexity": 31.118271733151648, "lr": 0.0026291804804649314, "grad_norm": 0.167784, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:00.985475+00:00", "epoch": 0, "step": 11465, "train_loss": 3.538120985031128, "perplexity": 34.40221614180295, "lr": 0.0026291804804649314, "grad_norm": 0.160368, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:01.289663+00:00", "epoch": 0, "step": 11466, "train_loss": 3.5646555423736572, "perplexity": 35.327282556334836, "lr": 0.0026291804804649314, "grad_norm": 0.166504, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:01.594816+00:00", "epoch": 0, "step": 11467, "train_loss": 3.514422655105591, "perplexity": 33.59652554103736, "lr": 0.0026291804804649314, "grad_norm": 0.175561, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:01.899909+00:00", "epoch": 0, "step": 11468, "train_loss": 3.5265097618103027, "perplexity": 34.005074443239224, "lr": 0.0026291804804649314, "grad_norm": 0.160021, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:02.204392+00:00", "epoch": 0, "step": 11469, "train_loss": 3.5109007358551025, "perplexity": 33.47840941081312, "lr": 0.0026291804804649314, "grad_norm": 0.170963, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:02.508335+00:00", "epoch": 0, "step": 11470, "train_loss": 3.596848249435425, "perplexity": 36.48306752164794, "lr": 0.0026291804804649314, "grad_norm": 0.153568, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:02.812866+00:00", "epoch": 0, "step": 11471, "train_loss": 3.480180025100708, "perplexity": 32.46556616662519, "lr": 0.0026291804804649314, "grad_norm": 0.145307, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:03.117886+00:00", "epoch": 0, "step": 11472, "train_loss": 3.5513696670532227, "perplexity": 34.86103281117204, "lr": 0.0026291804804649314, "grad_norm": 0.163696, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:03.422210+00:00", "epoch": 0, "step": 11473, "train_loss": 3.507460594177246, "perplexity": 33.36343681368985, "lr": 0.0026291804804649314, "grad_norm": 0.183915, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:03.725754+00:00", "epoch": 0, "step": 11474, "train_loss": 3.6384406089782715, "perplexity": 38.03248294711032, "lr": 0.0026291804804649314, "grad_norm": 0.151017, "tokens_per_sec": 107952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:04.029831+00:00", "epoch": 0, "step": 11475, "train_loss": 3.466623306274414, "perplexity": 32.02840951467668, "lr": 0.0026291804804649314, "grad_norm": 0.154515, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:04.334716+00:00", "epoch": 0, "step": 11476, "train_loss": 3.574012279510498, "perplexity": 35.6593819132691, "lr": 0.0026291804804649314, "grad_norm": 0.170529, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:04.639629+00:00", "epoch": 0, "step": 11477, "train_loss": 3.5581681728363037, "perplexity": 35.098843206056074, "lr": 0.0026291804804649314, "grad_norm": 0.164411, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:04.943776+00:00", "epoch": 0, "step": 11478, "train_loss": 3.523843765258789, "perplexity": 33.91453777089211, "lr": 0.0026291804804649314, "grad_norm": 0.169193, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:05.247780+00:00", "epoch": 0, "step": 11479, "train_loss": 3.5994248390197754, "perplexity": 36.577190619654225, "lr": 0.0026291804804649314, "grad_norm": 0.174919, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:05.551401+00:00", "epoch": 0, "step": 11480, "train_loss": 3.4879302978515625, "perplexity": 32.71816073344541, "lr": 0.0026291804804649314, "grad_norm": 0.174602, "tokens_per_sec": 107983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:05.855468+00:00", "epoch": 0, "step": 11481, "train_loss": 3.4074039459228516, "perplexity": 30.18677594670586, "lr": 0.0026291804804649314, "grad_norm": 0.163439, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:06.159666+00:00", "epoch": 0, "step": 11482, "train_loss": 3.465845823287964, "perplexity": 32.0035176489509, "lr": 0.0026291804804649314, "grad_norm": 0.151377, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:06.463947+00:00", "epoch": 0, "step": 11483, "train_loss": 3.500431776046753, "perplexity": 33.129753504935834, "lr": 0.0026291804804649314, "grad_norm": 0.143343, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:06.768186+00:00", "epoch": 0, "step": 11484, "train_loss": 3.6129212379455566, "perplexity": 37.07419733203081, "lr": 0.0026291804804649314, "grad_norm": 0.175133, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:07.071884+00:00", "epoch": 0, "step": 11485, "train_loss": 3.5491302013397217, "perplexity": 34.78305007587678, "lr": 0.0026291804804649314, "grad_norm": 0.168802, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:07.376264+00:00", "epoch": 0, "step": 11486, "train_loss": 3.550185203552246, "perplexity": 34.819765634757644, "lr": 0.0026291804804649314, "grad_norm": 0.164009, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:07.679952+00:00", "epoch": 0, "step": 11487, "train_loss": 3.5566322803497314, "perplexity": 35.04497653378626, "lr": 0.0026291804804649314, "grad_norm": 0.155176, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:07.984746+00:00", "epoch": 0, "step": 11488, "train_loss": 3.4889121055603027, "perplexity": 32.75029945030763, "lr": 0.0026291804804649314, "grad_norm": 0.139312, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:08.289107+00:00", "epoch": 0, "step": 11489, "train_loss": 3.461568593978882, "perplexity": 31.86692359555882, "lr": 0.0026291804804649314, "grad_norm": 0.156824, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:08.592825+00:00", "epoch": 0, "step": 11490, "train_loss": 3.5764684677124023, "perplexity": 35.74707571856962, "lr": 0.0026291804804649314, "grad_norm": 0.139285, "tokens_per_sec": 107948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:08.896703+00:00", "epoch": 0, "step": 11491, "train_loss": 3.630455493927002, "perplexity": 37.72999848894307, "lr": 0.0026291804804649314, "grad_norm": 0.170155, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:09.201450+00:00", "epoch": 0, "step": 11492, "train_loss": 3.591658592224121, "perplexity": 36.29422334929179, "lr": 0.0026291804804649314, "grad_norm": 0.155161, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:09.506991+00:00", "epoch": 0, "step": 11493, "train_loss": 3.5107152462005615, "perplexity": 33.47220008811729, "lr": 0.0026291804804649314, "grad_norm": 0.155853, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:09.812955+00:00", "epoch": 0, "step": 11494, "train_loss": 3.560746908187866, "perplexity": 35.18947063568703, "lr": 0.0026291804804649314, "grad_norm": 0.170157, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:10.117176+00:00", "epoch": 0, "step": 11495, "train_loss": 3.5081615447998047, "perplexity": 33.38683113366882, "lr": 0.0026291804804649314, "grad_norm": 0.190949, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:10.421091+00:00", "epoch": 0, "step": 11496, "train_loss": 3.579174280166626, "perplexity": 35.84393157905424, "lr": 0.0026291804804649314, "grad_norm": 0.160833, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:10.725696+00:00", "epoch": 0, "step": 11497, "train_loss": 3.537848949432373, "perplexity": 34.39285878716081, "lr": 0.0026291804804649314, "grad_norm": 0.169126, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:11.031897+00:00", "epoch": 0, "step": 11498, "train_loss": 3.6746902465820312, "perplexity": 39.43643939320303, "lr": 0.0026291804804649314, "grad_norm": 0.202587, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:11.336890+00:00", "epoch": 0, "step": 11499, "train_loss": 3.5378470420837402, "perplexity": 34.392793188051186, "lr": 0.0026291804804649314, "grad_norm": 0.179436, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:11.642984+00:00", "epoch": 0, "step": 11500, "train_loss": 3.5378429889678955, "perplexity": 34.39265379035867, "lr": 0.0026291804804649314, "grad_norm": 0.197506, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T21:58:14.716154+00:00", "step": 11500, "epoch": 0, "val_loss": 3.4945504784584047, "val_ppl": 32.93547941739521, "eval_train_loss": 3.5378429889678955, "eval_train_ppl": 34.39265379035867} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T21:58:15.661170+00:00", "step": 11500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4946_epoch_0000_step_0011500.pt", "val_loss": 3.4945504784584047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:16.844019+00:00", "epoch": 0, "step": 11501, "train_loss": 3.4634811878204346, "perplexity": 31.92793039939859, "lr": 0.0026291804804649314, "grad_norm": 0.178597, "tokens_per_sec": 6300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:17.146995+00:00", "epoch": 0, "step": 11502, "train_loss": 3.5047693252563477, "perplexity": 33.27376754927339, "lr": 0.0026291804804649314, "grad_norm": 0.161962, "tokens_per_sec": 108149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:17.449488+00:00", "epoch": 0, "step": 11503, "train_loss": 3.4846596717834473, "perplexity": 32.611326666425704, "lr": 0.0026291804804649314, "grad_norm": 0.154991, "tokens_per_sec": 108326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:17.752790+00:00", "epoch": 0, "step": 11504, "train_loss": 3.4985342025756836, "perplexity": 33.06694697243282, "lr": 0.0026291804804649314, "grad_norm": 0.18292, "tokens_per_sec": 108041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:18.057024+00:00", "epoch": 0, "step": 11505, "train_loss": 3.5877792835235596, "perplexity": 36.15369959660984, "lr": 0.0026291804804649314, "grad_norm": 0.150549, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:18.361542+00:00", "epoch": 0, "step": 11506, "train_loss": 3.460692882537842, "perplexity": 31.83902958131174, "lr": 0.0026291804804649314, "grad_norm": 0.176508, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:18.666045+00:00", "epoch": 0, "step": 11507, "train_loss": 3.490525007247925, "perplexity": 32.803165085636756, "lr": 0.0026291804804649314, "grad_norm": 0.162154, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:18.969949+00:00", "epoch": 0, "step": 11508, "train_loss": 3.5694775581359863, "perplexity": 35.498042642702096, "lr": 0.0026291804804649314, "grad_norm": 0.159916, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:19.273919+00:00", "epoch": 0, "step": 11509, "train_loss": 3.514075994491577, "perplexity": 33.58488096733873, "lr": 0.0026291804804649314, "grad_norm": 0.150278, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:19.584294+00:00", "epoch": 0, "step": 11510, "train_loss": 3.52888822555542, "perplexity": 34.086050541134924, "lr": 0.0026291804804649314, "grad_norm": 0.158157, "tokens_per_sec": 105642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:19.889382+00:00", "epoch": 0, "step": 11511, "train_loss": 3.437743663787842, "perplexity": 31.11666923280288, "lr": 0.0026291804804649314, "grad_norm": 0.156823, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:20.210409+00:00", "epoch": 0, "step": 11512, "train_loss": 3.406911849975586, "perplexity": 30.171924810993012, "lr": 0.0026291804804649314, "grad_norm": 0.176135, "tokens_per_sec": 102072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:20.513712+00:00", "epoch": 0, "step": 11513, "train_loss": 3.510671854019165, "perplexity": 33.47074768785096, "lr": 0.0026291804804649314, "grad_norm": 0.149399, "tokens_per_sec": 108036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:20.817989+00:00", "epoch": 0, "step": 11514, "train_loss": 3.586337089538574, "perplexity": 36.10159652891292, "lr": 0.0026291804804649314, "grad_norm": 0.159799, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:21.121951+00:00", "epoch": 0, "step": 11515, "train_loss": 3.5823512077331543, "perplexity": 35.957986229133255, "lr": 0.0026291804804649314, "grad_norm": 0.148638, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:21.426145+00:00", "epoch": 0, "step": 11516, "train_loss": 3.5134575366973877, "perplexity": 33.56411655755493, "lr": 0.0026291804804649314, "grad_norm": 0.153414, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:21.730005+00:00", "epoch": 0, "step": 11517, "train_loss": 3.5380513668060303, "perplexity": 34.399821203942366, "lr": 0.0026291804804649314, "grad_norm": 0.1595, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:22.033078+00:00", "epoch": 0, "step": 11518, "train_loss": 3.5796456336975098, "perplexity": 35.86083072518801, "lr": 0.0026291804804649314, "grad_norm": 0.169133, "tokens_per_sec": 108120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:22.335540+00:00", "epoch": 0, "step": 11519, "train_loss": 3.5102319717407227, "perplexity": 33.456027736857, "lr": 0.0026291804804649314, "grad_norm": 0.146821, "tokens_per_sec": 108337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:22.639874+00:00", "epoch": 0, "step": 11520, "train_loss": 3.5948987007141113, "perplexity": 36.412011290309515, "lr": 0.0026291804804649314, "grad_norm": 0.168372, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:22.943890+00:00", "epoch": 0, "step": 11521, "train_loss": 3.528282880783081, "perplexity": 34.06542297265545, "lr": 0.0026291804804649314, "grad_norm": 0.168762, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:23.247953+00:00", "epoch": 0, "step": 11522, "train_loss": 3.470667600631714, "perplexity": 32.15820411742651, "lr": 0.0026291804804649314, "grad_norm": 0.174118, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:23.551328+00:00", "epoch": 0, "step": 11523, "train_loss": 3.6150240898132324, "perplexity": 37.15224090541111, "lr": 0.0026291804804649314, "grad_norm": 0.177244, "tokens_per_sec": 108003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:23.854223+00:00", "epoch": 0, "step": 11524, "train_loss": 3.5628316402435303, "perplexity": 35.262907774925075, "lr": 0.0026291804804649314, "grad_norm": 0.183596, "tokens_per_sec": 108183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:24.157361+00:00", "epoch": 0, "step": 11525, "train_loss": 3.539947986602783, "perplexity": 34.465126495921155, "lr": 0.0026291804804649314, "grad_norm": 0.163856, "tokens_per_sec": 108097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:24.461972+00:00", "epoch": 0, "step": 11526, "train_loss": 3.486506938934326, "perplexity": 32.67162417457964, "lr": 0.0026291804804649314, "grad_norm": 0.163642, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:24.765919+00:00", "epoch": 0, "step": 11527, "train_loss": 3.4612183570861816, "perplexity": 31.855764577522766, "lr": 0.0026291804804649314, "grad_norm": 0.189757, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:25.069721+00:00", "epoch": 0, "step": 11528, "train_loss": 3.477839469909668, "perplexity": 32.38966757427496, "lr": 0.0026291804804649314, "grad_norm": 0.168036, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:25.374087+00:00", "epoch": 0, "step": 11529, "train_loss": 3.5225088596343994, "perplexity": 33.869295267627614, "lr": 0.0026291804804649314, "grad_norm": 0.160659, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:25.677481+00:00", "epoch": 0, "step": 11530, "train_loss": 3.55883526802063, "perplexity": 35.1222652868443, "lr": 0.0026291804804649314, "grad_norm": 0.164661, "tokens_per_sec": 108097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:25.981105+00:00", "epoch": 0, "step": 11531, "train_loss": 3.497326374053955, "perplexity": 33.02703188089854, "lr": 0.0026291804804649314, "grad_norm": 0.152374, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:26.285367+00:00", "epoch": 0, "step": 11532, "train_loss": 3.4803805351257324, "perplexity": 32.47207649078051, "lr": 0.0026291804804649314, "grad_norm": 0.163469, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:26.589260+00:00", "epoch": 0, "step": 11533, "train_loss": 3.4567008018493652, "perplexity": 31.71217897357613, "lr": 0.0026291804804649314, "grad_norm": 0.169553, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:26.892337+00:00", "epoch": 0, "step": 11534, "train_loss": 3.5144622325897217, "perplexity": 33.597855233306575, "lr": 0.0026291804804649314, "grad_norm": 0.163906, "tokens_per_sec": 108112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:27.195828+00:00", "epoch": 0, "step": 11535, "train_loss": 3.5171027183532715, "perplexity": 33.68668711978919, "lr": 0.0026291804804649314, "grad_norm": 0.191064, "tokens_per_sec": 107971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:27.499844+00:00", "epoch": 0, "step": 11536, "train_loss": 3.5969481468200684, "perplexity": 36.486712266724346, "lr": 0.0026291804804649314, "grad_norm": 0.149184, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:27.803958+00:00", "epoch": 0, "step": 11537, "train_loss": 3.5263168811798096, "perplexity": 33.99851615554431, "lr": 0.0026291804804649314, "grad_norm": 0.152704, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:28.106760+00:00", "epoch": 0, "step": 11538, "train_loss": 3.5313329696655273, "perplexity": 34.16948415779834, "lr": 0.0026291804804649314, "grad_norm": 0.175495, "tokens_per_sec": 108216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:28.409548+00:00", "epoch": 0, "step": 11539, "train_loss": 3.538602352142334, "perplexity": 34.41878022358847, "lr": 0.0026291804804649314, "grad_norm": 0.168351, "tokens_per_sec": 108221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:28.712707+00:00", "epoch": 0, "step": 11540, "train_loss": 3.5777342319488525, "perplexity": 35.7923517369093, "lr": 0.0026291804804649314, "grad_norm": 0.168046, "tokens_per_sec": 108158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:29.016491+00:00", "epoch": 0, "step": 11541, "train_loss": 3.5805087089538574, "perplexity": 35.89179468104897, "lr": 0.0026291804804649314, "grad_norm": 0.162394, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:29.320290+00:00", "epoch": 0, "step": 11542, "train_loss": 3.492424726486206, "perplexity": 32.86554111914264, "lr": 0.0026291804804649314, "grad_norm": 0.151913, "tokens_per_sec": 107862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:29.623586+00:00", "epoch": 0, "step": 11543, "train_loss": 3.6570968627929688, "perplexity": 38.748686668259, "lr": 0.0026291804804649314, "grad_norm": 0.146775, "tokens_per_sec": 108039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:29.927368+00:00", "epoch": 0, "step": 11544, "train_loss": 3.60330867767334, "perplexity": 36.71952675279059, "lr": 0.0026291804804649314, "grad_norm": 0.173571, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:30.231403+00:00", "epoch": 0, "step": 11545, "train_loss": 3.624622106552124, "perplexity": 37.51054549157827, "lr": 0.0026291804804649314, "grad_norm": 0.156116, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:30.535343+00:00", "epoch": 0, "step": 11546, "train_loss": 3.633389472961426, "perplexity": 37.84086006672409, "lr": 0.0026291804804649314, "grad_norm": 0.162607, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:30.841671+00:00", "epoch": 0, "step": 11547, "train_loss": 3.4755120277404785, "perplexity": 32.31437015530106, "lr": 0.0026291804804649314, "grad_norm": 0.168052, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:31.146574+00:00", "epoch": 0, "step": 11548, "train_loss": 3.504535436630249, "perplexity": 33.26598610352746, "lr": 0.0026291804804649314, "grad_norm": 0.181006, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:31.450361+00:00", "epoch": 0, "step": 11549, "train_loss": 3.5396578311920166, "perplexity": 34.455127703657595, "lr": 0.0026291804804649314, "grad_norm": 0.174723, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:31.754394+00:00", "epoch": 0, "step": 11550, "train_loss": 3.4723315238952637, "perplexity": 32.211757443328054, "lr": 0.0026291804804649314, "grad_norm": 0.17762, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:32.059289+00:00", "epoch": 0, "step": 11551, "train_loss": 3.4893813133239746, "perplexity": 32.765669750722445, "lr": 0.0026291804804649314, "grad_norm": 0.16583, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:32.364532+00:00", "epoch": 0, "step": 11552, "train_loss": 3.4684243202209473, "perplexity": 32.08614510259124, "lr": 0.0026291804804649314, "grad_norm": 0.158765, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:32.670081+00:00", "epoch": 0, "step": 11553, "train_loss": 3.514578342437744, "perplexity": 33.601756501655046, "lr": 0.0026291804804649314, "grad_norm": 0.174901, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:32.973751+00:00", "epoch": 0, "step": 11554, "train_loss": 3.5331716537475586, "perplexity": 34.23236883922537, "lr": 0.0026291804804649314, "grad_norm": 0.164637, "tokens_per_sec": 107906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:33.277048+00:00", "epoch": 0, "step": 11555, "train_loss": 3.4872524738311768, "perplexity": 32.6959910926038, "lr": 0.0026291804804649314, "grad_norm": 0.159542, "tokens_per_sec": 108040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:33.580808+00:00", "epoch": 0, "step": 11556, "train_loss": 3.453956365585327, "perplexity": 31.62526623737362, "lr": 0.0026291804804649314, "grad_norm": 0.163103, "tokens_per_sec": 107875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:33.885508+00:00", "epoch": 0, "step": 11557, "train_loss": 3.490201473236084, "perplexity": 32.792553862673856, "lr": 0.0026291804804649314, "grad_norm": 0.169261, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:34.189530+00:00", "epoch": 0, "step": 11558, "train_loss": 3.549222707748413, "perplexity": 34.78626787975408, "lr": 0.0026291804804649314, "grad_norm": 0.152284, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:34.493178+00:00", "epoch": 0, "step": 11559, "train_loss": 3.556687831878662, "perplexity": 35.04692338988897, "lr": 0.0026291804804649314, "grad_norm": 0.154653, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:34.797096+00:00", "epoch": 0, "step": 11560, "train_loss": 3.6675825119018555, "perplexity": 39.15712945163577, "lr": 0.0026291804804649314, "grad_norm": 0.155908, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:35.100758+00:00", "epoch": 0, "step": 11561, "train_loss": 3.4214134216308594, "perplexity": 30.612653043063208, "lr": 0.0026291804804649314, "grad_norm": 0.155475, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:35.405018+00:00", "epoch": 0, "step": 11562, "train_loss": 3.5902957916259766, "perplexity": 36.244795247975134, "lr": 0.0026291804804649314, "grad_norm": 0.167815, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:35.710183+00:00", "epoch": 0, "step": 11563, "train_loss": 3.476346969604492, "perplexity": 32.341362042513204, "lr": 0.0026291804804649314, "grad_norm": 0.183017, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:36.014096+00:00", "epoch": 0, "step": 11564, "train_loss": 3.5023419857025146, "perplexity": 33.19309876207318, "lr": 0.0026291804804649314, "grad_norm": 0.174574, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:36.318028+00:00", "epoch": 0, "step": 11565, "train_loss": 3.5147571563720703, "perplexity": 33.607765501165794, "lr": 0.0026291804804649314, "grad_norm": 0.168321, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:36.621482+00:00", "epoch": 0, "step": 11566, "train_loss": 3.5080339908599854, "perplexity": 33.38257278341008, "lr": 0.0026291804804649314, "grad_norm": 0.184549, "tokens_per_sec": 107985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:36.925281+00:00", "epoch": 0, "step": 11567, "train_loss": 3.5755059719085693, "perplexity": 35.71268586089294, "lr": 0.0026291804804649314, "grad_norm": 0.185645, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:37.230462+00:00", "epoch": 0, "step": 11568, "train_loss": 3.640143871307373, "perplexity": 38.09731744199841, "lr": 0.0026291804804649314, "grad_norm": 0.187026, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:37.534595+00:00", "epoch": 0, "step": 11569, "train_loss": 3.5229148864746094, "perplexity": 33.88304990275182, "lr": 0.0026291804804649314, "grad_norm": 0.175851, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:37.838591+00:00", "epoch": 0, "step": 11570, "train_loss": 3.5622520446777344, "perplexity": 35.242475471749806, "lr": 0.0026291804804649314, "grad_norm": 0.193055, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:38.143151+00:00", "epoch": 0, "step": 11571, "train_loss": 3.5698702335357666, "perplexity": 35.51198458793855, "lr": 0.0026291804804649314, "grad_norm": 0.168554, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:38.447070+00:00", "epoch": 0, "step": 11572, "train_loss": 3.5223162174224854, "perplexity": 33.86277124009301, "lr": 0.0026291804804649314, "grad_norm": 0.187777, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:38.751137+00:00", "epoch": 0, "step": 11573, "train_loss": 3.4816737174987793, "perplexity": 32.51409597128453, "lr": 0.0026291804804649314, "grad_norm": 0.182527, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:39.054427+00:00", "epoch": 0, "step": 11574, "train_loss": 3.6019551753997803, "perplexity": 36.669860409183414, "lr": 0.0026291804804649314, "grad_norm": 0.166088, "tokens_per_sec": 108051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:39.358470+00:00", "epoch": 0, "step": 11575, "train_loss": 3.504436731338501, "perplexity": 33.26270273670933, "lr": 0.0026291804804649314, "grad_norm": 0.157015, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:39.662775+00:00", "epoch": 0, "step": 11576, "train_loss": 3.5240259170532227, "perplexity": 33.920715927468535, "lr": 0.0026291804804649314, "grad_norm": 0.165771, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:39.967208+00:00", "epoch": 0, "step": 11577, "train_loss": 3.627828598022461, "perplexity": 37.63101577599561, "lr": 0.0026291804804649314, "grad_norm": 0.173041, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:40.271197+00:00", "epoch": 0, "step": 11578, "train_loss": 3.5297675132751465, "perplexity": 34.11603516741469, "lr": 0.0026291804804649314, "grad_norm": 0.189553, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:40.575106+00:00", "epoch": 0, "step": 11579, "train_loss": 3.5955026149749756, "perplexity": 36.43400766448798, "lr": 0.0026291804804649314, "grad_norm": 0.190058, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:40.879106+00:00", "epoch": 0, "step": 11580, "train_loss": 3.608607053756714, "perplexity": 36.91459693638125, "lr": 0.0026291804804649314, "grad_norm": 0.156486, "tokens_per_sec": 107887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:41.182440+00:00", "epoch": 0, "step": 11581, "train_loss": 3.511514663696289, "perplexity": 33.4989690488504, "lr": 0.0026291804804649314, "grad_norm": 0.219915, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:41.486521+00:00", "epoch": 0, "step": 11582, "train_loss": 3.573901891708374, "perplexity": 35.65544576972971, "lr": 0.0026291804804649314, "grad_norm": 0.184313, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:41.792800+00:00", "epoch": 0, "step": 11583, "train_loss": 3.563208818435669, "perplexity": 35.27621068335494, "lr": 0.0026291804804649314, "grad_norm": 0.169873, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:42.096707+00:00", "epoch": 0, "step": 11584, "train_loss": 3.5411267280578613, "perplexity": 34.505775922151685, "lr": 0.0026291804804649314, "grad_norm": 0.225575, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:42.400163+00:00", "epoch": 0, "step": 11585, "train_loss": 3.521818161010742, "perplexity": 33.84590986906302, "lr": 0.0026291804804649314, "grad_norm": 0.180594, "tokens_per_sec": 107989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:42.702996+00:00", "epoch": 0, "step": 11586, "train_loss": 3.6425046920776367, "perplexity": 38.187364631121035, "lr": 0.0026291804804649314, "grad_norm": 0.181726, "tokens_per_sec": 108199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:43.006588+00:00", "epoch": 0, "step": 11587, "train_loss": 3.5600244998931885, "perplexity": 35.16405865023418, "lr": 0.0026291804804649314, "grad_norm": 0.179519, "tokens_per_sec": 107936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:43.311569+00:00", "epoch": 0, "step": 11588, "train_loss": 3.5336084365844727, "perplexity": 34.247324216289506, "lr": 0.0026291804804649314, "grad_norm": 0.162331, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:43.616702+00:00", "epoch": 0, "step": 11589, "train_loss": 3.414562463760376, "perplexity": 30.40364382094663, "lr": 0.0026291804804649314, "grad_norm": 0.172746, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:43.920904+00:00", "epoch": 0, "step": 11590, "train_loss": 3.588886022567749, "perplexity": 36.193734457530695, "lr": 0.0026291804804649314, "grad_norm": 0.140214, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:44.225256+00:00", "epoch": 0, "step": 11591, "train_loss": 3.4196746349334717, "perplexity": 30.559470419198906, "lr": 0.0026291804804649314, "grad_norm": 0.160789, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:44.529285+00:00", "epoch": 0, "step": 11592, "train_loss": 3.5812246799468994, "perplexity": 35.917501366451226, "lr": 0.0026291804804649314, "grad_norm": 0.162601, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:44.834952+00:00", "epoch": 0, "step": 11593, "train_loss": 3.5353481769561768, "perplexity": 34.30695752704954, "lr": 0.0026291804804649314, "grad_norm": 0.16666, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:45.140333+00:00", "epoch": 0, "step": 11594, "train_loss": 3.5858068466186523, "perplexity": 36.082458987176985, "lr": 0.0026291804804649314, "grad_norm": 0.16093, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:45.445048+00:00", "epoch": 0, "step": 11595, "train_loss": 3.4848501682281494, "perplexity": 32.617539599964765, "lr": 0.0026291804804649314, "grad_norm": 0.165218, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:45.748847+00:00", "epoch": 0, "step": 11596, "train_loss": 3.5526123046875, "perplexity": 34.9043793689672, "lr": 0.0026291804804649314, "grad_norm": 0.163827, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:46.053459+00:00", "epoch": 0, "step": 11597, "train_loss": 3.4631223678588867, "perplexity": 31.916476075783255, "lr": 0.0026291804804649314, "grad_norm": 0.159735, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:46.357545+00:00", "epoch": 0, "step": 11598, "train_loss": 3.4704341888427734, "perplexity": 32.15069888941277, "lr": 0.0026291804804649314, "grad_norm": 0.176682, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:46.663021+00:00", "epoch": 0, "step": 11599, "train_loss": 3.580397367477417, "perplexity": 35.88779865810277, "lr": 0.0026291804804649314, "grad_norm": 0.157821, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:46.967267+00:00", "epoch": 0, "step": 11600, "train_loss": 3.5831804275512695, "perplexity": 35.98781566980641, "lr": 0.0026291804804649314, "grad_norm": 0.160832, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:47.271416+00:00", "epoch": 0, "step": 11601, "train_loss": 3.486682176589966, "perplexity": 32.67734997507921, "lr": 0.0026291804804649314, "grad_norm": 0.161313, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:47.575458+00:00", "epoch": 0, "step": 11602, "train_loss": 3.4880995750427246, "perplexity": 32.72369964058645, "lr": 0.0026291804804649314, "grad_norm": 0.163778, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:47.879914+00:00", "epoch": 0, "step": 11603, "train_loss": 3.501185894012451, "perplexity": 33.154746669965995, "lr": 0.0026291804804649314, "grad_norm": 0.156326, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:48.184156+00:00", "epoch": 0, "step": 11604, "train_loss": 3.471092462539673, "perplexity": 32.171869816196235, "lr": 0.0026291804804649314, "grad_norm": 0.147092, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:48.488228+00:00", "epoch": 0, "step": 11605, "train_loss": 3.4324142932891846, "perplexity": 30.951278080596484, "lr": 0.0026291804804649314, "grad_norm": 0.164246, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:48.792254+00:00", "epoch": 0, "step": 11606, "train_loss": 3.495361566543579, "perplexity": 32.962203828790905, "lr": 0.0026291804804649314, "grad_norm": 0.159024, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:49.097145+00:00", "epoch": 0, "step": 11607, "train_loss": 3.4963622093200684, "perplexity": 32.99520372775108, "lr": 0.0026291804804649314, "grad_norm": 0.168357, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:49.401420+00:00", "epoch": 0, "step": 11608, "train_loss": 3.469754219055176, "perplexity": 32.12884481641454, "lr": 0.0026291804804649314, "grad_norm": 0.187617, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:49.705806+00:00", "epoch": 0, "step": 11609, "train_loss": 3.539372205734253, "perplexity": 34.44528784735856, "lr": 0.0026291804804649314, "grad_norm": 0.16237, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:50.009380+00:00", "epoch": 0, "step": 11610, "train_loss": 3.5317866802215576, "perplexity": 34.18499073093679, "lr": 0.0026291804804649314, "grad_norm": 0.181642, "tokens_per_sec": 107999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:50.313032+00:00", "epoch": 0, "step": 11611, "train_loss": 3.462329149246216, "perplexity": 31.891169371110568, "lr": 0.0026291804804649314, "grad_norm": 0.1505, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:50.617088+00:00", "epoch": 0, "step": 11612, "train_loss": 3.504378318786621, "perplexity": 33.26075983410556, "lr": 0.0026291804804649314, "grad_norm": 0.151842, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:50.921237+00:00", "epoch": 0, "step": 11613, "train_loss": 3.517703056335449, "perplexity": 33.706916589212796, "lr": 0.0026291804804649314, "grad_norm": 0.175107, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:51.225340+00:00", "epoch": 0, "step": 11614, "train_loss": 3.5510849952697754, "perplexity": 34.851110271189455, "lr": 0.0026291804804649314, "grad_norm": 0.163424, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:51.529111+00:00", "epoch": 0, "step": 11615, "train_loss": 3.592329263687134, "perplexity": 36.31857301356414, "lr": 0.0026291804804649314, "grad_norm": 0.147935, "tokens_per_sec": 107870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:51.834459+00:00", "epoch": 0, "step": 11616, "train_loss": 3.52547025680542, "perplexity": 33.96974436424242, "lr": 0.0026291804804649314, "grad_norm": 0.143379, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:52.138075+00:00", "epoch": 0, "step": 11617, "train_loss": 3.425737142562866, "perplexity": 30.745300169740585, "lr": 0.0026291804804649314, "grad_norm": 0.138193, "tokens_per_sec": 107925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:52.443294+00:00", "epoch": 0, "step": 11618, "train_loss": 3.644286870956421, "perplexity": 38.25548202646536, "lr": 0.0026291804804649314, "grad_norm": 0.16462, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:52.748492+00:00", "epoch": 0, "step": 11619, "train_loss": 3.6323275566101074, "perplexity": 37.800697567053, "lr": 0.0026291804804649314, "grad_norm": 0.172369, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:53.052979+00:00", "epoch": 0, "step": 11620, "train_loss": 3.544039487838745, "perplexity": 34.606429477115796, "lr": 0.0026291804804649314, "grad_norm": 0.155229, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:53.356930+00:00", "epoch": 0, "step": 11621, "train_loss": 3.4833920001983643, "perplexity": 32.570012406248296, "lr": 0.0026291804804649314, "grad_norm": 0.162493, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:53.661141+00:00", "epoch": 0, "step": 11622, "train_loss": 3.43397855758667, "perplexity": 30.99973194730615, "lr": 0.0026291804804649314, "grad_norm": 0.158289, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:53.965575+00:00", "epoch": 0, "step": 11623, "train_loss": 3.597074031829834, "perplexity": 36.49130568596961, "lr": 0.0026291804804649314, "grad_norm": 0.17758, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:54.268257+00:00", "epoch": 0, "step": 11624, "train_loss": 3.4894652366638184, "perplexity": 32.768419670549754, "lr": 0.0026291804804649314, "grad_norm": 0.178663, "tokens_per_sec": 108259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:54.571916+00:00", "epoch": 0, "step": 11625, "train_loss": 3.558285713195801, "perplexity": 35.102968979172125, "lr": 0.0026291804804649314, "grad_norm": 0.17466, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:54.875594+00:00", "epoch": 0, "step": 11626, "train_loss": 3.462289333343506, "perplexity": 31.889899620691878, "lr": 0.0026291804804649314, "grad_norm": 0.157317, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:55.180604+00:00", "epoch": 0, "step": 11627, "train_loss": 3.4698283672332764, "perplexity": 32.13122720004577, "lr": 0.0026291804804649314, "grad_norm": 0.183026, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:55.485610+00:00", "epoch": 0, "step": 11628, "train_loss": 3.571934223175049, "perplexity": 35.58535664973868, "lr": 0.0026291804804649314, "grad_norm": 0.163628, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:55.789412+00:00", "epoch": 0, "step": 11629, "train_loss": 3.459432601928711, "perplexity": 31.798928744178294, "lr": 0.0026291804804649314, "grad_norm": 0.160919, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:56.093235+00:00", "epoch": 0, "step": 11630, "train_loss": 3.546992778778076, "perplexity": 34.70878339779282, "lr": 0.0026291804804649314, "grad_norm": 0.187353, "tokens_per_sec": 107909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:56.397457+00:00", "epoch": 0, "step": 11631, "train_loss": 3.5376510620117188, "perplexity": 34.386053546404476, "lr": 0.0026291804804649314, "grad_norm": 0.165376, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:56.702489+00:00", "epoch": 0, "step": 11632, "train_loss": 3.547943353652954, "perplexity": 34.741792381498065, "lr": 0.0026291804804649314, "grad_norm": 0.197323, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:57.006088+00:00", "epoch": 0, "step": 11633, "train_loss": 3.5922040939331055, "perplexity": 36.31402731121153, "lr": 0.0026291804804649314, "grad_norm": 0.162983, "tokens_per_sec": 107932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:57.309771+00:00", "epoch": 0, "step": 11634, "train_loss": 3.6315572261810303, "perplexity": 37.7715897522339, "lr": 0.0026291804804649314, "grad_norm": 0.212932, "tokens_per_sec": 107901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:57.613657+00:00", "epoch": 0, "step": 11635, "train_loss": 3.6010563373565674, "perplexity": 36.63691495214149, "lr": 0.0026291804804649314, "grad_norm": 0.213492, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:57.917594+00:00", "epoch": 0, "step": 11636, "train_loss": 3.454824447631836, "perplexity": 31.65273148252871, "lr": 0.0026291804804649314, "grad_norm": 0.181588, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:58.222153+00:00", "epoch": 0, "step": 11637, "train_loss": 3.5255112648010254, "perplexity": 33.97113742393313, "lr": 0.0026291804804649314, "grad_norm": 0.18885, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:58.526632+00:00", "epoch": 0, "step": 11638, "train_loss": 3.498061418533325, "perplexity": 33.05131714263293, "lr": 0.0026291804804649314, "grad_norm": 0.18218, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:58.830099+00:00", "epoch": 0, "step": 11639, "train_loss": 3.535554885864258, "perplexity": 34.314049813773714, "lr": 0.0026291804804649314, "grad_norm": 0.179151, "tokens_per_sec": 107976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:59.133965+00:00", "epoch": 0, "step": 11640, "train_loss": 3.535404682159424, "perplexity": 34.30889610342671, "lr": 0.0026291804804649314, "grad_norm": 0.152992, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:59.438107+00:00", "epoch": 0, "step": 11641, "train_loss": 3.480184555053711, "perplexity": 32.46571323444724, "lr": 0.0026291804804649314, "grad_norm": 0.148155, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:58:59.743635+00:00", "epoch": 0, "step": 11642, "train_loss": 3.5608348846435547, "perplexity": 35.192566616775835, "lr": 0.0026291804804649314, "grad_norm": 0.149585, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:00.046279+00:00", "epoch": 0, "step": 11643, "train_loss": 3.57743763923645, "perplexity": 35.78173756034591, "lr": 0.0026291804804649314, "grad_norm": 0.162285, "tokens_per_sec": 108268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:00.349667+00:00", "epoch": 0, "step": 11644, "train_loss": 3.589393377304077, "perplexity": 36.21210217921442, "lr": 0.0026291804804649314, "grad_norm": 0.163367, "tokens_per_sec": 108007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:00.654300+00:00", "epoch": 0, "step": 11645, "train_loss": 3.564481496810913, "perplexity": 35.32113453459557, "lr": 0.0026291804804649314, "grad_norm": 0.179625, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:00.959154+00:00", "epoch": 0, "step": 11646, "train_loss": 3.480956554412842, "perplexity": 32.49078642125477, "lr": 0.0026291804804649314, "grad_norm": 0.14682, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:01.264237+00:00", "epoch": 0, "step": 11647, "train_loss": 3.3619372844696045, "perplexity": 28.84501779046937, "lr": 0.0026291804804649314, "grad_norm": 0.183692, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:01.568256+00:00", "epoch": 0, "step": 11648, "train_loss": 3.5151689052581787, "perplexity": 33.62160631044885, "lr": 0.0026291804804649314, "grad_norm": 0.206595, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:01.872241+00:00", "epoch": 0, "step": 11649, "train_loss": 3.5142011642456055, "perplexity": 33.58908504173447, "lr": 0.0026291804804649314, "grad_norm": 0.187445, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:02.176179+00:00", "epoch": 0, "step": 11650, "train_loss": 3.5584659576416016, "perplexity": 35.10929666460968, "lr": 0.0026291804804649314, "grad_norm": 0.15587, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:02.480781+00:00", "epoch": 0, "step": 11651, "train_loss": 3.4876866340637207, "perplexity": 32.710189473663, "lr": 0.0026291804804649314, "grad_norm": 0.178586, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:02.784966+00:00", "epoch": 0, "step": 11652, "train_loss": 3.597618818283081, "perplexity": 36.511191071121026, "lr": 0.0026291804804649314, "grad_norm": 0.139905, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:03.089908+00:00", "epoch": 0, "step": 11653, "train_loss": 3.463106632232666, "perplexity": 31.915973853996842, "lr": 0.0026291804804649314, "grad_norm": 0.162982, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:03.392709+00:00", "epoch": 0, "step": 11654, "train_loss": 3.414292097091675, "perplexity": 30.395424800175086, "lr": 0.0026291804804649314, "grad_norm": 0.156606, "tokens_per_sec": 108215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:03.696297+00:00", "epoch": 0, "step": 11655, "train_loss": 3.464872121810913, "perplexity": 31.972370942803018, "lr": 0.0026291804804649314, "grad_norm": 0.152086, "tokens_per_sec": 107936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:04.000295+00:00", "epoch": 0, "step": 11656, "train_loss": 3.5144271850585938, "perplexity": 33.59667773206383, "lr": 0.0026291804804649314, "grad_norm": 0.157431, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:04.304681+00:00", "epoch": 0, "step": 11657, "train_loss": 3.5589439868927, "perplexity": 35.126083947487324, "lr": 0.0026291804804649314, "grad_norm": 0.15453, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:04.609236+00:00", "epoch": 0, "step": 11658, "train_loss": 3.5097172260284424, "perplexity": 33.43881082157634, "lr": 0.0026291804804649314, "grad_norm": 0.149286, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:04.913635+00:00", "epoch": 0, "step": 11659, "train_loss": 3.6165928840637207, "perplexity": 37.210570869229144, "lr": 0.0026291804804649314, "grad_norm": 0.176236, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:05.218599+00:00", "epoch": 0, "step": 11660, "train_loss": 3.448957681655884, "perplexity": 31.467575977500736, "lr": 0.0026291804804649314, "grad_norm": 0.178524, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:05.523114+00:00", "epoch": 0, "step": 11661, "train_loss": 3.571455478668213, "perplexity": 35.568324433084214, "lr": 0.0026291804804649314, "grad_norm": 0.156251, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:05.827620+00:00", "epoch": 0, "step": 11662, "train_loss": 3.4784035682678223, "perplexity": 32.40794368685269, "lr": 0.0026291804804649314, "grad_norm": 0.185224, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:06.131938+00:00", "epoch": 0, "step": 11663, "train_loss": 3.4687986373901367, "perplexity": 32.09815774572527, "lr": 0.0026291804804649314, "grad_norm": 0.204082, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:06.435721+00:00", "epoch": 0, "step": 11664, "train_loss": 3.4540603160858154, "perplexity": 31.628553870499662, "lr": 0.0026291804804649314, "grad_norm": 0.161551, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:06.739867+00:00", "epoch": 0, "step": 11665, "train_loss": 3.551530122756958, "perplexity": 34.86662691151535, "lr": 0.0026291804804649314, "grad_norm": 0.169627, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:07.044940+00:00", "epoch": 0, "step": 11666, "train_loss": 3.6181130409240723, "perplexity": 37.26717979013198, "lr": 0.0026291804804649314, "grad_norm": 0.200263, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:07.348913+00:00", "epoch": 0, "step": 11667, "train_loss": 3.591488838195801, "perplexity": 36.28806278157888, "lr": 0.0026291804804649314, "grad_norm": 0.16924, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:07.653648+00:00", "epoch": 0, "step": 11668, "train_loss": 3.5171473026275635, "perplexity": 33.68818904976871, "lr": 0.0026291804804649314, "grad_norm": 0.171235, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:07.957404+00:00", "epoch": 0, "step": 11669, "train_loss": 3.4808030128479004, "perplexity": 32.48579811802721, "lr": 0.0026291804804649314, "grad_norm": 0.181884, "tokens_per_sec": 107874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:08.261071+00:00", "epoch": 0, "step": 11670, "train_loss": 3.4947690963745117, "perplexity": 32.942680490384525, "lr": 0.0026291804804649314, "grad_norm": 0.163096, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:08.564053+00:00", "epoch": 0, "step": 11671, "train_loss": 3.622091293334961, "perplexity": 37.415733333802315, "lr": 0.0026291804804649314, "grad_norm": 0.154635, "tokens_per_sec": 108097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:08.867633+00:00", "epoch": 0, "step": 11672, "train_loss": 3.528580904006958, "perplexity": 34.075576772791464, "lr": 0.0026291804804649314, "grad_norm": 0.162226, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:09.171651+00:00", "epoch": 0, "step": 11673, "train_loss": 3.524378776550293, "perplexity": 33.932687286210566, "lr": 0.0026291804804649314, "grad_norm": 0.155603, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:09.475799+00:00", "epoch": 0, "step": 11674, "train_loss": 3.456912040710449, "perplexity": 31.71887852572297, "lr": 0.0026291804804649314, "grad_norm": 0.154513, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:09.778920+00:00", "epoch": 0, "step": 11675, "train_loss": 3.5414507389068604, "perplexity": 34.51695797935969, "lr": 0.0026291804804649314, "grad_norm": 0.202516, "tokens_per_sec": 108102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:10.083239+00:00", "epoch": 0, "step": 11676, "train_loss": 3.4891927242279053, "perplexity": 32.75949108531482, "lr": 0.0026291804804649314, "grad_norm": 0.201272, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:10.387771+00:00", "epoch": 0, "step": 11677, "train_loss": 3.5564775466918945, "perplexity": 35.039554315889, "lr": 0.0026291804804649314, "grad_norm": 0.167269, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:10.692660+00:00", "epoch": 0, "step": 11678, "train_loss": 3.5093133449554443, "perplexity": 33.42530824568291, "lr": 0.0026291804804649314, "grad_norm": 0.163786, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:10.997798+00:00", "epoch": 0, "step": 11679, "train_loss": 3.5990211963653564, "perplexity": 36.56242948465359, "lr": 0.0026291804804649314, "grad_norm": 0.180287, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:11.301882+00:00", "epoch": 0, "step": 11680, "train_loss": 3.4365923404693604, "perplexity": 31.0808645012879, "lr": 0.0026291804804649314, "grad_norm": 0.167735, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:11.605553+00:00", "epoch": 0, "step": 11681, "train_loss": 3.5295960903167725, "perplexity": 34.110187396973664, "lr": 0.0026291804804649314, "grad_norm": 0.157608, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:11.908652+00:00", "epoch": 0, "step": 11682, "train_loss": 3.5941498279571533, "perplexity": 36.38475353459736, "lr": 0.0026291804804649314, "grad_norm": 0.145486, "tokens_per_sec": 108111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:12.213782+00:00", "epoch": 0, "step": 11683, "train_loss": 3.465752363204956, "perplexity": 32.000526737302486, "lr": 0.0026291804804649314, "grad_norm": 0.171946, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:12.519278+00:00", "epoch": 0, "step": 11684, "train_loss": 3.493288516998291, "perplexity": 32.89394232631662, "lr": 0.0026291804804649314, "grad_norm": 0.157743, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:12.822932+00:00", "epoch": 0, "step": 11685, "train_loss": 3.5052952766418457, "perplexity": 33.291272536399255, "lr": 0.0026291804804649314, "grad_norm": 0.179903, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:13.126945+00:00", "epoch": 0, "step": 11686, "train_loss": 3.585761785507202, "perplexity": 36.0808331081034, "lr": 0.0026291804804649314, "grad_norm": 0.14859, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:13.431096+00:00", "epoch": 0, "step": 11687, "train_loss": 3.536548376083374, "perplexity": 34.34815742662643, "lr": 0.0026291804804649314, "grad_norm": 0.17051, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:13.736077+00:00", "epoch": 0, "step": 11688, "train_loss": 3.5608773231506348, "perplexity": 35.194060168455195, "lr": 0.0026291804804649314, "grad_norm": 0.173466, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:14.041827+00:00", "epoch": 0, "step": 11689, "train_loss": 3.512223243713379, "perplexity": 33.522714160573855, "lr": 0.0026291804804649314, "grad_norm": 0.162437, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:14.345645+00:00", "epoch": 0, "step": 11690, "train_loss": 3.600700616836548, "perplexity": 36.62388476739225, "lr": 0.0026291804804649314, "grad_norm": 0.181528, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:14.648812+00:00", "epoch": 0, "step": 11691, "train_loss": 3.5829060077667236, "perplexity": 35.97794125611323, "lr": 0.0026291804804649314, "grad_norm": 0.162465, "tokens_per_sec": 108020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:14.953580+00:00", "epoch": 0, "step": 11692, "train_loss": 3.3981003761291504, "perplexity": 29.90723355725008, "lr": 0.0026291804804649314, "grad_norm": 0.170394, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:15.258567+00:00", "epoch": 0, "step": 11693, "train_loss": 3.562203884124756, "perplexity": 35.24077821551348, "lr": 0.0026291804804649314, "grad_norm": 0.165716, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:15.563341+00:00", "epoch": 0, "step": 11694, "train_loss": 3.5553953647613525, "perplexity": 35.00165565367647, "lr": 0.0026291804804649314, "grad_norm": 0.171316, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:15.868621+00:00", "epoch": 0, "step": 11695, "train_loss": 3.539130687713623, "perplexity": 34.43696969415002, "lr": 0.0026291804804649314, "grad_norm": 0.191772, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:16.171660+00:00", "epoch": 0, "step": 11696, "train_loss": 3.4428298473358154, "perplexity": 31.27533748980926, "lr": 0.0026291804804649314, "grad_norm": 0.157939, "tokens_per_sec": 108127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:16.474673+00:00", "epoch": 0, "step": 11697, "train_loss": 3.4075284004211426, "perplexity": 30.190533060551342, "lr": 0.0026291804804649314, "grad_norm": 0.173374, "tokens_per_sec": 108141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:16.777893+00:00", "epoch": 0, "step": 11698, "train_loss": 3.478529691696167, "perplexity": 32.412031345585426, "lr": 0.0026291804804649314, "grad_norm": 0.148978, "tokens_per_sec": 108066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:17.083176+00:00", "epoch": 0, "step": 11699, "train_loss": 3.517814874649048, "perplexity": 33.71068585051471, "lr": 0.0026291804804649314, "grad_norm": 0.160154, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:17.388714+00:00", "epoch": 0, "step": 11700, "train_loss": 3.499851703643799, "perplexity": 33.110541421949236, "lr": 0.0026291804804649314, "grad_norm": 0.161809, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:17.692109+00:00", "epoch": 0, "step": 11701, "train_loss": 3.505070209503174, "perplexity": 33.28378060807181, "lr": 0.0026291804804649314, "grad_norm": 0.173939, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:17.995377+00:00", "epoch": 0, "step": 11702, "train_loss": 3.607409954071045, "perplexity": 36.870432923629096, "lr": 0.0026291804804649314, "grad_norm": 0.187355, "tokens_per_sec": 108049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:18.299217+00:00", "epoch": 0, "step": 11703, "train_loss": 3.4318156242370605, "perplexity": 30.93275405372001, "lr": 0.0026291804804649314, "grad_norm": 0.167336, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:18.604540+00:00", "epoch": 0, "step": 11704, "train_loss": 3.5806736946105957, "perplexity": 35.89771680088492, "lr": 0.0026291804804649314, "grad_norm": 0.186451, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:18.909456+00:00", "epoch": 0, "step": 11705, "train_loss": 3.579514503479004, "perplexity": 35.856128594921636, "lr": 0.0026291804804649314, "grad_norm": 0.207976, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:19.214442+00:00", "epoch": 0, "step": 11706, "train_loss": 3.5982601642608643, "perplexity": 36.53461488724079, "lr": 0.0026291804804649314, "grad_norm": 0.164167, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:19.518156+00:00", "epoch": 0, "step": 11707, "train_loss": 3.520752191543579, "perplexity": 33.80985038512319, "lr": 0.0026291804804649314, "grad_norm": 0.167483, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:19.821664+00:00", "epoch": 0, "step": 11708, "train_loss": 3.673563003540039, "perplexity": 39.39200998737001, "lr": 0.0026291804804649314, "grad_norm": 0.164303, "tokens_per_sec": 107964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:20.126313+00:00", "epoch": 0, "step": 11709, "train_loss": 3.5951874256134033, "perplexity": 36.422525862438675, "lr": 0.0026291804804649314, "grad_norm": 0.149821, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:20.431465+00:00", "epoch": 0, "step": 11710, "train_loss": 3.5964810848236084, "perplexity": 36.46967468916101, "lr": 0.0026291804804649314, "grad_norm": 0.173659, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:20.736012+00:00", "epoch": 0, "step": 11711, "train_loss": 3.5773098468780518, "perplexity": 35.777165219876785, "lr": 0.0026291804804649314, "grad_norm": 0.162753, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:21.040361+00:00", "epoch": 0, "step": 11712, "train_loss": 3.5995845794677734, "perplexity": 36.583033943165454, "lr": 0.0026291804804649314, "grad_norm": 0.183796, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:21.343779+00:00", "epoch": 0, "step": 11713, "train_loss": 3.617804527282715, "perplexity": 37.25568413016695, "lr": 0.0026291804804649314, "grad_norm": 0.158365, "tokens_per_sec": 107995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:21.647564+00:00", "epoch": 0, "step": 11714, "train_loss": 3.5742721557617188, "perplexity": 35.66865014400594, "lr": 0.0026291804804649314, "grad_norm": 0.168347, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:21.952719+00:00", "epoch": 0, "step": 11715, "train_loss": 3.4954440593719482, "perplexity": 32.96492308637211, "lr": 0.0026291804804649314, "grad_norm": 0.184162, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:22.257187+00:00", "epoch": 0, "step": 11716, "train_loss": 3.6477417945861816, "perplexity": 38.38788037668315, "lr": 0.0026291804804649314, "grad_norm": 0.17937, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:22.561744+00:00", "epoch": 0, "step": 11717, "train_loss": 3.4914944171905518, "perplexity": 32.834980218481924, "lr": 0.0026291804804649314, "grad_norm": 0.17339, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:22.867410+00:00", "epoch": 0, "step": 11718, "train_loss": 3.5624544620513916, "perplexity": 35.24960988311602, "lr": 0.0026291804804649314, "grad_norm": 0.187813, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:23.172161+00:00", "epoch": 0, "step": 11719, "train_loss": 3.4663498401641846, "perplexity": 32.019652027602405, "lr": 0.0026291804804649314, "grad_norm": 0.167717, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:23.477548+00:00", "epoch": 0, "step": 11720, "train_loss": 3.5081348419189453, "perplexity": 33.38593962099782, "lr": 0.0026291804804649314, "grad_norm": 0.144801, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:23.782281+00:00", "epoch": 0, "step": 11721, "train_loss": 3.4645984172821045, "perplexity": 31.96362115756177, "lr": 0.0026291804804649314, "grad_norm": 0.169972, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:24.086019+00:00", "epoch": 0, "step": 11722, "train_loss": 3.6260178089141846, "perplexity": 37.56293560051785, "lr": 0.0026291804804649314, "grad_norm": 0.170599, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:24.390779+00:00", "epoch": 0, "step": 11723, "train_loss": 3.5151798725128174, "perplexity": 33.621975049188634, "lr": 0.0026291804804649314, "grad_norm": 0.164834, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:24.694383+00:00", "epoch": 0, "step": 11724, "train_loss": 3.603482961654663, "perplexity": 36.72592693581386, "lr": 0.0026291804804649314, "grad_norm": 0.191155, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:24.999308+00:00", "epoch": 0, "step": 11725, "train_loss": 3.618136167526245, "perplexity": 37.268041663339154, "lr": 0.0026291804804649314, "grad_norm": 0.151351, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:25.303624+00:00", "epoch": 0, "step": 11726, "train_loss": 3.50569748878479, "perplexity": 33.304665383669814, "lr": 0.0026291804804649314, "grad_norm": 0.18884, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:25.608715+00:00", "epoch": 0, "step": 11727, "train_loss": 3.6258151531219482, "perplexity": 37.55532402533595, "lr": 0.0026291804804649314, "grad_norm": 0.179144, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:25.912562+00:00", "epoch": 0, "step": 11728, "train_loss": 3.564418315887451, "perplexity": 35.318902983194434, "lr": 0.0026291804804649314, "grad_norm": 0.154161, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:26.215984+00:00", "epoch": 0, "step": 11729, "train_loss": 3.5119099617004395, "perplexity": 33.51221374208444, "lr": 0.0026291804804649314, "grad_norm": 0.16603, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:26.520360+00:00", "epoch": 0, "step": 11730, "train_loss": 3.5707459449768066, "perplexity": 35.54309645965902, "lr": 0.0026291804804649314, "grad_norm": 0.146058, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:26.825568+00:00", "epoch": 0, "step": 11731, "train_loss": 3.479907512664795, "perplexity": 32.456720101491946, "lr": 0.0026291804804649314, "grad_norm": 0.162597, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:27.130795+00:00", "epoch": 0, "step": 11732, "train_loss": 3.579133987426758, "perplexity": 35.842487357939284, "lr": 0.0026291804804649314, "grad_norm": 0.146694, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:27.434811+00:00", "epoch": 0, "step": 11733, "train_loss": 3.4642746448516846, "perplexity": 31.953273893424388, "lr": 0.0026291804804649314, "grad_norm": 0.215318, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:27.738716+00:00", "epoch": 0, "step": 11734, "train_loss": 3.504380226135254, "perplexity": 33.260823274030855, "lr": 0.0026291804804649314, "grad_norm": 0.22649, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:28.043070+00:00", "epoch": 0, "step": 11735, "train_loss": 3.529512405395508, "perplexity": 34.10733300806335, "lr": 0.0026291804804649314, "grad_norm": 0.22519, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:28.347414+00:00", "epoch": 0, "step": 11736, "train_loss": 3.520599365234375, "perplexity": 33.80468374528339, "lr": 0.0026291804804649314, "grad_norm": 0.164253, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:28.652630+00:00", "epoch": 0, "step": 11737, "train_loss": 3.544564723968506, "perplexity": 34.62461079852466, "lr": 0.0026291804804649314, "grad_norm": 0.18035, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:28.955801+00:00", "epoch": 0, "step": 11738, "train_loss": 3.466826915740967, "perplexity": 32.03493146599548, "lr": 0.0026291804804649314, "grad_norm": 0.219883, "tokens_per_sec": 108084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:29.259697+00:00", "epoch": 0, "step": 11739, "train_loss": 3.6076819896698, "perplexity": 36.88046435831764, "lr": 0.0026291804804649314, "grad_norm": 0.162439, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:29.564279+00:00", "epoch": 0, "step": 11740, "train_loss": 3.528416395187378, "perplexity": 34.06997150095103, "lr": 0.0026291804804649314, "grad_norm": 0.176456, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:29.867940+00:00", "epoch": 0, "step": 11741, "train_loss": 3.5637738704681396, "perplexity": 35.296149210521584, "lr": 0.0026291804804649314, "grad_norm": 0.148348, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:30.172618+00:00", "epoch": 0, "step": 11742, "train_loss": 3.605483293533325, "perplexity": 36.799464503486774, "lr": 0.0026291804804649314, "grad_norm": 0.171019, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:30.477145+00:00", "epoch": 0, "step": 11743, "train_loss": 3.4781758785247803, "perplexity": 32.4005655704747, "lr": 0.0026291804804649314, "grad_norm": 0.165728, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:30.780797+00:00", "epoch": 0, "step": 11744, "train_loss": 3.4840829372406006, "perplexity": 32.592524010440904, "lr": 0.0026291804804649314, "grad_norm": 0.148234, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:31.085293+00:00", "epoch": 0, "step": 11745, "train_loss": 3.4772095680236816, "perplexity": 32.3692716859621, "lr": 0.0026291804804649314, "grad_norm": 0.155369, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:31.389052+00:00", "epoch": 0, "step": 11746, "train_loss": 3.531158924102783, "perplexity": 34.16353762819797, "lr": 0.0026291804804649314, "grad_norm": 0.161654, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:31.694117+00:00", "epoch": 0, "step": 11747, "train_loss": 3.40694522857666, "perplexity": 30.172931924442846, "lr": 0.0026291804804649314, "grad_norm": 0.160268, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:31.998288+00:00", "epoch": 0, "step": 11748, "train_loss": 3.559920072555542, "perplexity": 35.16038675293506, "lr": 0.0026291804804649314, "grad_norm": 0.161805, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:32.301627+00:00", "epoch": 0, "step": 11749, "train_loss": 3.481966495513916, "perplexity": 32.523616777440324, "lr": 0.0026291804804649314, "grad_norm": 0.146091, "tokens_per_sec": 108024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:32.604836+00:00", "epoch": 0, "step": 11750, "train_loss": 3.4972009658813477, "perplexity": 33.02289028088438, "lr": 0.0026291804804649314, "grad_norm": 0.159134, "tokens_per_sec": 108137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:32.909261+00:00", "epoch": 0, "step": 11751, "train_loss": 3.5926504135131836, "perplexity": 36.3302385900685, "lr": 0.0026291804804649314, "grad_norm": 0.141862, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:33.213102+00:00", "epoch": 0, "step": 11752, "train_loss": 3.505837917327881, "perplexity": 33.30934263771007, "lr": 0.0026291804804649314, "grad_norm": 0.164369, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:33.517165+00:00", "epoch": 0, "step": 11753, "train_loss": 3.496617317199707, "perplexity": 33.00362213796805, "lr": 0.0026291804804649314, "grad_norm": 0.153825, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:33.820869+00:00", "epoch": 0, "step": 11754, "train_loss": 3.585038900375366, "perplexity": 36.05476023528674, "lr": 0.0026291804804649314, "grad_norm": 0.147492, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:34.124395+00:00", "epoch": 0, "step": 11755, "train_loss": 3.5545685291290283, "perplexity": 34.97272699886107, "lr": 0.0026291804804649314, "grad_norm": 0.140004, "tokens_per_sec": 107960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:34.428278+00:00", "epoch": 0, "step": 11756, "train_loss": 3.603192090988159, "perplexity": 36.715245994429644, "lr": 0.0026291804804649314, "grad_norm": 0.1566, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:34.732312+00:00", "epoch": 0, "step": 11757, "train_loss": 3.43854022026062, "perplexity": 31.141465291511008, "lr": 0.0026291804804649314, "grad_norm": 0.151711, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:35.037681+00:00", "epoch": 0, "step": 11758, "train_loss": 3.5619540214538574, "perplexity": 35.23197396051715, "lr": 0.0026291804804649314, "grad_norm": 0.133808, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:35.341632+00:00", "epoch": 0, "step": 11759, "train_loss": 3.5567164421081543, "perplexity": 35.047926104754026, "lr": 0.0026291804804649314, "grad_norm": 0.163864, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:35.645517+00:00", "epoch": 0, "step": 11760, "train_loss": 3.4946279525756836, "perplexity": 32.93803116343609, "lr": 0.0026291804804649314, "grad_norm": 0.159649, "tokens_per_sec": 107926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:35.949928+00:00", "epoch": 0, "step": 11761, "train_loss": 3.4963531494140625, "perplexity": 32.994904795660815, "lr": 0.0026291804804649314, "grad_norm": 0.172147, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:36.254391+00:00", "epoch": 0, "step": 11762, "train_loss": 3.592115640640259, "perplexity": 36.31081535797534, "lr": 0.0026291804804649314, "grad_norm": 0.156337, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:36.559776+00:00", "epoch": 0, "step": 11763, "train_loss": 3.5596187114715576, "perplexity": 35.149792377116356, "lr": 0.0026291804804649314, "grad_norm": 0.174457, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:36.864258+00:00", "epoch": 0, "step": 11764, "train_loss": 3.375779151916504, "perplexity": 29.24706281096149, "lr": 0.0026291804804649314, "grad_norm": 0.181668, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:37.167335+00:00", "epoch": 0, "step": 11765, "train_loss": 3.510270357131958, "perplexity": 33.457311984218876, "lr": 0.0026291804804649314, "grad_norm": 0.149959, "tokens_per_sec": 108117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:37.471551+00:00", "epoch": 0, "step": 11766, "train_loss": 3.497319221496582, "perplexity": 33.02679565400296, "lr": 0.0026291804804649314, "grad_norm": 0.162085, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:37.776102+00:00", "epoch": 0, "step": 11767, "train_loss": 3.485032081604004, "perplexity": 32.62347370643577, "lr": 0.0026291804804649314, "grad_norm": 0.167077, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:38.081747+00:00", "epoch": 0, "step": 11768, "train_loss": 3.6306076049804688, "perplexity": 37.735738075276714, "lr": 0.0026291804804649314, "grad_norm": 0.192841, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:38.386463+00:00", "epoch": 0, "step": 11769, "train_loss": 3.470339775085449, "perplexity": 32.1476635644205, "lr": 0.0026291804804649314, "grad_norm": 0.205404, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:38.691069+00:00", "epoch": 0, "step": 11770, "train_loss": 3.6212759017944336, "perplexity": 37.38523729615451, "lr": 0.0026291804804649314, "grad_norm": 0.195788, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:38.995535+00:00", "epoch": 0, "step": 11771, "train_loss": 3.4843087196350098, "perplexity": 32.59988365936013, "lr": 0.0026291804804649314, "grad_norm": 0.179402, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:39.299503+00:00", "epoch": 0, "step": 11772, "train_loss": 3.5077590942382812, "perplexity": 33.37339728814224, "lr": 0.0026291804804649314, "grad_norm": 0.189339, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:39.602694+00:00", "epoch": 0, "step": 11773, "train_loss": 3.531989812850952, "perplexity": 34.19193552332288, "lr": 0.0026291804804649314, "grad_norm": 0.214111, "tokens_per_sec": 108077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:39.907211+00:00", "epoch": 0, "step": 11774, "train_loss": 3.5140860080718994, "perplexity": 33.58521727392572, "lr": 0.0026291804804649314, "grad_norm": 0.177547, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:40.212120+00:00", "epoch": 0, "step": 11775, "train_loss": 3.478317975997925, "perplexity": 32.40516993609733, "lr": 0.0026291804804649314, "grad_norm": 0.165486, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:40.516985+00:00", "epoch": 0, "step": 11776, "train_loss": 3.436584711074829, "perplexity": 31.080627374014817, "lr": 0.0026291804804649314, "grad_norm": 0.197251, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:40.820800+00:00", "epoch": 0, "step": 11777, "train_loss": 3.5911781787872314, "perplexity": 36.2767913043429, "lr": 0.0026291804804649314, "grad_norm": 0.180462, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:41.125654+00:00", "epoch": 0, "step": 11778, "train_loss": 3.51251482963562, "perplexity": 33.532490337326884, "lr": 0.0026291804804649314, "grad_norm": 0.159293, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:41.431255+00:00", "epoch": 0, "step": 11779, "train_loss": 3.532932996749878, "perplexity": 34.224200019666455, "lr": 0.0026291804804649314, "grad_norm": 0.18914, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:41.736405+00:00", "epoch": 0, "step": 11780, "train_loss": 3.646780490875244, "perplexity": 38.3509956963525, "lr": 0.0026291804804649314, "grad_norm": 0.169573, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:42.040674+00:00", "epoch": 0, "step": 11781, "train_loss": 3.492063283920288, "perplexity": 32.85366426016075, "lr": 0.0026291804804649314, "grad_norm": 0.164738, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:42.344004+00:00", "epoch": 0, "step": 11782, "train_loss": 3.527188777923584, "perplexity": 34.02817227773295, "lr": 0.0026291804804649314, "grad_norm": 0.174277, "tokens_per_sec": 108028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:42.647599+00:00", "epoch": 0, "step": 11783, "train_loss": 3.5236425399780273, "perplexity": 33.90771399508838, "lr": 0.0026291804804649314, "grad_norm": 0.169108, "tokens_per_sec": 107934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:42.951857+00:00", "epoch": 0, "step": 11784, "train_loss": 3.4995298385620117, "perplexity": 33.099886009721, "lr": 0.0026291804804649314, "grad_norm": 0.179573, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:43.256205+00:00", "epoch": 0, "step": 11785, "train_loss": 3.402250289916992, "perplexity": 30.031603882719892, "lr": 0.0026291804804649314, "grad_norm": 0.176683, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:43.560774+00:00", "epoch": 0, "step": 11786, "train_loss": 3.5085482597351074, "perplexity": 33.39974481670675, "lr": 0.0026291804804649314, "grad_norm": 0.177564, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:43.865406+00:00", "epoch": 0, "step": 11787, "train_loss": 3.5278372764587402, "perplexity": 34.050246654435966, "lr": 0.0026291804804649314, "grad_norm": 0.164306, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:44.168716+00:00", "epoch": 0, "step": 11788, "train_loss": 3.5416290760040283, "perplexity": 34.52311418237217, "lr": 0.0026291804804649314, "grad_norm": 0.185794, "tokens_per_sec": 108032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:44.472459+00:00", "epoch": 0, "step": 11789, "train_loss": 3.4985105991363525, "perplexity": 33.06616648796719, "lr": 0.0026291804804649314, "grad_norm": 0.177885, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:44.776536+00:00", "epoch": 0, "step": 11790, "train_loss": 3.449591875076294, "perplexity": 31.48753883662979, "lr": 0.0026291804804649314, "grad_norm": 0.145184, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:45.081440+00:00", "epoch": 0, "step": 11791, "train_loss": 3.4841253757476807, "perplexity": 32.5939072178523, "lr": 0.0026291804804649314, "grad_norm": 0.185518, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:45.384851+00:00", "epoch": 0, "step": 11792, "train_loss": 3.428189277648926, "perplexity": 30.820784309848808, "lr": 0.0026291804804649314, "grad_norm": 0.163799, "tokens_per_sec": 107999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:45.688794+00:00", "epoch": 0, "step": 11793, "train_loss": 3.526798963546753, "perplexity": 34.01491019200559, "lr": 0.0026291804804649314, "grad_norm": 0.147705, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:45.992746+00:00", "epoch": 0, "step": 11794, "train_loss": 3.550593137741089, "perplexity": 34.83397270518542, "lr": 0.0026291804804649314, "grad_norm": 0.150099, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:46.296992+00:00", "epoch": 0, "step": 11795, "train_loss": 3.451104164123535, "perplexity": 31.535193121181113, "lr": 0.0026291804804649314, "grad_norm": 0.150274, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:46.601508+00:00", "epoch": 0, "step": 11796, "train_loss": 3.4501185417175293, "perplexity": 31.504126640687744, "lr": 0.0026291804804649314, "grad_norm": 0.149722, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:46.905899+00:00", "epoch": 0, "step": 11797, "train_loss": 3.550035238265991, "perplexity": 34.81454427015841, "lr": 0.0026291804804649314, "grad_norm": 0.161933, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:47.210199+00:00", "epoch": 0, "step": 11798, "train_loss": 3.489368438720703, "perplexity": 32.76524790843901, "lr": 0.0026291804804649314, "grad_norm": 0.165947, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:47.514517+00:00", "epoch": 0, "step": 11799, "train_loss": 3.5128257274627686, "perplexity": 33.54291713646225, "lr": 0.0026291804804649314, "grad_norm": 0.185132, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:47.821038+00:00", "epoch": 0, "step": 11800, "train_loss": 3.5155375003814697, "perplexity": 33.634001354808206, "lr": 0.0026291804804649314, "grad_norm": 0.155038, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:48.125441+00:00", "epoch": 0, "step": 11801, "train_loss": 3.5609207153320312, "perplexity": 35.195587348631705, "lr": 0.0026291804804649314, "grad_norm": 0.184368, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:48.429495+00:00", "epoch": 0, "step": 11802, "train_loss": 3.5217952728271484, "perplexity": 33.84513520652938, "lr": 0.0026291804804649314, "grad_norm": 0.163782, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:48.734154+00:00", "epoch": 0, "step": 11803, "train_loss": 3.5742781162261963, "perplexity": 35.66886274636169, "lr": 0.0026291804804649314, "grad_norm": 0.158157, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:49.038897+00:00", "epoch": 0, "step": 11804, "train_loss": 3.502105712890625, "perplexity": 33.1852570617201, "lr": 0.0026291804804649314, "grad_norm": 0.181472, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:49.344618+00:00", "epoch": 0, "step": 11805, "train_loss": 3.508580446243286, "perplexity": 33.400819855167235, "lr": 0.0026291804804649314, "grad_norm": 0.160565, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:49.649982+00:00", "epoch": 0, "step": 11806, "train_loss": 3.5251617431640625, "perplexity": 33.959265851177854, "lr": 0.0026291804804649314, "grad_norm": 0.156088, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:49.953750+00:00", "epoch": 0, "step": 11807, "train_loss": 3.5955467224121094, "perplexity": 36.435614710631654, "lr": 0.0026291804804649314, "grad_norm": 0.151982, "tokens_per_sec": 107870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:50.257410+00:00", "epoch": 0, "step": 11808, "train_loss": 3.4714841842651367, "perplexity": 32.184474705195285, "lr": 0.0026291804804649314, "grad_norm": 0.152981, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:50.562319+00:00", "epoch": 0, "step": 11809, "train_loss": 3.5250933170318604, "perplexity": 33.95694222946236, "lr": 0.0026291804804649314, "grad_norm": 0.146772, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:50.867355+00:00", "epoch": 0, "step": 11810, "train_loss": 3.537309169769287, "perplexity": 34.374299230918254, "lr": 0.0026291804804649314, "grad_norm": 0.168849, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:51.172830+00:00", "epoch": 0, "step": 11811, "train_loss": 3.5424106121063232, "perplexity": 34.55010578855282, "lr": 0.0026291804804649314, "grad_norm": 0.161607, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:51.476250+00:00", "epoch": 0, "step": 11812, "train_loss": 3.519667387008667, "perplexity": 33.77319319263988, "lr": 0.0026291804804649314, "grad_norm": 0.17252, "tokens_per_sec": 107996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:51.780672+00:00", "epoch": 0, "step": 11813, "train_loss": 3.4767696857452393, "perplexity": 32.35503614819129, "lr": 0.0026291804804649314, "grad_norm": 0.165728, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:52.085315+00:00", "epoch": 0, "step": 11814, "train_loss": 3.593020439147949, "perplexity": 36.343684197120204, "lr": 0.0026291804804649314, "grad_norm": 0.16401, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:52.389579+00:00", "epoch": 0, "step": 11815, "train_loss": 3.516481637954712, "perplexity": 33.665771474547654, "lr": 0.0026291804804649314, "grad_norm": 0.15684, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:52.694406+00:00", "epoch": 0, "step": 11816, "train_loss": 3.5106661319732666, "perplexity": 33.47055616724438, "lr": 0.0026291804804649314, "grad_norm": 0.170395, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:52.998843+00:00", "epoch": 0, "step": 11817, "train_loss": 3.558629274368286, "perplexity": 35.11503106826642, "lr": 0.0026291804804649314, "grad_norm": 0.161343, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:53.303044+00:00", "epoch": 0, "step": 11818, "train_loss": 3.542306661605835, "perplexity": 34.54651447442686, "lr": 0.0026291804804649314, "grad_norm": 0.158161, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:53.607756+00:00", "epoch": 0, "step": 11819, "train_loss": 3.692553758621216, "perplexity": 40.14724252156741, "lr": 0.0026291804804649314, "grad_norm": 0.163254, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:53.912742+00:00", "epoch": 0, "step": 11820, "train_loss": 3.494839668273926, "perplexity": 32.9450053999542, "lr": 0.0026291804804649314, "grad_norm": 0.16488, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:54.217490+00:00", "epoch": 0, "step": 11821, "train_loss": 3.6013410091400146, "perplexity": 36.6473459326934, "lr": 0.0026291804804649314, "grad_norm": 0.151983, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:54.521412+00:00", "epoch": 0, "step": 11822, "train_loss": 3.511584997177124, "perplexity": 33.50132523080626, "lr": 0.0026291804804649314, "grad_norm": 0.174123, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:54.825778+00:00", "epoch": 0, "step": 11823, "train_loss": 3.571148633956909, "perplexity": 35.557412155114996, "lr": 0.0026291804804649314, "grad_norm": 0.153057, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:55.130013+00:00", "epoch": 0, "step": 11824, "train_loss": 3.5306599140167236, "perplexity": 34.146493931188054, "lr": 0.0026291804804649314, "grad_norm": 0.168066, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:55.434878+00:00", "epoch": 0, "step": 11825, "train_loss": 3.534475803375244, "perplexity": 34.2770420942746, "lr": 0.0026291804804649314, "grad_norm": 0.162006, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:55.740099+00:00", "epoch": 0, "step": 11826, "train_loss": 3.50052547454834, "perplexity": 33.13285785863157, "lr": 0.0026291804804649314, "grad_norm": 0.181428, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:56.043901+00:00", "epoch": 0, "step": 11827, "train_loss": 3.5785000324249268, "perplexity": 35.819772034807784, "lr": 0.0026291804804649314, "grad_norm": 0.207401, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:56.347765+00:00", "epoch": 0, "step": 11828, "train_loss": 3.4665753841400146, "perplexity": 32.02687468170785, "lr": 0.0026291804804649314, "grad_norm": 0.175983, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:56.653118+00:00", "epoch": 0, "step": 11829, "train_loss": 3.493091106414795, "perplexity": 32.88744935488022, "lr": 0.0026291804804649314, "grad_norm": 0.15665, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:56.958252+00:00", "epoch": 0, "step": 11830, "train_loss": 3.534620761871338, "perplexity": 34.28201120289568, "lr": 0.0026291804804649314, "grad_norm": 0.187446, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:57.262306+00:00", "epoch": 0, "step": 11831, "train_loss": 3.5812535285949707, "perplexity": 35.91853755275396, "lr": 0.0026291804804649314, "grad_norm": 0.165786, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:57.565446+00:00", "epoch": 0, "step": 11832, "train_loss": 3.5296430587768555, "perplexity": 34.11178953757359, "lr": 0.0026291804804649314, "grad_norm": 0.167289, "tokens_per_sec": 108095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:57.869586+00:00", "epoch": 0, "step": 11833, "train_loss": 3.537726879119873, "perplexity": 34.38866069637724, "lr": 0.0026291804804649314, "grad_norm": 0.172862, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:58.173882+00:00", "epoch": 0, "step": 11834, "train_loss": 3.4749934673309326, "perplexity": 32.29761754627442, "lr": 0.0026291804804649314, "grad_norm": 0.15255, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:58.478130+00:00", "epoch": 0, "step": 11835, "train_loss": 3.5786683559417725, "perplexity": 35.8258018522751, "lr": 0.0026291804804649314, "grad_norm": 0.169805, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:58.782504+00:00", "epoch": 0, "step": 11836, "train_loss": 3.5340864658355713, "perplexity": 34.263699352622005, "lr": 0.0026291804804649314, "grad_norm": 0.160136, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:59.086532+00:00", "epoch": 0, "step": 11837, "train_loss": 3.471907138824463, "perplexity": 32.19809015466661, "lr": 0.0026291804804649314, "grad_norm": 0.146635, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:59.390245+00:00", "epoch": 0, "step": 11838, "train_loss": 3.556849479675293, "perplexity": 35.052589105746534, "lr": 0.0026291804804649314, "grad_norm": 0.15724, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:59.694907+00:00", "epoch": 0, "step": 11839, "train_loss": 3.458298921585083, "perplexity": 31.762899350479152, "lr": 0.0026291804804649314, "grad_norm": 0.148143, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T21:59:59.999396+00:00", "epoch": 0, "step": 11840, "train_loss": 3.607767343521118, "perplexity": 36.8836123823351, "lr": 0.0026291804804649314, "grad_norm": 0.168315, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:00.305052+00:00", "epoch": 0, "step": 11841, "train_loss": 3.5574405193328857, "perplexity": 35.07331269963898, "lr": 0.0026291804804649314, "grad_norm": 0.148482, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:00.609752+00:00", "epoch": 0, "step": 11842, "train_loss": 3.4176430702209473, "perplexity": 30.49744989814405, "lr": 0.0026291804804649314, "grad_norm": 0.189926, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:00.913767+00:00", "epoch": 0, "step": 11843, "train_loss": 3.5150890350341797, "perplexity": 33.61892105245913, "lr": 0.0026291804804649314, "grad_norm": 0.180939, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:01.217985+00:00", "epoch": 0, "step": 11844, "train_loss": 3.640204429626465, "perplexity": 38.099624621363354, "lr": 0.0026291804804649314, "grad_norm": 0.162279, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:01.522837+00:00", "epoch": 0, "step": 11845, "train_loss": 3.538560390472412, "perplexity": 34.41733598439517, "lr": 0.0026291804804649314, "grad_norm": 0.152215, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:01.826574+00:00", "epoch": 0, "step": 11846, "train_loss": 3.5346274375915527, "perplexity": 34.282240060774775, "lr": 0.0026291804804649314, "grad_norm": 0.157542, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:02.130440+00:00", "epoch": 0, "step": 11847, "train_loss": 3.5676703453063965, "perplexity": 35.43394805833497, "lr": 0.0026291804804649314, "grad_norm": 0.158053, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:02.435272+00:00", "epoch": 0, "step": 11848, "train_loss": 3.5352730751037598, "perplexity": 34.30438110773661, "lr": 0.0026291804804649314, "grad_norm": 0.147446, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:02.740045+00:00", "epoch": 0, "step": 11849, "train_loss": 3.5475151538848877, "perplexity": 34.726919138644924, "lr": 0.0026291804804649314, "grad_norm": 0.138886, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:03.044450+00:00", "epoch": 0, "step": 11850, "train_loss": 3.606473922729492, "perplexity": 36.83593718989856, "lr": 0.0026291804804649314, "grad_norm": 0.143558, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:03.348307+00:00", "epoch": 0, "step": 11851, "train_loss": 3.522071361541748, "perplexity": 33.85448075644507, "lr": 0.0026291804804649314, "grad_norm": 0.135164, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:03.651848+00:00", "epoch": 0, "step": 11852, "train_loss": 3.5564188957214355, "perplexity": 35.037499272289665, "lr": 0.0026291804804649314, "grad_norm": 0.144154, "tokens_per_sec": 107950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:03.957572+00:00", "epoch": 0, "step": 11853, "train_loss": 3.5660204887390137, "perplexity": 35.37553532598775, "lr": 0.0026291804804649314, "grad_norm": 0.17123, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:04.261727+00:00", "epoch": 0, "step": 11854, "train_loss": 3.5675041675567627, "perplexity": 35.42806021381392, "lr": 0.0026291804804649314, "grad_norm": 0.158795, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:04.565937+00:00", "epoch": 0, "step": 11855, "train_loss": 3.5083000659942627, "perplexity": 33.39145623772669, "lr": 0.0026291804804649314, "grad_norm": 0.17821, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:04.869635+00:00", "epoch": 0, "step": 11856, "train_loss": 3.4694008827209473, "perplexity": 32.117494533515874, "lr": 0.0026291804804649314, "grad_norm": 0.160758, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:05.173249+00:00", "epoch": 0, "step": 11857, "train_loss": 3.550419807434082, "perplexity": 34.82793544523733, "lr": 0.0026291804804649314, "grad_norm": 0.179397, "tokens_per_sec": 107927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:05.477301+00:00", "epoch": 0, "step": 11858, "train_loss": 3.5140371322631836, "perplexity": 33.58357580938484, "lr": 0.0026291804804649314, "grad_norm": 0.148564, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:05.782253+00:00", "epoch": 0, "step": 11859, "train_loss": 3.6556520462036133, "perplexity": 38.69274234731841, "lr": 0.0026291804804649314, "grad_norm": 0.167396, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:06.086663+00:00", "epoch": 0, "step": 11860, "train_loss": 3.5672411918640137, "perplexity": 35.41874472006433, "lr": 0.0026291804804649314, "grad_norm": 0.170093, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:06.389822+00:00", "epoch": 0, "step": 11861, "train_loss": 3.4719018936157227, "perplexity": 32.19792126940563, "lr": 0.0026291804804649314, "grad_norm": 0.181267, "tokens_per_sec": 107987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:06.693928+00:00", "epoch": 0, "step": 11862, "train_loss": 3.5701067447662354, "perplexity": 35.52038456441505, "lr": 0.0026291804804649314, "grad_norm": 0.143366, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:06.999009+00:00", "epoch": 0, "step": 11863, "train_loss": 3.5512917041778564, "perplexity": 34.85831505075943, "lr": 0.0026291804804649314, "grad_norm": 0.194475, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:07.303762+00:00", "epoch": 0, "step": 11864, "train_loss": 3.4824743270874023, "perplexity": 32.54013749143893, "lr": 0.0026291804804649314, "grad_norm": 0.185571, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:07.607915+00:00", "epoch": 0, "step": 11865, "train_loss": 3.572458028793335, "perplexity": 35.60400134215155, "lr": 0.0026291804804649314, "grad_norm": 0.147036, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:07.911473+00:00", "epoch": 0, "step": 11866, "train_loss": 3.4995572566986084, "perplexity": 33.10079355935855, "lr": 0.0026291804804649314, "grad_norm": 0.177366, "tokens_per_sec": 107952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:08.214905+00:00", "epoch": 0, "step": 11867, "train_loss": 3.4960179328918457, "perplexity": 32.98384621203787, "lr": 0.0026291804804649314, "grad_norm": 0.16867, "tokens_per_sec": 107987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:08.518789+00:00", "epoch": 0, "step": 11868, "train_loss": 3.4858078956604004, "perplexity": 32.64879327627484, "lr": 0.0026291804804649314, "grad_norm": 0.153559, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:08.823600+00:00", "epoch": 0, "step": 11869, "train_loss": 3.6152470111846924, "perplexity": 37.16052385709573, "lr": 0.0026291804804649314, "grad_norm": 0.168245, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:09.127326+00:00", "epoch": 0, "step": 11870, "train_loss": 3.4879393577575684, "perplexity": 32.71845715824913, "lr": 0.0026291804804649314, "grad_norm": 0.163771, "tokens_per_sec": 107953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:09.432483+00:00", "epoch": 0, "step": 11871, "train_loss": 3.5741255283355713, "perplexity": 35.66342052505332, "lr": 0.0026291804804649314, "grad_norm": 0.168264, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:09.737486+00:00", "epoch": 0, "step": 11872, "train_loss": 3.5414927005767822, "perplexity": 34.51840639894597, "lr": 0.0026291804804649314, "grad_norm": 0.182596, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:10.042223+00:00", "epoch": 0, "step": 11873, "train_loss": 3.5264570713043213, "perplexity": 34.003282745864006, "lr": 0.0026291804804649314, "grad_norm": 0.186329, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:10.347389+00:00", "epoch": 0, "step": 11874, "train_loss": 3.4832146167755127, "perplexity": 32.56423553834081, "lr": 0.0026291804804649314, "grad_norm": 0.151039, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:10.651572+00:00", "epoch": 0, "step": 11875, "train_loss": 3.4238836765289307, "perplexity": 30.688367577784074, "lr": 0.0026291804804649314, "grad_norm": 0.162373, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:10.955798+00:00", "epoch": 0, "step": 11876, "train_loss": 3.5537257194519043, "perplexity": 34.94326406367849, "lr": 0.0026291804804649314, "grad_norm": 0.183774, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:11.260193+00:00", "epoch": 0, "step": 11877, "train_loss": 3.509676456451416, "perplexity": 33.437447563192805, "lr": 0.0026291804804649314, "grad_norm": 0.209936, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:11.564511+00:00", "epoch": 0, "step": 11878, "train_loss": 3.4510233402252197, "perplexity": 31.53264442693802, "lr": 0.0026291804804649314, "grad_norm": 0.149373, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:11.869739+00:00", "epoch": 0, "step": 11879, "train_loss": 3.6111679077148438, "perplexity": 37.00925097391367, "lr": 0.0026291804804649314, "grad_norm": 0.1759, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:12.175764+00:00", "epoch": 0, "step": 11880, "train_loss": 3.52315354347229, "perplexity": 33.891137294731834, "lr": 0.0026291804804649314, "grad_norm": 0.192802, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:12.480471+00:00", "epoch": 0, "step": 11881, "train_loss": 3.465636968612671, "perplexity": 31.996834262616627, "lr": 0.0026291804804649314, "grad_norm": 0.174266, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:12.784653+00:00", "epoch": 0, "step": 11882, "train_loss": 3.427306652069092, "perplexity": 30.793593098820814, "lr": 0.0026291804804649314, "grad_norm": 0.158044, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:13.088795+00:00", "epoch": 0, "step": 11883, "train_loss": 3.5509145259857178, "perplexity": 34.84516973372688, "lr": 0.0026291804804649314, "grad_norm": 0.170726, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:13.393804+00:00", "epoch": 0, "step": 11884, "train_loss": 3.491417646408081, "perplexity": 32.8324595481163, "lr": 0.0026291804804649314, "grad_norm": 0.148662, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:13.698413+00:00", "epoch": 0, "step": 11885, "train_loss": 3.5398495197296143, "perplexity": 34.46173298975844, "lr": 0.0026291804804649314, "grad_norm": 0.160164, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:14.002964+00:00", "epoch": 0, "step": 11886, "train_loss": 3.4686813354492188, "perplexity": 32.09439279034441, "lr": 0.0026291804804649314, "grad_norm": 0.148354, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:14.308081+00:00", "epoch": 0, "step": 11887, "train_loss": 3.580671548843384, "perplexity": 35.89763977282387, "lr": 0.0026291804804649314, "grad_norm": 0.15944, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:14.612150+00:00", "epoch": 0, "step": 11888, "train_loss": 3.5163559913635254, "perplexity": 33.66154175085296, "lr": 0.0026291804804649314, "grad_norm": 0.155276, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:14.916315+00:00", "epoch": 0, "step": 11889, "train_loss": 3.466049909591675, "perplexity": 32.010049795111726, "lr": 0.0026291804804649314, "grad_norm": 0.156892, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:15.220933+00:00", "epoch": 0, "step": 11890, "train_loss": 3.582282304763794, "perplexity": 35.9555087024653, "lr": 0.0026291804804649314, "grad_norm": 0.182405, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:15.524835+00:00", "epoch": 0, "step": 11891, "train_loss": 3.468221426010132, "perplexity": 32.07963566988798, "lr": 0.0026291804804649314, "grad_norm": 0.161854, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:15.828902+00:00", "epoch": 0, "step": 11892, "train_loss": 3.4700675010681152, "perplexity": 32.1389117824094, "lr": 0.0026291804804649314, "grad_norm": 0.180205, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:16.133267+00:00", "epoch": 0, "step": 11893, "train_loss": 3.5542922019958496, "perplexity": 34.9630644205478, "lr": 0.0026291804804649314, "grad_norm": 0.172099, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:16.436797+00:00", "epoch": 0, "step": 11894, "train_loss": 3.5125908851623535, "perplexity": 33.535040765528024, "lr": 0.0026291804804649314, "grad_norm": 0.166618, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:16.741303+00:00", "epoch": 0, "step": 11895, "train_loss": 3.502692699432373, "perplexity": 33.20474207916182, "lr": 0.0026291804804649314, "grad_norm": 0.165236, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:17.045286+00:00", "epoch": 0, "step": 11896, "train_loss": 3.5760018825531006, "perplexity": 35.73040055404556, "lr": 0.0026291804804649314, "grad_norm": 0.155192, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:17.348467+00:00", "epoch": 0, "step": 11897, "train_loss": 3.4603049755096436, "perplexity": 31.82668139309547, "lr": 0.0026291804804649314, "grad_norm": 0.170042, "tokens_per_sec": 108079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:17.653239+00:00", "epoch": 0, "step": 11898, "train_loss": 3.473201274871826, "perplexity": 32.23978583791609, "lr": 0.0026291804804649314, "grad_norm": 0.1605, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:17.957314+00:00", "epoch": 0, "step": 11899, "train_loss": 3.476557731628418, "perplexity": 32.34817909179612, "lr": 0.0026291804804649314, "grad_norm": 0.165792, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:18.261142+00:00", "epoch": 0, "step": 11900, "train_loss": 3.6270530223846436, "perplexity": 37.60184139186662, "lr": 0.0026291804804649314, "grad_norm": 0.165205, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:18.566929+00:00", "epoch": 0, "step": 11901, "train_loss": 3.5404531955718994, "perplexity": 34.48254298606169, "lr": 0.0026291804804649314, "grad_norm": 0.151516, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:18.870168+00:00", "epoch": 0, "step": 11902, "train_loss": 3.5072691440582275, "perplexity": 33.35704999113948, "lr": 0.0026291804804649314, "grad_norm": 0.168829, "tokens_per_sec": 108060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:19.174094+00:00", "epoch": 0, "step": 11903, "train_loss": 3.4723398685455322, "perplexity": 32.21202624029996, "lr": 0.0026291804804649314, "grad_norm": 0.158352, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:19.477600+00:00", "epoch": 0, "step": 11904, "train_loss": 3.552640199661255, "perplexity": 34.90535303929382, "lr": 0.0026291804804649314, "grad_norm": 0.167117, "tokens_per_sec": 107965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:19.781782+00:00", "epoch": 0, "step": 11905, "train_loss": 3.5620310306549072, "perplexity": 35.23468725115609, "lr": 0.0026291804804649314, "grad_norm": 0.190307, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:20.086334+00:00", "epoch": 0, "step": 11906, "train_loss": 3.568312406539917, "perplexity": 35.45670612798912, "lr": 0.0026291804804649314, "grad_norm": 0.156964, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:20.390117+00:00", "epoch": 0, "step": 11907, "train_loss": 3.5513477325439453, "perplexity": 34.860268159910575, "lr": 0.0026291804804649314, "grad_norm": 0.166454, "tokens_per_sec": 107870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:20.693647+00:00", "epoch": 0, "step": 11908, "train_loss": 3.420255184173584, "perplexity": 30.577216847368266, "lr": 0.0026291804804649314, "grad_norm": 0.177948, "tokens_per_sec": 107954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:20.997608+00:00", "epoch": 0, "step": 11909, "train_loss": 3.3989291191101074, "perplexity": 29.932029240345486, "lr": 0.0026291804804649314, "grad_norm": 0.166112, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:21.302672+00:00", "epoch": 0, "step": 11910, "train_loss": 3.499103307723999, "perplexity": 33.085770898082345, "lr": 0.0026291804804649314, "grad_norm": 0.159752, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:21.607431+00:00", "epoch": 0, "step": 11911, "train_loss": 3.617800235748291, "perplexity": 37.25552424645909, "lr": 0.0026291804804649314, "grad_norm": 0.195691, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:21.911729+00:00", "epoch": 0, "step": 11912, "train_loss": 3.4074695110321045, "perplexity": 30.188755210853426, "lr": 0.0026291804804649314, "grad_norm": 0.187902, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:22.216266+00:00", "epoch": 0, "step": 11913, "train_loss": 3.4435312747955322, "perplexity": 31.29728256587625, "lr": 0.0026291804804649314, "grad_norm": 0.157328, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:22.520216+00:00", "epoch": 0, "step": 11914, "train_loss": 3.6078875064849854, "perplexity": 36.88804469281154, "lr": 0.0026291804804649314, "grad_norm": 0.184222, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:22.825426+00:00", "epoch": 0, "step": 11915, "train_loss": 3.565607786178589, "perplexity": 35.360938764208704, "lr": 0.0026291804804649314, "grad_norm": 0.197111, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:23.130035+00:00", "epoch": 0, "step": 11916, "train_loss": 3.4998769760131836, "perplexity": 33.1113782143564, "lr": 0.0026291804804649314, "grad_norm": 0.17596, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:23.435120+00:00", "epoch": 0, "step": 11917, "train_loss": 3.4832358360290527, "perplexity": 32.56492653444222, "lr": 0.0026291804804649314, "grad_norm": 0.168693, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:23.738496+00:00", "epoch": 0, "step": 11918, "train_loss": 3.461413860321045, "perplexity": 31.86199309137418, "lr": 0.0026291804804649314, "grad_norm": 0.150197, "tokens_per_sec": 108011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:24.042891+00:00", "epoch": 0, "step": 11919, "train_loss": 3.493563175201416, "perplexity": 32.902978158235534, "lr": 0.0026291804804649314, "grad_norm": 0.193117, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:24.346809+00:00", "epoch": 0, "step": 11920, "train_loss": 3.527883768081665, "perplexity": 34.0518297424638, "lr": 0.0026291804804649314, "grad_norm": 0.163516, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:24.651008+00:00", "epoch": 0, "step": 11921, "train_loss": 3.4925785064697266, "perplexity": 32.87059557014107, "lr": 0.0026291804804649314, "grad_norm": 0.157334, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:24.955764+00:00", "epoch": 0, "step": 11922, "train_loss": 3.5752832889556885, "perplexity": 35.70473413993934, "lr": 0.0026291804804649314, "grad_norm": 0.161067, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:25.259304+00:00", "epoch": 0, "step": 11923, "train_loss": 3.4808690547943115, "perplexity": 32.48794361421124, "lr": 0.0026291804804649314, "grad_norm": 0.200824, "tokens_per_sec": 107954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:25.563588+00:00", "epoch": 0, "step": 11924, "train_loss": 3.570963144302368, "perplexity": 35.55081723468164, "lr": 0.0026291804804649314, "grad_norm": 0.168196, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:25.868986+00:00", "epoch": 0, "step": 11925, "train_loss": 3.42427921295166, "perplexity": 30.700508345814892, "lr": 0.0026291804804649314, "grad_norm": 0.15465, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:26.174407+00:00", "epoch": 0, "step": 11926, "train_loss": 3.5006661415100098, "perplexity": 33.1375188848965, "lr": 0.0026291804804649314, "grad_norm": 0.160979, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:26.479419+00:00", "epoch": 0, "step": 11927, "train_loss": 3.5137274265289307, "perplexity": 33.573176393842395, "lr": 0.0026291804804649314, "grad_norm": 0.159308, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:26.782889+00:00", "epoch": 0, "step": 11928, "train_loss": 3.4948744773864746, "perplexity": 32.94615220631463, "lr": 0.0026291804804649314, "grad_norm": 0.164354, "tokens_per_sec": 107978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:27.086777+00:00", "epoch": 0, "step": 11929, "train_loss": 3.615170478820801, "perplexity": 37.15767998318708, "lr": 0.0026291804804649314, "grad_norm": 0.15822, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:27.390399+00:00", "epoch": 0, "step": 11930, "train_loss": 3.5043327808380127, "perplexity": 33.259245241819535, "lr": 0.0026291804804649314, "grad_norm": 0.158081, "tokens_per_sec": 107993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:27.695500+00:00", "epoch": 0, "step": 11931, "train_loss": 3.4014892578125, "perplexity": 30.00875756250998, "lr": 0.0026291804804649314, "grad_norm": 0.17887, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:28.000619+00:00", "epoch": 0, "step": 11932, "train_loss": 3.57938814163208, "perplexity": 35.851598034539855, "lr": 0.0026291804804649314, "grad_norm": 0.172053, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:28.305123+00:00", "epoch": 0, "step": 11933, "train_loss": 3.516908645629883, "perplexity": 33.6801500870283, "lr": 0.0026291804804649314, "grad_norm": 0.167082, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:28.608297+00:00", "epoch": 0, "step": 11934, "train_loss": 3.544638156890869, "perplexity": 34.62715347823835, "lr": 0.0026291804804649314, "grad_norm": 0.161256, "tokens_per_sec": 108078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:28.912555+00:00", "epoch": 0, "step": 11935, "train_loss": 3.4560887813568115, "perplexity": 31.69277640817022, "lr": 0.0026291804804649314, "grad_norm": 0.166205, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:29.216305+00:00", "epoch": 0, "step": 11936, "train_loss": 3.582754373550415, "perplexity": 35.9724861827848, "lr": 0.0026291804804649314, "grad_norm": 0.171983, "tokens_per_sec": 107877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:29.520937+00:00", "epoch": 0, "step": 11937, "train_loss": 3.4514431953430176, "perplexity": 31.545886348728164, "lr": 0.0026291804804649314, "grad_norm": 0.191684, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:29.824844+00:00", "epoch": 0, "step": 11938, "train_loss": 3.4924027919769287, "perplexity": 32.864820237532186, "lr": 0.0026291804804649314, "grad_norm": 0.143197, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:30.128452+00:00", "epoch": 0, "step": 11939, "train_loss": 3.5246169567108154, "perplexity": 33.94077034168887, "lr": 0.0026291804804649314, "grad_norm": 0.165904, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:30.432533+00:00", "epoch": 0, "step": 11940, "train_loss": 3.5156350135803223, "perplexity": 33.63728127378562, "lr": 0.0026291804804649314, "grad_norm": 0.19068, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:30.736681+00:00", "epoch": 0, "step": 11941, "train_loss": 3.568720817565918, "perplexity": 35.47118999520168, "lr": 0.0026291804804649314, "grad_norm": 0.174266, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:31.042016+00:00", "epoch": 0, "step": 11942, "train_loss": 3.4663710594177246, "perplexity": 32.02033146792562, "lr": 0.0026291804804649314, "grad_norm": 0.157727, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:31.346470+00:00", "epoch": 0, "step": 11943, "train_loss": 3.4770402908325195, "perplexity": 32.36379277031194, "lr": 0.0026291804804649314, "grad_norm": 0.167387, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:31.649973+00:00", "epoch": 0, "step": 11944, "train_loss": 3.596817970275879, "perplexity": 36.4819628617499, "lr": 0.0026291804804649314, "grad_norm": 0.181591, "tokens_per_sec": 107965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:31.953475+00:00", "epoch": 0, "step": 11945, "train_loss": 3.5343544483184814, "perplexity": 34.272882654275236, "lr": 0.0026291804804649314, "grad_norm": 0.177895, "tokens_per_sec": 107967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:32.256772+00:00", "epoch": 0, "step": 11946, "train_loss": 3.5166358947753906, "perplexity": 33.67096504998293, "lr": 0.0026291804804649314, "grad_norm": 0.170326, "tokens_per_sec": 108039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:32.560665+00:00", "epoch": 0, "step": 11947, "train_loss": 3.6464760303497314, "perplexity": 38.339321109364484, "lr": 0.0026291804804649314, "grad_norm": 0.155284, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:32.864931+00:00", "epoch": 0, "step": 11948, "train_loss": 3.5726184844970703, "perplexity": 35.609714665598105, "lr": 0.0026291804804649314, "grad_norm": 0.162983, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:33.168846+00:00", "epoch": 0, "step": 11949, "train_loss": 3.6746227741241455, "perplexity": 39.433778609472725, "lr": 0.0026291804804649314, "grad_norm": 0.209876, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:33.473346+00:00", "epoch": 0, "step": 11950, "train_loss": 3.4106359481811523, "perplexity": 30.284497507824256, "lr": 0.0026291804804649314, "grad_norm": 0.145335, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:33.776772+00:00", "epoch": 0, "step": 11951, "train_loss": 3.4685962200164795, "perplexity": 32.09166117846637, "lr": 0.0026291804804649314, "grad_norm": 0.172703, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:34.080423+00:00", "epoch": 0, "step": 11952, "train_loss": 3.583883762359619, "perplexity": 36.013136056554345, "lr": 0.0026291804804649314, "grad_norm": 0.16858, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:34.384628+00:00", "epoch": 0, "step": 11953, "train_loss": 3.5282676219940186, "perplexity": 34.0649031795177, "lr": 0.0026291804804649314, "grad_norm": 0.210223, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:34.688762+00:00", "epoch": 0, "step": 11954, "train_loss": 3.646326780319214, "perplexity": 38.333599391512784, "lr": 0.0026291804804649314, "grad_norm": 0.220996, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:34.993234+00:00", "epoch": 0, "step": 11955, "train_loss": 3.486511707305908, "perplexity": 32.67177996539533, "lr": 0.0026291804804649314, "grad_norm": 0.181203, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:35.296414+00:00", "epoch": 0, "step": 11956, "train_loss": 3.5177698135375977, "perplexity": 33.70916684376686, "lr": 0.0026291804804649314, "grad_norm": 0.185857, "tokens_per_sec": 108081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:35.601367+00:00", "epoch": 0, "step": 11957, "train_loss": 3.5220465660095215, "perplexity": 33.85364132698355, "lr": 0.0026291804804649314, "grad_norm": 0.168049, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:35.907388+00:00", "epoch": 0, "step": 11958, "train_loss": 3.481663227081299, "perplexity": 32.51375488663286, "lr": 0.0026291804804649314, "grad_norm": 0.182296, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:36.212715+00:00", "epoch": 0, "step": 11959, "train_loss": 3.5705037117004395, "perplexity": 35.53448778164759, "lr": 0.0026291804804649314, "grad_norm": 0.176226, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:36.516755+00:00", "epoch": 0, "step": 11960, "train_loss": 3.5350427627563477, "perplexity": 34.29648129494431, "lr": 0.0026291804804649314, "grad_norm": 0.176903, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:36.820566+00:00", "epoch": 0, "step": 11961, "train_loss": 3.4936325550079346, "perplexity": 32.90526103968606, "lr": 0.0026291804804649314, "grad_norm": 0.162208, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:37.125047+00:00", "epoch": 0, "step": 11962, "train_loss": 3.5831711292266846, "perplexity": 35.98748104497094, "lr": 0.0026291804804649314, "grad_norm": 0.174461, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:37.429128+00:00", "epoch": 0, "step": 11963, "train_loss": 3.4531893730163574, "perplexity": 31.60101919301795, "lr": 0.0026291804804649314, "grad_norm": 0.180976, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:37.733633+00:00", "epoch": 0, "step": 11964, "train_loss": 3.4499623775482178, "perplexity": 31.499207209050493, "lr": 0.0026291804804649314, "grad_norm": 0.176874, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:38.037672+00:00", "epoch": 0, "step": 11965, "train_loss": 3.576265573501587, "perplexity": 35.73982357958612, "lr": 0.0026291804804649314, "grad_norm": 0.162164, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:38.341709+00:00", "epoch": 0, "step": 11966, "train_loss": 3.488168954849243, "perplexity": 32.725970083296616, "lr": 0.0026291804804649314, "grad_norm": 0.159392, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:38.646039+00:00", "epoch": 0, "step": 11967, "train_loss": 3.437732696533203, "perplexity": 31.116327970239258, "lr": 0.0026291804804649314, "grad_norm": 0.185852, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:38.950048+00:00", "epoch": 0, "step": 11968, "train_loss": 3.502981424331665, "perplexity": 33.214330499115825, "lr": 0.0026291804804649314, "grad_norm": 0.167771, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:39.254011+00:00", "epoch": 0, "step": 11969, "train_loss": 3.528839349746704, "perplexity": 34.08438459856128, "lr": 0.0026291804804649314, "grad_norm": 0.200673, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:39.557526+00:00", "epoch": 0, "step": 11970, "train_loss": 3.5029494762420654, "perplexity": 33.21326938165947, "lr": 0.0026291804804649314, "grad_norm": 0.198605, "tokens_per_sec": 108018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:39.861324+00:00", "epoch": 0, "step": 11971, "train_loss": 3.5858042240142822, "perplexity": 36.08236435728645, "lr": 0.0026291804804649314, "grad_norm": 0.157034, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:40.164668+00:00", "epoch": 0, "step": 11972, "train_loss": 3.552483320236206, "perplexity": 34.899877537085935, "lr": 0.0026291804804649314, "grad_norm": 0.177333, "tokens_per_sec": 108013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:40.469779+00:00", "epoch": 0, "step": 11973, "train_loss": 3.522313117980957, "perplexity": 33.862666284576214, "lr": 0.0026291804804649314, "grad_norm": 0.187618, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:40.775422+00:00", "epoch": 0, "step": 11974, "train_loss": 3.6338109970092773, "perplexity": 37.85681426153575, "lr": 0.0026291804804649314, "grad_norm": 0.152604, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:41.079794+00:00", "epoch": 0, "step": 11975, "train_loss": 3.436760425567627, "perplexity": 31.08608917053403, "lr": 0.0026291804804649314, "grad_norm": 0.177732, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:41.382835+00:00", "epoch": 0, "step": 11976, "train_loss": 3.659024238586426, "perplexity": 38.82344196660628, "lr": 0.0026291804804649314, "grad_norm": 0.158757, "tokens_per_sec": 108130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:41.687393+00:00", "epoch": 0, "step": 11977, "train_loss": 3.5986030101776123, "perplexity": 36.54714277822026, "lr": 0.0026291804804649314, "grad_norm": 0.191323, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:41.992495+00:00", "epoch": 0, "step": 11978, "train_loss": 3.521390676498413, "perplexity": 33.831444358899375, "lr": 0.0026291804804649314, "grad_norm": 0.1662, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:42.298210+00:00", "epoch": 0, "step": 11979, "train_loss": 3.555042028427124, "perplexity": 34.98929048163679, "lr": 0.0026291804804649314, "grad_norm": 0.207976, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:42.603287+00:00", "epoch": 0, "step": 11980, "train_loss": 3.6591029167175293, "perplexity": 38.82649664262974, "lr": 0.0026291804804649314, "grad_norm": 0.167849, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:42.906611+00:00", "epoch": 0, "step": 11981, "train_loss": 3.5057294368743896, "perplexity": 33.305729421100466, "lr": 0.0026291804804649314, "grad_norm": 0.181467, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:43.210810+00:00", "epoch": 0, "step": 11982, "train_loss": 3.5172765254974365, "perplexity": 33.69254261552262, "lr": 0.0026291804804649314, "grad_norm": 0.174621, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:43.515007+00:00", "epoch": 0, "step": 11983, "train_loss": 3.517533540725708, "perplexity": 33.70120322496057, "lr": 0.0026291804804649314, "grad_norm": 0.189079, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:43.819301+00:00", "epoch": 0, "step": 11984, "train_loss": 3.447465181350708, "perplexity": 31.420645641233442, "lr": 0.0026291804804649314, "grad_norm": 0.171674, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:44.125071+00:00", "epoch": 0, "step": 11985, "train_loss": 3.503021478652954, "perplexity": 33.21566090322507, "lr": 0.0026291804804649314, "grad_norm": 0.155977, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:44.428669+00:00", "epoch": 0, "step": 11986, "train_loss": 3.48972225189209, "perplexity": 32.77684273579313, "lr": 0.0026291804804649314, "grad_norm": 0.148446, "tokens_per_sec": 107932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:44.732768+00:00", "epoch": 0, "step": 11987, "train_loss": 3.6160411834716797, "perplexity": 37.19004743716579, "lr": 0.0026291804804649314, "grad_norm": 0.1609, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:45.036483+00:00", "epoch": 0, "step": 11988, "train_loss": 3.63997745513916, "perplexity": 38.09097795992132, "lr": 0.0026291804804649314, "grad_norm": 0.16056, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:45.341438+00:00", "epoch": 0, "step": 11989, "train_loss": 3.64510440826416, "perplexity": 38.2867700980878, "lr": 0.0026291804804649314, "grad_norm": 0.180545, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:45.645758+00:00", "epoch": 0, "step": 11990, "train_loss": 3.542879819869995, "perplexity": 34.56632077022465, "lr": 0.0026291804804649314, "grad_norm": 0.182322, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:45.949860+00:00", "epoch": 0, "step": 11991, "train_loss": 3.633235454559326, "perplexity": 37.835032326723656, "lr": 0.0026291804804649314, "grad_norm": 0.215422, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:46.253943+00:00", "epoch": 0, "step": 11992, "train_loss": 3.597332000732422, "perplexity": 36.50072052236671, "lr": 0.0026291804804649314, "grad_norm": 0.165504, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:46.557534+00:00", "epoch": 0, "step": 11993, "train_loss": 3.5290870666503906, "perplexity": 34.09282892263582, "lr": 0.0026291804804649314, "grad_norm": 0.151288, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:46.862264+00:00", "epoch": 0, "step": 11994, "train_loss": 3.5791828632354736, "perplexity": 35.84423923130706, "lr": 0.0026291804804649314, "grad_norm": 0.163942, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:47.166586+00:00", "epoch": 0, "step": 11995, "train_loss": 3.530275583267212, "perplexity": 34.13337290515068, "lr": 0.0026291804804649314, "grad_norm": 0.151836, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:47.471468+00:00", "epoch": 0, "step": 11996, "train_loss": 3.5733554363250732, "perplexity": 35.63596698207376, "lr": 0.0026291804804649314, "grad_norm": 0.157337, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:47.777203+00:00", "epoch": 0, "step": 11997, "train_loss": 3.487060070037842, "perplexity": 32.68970086504196, "lr": 0.0026291804804649314, "grad_norm": 0.146276, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:48.082359+00:00", "epoch": 0, "step": 11998, "train_loss": 3.486820936203003, "perplexity": 32.68188458611999, "lr": 0.0026291804804649314, "grad_norm": 0.168894, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:48.387076+00:00", "epoch": 0, "step": 11999, "train_loss": 3.528925895690918, "perplexity": 34.08733459146243, "lr": 0.0026291804804649314, "grad_norm": 0.168865, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:48.693072+00:00", "epoch": 0, "step": 12000, "train_loss": 3.569145917892456, "perplexity": 35.48627201511, "lr": 0.0026291804804649314, "grad_norm": 0.175667, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:00:51.766381+00:00", "step": 12000, "epoch": 0, "val_loss": 3.4846145033836367, "val_ppl": 32.60985369825054, "eval_train_loss": 3.569145917892456, "eval_train_ppl": 35.48627201511} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:00:52.711866+00:00", "step": 12000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4846_epoch_0000_step_0012000.pt", "val_loss": 3.4846145033836367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:53.977797+00:00", "epoch": 0, "step": 12001, "train_loss": 3.4639663696289062, "perplexity": 31.94342500895792, "lr": 0.0026291804804649314, "grad_norm": 0.150911, "tokens_per_sec": 6200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:54.281400+00:00", "epoch": 0, "step": 12002, "train_loss": 3.4924142360687256, "perplexity": 32.865196347703986, "lr": 0.0026291804804649314, "grad_norm": 0.162859, "tokens_per_sec": 107926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:54.585699+00:00", "epoch": 0, "step": 12003, "train_loss": 3.527085542678833, "perplexity": 34.02465955236108, "lr": 0.0026291804804649314, "grad_norm": 0.150605, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:54.889821+00:00", "epoch": 0, "step": 12004, "train_loss": 3.5308496952056885, "perplexity": 34.15297490836962, "lr": 0.0026291804804649314, "grad_norm": 0.165172, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:55.193627+00:00", "epoch": 0, "step": 12005, "train_loss": 3.5034961700439453, "perplexity": 33.2314318343684, "lr": 0.0026291804804649314, "grad_norm": 0.170967, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:55.497756+00:00", "epoch": 0, "step": 12006, "train_loss": 3.528223752975464, "perplexity": 34.06340881842637, "lr": 0.0026291804804649314, "grad_norm": 0.156641, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:55.800628+00:00", "epoch": 0, "step": 12007, "train_loss": 3.551713228225708, "perplexity": 34.87301176611277, "lr": 0.0026291804804649314, "grad_norm": 0.192946, "tokens_per_sec": 108192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:56.106730+00:00", "epoch": 0, "step": 12008, "train_loss": 3.6899476051330566, "perplexity": 40.04274886782367, "lr": 0.0026291804804649314, "grad_norm": 0.190669, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:56.411937+00:00", "epoch": 0, "step": 12009, "train_loss": 3.5716135501861572, "perplexity": 35.573947216507335, "lr": 0.0026291804804649314, "grad_norm": 0.201465, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:56.715959+00:00", "epoch": 0, "step": 12010, "train_loss": 3.555708169937134, "perplexity": 35.012606065309285, "lr": 0.0026291804804649314, "grad_norm": 0.140509, "tokens_per_sec": 107880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:57.019715+00:00", "epoch": 0, "step": 12011, "train_loss": 3.6034469604492188, "perplexity": 36.72460478197282, "lr": 0.0026291804804649314, "grad_norm": 0.174796, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:57.323444+00:00", "epoch": 0, "step": 12012, "train_loss": 3.481067419052124, "perplexity": 32.49438870025035, "lr": 0.0026291804804649314, "grad_norm": 0.16338, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:57.627488+00:00", "epoch": 0, "step": 12013, "train_loss": 3.555673360824585, "perplexity": 35.01138732877582, "lr": 0.0026291804804649314, "grad_norm": 0.147924, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:57.931849+00:00", "epoch": 0, "step": 12014, "train_loss": 3.4623465538024902, "perplexity": 31.891724427592788, "lr": 0.0026291804804649314, "grad_norm": 0.153412, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:58.235651+00:00", "epoch": 0, "step": 12015, "train_loss": 3.543839693069458, "perplexity": 34.59951598418542, "lr": 0.0026291804804649314, "grad_norm": 0.172452, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:58.539139+00:00", "epoch": 0, "step": 12016, "train_loss": 3.3996591567993164, "perplexity": 29.953888727959512, "lr": 0.0026291804804649314, "grad_norm": 0.152923, "tokens_per_sec": 107971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:58.842160+00:00", "epoch": 0, "step": 12017, "train_loss": 3.4766316413879395, "perplexity": 32.35057002628936, "lr": 0.0026291804804649314, "grad_norm": 0.152314, "tokens_per_sec": 108139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:59.145939+00:00", "epoch": 0, "step": 12018, "train_loss": 3.5278868675231934, "perplexity": 34.05193528428258, "lr": 0.0026291804804649314, "grad_norm": 0.163628, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:59.450267+00:00", "epoch": 0, "step": 12019, "train_loss": 3.5091116428375244, "perplexity": 33.41856697010527, "lr": 0.0026291804804649314, "grad_norm": 0.170384, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:00:59.753667+00:00", "epoch": 0, "step": 12020, "train_loss": 3.4808924198150635, "perplexity": 32.48870270455602, "lr": 0.0026291804804649314, "grad_norm": 0.182593, "tokens_per_sec": 108061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:00.056217+00:00", "epoch": 0, "step": 12021, "train_loss": 3.577728033065796, "perplexity": 35.792129864994244, "lr": 0.0026291804804649314, "grad_norm": 0.155753, "tokens_per_sec": 108248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:00.359117+00:00", "epoch": 0, "step": 12022, "train_loss": 3.442934513092041, "perplexity": 31.278611117974236, "lr": 0.0026291804804649314, "grad_norm": 0.169358, "tokens_per_sec": 108182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:00.661809+00:00", "epoch": 0, "step": 12023, "train_loss": 3.512636184692383, "perplexity": 33.53655992152248, "lr": 0.0026291804804649314, "grad_norm": 0.16571, "tokens_per_sec": 108254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:00.965612+00:00", "epoch": 0, "step": 12024, "train_loss": 3.4941861629486084, "perplexity": 32.92348269684924, "lr": 0.0026291804804649314, "grad_norm": 0.159573, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:01.270038+00:00", "epoch": 0, "step": 12025, "train_loss": 3.5421712398529053, "perplexity": 34.54183644163997, "lr": 0.0026291804804649314, "grad_norm": 0.162701, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:01.573042+00:00", "epoch": 0, "step": 12026, "train_loss": 3.4862465858459473, "perplexity": 32.66311912352963, "lr": 0.0026291804804649314, "grad_norm": 0.161861, "tokens_per_sec": 108144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:01.876161+00:00", "epoch": 0, "step": 12027, "train_loss": 3.603074073791504, "perplexity": 36.71091321969887, "lr": 0.0026291804804649314, "grad_norm": 0.151256, "tokens_per_sec": 108104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:02.179805+00:00", "epoch": 0, "step": 12028, "train_loss": 3.596442461013794, "perplexity": 36.468266118584175, "lr": 0.0026291804804649314, "grad_norm": 0.167092, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:02.483689+00:00", "epoch": 0, "step": 12029, "train_loss": 3.472520112991333, "perplexity": 32.21783280240233, "lr": 0.0026291804804649314, "grad_norm": 0.168213, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:02.787171+00:00", "epoch": 0, "step": 12030, "train_loss": 3.5482449531555176, "perplexity": 34.752272069053326, "lr": 0.0026291804804649314, "grad_norm": 0.162927, "tokens_per_sec": 108034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:03.090326+00:00", "epoch": 0, "step": 12031, "train_loss": 3.512887716293335, "perplexity": 33.54499648711692, "lr": 0.0026291804804649314, "grad_norm": 0.15684, "tokens_per_sec": 108030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:03.393539+00:00", "epoch": 0, "step": 12032, "train_loss": 3.6128711700439453, "perplexity": 37.0723411512344, "lr": 0.0026291804804649314, "grad_norm": 0.156089, "tokens_per_sec": 108069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:03.697253+00:00", "epoch": 0, "step": 12033, "train_loss": 3.541659116744995, "perplexity": 34.52415129788047, "lr": 0.0026291804804649314, "grad_norm": 0.148334, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:04.000953+00:00", "epoch": 0, "step": 12034, "train_loss": 3.479200839996338, "perplexity": 32.43379192679983, "lr": 0.0026291804804649314, "grad_norm": 0.161094, "tokens_per_sec": 107896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:04.304330+00:00", "epoch": 0, "step": 12035, "train_loss": 3.3979289531707764, "perplexity": 29.902107210196256, "lr": 0.0026291804804649314, "grad_norm": 0.177974, "tokens_per_sec": 108012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:04.607244+00:00", "epoch": 0, "step": 12036, "train_loss": 3.591999053955078, "perplexity": 36.30658224714353, "lr": 0.0026291804804649314, "grad_norm": 0.155803, "tokens_per_sec": 108175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:04.909795+00:00", "epoch": 0, "step": 12037, "train_loss": 3.4186453819274902, "perplexity": 30.528033173623783, "lr": 0.0026291804804649314, "grad_norm": 0.163895, "tokens_per_sec": 108308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:05.213991+00:00", "epoch": 0, "step": 12038, "train_loss": 3.463393449783325, "perplexity": 31.925129228342804, "lr": 0.0026291804804649314, "grad_norm": 0.183223, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:05.517960+00:00", "epoch": 0, "step": 12039, "train_loss": 3.550703763961792, "perplexity": 34.8378264690976, "lr": 0.0026291804804649314, "grad_norm": 0.198753, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:05.821065+00:00", "epoch": 0, "step": 12040, "train_loss": 3.5003316402435303, "perplexity": 33.126436196551495, "lr": 0.0026291804804649314, "grad_norm": 0.205305, "tokens_per_sec": 108167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:06.124537+00:00", "epoch": 0, "step": 12041, "train_loss": 3.421988010406494, "perplexity": 30.63024778428048, "lr": 0.0026291804804649314, "grad_norm": 0.178839, "tokens_per_sec": 107918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:06.427553+00:00", "epoch": 0, "step": 12042, "train_loss": 3.473501443862915, "perplexity": 32.2494646744746, "lr": 0.0026291804804649314, "grad_norm": 0.162353, "tokens_per_sec": 108140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:06.732161+00:00", "epoch": 0, "step": 12043, "train_loss": 3.525158166885376, "perplexity": 33.95914440359634, "lr": 0.0026291804804649314, "grad_norm": 0.192924, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:07.035473+00:00", "epoch": 0, "step": 12044, "train_loss": 3.4822564125061035, "perplexity": 32.53304729355888, "lr": 0.0026291804804649314, "grad_norm": 0.177344, "tokens_per_sec": 108032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:07.340351+00:00", "epoch": 0, "step": 12045, "train_loss": 3.4742283821105957, "perplexity": 32.27291656681973, "lr": 0.0026291804804649314, "grad_norm": 0.183822, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:07.643783+00:00", "epoch": 0, "step": 12046, "train_loss": 3.6050052642822266, "perplexity": 36.781877486918425, "lr": 0.0026291804804649314, "grad_norm": 0.171698, "tokens_per_sec": 107992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:07.947362+00:00", "epoch": 0, "step": 12047, "train_loss": 3.517716407775879, "perplexity": 33.70736662810603, "lr": 0.0026291804804649314, "grad_norm": 0.179134, "tokens_per_sec": 107939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:08.251870+00:00", "epoch": 0, "step": 12048, "train_loss": 3.4379613399505615, "perplexity": 31.123443327210786, "lr": 0.0026291804804649314, "grad_norm": 0.166618, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:08.555683+00:00", "epoch": 0, "step": 12049, "train_loss": 3.413923978805542, "perplexity": 30.38423774769699, "lr": 0.0026291804804649314, "grad_norm": 0.159561, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:08.859829+00:00", "epoch": 0, "step": 12050, "train_loss": 3.5561976432800293, "perplexity": 35.02974799756075, "lr": 0.0026291804804649314, "grad_norm": 0.169471, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:09.163148+00:00", "epoch": 0, "step": 12051, "train_loss": 3.536046028137207, "perplexity": 34.33090703350109, "lr": 0.0026291804804649314, "grad_norm": 0.190871, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:09.467639+00:00", "epoch": 0, "step": 12052, "train_loss": 3.6234073638916016, "perplexity": 37.465007495822896, "lr": 0.0026291804804649314, "grad_norm": 0.21001, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:09.771930+00:00", "epoch": 0, "step": 12053, "train_loss": 3.5159740447998047, "perplexity": 33.64868729566545, "lr": 0.0026291804804649314, "grad_norm": 0.161852, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:10.075460+00:00", "epoch": 0, "step": 12054, "train_loss": 3.493271589279175, "perplexity": 32.893385511613104, "lr": 0.0026291804804649314, "grad_norm": 0.179941, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:10.378988+00:00", "epoch": 0, "step": 12055, "train_loss": 3.5256855487823486, "perplexity": 33.97705856497848, "lr": 0.0026291804804649314, "grad_norm": 0.171595, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:10.684406+00:00", "epoch": 0, "step": 12056, "train_loss": 3.531402111053467, "perplexity": 34.17184676503421, "lr": 0.0026291804804649314, "grad_norm": 0.180603, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:10.989693+00:00", "epoch": 0, "step": 12057, "train_loss": 3.518454074859619, "perplexity": 33.732240616180334, "lr": 0.0026291804804649314, "grad_norm": 0.173649, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:11.293542+00:00", "epoch": 0, "step": 12058, "train_loss": 3.5710232257843018, "perplexity": 35.55295324463172, "lr": 0.0026291804804649314, "grad_norm": 0.169869, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:11.597313+00:00", "epoch": 0, "step": 12059, "train_loss": 3.4557456970214844, "perplexity": 31.681904978056878, "lr": 0.0026291804804649314, "grad_norm": 0.165378, "tokens_per_sec": 107872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:11.900998+00:00", "epoch": 0, "step": 12060, "train_loss": 3.487532377243042, "perplexity": 32.70514409298319, "lr": 0.0026291804804649314, "grad_norm": 0.175391, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:12.205115+00:00", "epoch": 0, "step": 12061, "train_loss": 3.669816017150879, "perplexity": 39.24468484712255, "lr": 0.0026291804804649314, "grad_norm": 0.172716, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:12.509393+00:00", "epoch": 0, "step": 12062, "train_loss": 3.529376268386841, "perplexity": 34.10269005382015, "lr": 0.0026291804804649314, "grad_norm": 0.160833, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:12.813057+00:00", "epoch": 0, "step": 12063, "train_loss": 3.4839019775390625, "perplexity": 32.58662661063555, "lr": 0.0026291804804649314, "grad_norm": 0.158641, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:13.115656+00:00", "epoch": 0, "step": 12064, "train_loss": 3.6283183097839355, "perplexity": 37.649448640044135, "lr": 0.0026291804804649314, "grad_norm": 0.151788, "tokens_per_sec": 108289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:13.418729+00:00", "epoch": 0, "step": 12065, "train_loss": 3.507051944732666, "perplexity": 33.349805649140265, "lr": 0.0026291804804649314, "grad_norm": 0.159555, "tokens_per_sec": 108118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:13.724279+00:00", "epoch": 0, "step": 12066, "train_loss": 3.4981777667999268, "perplexity": 33.055162829806555, "lr": 0.0026291804804649314, "grad_norm": 0.173087, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:14.028046+00:00", "epoch": 0, "step": 12067, "train_loss": 3.5896055698394775, "perplexity": 36.219786932282794, "lr": 0.0026291804804649314, "grad_norm": 0.166702, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:14.331732+00:00", "epoch": 0, "step": 12068, "train_loss": 3.5896871089935303, "perplexity": 36.22274038347854, "lr": 0.0026291804804649314, "grad_norm": 0.154624, "tokens_per_sec": 107901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:14.635096+00:00", "epoch": 0, "step": 12069, "train_loss": 3.6107237339019775, "perplexity": 36.99281608404132, "lr": 0.0026291804804649314, "grad_norm": 0.167507, "tokens_per_sec": 108016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:14.938487+00:00", "epoch": 0, "step": 12070, "train_loss": 3.531129837036133, "perplexity": 34.16254392555398, "lr": 0.0026291804804649314, "grad_norm": 0.164293, "tokens_per_sec": 108075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:15.242770+00:00", "epoch": 0, "step": 12071, "train_loss": 3.487372398376465, "perplexity": 32.6999123795929, "lr": 0.0026291804804649314, "grad_norm": 0.181832, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:15.546765+00:00", "epoch": 0, "step": 12072, "train_loss": 3.4637973308563232, "perplexity": 31.93802578795401, "lr": 0.0026291804804649314, "grad_norm": 0.182594, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:15.850631+00:00", "epoch": 0, "step": 12073, "train_loss": 3.5956673622131348, "perplexity": 36.440010561092514, "lr": 0.0026291804804649314, "grad_norm": 0.177764, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:16.154372+00:00", "epoch": 0, "step": 12074, "train_loss": 3.5267984867095947, "perplexity": 34.01489397243634, "lr": 0.0026291804804649314, "grad_norm": 0.158999, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:16.456720+00:00", "epoch": 0, "step": 12075, "train_loss": 3.5480763912200928, "perplexity": 34.74641465249556, "lr": 0.0026291804804649314, "grad_norm": 0.180469, "tokens_per_sec": 108378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:16.761461+00:00", "epoch": 0, "step": 12076, "train_loss": 3.4603514671325684, "perplexity": 31.828161101562504, "lr": 0.0026291804804649314, "grad_norm": 0.141702, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:17.067007+00:00", "epoch": 0, "step": 12077, "train_loss": 3.5231809616088867, "perplexity": 33.89206653930262, "lr": 0.0026291804804649314, "grad_norm": 0.185878, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:17.370094+00:00", "epoch": 0, "step": 12078, "train_loss": 3.5056121349334717, "perplexity": 33.3018228235259, "lr": 0.0026291804804649314, "grad_norm": 0.161081, "tokens_per_sec": 108114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:17.672920+00:00", "epoch": 0, "step": 12079, "train_loss": 3.5074760913848877, "perplexity": 33.363953857804155, "lr": 0.0026291804804649314, "grad_norm": 0.166186, "tokens_per_sec": 108207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:17.977338+00:00", "epoch": 0, "step": 12080, "train_loss": 3.5178582668304443, "perplexity": 33.7121486624472, "lr": 0.0026291804804649314, "grad_norm": 0.15409, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:18.281686+00:00", "epoch": 0, "step": 12081, "train_loss": 3.4576780796051025, "perplexity": 31.743185729311488, "lr": 0.0026291804804649314, "grad_norm": 0.176517, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:18.586943+00:00", "epoch": 0, "step": 12082, "train_loss": 3.4806628227233887, "perplexity": 32.48124424915537, "lr": 0.0026291804804649314, "grad_norm": 0.162765, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:18.891031+00:00", "epoch": 0, "step": 12083, "train_loss": 3.529873847961426, "perplexity": 34.119663078194264, "lr": 0.0026291804804649314, "grad_norm": 0.156745, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:19.193841+00:00", "epoch": 0, "step": 12084, "train_loss": 3.487311601638794, "perplexity": 32.69792439203029, "lr": 0.0026291804804649314, "grad_norm": 0.180994, "tokens_per_sec": 108214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:19.497135+00:00", "epoch": 0, "step": 12085, "train_loss": 3.5186820030212402, "perplexity": 33.7399300200539, "lr": 0.0026291804804649314, "grad_norm": 0.178086, "tokens_per_sec": 108040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:19.801310+00:00", "epoch": 0, "step": 12086, "train_loss": 3.498913526535034, "perplexity": 33.07949243692924, "lr": 0.0026291804804649314, "grad_norm": 0.165917, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:20.105631+00:00", "epoch": 0, "step": 12087, "train_loss": 3.3842668533325195, "perplexity": 29.49635962924271, "lr": 0.0026291804804649314, "grad_norm": 0.176413, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:20.409782+00:00", "epoch": 0, "step": 12088, "train_loss": 3.512225866317749, "perplexity": 33.522802077505794, "lr": 0.0026291804804649314, "grad_norm": 0.187601, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:20.713477+00:00", "epoch": 0, "step": 12089, "train_loss": 3.5200724601745605, "perplexity": 33.786876578127796, "lr": 0.0026291804804649314, "grad_norm": 0.169585, "tokens_per_sec": 107898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:21.017167+00:00", "epoch": 0, "step": 12090, "train_loss": 3.440115213394165, "perplexity": 31.190551530611476, "lr": 0.0026291804804649314, "grad_norm": 0.172278, "tokens_per_sec": 107961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:21.321401+00:00", "epoch": 0, "step": 12091, "train_loss": 3.4961752891540527, "perplexity": 32.989036835168825, "lr": 0.0026291804804649314, "grad_norm": 0.181518, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:21.626066+00:00", "epoch": 0, "step": 12092, "train_loss": 3.434295177459717, "perplexity": 31.009548632496593, "lr": 0.0026291804804649314, "grad_norm": 0.152805, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:21.930772+00:00", "epoch": 0, "step": 12093, "train_loss": 3.5332372188568115, "perplexity": 34.234613361808684, "lr": 0.0026291804804649314, "grad_norm": 0.161075, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:22.234471+00:00", "epoch": 0, "step": 12094, "train_loss": 3.5678300857543945, "perplexity": 35.439608745180436, "lr": 0.0026291804804649314, "grad_norm": 0.182711, "tokens_per_sec": 107896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:22.538260+00:00", "epoch": 0, "step": 12095, "train_loss": 3.5598459243774414, "perplexity": 35.157779770968695, "lr": 0.0026291804804649314, "grad_norm": 0.158104, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:22.842963+00:00", "epoch": 0, "step": 12096, "train_loss": 3.517021417617798, "perplexity": 33.68394847867896, "lr": 0.0026291804804649314, "grad_norm": 0.161892, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:23.147821+00:00", "epoch": 0, "step": 12097, "train_loss": 3.4860470294952393, "perplexity": 32.656601640998744, "lr": 0.0026291804804649314, "grad_norm": 0.173433, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:23.453328+00:00", "epoch": 0, "step": 12098, "train_loss": 3.5819265842437744, "perplexity": 35.94272086479507, "lr": 0.0026291804804649314, "grad_norm": 0.168343, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:23.758003+00:00", "epoch": 0, "step": 12099, "train_loss": 3.433912754058838, "perplexity": 30.99769212269673, "lr": 0.0026291804804649314, "grad_norm": 0.189528, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:24.062049+00:00", "epoch": 0, "step": 12100, "train_loss": 3.4640564918518066, "perplexity": 31.946303951153165, "lr": 0.0026291804804649314, "grad_norm": 0.157865, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:24.365284+00:00", "epoch": 0, "step": 12101, "train_loss": 3.4813427925109863, "perplexity": 32.503338024607004, "lr": 0.0026291804804649314, "grad_norm": 0.190396, "tokens_per_sec": 108007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:24.669692+00:00", "epoch": 0, "step": 12102, "train_loss": 3.4447717666625977, "perplexity": 31.336130680764068, "lr": 0.0026291804804649314, "grad_norm": 0.16533, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:24.973793+00:00", "epoch": 0, "step": 12103, "train_loss": 3.440108299255371, "perplexity": 31.190335875554666, "lr": 0.0026291804804649314, "grad_norm": 0.180837, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:25.277668+00:00", "epoch": 0, "step": 12104, "train_loss": 3.4735381603240967, "perplexity": 32.250648782430446, "lr": 0.0026291804804649314, "grad_norm": 0.172411, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:25.581664+00:00", "epoch": 0, "step": 12105, "train_loss": 3.620288848876953, "perplexity": 37.34835429434242, "lr": 0.0026291804804649314, "grad_norm": 0.17223, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:25.885979+00:00", "epoch": 0, "step": 12106, "train_loss": 3.5032308101654053, "perplexity": 33.222614715559644, "lr": 0.0026291804804649314, "grad_norm": 0.143013, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:26.189920+00:00", "epoch": 0, "step": 12107, "train_loss": 3.4845409393310547, "perplexity": 32.60745487349311, "lr": 0.0026291804804649314, "grad_norm": 0.154727, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:26.494377+00:00", "epoch": 0, "step": 12108, "train_loss": 3.50239896774292, "perplexity": 33.19499022645725, "lr": 0.0026291804804649314, "grad_norm": 0.171368, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:26.798234+00:00", "epoch": 0, "step": 12109, "train_loss": 3.4755215644836426, "perplexity": 32.31467833061924, "lr": 0.0026291804804649314, "grad_norm": 0.160642, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:27.102908+00:00", "epoch": 0, "step": 12110, "train_loss": 3.4184091091156006, "perplexity": 30.520821081428593, "lr": 0.0026291804804649314, "grad_norm": 0.161963, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:27.406374+00:00", "epoch": 0, "step": 12111, "train_loss": 3.544445276260376, "perplexity": 34.620475215117786, "lr": 0.0026291804804649314, "grad_norm": 0.148337, "tokens_per_sec": 107921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:27.710461+00:00", "epoch": 0, "step": 12112, "train_loss": 3.5113954544067383, "perplexity": 33.494975898564434, "lr": 0.0026291804804649314, "grad_norm": 0.1626, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:28.015437+00:00", "epoch": 0, "step": 12113, "train_loss": 3.4997353553771973, "perplexity": 33.10668929194723, "lr": 0.0026291804804649314, "grad_norm": 0.169951, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:28.320714+00:00", "epoch": 0, "step": 12114, "train_loss": 3.58178448677063, "perplexity": 35.93761385783723, "lr": 0.0026291804804649314, "grad_norm": 0.148022, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:28.624264+00:00", "epoch": 0, "step": 12115, "train_loss": 3.5255284309387207, "perplexity": 33.97172058216108, "lr": 0.0026291804804649314, "grad_norm": 0.17107, "tokens_per_sec": 107949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:28.928451+00:00", "epoch": 0, "step": 12116, "train_loss": 3.4803411960601807, "perplexity": 32.470799094760785, "lr": 0.0026291804804649314, "grad_norm": 0.178879, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:29.232387+00:00", "epoch": 0, "step": 12117, "train_loss": 3.4988813400268555, "perplexity": 33.07842774070989, "lr": 0.0026291804804649314, "grad_norm": 0.20071, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:29.537601+00:00", "epoch": 0, "step": 12118, "train_loss": 3.5215041637420654, "perplexity": 33.83528401414027, "lr": 0.0026291804804649314, "grad_norm": 0.187609, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:29.842329+00:00", "epoch": 0, "step": 12119, "train_loss": 3.595878839492798, "perplexity": 36.44771761030105, "lr": 0.0026291804804649314, "grad_norm": 0.158621, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:30.146394+00:00", "epoch": 0, "step": 12120, "train_loss": 3.546283721923828, "perplexity": 34.68418162008229, "lr": 0.0026291804804649314, "grad_norm": 0.168209, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:30.449786+00:00", "epoch": 0, "step": 12121, "train_loss": 3.5010125637054443, "perplexity": 33.14900044555876, "lr": 0.0026291804804649314, "grad_norm": 0.154793, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:30.754284+00:00", "epoch": 0, "step": 12122, "train_loss": 3.4745495319366455, "perplexity": 32.28328267280875, "lr": 0.0026291804804649314, "grad_norm": 0.168423, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:31.059682+00:00", "epoch": 0, "step": 12123, "train_loss": 3.6108946800231934, "perplexity": 36.99914040300723, "lr": 0.0026291804804649314, "grad_norm": 0.174611, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:31.365568+00:00", "epoch": 0, "step": 12124, "train_loss": 3.478459119796753, "perplexity": 32.40974404767993, "lr": 0.0026291804804649314, "grad_norm": 0.19079, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:31.669806+00:00", "epoch": 0, "step": 12125, "train_loss": 3.508395195007324, "perplexity": 33.39463288509655, "lr": 0.0026291804804649314, "grad_norm": 0.164459, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:31.974145+00:00", "epoch": 0, "step": 12126, "train_loss": 3.563340187072754, "perplexity": 35.28084517548073, "lr": 0.0026291804804649314, "grad_norm": 0.196186, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:32.277969+00:00", "epoch": 0, "step": 12127, "train_loss": 3.56589674949646, "perplexity": 35.3711582548549, "lr": 0.0026291804804649314, "grad_norm": 0.17392, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:32.582605+00:00", "epoch": 0, "step": 12128, "train_loss": 3.4742214679718018, "perplexity": 32.272693428166704, "lr": 0.0026291804804649314, "grad_norm": 0.173677, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:32.889081+00:00", "epoch": 0, "step": 12129, "train_loss": 3.485318422317505, "perplexity": 32.632816472717096, "lr": 0.0026291804804649314, "grad_norm": 0.172228, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:33.193654+00:00", "epoch": 0, "step": 12130, "train_loss": 3.5753488540649414, "perplexity": 35.707075201479206, "lr": 0.0026291804804649314, "grad_norm": 0.151133, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:33.497598+00:00", "epoch": 0, "step": 12131, "train_loss": 3.3807733058929443, "perplexity": 29.39349248782579, "lr": 0.0026291804804649314, "grad_norm": 0.158147, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:33.801660+00:00", "epoch": 0, "step": 12132, "train_loss": 3.4669926166534424, "perplexity": 32.04024012318301, "lr": 0.0026291804804649314, "grad_norm": 0.17016, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:34.105364+00:00", "epoch": 0, "step": 12133, "train_loss": 3.6313259601593018, "perplexity": 37.76285547694702, "lr": 0.0026291804804649314, "grad_norm": 0.192482, "tokens_per_sec": 107895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:34.410338+00:00", "epoch": 0, "step": 12134, "train_loss": 3.510356903076172, "perplexity": 33.460207704180014, "lr": 0.0026291804804649314, "grad_norm": 0.160594, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:34.715899+00:00", "epoch": 0, "step": 12135, "train_loss": 3.477138042449951, "perplexity": 32.36695653803065, "lr": 0.0026291804804649314, "grad_norm": 0.157722, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:35.019851+00:00", "epoch": 0, "step": 12136, "train_loss": 3.479616165161133, "perplexity": 32.447265294497406, "lr": 0.0026291804804649314, "grad_norm": 0.163254, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:35.324414+00:00", "epoch": 0, "step": 12137, "train_loss": 3.5700459480285645, "perplexity": 35.51822510655739, "lr": 0.0026291804804649314, "grad_norm": 0.177437, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:35.628270+00:00", "epoch": 0, "step": 12138, "train_loss": 3.5514111518859863, "perplexity": 34.86247904528636, "lr": 0.0026291804804649314, "grad_norm": 0.202541, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:35.933607+00:00", "epoch": 0, "step": 12139, "train_loss": 3.5205485820770264, "perplexity": 33.80296708029884, "lr": 0.0026291804804649314, "grad_norm": 0.182478, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:36.239984+00:00", "epoch": 0, "step": 12140, "train_loss": 3.535224676132202, "perplexity": 34.302720851148756, "lr": 0.0026291804804649314, "grad_norm": 0.171089, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:36.545540+00:00", "epoch": 0, "step": 12141, "train_loss": 3.4288430213928223, "perplexity": 30.840939792315798, "lr": 0.0026291804804649314, "grad_norm": 0.146307, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:36.849019+00:00", "epoch": 0, "step": 12142, "train_loss": 3.3895010948181152, "perplexity": 29.651155464464107, "lr": 0.0026291804804649314, "grad_norm": 0.147187, "tokens_per_sec": 107975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:37.152199+00:00", "epoch": 0, "step": 12143, "train_loss": 3.468290328979492, "perplexity": 32.08184612819432, "lr": 0.0026291804804649314, "grad_norm": 0.172921, "tokens_per_sec": 108080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:37.455982+00:00", "epoch": 0, "step": 12144, "train_loss": 3.6361122131347656, "perplexity": 37.944031287098, "lr": 0.0026291804804649314, "grad_norm": 0.172566, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:37.760451+00:00", "epoch": 0, "step": 12145, "train_loss": 3.573061466217041, "perplexity": 35.62549261266139, "lr": 0.0026291804804649314, "grad_norm": 0.1598, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:38.064651+00:00", "epoch": 0, "step": 12146, "train_loss": 3.530553102493286, "perplexity": 34.142846886927856, "lr": 0.0026291804804649314, "grad_norm": 0.168793, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:38.368984+00:00", "epoch": 0, "step": 12147, "train_loss": 3.4895143508911133, "perplexity": 32.77002910568431, "lr": 0.0026291804804649314, "grad_norm": 0.172243, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:38.672339+00:00", "epoch": 0, "step": 12148, "train_loss": 3.535390615463257, "perplexity": 34.308413494003766, "lr": 0.0026291804804649314, "grad_norm": 0.173075, "tokens_per_sec": 108018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:38.975933+00:00", "epoch": 0, "step": 12149, "train_loss": 3.528895378112793, "perplexity": 34.08629434443895, "lr": 0.0026291804804649314, "grad_norm": 0.163899, "tokens_per_sec": 107934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:39.280413+00:00", "epoch": 0, "step": 12150, "train_loss": 3.5252580642700195, "perplexity": 33.962537002760065, "lr": 0.0026291804804649314, "grad_norm": 0.184572, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:39.584799+00:00", "epoch": 0, "step": 12151, "train_loss": 3.4947669506073, "perplexity": 32.9426098031367, "lr": 0.0026291804804649314, "grad_norm": 0.158688, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:39.889107+00:00", "epoch": 0, "step": 12152, "train_loss": 3.5252976417541504, "perplexity": 33.96388118112876, "lr": 0.0026291804804649314, "grad_norm": 0.15814, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:40.192684+00:00", "epoch": 0, "step": 12153, "train_loss": 3.498091697692871, "perplexity": 33.05231792388923, "lr": 0.0026291804804649314, "grad_norm": 0.163068, "tokens_per_sec": 107942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:40.495852+00:00", "epoch": 0, "step": 12154, "train_loss": 3.5029869079589844, "perplexity": 33.21451263462532, "lr": 0.0026291804804649314, "grad_norm": 0.152671, "tokens_per_sec": 108082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:40.799672+00:00", "epoch": 0, "step": 12155, "train_loss": 3.5734403133392334, "perplexity": 35.63899178491421, "lr": 0.0026291804804649314, "grad_norm": 0.160628, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:41.104008+00:00", "epoch": 0, "step": 12156, "train_loss": 3.5436863899230957, "perplexity": 34.59421217607806, "lr": 0.0026291804804649314, "grad_norm": 0.175102, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:41.408205+00:00", "epoch": 0, "step": 12157, "train_loss": 3.467209815979004, "perplexity": 32.04719999754124, "lr": 0.0026291804804649314, "grad_norm": 0.177551, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:41.712296+00:00", "epoch": 0, "step": 12158, "train_loss": 3.5749027729034424, "perplexity": 35.69115050001833, "lr": 0.0026291804804649314, "grad_norm": 0.16267, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:42.016513+00:00", "epoch": 0, "step": 12159, "train_loss": 3.483903169631958, "perplexity": 32.58666545694477, "lr": 0.0026291804804649314, "grad_norm": 0.172658, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:42.321063+00:00", "epoch": 0, "step": 12160, "train_loss": 3.5203816890716553, "perplexity": 33.79732607226783, "lr": 0.0026291804804649314, "grad_norm": 0.146645, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:42.626807+00:00", "epoch": 0, "step": 12161, "train_loss": 3.5597736835479736, "perplexity": 35.1552400355332, "lr": 0.0026291804804649314, "grad_norm": 0.149593, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:42.931229+00:00", "epoch": 0, "step": 12162, "train_loss": 3.4507675170898438, "perplexity": 31.524578678720513, "lr": 0.0026291804804649314, "grad_norm": 0.158125, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:43.234863+00:00", "epoch": 0, "step": 12163, "train_loss": 3.4750030040740967, "perplexity": 32.297925561826496, "lr": 0.0026291804804649314, "grad_norm": 0.162085, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:43.539550+00:00", "epoch": 0, "step": 12164, "train_loss": 3.4725871086120605, "perplexity": 32.21999132841451, "lr": 0.0026291804804649314, "grad_norm": 0.170769, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:43.843982+00:00", "epoch": 0, "step": 12165, "train_loss": 3.6436660289764404, "perplexity": 38.23173878842104, "lr": 0.0026291804804649314, "grad_norm": 0.16784, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:44.148985+00:00", "epoch": 0, "step": 12166, "train_loss": 3.5486013889312744, "perplexity": 34.76466122994649, "lr": 0.0026291804804649314, "grad_norm": 0.151542, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:44.453954+00:00", "epoch": 0, "step": 12167, "train_loss": 3.516045331954956, "perplexity": 33.6510861003583, "lr": 0.0026291804804649314, "grad_norm": 0.185831, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:44.758756+00:00", "epoch": 0, "step": 12168, "train_loss": 3.531409740447998, "perplexity": 34.172107476529575, "lr": 0.0026291804804649314, "grad_norm": 0.173529, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:45.063097+00:00", "epoch": 0, "step": 12169, "train_loss": 3.53083872795105, "perplexity": 34.152600346051095, "lr": 0.0026291804804649314, "grad_norm": 0.143886, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:45.367858+00:00", "epoch": 0, "step": 12170, "train_loss": 3.5567567348480225, "perplexity": 35.04933831017411, "lr": 0.0026291804804649314, "grad_norm": 0.166918, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:45.672620+00:00", "epoch": 0, "step": 12171, "train_loss": 3.554499626159668, "perplexity": 34.97031735714091, "lr": 0.0026291804804649314, "grad_norm": 0.196697, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:45.976626+00:00", "epoch": 0, "step": 12172, "train_loss": 3.6528820991516113, "perplexity": 38.58571379985666, "lr": 0.0026291804804649314, "grad_norm": 0.171394, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:46.281777+00:00", "epoch": 0, "step": 12173, "train_loss": 3.5760622024536133, "perplexity": 35.732555873255926, "lr": 0.0026291804804649314, "grad_norm": 0.180928, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:46.586282+00:00", "epoch": 0, "step": 12174, "train_loss": 3.4483017921447754, "perplexity": 31.446943491532313, "lr": 0.0026291804804649314, "grad_norm": 0.164389, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:46.890365+00:00", "epoch": 0, "step": 12175, "train_loss": 3.489598035812378, "perplexity": 32.77277157774004, "lr": 0.0026291804804649314, "grad_norm": 0.151804, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:47.195318+00:00", "epoch": 0, "step": 12176, "train_loss": 3.5049118995666504, "perplexity": 33.27851187193413, "lr": 0.0026291804804649314, "grad_norm": 0.161417, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:47.499773+00:00", "epoch": 0, "step": 12177, "train_loss": 3.5750787258148193, "perplexity": 35.69743101438039, "lr": 0.0026291804804649314, "grad_norm": 0.156839, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:47.804104+00:00", "epoch": 0, "step": 12178, "train_loss": 3.5649502277374268, "perplexity": 35.33769452349776, "lr": 0.0026291804804649314, "grad_norm": 0.156152, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:48.107735+00:00", "epoch": 0, "step": 12179, "train_loss": 3.646700859069824, "perplexity": 38.34794185891845, "lr": 0.0026291804804649314, "grad_norm": 0.156319, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:48.411481+00:00", "epoch": 0, "step": 12180, "train_loss": 3.5493741035461426, "perplexity": 34.791534773212476, "lr": 0.0026291804804649314, "grad_norm": 0.162828, "tokens_per_sec": 107965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:48.715100+00:00", "epoch": 0, "step": 12181, "train_loss": 3.459676742553711, "perplexity": 31.806693102275283, "lr": 0.0026291804804649314, "grad_norm": 0.181172, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:49.020155+00:00", "epoch": 0, "step": 12182, "train_loss": 3.427328109741211, "perplexity": 30.794253864734124, "lr": 0.0026291804804649314, "grad_norm": 0.171933, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:49.324390+00:00", "epoch": 0, "step": 12183, "train_loss": 3.5115866661071777, "perplexity": 33.501381142221426, "lr": 0.0026291804804649314, "grad_norm": 0.164429, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:49.628664+00:00", "epoch": 0, "step": 12184, "train_loss": 3.5524070262908936, "perplexity": 34.89721498930714, "lr": 0.0026291804804649314, "grad_norm": 0.151841, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:49.932885+00:00", "epoch": 0, "step": 12185, "train_loss": 3.483100652694702, "perplexity": 32.56052459663143, "lr": 0.0026291804804649314, "grad_norm": 0.17554, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:50.236547+00:00", "epoch": 0, "step": 12186, "train_loss": 3.487572193145752, "perplexity": 32.70644630374269, "lr": 0.0026291804804649314, "grad_norm": 0.163867, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:50.541045+00:00", "epoch": 0, "step": 12187, "train_loss": 3.446934461593628, "perplexity": 31.403974508052606, "lr": 0.0026291804804649314, "grad_norm": 0.187768, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:50.845844+00:00", "epoch": 0, "step": 12188, "train_loss": 3.4592511653900146, "perplexity": 31.793159779978975, "lr": 0.0026291804804649314, "grad_norm": 0.14922, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:51.150084+00:00", "epoch": 0, "step": 12189, "train_loss": 3.6054093837738037, "perplexity": 36.79674476442367, "lr": 0.0026291804804649314, "grad_norm": 0.148431, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:51.454933+00:00", "epoch": 0, "step": 12190, "train_loss": 3.457174062728882, "perplexity": 31.727190659230143, "lr": 0.0026291804804649314, "grad_norm": 0.16512, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:51.758390+00:00", "epoch": 0, "step": 12191, "train_loss": 3.54891300201416, "perplexity": 34.775496041254634, "lr": 0.0026291804804649314, "grad_norm": 0.168705, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:52.063551+00:00", "epoch": 0, "step": 12192, "train_loss": 3.5415518283843994, "perplexity": 34.520447456979824, "lr": 0.0026291804804649314, "grad_norm": 0.174684, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:52.368880+00:00", "epoch": 0, "step": 12193, "train_loss": 3.5256612300872803, "perplexity": 33.976232297298836, "lr": 0.0026291804804649314, "grad_norm": 0.164841, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:52.673314+00:00", "epoch": 0, "step": 12194, "train_loss": 3.5742580890655518, "perplexity": 35.668148407470575, "lr": 0.0026291804804649314, "grad_norm": 0.169512, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:52.977504+00:00", "epoch": 0, "step": 12195, "train_loss": 3.6087260246276855, "perplexity": 36.91898895938656, "lr": 0.0026291804804649314, "grad_norm": 0.164619, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:53.282283+00:00", "epoch": 0, "step": 12196, "train_loss": 3.446809768676758, "perplexity": 31.400058898999315, "lr": 0.0026291804804649314, "grad_norm": 0.194599, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:53.586294+00:00", "epoch": 0, "step": 12197, "train_loss": 3.517160654067993, "perplexity": 33.68863883862065, "lr": 0.0026291804804649314, "grad_norm": 0.162975, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:53.892785+00:00", "epoch": 0, "step": 12198, "train_loss": 3.484262704849243, "perplexity": 32.598383617209855, "lr": 0.0026291804804649314, "grad_norm": 0.166677, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:54.198098+00:00", "epoch": 0, "step": 12199, "train_loss": 3.5109620094299316, "perplexity": 33.48046081548512, "lr": 0.0026291804804649314, "grad_norm": 0.183938, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:54.502648+00:00", "epoch": 0, "step": 12200, "train_loss": 3.6376075744628906, "perplexity": 38.000813768700475, "lr": 0.0026291804804649314, "grad_norm": 0.160493, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:54.807259+00:00", "epoch": 0, "step": 12201, "train_loss": 3.5337276458740234, "perplexity": 34.25140705882989, "lr": 0.0026291804804649314, "grad_norm": 0.160427, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:55.111581+00:00", "epoch": 0, "step": 12202, "train_loss": 3.5360875129699707, "perplexity": 34.33233127498001, "lr": 0.0026291804804649314, "grad_norm": 0.179733, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:55.417679+00:00", "epoch": 0, "step": 12203, "train_loss": 3.6332807540893555, "perplexity": 37.83674627472693, "lr": 0.0026291804804649314, "grad_norm": 0.163355, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:55.723142+00:00", "epoch": 0, "step": 12204, "train_loss": 3.5223746299743652, "perplexity": 33.864749308746326, "lr": 0.0026291804804649314, "grad_norm": 0.163004, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:56.028063+00:00", "epoch": 0, "step": 12205, "train_loss": 3.484177589416504, "perplexity": 32.595609109760055, "lr": 0.0026291804804649314, "grad_norm": 0.183749, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:56.333834+00:00", "epoch": 0, "step": 12206, "train_loss": 3.524515151977539, "perplexity": 33.937315186495596, "lr": 0.0026291804804649314, "grad_norm": 0.19315, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:56.638476+00:00", "epoch": 0, "step": 12207, "train_loss": 3.516968011856079, "perplexity": 33.68214960978817, "lr": 0.0026291804804649314, "grad_norm": 0.204449, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:56.943975+00:00", "epoch": 0, "step": 12208, "train_loss": 3.612210512161255, "perplexity": 37.047857105502494, "lr": 0.0026291804804649314, "grad_norm": 0.171909, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:57.249395+00:00", "epoch": 0, "step": 12209, "train_loss": 3.439810037612915, "perplexity": 31.181034381956003, "lr": 0.0026291804804649314, "grad_norm": 0.171128, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:57.554275+00:00", "epoch": 0, "step": 12210, "train_loss": 3.650000810623169, "perplexity": 38.47469723750048, "lr": 0.0026291804804649314, "grad_norm": 0.165515, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:57.859145+00:00", "epoch": 0, "step": 12211, "train_loss": 3.574737548828125, "perplexity": 35.6852539498194, "lr": 0.0026291804804649314, "grad_norm": 0.170797, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:58.164399+00:00", "epoch": 0, "step": 12212, "train_loss": 3.4717636108398438, "perplexity": 32.193469159307135, "lr": 0.0026291804804649314, "grad_norm": 0.144627, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:58.468001+00:00", "epoch": 0, "step": 12213, "train_loss": 3.509852409362793, "perplexity": 33.44333149707303, "lr": 0.0026291804804649314, "grad_norm": 0.162077, "tokens_per_sec": 107932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:58.772653+00:00", "epoch": 0, "step": 12214, "train_loss": 3.585146188735962, "perplexity": 36.058628698920906, "lr": 0.0026291804804649314, "grad_norm": 0.152812, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:59.077695+00:00", "epoch": 0, "step": 12215, "train_loss": 3.4233102798461914, "perplexity": 30.670776013570556, "lr": 0.0026291804804649314, "grad_norm": 0.15624, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:59.382453+00:00", "epoch": 0, "step": 12216, "train_loss": 3.464336633682251, "perplexity": 31.955254700899143, "lr": 0.0026291804804649314, "grad_norm": 0.166644, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:59.686346+00:00", "epoch": 0, "step": 12217, "train_loss": 3.52860689163208, "perplexity": 34.076462327613136, "lr": 0.0026291804804649314, "grad_norm": 0.165677, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:01:59.991447+00:00", "epoch": 0, "step": 12218, "train_loss": 3.5599687099456787, "perplexity": 35.162096903971225, "lr": 0.0026291804804649314, "grad_norm": 0.146072, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:00.296791+00:00", "epoch": 0, "step": 12219, "train_loss": 3.509281873703003, "perplexity": 33.42425632592188, "lr": 0.0026291804804649314, "grad_norm": 0.149828, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:00.601987+00:00", "epoch": 0, "step": 12220, "train_loss": 3.4765264987945557, "perplexity": 32.347168782270295, "lr": 0.0026291804804649314, "grad_norm": 0.151205, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:00.907638+00:00", "epoch": 0, "step": 12221, "train_loss": 3.517521619796753, "perplexity": 33.70080147770583, "lr": 0.0026291804804649314, "grad_norm": 0.143452, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:01.212762+00:00", "epoch": 0, "step": 12222, "train_loss": 3.471156597137451, "perplexity": 32.173933212293505, "lr": 0.0026291804804649314, "grad_norm": 0.159538, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:01.517756+00:00", "epoch": 0, "step": 12223, "train_loss": 3.5776631832122803, "perplexity": 35.789808825875824, "lr": 0.0026291804804649314, "grad_norm": 0.156094, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:01.822925+00:00", "epoch": 0, "step": 12224, "train_loss": 3.5734829902648926, "perplexity": 35.640512779972646, "lr": 0.0026291804804649314, "grad_norm": 0.161894, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:02.128559+00:00", "epoch": 0, "step": 12225, "train_loss": 3.5306692123413086, "perplexity": 34.1468114378482, "lr": 0.0026291804804649314, "grad_norm": 0.174777, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:02.433141+00:00", "epoch": 0, "step": 12226, "train_loss": 3.5602262020111084, "perplexity": 35.17115203068948, "lr": 0.0026291804804649314, "grad_norm": 0.163605, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:02.737519+00:00", "epoch": 0, "step": 12227, "train_loss": 3.432335615158081, "perplexity": 30.948842987677374, "lr": 0.0026291804804649314, "grad_norm": 0.164352, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:03.042726+00:00", "epoch": 0, "step": 12228, "train_loss": 3.5164387226104736, "perplexity": 33.66432672737691, "lr": 0.0026291804804649314, "grad_norm": 0.185959, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:03.346261+00:00", "epoch": 0, "step": 12229, "train_loss": 3.5017154216766357, "perplexity": 33.17230767463532, "lr": 0.0026291804804649314, "grad_norm": 0.195975, "tokens_per_sec": 107945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:03.651920+00:00", "epoch": 0, "step": 12230, "train_loss": 3.6120922565460205, "perplexity": 37.043476247403014, "lr": 0.0026291804804649314, "grad_norm": 0.148701, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:03.958101+00:00", "epoch": 0, "step": 12231, "train_loss": 3.482773542404175, "perplexity": 32.549875455787635, "lr": 0.0026291804804649314, "grad_norm": 0.186216, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:04.263022+00:00", "epoch": 0, "step": 12232, "train_loss": 3.4223134517669678, "perplexity": 30.64021775602363, "lr": 0.0026291804804649314, "grad_norm": 0.162356, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:04.567913+00:00", "epoch": 0, "step": 12233, "train_loss": 3.524817705154419, "perplexity": 33.94758458246089, "lr": 0.0026291804804649314, "grad_norm": 0.156244, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:04.872025+00:00", "epoch": 0, "step": 12234, "train_loss": 3.490403652191162, "perplexity": 32.79918449721295, "lr": 0.0026291804804649314, "grad_norm": 0.156842, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:05.177987+00:00", "epoch": 0, "step": 12235, "train_loss": 3.5325558185577393, "perplexity": 34.211293831894935, "lr": 0.0026291804804649314, "grad_norm": 0.183391, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:05.482365+00:00", "epoch": 0, "step": 12236, "train_loss": 3.516453981399536, "perplexity": 33.66484040815644, "lr": 0.0026291804804649314, "grad_norm": 0.200448, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:05.785796+00:00", "epoch": 0, "step": 12237, "train_loss": 3.5486559867858887, "perplexity": 34.76655935768242, "lr": 0.0026291804804649314, "grad_norm": 0.214889, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:06.090469+00:00", "epoch": 0, "step": 12238, "train_loss": 3.479602575302124, "perplexity": 32.44682434373307, "lr": 0.0026291804804649314, "grad_norm": 0.186534, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:06.395690+00:00", "epoch": 0, "step": 12239, "train_loss": 3.5251402854919434, "perplexity": 33.95853717220371, "lr": 0.0026291804804649314, "grad_norm": 0.194412, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:06.700552+00:00", "epoch": 0, "step": 12240, "train_loss": 3.542948007583618, "perplexity": 34.568677848967326, "lr": 0.0026291804804649314, "grad_norm": 0.209147, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:07.005988+00:00", "epoch": 0, "step": 12241, "train_loss": 3.4981701374053955, "perplexity": 33.05491063989006, "lr": 0.0026291804804649314, "grad_norm": 0.17381, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:07.310913+00:00", "epoch": 0, "step": 12242, "train_loss": 3.5848608016967773, "perplexity": 36.04833950191104, "lr": 0.0026291804804649314, "grad_norm": 0.170454, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:07.615407+00:00", "epoch": 0, "step": 12243, "train_loss": 3.6312665939331055, "perplexity": 37.76061370527037, "lr": 0.0026291804804649314, "grad_norm": 0.161676, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:07.920256+00:00", "epoch": 0, "step": 12244, "train_loss": 3.4588100910186768, "perplexity": 31.779139724187118, "lr": 0.0026291804804649314, "grad_norm": 0.180369, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:08.225691+00:00", "epoch": 0, "step": 12245, "train_loss": 3.565910816192627, "perplexity": 35.37165581369064, "lr": 0.0026291804804649314, "grad_norm": 0.210769, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:08.531380+00:00", "epoch": 0, "step": 12246, "train_loss": 3.5919103622436523, "perplexity": 36.303362297021565, "lr": 0.0026291804804649314, "grad_norm": 0.167034, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:08.835816+00:00", "epoch": 0, "step": 12247, "train_loss": 3.576843738555908, "perplexity": 35.76049307124056, "lr": 0.0026291804804649314, "grad_norm": 0.155541, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:09.140197+00:00", "epoch": 0, "step": 12248, "train_loss": 3.494905710220337, "perplexity": 32.94718122408238, "lr": 0.0026291804804649314, "grad_norm": 0.155061, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:09.444370+00:00", "epoch": 0, "step": 12249, "train_loss": 3.4808928966522217, "perplexity": 32.48871819638038, "lr": 0.0026291804804649314, "grad_norm": 0.155328, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:09.749508+00:00", "epoch": 0, "step": 12250, "train_loss": 3.513439893722534, "perplexity": 33.56352439191433, "lr": 0.0026291804804649314, "grad_norm": 0.147448, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:10.055678+00:00", "epoch": 0, "step": 12251, "train_loss": 3.3962831497192383, "perplexity": 29.852934694189077, "lr": 0.0026291804804649314, "grad_norm": 0.155834, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:10.360590+00:00", "epoch": 0, "step": 12252, "train_loss": 3.5121471881866455, "perplexity": 33.520164669843496, "lr": 0.0026291804804649314, "grad_norm": 0.146255, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:10.665896+00:00", "epoch": 0, "step": 12253, "train_loss": 3.5363786220550537, "perplexity": 34.342327183405516, "lr": 0.0026291804804649314, "grad_norm": 0.144363, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:10.971220+00:00", "epoch": 0, "step": 12254, "train_loss": 3.5844178199768066, "perplexity": 36.03237428288756, "lr": 0.0026291804804649314, "grad_norm": 0.15951, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:11.276788+00:00", "epoch": 0, "step": 12255, "train_loss": 3.50272798538208, "perplexity": 33.20591376069268, "lr": 0.0026291804804649314, "grad_norm": 0.184894, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:11.582575+00:00", "epoch": 0, "step": 12256, "train_loss": 3.411956548690796, "perplexity": 30.324517650161354, "lr": 0.0026291804804649314, "grad_norm": 0.164046, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:11.887232+00:00", "epoch": 0, "step": 12257, "train_loss": 3.4057841300964355, "perplexity": 30.13791850998622, "lr": 0.0026291804804649314, "grad_norm": 0.188984, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:12.192324+00:00", "epoch": 0, "step": 12258, "train_loss": 3.5799436569213867, "perplexity": 35.871519678270644, "lr": 0.0026291804804649314, "grad_norm": 0.173706, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:12.496464+00:00", "epoch": 0, "step": 12259, "train_loss": 3.4554238319396973, "perplexity": 31.671709320021115, "lr": 0.0026291804804649314, "grad_norm": 0.159092, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:12.802287+00:00", "epoch": 0, "step": 12260, "train_loss": 3.4821510314941406, "perplexity": 32.52961910874881, "lr": 0.0026291804804649314, "grad_norm": 0.159224, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:13.107879+00:00", "epoch": 0, "step": 12261, "train_loss": 3.433785915374756, "perplexity": 30.993760665554007, "lr": 0.0026291804804649314, "grad_norm": 0.160215, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:13.412488+00:00", "epoch": 0, "step": 12262, "train_loss": 3.557399272918701, "perplexity": 35.07186608109066, "lr": 0.0026291804804649314, "grad_norm": 0.158239, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:13.716937+00:00", "epoch": 0, "step": 12263, "train_loss": 3.5213778018951416, "perplexity": 33.83100879527901, "lr": 0.0026291804804649314, "grad_norm": 0.194023, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:14.021280+00:00", "epoch": 0, "step": 12264, "train_loss": 3.4990785121917725, "perplexity": 33.084950528954586, "lr": 0.0026291804804649314, "grad_norm": 0.163256, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:14.326464+00:00", "epoch": 0, "step": 12265, "train_loss": 3.4612133502960205, "perplexity": 31.855605082793385, "lr": 0.0026291804804649314, "grad_norm": 0.17084, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:14.632798+00:00", "epoch": 0, "step": 12266, "train_loss": 3.561825752258301, "perplexity": 35.22745507338257, "lr": 0.0026291804804649314, "grad_norm": 0.166343, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:14.937442+00:00", "epoch": 0, "step": 12267, "train_loss": 3.4479353427886963, "perplexity": 31.435421890509755, "lr": 0.0026291804804649314, "grad_norm": 0.192018, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:15.242070+00:00", "epoch": 0, "step": 12268, "train_loss": 3.4472038745880127, "perplexity": 31.41243628666578, "lr": 0.0026291804804649314, "grad_norm": 0.148552, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:15.546025+00:00", "epoch": 0, "step": 12269, "train_loss": 3.521662950515747, "perplexity": 33.84065703629661, "lr": 0.0026291804804649314, "grad_norm": 0.193275, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:15.851335+00:00", "epoch": 0, "step": 12270, "train_loss": 3.5505073070526123, "perplexity": 34.8309830096314, "lr": 0.0026291804804649314, "grad_norm": 0.164639, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:16.156341+00:00", "epoch": 0, "step": 12271, "train_loss": 3.511758804321289, "perplexity": 33.507148506519165, "lr": 0.0026291804804649314, "grad_norm": 0.186586, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:16.461006+00:00", "epoch": 0, "step": 12272, "train_loss": 3.558450937271118, "perplexity": 35.10876931392688, "lr": 0.0026291804804649314, "grad_norm": 0.159473, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:16.765928+00:00", "epoch": 0, "step": 12273, "train_loss": 3.48520827293396, "perplexity": 32.629222186057156, "lr": 0.0026291804804649314, "grad_norm": 0.191264, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:17.071117+00:00", "epoch": 0, "step": 12274, "train_loss": 3.4892256259918213, "perplexity": 32.76056894808821, "lr": 0.0026291804804649314, "grad_norm": 0.182407, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:17.376279+00:00", "epoch": 0, "step": 12275, "train_loss": 3.537523031234741, "perplexity": 34.381651355065735, "lr": 0.0026291804804649314, "grad_norm": 0.166487, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:17.680562+00:00", "epoch": 0, "step": 12276, "train_loss": 3.5167367458343506, "perplexity": 33.674360973702804, "lr": 0.0026291804804649314, "grad_norm": 0.162918, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:17.984486+00:00", "epoch": 0, "step": 12277, "train_loss": 3.5342979431152344, "perplexity": 34.27094611278785, "lr": 0.0026291804804649314, "grad_norm": 0.185594, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:18.289112+00:00", "epoch": 0, "step": 12278, "train_loss": 3.5038044452667236, "perplexity": 33.24167784063417, "lr": 0.0026291804804649314, "grad_norm": 0.18068, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:18.594418+00:00", "epoch": 0, "step": 12279, "train_loss": 3.633547067642212, "perplexity": 37.84682405492104, "lr": 0.0026291804804649314, "grad_norm": 0.161522, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:18.899083+00:00", "epoch": 0, "step": 12280, "train_loss": 3.6272494792938232, "perplexity": 37.609229259080955, "lr": 0.0026291804804649314, "grad_norm": 0.162698, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:19.203106+00:00", "epoch": 0, "step": 12281, "train_loss": 3.540118455886841, "perplexity": 34.47100224216277, "lr": 0.0026291804804649314, "grad_norm": 0.181642, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:19.507519+00:00", "epoch": 0, "step": 12282, "train_loss": 3.5667002201080322, "perplexity": 35.39958936125966, "lr": 0.0026291804804649314, "grad_norm": 0.155177, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:19.811554+00:00", "epoch": 0, "step": 12283, "train_loss": 3.48557186126709, "perplexity": 32.64108794755862, "lr": 0.0026291804804649314, "grad_norm": 0.142112, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:20.117557+00:00", "epoch": 0, "step": 12284, "train_loss": 3.5723202228546143, "perplexity": 35.599095237377334, "lr": 0.0026291804804649314, "grad_norm": 0.164606, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:20.421768+00:00", "epoch": 0, "step": 12285, "train_loss": 3.5516772270202637, "perplexity": 34.87175631825067, "lr": 0.0026291804804649314, "grad_norm": 0.146497, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:20.726044+00:00", "epoch": 0, "step": 12286, "train_loss": 3.5057873725891113, "perplexity": 33.307659068236006, "lr": 0.0026291804804649314, "grad_norm": 0.177453, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:21.030494+00:00", "epoch": 0, "step": 12287, "train_loss": 3.4849135875701904, "perplexity": 32.619608248460665, "lr": 0.0026291804804649314, "grad_norm": 0.157009, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:21.334570+00:00", "epoch": 0, "step": 12288, "train_loss": 3.4403645992279053, "perplexity": 31.19833098231196, "lr": 0.0026291804804649314, "grad_norm": 0.177573, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:21.639030+00:00", "epoch": 0, "step": 12289, "train_loss": 3.6179590225219727, "perplexity": 37.261440400647096, "lr": 0.0026291804804649314, "grad_norm": 0.200566, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:21.945552+00:00", "epoch": 0, "step": 12290, "train_loss": 3.524101972579956, "perplexity": 33.92329588349428, "lr": 0.0026291804804649314, "grad_norm": 0.154664, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:22.249784+00:00", "epoch": 0, "step": 12291, "train_loss": 3.597947597503662, "perplexity": 36.52319716563292, "lr": 0.0026291804804649314, "grad_norm": 0.175121, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:22.553831+00:00", "epoch": 0, "step": 12292, "train_loss": 3.4825942516326904, "perplexity": 32.544040086635, "lr": 0.0026291804804649314, "grad_norm": 0.162152, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:22.858724+00:00", "epoch": 0, "step": 12293, "train_loss": 3.4966514110565186, "perplexity": 33.00474737791726, "lr": 0.0026291804804649314, "grad_norm": 0.16944, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:23.163317+00:00", "epoch": 0, "step": 12294, "train_loss": 3.4844257831573486, "perplexity": 32.60370013995011, "lr": 0.0026291804804649314, "grad_norm": 0.17129, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:23.467274+00:00", "epoch": 0, "step": 12295, "train_loss": 3.6513710021972656, "perplexity": 38.52745107665073, "lr": 0.0026291804804649314, "grad_norm": 0.165636, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:23.771558+00:00", "epoch": 0, "step": 12296, "train_loss": 3.4337379932403564, "perplexity": 30.992275413978344, "lr": 0.0026291804804649314, "grad_norm": 0.168577, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:24.075676+00:00", "epoch": 0, "step": 12297, "train_loss": 3.5789146423339844, "perplexity": 35.83462634639319, "lr": 0.0026291804804649314, "grad_norm": 0.15749, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:24.380343+00:00", "epoch": 0, "step": 12298, "train_loss": 3.5502498149871826, "perplexity": 34.822015462460975, "lr": 0.0026291804804649314, "grad_norm": 0.155193, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:24.684904+00:00", "epoch": 0, "step": 12299, "train_loss": 3.594597101211548, "perplexity": 36.40103110170996, "lr": 0.0026291804804649314, "grad_norm": 0.156914, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:24.989702+00:00", "epoch": 0, "step": 12300, "train_loss": 3.4976770877838135, "perplexity": 33.03861694583756, "lr": 0.0026291804804649314, "grad_norm": 0.156898, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:25.294477+00:00", "epoch": 0, "step": 12301, "train_loss": 3.5837595462799072, "perplexity": 36.00866292379853, "lr": 0.0026291804804649314, "grad_norm": 0.172789, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:25.599143+00:00", "epoch": 0, "step": 12302, "train_loss": 3.577359914779663, "perplexity": 35.778956552308706, "lr": 0.0026291804804649314, "grad_norm": 0.160197, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:25.903116+00:00", "epoch": 0, "step": 12303, "train_loss": 3.5100741386413574, "perplexity": 33.45074768500128, "lr": 0.0026291804804649314, "grad_norm": 0.166081, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:26.206968+00:00", "epoch": 0, "step": 12304, "train_loss": 3.5146872997283936, "perplexity": 33.60541785746661, "lr": 0.0026291804804649314, "grad_norm": 0.158972, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:26.511400+00:00", "epoch": 0, "step": 12305, "train_loss": 3.551969289779663, "perplexity": 34.88194254706272, "lr": 0.0026291804804649314, "grad_norm": 0.164876, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:26.816507+00:00", "epoch": 0, "step": 12306, "train_loss": 3.4933929443359375, "perplexity": 32.897377532500045, "lr": 0.0026291804804649314, "grad_norm": 0.157527, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:27.120098+00:00", "epoch": 0, "step": 12307, "train_loss": 3.56441068649292, "perplexity": 35.31863352237708, "lr": 0.0026291804804649314, "grad_norm": 0.15651, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:27.424527+00:00", "epoch": 0, "step": 12308, "train_loss": 3.6049585342407227, "perplexity": 36.78015870841648, "lr": 0.0026291804804649314, "grad_norm": 0.186446, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:27.729166+00:00", "epoch": 0, "step": 12309, "train_loss": 3.573650360107422, "perplexity": 35.646478426205, "lr": 0.0026291804804649314, "grad_norm": 0.186157, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:28.034588+00:00", "epoch": 0, "step": 12310, "train_loss": 3.5626204013824463, "perplexity": 35.255459665141004, "lr": 0.0026291804804649314, "grad_norm": 0.174246, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:28.340302+00:00", "epoch": 0, "step": 12311, "train_loss": 3.450482130050659, "perplexity": 31.515583256199406, "lr": 0.0026291804804649314, "grad_norm": 0.182359, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:28.645698+00:00", "epoch": 0, "step": 12312, "train_loss": 3.521968126296997, "perplexity": 33.850985961234386, "lr": 0.0026291804804649314, "grad_norm": 0.159994, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:28.951481+00:00", "epoch": 0, "step": 12313, "train_loss": 3.5049898624420166, "perplexity": 33.28110646154709, "lr": 0.0026291804804649314, "grad_norm": 0.149677, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:29.256188+00:00", "epoch": 0, "step": 12314, "train_loss": 3.416088342666626, "perplexity": 30.450071512235397, "lr": 0.0026291804804649314, "grad_norm": 0.146379, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:29.561085+00:00", "epoch": 0, "step": 12315, "train_loss": 3.52705717086792, "perplexity": 34.023694224847986, "lr": 0.0026291804804649314, "grad_norm": 0.182259, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:29.865368+00:00", "epoch": 0, "step": 12316, "train_loss": 3.5834896564483643, "perplexity": 35.99894586315485, "lr": 0.0026291804804649314, "grad_norm": 0.202656, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:30.169740+00:00", "epoch": 0, "step": 12317, "train_loss": 3.443192958831787, "perplexity": 31.28669598646776, "lr": 0.0026291804804649314, "grad_norm": 0.167488, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:30.473224+00:00", "epoch": 0, "step": 12318, "train_loss": 3.592982053756714, "perplexity": 36.3422891573581, "lr": 0.0026291804804649314, "grad_norm": 0.15705, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:30.777043+00:00", "epoch": 0, "step": 12319, "train_loss": 3.4790098667144775, "perplexity": 32.42759853051782, "lr": 0.0026291804804649314, "grad_norm": 0.192583, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:31.080954+00:00", "epoch": 0, "step": 12320, "train_loss": 3.4951393604278564, "perplexity": 32.95488023921578, "lr": 0.0026291804804649314, "grad_norm": 0.196604, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:31.385704+00:00", "epoch": 0, "step": 12321, "train_loss": 3.4646778106689453, "perplexity": 31.96615895844214, "lr": 0.0026291804804649314, "grad_norm": 0.185514, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:31.690680+00:00", "epoch": 0, "step": 12322, "train_loss": 3.5331547260284424, "perplexity": 34.231789368205554, "lr": 0.0026291804804649314, "grad_norm": 0.156108, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:31.994654+00:00", "epoch": 0, "step": 12323, "train_loss": 3.479314088821411, "perplexity": 32.43746522362256, "lr": 0.0026291804804649314, "grad_norm": 0.151474, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:32.298586+00:00", "epoch": 0, "step": 12324, "train_loss": 3.5462496280670166, "perplexity": 34.68299912271858, "lr": 0.0026291804804649314, "grad_norm": 0.167481, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:32.603021+00:00", "epoch": 0, "step": 12325, "train_loss": 3.5280368328094482, "perplexity": 34.057042275431144, "lr": 0.0026291804804649314, "grad_norm": 0.170935, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:32.907827+00:00", "epoch": 0, "step": 12326, "train_loss": 3.5865540504455566, "perplexity": 36.10943001378861, "lr": 0.0026291804804649314, "grad_norm": 0.170471, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:33.213638+00:00", "epoch": 0, "step": 12327, "train_loss": 3.5478034019470215, "perplexity": 34.736930548605955, "lr": 0.0026291804804649314, "grad_norm": 0.174002, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:33.518168+00:00", "epoch": 0, "step": 12328, "train_loss": 3.493783473968506, "perplexity": 32.91022744223223, "lr": 0.0026291804804649314, "grad_norm": 0.157576, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:33.822780+00:00", "epoch": 0, "step": 12329, "train_loss": 3.542076826095581, "perplexity": 34.538575371024024, "lr": 0.0026291804804649314, "grad_norm": 0.199781, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:34.127189+00:00", "epoch": 0, "step": 12330, "train_loss": 3.462430953979492, "perplexity": 31.89441620837116, "lr": 0.0026291804804649314, "grad_norm": 0.158874, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:34.432086+00:00", "epoch": 0, "step": 12331, "train_loss": 3.49029803276062, "perplexity": 32.79572044896273, "lr": 0.0026291804804649314, "grad_norm": 0.16852, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:34.736687+00:00", "epoch": 0, "step": 12332, "train_loss": 3.531994581222534, "perplexity": 34.19209856356529, "lr": 0.0026291804804649314, "grad_norm": 0.156911, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:35.042067+00:00", "epoch": 0, "step": 12333, "train_loss": 3.4785492420196533, "perplexity": 32.41266501747731, "lr": 0.0026291804804649314, "grad_norm": 0.166665, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:35.346396+00:00", "epoch": 0, "step": 12334, "train_loss": 3.4231770038604736, "perplexity": 30.666688608047167, "lr": 0.0026291804804649314, "grad_norm": 0.156663, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:35.650413+00:00", "epoch": 0, "step": 12335, "train_loss": 3.5084152221679688, "perplexity": 33.395301691471126, "lr": 0.0026291804804649314, "grad_norm": 0.168131, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:35.954314+00:00", "epoch": 0, "step": 12336, "train_loss": 3.60672926902771, "perplexity": 36.845344311087054, "lr": 0.0026291804804649314, "grad_norm": 0.152441, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:36.259626+00:00", "epoch": 0, "step": 12337, "train_loss": 3.5068681240081787, "perplexity": 33.34367582711575, "lr": 0.0026291804804649314, "grad_norm": 0.179933, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:36.565550+00:00", "epoch": 0, "step": 12338, "train_loss": 3.504044771194458, "perplexity": 33.24966763773991, "lr": 0.0026291804804649314, "grad_norm": 0.183209, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:36.870396+00:00", "epoch": 0, "step": 12339, "train_loss": 3.5319254398345947, "perplexity": 34.18973455614036, "lr": 0.0026291804804649314, "grad_norm": 0.187418, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:37.174901+00:00", "epoch": 0, "step": 12340, "train_loss": 3.5220842361450195, "perplexity": 33.854916622259566, "lr": 0.0026291804804649314, "grad_norm": 0.193839, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:37.479309+00:00", "epoch": 0, "step": 12341, "train_loss": 3.637183904647827, "perplexity": 37.984717380976406, "lr": 0.0026291804804649314, "grad_norm": 0.172481, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:37.785085+00:00", "epoch": 0, "step": 12342, "train_loss": 3.5178146362304688, "perplexity": 33.71067781326184, "lr": 0.0026291804804649314, "grad_norm": 0.187319, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:38.088914+00:00", "epoch": 0, "step": 12343, "train_loss": 3.4314205646514893, "perplexity": 30.9205361862747, "lr": 0.0026291804804649314, "grad_norm": 0.181011, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:38.394074+00:00", "epoch": 0, "step": 12344, "train_loss": 3.5677883625030518, "perplexity": 35.438130120324004, "lr": 0.0026291804804649314, "grad_norm": 0.190331, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:38.698162+00:00", "epoch": 0, "step": 12345, "train_loss": 3.4349801540374756, "perplexity": 31.030796723388228, "lr": 0.0026291804804649314, "grad_norm": 0.157571, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:39.002200+00:00", "epoch": 0, "step": 12346, "train_loss": 3.4840822219848633, "perplexity": 32.59250069845945, "lr": 0.0026291804804649314, "grad_norm": 0.14404, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:39.306748+00:00", "epoch": 0, "step": 12347, "train_loss": 3.421041250228882, "perplexity": 30.601262008901976, "lr": 0.0026291804804649314, "grad_norm": 0.157406, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:39.612610+00:00", "epoch": 0, "step": 12348, "train_loss": 3.616425037384033, "perplexity": 37.2043257225877, "lr": 0.0026291804804649314, "grad_norm": 0.147961, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:39.917425+00:00", "epoch": 0, "step": 12349, "train_loss": 3.5472726821899414, "perplexity": 34.7184998644602, "lr": 0.0026291804804649314, "grad_norm": 0.139184, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:40.222383+00:00", "epoch": 0, "step": 12350, "train_loss": 3.5130953788757324, "perplexity": 33.551963251057, "lr": 0.0026291804804649314, "grad_norm": 0.149304, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:40.526960+00:00", "epoch": 0, "step": 12351, "train_loss": 3.4390459060668945, "perplexity": 31.15721707088547, "lr": 0.0026291804804649314, "grad_norm": 0.168386, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:40.830969+00:00", "epoch": 0, "step": 12352, "train_loss": 3.4561057090759277, "perplexity": 31.693312899128042, "lr": 0.0026291804804649314, "grad_norm": 0.156411, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:41.135330+00:00", "epoch": 0, "step": 12353, "train_loss": 3.5554287433624268, "perplexity": 35.002823979475906, "lr": 0.0026291804804649314, "grad_norm": 0.151448, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:41.440306+00:00", "epoch": 0, "step": 12354, "train_loss": 3.529578924179077, "perplexity": 34.1096018618257, "lr": 0.0026291804804649314, "grad_norm": 0.158166, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:41.745029+00:00", "epoch": 0, "step": 12355, "train_loss": 3.499405860900879, "perplexity": 33.095782617639486, "lr": 0.0026291804804649314, "grad_norm": 0.195857, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:42.048925+00:00", "epoch": 0, "step": 12356, "train_loss": 3.530914545059204, "perplexity": 34.15518979560633, "lr": 0.0026291804804649314, "grad_norm": 0.202897, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:42.352500+00:00", "epoch": 0, "step": 12357, "train_loss": 3.564220905303955, "perplexity": 35.311931346108196, "lr": 0.0026291804804649314, "grad_norm": 0.175814, "tokens_per_sec": 107942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:42.657002+00:00", "epoch": 0, "step": 12358, "train_loss": 3.5376296043395996, "perplexity": 34.38531570965816, "lr": 0.0026291804804649314, "grad_norm": 0.165894, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:42.961462+00:00", "epoch": 0, "step": 12359, "train_loss": 3.52825665473938, "perplexity": 34.06452958309895, "lr": 0.0026291804804649314, "grad_norm": 0.163503, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:43.266971+00:00", "epoch": 0, "step": 12360, "train_loss": 3.512956142425537, "perplexity": 33.54729192001417, "lr": 0.0026291804804649314, "grad_norm": 0.167721, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:43.570746+00:00", "epoch": 0, "step": 12361, "train_loss": 3.5366740226745605, "perplexity": 34.35247342666029, "lr": 0.0026291804804649314, "grad_norm": 0.180852, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:43.874716+00:00", "epoch": 0, "step": 12362, "train_loss": 3.524790048599243, "perplexity": 33.94664572219769, "lr": 0.0026291804804649314, "grad_norm": 0.175362, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:44.179246+00:00", "epoch": 0, "step": 12363, "train_loss": 3.503878593444824, "perplexity": 33.2441427418659, "lr": 0.0026291804804649314, "grad_norm": 0.194387, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:44.485138+00:00", "epoch": 0, "step": 12364, "train_loss": 3.5353000164031982, "perplexity": 34.305305324789835, "lr": 0.0026291804804649314, "grad_norm": 0.156433, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:44.790660+00:00", "epoch": 0, "step": 12365, "train_loss": 3.562350273132324, "perplexity": 35.2459374556812, "lr": 0.0026291804804649314, "grad_norm": 0.1799, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:45.094700+00:00", "epoch": 0, "step": 12366, "train_loss": 3.6743271350860596, "perplexity": 39.42212216823097, "lr": 0.0026291804804649314, "grad_norm": 0.180488, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:45.399397+00:00", "epoch": 0, "step": 12367, "train_loss": 3.563922882080078, "perplexity": 35.301409138496105, "lr": 0.0026291804804649314, "grad_norm": 0.186181, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:45.704102+00:00", "epoch": 0, "step": 12368, "train_loss": 3.5628418922424316, "perplexity": 35.26326929206998, "lr": 0.0026291804804649314, "grad_norm": 0.152213, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:46.009774+00:00", "epoch": 0, "step": 12369, "train_loss": 3.454958438873291, "perplexity": 31.656972955469268, "lr": 0.0026291804804649314, "grad_norm": 0.157364, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:46.315352+00:00", "epoch": 0, "step": 12370, "train_loss": 3.545612335205078, "perplexity": 34.660902936565755, "lr": 0.0026291804804649314, "grad_norm": 0.169177, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:46.620924+00:00", "epoch": 0, "step": 12371, "train_loss": 3.5760300159454346, "perplexity": 35.73140578556282, "lr": 0.0026291804804649314, "grad_norm": 0.154477, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:46.925678+00:00", "epoch": 0, "step": 12372, "train_loss": 3.5432708263397217, "perplexity": 34.579839067979165, "lr": 0.0026291804804649314, "grad_norm": 0.172512, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:47.230109+00:00", "epoch": 0, "step": 12373, "train_loss": 3.4735841751098633, "perplexity": 32.25213282326865, "lr": 0.0026291804804649314, "grad_norm": 0.152496, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:47.536357+00:00", "epoch": 0, "step": 12374, "train_loss": 3.7153542041778564, "perplexity": 41.07313278339652, "lr": 0.0026291804804649314, "grad_norm": 0.183348, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:47.841524+00:00", "epoch": 0, "step": 12375, "train_loss": 3.537263870239258, "perplexity": 34.372742126586324, "lr": 0.0026291804804649314, "grad_norm": 0.201622, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:48.147378+00:00", "epoch": 0, "step": 12376, "train_loss": 3.6184980869293213, "perplexity": 37.281532131815666, "lr": 0.0026291804804649314, "grad_norm": 0.181224, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:48.451911+00:00", "epoch": 0, "step": 12377, "train_loss": 3.419640302658081, "perplexity": 30.55842126105478, "lr": 0.0026291804804649314, "grad_norm": 0.162025, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:48.755521+00:00", "epoch": 0, "step": 12378, "train_loss": 3.4933416843414307, "perplexity": 32.895691256328064, "lr": 0.0026291804804649314, "grad_norm": 0.173364, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:49.059724+00:00", "epoch": 0, "step": 12379, "train_loss": 3.439098834991455, "perplexity": 31.158866232521124, "lr": 0.0026291804804649314, "grad_norm": 0.19064, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:49.365775+00:00", "epoch": 0, "step": 12380, "train_loss": 3.5419209003448486, "perplexity": 34.53319033757367, "lr": 0.0026291804804649314, "grad_norm": 0.154994, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:49.671773+00:00", "epoch": 0, "step": 12381, "train_loss": 3.428175926208496, "perplexity": 30.820372810730156, "lr": 0.0026291804804649314, "grad_norm": 0.175235, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:49.976642+00:00", "epoch": 0, "step": 12382, "train_loss": 3.433076858520508, "perplexity": 30.97179211651072, "lr": 0.0026291804804649314, "grad_norm": 0.197726, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:50.280901+00:00", "epoch": 0, "step": 12383, "train_loss": 3.522066116333008, "perplexity": 33.85430318309241, "lr": 0.0026291804804649314, "grad_norm": 0.192351, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:50.584858+00:00", "epoch": 0, "step": 12384, "train_loss": 3.412907361984253, "perplexity": 30.353364316384432, "lr": 0.0026291804804649314, "grad_norm": 0.163459, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:50.890500+00:00", "epoch": 0, "step": 12385, "train_loss": 3.5234439373016357, "perplexity": 33.90098050100574, "lr": 0.0026291804804649314, "grad_norm": 0.15152, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:51.195098+00:00", "epoch": 0, "step": 12386, "train_loss": 3.512085437774658, "perplexity": 33.518094849772034, "lr": 0.0026291804804649314, "grad_norm": 0.145432, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:51.500053+00:00", "epoch": 0, "step": 12387, "train_loss": 3.5377211570739746, "perplexity": 34.38846392344532, "lr": 0.0026291804804649314, "grad_norm": 0.141167, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:51.804929+00:00", "epoch": 0, "step": 12388, "train_loss": 3.519962787628174, "perplexity": 33.78317128852701, "lr": 0.0026291804804649314, "grad_norm": 0.176651, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:52.110270+00:00", "epoch": 0, "step": 12389, "train_loss": 3.500051736831665, "perplexity": 33.11716529157664, "lr": 0.0026291804804649314, "grad_norm": 0.189567, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:52.414287+00:00", "epoch": 0, "step": 12390, "train_loss": 3.5459609031677246, "perplexity": 34.672986722773864, "lr": 0.0026291804804649314, "grad_norm": 0.171543, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:52.719176+00:00", "epoch": 0, "step": 12391, "train_loss": 3.548903465270996, "perplexity": 34.775164397861886, "lr": 0.0026291804804649314, "grad_norm": 0.189027, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:53.024802+00:00", "epoch": 0, "step": 12392, "train_loss": 3.4964840412139893, "perplexity": 32.999223840795565, "lr": 0.0026291804804649314, "grad_norm": 0.174849, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:53.328619+00:00", "epoch": 0, "step": 12393, "train_loss": 3.424957513809204, "perplexity": 30.721339591079964, "lr": 0.0026291804804649314, "grad_norm": 0.166716, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:53.633263+00:00", "epoch": 0, "step": 12394, "train_loss": 3.5220181941986084, "perplexity": 33.852680851498384, "lr": 0.0026291804804649314, "grad_norm": 0.169688, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:53.937951+00:00", "epoch": 0, "step": 12395, "train_loss": 3.4256551265716553, "perplexity": 30.742778666875274, "lr": 0.0026291804804649314, "grad_norm": 0.183551, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:54.244298+00:00", "epoch": 0, "step": 12396, "train_loss": 3.5865039825439453, "perplexity": 36.107622135658154, "lr": 0.0026291804804649314, "grad_norm": 0.184182, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:54.548788+00:00", "epoch": 0, "step": 12397, "train_loss": 3.5640268325805664, "perplexity": 35.30507892837893, "lr": 0.0026291804804649314, "grad_norm": 0.156445, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:54.854143+00:00", "epoch": 0, "step": 12398, "train_loss": 3.566842555999756, "perplexity": 35.40462835198416, "lr": 0.0026291804804649314, "grad_norm": 0.181917, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:55.158772+00:00", "epoch": 0, "step": 12399, "train_loss": 3.5209789276123047, "perplexity": 33.81751716681884, "lr": 0.0026291804804649314, "grad_norm": 0.167449, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:55.464096+00:00", "epoch": 0, "step": 12400, "train_loss": 3.4748682975769043, "perplexity": 32.29357511443083, "lr": 0.0026291804804649314, "grad_norm": 0.175555, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:55.768635+00:00", "epoch": 0, "step": 12401, "train_loss": 3.558767795562744, "perplexity": 35.11989558122472, "lr": 0.0026291804804649314, "grad_norm": 0.162857, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:56.073400+00:00", "epoch": 0, "step": 12402, "train_loss": 3.643740653991699, "perplexity": 38.23459193896837, "lr": 0.0026291804804649314, "grad_norm": 0.162719, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:56.379055+00:00", "epoch": 0, "step": 12403, "train_loss": 3.4167115688323975, "perplexity": 30.469054708349258, "lr": 0.0026291804804649314, "grad_norm": 0.154462, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:56.683806+00:00", "epoch": 0, "step": 12404, "train_loss": 3.557417631149292, "perplexity": 35.07250994440551, "lr": 0.0026291804804649314, "grad_norm": 0.169356, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:56.988837+00:00", "epoch": 0, "step": 12405, "train_loss": 3.516380786895752, "perplexity": 33.66237641704419, "lr": 0.0026291804804649314, "grad_norm": 0.181709, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:57.293638+00:00", "epoch": 0, "step": 12406, "train_loss": 3.602626323699951, "perplexity": 36.69447958429919, "lr": 0.0026291804804649314, "grad_norm": 0.165215, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:57.599538+00:00", "epoch": 0, "step": 12407, "train_loss": 3.544518232345581, "perplexity": 34.623001081594964, "lr": 0.0026291804804649314, "grad_norm": 0.149696, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:57.906235+00:00", "epoch": 0, "step": 12408, "train_loss": 3.4869093894958496, "perplexity": 32.684775534283126, "lr": 0.0026291804804649314, "grad_norm": 0.157353, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:58.211492+00:00", "epoch": 0, "step": 12409, "train_loss": 3.48066782951355, "perplexity": 32.481406876336614, "lr": 0.0026291804804649314, "grad_norm": 0.142997, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:58.517193+00:00", "epoch": 0, "step": 12410, "train_loss": 3.4924063682556152, "perplexity": 32.864937771498504, "lr": 0.0026291804804649314, "grad_norm": 0.161193, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:58.822694+00:00", "epoch": 0, "step": 12411, "train_loss": 3.54372501373291, "perplexity": 34.59554836215396, "lr": 0.0026291804804649314, "grad_norm": 0.162682, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:59.127191+00:00", "epoch": 0, "step": 12412, "train_loss": 3.513033866882324, "perplexity": 33.54989946638908, "lr": 0.0026291804804649314, "grad_norm": 0.167734, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:59.432942+00:00", "epoch": 0, "step": 12413, "train_loss": 3.4749555587768555, "perplexity": 32.29639321349959, "lr": 0.0026291804804649314, "grad_norm": 0.160438, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:02:59.739160+00:00", "epoch": 0, "step": 12414, "train_loss": 3.5270025730133057, "perplexity": 34.02183665484739, "lr": 0.0026291804804649314, "grad_norm": 0.15285, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:00.045131+00:00", "epoch": 0, "step": 12415, "train_loss": 3.436305522918701, "perplexity": 31.07195124215995, "lr": 0.0026291804804649314, "grad_norm": 0.173682, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:00.350958+00:00", "epoch": 0, "step": 12416, "train_loss": 3.5505685806274414, "perplexity": 34.83311729386216, "lr": 0.0026291804804649314, "grad_norm": 0.144016, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:00.656347+00:00", "epoch": 0, "step": 12417, "train_loss": 3.496509313583374, "perplexity": 33.00005781990815, "lr": 0.0026291804804649314, "grad_norm": 0.15923, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:00.961843+00:00", "epoch": 0, "step": 12418, "train_loss": 3.512035846710205, "perplexity": 33.516432692984395, "lr": 0.0026291804804649314, "grad_norm": 0.177238, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:01.268397+00:00", "epoch": 0, "step": 12419, "train_loss": 3.553208351135254, "perplexity": 34.9251902017965, "lr": 0.0026291804804649314, "grad_norm": 0.148596, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:01.573926+00:00", "epoch": 0, "step": 12420, "train_loss": 3.504509210586548, "perplexity": 33.26511367976231, "lr": 0.0026291804804649314, "grad_norm": 0.168887, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:01.878343+00:00", "epoch": 0, "step": 12421, "train_loss": 3.5171828269958496, "perplexity": 33.689385822660576, "lr": 0.0026291804804649314, "grad_norm": 0.174682, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:02.183838+00:00", "epoch": 0, "step": 12422, "train_loss": 3.543592691421509, "perplexity": 34.59097090208721, "lr": 0.0026291804804649314, "grad_norm": 0.168369, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:02.489174+00:00", "epoch": 0, "step": 12423, "train_loss": 3.4449985027313232, "perplexity": 31.343236517387982, "lr": 0.0026291804804649314, "grad_norm": 0.165599, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:02.794129+00:00", "epoch": 0, "step": 12424, "train_loss": 3.7047979831695557, "perplexity": 40.64183615258702, "lr": 0.0026291804804649314, "grad_norm": 0.16852, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:03.098869+00:00", "epoch": 0, "step": 12425, "train_loss": 3.5496866703033447, "perplexity": 34.802411150123, "lr": 0.0026291804804649314, "grad_norm": 0.175639, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:03.403182+00:00", "epoch": 0, "step": 12426, "train_loss": 3.4844162464141846, "perplexity": 32.60338920831831, "lr": 0.0026291804804649314, "grad_norm": 0.172731, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:03.706595+00:00", "epoch": 0, "step": 12427, "train_loss": 3.605830192565918, "perplexity": 36.81223241658328, "lr": 0.0026291804804649314, "grad_norm": 0.184338, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:04.012591+00:00", "epoch": 0, "step": 12428, "train_loss": 3.5112602710723877, "perplexity": 33.49044824207722, "lr": 0.0026291804804649314, "grad_norm": 0.175283, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:04.316990+00:00", "epoch": 0, "step": 12429, "train_loss": 3.5078580379486084, "perplexity": 33.37669953926195, "lr": 0.0026291804804649314, "grad_norm": 0.169139, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:04.622325+00:00", "epoch": 0, "step": 12430, "train_loss": 3.386157751083374, "perplexity": 29.552186994609645, "lr": 0.0026291804804649314, "grad_norm": 0.148668, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:04.927842+00:00", "epoch": 0, "step": 12431, "train_loss": 3.5736000537872314, "perplexity": 35.64468522815261, "lr": 0.0026291804804649314, "grad_norm": 0.198057, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:05.233027+00:00", "epoch": 0, "step": 12432, "train_loss": 3.6852376461029053, "perplexity": 39.85459261316053, "lr": 0.0026291804804649314, "grad_norm": 0.182153, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:05.538021+00:00", "epoch": 0, "step": 12433, "train_loss": 3.5138847827911377, "perplexity": 33.57845975906454, "lr": 0.0026291804804649314, "grad_norm": 0.173172, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:05.844623+00:00", "epoch": 0, "step": 12434, "train_loss": 3.4995081424713135, "perplexity": 33.09916787938237, "lr": 0.0026291804804649314, "grad_norm": 0.156162, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:06.149810+00:00", "epoch": 0, "step": 12435, "train_loss": 3.4965546131134033, "perplexity": 33.001552740877685, "lr": 0.0026291804804649314, "grad_norm": 0.17651, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:06.455659+00:00", "epoch": 0, "step": 12436, "train_loss": 3.511589765548706, "perplexity": 33.50148497795431, "lr": 0.0026291804804649314, "grad_norm": 0.17772, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:06.761425+00:00", "epoch": 0, "step": 12437, "train_loss": 3.573457717895508, "perplexity": 35.63961207115018, "lr": 0.0026291804804649314, "grad_norm": 0.198598, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:07.066880+00:00", "epoch": 0, "step": 12438, "train_loss": 3.5154194831848145, "perplexity": 33.63003219847497, "lr": 0.0026291804804649314, "grad_norm": 0.184731, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:07.371955+00:00", "epoch": 0, "step": 12439, "train_loss": 3.6466803550720215, "perplexity": 38.34715558086379, "lr": 0.0026291804804649314, "grad_norm": 0.158267, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:07.677803+00:00", "epoch": 0, "step": 12440, "train_loss": 3.5421407222747803, "perplexity": 34.540782324532415, "lr": 0.0026291804804649314, "grad_norm": 0.181031, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:07.983444+00:00", "epoch": 0, "step": 12441, "train_loss": 3.553959369659424, "perplexity": 34.951429518471336, "lr": 0.0026291804804649314, "grad_norm": 0.185115, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:08.288623+00:00", "epoch": 0, "step": 12442, "train_loss": 3.5061585903167725, "perplexity": 33.32002575697406, "lr": 0.0026291804804649314, "grad_norm": 0.15874, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:08.593932+00:00", "epoch": 0, "step": 12443, "train_loss": 3.5836148262023926, "perplexity": 36.00345212437174, "lr": 0.0026291804804649314, "grad_norm": 0.14891, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:08.900056+00:00", "epoch": 0, "step": 12444, "train_loss": 3.518840789794922, "perplexity": 33.745287900055004, "lr": 0.0026291804804649314, "grad_norm": 0.159893, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:09.206251+00:00", "epoch": 0, "step": 12445, "train_loss": 3.552049160003662, "perplexity": 34.8847286868908, "lr": 0.0026291804804649314, "grad_norm": 0.15478, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:09.511452+00:00", "epoch": 0, "step": 12446, "train_loss": 3.5262134075164795, "perplexity": 33.99499838653127, "lr": 0.0026291804804649314, "grad_norm": 0.153696, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:09.815999+00:00", "epoch": 0, "step": 12447, "train_loss": 3.647655487060547, "perplexity": 38.38456735668481, "lr": 0.0026291804804649314, "grad_norm": 0.168464, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:10.121451+00:00", "epoch": 0, "step": 12448, "train_loss": 3.5091261863708496, "perplexity": 33.419052997681945, "lr": 0.0026291804804649314, "grad_norm": 0.166296, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:10.425694+00:00", "epoch": 0, "step": 12449, "train_loss": 3.6815998554229736, "perplexity": 39.70987333641341, "lr": 0.0026291804804649314, "grad_norm": 0.142861, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:10.732003+00:00", "epoch": 0, "step": 12450, "train_loss": 3.5706324577331543, "perplexity": 35.539063000488355, "lr": 0.0026291804804649314, "grad_norm": 0.145653, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:11.036799+00:00", "epoch": 0, "step": 12451, "train_loss": 3.456944227218628, "perplexity": 31.71989946209616, "lr": 0.0026291804804649314, "grad_norm": 0.154, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:11.341713+00:00", "epoch": 0, "step": 12452, "train_loss": 3.625746726989746, "perplexity": 37.55275434768686, "lr": 0.0026291804804649314, "grad_norm": 0.157363, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:11.646496+00:00", "epoch": 0, "step": 12453, "train_loss": 3.5551645755767822, "perplexity": 34.99357858219577, "lr": 0.0026291804804649314, "grad_norm": 0.170003, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:11.951183+00:00", "epoch": 0, "step": 12454, "train_loss": 3.617424726486206, "perplexity": 37.24153707836066, "lr": 0.0026291804804649314, "grad_norm": 0.171946, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:12.257204+00:00", "epoch": 0, "step": 12455, "train_loss": 3.5547332763671875, "perplexity": 34.97848913367989, "lr": 0.0026291804804649314, "grad_norm": 0.158453, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:12.562103+00:00", "epoch": 0, "step": 12456, "train_loss": 3.530832052230835, "perplexity": 34.15237235360758, "lr": 0.0026291804804649314, "grad_norm": 0.197447, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:12.867074+00:00", "epoch": 0, "step": 12457, "train_loss": 3.523775815963745, "perplexity": 33.91223338025058, "lr": 0.0026291804804649314, "grad_norm": 0.190362, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:13.172716+00:00", "epoch": 0, "step": 12458, "train_loss": 3.488990545272827, "perplexity": 32.75286847513706, "lr": 0.0026291804804649314, "grad_norm": 0.156602, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:13.477416+00:00", "epoch": 0, "step": 12459, "train_loss": 3.503530502319336, "perplexity": 33.232572764623015, "lr": 0.0026291804804649314, "grad_norm": 0.171922, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:13.781845+00:00", "epoch": 0, "step": 12460, "train_loss": 3.5270047187805176, "perplexity": 34.0219096578673, "lr": 0.0026291804804649314, "grad_norm": 0.16197, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:14.088113+00:00", "epoch": 0, "step": 12461, "train_loss": 3.407348155975342, "perplexity": 30.185091875037877, "lr": 0.0026291804804649314, "grad_norm": 0.163479, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:14.394285+00:00", "epoch": 0, "step": 12462, "train_loss": 3.476135492324829, "perplexity": 32.33452330239242, "lr": 0.0026291804804649314, "grad_norm": 0.184137, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:14.699277+00:00", "epoch": 0, "step": 12463, "train_loss": 3.4637842178344727, "perplexity": 31.93760698666987, "lr": 0.0026291804804649314, "grad_norm": 0.152802, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:15.004390+00:00", "epoch": 0, "step": 12464, "train_loss": 3.5820696353912354, "perplexity": 35.94786288003447, "lr": 0.0026291804804649314, "grad_norm": 0.180442, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:15.308770+00:00", "epoch": 0, "step": 12465, "train_loss": 3.3550333976745605, "perplexity": 28.646560903248865, "lr": 0.0026291804804649314, "grad_norm": 0.18498, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:15.615011+00:00", "epoch": 0, "step": 12466, "train_loss": 3.4747049808502197, "perplexity": 32.288301464099575, "lr": 0.0026291804804649314, "grad_norm": 0.161755, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:15.920063+00:00", "epoch": 0, "step": 12467, "train_loss": 3.6426994800567627, "perplexity": 38.194803795212025, "lr": 0.0026291804804649314, "grad_norm": 0.157572, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:16.225140+00:00", "epoch": 0, "step": 12468, "train_loss": 3.5540812015533447, "perplexity": 34.95568797672756, "lr": 0.0026291804804649314, "grad_norm": 0.153459, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:16.530386+00:00", "epoch": 0, "step": 12469, "train_loss": 3.561542510986328, "perplexity": 35.217478617138, "lr": 0.0026291804804649314, "grad_norm": 0.156125, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:16.836181+00:00", "epoch": 0, "step": 12470, "train_loss": 3.435180425643921, "perplexity": 31.037011933241555, "lr": 0.0026291804804649314, "grad_norm": 0.158819, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:17.141230+00:00", "epoch": 0, "step": 12471, "train_loss": 3.5393636226654053, "perplexity": 34.44499220235026, "lr": 0.0026291804804649314, "grad_norm": 0.172339, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:17.447594+00:00", "epoch": 0, "step": 12472, "train_loss": 3.5645804405212402, "perplexity": 35.32462951159953, "lr": 0.0026291804804649314, "grad_norm": 0.168497, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:17.752645+00:00", "epoch": 0, "step": 12473, "train_loss": 3.4982755184173584, "perplexity": 33.05839418337008, "lr": 0.0026291804804649314, "grad_norm": 0.162415, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:18.057743+00:00", "epoch": 0, "step": 12474, "train_loss": 3.5092809200286865, "perplexity": 33.42422445008228, "lr": 0.0026291804804649314, "grad_norm": 0.170463, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:18.362549+00:00", "epoch": 0, "step": 12475, "train_loss": 3.4485368728637695, "perplexity": 31.45433693061191, "lr": 0.0026291804804649314, "grad_norm": 0.200901, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:18.667438+00:00", "epoch": 0, "step": 12476, "train_loss": 3.6938180923461914, "perplexity": 40.19803413626187, "lr": 0.0026291804804649314, "grad_norm": 0.206651, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:18.973270+00:00", "epoch": 0, "step": 12477, "train_loss": 3.5375900268554688, "perplexity": 34.38395485230116, "lr": 0.0026291804804649314, "grad_norm": 0.14973, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:19.277240+00:00", "epoch": 0, "step": 12478, "train_loss": 3.591778039932251, "perplexity": 36.29855887001807, "lr": 0.0026291804804649314, "grad_norm": 0.213639, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:19.581699+00:00", "epoch": 0, "step": 12479, "train_loss": 3.563883066177368, "perplexity": 35.30000360900572, "lr": 0.0026291804804649314, "grad_norm": 0.170219, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:19.887360+00:00", "epoch": 0, "step": 12480, "train_loss": 3.58807110786438, "perplexity": 36.16425166576326, "lr": 0.0026291804804649314, "grad_norm": 0.181023, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:20.192127+00:00", "epoch": 0, "step": 12481, "train_loss": 3.4790422916412354, "perplexity": 32.42865001007209, "lr": 0.0026291804804649314, "grad_norm": 0.154826, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:20.498177+00:00", "epoch": 0, "step": 12482, "train_loss": 3.504960536956787, "perplexity": 33.2801304912616, "lr": 0.0026291804804649314, "grad_norm": 0.157771, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:20.803630+00:00", "epoch": 0, "step": 12483, "train_loss": 3.4711813926696777, "perplexity": 32.174730991981974, "lr": 0.0026291804804649314, "grad_norm": 0.167546, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:21.108681+00:00", "epoch": 0, "step": 12484, "train_loss": 3.5114145278930664, "perplexity": 33.49561477062204, "lr": 0.0026291804804649314, "grad_norm": 0.159649, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:21.414066+00:00", "epoch": 0, "step": 12485, "train_loss": 3.5248336791992188, "perplexity": 33.94812686702909, "lr": 0.0026291804804649314, "grad_norm": 0.166851, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:21.720679+00:00", "epoch": 0, "step": 12486, "train_loss": 3.5446295738220215, "perplexity": 34.62685627227152, "lr": 0.0026291804804649314, "grad_norm": 0.197457, "tokens_per_sec": 106870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:22.026828+00:00", "epoch": 0, "step": 12487, "train_loss": 3.430521249771118, "perplexity": 30.892741387984472, "lr": 0.0026291804804649314, "grad_norm": 0.158842, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:22.332328+00:00", "epoch": 0, "step": 12488, "train_loss": 3.5107381343841553, "perplexity": 33.47296621474578, "lr": 0.0026291804804649314, "grad_norm": 0.18463, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:22.637346+00:00", "epoch": 0, "step": 12489, "train_loss": 3.579946279525757, "perplexity": 35.87161375519828, "lr": 0.0026291804804649314, "grad_norm": 0.185807, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:22.944491+00:00", "epoch": 0, "step": 12490, "train_loss": 3.485778331756592, "perplexity": 32.64782806475873, "lr": 0.0026291804804649314, "grad_norm": 0.142149, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:23.251721+00:00", "epoch": 0, "step": 12491, "train_loss": 3.551255226135254, "perplexity": 34.857043510849735, "lr": 0.0026291804804649314, "grad_norm": 0.163354, "tokens_per_sec": 106597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:23.557575+00:00", "epoch": 0, "step": 12492, "train_loss": 3.5230841636657715, "perplexity": 33.88878601575021, "lr": 0.0026291804804649314, "grad_norm": 0.174588, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:23.863169+00:00", "epoch": 0, "step": 12493, "train_loss": 3.5113532543182373, "perplexity": 33.49356243744148, "lr": 0.0026291804804649314, "grad_norm": 0.19953, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:24.169177+00:00", "epoch": 0, "step": 12494, "train_loss": 3.498976469039917, "perplexity": 33.08157460857134, "lr": 0.0026291804804649314, "grad_norm": 0.153306, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:24.474543+00:00", "epoch": 0, "step": 12495, "train_loss": 3.5431783199310303, "perplexity": 34.57664035920667, "lr": 0.0026291804804649314, "grad_norm": 0.156243, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:24.780685+00:00", "epoch": 0, "step": 12496, "train_loss": 3.5256729125976562, "perplexity": 33.97662922730375, "lr": 0.0026291804804649314, "grad_norm": 0.182226, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:25.085993+00:00", "epoch": 0, "step": 12497, "train_loss": 3.4619290828704834, "perplexity": 31.87841333836243, "lr": 0.0026291804804649314, "grad_norm": 0.139967, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:25.391480+00:00", "epoch": 0, "step": 12498, "train_loss": 3.4744038581848145, "perplexity": 32.27858018842297, "lr": 0.0026291804804649314, "grad_norm": 0.157819, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:25.697534+00:00", "epoch": 0, "step": 12499, "train_loss": 3.4782915115356445, "perplexity": 32.404312362047534, "lr": 0.0026291804804649314, "grad_norm": 0.141385, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:26.004358+00:00", "epoch": 0, "step": 12500, "train_loss": 3.54378604888916, "perplexity": 34.59765997129434, "lr": 0.0026291804804649314, "grad_norm": 0.14552, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:03:29.077846+00:00", "step": 12500, "epoch": 0, "val_loss": 3.481094980239868, "val_ppl": 32.49528429653974, "eval_train_loss": 3.54378604888916, "eval_train_ppl": 34.59765997129434} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:03:30.097158+00:00", "step": 12500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4811_epoch_0000_step_0012500.pt", "val_loss": 3.481094980239868} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T22:03:32.115573+00:00", "step": 12500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0012500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:33.393812+00:00", "epoch": 0, "step": 12501, "train_loss": 3.5046279430389404, "perplexity": 33.26906356277363, "lr": 0.0026291804804649314, "grad_norm": 0.149024, "tokens_per_sec": 4434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:33.698958+00:00", "epoch": 0, "step": 12502, "train_loss": 3.436631441116333, "perplexity": 31.082079806957832, "lr": 0.0026291804804649314, "grad_norm": 0.152818, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:34.001886+00:00", "epoch": 0, "step": 12503, "train_loss": 3.565729856491089, "perplexity": 35.36525554852419, "lr": 0.0026291804804649314, "grad_norm": 0.152606, "tokens_per_sec": 108172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:34.306349+00:00", "epoch": 0, "step": 12504, "train_loss": 3.587029218673706, "perplexity": 36.12659214479453, "lr": 0.0026291804804649314, "grad_norm": 0.153253, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:34.609710+00:00", "epoch": 0, "step": 12505, "train_loss": 3.5862863063812256, "perplexity": 36.0997632224068, "lr": 0.0026291804804649314, "grad_norm": 0.147713, "tokens_per_sec": 108018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:34.914795+00:00", "epoch": 0, "step": 12506, "train_loss": 3.500478744506836, "perplexity": 33.13130959498418, "lr": 0.0026291804804649314, "grad_norm": 0.157855, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:35.218968+00:00", "epoch": 0, "step": 12507, "train_loss": 3.5593719482421875, "perplexity": 35.141119770921854, "lr": 0.0026291804804649314, "grad_norm": 0.145827, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:35.522692+00:00", "epoch": 0, "step": 12508, "train_loss": 3.545593500137329, "perplexity": 34.660250102258814, "lr": 0.0026291804804649314, "grad_norm": 0.163787, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:35.827060+00:00", "epoch": 0, "step": 12509, "train_loss": 3.508502244949341, "perplexity": 33.39820796996346, "lr": 0.0026291804804649314, "grad_norm": 0.141341, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:36.130781+00:00", "epoch": 0, "step": 12510, "train_loss": 3.5517632961273193, "perplexity": 34.87475782834524, "lr": 0.0026291804804649314, "grad_norm": 0.185153, "tokens_per_sec": 107950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:36.435775+00:00", "epoch": 0, "step": 12511, "train_loss": 3.561292886734009, "perplexity": 35.20868857751881, "lr": 0.0026291804804649314, "grad_norm": 0.180954, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:36.741352+00:00", "epoch": 0, "step": 12512, "train_loss": 3.547370433807373, "perplexity": 34.72189381985636, "lr": 0.0026291804804649314, "grad_norm": 0.172869, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:37.046202+00:00", "epoch": 0, "step": 12513, "train_loss": 3.5101001262664795, "perplexity": 33.45161700178787, "lr": 0.0026291804804649314, "grad_norm": 0.164972, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:37.453777+00:00", "epoch": 0, "step": 12514, "train_loss": 3.535031318664551, "perplexity": 34.2960888051099, "lr": 0.0026291804804649314, "grad_norm": 0.140779, "tokens_per_sec": 80398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:37.758251+00:00", "epoch": 0, "step": 12515, "train_loss": 3.5645909309387207, "perplexity": 35.325000083654174, "lr": 0.0026291804804649314, "grad_norm": 0.185445, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:38.061777+00:00", "epoch": 0, "step": 12516, "train_loss": 3.621112108230591, "perplexity": 37.37911433636705, "lr": 0.0026291804804649314, "grad_norm": 0.187162, "tokens_per_sec": 107958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:38.365532+00:00", "epoch": 0, "step": 12517, "train_loss": 3.584096908569336, "perplexity": 36.020812938124855, "lr": 0.0026291804804649314, "grad_norm": 0.222336, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:38.669673+00:00", "epoch": 0, "step": 12518, "train_loss": 3.5871083736419678, "perplexity": 36.129451857227885, "lr": 0.0026291804804649314, "grad_norm": 0.171155, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:38.973505+00:00", "epoch": 0, "step": 12519, "train_loss": 3.5496082305908203, "perplexity": 34.79968136606037, "lr": 0.0026291804804649314, "grad_norm": 0.187363, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:39.278901+00:00", "epoch": 0, "step": 12520, "train_loss": 3.51496958732605, "perplexity": 33.61490558921333, "lr": 0.0026291804804649314, "grad_norm": 0.168234, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:39.583332+00:00", "epoch": 0, "step": 12521, "train_loss": 3.4658446311950684, "perplexity": 32.00347949780762, "lr": 0.0026291804804649314, "grad_norm": 0.17849, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:39.886811+00:00", "epoch": 0, "step": 12522, "train_loss": 3.455986976623535, "perplexity": 31.68955009775082, "lr": 0.0026291804804649314, "grad_norm": 0.183929, "tokens_per_sec": 107975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:40.190838+00:00", "epoch": 0, "step": 12523, "train_loss": 3.516177177429199, "perplexity": 33.65552313625912, "lr": 0.0026291804804649314, "grad_norm": 0.20848, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:40.494795+00:00", "epoch": 0, "step": 12524, "train_loss": 3.493821382522583, "perplexity": 32.91147504501617, "lr": 0.0026291804804649314, "grad_norm": 0.163517, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:40.799349+00:00", "epoch": 0, "step": 12525, "train_loss": 3.554766893386841, "perplexity": 34.979665026001435, "lr": 0.0026291804804649314, "grad_norm": 0.160196, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:41.104371+00:00", "epoch": 0, "step": 12526, "train_loss": 3.448529005050659, "perplexity": 31.454089454740977, "lr": 0.0026291804804649314, "grad_norm": 0.171803, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:41.408020+00:00", "epoch": 0, "step": 12527, "train_loss": 3.5301640033721924, "perplexity": 34.129564519458846, "lr": 0.0026291804804649314, "grad_norm": 0.163233, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:41.711810+00:00", "epoch": 0, "step": 12528, "train_loss": 3.5705618858337402, "perplexity": 35.53655502980618, "lr": 0.0026291804804649314, "grad_norm": 0.208051, "tokens_per_sec": 107864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:42.015842+00:00", "epoch": 0, "step": 12529, "train_loss": 3.488647699356079, "perplexity": 32.7416412126391, "lr": 0.0026291804804649314, "grad_norm": 0.157011, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:42.319971+00:00", "epoch": 0, "step": 12530, "train_loss": 3.5049643516540527, "perplexity": 33.28025744512653, "lr": 0.0026291804804649314, "grad_norm": 0.222776, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:42.624274+00:00", "epoch": 0, "step": 12531, "train_loss": 3.585284471511841, "perplexity": 36.0636153309665, "lr": 0.0026291804804649314, "grad_norm": 0.189566, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:42.928598+00:00", "epoch": 0, "step": 12532, "train_loss": 3.4682765007019043, "perplexity": 32.081402494587884, "lr": 0.0026291804804649314, "grad_norm": 0.158102, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:43.232290+00:00", "epoch": 0, "step": 12533, "train_loss": 3.6327078342437744, "perplexity": 37.81507506042107, "lr": 0.0026291804804649314, "grad_norm": 0.165052, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:43.535964+00:00", "epoch": 0, "step": 12534, "train_loss": 3.5223653316497803, "perplexity": 33.86443442477921, "lr": 0.0026291804804649314, "grad_norm": 0.15016, "tokens_per_sec": 107906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:43.839799+00:00", "epoch": 0, "step": 12535, "train_loss": 3.5273995399475098, "perplexity": 34.03534488002224, "lr": 0.0026291804804649314, "grad_norm": 0.14612, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:44.144609+00:00", "epoch": 0, "step": 12536, "train_loss": 3.5075769424438477, "perplexity": 33.36731881755883, "lr": 0.0026291804804649314, "grad_norm": 0.170213, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:44.449570+00:00", "epoch": 0, "step": 12537, "train_loss": 3.528362512588501, "perplexity": 34.068135771800264, "lr": 0.0026291804804649314, "grad_norm": 0.166288, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:44.755001+00:00", "epoch": 0, "step": 12538, "train_loss": 3.4613797664642334, "perplexity": 31.86090681166183, "lr": 0.0026291804804649314, "grad_norm": 0.166304, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:45.058859+00:00", "epoch": 0, "step": 12539, "train_loss": 3.4572410583496094, "perplexity": 31.72931631326626, "lr": 0.0026291804804649314, "grad_norm": 0.166293, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:45.364506+00:00", "epoch": 0, "step": 12540, "train_loss": 3.477649450302124, "perplexity": 32.383513487071156, "lr": 0.0026291804804649314, "grad_norm": 0.155361, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:45.669080+00:00", "epoch": 0, "step": 12541, "train_loss": 3.5188088417053223, "perplexity": 33.744209819795, "lr": 0.0026291804804649314, "grad_norm": 0.151892, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:45.974432+00:00", "epoch": 0, "step": 12542, "train_loss": 3.582479953765869, "perplexity": 35.962615975229, "lr": 0.0026291804804649314, "grad_norm": 0.147289, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:46.279014+00:00", "epoch": 0, "step": 12543, "train_loss": 3.621678113937378, "perplexity": 37.4002771169584, "lr": 0.0026291804804649314, "grad_norm": 0.159498, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:46.583132+00:00", "epoch": 0, "step": 12544, "train_loss": 3.527287721633911, "perplexity": 34.03153931792474, "lr": 0.0026291804804649314, "grad_norm": 0.175581, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:46.888311+00:00", "epoch": 0, "step": 12545, "train_loss": 3.5794007778167725, "perplexity": 35.85205106481642, "lr": 0.0026291804804649314, "grad_norm": 0.158681, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:47.192943+00:00", "epoch": 0, "step": 12546, "train_loss": 3.6471951007843018, "perplexity": 38.36689969594052, "lr": 0.0026291804804649314, "grad_norm": 0.158168, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:47.498367+00:00", "epoch": 0, "step": 12547, "train_loss": 3.477268695831299, "perplexity": 32.37118566661523, "lr": 0.0026291804804649314, "grad_norm": 0.178166, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:47.803473+00:00", "epoch": 0, "step": 12548, "train_loss": 3.5425875186920166, "perplexity": 34.55621847047412, "lr": 0.0026291804804649314, "grad_norm": 0.177996, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:48.107539+00:00", "epoch": 0, "step": 12549, "train_loss": 3.4892351627349854, "perplexity": 32.76088137870997, "lr": 0.0026291804804649314, "grad_norm": 0.154337, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:48.412562+00:00", "epoch": 0, "step": 12550, "train_loss": 3.5100579261779785, "perplexity": 33.450205370375585, "lr": 0.0026291804804649314, "grad_norm": 0.157204, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:48.716888+00:00", "epoch": 0, "step": 12551, "train_loss": 3.4307944774627686, "perplexity": 30.901183293631345, "lr": 0.0026291804804649314, "grad_norm": 0.1467, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:49.022258+00:00", "epoch": 0, "step": 12552, "train_loss": 3.608743667602539, "perplexity": 36.919640325926395, "lr": 0.0026291804804649314, "grad_norm": 0.146256, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:49.326887+00:00", "epoch": 0, "step": 12553, "train_loss": 3.44858980178833, "perplexity": 31.456001818898397, "lr": 0.0026291804804649314, "grad_norm": 0.150279, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:49.630973+00:00", "epoch": 0, "step": 12554, "train_loss": 3.5048940181732178, "perplexity": 33.27791681109077, "lr": 0.0026291804804649314, "grad_norm": 0.156625, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:49.936664+00:00", "epoch": 0, "step": 12555, "train_loss": 3.547074317932129, "perplexity": 34.7116136380154, "lr": 0.0026291804804649314, "grad_norm": 0.15258, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:50.241379+00:00", "epoch": 0, "step": 12556, "train_loss": 3.5016491413116455, "perplexity": 33.17010907483788, "lr": 0.0026291804804649314, "grad_norm": 0.143395, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:50.546082+00:00", "epoch": 0, "step": 12557, "train_loss": 3.4653425216674805, "perplexity": 31.987414279422932, "lr": 0.0026291804804649314, "grad_norm": 0.15616, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:50.850942+00:00", "epoch": 0, "step": 12558, "train_loss": 3.485419750213623, "perplexity": 32.6361232548865, "lr": 0.0026291804804649314, "grad_norm": 0.157537, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:51.155343+00:00", "epoch": 0, "step": 12559, "train_loss": 3.5528411865234375, "perplexity": 34.91236926173521, "lr": 0.0026291804804649314, "grad_norm": 0.168229, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:51.460261+00:00", "epoch": 0, "step": 12560, "train_loss": 3.4186599254608154, "perplexity": 30.528477162320172, "lr": 0.0026291804804649314, "grad_norm": 0.17749, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:51.765053+00:00", "epoch": 0, "step": 12561, "train_loss": 3.594987154006958, "perplexity": 36.415232195055026, "lr": 0.0026291804804649314, "grad_norm": 0.167549, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:52.070188+00:00", "epoch": 0, "step": 12562, "train_loss": 3.475349187850952, "perplexity": 32.30910851524893, "lr": 0.0026291804804649314, "grad_norm": 0.244262, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:52.375382+00:00", "epoch": 0, "step": 12563, "train_loss": 3.4796342849731445, "perplexity": 32.44785323817153, "lr": 0.0026291804804649314, "grad_norm": 0.198956, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:52.679552+00:00", "epoch": 0, "step": 12564, "train_loss": 3.4678499698638916, "perplexity": 32.06772170494399, "lr": 0.0026291804804649314, "grad_norm": 0.160601, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:52.983960+00:00", "epoch": 0, "step": 12565, "train_loss": 3.4645416736602783, "perplexity": 31.96180747738849, "lr": 0.0026291804804649314, "grad_norm": 0.190455, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:53.287750+00:00", "epoch": 0, "step": 12566, "train_loss": 3.508387565612793, "perplexity": 33.39437810523895, "lr": 0.0026291804804649314, "grad_norm": 0.184545, "tokens_per_sec": 107864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:53.593156+00:00", "epoch": 0, "step": 12567, "train_loss": 3.5453662872314453, "perplexity": 34.652375740726576, "lr": 0.0026291804804649314, "grad_norm": 0.159152, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:53.898836+00:00", "epoch": 0, "step": 12568, "train_loss": 3.498685121536255, "perplexity": 33.071937778293304, "lr": 0.0026291804804649314, "grad_norm": 0.181137, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:54.204788+00:00", "epoch": 0, "step": 12569, "train_loss": 3.5100278854370117, "perplexity": 33.44920051651412, "lr": 0.0026291804804649314, "grad_norm": 0.187415, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:54.509856+00:00", "epoch": 0, "step": 12570, "train_loss": 3.454151153564453, "perplexity": 31.631427059080842, "lr": 0.0026291804804649314, "grad_norm": 0.203453, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:54.814962+00:00", "epoch": 0, "step": 12571, "train_loss": 3.5718188285827637, "perplexity": 35.58125052893339, "lr": 0.0026291804804649314, "grad_norm": 0.173511, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:55.119918+00:00", "epoch": 0, "step": 12572, "train_loss": 3.6073861122131348, "perplexity": 36.86955387448537, "lr": 0.0026291804804649314, "grad_norm": 0.18349, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:55.425179+00:00", "epoch": 0, "step": 12573, "train_loss": 3.455544948577881, "perplexity": 31.675545523288818, "lr": 0.0026291804804649314, "grad_norm": 0.161356, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:55.730585+00:00", "epoch": 0, "step": 12574, "train_loss": 3.518644332885742, "perplexity": 33.738659056257184, "lr": 0.0026291804804649314, "grad_norm": 0.162392, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:56.034159+00:00", "epoch": 0, "step": 12575, "train_loss": 3.565683603286743, "perplexity": 35.363619829961465, "lr": 0.0026291804804649314, "grad_norm": 0.180029, "tokens_per_sec": 107941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:56.338416+00:00", "epoch": 0, "step": 12576, "train_loss": 3.5108344554901123, "perplexity": 33.476190523153235, "lr": 0.0026291804804649314, "grad_norm": 0.170966, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:56.642142+00:00", "epoch": 0, "step": 12577, "train_loss": 3.508049726486206, "perplexity": 33.38309808323064, "lr": 0.0026291804804649314, "grad_norm": 0.169986, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:56.947498+00:00", "epoch": 0, "step": 12578, "train_loss": 3.511746883392334, "perplexity": 33.50674907256315, "lr": 0.0026291804804649314, "grad_norm": 0.188033, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:57.252213+00:00", "epoch": 0, "step": 12579, "train_loss": 3.5579748153686523, "perplexity": 35.092057238696135, "lr": 0.0026291804804649314, "grad_norm": 0.163146, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:57.557825+00:00", "epoch": 0, "step": 12580, "train_loss": 3.5172648429870605, "perplexity": 33.69214900434311, "lr": 0.0026291804804649314, "grad_norm": 0.157591, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:57.863226+00:00", "epoch": 0, "step": 12581, "train_loss": 3.5879483222961426, "perplexity": 36.15981149017293, "lr": 0.0026291804804649314, "grad_norm": 0.147162, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:58.167273+00:00", "epoch": 0, "step": 12582, "train_loss": 3.5446133613586426, "perplexity": 34.62629489018298, "lr": 0.0026291804804649314, "grad_norm": 0.162998, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:58.472549+00:00", "epoch": 0, "step": 12583, "train_loss": 3.5502965450286865, "perplexity": 34.823642734709736, "lr": 0.0026291804804649314, "grad_norm": 0.151367, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:58.778557+00:00", "epoch": 0, "step": 12584, "train_loss": 3.5583698749542236, "perplexity": 35.10592343109146, "lr": 0.0026291804804649314, "grad_norm": 0.167083, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:59.084447+00:00", "epoch": 0, "step": 12585, "train_loss": 3.446584939956665, "perplexity": 31.3930000574911, "lr": 0.0026291804804649314, "grad_norm": 0.163504, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:59.388582+00:00", "epoch": 0, "step": 12586, "train_loss": 3.4755687713623047, "perplexity": 32.31620384172522, "lr": 0.0026291804804649314, "grad_norm": 0.161323, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:59.692829+00:00", "epoch": 0, "step": 12587, "train_loss": 3.4970693588256836, "perplexity": 33.01854452149757, "lr": 0.0026291804804649314, "grad_norm": 0.150909, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:03:59.996982+00:00", "epoch": 0, "step": 12588, "train_loss": 3.5771751403808594, "perplexity": 35.772346127859336, "lr": 0.0026291804804649314, "grad_norm": 0.147707, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:00.302235+00:00", "epoch": 0, "step": 12589, "train_loss": 3.53291654586792, "perplexity": 34.22363700602287, "lr": 0.0026291804804649314, "grad_norm": 0.147741, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:00.607693+00:00", "epoch": 0, "step": 12590, "train_loss": 3.504718780517578, "perplexity": 33.27208577788743, "lr": 0.0026291804804649314, "grad_norm": 0.163828, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:00.912090+00:00", "epoch": 0, "step": 12591, "train_loss": 3.5650980472564697, "perplexity": 35.34291851060043, "lr": 0.0026291804804649314, "grad_norm": 0.190078, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:01.217700+00:00", "epoch": 0, "step": 12592, "train_loss": 3.63370418548584, "perplexity": 37.85277093347285, "lr": 0.0026291804804649314, "grad_norm": 0.179563, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:01.521892+00:00", "epoch": 0, "step": 12593, "train_loss": 3.453350782394409, "perplexity": 31.606120305544323, "lr": 0.0026291804804649314, "grad_norm": 0.180842, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:01.826940+00:00", "epoch": 0, "step": 12594, "train_loss": 3.5169804096221924, "perplexity": 33.68256719578979, "lr": 0.0026291804804649314, "grad_norm": 0.178728, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:02.132094+00:00", "epoch": 0, "step": 12595, "train_loss": 3.5344038009643555, "perplexity": 34.27457415345559, "lr": 0.0026291804804649314, "grad_norm": 0.174145, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:02.437197+00:00", "epoch": 0, "step": 12596, "train_loss": 3.52730131149292, "perplexity": 34.03200180488848, "lr": 0.0026291804804649314, "grad_norm": 0.193879, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:02.741701+00:00", "epoch": 0, "step": 12597, "train_loss": 3.5222902297973633, "perplexity": 33.86189123852305, "lr": 0.0026291804804649314, "grad_norm": 0.186633, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:03.046902+00:00", "epoch": 0, "step": 12598, "train_loss": 3.4172821044921875, "perplexity": 30.486443350529, "lr": 0.0026291804804649314, "grad_norm": 0.18628, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:03.352879+00:00", "epoch": 0, "step": 12599, "train_loss": 3.483647346496582, "perplexity": 32.578330100250135, "lr": 0.0026291804804649314, "grad_norm": 0.174159, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:03.657870+00:00", "epoch": 0, "step": 12600, "train_loss": 3.446388006210327, "perplexity": 31.38681832509683, "lr": 0.0026291804804649314, "grad_norm": 0.200065, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:03.963384+00:00", "epoch": 0, "step": 12601, "train_loss": 3.3737053871154785, "perplexity": 29.18647412662467, "lr": 0.0026291804804649314, "grad_norm": 0.188645, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:04.267533+00:00", "epoch": 0, "step": 12602, "train_loss": 3.4407756328582764, "perplexity": 31.21115718137614, "lr": 0.0026291804804649314, "grad_norm": 0.174835, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:04.572402+00:00", "epoch": 0, "step": 12603, "train_loss": 3.434349298477173, "perplexity": 31.01122694623505, "lr": 0.0026291804804649314, "grad_norm": 0.194916, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:04.877202+00:00", "epoch": 0, "step": 12604, "train_loss": 3.5021345615386963, "perplexity": 33.18621442533149, "lr": 0.0026291804804649314, "grad_norm": 0.190691, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:05.181843+00:00", "epoch": 0, "step": 12605, "train_loss": 3.5825207233428955, "perplexity": 35.96408218575926, "lr": 0.0026291804804649314, "grad_norm": 0.177537, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:05.487223+00:00", "epoch": 0, "step": 12606, "train_loss": 3.544645309448242, "perplexity": 34.627401151826014, "lr": 0.0026291804804649314, "grad_norm": 0.162731, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:05.793834+00:00", "epoch": 0, "step": 12607, "train_loss": 3.4724724292755127, "perplexity": 32.21629657304548, "lr": 0.0026291804804649314, "grad_norm": 0.173169, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:06.098640+00:00", "epoch": 0, "step": 12608, "train_loss": 3.4798591136932373, "perplexity": 32.45514926763258, "lr": 0.0026291804804649314, "grad_norm": 0.157224, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:06.402775+00:00", "epoch": 0, "step": 12609, "train_loss": 3.4415245056152344, "perplexity": 31.234539120658336, "lr": 0.0026291804804649314, "grad_norm": 0.181484, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:06.707482+00:00", "epoch": 0, "step": 12610, "train_loss": 3.400827646255493, "perplexity": 29.988909988111498, "lr": 0.0026291804804649314, "grad_norm": 0.151792, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:07.013167+00:00", "epoch": 0, "step": 12611, "train_loss": 3.566720962524414, "perplexity": 35.40032364189729, "lr": 0.0026291804804649314, "grad_norm": 0.175295, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:07.318162+00:00", "epoch": 0, "step": 12612, "train_loss": 3.595593214035034, "perplexity": 36.437308700869686, "lr": 0.0026291804804649314, "grad_norm": 0.173175, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:07.624427+00:00", "epoch": 0, "step": 12613, "train_loss": 3.6111338138580322, "perplexity": 37.00798920731963, "lr": 0.0026291804804649314, "grad_norm": 0.183566, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:07.928866+00:00", "epoch": 0, "step": 12614, "train_loss": 3.514488935470581, "perplexity": 33.598752404810476, "lr": 0.0026291804804649314, "grad_norm": 0.162946, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:08.234775+00:00", "epoch": 0, "step": 12615, "train_loss": 3.523732900619507, "perplexity": 33.91077805630926, "lr": 0.0026291804804649314, "grad_norm": 0.150092, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:08.539585+00:00", "epoch": 0, "step": 12616, "train_loss": 3.5112476348876953, "perplexity": 33.490025053261554, "lr": 0.0026291804804649314, "grad_norm": 0.158998, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:08.845695+00:00", "epoch": 0, "step": 12617, "train_loss": 3.4573569297790527, "perplexity": 31.73299304751285, "lr": 0.0026291804804649314, "grad_norm": 0.156316, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:09.149910+00:00", "epoch": 0, "step": 12618, "train_loss": 3.437523126602173, "perplexity": 31.109807606792543, "lr": 0.0026291804804649314, "grad_norm": 0.156958, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:09.454154+00:00", "epoch": 0, "step": 12619, "train_loss": 3.5842232704162598, "perplexity": 36.025364882165384, "lr": 0.0026291804804649314, "grad_norm": 0.167682, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:09.758365+00:00", "epoch": 0, "step": 12620, "train_loss": 3.604764699935913, "perplexity": 36.77303014282504, "lr": 0.0026291804804649314, "grad_norm": 0.152473, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:10.063555+00:00", "epoch": 0, "step": 12621, "train_loss": 3.5343992710113525, "perplexity": 34.27441889159714, "lr": 0.0026291804804649314, "grad_norm": 0.175304, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:10.368855+00:00", "epoch": 0, "step": 12622, "train_loss": 3.5649120807647705, "perplexity": 35.33634652314227, "lr": 0.0026291804804649314, "grad_norm": 0.167959, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:10.674683+00:00", "epoch": 0, "step": 12623, "train_loss": 3.5157551765441895, "perplexity": 33.641323472056314, "lr": 0.0026291804804649314, "grad_norm": 0.190606, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:10.979511+00:00", "epoch": 0, "step": 12624, "train_loss": 3.5348668098449707, "perplexity": 34.29044726007904, "lr": 0.0026291804804649314, "grad_norm": 0.151795, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:11.284690+00:00", "epoch": 0, "step": 12625, "train_loss": 3.625563383102417, "perplexity": 37.545869910853824, "lr": 0.0026291804804649314, "grad_norm": 0.167234, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:11.590148+00:00", "epoch": 0, "step": 12626, "train_loss": 3.5052554607391357, "perplexity": 33.289947040718936, "lr": 0.0026291804804649314, "grad_norm": 0.162263, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:11.895623+00:00", "epoch": 0, "step": 12627, "train_loss": 3.52596116065979, "perplexity": 33.98642433647915, "lr": 0.0026291804804649314, "grad_norm": 0.166831, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:12.201776+00:00", "epoch": 0, "step": 12628, "train_loss": 3.5394270420074463, "perplexity": 34.44717675036291, "lr": 0.0026291804804649314, "grad_norm": 0.185694, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:12.506678+00:00", "epoch": 0, "step": 12629, "train_loss": 3.52622389793396, "perplexity": 33.995355010127156, "lr": 0.0026291804804649314, "grad_norm": 0.197506, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:12.811589+00:00", "epoch": 0, "step": 12630, "train_loss": 3.5346531867980957, "perplexity": 34.28312281261989, "lr": 0.0026291804804649314, "grad_norm": 0.176105, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:13.116750+00:00", "epoch": 0, "step": 12631, "train_loss": 3.4607014656066895, "perplexity": 31.839302859067455, "lr": 0.0026291804804649314, "grad_norm": 0.156186, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:13.421892+00:00", "epoch": 0, "step": 12632, "train_loss": 3.4275896549224854, "perplexity": 30.80230900678926, "lr": 0.0026291804804649314, "grad_norm": 0.170361, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:13.727578+00:00", "epoch": 0, "step": 12633, "train_loss": 3.559554100036621, "perplexity": 35.14752137196039, "lr": 0.0026291804804649314, "grad_norm": 0.16894, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:14.032186+00:00", "epoch": 0, "step": 12634, "train_loss": 3.492870330810547, "perplexity": 32.880189409815586, "lr": 0.0026291804804649314, "grad_norm": 0.142785, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:14.337563+00:00", "epoch": 0, "step": 12635, "train_loss": 3.546781301498413, "perplexity": 34.70144405477894, "lr": 0.0026291804804649314, "grad_norm": 0.171123, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:14.642272+00:00", "epoch": 0, "step": 12636, "train_loss": 3.519186019897461, "perplexity": 33.75693980043461, "lr": 0.0026291804804649314, "grad_norm": 0.171481, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:14.947401+00:00", "epoch": 0, "step": 12637, "train_loss": 3.4745309352874756, "perplexity": 32.28268231750915, "lr": 0.0026291804804649314, "grad_norm": 0.157497, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:15.253465+00:00", "epoch": 0, "step": 12638, "train_loss": 3.5299623012542725, "perplexity": 34.12268120822415, "lr": 0.0026291804804649314, "grad_norm": 0.155947, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:15.558150+00:00", "epoch": 0, "step": 12639, "train_loss": 3.5202667713165283, "perplexity": 33.79344238258304, "lr": 0.0026291804804649314, "grad_norm": 0.155259, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:15.862168+00:00", "epoch": 0, "step": 12640, "train_loss": 3.377575397491455, "perplexity": 29.299644929176573, "lr": 0.0026291804804649314, "grad_norm": 0.170711, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:16.167295+00:00", "epoch": 0, "step": 12641, "train_loss": 3.576192617416382, "perplexity": 35.73721623708374, "lr": 0.0026291804804649314, "grad_norm": 0.177952, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:16.471482+00:00", "epoch": 0, "step": 12642, "train_loss": 3.3862714767456055, "perplexity": 29.55554802776023, "lr": 0.0026291804804649314, "grad_norm": 0.159172, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:16.776429+00:00", "epoch": 0, "step": 12643, "train_loss": 3.48435378074646, "perplexity": 32.60135267944856, "lr": 0.0026291804804649314, "grad_norm": 0.150562, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:17.082410+00:00", "epoch": 0, "step": 12644, "train_loss": 3.482901096343994, "perplexity": 32.55402758544727, "lr": 0.0026291804804649314, "grad_norm": 0.16274, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:17.386821+00:00", "epoch": 0, "step": 12645, "train_loss": 3.4301295280456543, "perplexity": 30.88064239989662, "lr": 0.0026291804804649314, "grad_norm": 0.188189, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:17.691331+00:00", "epoch": 0, "step": 12646, "train_loss": 3.4689254760742188, "perplexity": 32.10222929202452, "lr": 0.0026291804804649314, "grad_norm": 0.180262, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:17.997411+00:00", "epoch": 0, "step": 12647, "train_loss": 3.5972111225128174, "perplexity": 36.49630864691088, "lr": 0.0026291804804649314, "grad_norm": 0.150404, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:18.302866+00:00", "epoch": 0, "step": 12648, "train_loss": 3.585963726043701, "perplexity": 36.08812002663594, "lr": 0.0026291804804649314, "grad_norm": 0.180145, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:18.607543+00:00", "epoch": 0, "step": 12649, "train_loss": 3.47328782081604, "perplexity": 32.2425761813674, "lr": 0.0026291804804649314, "grad_norm": 0.151834, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:18.913106+00:00", "epoch": 0, "step": 12650, "train_loss": 3.558023452758789, "perplexity": 35.09376406628223, "lr": 0.0026291804804649314, "grad_norm": 0.165422, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:19.216965+00:00", "epoch": 0, "step": 12651, "train_loss": 3.5013952255249023, "perplexity": 33.16168772969727, "lr": 0.0026291804804649314, "grad_norm": 0.156236, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:19.521280+00:00", "epoch": 0, "step": 12652, "train_loss": 3.6016597747802734, "perplexity": 36.65902970947778, "lr": 0.0026291804804649314, "grad_norm": 0.174954, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:19.827559+00:00", "epoch": 0, "step": 12653, "train_loss": 3.395664930343628, "perplexity": 29.834484735190987, "lr": 0.0026291804804649314, "grad_norm": 0.15236, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:20.133994+00:00", "epoch": 0, "step": 12654, "train_loss": 3.4913132190704346, "perplexity": 32.829031120791186, "lr": 0.0026291804804649314, "grad_norm": 0.147532, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:20.440415+00:00", "epoch": 0, "step": 12655, "train_loss": 3.4618213176727295, "perplexity": 31.874978139945686, "lr": 0.0026291804804649314, "grad_norm": 0.157321, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:20.745051+00:00", "epoch": 0, "step": 12656, "train_loss": 3.608137845993042, "perplexity": 36.897280383754044, "lr": 0.0026291804804649314, "grad_norm": 0.175435, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:21.048950+00:00", "epoch": 0, "step": 12657, "train_loss": 3.5920510292053223, "perplexity": 36.30846933988197, "lr": 0.0026291804804649314, "grad_norm": 0.179892, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:21.354230+00:00", "epoch": 0, "step": 12658, "train_loss": 3.4607534408569336, "perplexity": 31.840957757807672, "lr": 0.0026291804804649314, "grad_norm": 0.153179, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:21.660358+00:00", "epoch": 0, "step": 12659, "train_loss": 3.615934133529663, "perplexity": 37.18606645782965, "lr": 0.0026291804804649314, "grad_norm": 0.15416, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:21.966082+00:00", "epoch": 0, "step": 12660, "train_loss": 3.5456955432891846, "perplexity": 34.663787123884696, "lr": 0.0026291804804649314, "grad_norm": 0.162892, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:22.270830+00:00", "epoch": 0, "step": 12661, "train_loss": 3.5597496032714844, "perplexity": 35.154393497825566, "lr": 0.0026291804804649314, "grad_norm": 0.135317, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:22.576377+00:00", "epoch": 0, "step": 12662, "train_loss": 3.4947710037231445, "perplexity": 32.942743323621045, "lr": 0.0026291804804649314, "grad_norm": 0.161175, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:22.881188+00:00", "epoch": 0, "step": 12663, "train_loss": 3.5036303997039795, "perplexity": 33.23589277755472, "lr": 0.0026291804804649314, "grad_norm": 0.170868, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:23.186496+00:00", "epoch": 0, "step": 12664, "train_loss": 3.520864486694336, "perplexity": 33.813647280551734, "lr": 0.0026291804804649314, "grad_norm": 0.170236, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:23.492483+00:00", "epoch": 0, "step": 12665, "train_loss": 3.5517711639404297, "perplexity": 34.87503221750152, "lr": 0.0026291804804649314, "grad_norm": 0.182675, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:23.797461+00:00", "epoch": 0, "step": 12666, "train_loss": 3.5914347171783447, "perplexity": 36.28609888784407, "lr": 0.0026291804804649314, "grad_norm": 0.189778, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:24.102576+00:00", "epoch": 0, "step": 12667, "train_loss": 3.6632254123687744, "perplexity": 38.986889087615204, "lr": 0.0026291804804649314, "grad_norm": 0.180461, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:24.407326+00:00", "epoch": 0, "step": 12668, "train_loss": 3.5035781860351562, "perplexity": 33.23415745496036, "lr": 0.0026291804804649314, "grad_norm": 0.166997, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:24.712960+00:00", "epoch": 0, "step": 12669, "train_loss": 3.659773588180542, "perplexity": 38.852545199972354, "lr": 0.0026291804804649314, "grad_norm": 0.177573, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:25.018445+00:00", "epoch": 0, "step": 12670, "train_loss": 3.5242555141448975, "perplexity": 33.92850491932373, "lr": 0.0026291804804649314, "grad_norm": 0.169548, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:25.324020+00:00", "epoch": 0, "step": 12671, "train_loss": 3.575927734375, "perplexity": 35.72775130816127, "lr": 0.0026291804804649314, "grad_norm": 0.178694, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:25.629519+00:00", "epoch": 0, "step": 12672, "train_loss": 3.4370150566101074, "perplexity": 31.094005661675588, "lr": 0.0026291804804649314, "grad_norm": 0.155274, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:25.934953+00:00", "epoch": 0, "step": 12673, "train_loss": 3.5156004428863525, "perplexity": 33.63611842972901, "lr": 0.0026291804804649314, "grad_norm": 0.151653, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:26.239656+00:00", "epoch": 0, "step": 12674, "train_loss": 3.434210777282715, "perplexity": 31.006931531546705, "lr": 0.0026291804804649314, "grad_norm": 0.148303, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:26.546173+00:00", "epoch": 0, "step": 12675, "train_loss": 3.550981044769287, "perplexity": 34.8474876691231, "lr": 0.0026291804804649314, "grad_norm": 0.160405, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:26.851015+00:00", "epoch": 0, "step": 12676, "train_loss": 3.5889577865600586, "perplexity": 36.19633195761433, "lr": 0.0026291804804649314, "grad_norm": 0.14559, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:27.157275+00:00", "epoch": 0, "step": 12677, "train_loss": 3.5121891498565674, "perplexity": 33.521571261440364, "lr": 0.0026291804804649314, "grad_norm": 0.162466, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:27.461738+00:00", "epoch": 0, "step": 12678, "train_loss": 3.5630075931549072, "perplexity": 35.26911293210342, "lr": 0.0026291804804649314, "grad_norm": 0.192155, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:27.766710+00:00", "epoch": 0, "step": 12679, "train_loss": 3.51883864402771, "perplexity": 33.74521549060036, "lr": 0.0026291804804649314, "grad_norm": 0.189654, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:28.072960+00:00", "epoch": 0, "step": 12680, "train_loss": 3.5509798526763916, "perplexity": 34.84744612770538, "lr": 0.0026291804804649314, "grad_norm": 0.171535, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:28.378991+00:00", "epoch": 0, "step": 12681, "train_loss": 3.4775524139404297, "perplexity": 32.380371261201034, "lr": 0.0026291804804649314, "grad_norm": 0.180278, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:28.684751+00:00", "epoch": 0, "step": 12682, "train_loss": 3.505941152572632, "perplexity": 33.3127815133533, "lr": 0.0026291804804649314, "grad_norm": 0.182602, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:28.991121+00:00", "epoch": 0, "step": 12683, "train_loss": 3.5674452781677246, "perplexity": 35.42597393842347, "lr": 0.0026291804804649314, "grad_norm": 0.184988, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:29.295818+00:00", "epoch": 0, "step": 12684, "train_loss": 3.5954976081848145, "perplexity": 36.43382524751354, "lr": 0.0026291804804649314, "grad_norm": 0.19691, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:29.601370+00:00", "epoch": 0, "step": 12685, "train_loss": 3.502100944519043, "perplexity": 33.18509882246066, "lr": 0.0026291804804649314, "grad_norm": 0.183442, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:29.906018+00:00", "epoch": 0, "step": 12686, "train_loss": 3.3957359790802, "perplexity": 29.836604512940582, "lr": 0.0026291804804649314, "grad_norm": 0.179922, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:30.211620+00:00", "epoch": 0, "step": 12687, "train_loss": 3.515496253967285, "perplexity": 33.63261410146745, "lr": 0.0026291804804649314, "grad_norm": 0.173603, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:30.516021+00:00", "epoch": 0, "step": 12688, "train_loss": 3.422318458557129, "perplexity": 30.64037116554847, "lr": 0.0026291804804649314, "grad_norm": 0.182942, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:30.821443+00:00", "epoch": 0, "step": 12689, "train_loss": 3.576677083969116, "perplexity": 35.754533917618225, "lr": 0.0026291804804649314, "grad_norm": 0.199114, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:31.126565+00:00", "epoch": 0, "step": 12690, "train_loss": 3.4528279304504395, "perplexity": 31.589599303496612, "lr": 0.0026291804804649314, "grad_norm": 0.190871, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:31.431982+00:00", "epoch": 0, "step": 12691, "train_loss": 3.631319284439087, "perplexity": 37.76260338353079, "lr": 0.0026291804804649314, "grad_norm": 0.205249, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:31.736401+00:00", "epoch": 0, "step": 12692, "train_loss": 3.5530331134796143, "perplexity": 34.91907052955693, "lr": 0.0026291804804649314, "grad_norm": 0.175381, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:32.042300+00:00", "epoch": 0, "step": 12693, "train_loss": 3.5854508876800537, "perplexity": 36.06961739904852, "lr": 0.0026291804804649314, "grad_norm": 0.161766, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:32.348010+00:00", "epoch": 0, "step": 12694, "train_loss": 3.57881236076355, "perplexity": 35.83096131197048, "lr": 0.0026291804804649314, "grad_norm": 0.187306, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:32.653011+00:00", "epoch": 0, "step": 12695, "train_loss": 3.563143014907837, "perplexity": 35.27388946061659, "lr": 0.0026291804804649314, "grad_norm": 0.155867, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:32.957738+00:00", "epoch": 0, "step": 12696, "train_loss": 3.5372321605682373, "perplexity": 34.37165219552219, "lr": 0.0026291804804649314, "grad_norm": 0.147435, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:33.262601+00:00", "epoch": 0, "step": 12697, "train_loss": 3.510535478591919, "perplexity": 33.466183411569425, "lr": 0.0026291804804649314, "grad_norm": 0.150531, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:33.568422+00:00", "epoch": 0, "step": 12698, "train_loss": 3.4597885608673096, "perplexity": 31.810249871911413, "lr": 0.0026291804804649314, "grad_norm": 0.149309, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:33.873837+00:00", "epoch": 0, "step": 12699, "train_loss": 3.4913132190704346, "perplexity": 32.829031120791186, "lr": 0.0026291804804649314, "grad_norm": 0.148565, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:34.179007+00:00", "epoch": 0, "step": 12700, "train_loss": 3.5235588550567627, "perplexity": 33.904876549439805, "lr": 0.0026291804804649314, "grad_norm": 0.171567, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:34.483929+00:00", "epoch": 0, "step": 12701, "train_loss": 3.53739595413208, "perplexity": 34.37728251202267, "lr": 0.0026291804804649314, "grad_norm": 0.179854, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:34.788611+00:00", "epoch": 0, "step": 12702, "train_loss": 3.59586238861084, "perplexity": 36.447118018132926, "lr": 0.0026291804804649314, "grad_norm": 0.175319, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:35.094539+00:00", "epoch": 0, "step": 12703, "train_loss": 3.5095341205596924, "perplexity": 33.43268855297395, "lr": 0.0026291804804649314, "grad_norm": 0.160885, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:35.400695+00:00", "epoch": 0, "step": 12704, "train_loss": 3.57651424407959, "perplexity": 35.748712127287725, "lr": 0.0026291804804649314, "grad_norm": 0.176094, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:35.704989+00:00", "epoch": 0, "step": 12705, "train_loss": 3.5823135375976562, "perplexity": 35.95663171243234, "lr": 0.0026291804804649314, "grad_norm": 0.148173, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:36.008938+00:00", "epoch": 0, "step": 12706, "train_loss": 3.555723190307617, "perplexity": 35.013131971573614, "lr": 0.0026291804804649314, "grad_norm": 0.143899, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:36.313944+00:00", "epoch": 0, "step": 12707, "train_loss": 3.5997722148895264, "perplexity": 36.589898860199035, "lr": 0.0026291804804649314, "grad_norm": 0.150177, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:36.618102+00:00", "epoch": 0, "step": 12708, "train_loss": 3.557218313217163, "perplexity": 35.06552006087652, "lr": 0.0026291804804649314, "grad_norm": 0.158094, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:36.924156+00:00", "epoch": 0, "step": 12709, "train_loss": 3.615795135498047, "perplexity": 37.18089802699764, "lr": 0.0026291804804649314, "grad_norm": 0.179565, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:37.228856+00:00", "epoch": 0, "step": 12710, "train_loss": 3.501722812652588, "perplexity": 33.17255285126967, "lr": 0.0026291804804649314, "grad_norm": 0.168375, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:37.534131+00:00", "epoch": 0, "step": 12711, "train_loss": 3.505110025405884, "perplexity": 33.28510585822516, "lr": 0.0026291804804649314, "grad_norm": 0.181478, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:37.839377+00:00", "epoch": 0, "step": 12712, "train_loss": 3.5482187271118164, "perplexity": 34.751360666398625, "lr": 0.0026291804804649314, "grad_norm": 0.160307, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:38.145375+00:00", "epoch": 0, "step": 12713, "train_loss": 3.566648244857788, "perplexity": 35.397749506557936, "lr": 0.0026291804804649314, "grad_norm": 0.184642, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:38.450657+00:00", "epoch": 0, "step": 12714, "train_loss": 3.5152132511138916, "perplexity": 33.623097322410985, "lr": 0.0026291804804649314, "grad_norm": 0.196866, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:38.756939+00:00", "epoch": 0, "step": 12715, "train_loss": 3.604146957397461, "perplexity": 36.75032089279509, "lr": 0.0026291804804649314, "grad_norm": 0.146078, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:39.061903+00:00", "epoch": 0, "step": 12716, "train_loss": 3.592881917953491, "perplexity": 36.338650175241625, "lr": 0.0026291804804649314, "grad_norm": 0.157231, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:39.367310+00:00", "epoch": 0, "step": 12717, "train_loss": 3.4525585174560547, "perplexity": 31.581089801293547, "lr": 0.0026291804804649314, "grad_norm": 0.148523, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:39.672359+00:00", "epoch": 0, "step": 12718, "train_loss": 3.5476272106170654, "perplexity": 34.73081074175829, "lr": 0.0026291804804649314, "grad_norm": 0.157883, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:39.978275+00:00", "epoch": 0, "step": 12719, "train_loss": 3.4234118461608887, "perplexity": 30.673891289460027, "lr": 0.0026291804804649314, "grad_norm": 0.16893, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:40.284020+00:00", "epoch": 0, "step": 12720, "train_loss": 3.5054843425750732, "perplexity": 33.297567376960885, "lr": 0.0026291804804649314, "grad_norm": 0.148707, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:40.590564+00:00", "epoch": 0, "step": 12721, "train_loss": 3.427013397216797, "perplexity": 30.784564052195563, "lr": 0.0026291804804649314, "grad_norm": 0.145962, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:40.896120+00:00", "epoch": 0, "step": 12722, "train_loss": 3.4818224906921387, "perplexity": 32.518933557013995, "lr": 0.0026291804804649314, "grad_norm": 0.182747, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:41.201719+00:00", "epoch": 0, "step": 12723, "train_loss": 3.6046788692474365, "perplexity": 36.76987402377839, "lr": 0.0026291804804649314, "grad_norm": 0.186235, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:41.507102+00:00", "epoch": 0, "step": 12724, "train_loss": 3.5665292739868164, "perplexity": 35.393538455969846, "lr": 0.0026291804804649314, "grad_norm": 0.191957, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:41.811882+00:00", "epoch": 0, "step": 12725, "train_loss": 3.5966923236846924, "perplexity": 36.477379315436046, "lr": 0.0026291804804649314, "grad_norm": 0.167049, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:42.118377+00:00", "epoch": 0, "step": 12726, "train_loss": 3.5559470653533936, "perplexity": 35.020971415591724, "lr": 0.0026291804804649314, "grad_norm": 0.176219, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:42.423162+00:00", "epoch": 0, "step": 12727, "train_loss": 3.5981104373931885, "perplexity": 36.52914508329051, "lr": 0.0026291804804649314, "grad_norm": 0.15513, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:42.727032+00:00", "epoch": 0, "step": 12728, "train_loss": 3.5267560482025146, "perplexity": 34.0134504617481, "lr": 0.0026291804804649314, "grad_norm": 0.187192, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:43.032160+00:00", "epoch": 0, "step": 12729, "train_loss": 3.569849729537964, "perplexity": 35.5112564577494, "lr": 0.0026291804804649314, "grad_norm": 0.175017, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:43.337963+00:00", "epoch": 0, "step": 12730, "train_loss": 3.5447607040405273, "perplexity": 34.6313971972204, "lr": 0.0026291804804649314, "grad_norm": 0.173019, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:43.644711+00:00", "epoch": 0, "step": 12731, "train_loss": 3.423753499984741, "perplexity": 30.684372932156204, "lr": 0.0026291804804649314, "grad_norm": 0.19115, "tokens_per_sec": 106765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:43.950129+00:00", "epoch": 0, "step": 12732, "train_loss": 3.4221742153167725, "perplexity": 30.635951817864004, "lr": 0.0026291804804649314, "grad_norm": 0.182626, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:44.254784+00:00", "epoch": 0, "step": 12733, "train_loss": 3.6593124866485596, "perplexity": 38.8346343615341, "lr": 0.0026291804804649314, "grad_norm": 0.148427, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:44.559887+00:00", "epoch": 0, "step": 12734, "train_loss": 3.605954170227051, "perplexity": 36.816796593981536, "lr": 0.0026291804804649314, "grad_norm": 0.172503, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:44.864753+00:00", "epoch": 0, "step": 12735, "train_loss": 3.4747071266174316, "perplexity": 32.288370747352516, "lr": 0.0026291804804649314, "grad_norm": 0.173258, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:45.169939+00:00", "epoch": 0, "step": 12736, "train_loss": 3.6071293354034424, "perplexity": 36.86008784344734, "lr": 0.0026291804804649314, "grad_norm": 0.173536, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:45.476320+00:00", "epoch": 0, "step": 12737, "train_loss": 3.742464780807495, "perplexity": 42.20188046400153, "lr": 0.0026291804804649314, "grad_norm": 0.164161, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:45.781668+00:00", "epoch": 0, "step": 12738, "train_loss": 3.5165674686431885, "perplexity": 33.668661154901265, "lr": 0.0026291804804649314, "grad_norm": 0.170556, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:46.086548+00:00", "epoch": 0, "step": 12739, "train_loss": 3.520481824874878, "perplexity": 33.800710564112435, "lr": 0.0026291804804649314, "grad_norm": 0.152189, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:46.391800+00:00", "epoch": 0, "step": 12740, "train_loss": 3.498011827468872, "perplexity": 33.049678133274696, "lr": 0.0026291804804649314, "grad_norm": 0.150645, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:46.696784+00:00", "epoch": 0, "step": 12741, "train_loss": 3.486522912979126, "perplexity": 32.67214607673632, "lr": 0.0026291804804649314, "grad_norm": 0.131503, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:47.001836+00:00", "epoch": 0, "step": 12742, "train_loss": 3.490375518798828, "perplexity": 32.79826175786721, "lr": 0.0026291804804649314, "grad_norm": 0.150517, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:47.307399+00:00", "epoch": 0, "step": 12743, "train_loss": 3.602566957473755, "perplexity": 36.692301236184825, "lr": 0.0026291804804649314, "grad_norm": 0.153912, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:47.611653+00:00", "epoch": 0, "step": 12744, "train_loss": 3.5418848991394043, "perplexity": 34.53194712347242, "lr": 0.0026291804804649314, "grad_norm": 0.1667, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:47.916204+00:00", "epoch": 0, "step": 12745, "train_loss": 3.535693407058716, "perplexity": 34.31880336616688, "lr": 0.0026291804804649314, "grad_norm": 0.161986, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:48.221800+00:00", "epoch": 0, "step": 12746, "train_loss": 3.52143931388855, "perplexity": 33.833089872073955, "lr": 0.0026291804804649314, "grad_norm": 0.162597, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:48.528213+00:00", "epoch": 0, "step": 12747, "train_loss": 3.4606103897094727, "perplexity": 31.836403198039427, "lr": 0.0026291804804649314, "grad_norm": 0.158954, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:48.834528+00:00", "epoch": 0, "step": 12748, "train_loss": 3.580564260482788, "perplexity": 35.89378858050114, "lr": 0.0026291804804649314, "grad_norm": 0.175097, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:49.140648+00:00", "epoch": 0, "step": 12749, "train_loss": 3.5634303092956543, "perplexity": 35.284024906953825, "lr": 0.0026291804804649314, "grad_norm": 0.161322, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:49.445536+00:00", "epoch": 0, "step": 12750, "train_loss": 3.5892422199249268, "perplexity": 36.206628866431714, "lr": 0.0026291804804649314, "grad_norm": 0.173882, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:49.750568+00:00", "epoch": 0, "step": 12751, "train_loss": 3.572498321533203, "perplexity": 35.60543595381792, "lr": 0.0026291804804649314, "grad_norm": 0.163818, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:50.056475+00:00", "epoch": 0, "step": 12752, "train_loss": 3.4840524196624756, "perplexity": 32.59152938072005, "lr": 0.0026291804804649314, "grad_norm": 0.162476, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:50.361116+00:00", "epoch": 0, "step": 12753, "train_loss": 3.547841787338257, "perplexity": 34.738263964867066, "lr": 0.0026291804804649314, "grad_norm": 0.182662, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:50.665584+00:00", "epoch": 0, "step": 12754, "train_loss": 3.5928597450256348, "perplexity": 36.33784444990557, "lr": 0.0026291804804649314, "grad_norm": 0.190759, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:50.970835+00:00", "epoch": 0, "step": 12755, "train_loss": 3.571711778640747, "perplexity": 35.57744176199516, "lr": 0.0026291804804649314, "grad_norm": 0.195021, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:51.276011+00:00", "epoch": 0, "step": 12756, "train_loss": 3.5591468811035156, "perplexity": 35.13321154961904, "lr": 0.0026291804804649314, "grad_norm": 0.171763, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:51.581738+00:00", "epoch": 0, "step": 12757, "train_loss": 3.432744026184082, "perplexity": 30.961485417873728, "lr": 0.0026291804804649314, "grad_norm": 0.176132, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:51.886347+00:00", "epoch": 0, "step": 12758, "train_loss": 3.580979108810425, "perplexity": 35.90868214773844, "lr": 0.0026291804804649314, "grad_norm": 0.189619, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:52.192218+00:00", "epoch": 0, "step": 12759, "train_loss": 3.562398672103882, "perplexity": 35.24764336408741, "lr": 0.0026291804804649314, "grad_norm": 0.140594, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:52.498008+00:00", "epoch": 0, "step": 12760, "train_loss": 3.5125818252563477, "perplexity": 33.534736942587095, "lr": 0.0026291804804649314, "grad_norm": 0.149693, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:52.803338+00:00", "epoch": 0, "step": 12761, "train_loss": 3.548877477645874, "perplexity": 34.77426068566868, "lr": 0.0026291804804649314, "grad_norm": 0.156872, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:53.109023+00:00", "epoch": 0, "step": 12762, "train_loss": 3.472161293029785, "perplexity": 32.20627447467838, "lr": 0.0026291804804649314, "grad_norm": 0.158069, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:53.414705+00:00", "epoch": 0, "step": 12763, "train_loss": 3.5083258152008057, "perplexity": 33.392316052299854, "lr": 0.0026291804804649314, "grad_norm": 0.157947, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:53.720332+00:00", "epoch": 0, "step": 12764, "train_loss": 3.513284683227539, "perplexity": 33.55831538493641, "lr": 0.0026291804804649314, "grad_norm": 0.165038, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:54.025618+00:00", "epoch": 0, "step": 12765, "train_loss": 3.449129104614258, "perplexity": 31.47297070484451, "lr": 0.0026291804804649314, "grad_norm": 0.162872, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:54.329991+00:00", "epoch": 0, "step": 12766, "train_loss": 3.4605305194854736, "perplexity": 31.833860518928205, "lr": 0.0026291804804649314, "grad_norm": 0.168845, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:54.636113+00:00", "epoch": 0, "step": 12767, "train_loss": 3.455467700958252, "perplexity": 31.673098757301336, "lr": 0.0026291804804649314, "grad_norm": 0.144034, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:54.941909+00:00", "epoch": 0, "step": 12768, "train_loss": 3.528245449066162, "perplexity": 34.06414786925081, "lr": 0.0026291804804649314, "grad_norm": 0.174457, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:55.248686+00:00", "epoch": 0, "step": 12769, "train_loss": 3.462258815765381, "perplexity": 31.888926433038545, "lr": 0.0026291804804649314, "grad_norm": 0.166428, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:55.554447+00:00", "epoch": 0, "step": 12770, "train_loss": 3.4663190841674805, "perplexity": 32.01866724643422, "lr": 0.0026291804804649314, "grad_norm": 0.155524, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:55.859824+00:00", "epoch": 0, "step": 12771, "train_loss": 3.4707822799682617, "perplexity": 32.16189221040925, "lr": 0.0026291804804649314, "grad_norm": 0.14733, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:56.164800+00:00", "epoch": 0, "step": 12772, "train_loss": 3.5460190773010254, "perplexity": 34.675003852397225, "lr": 0.0026291804804649314, "grad_norm": 0.148927, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:56.471068+00:00", "epoch": 0, "step": 12773, "train_loss": 3.6424787044525146, "perplexity": 38.186372245099534, "lr": 0.0026291804804649314, "grad_norm": 0.152518, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:56.776796+00:00", "epoch": 0, "step": 12774, "train_loss": 3.6302497386932373, "perplexity": 37.722236142883084, "lr": 0.0026291804804649314, "grad_norm": 0.183941, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:57.083975+00:00", "epoch": 0, "step": 12775, "train_loss": 3.5704874992370605, "perplexity": 35.533911684735735, "lr": 0.0026291804804649314, "grad_norm": 0.176831, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:57.389947+00:00", "epoch": 0, "step": 12776, "train_loss": 3.4803831577301025, "perplexity": 32.47216165230189, "lr": 0.0026291804804649314, "grad_norm": 0.167464, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:57.694596+00:00", "epoch": 0, "step": 12777, "train_loss": 3.5534138679504395, "perplexity": 34.93236865327749, "lr": 0.0026291804804649314, "grad_norm": 0.198683, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:57.999770+00:00", "epoch": 0, "step": 12778, "train_loss": 3.6022064685821533, "perplexity": 36.67907645301881, "lr": 0.0026291804804649314, "grad_norm": 0.153413, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:58.305625+00:00", "epoch": 0, "step": 12779, "train_loss": 3.5521163940429688, "perplexity": 34.88707420695905, "lr": 0.0026291804804649314, "grad_norm": 0.183913, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:58.610298+00:00", "epoch": 0, "step": 12780, "train_loss": 3.484098434448242, "perplexity": 32.59302910746685, "lr": 0.0026291804804649314, "grad_norm": 0.17637, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:58.914429+00:00", "epoch": 0, "step": 12781, "train_loss": 3.5308759212493896, "perplexity": 34.153870617527495, "lr": 0.0026291804804649314, "grad_norm": 0.162375, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:59.218701+00:00", "epoch": 0, "step": 12782, "train_loss": 3.489285707473755, "perplexity": 32.76253731075008, "lr": 0.0026291804804649314, "grad_norm": 0.168797, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:59.523765+00:00", "epoch": 0, "step": 12783, "train_loss": 3.535593032836914, "perplexity": 34.315358815860755, "lr": 0.0026291804804649314, "grad_norm": 0.164597, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:04:59.829399+00:00", "epoch": 0, "step": 12784, "train_loss": 3.439873218536377, "perplexity": 31.183004490738742, "lr": 0.0026291804804649314, "grad_norm": 0.181955, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:00.134637+00:00", "epoch": 0, "step": 12785, "train_loss": 3.529370069503784, "perplexity": 34.1024786558878, "lr": 0.0026291804804649314, "grad_norm": 0.17088, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:00.441132+00:00", "epoch": 0, "step": 12786, "train_loss": 3.5239791870117188, "perplexity": 33.9191308480411, "lr": 0.0026291804804649314, "grad_norm": 0.181438, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:00.746600+00:00", "epoch": 0, "step": 12787, "train_loss": 3.6233479976654053, "perplexity": 37.462783405732026, "lr": 0.0026291804804649314, "grad_norm": 0.152752, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:01.050515+00:00", "epoch": 0, "step": 12788, "train_loss": 3.589829444885254, "perplexity": 36.22789654647652, "lr": 0.0026291804804649314, "grad_norm": 0.181554, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:01.355180+00:00", "epoch": 0, "step": 12789, "train_loss": 3.591277837753296, "perplexity": 36.28040679201132, "lr": 0.0026291804804649314, "grad_norm": 0.161556, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:01.661669+00:00", "epoch": 0, "step": 12790, "train_loss": 3.520689010620117, "perplexity": 33.8077143150339, "lr": 0.0026291804804649314, "grad_norm": 0.159148, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:01.966968+00:00", "epoch": 0, "step": 12791, "train_loss": 3.496223211288452, "perplexity": 32.990617778106525, "lr": 0.0026291804804649314, "grad_norm": 0.187447, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:02.272543+00:00", "epoch": 0, "step": 12792, "train_loss": 3.614298105239868, "perplexity": 37.1252787398902, "lr": 0.0026291804804649314, "grad_norm": 0.204493, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:02.577500+00:00", "epoch": 0, "step": 12793, "train_loss": 3.574047088623047, "perplexity": 35.66062320631157, "lr": 0.0026291804804649314, "grad_norm": 0.179234, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:02.882126+00:00", "epoch": 0, "step": 12794, "train_loss": 3.6447832584381104, "perplexity": 38.27447628271479, "lr": 0.0026291804804649314, "grad_norm": 0.191968, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:03.188382+00:00", "epoch": 0, "step": 12795, "train_loss": 3.501694440841675, "perplexity": 33.171611699223824, "lr": 0.0026291804804649314, "grad_norm": 0.16751, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:03.494868+00:00", "epoch": 0, "step": 12796, "train_loss": 3.5649213790893555, "perplexity": 35.33667509348946, "lr": 0.0026291804804649314, "grad_norm": 0.176487, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:03.800603+00:00", "epoch": 0, "step": 12797, "train_loss": 3.49399995803833, "perplexity": 32.91735275343812, "lr": 0.0026291804804649314, "grad_norm": 0.158131, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:04.106735+00:00", "epoch": 0, "step": 12798, "train_loss": 3.5966291427612305, "perplexity": 36.47507471372962, "lr": 0.0026291804804649314, "grad_norm": 0.17941, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:04.411371+00:00", "epoch": 0, "step": 12799, "train_loss": 3.581854820251465, "perplexity": 35.940141564202904, "lr": 0.0026291804804649314, "grad_norm": 0.164853, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:04.716940+00:00", "epoch": 0, "step": 12800, "train_loss": 3.522303581237793, "perplexity": 33.8623433465649, "lr": 0.0026291804804649314, "grad_norm": 0.169118, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:05.021964+00:00", "epoch": 0, "step": 12801, "train_loss": 3.4339072704315186, "perplexity": 30.997522143371423, "lr": 0.0026291804804649314, "grad_norm": 0.165983, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:05.327328+00:00", "epoch": 0, "step": 12802, "train_loss": 3.5272698402404785, "perplexity": 34.030930792021735, "lr": 0.0026291804804649314, "grad_norm": 0.165681, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:05.633219+00:00", "epoch": 0, "step": 12803, "train_loss": 3.498927354812622, "perplexity": 33.07994987249589, "lr": 0.0026291804804649314, "grad_norm": 0.192812, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:05.937999+00:00", "epoch": 0, "step": 12804, "train_loss": 3.516444683074951, "perplexity": 33.66452738299853, "lr": 0.0026291804804649314, "grad_norm": 0.177085, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:06.243004+00:00", "epoch": 0, "step": 12805, "train_loss": 3.4310853481292725, "perplexity": 30.910172848747283, "lr": 0.0026291804804649314, "grad_norm": 0.174325, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:06.548044+00:00", "epoch": 0, "step": 12806, "train_loss": 3.5395877361297607, "perplexity": 34.45271265397869, "lr": 0.0026291804804649314, "grad_norm": 0.194189, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:06.854308+00:00", "epoch": 0, "step": 12807, "train_loss": 3.514355182647705, "perplexity": 33.594258777355606, "lr": 0.0026291804804649314, "grad_norm": 0.152026, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:07.158830+00:00", "epoch": 0, "step": 12808, "train_loss": 3.5332911014556885, "perplexity": 34.23645806144633, "lr": 0.0026291804804649314, "grad_norm": 0.153953, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:07.464698+00:00", "epoch": 0, "step": 12809, "train_loss": 3.507720470428467, "perplexity": 33.372108305285394, "lr": 0.0026291804804649314, "grad_norm": 0.169031, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:07.770593+00:00", "epoch": 0, "step": 12810, "train_loss": 3.3937602043151855, "perplexity": 29.777712300701648, "lr": 0.0026291804804649314, "grad_norm": 0.156052, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:08.075997+00:00", "epoch": 0, "step": 12811, "train_loss": 3.539614200592041, "perplexity": 34.453624438558066, "lr": 0.0026291804804649314, "grad_norm": 0.175223, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:08.381725+00:00", "epoch": 0, "step": 12812, "train_loss": 3.5904314517974854, "perplexity": 36.2497125566487, "lr": 0.0026291804804649314, "grad_norm": 0.191673, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:08.688261+00:00", "epoch": 0, "step": 12813, "train_loss": 3.4809892177581787, "perplexity": 32.49184769636422, "lr": 0.0026291804804649314, "grad_norm": 0.185182, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:08.993649+00:00", "epoch": 0, "step": 12814, "train_loss": 3.5118489265441895, "perplexity": 33.5101683813025, "lr": 0.0026291804804649314, "grad_norm": 0.201661, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:09.299437+00:00", "epoch": 0, "step": 12815, "train_loss": 3.4606122970581055, "perplexity": 31.83646392121745, "lr": 0.0026291804804649314, "grad_norm": 0.181315, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:09.604524+00:00", "epoch": 0, "step": 12816, "train_loss": 3.5006887912750244, "perplexity": 33.138269450412444, "lr": 0.0026291804804649314, "grad_norm": 0.196855, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:09.911226+00:00", "epoch": 0, "step": 12817, "train_loss": 3.5145509243011475, "perplexity": 33.60083521673541, "lr": 0.0026291804804649314, "grad_norm": 0.165326, "tokens_per_sec": 106839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:10.216913+00:00", "epoch": 0, "step": 12818, "train_loss": 3.498382806777954, "perplexity": 33.061941154556905, "lr": 0.0026291804804649314, "grad_norm": 0.173147, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:10.523877+00:00", "epoch": 0, "step": 12819, "train_loss": 3.5807976722717285, "perplexity": 35.90216759174753, "lr": 0.0026291804804649314, "grad_norm": 0.225476, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:10.829998+00:00", "epoch": 0, "step": 12820, "train_loss": 3.6034765243530273, "perplexity": 36.725690520705264, "lr": 0.0026291804804649314, "grad_norm": 0.211878, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:11.135243+00:00", "epoch": 0, "step": 12821, "train_loss": 3.5072600841522217, "perplexity": 33.35674778077093, "lr": 0.0026291804804649314, "grad_norm": 0.201024, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:11.440291+00:00", "epoch": 0, "step": 12822, "train_loss": 3.6039254665374756, "perplexity": 36.742181934001565, "lr": 0.0026291804804649314, "grad_norm": 0.179999, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:11.747379+00:00", "epoch": 0, "step": 12823, "train_loss": 3.4194984436035156, "perplexity": 30.554086579769837, "lr": 0.0026291804804649314, "grad_norm": 0.167495, "tokens_per_sec": 106704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:12.053330+00:00", "epoch": 0, "step": 12824, "train_loss": 3.5160374641418457, "perplexity": 33.65082134094344, "lr": 0.0026291804804649314, "grad_norm": 0.184456, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:12.358909+00:00", "epoch": 0, "step": 12825, "train_loss": 3.5502028465270996, "perplexity": 34.82037996442643, "lr": 0.0026291804804649314, "grad_norm": 0.170494, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:12.664557+00:00", "epoch": 0, "step": 12826, "train_loss": 3.4784650802612305, "perplexity": 32.409937225383764, "lr": 0.0026291804804649314, "grad_norm": 0.190043, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:12.969049+00:00", "epoch": 0, "step": 12827, "train_loss": 3.447087526321411, "perplexity": 31.408781716759655, "lr": 0.0026291804804649314, "grad_norm": 0.183849, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:13.274493+00:00", "epoch": 0, "step": 12828, "train_loss": 3.5086629390716553, "perplexity": 33.40357529691747, "lr": 0.0026291804804649314, "grad_norm": 0.162667, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:13.580038+00:00", "epoch": 0, "step": 12829, "train_loss": 3.4948058128356934, "perplexity": 32.94389005123923, "lr": 0.0026291804804649314, "grad_norm": 0.154821, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:13.885559+00:00", "epoch": 0, "step": 12830, "train_loss": 3.525097131729126, "perplexity": 33.9570717651641, "lr": 0.0026291804804649314, "grad_norm": 0.18103, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:14.190369+00:00", "epoch": 0, "step": 12831, "train_loss": 3.652784824371338, "perplexity": 38.581960565575564, "lr": 0.0026291804804649314, "grad_norm": 0.160324, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:14.497466+00:00", "epoch": 0, "step": 12832, "train_loss": 3.503636121749878, "perplexity": 33.23608295540277, "lr": 0.0026291804804649314, "grad_norm": 0.148943, "tokens_per_sec": 106702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:14.802848+00:00", "epoch": 0, "step": 12833, "train_loss": 3.5539064407348633, "perplexity": 34.94957962585191, "lr": 0.0026291804804649314, "grad_norm": 0.149984, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:15.107848+00:00", "epoch": 0, "step": 12834, "train_loss": 3.515507936477661, "perplexity": 33.63300701712578, "lr": 0.0026291804804649314, "grad_norm": 0.151846, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:15.412718+00:00", "epoch": 0, "step": 12835, "train_loss": 3.520659923553467, "perplexity": 33.80673096209581, "lr": 0.0026291804804649314, "grad_norm": 0.163963, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:15.718300+00:00", "epoch": 0, "step": 12836, "train_loss": 3.4690515995025635, "perplexity": 32.10627839057807, "lr": 0.0026291804804649314, "grad_norm": 0.136583, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:16.022591+00:00", "epoch": 0, "step": 12837, "train_loss": 3.5360960960388184, "perplexity": 34.33262595300766, "lr": 0.0026291804804649314, "grad_norm": 0.160958, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:16.328019+00:00", "epoch": 0, "step": 12838, "train_loss": 3.509921073913574, "perplexity": 33.44562794724836, "lr": 0.0026291804804649314, "grad_norm": 0.181377, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:16.633963+00:00", "epoch": 0, "step": 12839, "train_loss": 3.573807716369629, "perplexity": 35.65208806415525, "lr": 0.0026291804804649314, "grad_norm": 0.201114, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:16.939195+00:00", "epoch": 0, "step": 12840, "train_loss": 3.4637012481689453, "perplexity": 31.93495724402609, "lr": 0.0026291804804649314, "grad_norm": 0.168232, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:17.245080+00:00", "epoch": 0, "step": 12841, "train_loss": 3.527858018875122, "perplexity": 34.05095294615505, "lr": 0.0026291804804649314, "grad_norm": 0.167513, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:17.549547+00:00", "epoch": 0, "step": 12842, "train_loss": 3.5273120403289795, "perplexity": 34.032366930615304, "lr": 0.0026291804804649314, "grad_norm": 0.185182, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:17.853655+00:00", "epoch": 0, "step": 12843, "train_loss": 3.4669835567474365, "perplexity": 32.03994984293405, "lr": 0.0026291804804649314, "grad_norm": 0.200627, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:18.160280+00:00", "epoch": 0, "step": 12844, "train_loss": 3.440185308456421, "perplexity": 31.192737910889143, "lr": 0.0026291804804649314, "grad_norm": 0.206948, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:18.466617+00:00", "epoch": 0, "step": 12845, "train_loss": 3.5774004459381104, "perplexity": 35.780406744254584, "lr": 0.0026291804804649314, "grad_norm": 0.168493, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:18.771321+00:00", "epoch": 0, "step": 12846, "train_loss": 3.5231449604034424, "perplexity": 33.89084640601547, "lr": 0.0026291804804649314, "grad_norm": 0.167929, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:19.077102+00:00", "epoch": 0, "step": 12847, "train_loss": 3.4938576221466064, "perplexity": 32.91266776610961, "lr": 0.0026291804804649314, "grad_norm": 0.171275, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:19.382271+00:00", "epoch": 0, "step": 12848, "train_loss": 3.5530569553375244, "perplexity": 34.919903074999525, "lr": 0.0026291804804649314, "grad_norm": 0.157106, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:19.687907+00:00", "epoch": 0, "step": 12849, "train_loss": 3.517946243286133, "perplexity": 33.715114668267596, "lr": 0.0026291804804649314, "grad_norm": 0.147099, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:19.995544+00:00", "epoch": 0, "step": 12850, "train_loss": 3.550182580947876, "perplexity": 34.81967431640787, "lr": 0.0026291804804649314, "grad_norm": 0.15187, "tokens_per_sec": 106576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:20.301894+00:00", "epoch": 0, "step": 12851, "train_loss": 3.474912166595459, "perplexity": 32.29499183295152, "lr": 0.0026291804804649314, "grad_norm": 0.175452, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:20.607191+00:00", "epoch": 0, "step": 12852, "train_loss": 3.5370731353759766, "perplexity": 34.36618667151299, "lr": 0.0026291804804649314, "grad_norm": 0.170122, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:20.913804+00:00", "epoch": 0, "step": 12853, "train_loss": 3.688210964202881, "perplexity": 39.97326933914178, "lr": 0.0026291804804649314, "grad_norm": 0.147887, "tokens_per_sec": 106873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:21.218780+00:00", "epoch": 0, "step": 12854, "train_loss": 3.577988386154175, "perplexity": 35.80144966971158, "lr": 0.0026291804804649314, "grad_norm": 0.175017, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:21.524271+00:00", "epoch": 0, "step": 12855, "train_loss": 3.5385050773620605, "perplexity": 34.4154323071415, "lr": 0.0026291804804649314, "grad_norm": 0.151633, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:21.831793+00:00", "epoch": 0, "step": 12856, "train_loss": 3.5052366256713867, "perplexity": 33.28932002821599, "lr": 0.0026291804804649314, "grad_norm": 0.158133, "tokens_per_sec": 106555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:22.137458+00:00", "epoch": 0, "step": 12857, "train_loss": 3.5188186168670654, "perplexity": 33.74453967651609, "lr": 0.0026291804804649314, "grad_norm": 0.159967, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:22.441650+00:00", "epoch": 0, "step": 12858, "train_loss": 3.4813661575317383, "perplexity": 32.504097474646706, "lr": 0.0026291804804649314, "grad_norm": 0.146821, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:22.747027+00:00", "epoch": 0, "step": 12859, "train_loss": 3.5237417221069336, "perplexity": 33.91107720113097, "lr": 0.0026291804804649314, "grad_norm": 0.170868, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:23.052029+00:00", "epoch": 0, "step": 12860, "train_loss": 3.599137306213379, "perplexity": 36.566674989252064, "lr": 0.0026291804804649314, "grad_norm": 0.152663, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:23.357297+00:00", "epoch": 0, "step": 12861, "train_loss": 3.472665786743164, "perplexity": 32.22252643684389, "lr": 0.0026291804804649314, "grad_norm": 0.177884, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:23.663922+00:00", "epoch": 0, "step": 12862, "train_loss": 3.577056884765625, "perplexity": 35.76811609717698, "lr": 0.0026291804804649314, "grad_norm": 0.175045, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:23.968559+00:00", "epoch": 0, "step": 12863, "train_loss": 3.526362419128418, "perplexity": 34.00006441347774, "lr": 0.0026291804804649314, "grad_norm": 0.149579, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:24.272691+00:00", "epoch": 0, "step": 12864, "train_loss": 3.5911123752593994, "perplexity": 36.27440424203607, "lr": 0.0026291804804649314, "grad_norm": 0.162105, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:24.578426+00:00", "epoch": 0, "step": 12865, "train_loss": 3.5780341625213623, "perplexity": 35.80308856752861, "lr": 0.0026291804804649314, "grad_norm": 0.153811, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:24.882812+00:00", "epoch": 0, "step": 12866, "train_loss": 3.576913833618164, "perplexity": 35.76299979308207, "lr": 0.0026291804804649314, "grad_norm": 0.160183, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:25.189741+00:00", "epoch": 0, "step": 12867, "train_loss": 3.5555105209350586, "perplexity": 35.00568654250134, "lr": 0.0026291804804649314, "grad_norm": 0.166216, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:25.495091+00:00", "epoch": 0, "step": 12868, "train_loss": 3.5150251388549805, "perplexity": 33.61677300048196, "lr": 0.0026291804804649314, "grad_norm": 0.175466, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:25.799936+00:00", "epoch": 0, "step": 12869, "train_loss": 3.5161850452423096, "perplexity": 33.65578793266697, "lr": 0.0026291804804649314, "grad_norm": 0.144292, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:26.105323+00:00", "epoch": 0, "step": 12870, "train_loss": 3.502534866333008, "perplexity": 33.199501685370514, "lr": 0.0026291804804649314, "grad_norm": 0.214688, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:26.411103+00:00", "epoch": 0, "step": 12871, "train_loss": 3.533703088760376, "perplexity": 34.25056595346176, "lr": 0.0026291804804649314, "grad_norm": 0.209564, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:26.718058+00:00", "epoch": 0, "step": 12872, "train_loss": 3.5139241218566895, "perplexity": 33.57978073027685, "lr": 0.0026291804804649314, "grad_norm": 0.170692, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:27.025861+00:00", "epoch": 0, "step": 12873, "train_loss": 3.4565012454986572, "perplexity": 31.705851238258045, "lr": 0.0026291804804649314, "grad_norm": 0.198305, "tokens_per_sec": 106458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:27.332246+00:00", "epoch": 0, "step": 12874, "train_loss": 3.5968968868255615, "perplexity": 36.48484200598917, "lr": 0.0026291804804649314, "grad_norm": 0.218137, "tokens_per_sec": 106955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:27.636935+00:00", "epoch": 0, "step": 12875, "train_loss": 3.484015941619873, "perplexity": 32.59034052720647, "lr": 0.0026291804804649314, "grad_norm": 0.185469, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:27.942222+00:00", "epoch": 0, "step": 12876, "train_loss": 3.361224412918091, "perplexity": 28.824462325457258, "lr": 0.0026291804804649314, "grad_norm": 0.179547, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:28.248875+00:00", "epoch": 0, "step": 12877, "train_loss": 3.4985079765319824, "perplexity": 33.06607976860816, "lr": 0.0026291804804649314, "grad_norm": 0.201755, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:28.556374+00:00", "epoch": 0, "step": 12878, "train_loss": 3.5522501468658447, "perplexity": 34.89174076369172, "lr": 0.0026291804804649314, "grad_norm": 0.163009, "tokens_per_sec": 106564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:28.861424+00:00", "epoch": 0, "step": 12879, "train_loss": 3.5254931449890137, "perplexity": 33.970521878885975, "lr": 0.0026291804804649314, "grad_norm": 0.17806, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:29.167898+00:00", "epoch": 0, "step": 12880, "train_loss": 3.457571506500244, "perplexity": 31.739802939710614, "lr": 0.0026291804804649314, "grad_norm": 0.153868, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:29.474377+00:00", "epoch": 0, "step": 12881, "train_loss": 3.65875506401062, "perplexity": 38.81299308943273, "lr": 0.0026291804804649314, "grad_norm": 0.176069, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:29.781330+00:00", "epoch": 0, "step": 12882, "train_loss": 3.5410706996917725, "perplexity": 34.50384267406501, "lr": 0.0026291804804649314, "grad_norm": 0.16572, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:30.088503+00:00", "epoch": 0, "step": 12883, "train_loss": 3.4697265625, "perplexity": 32.12795625553247, "lr": 0.0026291804804649314, "grad_norm": 0.158181, "tokens_per_sec": 106675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:30.394275+00:00", "epoch": 0, "step": 12884, "train_loss": 3.5865108966827393, "perplexity": 36.107871789632185, "lr": 0.0026291804804649314, "grad_norm": 0.165476, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:30.699924+00:00", "epoch": 0, "step": 12885, "train_loss": 3.392843246459961, "perplexity": 29.75041990839597, "lr": 0.0026291804804649314, "grad_norm": 0.161878, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:31.005688+00:00", "epoch": 0, "step": 12886, "train_loss": 3.5996086597442627, "perplexity": 36.583914883344214, "lr": 0.0026291804804649314, "grad_norm": 0.17292, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:31.312215+00:00", "epoch": 0, "step": 12887, "train_loss": 3.4846110343933105, "perplexity": 32.609740575179735, "lr": 0.0026291804804649314, "grad_norm": 0.18664, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:31.618659+00:00", "epoch": 0, "step": 12888, "train_loss": 3.550232172012329, "perplexity": 34.8214011039374, "lr": 0.0026291804804649314, "grad_norm": 0.181797, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:31.924724+00:00", "epoch": 0, "step": 12889, "train_loss": 3.400139570236206, "perplexity": 29.968282435780008, "lr": 0.0026291804804649314, "grad_norm": 0.153373, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:32.230596+00:00", "epoch": 0, "step": 12890, "train_loss": 3.522279739379883, "perplexity": 33.86153601501051, "lr": 0.0026291804804649314, "grad_norm": 0.160327, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:32.535968+00:00", "epoch": 0, "step": 12891, "train_loss": 3.5140457153320312, "perplexity": 33.5838640607652, "lr": 0.0026291804804649314, "grad_norm": 0.166963, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:32.841790+00:00", "epoch": 0, "step": 12892, "train_loss": 3.459259271621704, "perplexity": 31.793417503742873, "lr": 0.0026291804804649314, "grad_norm": 0.170836, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:33.147915+00:00", "epoch": 0, "step": 12893, "train_loss": 3.5961642265319824, "perplexity": 36.45812080091216, "lr": 0.0026291804804649314, "grad_norm": 0.166238, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:33.454104+00:00", "epoch": 0, "step": 12894, "train_loss": 3.503875255584717, "perplexity": 33.24403177775323, "lr": 0.0026291804804649314, "grad_norm": 0.143389, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:33.759214+00:00", "epoch": 0, "step": 12895, "train_loss": 3.5603888034820557, "perplexity": 35.17687137671901, "lr": 0.0026291804804649314, "grad_norm": 0.167116, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:34.064293+00:00", "epoch": 0, "step": 12896, "train_loss": 3.568363666534424, "perplexity": 35.45852368513406, "lr": 0.0026291804804649314, "grad_norm": 0.150782, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:34.370813+00:00", "epoch": 0, "step": 12897, "train_loss": 3.529097557067871, "perplexity": 34.09318657252025, "lr": 0.0026291804804649314, "grad_norm": 0.150173, "tokens_per_sec": 106904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:34.678676+00:00", "epoch": 0, "step": 12898, "train_loss": 3.522353172302246, "perplexity": 33.86402265785541, "lr": 0.0026291804804649314, "grad_norm": 0.1613, "tokens_per_sec": 106436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:34.985035+00:00", "epoch": 0, "step": 12899, "train_loss": 3.4662044048309326, "perplexity": 32.014995577453554, "lr": 0.0026291804804649314, "grad_norm": 0.148923, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:35.290342+00:00", "epoch": 0, "step": 12900, "train_loss": 3.415769577026367, "perplexity": 30.440366622568735, "lr": 0.0026291804804649314, "grad_norm": 0.153945, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:35.595519+00:00", "epoch": 0, "step": 12901, "train_loss": 3.5595855712890625, "perplexity": 35.14862752588411, "lr": 0.0026291804804649314, "grad_norm": 0.15597, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:35.900719+00:00", "epoch": 0, "step": 12902, "train_loss": 3.56967830657959, "perplexity": 35.505169534845905, "lr": 0.0026291804804649314, "grad_norm": 0.146799, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:36.207050+00:00", "epoch": 0, "step": 12903, "train_loss": 3.496068000793457, "perplexity": 32.98549768534733, "lr": 0.0026291804804649314, "grad_norm": 0.161176, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:36.513171+00:00", "epoch": 0, "step": 12904, "train_loss": 3.5065112113952637, "perplexity": 33.33177717216843, "lr": 0.0026291804804649314, "grad_norm": 0.156011, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:36.818133+00:00", "epoch": 0, "step": 12905, "train_loss": 3.544278860092163, "perplexity": 34.61471428765981, "lr": 0.0026291804804649314, "grad_norm": 0.195121, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:37.123830+00:00", "epoch": 0, "step": 12906, "train_loss": 3.477414131164551, "perplexity": 32.37589392315555, "lr": 0.0026291804804649314, "grad_norm": 0.189544, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:37.428345+00:00", "epoch": 0, "step": 12907, "train_loss": 3.5237646102905273, "perplexity": 33.911853372974356, "lr": 0.0026291804804649314, "grad_norm": 0.141912, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:37.735158+00:00", "epoch": 0, "step": 12908, "train_loss": 3.602539539337158, "perplexity": 36.69129521544915, "lr": 0.0026291804804649314, "grad_norm": 0.159588, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:38.041581+00:00", "epoch": 0, "step": 12909, "train_loss": 3.542717456817627, "perplexity": 34.56070893246514, "lr": 0.0026291804804649314, "grad_norm": 0.163151, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:38.346804+00:00", "epoch": 0, "step": 12910, "train_loss": 3.5092387199401855, "perplexity": 33.42281397461371, "lr": 0.0026291804804649314, "grad_norm": 0.165646, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:38.652429+00:00", "epoch": 0, "step": 12911, "train_loss": 3.5372350215911865, "perplexity": 34.371750533748596, "lr": 0.0026291804804649314, "grad_norm": 0.158539, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:38.957362+00:00", "epoch": 0, "step": 12912, "train_loss": 3.6167550086975098, "perplexity": 37.216604108459535, "lr": 0.0026291804804649314, "grad_norm": 0.181229, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:39.264283+00:00", "epoch": 0, "step": 12913, "train_loss": 3.5361900329589844, "perplexity": 34.33585120563367, "lr": 0.0026291804804649314, "grad_norm": 0.177842, "tokens_per_sec": 106764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:39.568734+00:00", "epoch": 0, "step": 12914, "train_loss": 3.4826576709747314, "perplexity": 32.546104073692305, "lr": 0.0026291804804649314, "grad_norm": 0.177583, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:39.875023+00:00", "epoch": 0, "step": 12915, "train_loss": 3.5900378227233887, "perplexity": 36.23544642382529, "lr": 0.0026291804804649314, "grad_norm": 0.213782, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:40.181385+00:00", "epoch": 0, "step": 12916, "train_loss": 3.562976121902466, "perplexity": 35.268002986412725, "lr": 0.0026291804804649314, "grad_norm": 0.149002, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:40.486481+00:00", "epoch": 0, "step": 12917, "train_loss": 3.507578134536743, "perplexity": 33.367358594526245, "lr": 0.0026291804804649314, "grad_norm": 0.202543, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:40.792752+00:00", "epoch": 0, "step": 12918, "train_loss": 3.580836057662964, "perplexity": 35.90354573694688, "lr": 0.0026291804804649314, "grad_norm": 0.196754, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:41.098037+00:00", "epoch": 0, "step": 12919, "train_loss": 3.4795186519622803, "perplexity": 32.444101412127175, "lr": 0.0026291804804649314, "grad_norm": 0.18193, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:41.403805+00:00", "epoch": 0, "step": 12920, "train_loss": 3.555356979370117, "perplexity": 35.000312127216375, "lr": 0.0026291804804649314, "grad_norm": 0.226813, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:41.709953+00:00", "epoch": 0, "step": 12921, "train_loss": 3.5214827060699463, "perplexity": 33.834557995499196, "lr": 0.0026291804804649314, "grad_norm": 0.192732, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:42.017015+00:00", "epoch": 0, "step": 12922, "train_loss": 3.4467647075653076, "perplexity": 31.398644009324215, "lr": 0.0026291804804649314, "grad_norm": 0.21136, "tokens_per_sec": 106711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:42.322654+00:00", "epoch": 0, "step": 12923, "train_loss": 3.523852586746216, "perplexity": 33.914836948880236, "lr": 0.0026291804804649314, "grad_norm": 0.194179, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:42.627041+00:00", "epoch": 0, "step": 12924, "train_loss": 3.6450579166412354, "perplexity": 38.28499012538662, "lr": 0.0026291804804649314, "grad_norm": 0.211442, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:42.931906+00:00", "epoch": 0, "step": 12925, "train_loss": 3.5066561698913574, "perplexity": 33.336609246676105, "lr": 0.0026291804804649314, "grad_norm": 0.156632, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:43.237599+00:00", "epoch": 0, "step": 12926, "train_loss": 3.5619139671325684, "perplexity": 35.2305627959743, "lr": 0.0026291804804649314, "grad_norm": 0.195994, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:43.542683+00:00", "epoch": 0, "step": 12927, "train_loss": 3.591602325439453, "perplexity": 36.292181247493694, "lr": 0.0026291804804649314, "grad_norm": 0.179262, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:43.850544+00:00", "epoch": 0, "step": 12928, "train_loss": 3.538011074066162, "perplexity": 34.39843516881884, "lr": 0.0026291804804649314, "grad_norm": 0.148238, "tokens_per_sec": 106439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:44.155144+00:00", "epoch": 0, "step": 12929, "train_loss": 3.54126238822937, "perplexity": 34.510457299162354, "lr": 0.0026291804804649314, "grad_norm": 0.190329, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:44.460936+00:00", "epoch": 0, "step": 12930, "train_loss": 3.6217432022094727, "perplexity": 37.402711515596344, "lr": 0.0026291804804649314, "grad_norm": 0.165814, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:44.765659+00:00", "epoch": 0, "step": 12931, "train_loss": 3.4566173553466797, "perplexity": 31.70953281355618, "lr": 0.0026291804804649314, "grad_norm": 0.192688, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:45.071661+00:00", "epoch": 0, "step": 12932, "train_loss": 3.5089492797851562, "perplexity": 33.41314147002853, "lr": 0.0026291804804649314, "grad_norm": 0.147729, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:45.377766+00:00", "epoch": 0, "step": 12933, "train_loss": 3.4900574684143066, "perplexity": 32.78783191679888, "lr": 0.0026291804804649314, "grad_norm": 0.167037, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:45.684994+00:00", "epoch": 0, "step": 12934, "train_loss": 3.5753207206726074, "perplexity": 35.706070654454194, "lr": 0.0026291804804649314, "grad_norm": 0.174261, "tokens_per_sec": 106657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:45.989622+00:00", "epoch": 0, "step": 12935, "train_loss": 3.543710470199585, "perplexity": 34.59504522430216, "lr": 0.0026291804804649314, "grad_norm": 0.17655, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:46.294439+00:00", "epoch": 0, "step": 12936, "train_loss": 3.3960280418395996, "perplexity": 29.845319946650594, "lr": 0.0026291804804649314, "grad_norm": 0.153106, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:46.599505+00:00", "epoch": 0, "step": 12937, "train_loss": 3.4712963104248047, "perplexity": 32.17842865229873, "lr": 0.0026291804804649314, "grad_norm": 0.155259, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:46.905983+00:00", "epoch": 0, "step": 12938, "train_loss": 3.536388874053955, "perplexity": 34.342679262710824, "lr": 0.0026291804804649314, "grad_norm": 0.151373, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:47.211090+00:00", "epoch": 0, "step": 12939, "train_loss": 3.5004472732543945, "perplexity": 33.130266927583314, "lr": 0.0026291804804649314, "grad_norm": 0.157141, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:47.517712+00:00", "epoch": 0, "step": 12940, "train_loss": 3.4753103256225586, "perplexity": 32.30785293569209, "lr": 0.0026291804804649314, "grad_norm": 0.149092, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:47.821998+00:00", "epoch": 0, "step": 12941, "train_loss": 3.702176570892334, "perplexity": 40.53543666372068, "lr": 0.0026291804804649314, "grad_norm": 0.164089, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:48.126657+00:00", "epoch": 0, "step": 12942, "train_loss": 3.552213430404663, "perplexity": 34.89045968596488, "lr": 0.0026291804804649314, "grad_norm": 0.157054, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:48.431739+00:00", "epoch": 0, "step": 12943, "train_loss": 3.5196526050567627, "perplexity": 33.77269396261226, "lr": 0.0026291804804649314, "grad_norm": 0.152583, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:48.737822+00:00", "epoch": 0, "step": 12944, "train_loss": 3.6333272457122803, "perplexity": 37.838505407359584, "lr": 0.0026291804804649314, "grad_norm": 0.169636, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:49.044119+00:00", "epoch": 0, "step": 12945, "train_loss": 3.4625792503356934, "perplexity": 31.899146384803323, "lr": 0.0026291804804649314, "grad_norm": 0.167871, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:49.349429+00:00", "epoch": 0, "step": 12946, "train_loss": 3.542084217071533, "perplexity": 34.538830645747375, "lr": 0.0026291804804649314, "grad_norm": 0.150014, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:49.654380+00:00", "epoch": 0, "step": 12947, "train_loss": 3.532404899597168, "perplexity": 34.206131088577756, "lr": 0.0026291804804649314, "grad_norm": 0.147573, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:49.959769+00:00", "epoch": 0, "step": 12948, "train_loss": 3.485018253326416, "perplexity": 32.623022583104614, "lr": 0.0026291804804649314, "grad_norm": 0.167422, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:50.266128+00:00", "epoch": 0, "step": 12949, "train_loss": 3.569577693939209, "perplexity": 35.501597445693506, "lr": 0.0026291804804649314, "grad_norm": 0.199639, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:50.573062+00:00", "epoch": 0, "step": 12950, "train_loss": 3.6156437397003174, "perplexity": 37.17526942136491, "lr": 0.0026291804804649314, "grad_norm": 0.171393, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:50.879978+00:00", "epoch": 0, "step": 12951, "train_loss": 3.509349822998047, "perplexity": 33.42652755774002, "lr": 0.0026291804804649314, "grad_norm": 0.141795, "tokens_per_sec": 106706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:51.185773+00:00", "epoch": 0, "step": 12952, "train_loss": 3.55159330368042, "perplexity": 34.86882988679389, "lr": 0.0026291804804649314, "grad_norm": 0.16166, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:51.493383+00:00", "epoch": 0, "step": 12953, "train_loss": 3.564967393875122, "perplexity": 35.3383011404344, "lr": 0.0026291804804649314, "grad_norm": 0.155651, "tokens_per_sec": 106522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:51.799321+00:00", "epoch": 0, "step": 12954, "train_loss": 3.5986621379852295, "perplexity": 36.54930379453486, "lr": 0.0026291804804649314, "grad_norm": 0.158076, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:52.107124+00:00", "epoch": 0, "step": 12955, "train_loss": 3.5437562465667725, "perplexity": 34.59662889604231, "lr": 0.0026291804804649314, "grad_norm": 0.178293, "tokens_per_sec": 106458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:52.413249+00:00", "epoch": 0, "step": 12956, "train_loss": 3.5231575965881348, "perplexity": 33.891274659715776, "lr": 0.0026291804804649314, "grad_norm": 0.169371, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:52.718630+00:00", "epoch": 0, "step": 12957, "train_loss": 3.4975833892822266, "perplexity": 33.035521421960475, "lr": 0.0026291804804649314, "grad_norm": 0.146842, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:53.025401+00:00", "epoch": 0, "step": 12958, "train_loss": 3.561175584793091, "perplexity": 35.204558772233305, "lr": 0.0026291804804649314, "grad_norm": 0.163032, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:53.331164+00:00", "epoch": 0, "step": 12959, "train_loss": 3.5750317573547363, "perplexity": 35.695754400391024, "lr": 0.0026291804804649314, "grad_norm": 0.151217, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:53.638510+00:00", "epoch": 0, "step": 12960, "train_loss": 3.488150119781494, "perplexity": 32.725353693237835, "lr": 0.0026291804804649314, "grad_norm": 0.149353, "tokens_per_sec": 106672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:53.943730+00:00", "epoch": 0, "step": 12961, "train_loss": 3.648878335952759, "perplexity": 38.43153459341893, "lr": 0.0026291804804649314, "grad_norm": 0.163122, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:54.249344+00:00", "epoch": 0, "step": 12962, "train_loss": 3.382239818572998, "perplexity": 29.436630040423, "lr": 0.0026291804804649314, "grad_norm": 0.149371, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:54.553935+00:00", "epoch": 0, "step": 12963, "train_loss": 3.49985408782959, "perplexity": 33.11062036372574, "lr": 0.0026291804804649314, "grad_norm": 0.155098, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:54.858938+00:00", "epoch": 0, "step": 12964, "train_loss": 3.558912754058838, "perplexity": 35.12498687747555, "lr": 0.0026291804804649314, "grad_norm": 0.187004, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:55.164238+00:00", "epoch": 0, "step": 12965, "train_loss": 3.50852108001709, "perplexity": 33.39883703339748, "lr": 0.0026291804804649314, "grad_norm": 0.189067, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:55.469578+00:00", "epoch": 0, "step": 12966, "train_loss": 3.5156619548797607, "perplexity": 33.638187518060356, "lr": 0.0026291804804649314, "grad_norm": 0.176861, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:55.774867+00:00", "epoch": 0, "step": 12967, "train_loss": 3.5471854209899902, "perplexity": 34.715470418679985, "lr": 0.0026291804804649314, "grad_norm": 0.182228, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:56.080807+00:00", "epoch": 0, "step": 12968, "train_loss": 3.5503716468811035, "perplexity": 34.82625815299718, "lr": 0.0026291804804649314, "grad_norm": 0.221215, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:56.385947+00:00", "epoch": 0, "step": 12969, "train_loss": 3.5451836585998535, "perplexity": 34.64604780261257, "lr": 0.0026291804804649314, "grad_norm": 0.206791, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:56.692342+00:00", "epoch": 0, "step": 12970, "train_loss": 3.52895450592041, "perplexity": 34.08830985187901, "lr": 0.0026291804804649314, "grad_norm": 0.186029, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:56.998166+00:00", "epoch": 0, "step": 12971, "train_loss": 3.641707181930542, "perplexity": 38.156921961119544, "lr": 0.0026291804804649314, "grad_norm": 0.179472, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:57.303610+00:00", "epoch": 0, "step": 12972, "train_loss": 3.620352029800415, "perplexity": 37.35071407240221, "lr": 0.0026291804804649314, "grad_norm": 0.157646, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:57.608746+00:00", "epoch": 0, "step": 12973, "train_loss": 3.559537649154663, "perplexity": 35.14694316899117, "lr": 0.0026291804804649314, "grad_norm": 0.1644, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:57.914218+00:00", "epoch": 0, "step": 12974, "train_loss": 3.546517848968506, "perplexity": 34.69230307571117, "lr": 0.0026291804804649314, "grad_norm": 0.170256, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:58.220156+00:00", "epoch": 0, "step": 12975, "train_loss": 3.4432597160339355, "perplexity": 31.288784668472793, "lr": 0.0026291804804649314, "grad_norm": 0.164179, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:58.526231+00:00", "epoch": 0, "step": 12976, "train_loss": 3.536564588546753, "perplexity": 34.348714299384966, "lr": 0.0026291804804649314, "grad_norm": 0.160993, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:58.831723+00:00", "epoch": 0, "step": 12977, "train_loss": 3.4821524620056152, "perplexity": 32.5296656427755, "lr": 0.0026291804804649314, "grad_norm": 0.150383, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:59.137751+00:00", "epoch": 0, "step": 12978, "train_loss": 3.569868326187134, "perplexity": 35.51191685426789, "lr": 0.0026291804804649314, "grad_norm": 0.151965, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:59.442497+00:00", "epoch": 0, "step": 12979, "train_loss": 3.461087942123413, "perplexity": 31.851610380062066, "lr": 0.0026291804804649314, "grad_norm": 0.155185, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:05:59.748115+00:00", "epoch": 0, "step": 12980, "train_loss": 3.457012176513672, "perplexity": 31.72205488013221, "lr": 0.0026291804804649314, "grad_norm": 0.178547, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:00.052911+00:00", "epoch": 0, "step": 12981, "train_loss": 3.558565616607666, "perplexity": 35.1127957951715, "lr": 0.0026291804804649314, "grad_norm": 0.158053, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:00.359219+00:00", "epoch": 0, "step": 12982, "train_loss": 3.465287923812866, "perplexity": 31.985667882903805, "lr": 0.0026291804804649314, "grad_norm": 0.187846, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:00.665543+00:00", "epoch": 0, "step": 12983, "train_loss": 3.4862804412841797, "perplexity": 32.664224966460885, "lr": 0.0026291804804649314, "grad_norm": 0.159144, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:00.971664+00:00", "epoch": 0, "step": 12984, "train_loss": 3.5207605361938477, "perplexity": 33.81013251767743, "lr": 0.0026291804804649314, "grad_norm": 0.170383, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:01.276425+00:00", "epoch": 0, "step": 12985, "train_loss": 3.5700855255126953, "perplexity": 35.51963085636574, "lr": 0.0026291804804649314, "grad_norm": 0.167139, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:01.582370+00:00", "epoch": 0, "step": 12986, "train_loss": 3.477884292602539, "perplexity": 32.39111939893394, "lr": 0.0026291804804649314, "grad_norm": 0.162217, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:01.888075+00:00", "epoch": 0, "step": 12987, "train_loss": 3.5868592262268066, "perplexity": 36.12045141895099, "lr": 0.0026291804804649314, "grad_norm": 0.146004, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:02.193884+00:00", "epoch": 0, "step": 12988, "train_loss": 3.4362094402313232, "perplexity": 31.06896590900424, "lr": 0.0026291804804649314, "grad_norm": 0.163276, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:02.499324+00:00", "epoch": 0, "step": 12989, "train_loss": 3.4538190364837646, "perplexity": 31.620923466175825, "lr": 0.0026291804804649314, "grad_norm": 0.167695, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:02.805793+00:00", "epoch": 0, "step": 12990, "train_loss": 3.5542984008789062, "perplexity": 34.963281153167195, "lr": 0.0026291804804649314, "grad_norm": 0.156546, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:03.112038+00:00", "epoch": 0, "step": 12991, "train_loss": 3.4637465476989746, "perplexity": 31.93640391534727, "lr": 0.0026291804804649314, "grad_norm": 0.191436, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:03.416934+00:00", "epoch": 0, "step": 12992, "train_loss": 3.6108572483062744, "perplexity": 36.997755487577464, "lr": 0.0026291804804649314, "grad_norm": 0.213554, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:03.723407+00:00", "epoch": 0, "step": 12993, "train_loss": 3.5531768798828125, "perplexity": 34.924091079614556, "lr": 0.0026291804804649314, "grad_norm": 0.191853, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:04.029290+00:00", "epoch": 0, "step": 12994, "train_loss": 3.533442497253418, "perplexity": 34.24164170970735, "lr": 0.0026291804804649314, "grad_norm": 0.180116, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:04.334961+00:00", "epoch": 0, "step": 12995, "train_loss": 3.563863515853882, "perplexity": 35.29931348926215, "lr": 0.0026291804804649314, "grad_norm": 0.147206, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:04.640187+00:00", "epoch": 0, "step": 12996, "train_loss": 3.5505051612854004, "perplexity": 34.83090827053029, "lr": 0.0026291804804649314, "grad_norm": 0.178505, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:04.945916+00:00", "epoch": 0, "step": 12997, "train_loss": 3.6358656883239746, "perplexity": 37.93467829488399, "lr": 0.0026291804804649314, "grad_norm": 0.168925, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:05.253121+00:00", "epoch": 0, "step": 12998, "train_loss": 3.6271374225616455, "perplexity": 37.605015127865755, "lr": 0.0026291804804649314, "grad_norm": 0.162561, "tokens_per_sec": 106666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:05.558808+00:00", "epoch": 0, "step": 12999, "train_loss": 3.5088369846343994, "perplexity": 33.409389546935216, "lr": 0.0026291804804649314, "grad_norm": 0.163747, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:05.866094+00:00", "epoch": 0, "step": 13000, "train_loss": 3.507761240005493, "perplexity": 33.373468899760724, "lr": 0.0026291804804649314, "grad_norm": 0.178966, "tokens_per_sec": 106705} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:06:08.940918+00:00", "step": 13000, "epoch": 0, "val_loss": 3.478449583053589, "val_ppl": 32.40943496574875, "eval_train_loss": 3.507761240005493, "eval_train_ppl": 33.373468899760724} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:06:09.901767+00:00", "step": 13000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4784_epoch_0000_step_0013000.pt", "val_loss": 3.478449583053589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:11.210400+00:00", "epoch": 0, "step": 13001, "train_loss": 3.5592105388641357, "perplexity": 35.13544812237653, "lr": 0.0026291804804649314, "grad_norm": 0.153077, "tokens_per_sec": 6131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:11.513725+00:00", "epoch": 0, "step": 13002, "train_loss": 3.538909435272217, "perplexity": 34.42935127335873, "lr": 0.0026291804804649314, "grad_norm": 0.165023, "tokens_per_sec": 108031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:11.818095+00:00", "epoch": 0, "step": 13003, "train_loss": 3.595207929611206, "perplexity": 36.42327267748525, "lr": 0.0026291804804649314, "grad_norm": 0.145017, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:12.122647+00:00", "epoch": 0, "step": 13004, "train_loss": 3.5082945823669434, "perplexity": 33.39127313192708, "lr": 0.0026291804804649314, "grad_norm": 0.149411, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:12.427831+00:00", "epoch": 0, "step": 13005, "train_loss": 3.591437816619873, "perplexity": 36.28621135466016, "lr": 0.0026291804804649314, "grad_norm": 0.168958, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:12.733970+00:00", "epoch": 0, "step": 13006, "train_loss": 3.6125285625457764, "perplexity": 37.05964206470666, "lr": 0.0026291804804649314, "grad_norm": 0.163654, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:13.038485+00:00", "epoch": 0, "step": 13007, "train_loss": 3.556989908218384, "perplexity": 35.05751183540401, "lr": 0.0026291804804649314, "grad_norm": 0.164711, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:13.343010+00:00", "epoch": 0, "step": 13008, "train_loss": 3.501683473587036, "perplexity": 33.171247899706486, "lr": 0.0026291804804649314, "grad_norm": 0.197268, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:13.646937+00:00", "epoch": 0, "step": 13009, "train_loss": 3.5134878158569336, "perplexity": 33.56513286618161, "lr": 0.0026291804804649314, "grad_norm": 0.156095, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:13.951669+00:00", "epoch": 0, "step": 13010, "train_loss": 3.539017677307129, "perplexity": 34.4330781781015, "lr": 0.0026291804804649314, "grad_norm": 0.168723, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:14.256398+00:00", "epoch": 0, "step": 13011, "train_loss": 3.552086353302002, "perplexity": 34.8860261891414, "lr": 0.0026291804804649314, "grad_norm": 0.200385, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:14.560547+00:00", "epoch": 0, "step": 13012, "train_loss": 3.5129098892211914, "perplexity": 33.54574028615004, "lr": 0.0026291804804649314, "grad_norm": 0.160019, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:14.866304+00:00", "epoch": 0, "step": 13013, "train_loss": 3.4934957027435303, "perplexity": 32.900758188321696, "lr": 0.0026291804804649314, "grad_norm": 0.162269, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:15.176222+00:00", "epoch": 0, "step": 13014, "train_loss": 3.513275623321533, "perplexity": 33.55801135113057, "lr": 0.0026291804804649314, "grad_norm": 0.161214, "tokens_per_sec": 105736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:15.481348+00:00", "epoch": 0, "step": 13015, "train_loss": 3.552260160446167, "perplexity": 34.892090156689775, "lr": 0.0026291804804649314, "grad_norm": 0.170507, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:15.786711+00:00", "epoch": 0, "step": 13016, "train_loss": 3.4377965927124023, "perplexity": 31.118316248228272, "lr": 0.0026291804804649314, "grad_norm": 0.161492, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:16.091413+00:00", "epoch": 0, "step": 13017, "train_loss": 3.461388111114502, "perplexity": 31.861172680895706, "lr": 0.0026291804804649314, "grad_norm": 0.213786, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:16.396358+00:00", "epoch": 0, "step": 13018, "train_loss": 3.563006639480591, "perplexity": 35.26907929687229, "lr": 0.0026291804804649314, "grad_norm": 0.167605, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:16.699928+00:00", "epoch": 0, "step": 13019, "train_loss": 3.567351818084717, "perplexity": 35.4226631786729, "lr": 0.0026291804804649314, "grad_norm": 0.204119, "tokens_per_sec": 107941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:17.004856+00:00", "epoch": 0, "step": 13020, "train_loss": 3.4866104125976562, "perplexity": 32.67500500213022, "lr": 0.0026291804804649314, "grad_norm": 0.193751, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:17.309701+00:00", "epoch": 0, "step": 13021, "train_loss": 3.5858538150787354, "perplexity": 36.08415376451185, "lr": 0.0026291804804649314, "grad_norm": 0.181193, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:17.614221+00:00", "epoch": 0, "step": 13022, "train_loss": 3.5548336505889893, "perplexity": 34.98200024851625, "lr": 0.0026291804804649314, "grad_norm": 0.179669, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:17.918706+00:00", "epoch": 0, "step": 13023, "train_loss": 3.349891185760498, "perplexity": 28.499632309219844, "lr": 0.0026291804804649314, "grad_norm": 0.161325, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:18.222941+00:00", "epoch": 0, "step": 13024, "train_loss": 3.5962979793548584, "perplexity": 36.46299750361514, "lr": 0.0026291804804649314, "grad_norm": 0.183428, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:18.528631+00:00", "epoch": 0, "step": 13025, "train_loss": 3.4872817993164062, "perplexity": 32.6969499324668, "lr": 0.0026291804804649314, "grad_norm": 0.18029, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:18.833819+00:00", "epoch": 0, "step": 13026, "train_loss": 3.4807653427124023, "perplexity": 32.48457439665937, "lr": 0.0026291804804649314, "grad_norm": 0.16586, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:19.138806+00:00", "epoch": 0, "step": 13027, "train_loss": 3.562317132949829, "perplexity": 35.2447694182363, "lr": 0.0026291804804649314, "grad_norm": 0.172453, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:19.442439+00:00", "epoch": 0, "step": 13028, "train_loss": 3.528083086013794, "perplexity": 34.0586175591976, "lr": 0.0026291804804649314, "grad_norm": 0.169585, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:19.747234+00:00", "epoch": 0, "step": 13029, "train_loss": 3.610574722290039, "perplexity": 36.987304135568856, "lr": 0.0026291804804649314, "grad_norm": 0.179387, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:20.052950+00:00", "epoch": 0, "step": 13030, "train_loss": 3.5070853233337402, "perplexity": 33.35091883757716, "lr": 0.0026291804804649314, "grad_norm": 0.186267, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:20.358417+00:00", "epoch": 0, "step": 13031, "train_loss": 3.4719228744506836, "perplexity": 32.19859681576461, "lr": 0.0026291804804649314, "grad_norm": 0.157122, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:20.662935+00:00", "epoch": 0, "step": 13032, "train_loss": 3.5687506198883057, "perplexity": 35.47224713479392, "lr": 0.0026291804804649314, "grad_norm": 0.156831, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:20.968743+00:00", "epoch": 0, "step": 13033, "train_loss": 3.5034432411193848, "perplexity": 33.22967297696742, "lr": 0.0026291804804649314, "grad_norm": 0.169126, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:21.273907+00:00", "epoch": 0, "step": 13034, "train_loss": 3.452951431274414, "perplexity": 31.59350088595899, "lr": 0.0026291804804649314, "grad_norm": 0.153477, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:21.578981+00:00", "epoch": 0, "step": 13035, "train_loss": 3.6382575035095215, "perplexity": 38.025519629022824, "lr": 0.0026291804804649314, "grad_norm": 0.150329, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:21.884389+00:00", "epoch": 0, "step": 13036, "train_loss": 3.5946033000946045, "perplexity": 36.40125674814428, "lr": 0.0026291804804649314, "grad_norm": 0.173695, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:22.190686+00:00", "epoch": 0, "step": 13037, "train_loss": 3.487700939178467, "perplexity": 32.71065740002365, "lr": 0.0026291804804649314, "grad_norm": 0.214953, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:22.496793+00:00", "epoch": 0, "step": 13038, "train_loss": 3.456336736679077, "perplexity": 31.700635775103663, "lr": 0.0026291804804649314, "grad_norm": 0.183309, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:22.801405+00:00", "epoch": 0, "step": 13039, "train_loss": 3.4878153800964355, "perplexity": 32.71440105189326, "lr": 0.0026291804804649314, "grad_norm": 0.194561, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:23.106174+00:00", "epoch": 0, "step": 13040, "train_loss": 3.584547996520996, "perplexity": 36.03706515816499, "lr": 0.0026291804804649314, "grad_norm": 0.17976, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:23.411461+00:00", "epoch": 0, "step": 13041, "train_loss": 3.6030774116516113, "perplexity": 36.71103575579611, "lr": 0.0026291804804649314, "grad_norm": 0.166596, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:23.718181+00:00", "epoch": 0, "step": 13042, "train_loss": 3.586621046066284, "perplexity": 36.11184926850534, "lr": 0.0026291804804649314, "grad_norm": 0.205721, "tokens_per_sec": 106834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:24.023941+00:00", "epoch": 0, "step": 13043, "train_loss": 3.58996319770813, "perplexity": 36.23274245397615, "lr": 0.0026291804804649314, "grad_norm": 0.194027, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:24.329206+00:00", "epoch": 0, "step": 13044, "train_loss": 3.5401952266693115, "perplexity": 34.473648709561836, "lr": 0.0026291804804649314, "grad_norm": 0.173495, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:24.634015+00:00", "epoch": 0, "step": 13045, "train_loss": 3.570774555206299, "perplexity": 35.54411337035255, "lr": 0.0026291804804649314, "grad_norm": 0.179916, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:24.938841+00:00", "epoch": 0, "step": 13046, "train_loss": 3.502225399017334, "perplexity": 33.189229114296715, "lr": 0.0026291804804649314, "grad_norm": 0.17311, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:25.244177+00:00", "epoch": 0, "step": 13047, "train_loss": 3.4971861839294434, "perplexity": 33.02240214171634, "lr": 0.0026291804804649314, "grad_norm": 0.179615, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:25.549742+00:00", "epoch": 0, "step": 13048, "train_loss": 3.5783560276031494, "perplexity": 35.814614186306244, "lr": 0.0026291804804649314, "grad_norm": 0.187938, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:25.855398+00:00", "epoch": 0, "step": 13049, "train_loss": 3.539829730987549, "perplexity": 34.46105104216064, "lr": 0.0026291804804649314, "grad_norm": 0.165771, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:26.160298+00:00", "epoch": 0, "step": 13050, "train_loss": 3.5020997524261475, "perplexity": 33.185059262763694, "lr": 0.0026291804804649314, "grad_norm": 0.152884, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:26.465059+00:00", "epoch": 0, "step": 13051, "train_loss": 3.513629913330078, "perplexity": 33.56990272563211, "lr": 0.0026291804804649314, "grad_norm": 0.161992, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:26.770383+00:00", "epoch": 0, "step": 13052, "train_loss": 3.4857847690582275, "perplexity": 32.64803822935218, "lr": 0.0026291804804649314, "grad_norm": 0.154372, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:27.075191+00:00", "epoch": 0, "step": 13053, "train_loss": 3.4634957313537598, "perplexity": 31.928394747694984, "lr": 0.0026291804804649314, "grad_norm": 0.167588, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:27.380155+00:00", "epoch": 0, "step": 13054, "train_loss": 3.4273693561553955, "perplexity": 30.795524043478512, "lr": 0.0026291804804649314, "grad_norm": 0.160309, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:27.685452+00:00", "epoch": 0, "step": 13055, "train_loss": 3.5408105850219727, "perplexity": 34.49486888557863, "lr": 0.0026291804804649314, "grad_norm": 0.205976, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:27.990511+00:00", "epoch": 0, "step": 13056, "train_loss": 3.5350027084350586, "perplexity": 34.29510760017482, "lr": 0.0026291804804649314, "grad_norm": 0.210895, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:28.294812+00:00", "epoch": 0, "step": 13057, "train_loss": 3.3642737865448, "perplexity": 28.91249303171948, "lr": 0.0026291804804649314, "grad_norm": 0.166361, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:28.600114+00:00", "epoch": 0, "step": 13058, "train_loss": 3.492002487182617, "perplexity": 32.851666925269534, "lr": 0.0026291804804649314, "grad_norm": 0.180871, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:28.905677+00:00", "epoch": 0, "step": 13059, "train_loss": 3.4288249015808105, "perplexity": 30.840380965347432, "lr": 0.0026291804804649314, "grad_norm": 0.202333, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:29.210574+00:00", "epoch": 0, "step": 13060, "train_loss": 3.4349873065948486, "perplexity": 31.03101867373588, "lr": 0.0026291804804649314, "grad_norm": 0.166499, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:29.514333+00:00", "epoch": 0, "step": 13061, "train_loss": 3.499645948410034, "perplexity": 33.103729455581785, "lr": 0.0026291804804649314, "grad_norm": 0.156337, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:29.819633+00:00", "epoch": 0, "step": 13062, "train_loss": 3.488548755645752, "perplexity": 32.738401793438044, "lr": 0.0026291804804649314, "grad_norm": 0.149192, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:30.125785+00:00", "epoch": 0, "step": 13063, "train_loss": 3.5323638916015625, "perplexity": 34.20472839246547, "lr": 0.0026291804804649314, "grad_norm": 0.158346, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:30.431826+00:00", "epoch": 0, "step": 13064, "train_loss": 3.5625596046447754, "perplexity": 35.25331631336333, "lr": 0.0026291804804649314, "grad_norm": 0.151401, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:30.737592+00:00", "epoch": 0, "step": 13065, "train_loss": 3.4623467922210693, "perplexity": 31.891732031173316, "lr": 0.0026291804804649314, "grad_norm": 0.163144, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:31.042969+00:00", "epoch": 0, "step": 13066, "train_loss": 3.452805757522583, "perplexity": 31.588898877355522, "lr": 0.0026291804804649314, "grad_norm": 0.160351, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:31.348506+00:00", "epoch": 0, "step": 13067, "train_loss": 3.465228796005249, "perplexity": 31.983776696398124, "lr": 0.0026291804804649314, "grad_norm": 0.17585, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:31.653750+00:00", "epoch": 0, "step": 13068, "train_loss": 3.4534718990325928, "perplexity": 31.609948564409997, "lr": 0.0026291804804649314, "grad_norm": 0.191203, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:31.959179+00:00", "epoch": 0, "step": 13069, "train_loss": 3.710022211074829, "perplexity": 40.854713944021405, "lr": 0.0026291804804649314, "grad_norm": 0.172011, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:32.264929+00:00", "epoch": 0, "step": 13070, "train_loss": 3.462874412536621, "perplexity": 31.90856319673308, "lr": 0.0026291804804649314, "grad_norm": 0.16195, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:32.569809+00:00", "epoch": 0, "step": 13071, "train_loss": 3.5619568824768066, "perplexity": 35.2320747601474, "lr": 0.0026291804804649314, "grad_norm": 0.155451, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:32.874385+00:00", "epoch": 0, "step": 13072, "train_loss": 3.4827041625976562, "perplexity": 32.54761723006485, "lr": 0.0026291804804649314, "grad_norm": 0.17443, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:33.179476+00:00", "epoch": 0, "step": 13073, "train_loss": 3.5299620628356934, "perplexity": 34.12267307274395, "lr": 0.0026291804804649314, "grad_norm": 0.149905, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:33.485571+00:00", "epoch": 0, "step": 13074, "train_loss": 3.5369155406951904, "perplexity": 34.36077117003298, "lr": 0.0026291804804649314, "grad_norm": 0.178401, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:33.790999+00:00", "epoch": 0, "step": 13075, "train_loss": 3.5422158241271973, "perplexity": 34.54337649868134, "lr": 0.0026291804804649314, "grad_norm": 0.177441, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:34.095795+00:00", "epoch": 0, "step": 13076, "train_loss": 3.572101593017578, "perplexity": 35.59131306372563, "lr": 0.0026291804804649314, "grad_norm": 0.17823, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:34.400181+00:00", "epoch": 0, "step": 13077, "train_loss": 3.5199732780456543, "perplexity": 33.78352568995654, "lr": 0.0026291804804649314, "grad_norm": 0.15677, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:34.704725+00:00", "epoch": 0, "step": 13078, "train_loss": 3.6032962799072266, "perplexity": 36.71907151550809, "lr": 0.0026291804804649314, "grad_norm": 0.166294, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:35.010690+00:00", "epoch": 0, "step": 13079, "train_loss": 3.5528347492218018, "perplexity": 34.912144521006816, "lr": 0.0026291804804649314, "grad_norm": 0.181181, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:35.316967+00:00", "epoch": 0, "step": 13080, "train_loss": 3.5213897228240967, "perplexity": 33.831412094735185, "lr": 0.0026291804804649314, "grad_norm": 0.193245, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:35.622371+00:00", "epoch": 0, "step": 13081, "train_loss": 3.5011813640594482, "perplexity": 33.154596480861926, "lr": 0.0026291804804649314, "grad_norm": 0.161591, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:35.927161+00:00", "epoch": 0, "step": 13082, "train_loss": 3.531771421432495, "perplexity": 34.184469113353764, "lr": 0.0026291804804649314, "grad_norm": 0.182174, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:36.232207+00:00", "epoch": 0, "step": 13083, "train_loss": 3.518216133117676, "perplexity": 33.7242152629096, "lr": 0.0026291804804649314, "grad_norm": 0.155506, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:36.538261+00:00", "epoch": 0, "step": 13084, "train_loss": 3.5142242908477783, "perplexity": 33.58986185212404, "lr": 0.0026291804804649314, "grad_norm": 0.175155, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:36.844227+00:00", "epoch": 0, "step": 13085, "train_loss": 3.5445899963378906, "perplexity": 34.6254858555359, "lr": 0.0026291804804649314, "grad_norm": 0.185559, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:37.148852+00:00", "epoch": 0, "step": 13086, "train_loss": 3.471306324005127, "perplexity": 32.17875087519198, "lr": 0.0026291804804649314, "grad_norm": 0.160841, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:37.453527+00:00", "epoch": 0, "step": 13087, "train_loss": 3.5927939414978027, "perplexity": 36.33545337021855, "lr": 0.0026291804804649314, "grad_norm": 0.191392, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:37.758500+00:00", "epoch": 0, "step": 13088, "train_loss": 3.598360538482666, "perplexity": 36.53828220482792, "lr": 0.0026291804804649314, "grad_norm": 0.157495, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:38.063911+00:00", "epoch": 0, "step": 13089, "train_loss": 3.485262155532837, "perplexity": 32.6309803807155, "lr": 0.0026291804804649314, "grad_norm": 0.183105, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:38.369743+00:00", "epoch": 0, "step": 13090, "train_loss": 3.5975747108459473, "perplexity": 36.509580691571294, "lr": 0.0026291804804649314, "grad_norm": 0.214481, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:38.676417+00:00", "epoch": 0, "step": 13091, "train_loss": 3.578089475631714, "perplexity": 35.80506900248898, "lr": 0.0026291804804649314, "grad_norm": 0.163862, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:38.982114+00:00", "epoch": 0, "step": 13092, "train_loss": 3.456772565841675, "perplexity": 31.714454847806042, "lr": 0.0026291804804649314, "grad_norm": 0.176098, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:39.287831+00:00", "epoch": 0, "step": 13093, "train_loss": 3.6163904666900635, "perplexity": 37.20303956546059, "lr": 0.0026291804804649314, "grad_norm": 0.173027, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:39.591998+00:00", "epoch": 0, "step": 13094, "train_loss": 3.5636816024780273, "perplexity": 35.292892656015404, "lr": 0.0026291804804649314, "grad_norm": 0.199918, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:39.897855+00:00", "epoch": 0, "step": 13095, "train_loss": 3.446681499481201, "perplexity": 31.396031497005236, "lr": 0.0026291804804649314, "grad_norm": 0.16824, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:40.204105+00:00", "epoch": 0, "step": 13096, "train_loss": 3.5908963680267334, "perplexity": 36.266569554564185, "lr": 0.0026291804804649314, "grad_norm": 0.158167, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:40.509568+00:00", "epoch": 0, "step": 13097, "train_loss": 3.4234094619750977, "perplexity": 30.67381815729144, "lr": 0.0026291804804649314, "grad_norm": 0.179201, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:40.814809+00:00", "epoch": 0, "step": 13098, "train_loss": 3.600280284881592, "perplexity": 36.60849381319142, "lr": 0.0026291804804649314, "grad_norm": 0.173574, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:41.119040+00:00", "epoch": 0, "step": 13099, "train_loss": 3.5239410400390625, "perplexity": 33.91783696056322, "lr": 0.0026291804804649314, "grad_norm": 0.194192, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:41.425370+00:00", "epoch": 0, "step": 13100, "train_loss": 3.5678088665008545, "perplexity": 35.438856751115516, "lr": 0.0026291804804649314, "grad_norm": 0.19515, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:41.731464+00:00", "epoch": 0, "step": 13101, "train_loss": 3.6020686626434326, "perplexity": 36.67402220671737, "lr": 0.0026291804804649314, "grad_norm": 0.157833, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:42.038285+00:00", "epoch": 0, "step": 13102, "train_loss": 3.4678587913513184, "perplexity": 32.06800459119555, "lr": 0.0026291804804649314, "grad_norm": 0.158026, "tokens_per_sec": 106798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:42.344421+00:00", "epoch": 0, "step": 13103, "train_loss": 3.569906711578369, "perplexity": 35.5132800192525, "lr": 0.0026291804804649314, "grad_norm": 0.163839, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:42.648740+00:00", "epoch": 0, "step": 13104, "train_loss": 3.507661819458008, "perplexity": 33.37015105614487, "lr": 0.0026291804804649314, "grad_norm": 0.166407, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:42.955036+00:00", "epoch": 0, "step": 13105, "train_loss": 3.517836570739746, "perplexity": 33.71141724854663, "lr": 0.0026291804804649314, "grad_norm": 0.16051, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:43.261065+00:00", "epoch": 0, "step": 13106, "train_loss": 3.4489402770996094, "perplexity": 31.467028303069846, "lr": 0.0026291804804649314, "grad_norm": 0.147393, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:43.567557+00:00", "epoch": 0, "step": 13107, "train_loss": 3.5307297706604004, "perplexity": 34.148879373965954, "lr": 0.0026291804804649314, "grad_norm": 0.160075, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:43.873156+00:00", "epoch": 0, "step": 13108, "train_loss": 3.545274257659912, "perplexity": 34.649186844173194, "lr": 0.0026291804804649314, "grad_norm": 0.192859, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:44.178874+00:00", "epoch": 0, "step": 13109, "train_loss": 3.6069958209991455, "perplexity": 36.8551668192977, "lr": 0.0026291804804649314, "grad_norm": 0.188759, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:44.485016+00:00", "epoch": 0, "step": 13110, "train_loss": 3.457123041152954, "perplexity": 31.725571929258376, "lr": 0.0026291804804649314, "grad_norm": 0.187738, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:44.791362+00:00", "epoch": 0, "step": 13111, "train_loss": 3.5906028747558594, "perplexity": 36.25592712225976, "lr": 0.0026291804804649314, "grad_norm": 0.183483, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:45.096500+00:00", "epoch": 0, "step": 13112, "train_loss": 3.4778382778167725, "perplexity": 32.38962896280537, "lr": 0.0026291804804649314, "grad_norm": 0.161249, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:45.402116+00:00", "epoch": 0, "step": 13113, "train_loss": 3.5006449222564697, "perplexity": 33.136815738941735, "lr": 0.0026291804804649314, "grad_norm": 0.147105, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:45.706986+00:00", "epoch": 0, "step": 13114, "train_loss": 3.450772285461426, "perplexity": 31.524728999984013, "lr": 0.0026291804804649314, "grad_norm": 0.163573, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:46.012678+00:00", "epoch": 0, "step": 13115, "train_loss": 3.553311824798584, "perplexity": 34.928804226144116, "lr": 0.0026291804804649314, "grad_norm": 0.183721, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:46.319001+00:00", "epoch": 0, "step": 13116, "train_loss": 3.4246532917022705, "perplexity": 30.711994901924328, "lr": 0.0026291804804649314, "grad_norm": 0.160278, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:46.623458+00:00", "epoch": 0, "step": 13117, "train_loss": 3.548246383666992, "perplexity": 34.752321782612846, "lr": 0.0026291804804649314, "grad_norm": 0.180811, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:46.928833+00:00", "epoch": 0, "step": 13118, "train_loss": 3.5069093704223633, "perplexity": 33.34505116254299, "lr": 0.0026291804804649314, "grad_norm": 0.167677, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:47.235206+00:00", "epoch": 0, "step": 13119, "train_loss": 3.44779896736145, "perplexity": 31.431135163727518, "lr": 0.0026291804804649314, "grad_norm": 0.184819, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:47.540609+00:00", "epoch": 0, "step": 13120, "train_loss": 3.51926851272583, "perplexity": 33.759724620738105, "lr": 0.0026291804804649314, "grad_norm": 0.203721, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:47.846326+00:00", "epoch": 0, "step": 13121, "train_loss": 3.437382698059082, "perplexity": 31.105439208565564, "lr": 0.0026291804804649314, "grad_norm": 0.20388, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:48.151097+00:00", "epoch": 0, "step": 13122, "train_loss": 3.5439951419830322, "perplexity": 34.604894859414685, "lr": 0.0026291804804649314, "grad_norm": 0.173549, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:48.455934+00:00", "epoch": 0, "step": 13123, "train_loss": 3.470221996307373, "perplexity": 32.14387747485275, "lr": 0.0026291804804649314, "grad_norm": 0.199996, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:48.761717+00:00", "epoch": 0, "step": 13124, "train_loss": 3.5206944942474365, "perplexity": 33.807899704448026, "lr": 0.0026291804804649314, "grad_norm": 0.188123, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:49.066993+00:00", "epoch": 0, "step": 13125, "train_loss": 3.504971504211426, "perplexity": 33.28049548492859, "lr": 0.0026291804804649314, "grad_norm": 0.158311, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:49.370891+00:00", "epoch": 0, "step": 13126, "train_loss": 3.5939066410064697, "perplexity": 36.37590631314191, "lr": 0.0026291804804649314, "grad_norm": 0.173327, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:49.675293+00:00", "epoch": 0, "step": 13127, "train_loss": 3.3503634929656982, "perplexity": 28.513096070170416, "lr": 0.0026291804804649314, "grad_norm": 0.171404, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:49.980832+00:00", "epoch": 0, "step": 13128, "train_loss": 3.663630962371826, "perplexity": 39.002703427139636, "lr": 0.0026291804804649314, "grad_norm": 0.166562, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:50.286411+00:00", "epoch": 0, "step": 13129, "train_loss": 3.4803476333618164, "perplexity": 32.47100811976169, "lr": 0.0026291804804649314, "grad_norm": 0.164337, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:50.592706+00:00", "epoch": 0, "step": 13130, "train_loss": 3.578984022140503, "perplexity": 35.83711263208379, "lr": 0.0026291804804649314, "grad_norm": 0.16324, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:50.897591+00:00", "epoch": 0, "step": 13131, "train_loss": 3.559540271759033, "perplexity": 35.14703534563879, "lr": 0.0026291804804649314, "grad_norm": 0.168631, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:51.202259+00:00", "epoch": 0, "step": 13132, "train_loss": 3.5163633823394775, "perplexity": 33.661790543417965, "lr": 0.0026291804804649314, "grad_norm": 0.166525, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:51.507590+00:00", "epoch": 0, "step": 13133, "train_loss": 3.4785096645355225, "perplexity": 32.41138223112684, "lr": 0.0026291804804649314, "grad_norm": 0.169136, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:51.813632+00:00", "epoch": 0, "step": 13134, "train_loss": 3.5472357273101807, "perplexity": 34.71721687017883, "lr": 0.0026291804804649314, "grad_norm": 0.158334, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:52.119594+00:00", "epoch": 0, "step": 13135, "train_loss": 3.5161848068237305, "perplexity": 33.65577990850279, "lr": 0.0026291804804649314, "grad_norm": 0.162819, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:52.424499+00:00", "epoch": 0, "step": 13136, "train_loss": 3.4602484703063965, "perplexity": 31.824883070802436, "lr": 0.0026291804804649314, "grad_norm": 0.182408, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:52.729778+00:00", "epoch": 0, "step": 13137, "train_loss": 3.529829978942871, "perplexity": 34.11816631489261, "lr": 0.0026291804804649314, "grad_norm": 0.172553, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:53.034573+00:00", "epoch": 0, "step": 13138, "train_loss": 3.4969472885131836, "perplexity": 33.01451418344685, "lr": 0.0026291804804649314, "grad_norm": 0.157982, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:53.340390+00:00", "epoch": 0, "step": 13139, "train_loss": 3.4728610515594482, "perplexity": 32.22881897688463, "lr": 0.0026291804804649314, "grad_norm": 0.164562, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:53.646143+00:00", "epoch": 0, "step": 13140, "train_loss": 3.5048775672912598, "perplexity": 33.2773693645125, "lr": 0.0026291804804649314, "grad_norm": 0.161557, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:53.952187+00:00", "epoch": 0, "step": 13141, "train_loss": 3.5579423904418945, "perplexity": 35.09091939975766, "lr": 0.0026291804804649314, "grad_norm": 0.155828, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:54.257653+00:00", "epoch": 0, "step": 13142, "train_loss": 3.4994356632232666, "perplexity": 33.096768963520354, "lr": 0.0026291804804649314, "grad_norm": 0.151963, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:54.562678+00:00", "epoch": 0, "step": 13143, "train_loss": 3.5271012783050537, "perplexity": 34.02519495589853, "lr": 0.0026291804804649314, "grad_norm": 0.143037, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:54.867116+00:00", "epoch": 0, "step": 13144, "train_loss": 3.455537796020508, "perplexity": 31.675318962942384, "lr": 0.0026291804804649314, "grad_norm": 0.167355, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:55.173848+00:00", "epoch": 0, "step": 13145, "train_loss": 3.496732234954834, "perplexity": 33.00741505806796, "lr": 0.0026291804804649314, "grad_norm": 0.149295, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:55.479294+00:00", "epoch": 0, "step": 13146, "train_loss": 3.5405023097991943, "perplexity": 34.484236611105814, "lr": 0.0026291804804649314, "grad_norm": 0.172169, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:55.785408+00:00", "epoch": 0, "step": 13147, "train_loss": 3.5366854667663574, "perplexity": 34.35286656176917, "lr": 0.0026291804804649314, "grad_norm": 0.20053, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:56.090374+00:00", "epoch": 0, "step": 13148, "train_loss": 3.4236035346984863, "perplexity": 30.679771686408074, "lr": 0.0026291804804649314, "grad_norm": 0.206526, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:56.395170+00:00", "epoch": 0, "step": 13149, "train_loss": 3.511600971221924, "perplexity": 33.501860386750636, "lr": 0.0026291804804649314, "grad_norm": 0.165652, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:56.700022+00:00", "epoch": 0, "step": 13150, "train_loss": 3.464921236038208, "perplexity": 31.973941279659286, "lr": 0.0026291804804649314, "grad_norm": 0.180973, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:57.007241+00:00", "epoch": 0, "step": 13151, "train_loss": 3.5432214736938477, "perplexity": 34.578132503539344, "lr": 0.0026291804804649314, "grad_norm": 0.155916, "tokens_per_sec": 106601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:57.312843+00:00", "epoch": 0, "step": 13152, "train_loss": 3.558229446411133, "perplexity": 35.100993903541465, "lr": 0.0026291804804649314, "grad_norm": 0.163597, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:57.618864+00:00", "epoch": 0, "step": 13153, "train_loss": 3.547029972076416, "perplexity": 34.710074355936094, "lr": 0.0026291804804649314, "grad_norm": 0.166174, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:57.924382+00:00", "epoch": 0, "step": 13154, "train_loss": 3.5195930004119873, "perplexity": 33.77068101317667, "lr": 0.0026291804804649314, "grad_norm": 0.16044, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:58.229509+00:00", "epoch": 0, "step": 13155, "train_loss": 3.505861759185791, "perplexity": 33.31013680379148, "lr": 0.0026291804804649314, "grad_norm": 0.154767, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:58.534109+00:00", "epoch": 0, "step": 13156, "train_loss": 3.550814151763916, "perplexity": 34.84167235245771, "lr": 0.0026291804804649314, "grad_norm": 0.16238, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:58.839266+00:00", "epoch": 0, "step": 13157, "train_loss": 3.537346363067627, "perplexity": 34.37557774826077, "lr": 0.0026291804804649314, "grad_norm": 0.160837, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:59.144809+00:00", "epoch": 0, "step": 13158, "train_loss": 3.4188101291656494, "perplexity": 30.533062997088955, "lr": 0.0026291804804649314, "grad_norm": 0.165597, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:59.449659+00:00", "epoch": 0, "step": 13159, "train_loss": 3.4163143634796143, "perplexity": 30.456954639989906, "lr": 0.0026291804804649314, "grad_norm": 0.154303, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:06:59.754438+00:00", "epoch": 0, "step": 13160, "train_loss": 3.657078981399536, "perplexity": 38.747993793942484, "lr": 0.0026291804804649314, "grad_norm": 0.145071, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:00.059834+00:00", "epoch": 0, "step": 13161, "train_loss": 3.5877559185028076, "perplexity": 36.15285487453702, "lr": 0.0026291804804649314, "grad_norm": 0.163365, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:00.365811+00:00", "epoch": 0, "step": 13162, "train_loss": 3.4425570964813232, "perplexity": 31.26680827801216, "lr": 0.0026291804804649314, "grad_norm": 0.169597, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:00.670961+00:00", "epoch": 0, "step": 13163, "train_loss": 3.5942304134368896, "perplexity": 36.38768573556086, "lr": 0.0026291804804649314, "grad_norm": 0.153334, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:00.975810+00:00", "epoch": 0, "step": 13164, "train_loss": 3.5166451930999756, "perplexity": 33.67127813500064, "lr": 0.0026291804804649314, "grad_norm": 0.184638, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:01.280611+00:00", "epoch": 0, "step": 13165, "train_loss": 3.7058324813842773, "perplexity": 40.6839018142043, "lr": 0.0026291804804649314, "grad_norm": 0.191017, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:01.586397+00:00", "epoch": 0, "step": 13166, "train_loss": 3.613090991973877, "perplexity": 37.0804913605779, "lr": 0.0026291804804649314, "grad_norm": 0.192566, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:01.891423+00:00", "epoch": 0, "step": 13167, "train_loss": 3.4942734241485596, "perplexity": 32.926355764807795, "lr": 0.0026291804804649314, "grad_norm": 0.184285, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:02.197075+00:00", "epoch": 0, "step": 13168, "train_loss": 3.513353109359741, "perplexity": 33.560611729225506, "lr": 0.0026291804804649314, "grad_norm": 0.168686, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:02.503529+00:00", "epoch": 0, "step": 13169, "train_loss": 3.511427164077759, "perplexity": 33.49603803007085, "lr": 0.0026291804804649314, "grad_norm": 0.180382, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:02.809038+00:00", "epoch": 0, "step": 13170, "train_loss": 3.5043227672576904, "perplexity": 33.25891219936332, "lr": 0.0026291804804649314, "grad_norm": 0.173863, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:03.114177+00:00", "epoch": 0, "step": 13171, "train_loss": 3.497692346572876, "perplexity": 33.03912107897067, "lr": 0.0026291804804649314, "grad_norm": 0.162198, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:03.419952+00:00", "epoch": 0, "step": 13172, "train_loss": 3.5090582370758057, "perplexity": 33.41678227373764, "lr": 0.0026291804804649314, "grad_norm": 0.151419, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:03.725043+00:00", "epoch": 0, "step": 13173, "train_loss": 3.4826767444610596, "perplexity": 32.54672484728353, "lr": 0.0026291804804649314, "grad_norm": 0.18654, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:04.030356+00:00", "epoch": 0, "step": 13174, "train_loss": 3.5503451824188232, "perplexity": 34.82533650699741, "lr": 0.0026291804804649314, "grad_norm": 0.198349, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:04.336082+00:00", "epoch": 0, "step": 13175, "train_loss": 3.569815158843994, "perplexity": 35.51002883019001, "lr": 0.0026291804804649314, "grad_norm": 0.187086, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:04.640983+00:00", "epoch": 0, "step": 13176, "train_loss": 3.631408452987671, "perplexity": 37.765970770195516, "lr": 0.0026291804804649314, "grad_norm": 0.163279, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:04.946389+00:00", "epoch": 0, "step": 13177, "train_loss": 3.545311212539673, "perplexity": 34.65046732436668, "lr": 0.0026291804804649314, "grad_norm": 0.156973, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:05.251842+00:00", "epoch": 0, "step": 13178, "train_loss": 3.454525947570801, "perplexity": 31.64328455027432, "lr": 0.0026291804804649314, "grad_norm": 0.186676, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:05.558825+00:00", "epoch": 0, "step": 13179, "train_loss": 3.444387674331665, "perplexity": 31.324097024453977, "lr": 0.0026291804804649314, "grad_norm": 0.169501, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:05.864324+00:00", "epoch": 0, "step": 13180, "train_loss": 3.453108072280884, "perplexity": 31.59845011135137, "lr": 0.0026291804804649314, "grad_norm": 0.1571, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:06.168375+00:00", "epoch": 0, "step": 13181, "train_loss": 3.5273194313049316, "perplexity": 34.03261846395042, "lr": 0.0026291804804649314, "grad_norm": 0.171034, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:06.473171+00:00", "epoch": 0, "step": 13182, "train_loss": 3.520703077316284, "perplexity": 33.80818988122408, "lr": 0.0026291804804649314, "grad_norm": 0.169387, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:06.778129+00:00", "epoch": 0, "step": 13183, "train_loss": 3.584878444671631, "perplexity": 36.04897550746888, "lr": 0.0026291804804649314, "grad_norm": 0.152228, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:07.083726+00:00", "epoch": 0, "step": 13184, "train_loss": 3.459817886352539, "perplexity": 31.81118273660247, "lr": 0.0026291804804649314, "grad_norm": 0.171651, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:07.390947+00:00", "epoch": 0, "step": 13185, "train_loss": 3.5543055534362793, "perplexity": 34.96353123093594, "lr": 0.0026291804804649314, "grad_norm": 0.166893, "tokens_per_sec": 106664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:07.695682+00:00", "epoch": 0, "step": 13186, "train_loss": 3.5670769214630127, "perplexity": 35.4129269465233, "lr": 0.0026291804804649314, "grad_norm": 0.17334, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:08.000381+00:00", "epoch": 0, "step": 13187, "train_loss": 3.628340005874634, "perplexity": 37.65026549475781, "lr": 0.0026291804804649314, "grad_norm": 0.161297, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:08.304509+00:00", "epoch": 0, "step": 13188, "train_loss": 3.514346122741699, "perplexity": 33.593954417907476, "lr": 0.0026291804804649314, "grad_norm": 0.174253, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:08.608946+00:00", "epoch": 0, "step": 13189, "train_loss": 3.3883650302886963, "perplexity": 29.617488965763098, "lr": 0.0026291804804649314, "grad_norm": 0.154247, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:08.914475+00:00", "epoch": 0, "step": 13190, "train_loss": 3.583019733428955, "perplexity": 35.982033103977955, "lr": 0.0026291804804649314, "grad_norm": 0.159215, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:09.221490+00:00", "epoch": 0, "step": 13191, "train_loss": 3.641690969467163, "perplexity": 38.15630334843423, "lr": 0.0026291804804649314, "grad_norm": 0.17648, "tokens_per_sec": 106653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:09.526034+00:00", "epoch": 0, "step": 13192, "train_loss": 3.5396950244903564, "perplexity": 34.456409227333516, "lr": 0.0026291804804649314, "grad_norm": 0.170804, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:09.831265+00:00", "epoch": 0, "step": 13193, "train_loss": 3.345682382583618, "perplexity": 28.379935034046216, "lr": 0.0026291804804649314, "grad_norm": 0.187261, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:10.136568+00:00", "epoch": 0, "step": 13194, "train_loss": 3.457749843597412, "perplexity": 31.745463828789827, "lr": 0.0026291804804649314, "grad_norm": 0.169592, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:10.441787+00:00", "epoch": 0, "step": 13195, "train_loss": 3.6566338539123535, "perplexity": 38.7307498349964, "lr": 0.0026291804804649314, "grad_norm": 0.186132, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:10.747677+00:00", "epoch": 0, "step": 13196, "train_loss": 3.533111333847046, "perplexity": 34.2303040084185, "lr": 0.0026291804804649314, "grad_norm": 0.200345, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:11.053339+00:00", "epoch": 0, "step": 13197, "train_loss": 3.4778225421905518, "perplexity": 32.38911929572056, "lr": 0.0026291804804649314, "grad_norm": 0.180983, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:11.359028+00:00", "epoch": 0, "step": 13198, "train_loss": 3.497401714324951, "perplexity": 33.0295202401663, "lr": 0.0026291804804649314, "grad_norm": 0.192088, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:11.664069+00:00", "epoch": 0, "step": 13199, "train_loss": 3.5297789573669434, "perplexity": 34.11642559668694, "lr": 0.0026291804804649314, "grad_norm": 0.178068, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:11.969260+00:00", "epoch": 0, "step": 13200, "train_loss": 3.5684871673583984, "perplexity": 35.46290311245197, "lr": 0.0026291804804649314, "grad_norm": 0.158888, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:12.276221+00:00", "epoch": 0, "step": 13201, "train_loss": 3.4154417514801025, "perplexity": 30.430389128280307, "lr": 0.0026291804804649314, "grad_norm": 0.178768, "tokens_per_sec": 106682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:12.582425+00:00", "epoch": 0, "step": 13202, "train_loss": 3.4814584255218506, "perplexity": 32.507096700755255, "lr": 0.0026291804804649314, "grad_norm": 0.172076, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:12.886760+00:00", "epoch": 0, "step": 13203, "train_loss": 3.585611581802368, "perplexity": 36.07541404028929, "lr": 0.0026291804804649314, "grad_norm": 0.158095, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:13.192043+00:00", "epoch": 0, "step": 13204, "train_loss": 3.541372299194336, "perplexity": 34.51425058528357, "lr": 0.0026291804804649314, "grad_norm": 0.172248, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:13.498405+00:00", "epoch": 0, "step": 13205, "train_loss": 3.6519453525543213, "perplexity": 38.54958568783549, "lr": 0.0026291804804649314, "grad_norm": 0.179394, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:13.804312+00:00", "epoch": 0, "step": 13206, "train_loss": 3.523253917694092, "perplexity": 33.894539261995824, "lr": 0.0026291804804649314, "grad_norm": 0.192396, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:14.109284+00:00", "epoch": 0, "step": 13207, "train_loss": 3.504925489425659, "perplexity": 33.27896412529151, "lr": 0.0026291804804649314, "grad_norm": 0.174735, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:14.415024+00:00", "epoch": 0, "step": 13208, "train_loss": 3.495246648788452, "perplexity": 32.958416103965455, "lr": 0.0026291804804649314, "grad_norm": 0.159183, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:14.720704+00:00", "epoch": 0, "step": 13209, "train_loss": 3.5867624282836914, "perplexity": 36.11695520276524, "lr": 0.0026291804804649314, "grad_norm": 0.150626, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:15.026138+00:00", "epoch": 0, "step": 13210, "train_loss": 3.4597578048706055, "perplexity": 31.809271531016197, "lr": 0.0026291804804649314, "grad_norm": 0.151672, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:15.332597+00:00", "epoch": 0, "step": 13211, "train_loss": 3.506516218185425, "perplexity": 33.33194405780021, "lr": 0.0026291804804649314, "grad_norm": 0.159964, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:15.638101+00:00", "epoch": 0, "step": 13212, "train_loss": 3.533846855163574, "perplexity": 34.255490388112044, "lr": 0.0026291804804649314, "grad_norm": 0.157848, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:15.944053+00:00", "epoch": 0, "step": 13213, "train_loss": 3.4737155437469482, "perplexity": 32.25637002031203, "lr": 0.0026291804804649314, "grad_norm": 0.149132, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:16.250157+00:00", "epoch": 0, "step": 13214, "train_loss": 3.429964542388916, "perplexity": 30.875547957096323, "lr": 0.0026291804804649314, "grad_norm": 0.15551, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:16.555596+00:00", "epoch": 0, "step": 13215, "train_loss": 3.4432613849639893, "perplexity": 31.288836887309444, "lr": 0.0026291804804649314, "grad_norm": 0.171797, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:16.860810+00:00", "epoch": 0, "step": 13216, "train_loss": 3.38144588470459, "perplexity": 29.41326857781742, "lr": 0.0026291804804649314, "grad_norm": 0.181135, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:17.167259+00:00", "epoch": 0, "step": 13217, "train_loss": 3.5478367805480957, "perplexity": 34.73809003810424, "lr": 0.0026291804804649314, "grad_norm": 0.181873, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:17.472377+00:00", "epoch": 0, "step": 13218, "train_loss": 3.499223232269287, "perplexity": 33.089738932040376, "lr": 0.0026291804804649314, "grad_norm": 0.173248, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:17.777523+00:00", "epoch": 0, "step": 13219, "train_loss": 3.4600706100463867, "perplexity": 31.819223192173347, "lr": 0.0026291804804649314, "grad_norm": 0.170392, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:18.082713+00:00", "epoch": 0, "step": 13220, "train_loss": 3.627349615097046, "perplexity": 37.61299547802493, "lr": 0.0026291804804649314, "grad_norm": 0.167764, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:18.387109+00:00", "epoch": 0, "step": 13221, "train_loss": 3.5938432216644287, "perplexity": 36.37359945024802, "lr": 0.0026291804804649314, "grad_norm": 0.161878, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:18.692278+00:00", "epoch": 0, "step": 13222, "train_loss": 3.533256769180298, "perplexity": 34.23528266611687, "lr": 0.0026291804804649314, "grad_norm": 0.142906, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:18.998428+00:00", "epoch": 0, "step": 13223, "train_loss": 3.4953460693359375, "perplexity": 32.96169301063198, "lr": 0.0026291804804649314, "grad_norm": 0.162608, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:19.303451+00:00", "epoch": 0, "step": 13224, "train_loss": 3.525702476501465, "perplexity": 33.97763372395031, "lr": 0.0026291804804649314, "grad_norm": 0.186917, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:19.608661+00:00", "epoch": 0, "step": 13225, "train_loss": 3.624297857284546, "perplexity": 37.49838469634719, "lr": 0.0026291804804649314, "grad_norm": 0.16089, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:19.913238+00:00", "epoch": 0, "step": 13226, "train_loss": 3.417813301086426, "perplexity": 30.502641947346074, "lr": 0.0026291804804649314, "grad_norm": 0.176447, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:20.218560+00:00", "epoch": 0, "step": 13227, "train_loss": 3.5010454654693604, "perplexity": 33.150091124088, "lr": 0.0026291804804649314, "grad_norm": 0.166153, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:20.525394+00:00", "epoch": 0, "step": 13228, "train_loss": 3.5313479900360107, "perplexity": 34.169997399964146, "lr": 0.0026291804804649314, "grad_norm": 0.171991, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:20.832151+00:00", "epoch": 0, "step": 13229, "train_loss": 3.5936691761016846, "perplexity": 36.367269337542666, "lr": 0.0026291804804649314, "grad_norm": 0.187065, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:21.137970+00:00", "epoch": 0, "step": 13230, "train_loss": 3.50588321685791, "perplexity": 33.31085156945383, "lr": 0.0026291804804649314, "grad_norm": 0.152578, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:21.443860+00:00", "epoch": 0, "step": 13231, "train_loss": 3.52185320854187, "perplexity": 33.84709610542991, "lr": 0.0026291804804649314, "grad_norm": 0.225155, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:21.749278+00:00", "epoch": 0, "step": 13232, "train_loss": 3.420614242553711, "perplexity": 30.588197824606247, "lr": 0.0026291804804649314, "grad_norm": 0.202763, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:22.055294+00:00", "epoch": 0, "step": 13233, "train_loss": 3.5627405643463135, "perplexity": 35.25969632020628, "lr": 0.0026291804804649314, "grad_norm": 0.158137, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:22.361574+00:00", "epoch": 0, "step": 13234, "train_loss": 3.533298969268799, "perplexity": 34.236727428559576, "lr": 0.0026291804804649314, "grad_norm": 0.163123, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:22.667897+00:00", "epoch": 0, "step": 13235, "train_loss": 3.4664366245269775, "perplexity": 32.022430953282374, "lr": 0.0026291804804649314, "grad_norm": 0.182449, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:22.973368+00:00", "epoch": 0, "step": 13236, "train_loss": 3.5217819213867188, "perplexity": 33.844683328239455, "lr": 0.0026291804804649314, "grad_norm": 0.154104, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:23.279362+00:00", "epoch": 0, "step": 13237, "train_loss": 3.4390056133270264, "perplexity": 31.155961686534624, "lr": 0.0026291804804649314, "grad_norm": 0.182863, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:23.584896+00:00", "epoch": 0, "step": 13238, "train_loss": 3.637998342514038, "perplexity": 38.0156661743727, "lr": 0.0026291804804649314, "grad_norm": 0.1552, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:23.890662+00:00", "epoch": 0, "step": 13239, "train_loss": 3.53019642829895, "perplexity": 34.13067118603036, "lr": 0.0026291804804649314, "grad_norm": 0.16356, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:24.195852+00:00", "epoch": 0, "step": 13240, "train_loss": 3.593203067779541, "perplexity": 36.350322200558416, "lr": 0.0026291804804649314, "grad_norm": 0.155708, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:24.500637+00:00", "epoch": 0, "step": 13241, "train_loss": 3.4847326278686523, "perplexity": 32.61370594794311, "lr": 0.0026291804804649314, "grad_norm": 0.174308, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:24.806176+00:00", "epoch": 0, "step": 13242, "train_loss": 3.559406042098999, "perplexity": 35.142317887651615, "lr": 0.0026291804804649314, "grad_norm": 0.165587, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:25.111635+00:00", "epoch": 0, "step": 13243, "train_loss": 3.4789557456970215, "perplexity": 32.42584356338242, "lr": 0.0026291804804649314, "grad_norm": 0.168883, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:25.416922+00:00", "epoch": 0, "step": 13244, "train_loss": 3.4298763275146484, "perplexity": 30.87282439464647, "lr": 0.0026291804804649314, "grad_norm": 0.162896, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:25.722257+00:00", "epoch": 0, "step": 13245, "train_loss": 3.484703302383423, "perplexity": 32.61274954921455, "lr": 0.0026291804804649314, "grad_norm": 0.165935, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:26.028419+00:00", "epoch": 0, "step": 13246, "train_loss": 3.5789430141448975, "perplexity": 35.835643054058885, "lr": 0.0026291804804649314, "grad_norm": 0.20102, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:26.334472+00:00", "epoch": 0, "step": 13247, "train_loss": 3.436593532562256, "perplexity": 31.080901552587743, "lr": 0.0026291804804649314, "grad_norm": 0.204598, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:26.640492+00:00", "epoch": 0, "step": 13248, "train_loss": 3.5007643699645996, "perplexity": 33.14077409203987, "lr": 0.0026291804804649314, "grad_norm": 0.165887, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:26.945103+00:00", "epoch": 0, "step": 13249, "train_loss": 3.504837989807129, "perplexity": 33.27605235601667, "lr": 0.0026291804804649314, "grad_norm": 0.16859, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:27.250814+00:00", "epoch": 0, "step": 13250, "train_loss": 3.5076239109039307, "perplexity": 33.36888606594612, "lr": 0.0026291804804649314, "grad_norm": 0.184462, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:27.556905+00:00", "epoch": 0, "step": 13251, "train_loss": 3.5141561031341553, "perplexity": 33.58757151433076, "lr": 0.0026291804804649314, "grad_norm": 0.190504, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:27.862110+00:00", "epoch": 0, "step": 13252, "train_loss": 3.526057481765747, "perplexity": 33.989698104122695, "lr": 0.0026291804804649314, "grad_norm": 0.157229, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:28.166522+00:00", "epoch": 0, "step": 13253, "train_loss": 3.4997501373291016, "perplexity": 33.10717867705307, "lr": 0.0026291804804649314, "grad_norm": 0.172605, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:28.472493+00:00", "epoch": 0, "step": 13254, "train_loss": 3.600762367248535, "perplexity": 36.62614637719216, "lr": 0.0026291804804649314, "grad_norm": 0.198594, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:28.777484+00:00", "epoch": 0, "step": 13255, "train_loss": 3.750378370285034, "perplexity": 42.537173758108665, "lr": 0.0026291804804649314, "grad_norm": 0.17661, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:29.082419+00:00", "epoch": 0, "step": 13256, "train_loss": 3.518561840057373, "perplexity": 33.735875973640006, "lr": 0.0026291804804649314, "grad_norm": 0.168657, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:29.387806+00:00", "epoch": 0, "step": 13257, "train_loss": 3.4644525051116943, "perplexity": 31.95895761646648, "lr": 0.0026291804804649314, "grad_norm": 0.162448, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:29.693885+00:00", "epoch": 0, "step": 13258, "train_loss": 3.4918975830078125, "perplexity": 32.84822082901788, "lr": 0.0026291804804649314, "grad_norm": 0.208967, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:29.999720+00:00", "epoch": 0, "step": 13259, "train_loss": 3.551050901412964, "perplexity": 34.849922082681225, "lr": 0.0026291804804649314, "grad_norm": 0.175148, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:30.303997+00:00", "epoch": 0, "step": 13260, "train_loss": 3.489264726638794, "perplexity": 32.76184993257278, "lr": 0.0026291804804649314, "grad_norm": 0.159744, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:30.609574+00:00", "epoch": 0, "step": 13261, "train_loss": 3.4744670391082764, "perplexity": 32.28061964335396, "lr": 0.0026291804804649314, "grad_norm": 0.173433, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:30.915171+00:00", "epoch": 0, "step": 13262, "train_loss": 3.6474761962890625, "perplexity": 38.37768597489289, "lr": 0.0026291804804649314, "grad_norm": 0.167049, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:31.219929+00:00", "epoch": 0, "step": 13263, "train_loss": 3.5534796714782715, "perplexity": 34.93466740200247, "lr": 0.0026291804804649314, "grad_norm": 0.167547, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:31.525839+00:00", "epoch": 0, "step": 13264, "train_loss": 3.5393664836883545, "perplexity": 34.44509075040441, "lr": 0.0026291804804649314, "grad_norm": 0.174646, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:31.831137+00:00", "epoch": 0, "step": 13265, "train_loss": 3.5150725841522217, "perplexity": 33.618367996106485, "lr": 0.0026291804804649314, "grad_norm": 0.196136, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:32.136591+00:00", "epoch": 0, "step": 13266, "train_loss": 3.5498814582824707, "perplexity": 34.80919090174527, "lr": 0.0026291804804649314, "grad_norm": 0.162664, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:32.442150+00:00", "epoch": 0, "step": 13267, "train_loss": 3.5941781997680664, "perplexity": 36.38578585058903, "lr": 0.0026291804804649314, "grad_norm": 0.187679, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:32.748582+00:00", "epoch": 0, "step": 13268, "train_loss": 3.5420403480529785, "perplexity": 34.537315494379264, "lr": 0.0026291804804649314, "grad_norm": 0.187052, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:33.054078+00:00", "epoch": 0, "step": 13269, "train_loss": 3.4889039993286133, "perplexity": 32.750033969868404, "lr": 0.0026291804804649314, "grad_norm": 0.15276, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:33.359686+00:00", "epoch": 0, "step": 13270, "train_loss": 3.4542794227600098, "perplexity": 31.635484657010867, "lr": 0.0026291804804649314, "grad_norm": 0.172173, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:33.665506+00:00", "epoch": 0, "step": 13271, "train_loss": 3.5387513637542725, "perplexity": 34.42390940365475, "lr": 0.0026291804804649314, "grad_norm": 0.157651, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:33.970826+00:00", "epoch": 0, "step": 13272, "train_loss": 3.45426607131958, "perplexity": 31.63506228054168, "lr": 0.0026291804804649314, "grad_norm": 0.150193, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:34.277089+00:00", "epoch": 0, "step": 13273, "train_loss": 3.5293056964874268, "perplexity": 34.10028344712832, "lr": 0.0026291804804649314, "grad_norm": 0.15128, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:34.583049+00:00", "epoch": 0, "step": 13274, "train_loss": 3.4915125370025635, "perplexity": 32.83557518754124, "lr": 0.0026291804804649314, "grad_norm": 0.155913, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:34.888184+00:00", "epoch": 0, "step": 13275, "train_loss": 3.5328240394592285, "perplexity": 34.22047124669986, "lr": 0.0026291804804649314, "grad_norm": 0.163967, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:35.193228+00:00", "epoch": 0, "step": 13276, "train_loss": 3.592742919921875, "perplexity": 36.333599525419, "lr": 0.0026291804804649314, "grad_norm": 0.16025, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:35.499366+00:00", "epoch": 0, "step": 13277, "train_loss": 3.550832509994507, "perplexity": 34.84231198978421, "lr": 0.0026291804804649314, "grad_norm": 0.173132, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:35.805212+00:00", "epoch": 0, "step": 13278, "train_loss": 3.5749824047088623, "perplexity": 35.69399276393597, "lr": 0.0026291804804649314, "grad_norm": 0.17777, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:36.110911+00:00", "epoch": 0, "step": 13279, "train_loss": 3.5225868225097656, "perplexity": 33.87193591820832, "lr": 0.0026291804804649314, "grad_norm": 0.18913, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:36.416214+00:00", "epoch": 0, "step": 13280, "train_loss": 3.5163564682006836, "perplexity": 33.6615578019307, "lr": 0.0026291804804649314, "grad_norm": 0.152856, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:36.721972+00:00", "epoch": 0, "step": 13281, "train_loss": 3.4406161308288574, "perplexity": 31.20617933546401, "lr": 0.0026291804804649314, "grad_norm": 0.156016, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:37.027138+00:00", "epoch": 0, "step": 13282, "train_loss": 3.5190532207489014, "perplexity": 33.75245720522013, "lr": 0.0026291804804649314, "grad_norm": 0.155979, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:37.332376+00:00", "epoch": 0, "step": 13283, "train_loss": 3.4713470935821533, "perplexity": 32.18006281599785, "lr": 0.0026291804804649314, "grad_norm": 0.145748, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:37.639024+00:00", "epoch": 0, "step": 13284, "train_loss": 3.5430541038513184, "perplexity": 34.57234565123306, "lr": 0.0026291804804649314, "grad_norm": 0.179459, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:37.943784+00:00", "epoch": 0, "step": 13285, "train_loss": 3.47224497795105, "perplexity": 32.208969766998116, "lr": 0.0026291804804649314, "grad_norm": 0.181053, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:38.247746+00:00", "epoch": 0, "step": 13286, "train_loss": 3.490626573562622, "perplexity": 32.80649695142471, "lr": 0.0026291804804649314, "grad_norm": 0.169283, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:38.552892+00:00", "epoch": 0, "step": 13287, "train_loss": 3.5898702144622803, "perplexity": 36.22937357260393, "lr": 0.0026291804804649314, "grad_norm": 0.164518, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:38.858708+00:00", "epoch": 0, "step": 13288, "train_loss": 3.4313552379608154, "perplexity": 30.91851631594823, "lr": 0.0026291804804649314, "grad_norm": 0.171636, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:39.165512+00:00", "epoch": 0, "step": 13289, "train_loss": 3.497807741165161, "perplexity": 33.042933834858516, "lr": 0.0026291804804649314, "grad_norm": 0.150264, "tokens_per_sec": 106801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:39.472975+00:00", "epoch": 0, "step": 13290, "train_loss": 3.5009219646453857, "perplexity": 33.14599731331904, "lr": 0.0026291804804649314, "grad_norm": 0.207595, "tokens_per_sec": 106635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:39.778323+00:00", "epoch": 0, "step": 13291, "train_loss": 3.5502333641052246, "perplexity": 34.821442614307, "lr": 0.0026291804804649314, "grad_norm": 0.165383, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:40.083246+00:00", "epoch": 0, "step": 13292, "train_loss": 3.6201579570770264, "perplexity": 37.343466020948966, "lr": 0.0026291804804649314, "grad_norm": 0.153789, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:40.390488+00:00", "epoch": 0, "step": 13293, "train_loss": 3.5655078887939453, "perplexity": 35.357406475343765, "lr": 0.0026291804804649314, "grad_norm": 0.18129, "tokens_per_sec": 106663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:40.696650+00:00", "epoch": 0, "step": 13294, "train_loss": 3.613635301589966, "perplexity": 37.10068012256496, "lr": 0.0026291804804649314, "grad_norm": 0.17926, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:41.003364+00:00", "epoch": 0, "step": 13295, "train_loss": 3.4439775943756104, "perplexity": 31.311254273580065, "lr": 0.0026291804804649314, "grad_norm": 0.199989, "tokens_per_sec": 106835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:41.308756+00:00", "epoch": 0, "step": 13296, "train_loss": 3.4624686241149902, "perplexity": 31.895617697981358, "lr": 0.0026291804804649314, "grad_norm": 0.204497, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:41.613735+00:00", "epoch": 0, "step": 13297, "train_loss": 3.4754605293273926, "perplexity": 32.31270605936771, "lr": 0.0026291804804649314, "grad_norm": 0.21712, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:41.919475+00:00", "epoch": 0, "step": 13298, "train_loss": 3.443312168121338, "perplexity": 31.290425873582876, "lr": 0.0026291804804649314, "grad_norm": 0.159821, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:42.225357+00:00", "epoch": 0, "step": 13299, "train_loss": 3.518589496612549, "perplexity": 33.73680900465743, "lr": 0.0026291804804649314, "grad_norm": 0.18156, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:42.532088+00:00", "epoch": 0, "step": 13300, "train_loss": 3.5162975788116455, "perplexity": 33.659575551725, "lr": 0.0026291804804649314, "grad_norm": 0.162538, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:42.836825+00:00", "epoch": 0, "step": 13301, "train_loss": 3.5757973194122314, "perplexity": 35.723092178621314, "lr": 0.0026291804804649314, "grad_norm": 0.151276, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:43.142490+00:00", "epoch": 0, "step": 13302, "train_loss": 3.544750928878784, "perplexity": 34.631058671365984, "lr": 0.0026291804804649314, "grad_norm": 0.178051, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:43.447787+00:00", "epoch": 0, "step": 13303, "train_loss": 3.4477689266204834, "perplexity": 31.430190963320083, "lr": 0.0026291804804649314, "grad_norm": 0.173349, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:43.752817+00:00", "epoch": 0, "step": 13304, "train_loss": 3.487304925918579, "perplexity": 32.697706110564035, "lr": 0.0026291804804649314, "grad_norm": 0.17365, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:44.059663+00:00", "epoch": 0, "step": 13305, "train_loss": 3.5849053859710693, "perplexity": 36.049946726795376, "lr": 0.0026291804804649314, "grad_norm": 0.164103, "tokens_per_sec": 106796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:44.365347+00:00", "epoch": 0, "step": 13306, "train_loss": 3.568693161010742, "perplexity": 35.47020899784399, "lr": 0.0026291804804649314, "grad_norm": 0.15955, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:44.670693+00:00", "epoch": 0, "step": 13307, "train_loss": 3.5183069705963135, "perplexity": 33.72727882473413, "lr": 0.0026291804804649314, "grad_norm": 0.174557, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:44.976466+00:00", "epoch": 0, "step": 13308, "train_loss": 3.4772417545318604, "perplexity": 32.37031355655695, "lr": 0.0026291804804649314, "grad_norm": 0.170748, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:45.281346+00:00", "epoch": 0, "step": 13309, "train_loss": 3.482830762863159, "perplexity": 32.55173802788921, "lr": 0.0026291804804649314, "grad_norm": 0.170679, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:45.587111+00:00", "epoch": 0, "step": 13310, "train_loss": 3.438184976577759, "perplexity": 31.130404447454964, "lr": 0.0026291804804649314, "grad_norm": 0.171593, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:45.893887+00:00", "epoch": 0, "step": 13311, "train_loss": 3.5171315670013428, "perplexity": 33.68765894918851, "lr": 0.0026291804804649314, "grad_norm": 0.152176, "tokens_per_sec": 106744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:46.199918+00:00", "epoch": 0, "step": 13312, "train_loss": 3.6011486053466797, "perplexity": 36.64029552260486, "lr": 0.0026291804804649314, "grad_norm": 0.179555, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:46.504845+00:00", "epoch": 0, "step": 13313, "train_loss": 3.440549373626709, "perplexity": 31.20409616777583, "lr": 0.0026291804804649314, "grad_norm": 0.188477, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:46.810134+00:00", "epoch": 0, "step": 13314, "train_loss": 3.555244207382202, "perplexity": 34.99636529499038, "lr": 0.0026291804804649314, "grad_norm": 0.162456, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:47.116538+00:00", "epoch": 0, "step": 13315, "train_loss": 3.5327935218811035, "perplexity": 34.2194269367301, "lr": 0.0026291804804649314, "grad_norm": 0.156993, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:47.423528+00:00", "epoch": 0, "step": 13316, "train_loss": 3.5307464599609375, "perplexity": 34.14944929963265, "lr": 0.0026291804804649314, "grad_norm": 0.155003, "tokens_per_sec": 106741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:47.728797+00:00", "epoch": 0, "step": 13317, "train_loss": 3.628066062927246, "perplexity": 37.639952882656424, "lr": 0.0026291804804649314, "grad_norm": 0.158572, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:48.034349+00:00", "epoch": 0, "step": 13318, "train_loss": 3.6266257762908936, "perplexity": 37.58577958343093, "lr": 0.0026291804804649314, "grad_norm": 0.166722, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:48.340623+00:00", "epoch": 0, "step": 13319, "train_loss": 3.549853563308716, "perplexity": 34.80821991382154, "lr": 0.0026291804804649314, "grad_norm": 0.185384, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:48.647218+00:00", "epoch": 0, "step": 13320, "train_loss": 3.5701966285705566, "perplexity": 35.5235774152013, "lr": 0.0026291804804649314, "grad_norm": 0.192116, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:48.954106+00:00", "epoch": 0, "step": 13321, "train_loss": 3.5910983085632324, "perplexity": 36.273893984601784, "lr": 0.0026291804804649314, "grad_norm": 0.163199, "tokens_per_sec": 106720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:49.261445+00:00", "epoch": 0, "step": 13322, "train_loss": 3.5214240550994873, "perplexity": 33.832573624030935, "lr": 0.0026291804804649314, "grad_norm": 0.17011, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:49.566984+00:00", "epoch": 0, "step": 13323, "train_loss": 3.5899112224578857, "perplexity": 36.230859297059276, "lr": 0.0026291804804649314, "grad_norm": 0.183404, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:49.872214+00:00", "epoch": 0, "step": 13324, "train_loss": 3.5702857971191406, "perplexity": 35.52674514226864, "lr": 0.0026291804804649314, "grad_norm": 0.149431, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:50.178579+00:00", "epoch": 0, "step": 13325, "train_loss": 3.579728603363037, "perplexity": 35.86380620975458, "lr": 0.0026291804804649314, "grad_norm": 0.155135, "tokens_per_sec": 106958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:50.484483+00:00", "epoch": 0, "step": 13326, "train_loss": 3.537782669067383, "perplexity": 34.390579291471084, "lr": 0.0026291804804649314, "grad_norm": 0.168866, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:50.792169+00:00", "epoch": 0, "step": 13327, "train_loss": 3.5273399353027344, "perplexity": 34.03331627583857, "lr": 0.0026291804804649314, "grad_norm": 0.18829, "tokens_per_sec": 106497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:51.097814+00:00", "epoch": 0, "step": 13328, "train_loss": 3.5231995582580566, "perplexity": 33.89269682403427, "lr": 0.0026291804804649314, "grad_norm": 0.163259, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:51.404685+00:00", "epoch": 0, "step": 13329, "train_loss": 3.4486725330352783, "perplexity": 31.45860432080551, "lr": 0.0026291804804649314, "grad_norm": 0.158371, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:51.711456+00:00", "epoch": 0, "step": 13330, "train_loss": 3.5059070587158203, "perplexity": 33.31164577151141, "lr": 0.0026291804804649314, "grad_norm": 0.148149, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:52.018169+00:00", "epoch": 0, "step": 13331, "train_loss": 3.5486910343170166, "perplexity": 34.76777786110636, "lr": 0.0026291804804649314, "grad_norm": 0.17068, "tokens_per_sec": 106779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:52.323258+00:00", "epoch": 0, "step": 13332, "train_loss": 3.488443613052368, "perplexity": 32.73495977392484, "lr": 0.0026291804804649314, "grad_norm": 0.167932, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:52.629308+00:00", "epoch": 0, "step": 13333, "train_loss": 3.438889503479004, "perplexity": 31.15234438256458, "lr": 0.0026291804804649314, "grad_norm": 0.149084, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:52.935694+00:00", "epoch": 0, "step": 13334, "train_loss": 3.5894906520843506, "perplexity": 36.21562487482909, "lr": 0.0026291804804649314, "grad_norm": 0.147228, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:53.243070+00:00", "epoch": 0, "step": 13335, "train_loss": 3.581690549850464, "perplexity": 35.9342381476277, "lr": 0.0026291804804649314, "grad_norm": 0.170097, "tokens_per_sec": 106606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:53.550050+00:00", "epoch": 0, "step": 13336, "train_loss": 3.432921886444092, "perplexity": 30.966992725471503, "lr": 0.0026291804804649314, "grad_norm": 0.166287, "tokens_per_sec": 106743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:53.855100+00:00", "epoch": 0, "step": 13337, "train_loss": 3.435375928878784, "perplexity": 31.043080362654457, "lr": 0.0026291804804649314, "grad_norm": 0.188513, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:54.160597+00:00", "epoch": 0, "step": 13338, "train_loss": 3.5674984455108643, "perplexity": 35.427857493407274, "lr": 0.0026291804804649314, "grad_norm": 0.213886, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:54.466269+00:00", "epoch": 0, "step": 13339, "train_loss": 3.579396963119507, "perplexity": 35.851914300356114, "lr": 0.0026291804804649314, "grad_norm": 0.207547, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:54.771577+00:00", "epoch": 0, "step": 13340, "train_loss": 3.526322841644287, "perplexity": 33.99871880309608, "lr": 0.0026291804804649314, "grad_norm": 0.168406, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:55.076248+00:00", "epoch": 0, "step": 13341, "train_loss": 3.5290749073028564, "perplexity": 34.09241437860082, "lr": 0.0026291804804649314, "grad_norm": 0.199215, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:55.381925+00:00", "epoch": 0, "step": 13342, "train_loss": 3.462794542312622, "perplexity": 31.906014754416752, "lr": 0.0026291804804649314, "grad_norm": 0.20598, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:55.686527+00:00", "epoch": 0, "step": 13343, "train_loss": 3.5751492977142334, "perplexity": 35.69995033878699, "lr": 0.0026291804804649314, "grad_norm": 0.149116, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:55.993891+00:00", "epoch": 0, "step": 13344, "train_loss": 3.559298276901245, "perplexity": 35.13853097286739, "lr": 0.0026291804804649314, "grad_norm": 0.208184, "tokens_per_sec": 106610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:56.298844+00:00", "epoch": 0, "step": 13345, "train_loss": 3.598379611968994, "perplexity": 36.538979123900326, "lr": 0.0026291804804649314, "grad_norm": 0.188172, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:56.603918+00:00", "epoch": 0, "step": 13346, "train_loss": 3.566154718399048, "perplexity": 35.38028409077349, "lr": 0.0026291804804649314, "grad_norm": 0.159059, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:56.909294+00:00", "epoch": 0, "step": 13347, "train_loss": 3.530932664871216, "perplexity": 34.15580868683173, "lr": 0.0026291804804649314, "grad_norm": 0.183965, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:57.214098+00:00", "epoch": 0, "step": 13348, "train_loss": 3.4613513946533203, "perplexity": 31.860002872861504, "lr": 0.0026291804804649314, "grad_norm": 0.156687, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:57.519795+00:00", "epoch": 0, "step": 13349, "train_loss": 3.4566214084625244, "perplexity": 31.709661336226514, "lr": 0.0026291804804649314, "grad_norm": 0.195342, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:57.827510+00:00", "epoch": 0, "step": 13350, "train_loss": 3.4688868522644043, "perplexity": 32.10098940557045, "lr": 0.0026291804804649314, "grad_norm": 0.190783, "tokens_per_sec": 106561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:58.132598+00:00", "epoch": 0, "step": 13351, "train_loss": 3.4252777099609375, "perplexity": 30.73117802082274, "lr": 0.0026291804804649314, "grad_norm": 0.18399, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:58.438566+00:00", "epoch": 0, "step": 13352, "train_loss": 3.580983877182007, "perplexity": 35.90885337408618, "lr": 0.0026291804804649314, "grad_norm": 0.185855, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:58.743015+00:00", "epoch": 0, "step": 13353, "train_loss": 3.5284101963043213, "perplexity": 34.06976030583654, "lr": 0.0026291804804649314, "grad_norm": 0.16854, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:59.048198+00:00", "epoch": 0, "step": 13354, "train_loss": 3.5183475017547607, "perplexity": 33.728645858119705, "lr": 0.0026291804804649314, "grad_norm": 0.199217, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:59.354285+00:00", "epoch": 0, "step": 13355, "train_loss": 3.4495902061462402, "perplexity": 31.48748628617376, "lr": 0.0026291804804649314, "grad_norm": 0.168152, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:59.660831+00:00", "epoch": 0, "step": 13356, "train_loss": 3.6197750568389893, "perplexity": 37.329169936082224, "lr": 0.0026291804804649314, "grad_norm": 0.207729, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:07:59.966566+00:00", "epoch": 0, "step": 13357, "train_loss": 3.4096646308898926, "perplexity": 30.255095933178307, "lr": 0.0026291804804649314, "grad_norm": 0.169886, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:00.272265+00:00", "epoch": 0, "step": 13358, "train_loss": 3.508218288421631, "perplexity": 33.38872567713976, "lr": 0.0026291804804649314, "grad_norm": 0.208024, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:00.577610+00:00", "epoch": 0, "step": 13359, "train_loss": 3.5501630306243896, "perplexity": 34.81899358716556, "lr": 0.0026291804804649314, "grad_norm": 0.173236, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:00.883337+00:00", "epoch": 0, "step": 13360, "train_loss": 3.4807357788085938, "perplexity": 32.48361404002266, "lr": 0.0026291804804649314, "grad_norm": 0.179757, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:01.190158+00:00", "epoch": 0, "step": 13361, "train_loss": 3.4577343463897705, "perplexity": 31.744971866557222, "lr": 0.0026291804804649314, "grad_norm": 0.165098, "tokens_per_sec": 106715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:01.496570+00:00", "epoch": 0, "step": 13362, "train_loss": 3.562025308609009, "perplexity": 35.234485637235245, "lr": 0.0026291804804649314, "grad_norm": 0.165646, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:01.801218+00:00", "epoch": 0, "step": 13363, "train_loss": 3.495091676712036, "perplexity": 32.95330886553633, "lr": 0.0026291804804649314, "grad_norm": 0.16465, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:02.106145+00:00", "epoch": 0, "step": 13364, "train_loss": 3.6058032512664795, "perplexity": 36.811240660566405, "lr": 0.0026291804804649314, "grad_norm": 0.148419, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:02.411571+00:00", "epoch": 0, "step": 13365, "train_loss": 3.511164903640747, "perplexity": 33.487254496335964, "lr": 0.0026291804804649314, "grad_norm": 0.148416, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:02.717973+00:00", "epoch": 0, "step": 13366, "train_loss": 3.559204339981079, "perplexity": 35.135230322517536, "lr": 0.0026291804804649314, "grad_norm": 0.151532, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:03.024485+00:00", "epoch": 0, "step": 13367, "train_loss": 3.450805187225342, "perplexity": 31.525766236238443, "lr": 0.0026291804804649314, "grad_norm": 0.148133, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:03.330195+00:00", "epoch": 0, "step": 13368, "train_loss": 3.5540499687194824, "perplexity": 34.95459622858172, "lr": 0.0026291804804649314, "grad_norm": 0.15224, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:03.635529+00:00", "epoch": 0, "step": 13369, "train_loss": 3.5878820419311523, "perplexity": 36.157414884094216, "lr": 0.0026291804804649314, "grad_norm": 0.158266, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:03.940571+00:00", "epoch": 0, "step": 13370, "train_loss": 3.4766933917999268, "perplexity": 32.35256774899597, "lr": 0.0026291804804649314, "grad_norm": 0.166881, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:04.245882+00:00", "epoch": 0, "step": 13371, "train_loss": 3.541841506958008, "perplexity": 34.53044873946817, "lr": 0.0026291804804649314, "grad_norm": 0.156532, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:04.552507+00:00", "epoch": 0, "step": 13372, "train_loss": 3.398543357849121, "perplexity": 29.920484849834335, "lr": 0.0026291804804649314, "grad_norm": 0.159315, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:04.859586+00:00", "epoch": 0, "step": 13373, "train_loss": 3.384138822555542, "perplexity": 29.492583429141444, "lr": 0.0026291804804649314, "grad_norm": 0.162169, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:05.164183+00:00", "epoch": 0, "step": 13374, "train_loss": 3.5317764282226562, "perplexity": 34.18464026824585, "lr": 0.0026291804804649314, "grad_norm": 0.143872, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:05.468751+00:00", "epoch": 0, "step": 13375, "train_loss": 3.518002510070801, "perplexity": 33.71701176273589, "lr": 0.0026291804804649314, "grad_norm": 0.139989, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:05.773823+00:00", "epoch": 0, "step": 13376, "train_loss": 3.5910215377807617, "perplexity": 36.27110931626928, "lr": 0.0026291804804649314, "grad_norm": 0.146738, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:06.080102+00:00", "epoch": 0, "step": 13377, "train_loss": 3.4185471534729004, "perplexity": 30.52503459937854, "lr": 0.0026291804804649314, "grad_norm": 0.147125, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:06.386680+00:00", "epoch": 0, "step": 13378, "train_loss": 3.608243703842163, "perplexity": 36.90118645723451, "lr": 0.0026291804804649314, "grad_norm": 0.155958, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:06.694587+00:00", "epoch": 0, "step": 13379, "train_loss": 3.558367967605591, "perplexity": 35.10585647192026, "lr": 0.0026291804804649314, "grad_norm": 0.158452, "tokens_per_sec": 106421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:06.999187+00:00", "epoch": 0, "step": 13380, "train_loss": 3.5237977504730225, "perplexity": 33.91297723660631, "lr": 0.0026291804804649314, "grad_norm": 0.193644, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:07.306054+00:00", "epoch": 0, "step": 13381, "train_loss": 3.5611374378204346, "perplexity": 35.20321585050681, "lr": 0.0026291804804649314, "grad_norm": 0.185743, "tokens_per_sec": 106723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:07.612190+00:00", "epoch": 0, "step": 13382, "train_loss": 3.5807554721832275, "perplexity": 35.90065254906547, "lr": 0.0026291804804649314, "grad_norm": 0.169339, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:07.919152+00:00", "epoch": 0, "step": 13383, "train_loss": 3.4081811904907227, "perplexity": 30.21024757474806, "lr": 0.0026291804804649314, "grad_norm": 0.196109, "tokens_per_sec": 106750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:08.226780+00:00", "epoch": 0, "step": 13384, "train_loss": 3.5845370292663574, "perplexity": 36.03666993266224, "lr": 0.0026291804804649314, "grad_norm": 0.163813, "tokens_per_sec": 106518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:08.532905+00:00", "epoch": 0, "step": 13385, "train_loss": 3.5455074310302734, "perplexity": 34.657267053858185, "lr": 0.0026291804804649314, "grad_norm": 0.159288, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:08.838205+00:00", "epoch": 0, "step": 13386, "train_loss": 3.588772773742676, "perplexity": 36.18963579171729, "lr": 0.0026291804804649314, "grad_norm": 0.160288, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:09.144472+00:00", "epoch": 0, "step": 13387, "train_loss": 3.477621078491211, "perplexity": 32.38259472118338, "lr": 0.0026291804804649314, "grad_norm": 0.157589, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:09.450607+00:00", "epoch": 0, "step": 13388, "train_loss": 3.5186569690704346, "perplexity": 33.73908538687788, "lr": 0.0026291804804649314, "grad_norm": 0.151984, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:09.757395+00:00", "epoch": 0, "step": 13389, "train_loss": 3.5143797397613525, "perplexity": 33.59508376551592, "lr": 0.0026291804804649314, "grad_norm": 0.186649, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:10.063472+00:00", "epoch": 0, "step": 13390, "train_loss": 3.5550999641418457, "perplexity": 34.99131766991119, "lr": 0.0026291804804649314, "grad_norm": 0.17161, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:10.370296+00:00", "epoch": 0, "step": 13391, "train_loss": 3.5441040992736816, "perplexity": 34.60866552041855, "lr": 0.0026291804804649314, "grad_norm": 0.189365, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:10.675006+00:00", "epoch": 0, "step": 13392, "train_loss": 3.5284929275512695, "perplexity": 34.07257905618763, "lr": 0.0026291804804649314, "grad_norm": 0.175539, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:10.980694+00:00", "epoch": 0, "step": 13393, "train_loss": 3.439500570297241, "perplexity": 31.171386363896012, "lr": 0.0026291804804649314, "grad_norm": 0.194071, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:11.287482+00:00", "epoch": 0, "step": 13394, "train_loss": 3.58201265335083, "perplexity": 35.94581455581876, "lr": 0.0026291804804649314, "grad_norm": 0.204702, "tokens_per_sec": 106812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:11.593133+00:00", "epoch": 0, "step": 13395, "train_loss": 3.5856730937957764, "perplexity": 36.07763317917106, "lr": 0.0026291804804649314, "grad_norm": 0.172025, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:11.898772+00:00", "epoch": 0, "step": 13396, "train_loss": 3.5219709873199463, "perplexity": 33.85108280982062, "lr": 0.0026291804804649314, "grad_norm": 0.164001, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:12.203324+00:00", "epoch": 0, "step": 13397, "train_loss": 3.533019542694092, "perplexity": 34.22716211354888, "lr": 0.0026291804804649314, "grad_norm": 0.15613, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:12.508619+00:00", "epoch": 0, "step": 13398, "train_loss": 3.437795400619507, "perplexity": 31.118279152326664, "lr": 0.0026291804804649314, "grad_norm": 0.16803, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:12.814682+00:00", "epoch": 0, "step": 13399, "train_loss": 3.5593535900115967, "perplexity": 35.14047464806355, "lr": 0.0026291804804649314, "grad_norm": 0.159026, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:13.120819+00:00", "epoch": 0, "step": 13400, "train_loss": 3.4546399116516113, "perplexity": 31.64689095360824, "lr": 0.0026291804804649314, "grad_norm": 0.162693, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:13.425261+00:00", "epoch": 0, "step": 13401, "train_loss": 3.5082991123199463, "perplexity": 33.391424393167675, "lr": 0.0026291804804649314, "grad_norm": 0.179747, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:13.730416+00:00", "epoch": 0, "step": 13402, "train_loss": 3.508310556411743, "perplexity": 33.39180652988026, "lr": 0.0026291804804649314, "grad_norm": 0.160176, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:14.035180+00:00", "epoch": 0, "step": 13403, "train_loss": 3.564008951187134, "perplexity": 35.30444763001671, "lr": 0.0026291804804649314, "grad_norm": 0.154821, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:14.340752+00:00", "epoch": 0, "step": 13404, "train_loss": 3.4916396141052246, "perplexity": 32.839748102435635, "lr": 0.0026291804804649314, "grad_norm": 0.167776, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:14.648582+00:00", "epoch": 0, "step": 13405, "train_loss": 3.510622501373291, "perplexity": 33.46909585865459, "lr": 0.0026291804804649314, "grad_norm": 0.174918, "tokens_per_sec": 106449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:14.954415+00:00", "epoch": 0, "step": 13406, "train_loss": 3.562556505203247, "perplexity": 35.253207047940066, "lr": 0.0026291804804649314, "grad_norm": 0.190315, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:15.259357+00:00", "epoch": 0, "step": 13407, "train_loss": 3.537177562713623, "perplexity": 34.3697756282815, "lr": 0.0026291804804649314, "grad_norm": 0.186268, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:15.564050+00:00", "epoch": 0, "step": 13408, "train_loss": 3.512237310409546, "perplexity": 33.52318571772526, "lr": 0.0026291804804649314, "grad_norm": 0.160811, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:15.869401+00:00", "epoch": 0, "step": 13409, "train_loss": 3.4864444732666016, "perplexity": 32.66958338350028, "lr": 0.0026291804804649314, "grad_norm": 0.147402, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:16.175450+00:00", "epoch": 0, "step": 13410, "train_loss": 3.4653913974761963, "perplexity": 31.988977728371673, "lr": 0.0026291804804649314, "grad_norm": 0.172647, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:16.481106+00:00", "epoch": 0, "step": 13411, "train_loss": 3.510399341583252, "perplexity": 33.46162773557336, "lr": 0.0026291804804649314, "grad_norm": 0.153441, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:16.786599+00:00", "epoch": 0, "step": 13412, "train_loss": 3.570915460586548, "perplexity": 35.549122080031395, "lr": 0.0026291804804649314, "grad_norm": 0.15511, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:17.091575+00:00", "epoch": 0, "step": 13413, "train_loss": 3.611318349838257, "perplexity": 37.01481914304921, "lr": 0.0026291804804649314, "grad_norm": 0.15383, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:17.396735+00:00", "epoch": 0, "step": 13414, "train_loss": 3.5531005859375, "perplexity": 34.921426684559535, "lr": 0.0026291804804649314, "grad_norm": 0.182454, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:17.701831+00:00", "epoch": 0, "step": 13415, "train_loss": 3.4567956924438477, "perplexity": 31.715188303867578, "lr": 0.0026291804804649314, "grad_norm": 0.152132, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:18.007218+00:00", "epoch": 0, "step": 13416, "train_loss": 3.550199031829834, "perplexity": 34.82024713547154, "lr": 0.0026291804804649314, "grad_norm": 0.174041, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:18.312273+00:00", "epoch": 0, "step": 13417, "train_loss": 3.4533441066741943, "perplexity": 31.60590931263235, "lr": 0.0026291804804649314, "grad_norm": 0.182593, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:18.617020+00:00", "epoch": 0, "step": 13418, "train_loss": 3.689852476119995, "perplexity": 40.03893982182186, "lr": 0.0026291804804649314, "grad_norm": 0.167934, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:18.921714+00:00", "epoch": 0, "step": 13419, "train_loss": 3.537834405899048, "perplexity": 34.39235859711017, "lr": 0.0026291804804649314, "grad_norm": 0.180768, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:19.227376+00:00", "epoch": 0, "step": 13420, "train_loss": 3.535430669784546, "perplexity": 34.30978772174247, "lr": 0.0026291804804649314, "grad_norm": 0.172567, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:19.534438+00:00", "epoch": 0, "step": 13421, "train_loss": 3.596928358078003, "perplexity": 36.485990247730435, "lr": 0.0026291804804649314, "grad_norm": 0.186333, "tokens_per_sec": 106623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:19.839834+00:00", "epoch": 0, "step": 13422, "train_loss": 3.528376817703247, "perplexity": 34.06862312387747, "lr": 0.0026291804804649314, "grad_norm": 0.169171, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:20.144753+00:00", "epoch": 0, "step": 13423, "train_loss": 3.5759739875793457, "perplexity": 35.729403869361164, "lr": 0.0026291804804649314, "grad_norm": 0.157485, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:20.450139+00:00", "epoch": 0, "step": 13424, "train_loss": 3.5037550926208496, "perplexity": 33.240037316361885, "lr": 0.0026291804804649314, "grad_norm": 0.162563, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:20.755101+00:00", "epoch": 0, "step": 13425, "train_loss": 3.5409936904907227, "perplexity": 34.501185663015995, "lr": 0.0026291804804649314, "grad_norm": 0.184257, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:21.060740+00:00", "epoch": 0, "step": 13426, "train_loss": 3.5215070247650146, "perplexity": 33.83538081780281, "lr": 0.0026291804804649314, "grad_norm": 0.159101, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:21.365438+00:00", "epoch": 0, "step": 13427, "train_loss": 3.4634029865264893, "perplexity": 31.925433691552527, "lr": 0.0026291804804649314, "grad_norm": 0.177652, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:21.669943+00:00", "epoch": 0, "step": 13428, "train_loss": 3.5637381076812744, "perplexity": 35.29488694443141, "lr": 0.0026291804804649314, "grad_norm": 0.20828, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:21.973816+00:00", "epoch": 0, "step": 13429, "train_loss": 3.4868147373199463, "perplexity": 32.68168199556729, "lr": 0.0026291804804649314, "grad_norm": 0.170618, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:22.278214+00:00", "epoch": 0, "step": 13430, "train_loss": 3.4833319187164307, "perplexity": 32.56805561042052, "lr": 0.0026291804804649314, "grad_norm": 0.156661, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:22.584079+00:00", "epoch": 0, "step": 13431, "train_loss": 3.512556791305542, "perplexity": 33.53389744614019, "lr": 0.0026291804804649314, "grad_norm": 0.197648, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:22.891020+00:00", "epoch": 0, "step": 13432, "train_loss": 3.5037569999694824, "perplexity": 33.24010071676208, "lr": 0.0026291804804649314, "grad_norm": 0.193922, "tokens_per_sec": 106758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:23.197017+00:00", "epoch": 0, "step": 13433, "train_loss": 3.5383968353271484, "perplexity": 34.411707312320345, "lr": 0.0026291804804649314, "grad_norm": 0.172541, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:23.501323+00:00", "epoch": 0, "step": 13434, "train_loss": 3.5418179035186768, "perplexity": 34.529633711735045, "lr": 0.0026291804804649314, "grad_norm": 0.156767, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:23.806921+00:00", "epoch": 0, "step": 13435, "train_loss": 3.541715383529663, "perplexity": 34.526093915519304, "lr": 0.0026291804804649314, "grad_norm": 0.161771, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:24.111810+00:00", "epoch": 0, "step": 13436, "train_loss": 3.6227118968963623, "perplexity": 37.438960877964526, "lr": 0.0026291804804649314, "grad_norm": 0.182112, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:24.417774+00:00", "epoch": 0, "step": 13437, "train_loss": 3.52007794380188, "perplexity": 33.787061853275226, "lr": 0.0026291804804649314, "grad_norm": 0.149191, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:24.722708+00:00", "epoch": 0, "step": 13438, "train_loss": 3.5613040924072266, "perplexity": 35.209083116787966, "lr": 0.0026291804804649314, "grad_norm": 0.157707, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:25.026845+00:00", "epoch": 0, "step": 13439, "train_loss": 3.4855473041534424, "perplexity": 32.64028638649436, "lr": 0.0026291804804649314, "grad_norm": 0.166037, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:25.331788+00:00", "epoch": 0, "step": 13440, "train_loss": 3.5143086910247803, "perplexity": 33.592696962050034, "lr": 0.0026291804804649314, "grad_norm": 0.188451, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:25.636420+00:00", "epoch": 0, "step": 13441, "train_loss": 3.460709571838379, "perplexity": 31.83956095687936, "lr": 0.0026291804804649314, "grad_norm": 0.159701, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:25.940824+00:00", "epoch": 0, "step": 13442, "train_loss": 3.485442638397217, "perplexity": 32.636870245015935, "lr": 0.0026291804804649314, "grad_norm": 0.155163, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:26.246445+00:00", "epoch": 0, "step": 13443, "train_loss": 3.6392316818237305, "perplexity": 38.062581315046316, "lr": 0.0026291804804649314, "grad_norm": 0.160154, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:26.551236+00:00", "epoch": 0, "step": 13444, "train_loss": 3.4423320293426514, "perplexity": 31.259771938791697, "lr": 0.0026291804804649314, "grad_norm": 0.166846, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:26.855413+00:00", "epoch": 0, "step": 13445, "train_loss": 3.5250720977783203, "perplexity": 33.95622169614037, "lr": 0.0026291804804649314, "grad_norm": 0.164778, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:27.160443+00:00", "epoch": 0, "step": 13446, "train_loss": 3.563772201538086, "perplexity": 35.29609030376654, "lr": 0.0026291804804649314, "grad_norm": 0.161726, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:27.465364+00:00", "epoch": 0, "step": 13447, "train_loss": 3.511901617050171, "perplexity": 33.511934095547815, "lr": 0.0026291804804649314, "grad_norm": 0.164556, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:27.770837+00:00", "epoch": 0, "step": 13448, "train_loss": 3.491671085357666, "perplexity": 32.84078162670134, "lr": 0.0026291804804649314, "grad_norm": 0.163417, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:28.076631+00:00", "epoch": 0, "step": 13449, "train_loss": 3.4869325160980225, "perplexity": 32.68553143082464, "lr": 0.0026291804804649314, "grad_norm": 0.14995, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:28.381220+00:00", "epoch": 0, "step": 13450, "train_loss": 3.606942892074585, "perplexity": 36.853216166576885, "lr": 0.0026291804804649314, "grad_norm": 0.155674, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:28.686364+00:00", "epoch": 0, "step": 13451, "train_loss": 3.5021092891693115, "perplexity": 33.185375741659854, "lr": 0.0026291804804649314, "grad_norm": 0.155908, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:28.990842+00:00", "epoch": 0, "step": 13452, "train_loss": 3.473243236541748, "perplexity": 32.24113870155178, "lr": 0.0026291804804649314, "grad_norm": 0.149742, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:29.296007+00:00", "epoch": 0, "step": 13453, "train_loss": 3.4970502853393555, "perplexity": 33.017914748746065, "lr": 0.0026291804804649314, "grad_norm": 0.159381, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:29.601436+00:00", "epoch": 0, "step": 13454, "train_loss": 3.558640718460083, "perplexity": 35.11543293020488, "lr": 0.0026291804804649314, "grad_norm": 0.163423, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:29.906531+00:00", "epoch": 0, "step": 13455, "train_loss": 3.476623773574829, "perplexity": 32.35031549905167, "lr": 0.0026291804804649314, "grad_norm": 0.151652, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:30.211330+00:00", "epoch": 0, "step": 13456, "train_loss": 3.5353872776031494, "perplexity": 34.308298977510134, "lr": 0.0026291804804649314, "grad_norm": 0.154577, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:30.516234+00:00", "epoch": 0, "step": 13457, "train_loss": 3.4790241718292236, "perplexity": 32.42806241435369, "lr": 0.0026291804804649314, "grad_norm": 0.16521, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:30.820515+00:00", "epoch": 0, "step": 13458, "train_loss": 3.5141141414642334, "perplexity": 33.58616215331118, "lr": 0.0026291804804649314, "grad_norm": 0.204345, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:31.125356+00:00", "epoch": 0, "step": 13459, "train_loss": 3.633326768875122, "perplexity": 37.838487364558496, "lr": 0.0026291804804649314, "grad_norm": 0.195354, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:31.430165+00:00", "epoch": 0, "step": 13460, "train_loss": 3.5613226890563965, "perplexity": 35.20973789384259, "lr": 0.0026291804804649314, "grad_norm": 0.163383, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:31.734669+00:00", "epoch": 0, "step": 13461, "train_loss": 3.483826160430908, "perplexity": 32.58415608049684, "lr": 0.0026291804804649314, "grad_norm": 0.169352, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:32.039719+00:00", "epoch": 0, "step": 13462, "train_loss": 3.5558712482452393, "perplexity": 35.01831632746607, "lr": 0.0026291804804649314, "grad_norm": 0.171025, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:32.344636+00:00", "epoch": 0, "step": 13463, "train_loss": 3.435039520263672, "perplexity": 31.032638959368317, "lr": 0.0026291804804649314, "grad_norm": 0.182376, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:32.650243+00:00", "epoch": 0, "step": 13464, "train_loss": 3.4687607288360596, "perplexity": 32.09694097403976, "lr": 0.0026291804804649314, "grad_norm": 0.182361, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:32.955572+00:00", "epoch": 0, "step": 13465, "train_loss": 3.4602978229522705, "perplexity": 31.82645375174492, "lr": 0.0026291804804649314, "grad_norm": 0.19607, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:33.259356+00:00", "epoch": 0, "step": 13466, "train_loss": 3.6031947135925293, "perplexity": 36.7153422841205, "lr": 0.0026291804804649314, "grad_norm": 0.154728, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:33.563565+00:00", "epoch": 0, "step": 13467, "train_loss": 3.461947202682495, "perplexity": 31.878990974452666, "lr": 0.0026291804804649314, "grad_norm": 0.149009, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:33.867536+00:00", "epoch": 0, "step": 13468, "train_loss": 3.624011516571045, "perplexity": 37.4876489192365, "lr": 0.0026291804804649314, "grad_norm": 0.178182, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:34.173151+00:00", "epoch": 0, "step": 13469, "train_loss": 3.5173146724700928, "perplexity": 33.693827908539355, "lr": 0.0026291804804649314, "grad_norm": 0.168158, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:34.479011+00:00", "epoch": 0, "step": 13470, "train_loss": 3.5686936378479004, "perplexity": 35.47022591136168, "lr": 0.0026291804804649314, "grad_norm": 0.145853, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:34.783814+00:00", "epoch": 0, "step": 13471, "train_loss": 3.541313886642456, "perplexity": 34.51223457871128, "lr": 0.0026291804804649314, "grad_norm": 0.151896, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:35.088142+00:00", "epoch": 0, "step": 13472, "train_loss": 3.538026809692383, "perplexity": 34.39897645399596, "lr": 0.0026291804804649314, "grad_norm": 0.150519, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:35.393131+00:00", "epoch": 0, "step": 13473, "train_loss": 3.4513816833496094, "perplexity": 31.543945958054287, "lr": 0.0026291804804649314, "grad_norm": 0.147229, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:35.698381+00:00", "epoch": 0, "step": 13474, "train_loss": 3.6234843730926514, "perplexity": 37.467892757211885, "lr": 0.0026291804804649314, "grad_norm": 0.143585, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:36.003954+00:00", "epoch": 0, "step": 13475, "train_loss": 3.420363187789917, "perplexity": 30.580519475709874, "lr": 0.0026291804804649314, "grad_norm": 0.171299, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:36.309058+00:00", "epoch": 0, "step": 13476, "train_loss": 3.5890049934387207, "perplexity": 36.198040713797276, "lr": 0.0026291804804649314, "grad_norm": 0.164238, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:36.613963+00:00", "epoch": 0, "step": 13477, "train_loss": 3.5627737045288086, "perplexity": 35.26086485233963, "lr": 0.0026291804804649314, "grad_norm": 0.161475, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:36.918824+00:00", "epoch": 0, "step": 13478, "train_loss": 3.5223848819732666, "perplexity": 33.86509649189869, "lr": 0.0026291804804649314, "grad_norm": 0.181771, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:37.224049+00:00", "epoch": 0, "step": 13479, "train_loss": 3.4717490673065186, "perplexity": 32.19300095592023, "lr": 0.0026291804804649314, "grad_norm": 0.202383, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:37.529465+00:00", "epoch": 0, "step": 13480, "train_loss": 3.4364078044891357, "perplexity": 31.07512949266492, "lr": 0.0026291804804649314, "grad_norm": 0.186538, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:37.835215+00:00", "epoch": 0, "step": 13481, "train_loss": 3.4899990558624268, "perplexity": 32.785916751801395, "lr": 0.0026291804804649314, "grad_norm": 0.177333, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:38.140840+00:00", "epoch": 0, "step": 13482, "train_loss": 3.5487353801727295, "perplexity": 34.76931970215372, "lr": 0.0026291804804649314, "grad_norm": 0.173153, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:38.444721+00:00", "epoch": 0, "step": 13483, "train_loss": 3.6650261878967285, "perplexity": 39.057158974552394, "lr": 0.0026291804804649314, "grad_norm": 0.165157, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:38.749324+00:00", "epoch": 0, "step": 13484, "train_loss": 3.5727264881134033, "perplexity": 35.61356085125582, "lr": 0.0026291804804649314, "grad_norm": 0.175945, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:39.054987+00:00", "epoch": 0, "step": 13485, "train_loss": 3.4798638820648193, "perplexity": 32.455304026213014, "lr": 0.0026291804804649314, "grad_norm": 0.166458, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:39.361266+00:00", "epoch": 0, "step": 13486, "train_loss": 3.5393364429473877, "perplexity": 34.444056009897864, "lr": 0.0026291804804649314, "grad_norm": 0.186735, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:39.666423+00:00", "epoch": 0, "step": 13487, "train_loss": 3.547366142272949, "perplexity": 34.721744809973515, "lr": 0.0026291804804649314, "grad_norm": 0.181201, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:39.970357+00:00", "epoch": 0, "step": 13488, "train_loss": 3.592796802520752, "perplexity": 36.33555732693323, "lr": 0.0026291804804649314, "grad_norm": 0.160811, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:40.275248+00:00", "epoch": 0, "step": 13489, "train_loss": 3.5348474979400635, "perplexity": 34.289785052616594, "lr": 0.0026291804804649314, "grad_norm": 0.175528, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:40.580648+00:00", "epoch": 0, "step": 13490, "train_loss": 3.519911766052246, "perplexity": 33.781447661859474, "lr": 0.0026291804804649314, "grad_norm": 0.208829, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:40.886110+00:00", "epoch": 0, "step": 13491, "train_loss": 3.5139734745025635, "perplexity": 33.58143802219929, "lr": 0.0026291804804649314, "grad_norm": 0.194885, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:41.191105+00:00", "epoch": 0, "step": 13492, "train_loss": 3.45727801322937, "perplexity": 31.730488888001553, "lr": 0.0026291804804649314, "grad_norm": 0.166044, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:41.496939+00:00", "epoch": 0, "step": 13493, "train_loss": 3.4299476146698, "perplexity": 30.87502530891658, "lr": 0.0026291804804649314, "grad_norm": 0.210982, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:41.802245+00:00", "epoch": 0, "step": 13494, "train_loss": 3.5412344932556152, "perplexity": 34.5094946442884, "lr": 0.0026291804804649314, "grad_norm": 0.183755, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:42.106047+00:00", "epoch": 0, "step": 13495, "train_loss": 3.6023032665252686, "perplexity": 36.68262708401895, "lr": 0.0026291804804649314, "grad_norm": 0.217493, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:42.410931+00:00", "epoch": 0, "step": 13496, "train_loss": 3.446580648422241, "perplexity": 31.392865333639772, "lr": 0.0026291804804649314, "grad_norm": 0.189744, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:42.716807+00:00", "epoch": 0, "step": 13497, "train_loss": 3.503636121749878, "perplexity": 33.23608295540277, "lr": 0.0026291804804649314, "grad_norm": 0.199603, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:43.023076+00:00", "epoch": 0, "step": 13498, "train_loss": 3.465545177459717, "perplexity": 31.993897371101173, "lr": 0.0026291804804649314, "grad_norm": 0.177255, "tokens_per_sec": 106991} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:43.328265+00:00", "epoch": 0, "step": 13499, "train_loss": 3.4884555339813232, "perplexity": 32.735350007380625, "lr": 0.0026291804804649314, "grad_norm": 0.18672, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:43.633302+00:00", "epoch": 0, "step": 13500, "train_loss": 3.5645437240600586, "perplexity": 35.32333254002156, "lr": 0.0026291804804649314, "grad_norm": 0.176802, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:08:46.706605+00:00", "step": 13500, "epoch": 0, "val_loss": 3.483438265323639, "val_ppl": 32.57151929681045, "eval_train_loss": 3.5645437240600586, "eval_train_ppl": 35.32333254002156} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:08:47.659567+00:00", "step": 13500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4834_epoch_0000_step_0013500.pt", "val_loss": 3.483438265323639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:49.141756+00:00", "epoch": 0, "step": 13501, "train_loss": 3.4921767711639404, "perplexity": 32.85739294353652, "lr": 0.0026291804804649314, "grad_norm": 0.207209, "tokens_per_sec": 5949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:49.445668+00:00", "epoch": 0, "step": 13502, "train_loss": 3.531193256378174, "perplexity": 34.16471056031474, "lr": 0.0026291804804649314, "grad_norm": 0.163199, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:49.749721+00:00", "epoch": 0, "step": 13503, "train_loss": 3.5297138690948486, "perplexity": 34.11420508976006, "lr": 0.0026291804804649314, "grad_norm": 0.181641, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:50.053916+00:00", "epoch": 0, "step": 13504, "train_loss": 3.491680860519409, "perplexity": 32.84110265222255, "lr": 0.0026291804804649314, "grad_norm": 0.164377, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:50.358465+00:00", "epoch": 0, "step": 13505, "train_loss": 3.4934005737304688, "perplexity": 32.89762852052973, "lr": 0.0026291804804649314, "grad_norm": 0.175893, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:50.663081+00:00", "epoch": 0, "step": 13506, "train_loss": 3.438843250274658, "perplexity": 31.150903520136513, "lr": 0.0026291804804649314, "grad_norm": 0.183723, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:50.967718+00:00", "epoch": 0, "step": 13507, "train_loss": 3.5291483402252197, "perplexity": 34.09491797614118, "lr": 0.0026291804804649314, "grad_norm": 0.163323, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:51.271885+00:00", "epoch": 0, "step": 13508, "train_loss": 3.514613151550293, "perplexity": 33.602926169336385, "lr": 0.0026291804804649314, "grad_norm": 0.17382, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:51.576225+00:00", "epoch": 0, "step": 13509, "train_loss": 3.4282383918762207, "perplexity": 30.82229808602848, "lr": 0.0026291804804649314, "grad_norm": 0.186095, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:51.880280+00:00", "epoch": 0, "step": 13510, "train_loss": 3.519021987915039, "perplexity": 33.751403036794215, "lr": 0.0026291804804649314, "grad_norm": 0.180415, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:52.184895+00:00", "epoch": 0, "step": 13511, "train_loss": 3.5805819034576416, "perplexity": 35.894421859296905, "lr": 0.0026291804804649314, "grad_norm": 0.176528, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:52.489674+00:00", "epoch": 0, "step": 13512, "train_loss": 3.4773004055023193, "perplexity": 32.372212162538105, "lr": 0.0026291804804649314, "grad_norm": 0.192507, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:52.796775+00:00", "epoch": 0, "step": 13513, "train_loss": 3.5416483879089355, "perplexity": 34.5237808959081, "lr": 0.0026291804804649314, "grad_norm": 0.165271, "tokens_per_sec": 106703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:53.102301+00:00", "epoch": 0, "step": 13514, "train_loss": 3.466488838195801, "perplexity": 32.0241030055387, "lr": 0.0026291804804649314, "grad_norm": 0.208521, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:53.406593+00:00", "epoch": 0, "step": 13515, "train_loss": 3.5310521125793457, "perplexity": 34.15988876357175, "lr": 0.0026291804804649314, "grad_norm": 0.156641, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:53.711088+00:00", "epoch": 0, "step": 13516, "train_loss": 3.511103630065918, "perplexity": 33.485202675403606, "lr": 0.0026291804804649314, "grad_norm": 0.20017, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:54.016977+00:00", "epoch": 0, "step": 13517, "train_loss": 3.509941577911377, "perplexity": 33.44631372336086, "lr": 0.0026291804804649314, "grad_norm": 0.184672, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:54.322030+00:00", "epoch": 0, "step": 13518, "train_loss": 3.5925135612487793, "perplexity": 36.32526705484178, "lr": 0.0026291804804649314, "grad_norm": 0.166647, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:54.625413+00:00", "epoch": 0, "step": 13519, "train_loss": 3.515059232711792, "perplexity": 33.61791914546525, "lr": 0.0026291804804649314, "grad_norm": 0.19397, "tokens_per_sec": 108007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:54.929118+00:00", "epoch": 0, "step": 13520, "train_loss": 3.529269218444824, "perplexity": 34.09903955822343, "lr": 0.0026291804804649314, "grad_norm": 0.16021, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:55.234025+00:00", "epoch": 0, "step": 13521, "train_loss": 3.45143723487854, "perplexity": 31.545698321153537, "lr": 0.0026291804804649314, "grad_norm": 0.178081, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:55.537915+00:00", "epoch": 0, "step": 13522, "train_loss": 3.4044570922851562, "perplexity": 30.09795087771519, "lr": 0.0026291804804649314, "grad_norm": 0.175254, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:55.843512+00:00", "epoch": 0, "step": 13523, "train_loss": 3.55637264251709, "perplexity": 35.035878713154375, "lr": 0.0026291804804649314, "grad_norm": 0.170751, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:56.148397+00:00", "epoch": 0, "step": 13524, "train_loss": 3.4297797679901123, "perplexity": 30.869843473322327, "lr": 0.0026291804804649314, "grad_norm": 0.156448, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:56.452212+00:00", "epoch": 0, "step": 13525, "train_loss": 3.502964735031128, "perplexity": 33.2137761797976, "lr": 0.0026291804804649314, "grad_norm": 0.188276, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:56.756208+00:00", "epoch": 0, "step": 13526, "train_loss": 3.4620885848999023, "perplexity": 31.883498415513838, "lr": 0.0026291804804649314, "grad_norm": 0.17822, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:57.061743+00:00", "epoch": 0, "step": 13527, "train_loss": 3.4830739498138428, "perplexity": 32.55965514843084, "lr": 0.0026291804804649314, "grad_norm": 0.15221, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:57.366338+00:00", "epoch": 0, "step": 13528, "train_loss": 3.4829959869384766, "perplexity": 32.55711680304418, "lr": 0.0026291804804649314, "grad_norm": 0.186242, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:57.671490+00:00", "epoch": 0, "step": 13529, "train_loss": 3.4195210933685303, "perplexity": 30.55477863048847, "lr": 0.0026291804804649314, "grad_norm": 0.161136, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:57.976927+00:00", "epoch": 0, "step": 13530, "train_loss": 3.5479583740234375, "perplexity": 34.74231422000999, "lr": 0.0026291804804649314, "grad_norm": 0.169467, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:58.292549+00:00", "epoch": 0, "step": 13531, "train_loss": 3.4215281009674072, "perplexity": 30.616163883110712, "lr": 0.0026291804804649314, "grad_norm": 0.162631, "tokens_per_sec": 103768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:58.598316+00:00", "epoch": 0, "step": 13532, "train_loss": 3.5207700729370117, "perplexity": 33.81045495776511, "lr": 0.0026291804804649314, "grad_norm": 0.166824, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:58.904210+00:00", "epoch": 0, "step": 13533, "train_loss": 3.467705726623535, "perplexity": 32.06309648644069, "lr": 0.0026291804804649314, "grad_norm": 0.185698, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:59.209425+00:00", "epoch": 0, "step": 13534, "train_loss": 3.523977041244507, "perplexity": 33.91905806556036, "lr": 0.0026291804804649314, "grad_norm": 0.197574, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:59.513467+00:00", "epoch": 0, "step": 13535, "train_loss": 3.5637624263763428, "perplexity": 35.29574528046125, "lr": 0.0026291804804649314, "grad_norm": 0.161809, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:08:59.818408+00:00", "epoch": 0, "step": 13536, "train_loss": 3.5269534587860107, "perplexity": 34.020165739662126, "lr": 0.0026291804804649314, "grad_norm": 0.175651, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:00.124311+00:00", "epoch": 0, "step": 13537, "train_loss": 3.5699894428253174, "perplexity": 35.51621819872967, "lr": 0.0026291804804649314, "grad_norm": 0.158732, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:00.430938+00:00", "epoch": 0, "step": 13538, "train_loss": 3.5869317054748535, "perplexity": 36.12306949698597, "lr": 0.0026291804804649314, "grad_norm": 0.168022, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:00.734769+00:00", "epoch": 0, "step": 13539, "train_loss": 3.5042476654052734, "perplexity": 33.256414487240356, "lr": 0.0026291804804649314, "grad_norm": 0.165361, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:01.038899+00:00", "epoch": 0, "step": 13540, "train_loss": 3.494239091873169, "perplexity": 32.92522534749908, "lr": 0.0026291804804649314, "grad_norm": 0.164498, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:01.343542+00:00", "epoch": 0, "step": 13541, "train_loss": 3.419830083847046, "perplexity": 30.5642212249192, "lr": 0.0026291804804649314, "grad_norm": 0.188796, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:01.648573+00:00", "epoch": 0, "step": 13542, "train_loss": 3.3867409229278564, "perplexity": 29.56942602417755, "lr": 0.0026291804804649314, "grad_norm": 0.177846, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:01.953113+00:00", "epoch": 0, "step": 13543, "train_loss": 3.412532091140747, "perplexity": 30.341975720791773, "lr": 0.0026291804804649314, "grad_norm": 0.180633, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:02.257577+00:00", "epoch": 0, "step": 13544, "train_loss": 3.6264989376068115, "perplexity": 37.581012554936564, "lr": 0.0026291804804649314, "grad_norm": 0.180765, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:02.561997+00:00", "epoch": 0, "step": 13545, "train_loss": 3.3510656356811523, "perplexity": 28.53312336305724, "lr": 0.0026291804804649314, "grad_norm": 0.175278, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:02.867308+00:00", "epoch": 0, "step": 13546, "train_loss": 3.5332107543945312, "perplexity": 34.23370737316299, "lr": 0.0026291804804649314, "grad_norm": 0.193848, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:03.171913+00:00", "epoch": 0, "step": 13547, "train_loss": 3.5527100563049316, "perplexity": 34.90779149527368, "lr": 0.0026291804804649314, "grad_norm": 0.174905, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:03.476690+00:00", "epoch": 0, "step": 13548, "train_loss": 3.444817304611206, "perplexity": 31.33755769636403, "lr": 0.0026291804804649314, "grad_norm": 0.161286, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:03.780667+00:00", "epoch": 0, "step": 13549, "train_loss": 3.4643194675445557, "perplexity": 31.95470615730506, "lr": 0.0026291804804649314, "grad_norm": 0.178569, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:04.085721+00:00", "epoch": 0, "step": 13550, "train_loss": 3.582104206085205, "perplexity": 35.949105644082444, "lr": 0.0026291804804649314, "grad_norm": 0.191105, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:04.391076+00:00", "epoch": 0, "step": 13551, "train_loss": 3.5722694396972656, "perplexity": 35.59728744882542, "lr": 0.0026291804804649314, "grad_norm": 0.169842, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:04.696652+00:00", "epoch": 0, "step": 13552, "train_loss": 3.5599615573883057, "perplexity": 35.16184540595519, "lr": 0.0026291804804649314, "grad_norm": 0.165266, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:05.000791+00:00", "epoch": 0, "step": 13553, "train_loss": 3.5056190490722656, "perplexity": 33.302053077747, "lr": 0.0026291804804649314, "grad_norm": 0.165873, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:05.304872+00:00", "epoch": 0, "step": 13554, "train_loss": 3.5354199409484863, "perplexity": 34.309419619629416, "lr": 0.0026291804804649314, "grad_norm": 0.160328, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:05.608972+00:00", "epoch": 0, "step": 13555, "train_loss": 3.5170555114746094, "perplexity": 33.685096913972416, "lr": 0.0026291804804649314, "grad_norm": 0.158099, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:05.914092+00:00", "epoch": 0, "step": 13556, "train_loss": 3.5010852813720703, "perplexity": 33.15141105116789, "lr": 0.0026291804804649314, "grad_norm": 0.140443, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:06.219537+00:00", "epoch": 0, "step": 13557, "train_loss": 3.565103769302368, "perplexity": 35.34312074498093, "lr": 0.0026291804804649314, "grad_norm": 0.168925, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:06.523882+00:00", "epoch": 0, "step": 13558, "train_loss": 3.5519142150878906, "perplexity": 34.88002148772988, "lr": 0.0026291804804649314, "grad_norm": 0.153557, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:06.827425+00:00", "epoch": 0, "step": 13559, "train_loss": 3.527540445327759, "perplexity": 34.04014098112475, "lr": 0.0026291804804649314, "grad_norm": 0.139468, "tokens_per_sec": 107951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:07.132824+00:00", "epoch": 0, "step": 13560, "train_loss": 3.5578134059906006, "perplexity": 35.08639350866455, "lr": 0.0026291804804649314, "grad_norm": 0.144879, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:07.437248+00:00", "epoch": 0, "step": 13561, "train_loss": 3.536287307739258, "perplexity": 34.339191380469956, "lr": 0.0026291804804649314, "grad_norm": 0.159512, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:07.741493+00:00", "epoch": 0, "step": 13562, "train_loss": 3.5660324096679688, "perplexity": 35.37595703774471, "lr": 0.0026291804804649314, "grad_norm": 0.183274, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:08.046924+00:00", "epoch": 0, "step": 13563, "train_loss": 3.495234966278076, "perplexity": 32.958031069176435, "lr": 0.0026291804804649314, "grad_norm": 0.181916, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:08.351135+00:00", "epoch": 0, "step": 13564, "train_loss": 3.500094413757324, "perplexity": 33.118578660536734, "lr": 0.0026291804804649314, "grad_norm": 0.152157, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:08.655986+00:00", "epoch": 0, "step": 13565, "train_loss": 3.47359561920166, "perplexity": 32.25250192174932, "lr": 0.0026291804804649314, "grad_norm": 0.156712, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:08.960732+00:00", "epoch": 0, "step": 13566, "train_loss": 3.487330675125122, "perplexity": 32.69854806139189, "lr": 0.0026291804804649314, "grad_norm": 0.163044, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:09.265859+00:00", "epoch": 0, "step": 13567, "train_loss": 3.5202763080596924, "perplexity": 33.793764663500426, "lr": 0.0026291804804649314, "grad_norm": 0.157833, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:09.572197+00:00", "epoch": 0, "step": 13568, "train_loss": 3.536717414855957, "perplexity": 34.35396408775993, "lr": 0.0026291804804649314, "grad_norm": 0.180211, "tokens_per_sec": 106967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:09.876968+00:00", "epoch": 0, "step": 13569, "train_loss": 3.4115426540374756, "perplexity": 30.311969091511244, "lr": 0.0026291804804649314, "grad_norm": 0.182337, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:10.181547+00:00", "epoch": 0, "step": 13570, "train_loss": 3.5187487602233887, "perplexity": 33.742182478565795, "lr": 0.0026291804804649314, "grad_norm": 0.172591, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:10.485928+00:00", "epoch": 0, "step": 13571, "train_loss": 3.4673709869384766, "perplexity": 32.052365491765904, "lr": 0.0026291804804649314, "grad_norm": 0.172079, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:10.790711+00:00", "epoch": 0, "step": 13572, "train_loss": 3.576961040496826, "perplexity": 35.76468809252326, "lr": 0.0026291804804649314, "grad_norm": 0.177217, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:11.094975+00:00", "epoch": 0, "step": 13573, "train_loss": 3.557239294052124, "perplexity": 35.066255772483636, "lr": 0.0026291804804649314, "grad_norm": 0.160705, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:11.398917+00:00", "epoch": 0, "step": 13574, "train_loss": 3.4564051628112793, "perplexity": 31.702805001213132, "lr": 0.0026291804804649314, "grad_norm": 0.177822, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:11.703163+00:00", "epoch": 0, "step": 13575, "train_loss": 3.5581610202789307, "perplexity": 35.09859216046413, "lr": 0.0026291804804649314, "grad_norm": 0.184606, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:12.008533+00:00", "epoch": 0, "step": 13576, "train_loss": 3.54069185256958, "perplexity": 34.49077346833516, "lr": 0.0026291804804649314, "grad_norm": 0.157848, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:12.312420+00:00", "epoch": 0, "step": 13577, "train_loss": 3.585834503173828, "perplexity": 36.08345691749444, "lr": 0.0026291804804649314, "grad_norm": 0.184795, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:12.617874+00:00", "epoch": 0, "step": 13578, "train_loss": 3.589531898498535, "perplexity": 36.21711867029927, "lr": 0.0026291804804649314, "grad_norm": 0.169712, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:12.923386+00:00", "epoch": 0, "step": 13579, "train_loss": 3.4592859745025635, "perplexity": 31.794266490917742, "lr": 0.0026291804804649314, "grad_norm": 0.166703, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:13.228237+00:00", "epoch": 0, "step": 13580, "train_loss": 3.5742130279541016, "perplexity": 35.66654119727157, "lr": 0.0026291804804649314, "grad_norm": 0.191092, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:13.531560+00:00", "epoch": 0, "step": 13581, "train_loss": 3.4840705394744873, "perplexity": 32.59211993845598, "lr": 0.0026291804804649314, "grad_norm": 0.172919, "tokens_per_sec": 107973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:13.836466+00:00", "epoch": 0, "step": 13582, "train_loss": 3.5006103515625, "perplexity": 33.135670196026915, "lr": 0.0026291804804649314, "grad_norm": 0.16202, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:14.142151+00:00", "epoch": 0, "step": 13583, "train_loss": 3.4961676597595215, "perplexity": 32.988785149751706, "lr": 0.0026291804804649314, "grad_norm": 0.150438, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:14.447700+00:00", "epoch": 0, "step": 13584, "train_loss": 3.4971261024475098, "perplexity": 33.02041816645934, "lr": 0.0026291804804649314, "grad_norm": 0.1772, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:14.751716+00:00", "epoch": 0, "step": 13585, "train_loss": 3.5483455657958984, "perplexity": 34.75576876280827, "lr": 0.0026291804804649314, "grad_norm": 0.16331, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:15.055748+00:00", "epoch": 0, "step": 13586, "train_loss": 3.643876791000366, "perplexity": 38.23979743626492, "lr": 0.0026291804804649314, "grad_norm": 0.192203, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:15.360685+00:00", "epoch": 0, "step": 13587, "train_loss": 3.4996676445007324, "perplexity": 33.10444768488986, "lr": 0.0026291804804649314, "grad_norm": 0.175619, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:15.665589+00:00", "epoch": 0, "step": 13588, "train_loss": 3.4512546062469482, "perplexity": 31.539937699479776, "lr": 0.0026291804804649314, "grad_norm": 0.188568, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:15.970573+00:00", "epoch": 0, "step": 13589, "train_loss": 3.5061557292938232, "perplexity": 33.31993042775208, "lr": 0.0026291804804649314, "grad_norm": 0.154007, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:16.274909+00:00", "epoch": 0, "step": 13590, "train_loss": 3.433236598968506, "perplexity": 30.97673995963353, "lr": 0.0026291804804649314, "grad_norm": 0.179751, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:16.579585+00:00", "epoch": 0, "step": 13591, "train_loss": 3.513350248336792, "perplexity": 33.56051571168251, "lr": 0.0026291804804649314, "grad_norm": 0.182578, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:16.883555+00:00", "epoch": 0, "step": 13592, "train_loss": 3.5711216926574707, "perplexity": 35.55645420513113, "lr": 0.0026291804804649314, "grad_norm": 0.182483, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:17.187677+00:00", "epoch": 0, "step": 13593, "train_loss": 3.4668564796447754, "perplexity": 32.03587855762765, "lr": 0.0026291804804649314, "grad_norm": 0.1636, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:17.492672+00:00", "epoch": 0, "step": 13594, "train_loss": 3.518867254257202, "perplexity": 33.746180962770936, "lr": 0.0026291804804649314, "grad_norm": 0.176884, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:17.797101+00:00", "epoch": 0, "step": 13595, "train_loss": 3.564544439315796, "perplexity": 35.32335780524686, "lr": 0.0026291804804649314, "grad_norm": 0.193891, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:18.101203+00:00", "epoch": 0, "step": 13596, "train_loss": 3.6126608848571777, "perplexity": 37.064546206660914, "lr": 0.0026291804804649314, "grad_norm": 0.159615, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:18.404782+00:00", "epoch": 0, "step": 13597, "train_loss": 3.5600802898406982, "perplexity": 35.16602050594591, "lr": 0.0026291804804649314, "grad_norm": 0.165562, "tokens_per_sec": 107939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:18.710150+00:00", "epoch": 0, "step": 13598, "train_loss": 3.5119800567626953, "perplexity": 33.514562865123025, "lr": 0.0026291804804649314, "grad_norm": 0.170953, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:19.015047+00:00", "epoch": 0, "step": 13599, "train_loss": 3.5039467811584473, "perplexity": 33.24640966123797, "lr": 0.0026291804804649314, "grad_norm": 0.152826, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:19.321626+00:00", "epoch": 0, "step": 13600, "train_loss": 3.481846570968628, "perplexity": 32.51971663135347, "lr": 0.0026291804804649314, "grad_norm": 0.144241, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:19.625625+00:00", "epoch": 0, "step": 13601, "train_loss": 3.601186752319336, "perplexity": 36.64169326561594, "lr": 0.0026291804804649314, "grad_norm": 0.163583, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:19.929744+00:00", "epoch": 0, "step": 13602, "train_loss": 3.4124350547790527, "perplexity": 30.339031588707453, "lr": 0.0026291804804649314, "grad_norm": 0.142947, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:20.234444+00:00", "epoch": 0, "step": 13603, "train_loss": 3.55420184135437, "perplexity": 34.959905278351876, "lr": 0.0026291804804649314, "grad_norm": 0.142346, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:20.539838+00:00", "epoch": 0, "step": 13604, "train_loss": 3.58764386177063, "perplexity": 36.14880393073288, "lr": 0.0026291804804649314, "grad_norm": 0.146082, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:20.846222+00:00", "epoch": 0, "step": 13605, "train_loss": 3.521833896636963, "perplexity": 33.846442459840134, "lr": 0.0026291804804649314, "grad_norm": 0.136182, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:21.152158+00:00", "epoch": 0, "step": 13606, "train_loss": 3.4844706058502197, "perplexity": 32.60516155834005, "lr": 0.0026291804804649314, "grad_norm": 0.143763, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:21.456635+00:00", "epoch": 0, "step": 13607, "train_loss": 3.580996513366699, "perplexity": 35.90930712785636, "lr": 0.0026291804804649314, "grad_norm": 0.138593, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:21.760732+00:00", "epoch": 0, "step": 13608, "train_loss": 3.543266534805298, "perplexity": 34.57969066772787, "lr": 0.0026291804804649314, "grad_norm": 0.175485, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:22.066431+00:00", "epoch": 0, "step": 13609, "train_loss": 3.4698681831359863, "perplexity": 32.13250655933117, "lr": 0.0026291804804649314, "grad_norm": 0.164753, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:22.372581+00:00", "epoch": 0, "step": 13610, "train_loss": 3.541583299636841, "perplexity": 34.521533875791526, "lr": 0.0026291804804649314, "grad_norm": 0.177028, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:22.678164+00:00", "epoch": 0, "step": 13611, "train_loss": 3.6359825134277344, "perplexity": 37.93911027649019, "lr": 0.0026291804804649314, "grad_norm": 0.197642, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:22.983453+00:00", "epoch": 0, "step": 13612, "train_loss": 3.4774489402770996, "perplexity": 32.377020918905735, "lr": 0.0026291804804649314, "grad_norm": 0.191363, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:23.287878+00:00", "epoch": 0, "step": 13613, "train_loss": 3.482865810394287, "perplexity": 32.55287890593337, "lr": 0.0026291804804649314, "grad_norm": 0.151467, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:23.593031+00:00", "epoch": 0, "step": 13614, "train_loss": 3.4912383556365967, "perplexity": 32.826573518785324, "lr": 0.0026291804804649314, "grad_norm": 0.177753, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:23.898173+00:00", "epoch": 0, "step": 13615, "train_loss": 3.538910388946533, "perplexity": 34.42938410776243, "lr": 0.0026291804804649314, "grad_norm": 0.159831, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:24.202911+00:00", "epoch": 0, "step": 13616, "train_loss": 3.479055643081665, "perplexity": 32.42908298215131, "lr": 0.0026291804804649314, "grad_norm": 0.170519, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:24.506861+00:00", "epoch": 0, "step": 13617, "train_loss": 3.5732150077819824, "perplexity": 35.63096302650615, "lr": 0.0026291804804649314, "grad_norm": 0.203663, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:24.811724+00:00", "epoch": 0, "step": 13618, "train_loss": 3.56911039352417, "perplexity": 35.485011410105066, "lr": 0.0026291804804649314, "grad_norm": 0.221376, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:25.115702+00:00", "epoch": 0, "step": 13619, "train_loss": 3.513662338256836, "perplexity": 33.57099124491674, "lr": 0.0026291804804649314, "grad_norm": 0.172166, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:25.421085+00:00", "epoch": 0, "step": 13620, "train_loss": 3.514383316040039, "perplexity": 33.595203911112804, "lr": 0.0026291804804649314, "grad_norm": 0.193617, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:25.726064+00:00", "epoch": 0, "step": 13621, "train_loss": 3.5741584300994873, "perplexity": 35.66459393379938, "lr": 0.0026291804804649314, "grad_norm": 0.196534, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:26.030872+00:00", "epoch": 0, "step": 13622, "train_loss": 3.546891212463379, "perplexity": 34.7052583335924, "lr": 0.0026291804804649314, "grad_norm": 0.153888, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:26.335228+00:00", "epoch": 0, "step": 13623, "train_loss": 3.451389789581299, "perplexity": 31.544201661625017, "lr": 0.0026291804804649314, "grad_norm": 0.167542, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:26.639476+00:00", "epoch": 0, "step": 13624, "train_loss": 3.3593690395355225, "perplexity": 28.771031767436835, "lr": 0.0026291804804649314, "grad_norm": 0.153454, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:26.944075+00:00", "epoch": 0, "step": 13625, "train_loss": 3.533604145050049, "perplexity": 34.24717724303408, "lr": 0.0026291804804649314, "grad_norm": 0.170259, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:27.249768+00:00", "epoch": 0, "step": 13626, "train_loss": 3.502814769744873, "perplexity": 33.20879563980858, "lr": 0.0026291804804649314, "grad_norm": 0.163206, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:27.555444+00:00", "epoch": 0, "step": 13627, "train_loss": 3.4844281673431396, "perplexity": 32.603777873321384, "lr": 0.0026291804804649314, "grad_norm": 0.158124, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:27.859419+00:00", "epoch": 0, "step": 13628, "train_loss": 3.479266405105591, "perplexity": 32.435918521625446, "lr": 0.0026291804804649314, "grad_norm": 0.160508, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:28.163320+00:00", "epoch": 0, "step": 13629, "train_loss": 3.480072498321533, "perplexity": 32.46207543653805, "lr": 0.0026291804804649314, "grad_norm": 0.164663, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:28.468880+00:00", "epoch": 0, "step": 13630, "train_loss": 3.516709327697754, "perplexity": 33.673437698131124, "lr": 0.0026291804804649314, "grad_norm": 0.158378, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:28.773434+00:00", "epoch": 0, "step": 13631, "train_loss": 3.482954978942871, "perplexity": 32.55578172831595, "lr": 0.0026291804804649314, "grad_norm": 0.153585, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:29.077827+00:00", "epoch": 0, "step": 13632, "train_loss": 3.501739978790283, "perplexity": 33.173122300767226, "lr": 0.0026291804804649314, "grad_norm": 0.146023, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:29.382359+00:00", "epoch": 0, "step": 13633, "train_loss": 3.4822418689727783, "perplexity": 32.532574151541986, "lr": 0.0026291804804649314, "grad_norm": 0.141454, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:29.686898+00:00", "epoch": 0, "step": 13634, "train_loss": 3.544804334640503, "perplexity": 34.63290821882126, "lr": 0.0026291804804649314, "grad_norm": 0.14449, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:29.990485+00:00", "epoch": 0, "step": 13635, "train_loss": 3.4739694595336914, "perplexity": 32.26456146180737, "lr": 0.0026291804804649314, "grad_norm": 0.146078, "tokens_per_sec": 107937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:30.295560+00:00", "epoch": 0, "step": 13636, "train_loss": 3.6232945919036865, "perplexity": 37.4607827306724, "lr": 0.0026291804804649314, "grad_norm": 0.140418, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:30.600336+00:00", "epoch": 0, "step": 13637, "train_loss": 3.4803285598754883, "perplexity": 32.47038879033866, "lr": 0.0026291804804649314, "grad_norm": 0.140758, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:30.905225+00:00", "epoch": 0, "step": 13638, "train_loss": 3.536533832550049, "perplexity": 34.34765788668678, "lr": 0.0026291804804649314, "grad_norm": 0.143024, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:31.210010+00:00", "epoch": 0, "step": 13639, "train_loss": 3.576688289642334, "perplexity": 35.75493457348617, "lr": 0.0026291804804649314, "grad_norm": 0.148693, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:31.514462+00:00", "epoch": 0, "step": 13640, "train_loss": 3.5271003246307373, "perplexity": 34.02516250695946, "lr": 0.0026291804804649314, "grad_norm": 0.159446, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:31.819258+00:00", "epoch": 0, "step": 13641, "train_loss": 3.476677894592285, "perplexity": 32.35206637842076, "lr": 0.0026291804804649314, "grad_norm": 0.172424, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:32.125520+00:00", "epoch": 0, "step": 13642, "train_loss": 3.488590717315674, "perplexity": 32.739775580270866, "lr": 0.0026291804804649314, "grad_norm": 0.197556, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:32.430146+00:00", "epoch": 0, "step": 13643, "train_loss": 3.6507058143615723, "perplexity": 38.50183160667883, "lr": 0.0026291804804649314, "grad_norm": 0.189026, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:32.734217+00:00", "epoch": 0, "step": 13644, "train_loss": 3.4799017906188965, "perplexity": 32.45653438318116, "lr": 0.0026291804804649314, "grad_norm": 0.186591, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:33.038839+00:00", "epoch": 0, "step": 13645, "train_loss": 3.487896680831909, "perplexity": 32.71706086488027, "lr": 0.0026291804804649314, "grad_norm": 0.230371, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:33.343220+00:00", "epoch": 0, "step": 13646, "train_loss": 3.6051673889160156, "perplexity": 36.787841218756924, "lr": 0.0026291804804649314, "grad_norm": 0.210341, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:33.648906+00:00", "epoch": 0, "step": 13647, "train_loss": 3.443406343460083, "perplexity": 31.29337279880066, "lr": 0.0026291804804649314, "grad_norm": 0.166294, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:33.954722+00:00", "epoch": 0, "step": 13648, "train_loss": 3.5566794872283936, "perplexity": 35.0466309367905, "lr": 0.0026291804804649314, "grad_norm": 0.185887, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:34.259408+00:00", "epoch": 0, "step": 13649, "train_loss": 3.508972644805908, "perplexity": 33.41392217789295, "lr": 0.0026291804804649314, "grad_norm": 0.163643, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:34.563745+00:00", "epoch": 0, "step": 13650, "train_loss": 3.435196876525879, "perplexity": 31.037522523661018, "lr": 0.0026291804804649314, "grad_norm": 0.162782, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:34.867499+00:00", "epoch": 0, "step": 13651, "train_loss": 3.4353413581848145, "perplexity": 31.04200720037345, "lr": 0.0026291804804649314, "grad_norm": 0.162675, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:35.172573+00:00", "epoch": 0, "step": 13652, "train_loss": 3.5279996395111084, "perplexity": 34.05577560525319, "lr": 0.0026291804804649314, "grad_norm": 0.174139, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:35.476832+00:00", "epoch": 0, "step": 13653, "train_loss": 3.665004014968872, "perplexity": 39.056292972585105, "lr": 0.0026291804804649314, "grad_norm": 0.167934, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:35.782142+00:00", "epoch": 0, "step": 13654, "train_loss": 3.5010554790496826, "perplexity": 33.150423076850174, "lr": 0.0026291804804649314, "grad_norm": 0.169668, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:36.086350+00:00", "epoch": 0, "step": 13655, "train_loss": 3.5528926849365234, "perplexity": 34.914167239645366, "lr": 0.0026291804804649314, "grad_norm": 0.171274, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:36.389852+00:00", "epoch": 0, "step": 13656, "train_loss": 3.4723167419433594, "perplexity": 32.21128129419799, "lr": 0.0026291804804649314, "grad_norm": 0.1809, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:36.694837+00:00", "epoch": 0, "step": 13657, "train_loss": 3.6015255451202393, "perplexity": 36.65410931062181, "lr": 0.0026291804804649314, "grad_norm": 0.183307, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:36.999715+00:00", "epoch": 0, "step": 13658, "train_loss": 3.486863136291504, "perplexity": 32.683263793643036, "lr": 0.0026291804804649314, "grad_norm": 0.183777, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:37.304938+00:00", "epoch": 0, "step": 13659, "train_loss": 3.5151467323303223, "perplexity": 33.62086082926249, "lr": 0.0026291804804649314, "grad_norm": 0.17547, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:37.609746+00:00", "epoch": 0, "step": 13660, "train_loss": 3.577040433883667, "perplexity": 35.767527684961166, "lr": 0.0026291804804649314, "grad_norm": 0.158743, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:37.914871+00:00", "epoch": 0, "step": 13661, "train_loss": 3.579191207885742, "perplexity": 35.84453834019556, "lr": 0.0026291804804649314, "grad_norm": 0.18167, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:38.219292+00:00", "epoch": 0, "step": 13662, "train_loss": 3.5620548725128174, "perplexity": 35.23552732157742, "lr": 0.0026291804804649314, "grad_norm": 0.158256, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:38.524192+00:00", "epoch": 0, "step": 13663, "train_loss": 3.415837526321411, "perplexity": 30.442435094296417, "lr": 0.0026291804804649314, "grad_norm": 0.162824, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:38.830285+00:00", "epoch": 0, "step": 13664, "train_loss": 3.4843087196350098, "perplexity": 32.59988365936013, "lr": 0.0026291804804649314, "grad_norm": 0.188999, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:39.135377+00:00", "epoch": 0, "step": 13665, "train_loss": 3.4384429454803467, "perplexity": 31.13843615964884, "lr": 0.0026291804804649314, "grad_norm": 0.186167, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:39.439900+00:00", "epoch": 0, "step": 13666, "train_loss": 3.4708378314971924, "perplexity": 32.16367890232112, "lr": 0.0026291804804649314, "grad_norm": 0.154072, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:39.743986+00:00", "epoch": 0, "step": 13667, "train_loss": 3.4610023498535156, "perplexity": 31.848884245099452, "lr": 0.0026291804804649314, "grad_norm": 0.16687, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:40.048319+00:00", "epoch": 0, "step": 13668, "train_loss": 3.519779682159424, "perplexity": 33.77698597141234, "lr": 0.0026291804804649314, "grad_norm": 0.167891, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:40.354845+00:00", "epoch": 0, "step": 13669, "train_loss": 3.3850226402282715, "perplexity": 29.518661017809926, "lr": 0.0026291804804649314, "grad_norm": 0.166836, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:40.659465+00:00", "epoch": 0, "step": 13670, "train_loss": 3.5044124126434326, "perplexity": 33.26189384102001, "lr": 0.0026291804804649314, "grad_norm": 0.165788, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:40.963781+00:00", "epoch": 0, "step": 13671, "train_loss": 3.5238311290740967, "perplexity": 33.914109223236686, "lr": 0.0026291804804649314, "grad_norm": 0.148757, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:41.268512+00:00", "epoch": 0, "step": 13672, "train_loss": 3.585566282272339, "perplexity": 36.07377987800132, "lr": 0.0026291804804649314, "grad_norm": 0.150866, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:41.572922+00:00", "epoch": 0, "step": 13673, "train_loss": 3.4795873165130615, "perplexity": 32.44632924826195, "lr": 0.0026291804804649314, "grad_norm": 0.168197, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:41.878967+00:00", "epoch": 0, "step": 13674, "train_loss": 3.4992167949676514, "perplexity": 33.089525924095426, "lr": 0.0026291804804649314, "grad_norm": 0.171256, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:42.184498+00:00", "epoch": 0, "step": 13675, "train_loss": 3.497793674468994, "perplexity": 33.04246903321692, "lr": 0.0026291804804649314, "grad_norm": 0.175461, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:42.488690+00:00", "epoch": 0, "step": 13676, "train_loss": 3.4573588371276855, "perplexity": 31.733053573451475, "lr": 0.0026291804804649314, "grad_norm": 0.153879, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:42.793991+00:00", "epoch": 0, "step": 13677, "train_loss": 3.4078330993652344, "perplexity": 30.199733485705192, "lr": 0.0026291804804649314, "grad_norm": 0.155823, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:43.099441+00:00", "epoch": 0, "step": 13678, "train_loss": 3.53340744972229, "perplexity": 34.24044164573342, "lr": 0.0026291804804649314, "grad_norm": 0.15756, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:43.405065+00:00", "epoch": 0, "step": 13679, "train_loss": 3.5591132640838623, "perplexity": 35.13203049560776, "lr": 0.0026291804804649314, "grad_norm": 0.166686, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:43.710176+00:00", "epoch": 0, "step": 13680, "train_loss": 3.5292088985443115, "perplexity": 34.096982769582965, "lr": 0.0026291804804649314, "grad_norm": 0.1516, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:44.014910+00:00", "epoch": 0, "step": 13681, "train_loss": 3.5367798805236816, "perplexity": 34.35611009809094, "lr": 0.0026291804804649314, "grad_norm": 0.164356, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:44.320402+00:00", "epoch": 0, "step": 13682, "train_loss": 3.4813616275787354, "perplexity": 32.50395023294624, "lr": 0.0026291804804649314, "grad_norm": 0.185517, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:44.625726+00:00", "epoch": 0, "step": 13683, "train_loss": 3.6880555152893066, "perplexity": 39.9670560207903, "lr": 0.0026291804804649314, "grad_norm": 0.175724, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:44.930196+00:00", "epoch": 0, "step": 13684, "train_loss": 3.480455160140991, "perplexity": 32.474499810403124, "lr": 0.0026291804804649314, "grad_norm": 0.17219, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:45.235395+00:00", "epoch": 0, "step": 13685, "train_loss": 3.513401985168457, "perplexity": 33.56225207135096, "lr": 0.0026291804804649314, "grad_norm": 0.163661, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:45.541126+00:00", "epoch": 0, "step": 13686, "train_loss": 3.554844379425049, "perplexity": 34.982375566675316, "lr": 0.0026291804804649314, "grad_norm": 0.186072, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:45.845120+00:00", "epoch": 0, "step": 13687, "train_loss": 3.4763240814208984, "perplexity": 32.34062181595235, "lr": 0.0026291804804649314, "grad_norm": 0.166635, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:46.149487+00:00", "epoch": 0, "step": 13688, "train_loss": 3.5486602783203125, "perplexity": 34.766708559888855, "lr": 0.0026291804804649314, "grad_norm": 0.187851, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:46.454424+00:00", "epoch": 0, "step": 13689, "train_loss": 3.5352673530578613, "perplexity": 34.30418481705499, "lr": 0.0026291804804649314, "grad_norm": 0.180439, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:46.760119+00:00", "epoch": 0, "step": 13690, "train_loss": 3.452090263366699, "perplexity": 31.566305288572202, "lr": 0.0026291804804649314, "grad_norm": 0.174381, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:47.066519+00:00", "epoch": 0, "step": 13691, "train_loss": 3.539868116378784, "perplexity": 34.462373868475716, "lr": 0.0026291804804649314, "grad_norm": 0.180049, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:47.372271+00:00", "epoch": 0, "step": 13692, "train_loss": 3.5462584495544434, "perplexity": 34.68330507970876, "lr": 0.0026291804804649314, "grad_norm": 0.157214, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:47.678054+00:00", "epoch": 0, "step": 13693, "train_loss": 3.4537107944488525, "perplexity": 31.617500938308076, "lr": 0.0026291804804649314, "grad_norm": 0.187575, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:47.982370+00:00", "epoch": 0, "step": 13694, "train_loss": 3.453615188598633, "perplexity": 31.614478264743987, "lr": 0.0026291804804649314, "grad_norm": 0.156517, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:48.288414+00:00", "epoch": 0, "step": 13695, "train_loss": 3.4613616466522217, "perplexity": 31.860329503250256, "lr": 0.0026291804804649314, "grad_norm": 0.169106, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:48.593087+00:00", "epoch": 0, "step": 13696, "train_loss": 3.5155417919158936, "perplexity": 33.634145696592554, "lr": 0.0026291804804649314, "grad_norm": 0.161999, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:48.898571+00:00", "epoch": 0, "step": 13697, "train_loss": 3.569136619567871, "perplexity": 35.48594205376854, "lr": 0.0026291804804649314, "grad_norm": 0.154944, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:49.203407+00:00", "epoch": 0, "step": 13698, "train_loss": 3.5512173175811768, "perplexity": 34.85572215577633, "lr": 0.0026291804804649314, "grad_norm": 0.151661, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:49.507217+00:00", "epoch": 0, "step": 13699, "train_loss": 3.5920469760894775, "perplexity": 36.30832217774782, "lr": 0.0026291804804649314, "grad_norm": 0.141631, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:49.812490+00:00", "epoch": 0, "step": 13700, "train_loss": 3.6451261043548584, "perplexity": 38.28760078033564, "lr": 0.0026291804804649314, "grad_norm": 0.162914, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:50.116737+00:00", "epoch": 0, "step": 13701, "train_loss": 3.5015578269958496, "perplexity": 33.16708030730986, "lr": 0.0026291804804649314, "grad_norm": 0.16154, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:50.422902+00:00", "epoch": 0, "step": 13702, "train_loss": 3.584770679473877, "perplexity": 36.04509089181144, "lr": 0.0026291804804649314, "grad_norm": 0.154992, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:50.729058+00:00", "epoch": 0, "step": 13703, "train_loss": 3.396268844604492, "perplexity": 29.85250764758725, "lr": 0.0026291804804649314, "grad_norm": 0.158103, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:51.033787+00:00", "epoch": 0, "step": 13704, "train_loss": 3.4503936767578125, "perplexity": 31.5127957223695, "lr": 0.0026291804804649314, "grad_norm": 0.175825, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:51.338215+00:00", "epoch": 0, "step": 13705, "train_loss": 3.492241144180298, "perplexity": 32.85950814111003, "lr": 0.0026291804804649314, "grad_norm": 0.178927, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:51.643375+00:00", "epoch": 0, "step": 13706, "train_loss": 3.4654810428619385, "perplexity": 31.991845521159906, "lr": 0.0026291804804649314, "grad_norm": 0.16982, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:51.948460+00:00", "epoch": 0, "step": 13707, "train_loss": 3.493337392807007, "perplexity": 32.89555008363956, "lr": 0.0026291804804649314, "grad_norm": 0.159267, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:52.253837+00:00", "epoch": 0, "step": 13708, "train_loss": 3.4599413871765137, "perplexity": 31.815111686491324, "lr": 0.0026291804804649314, "grad_norm": 0.159564, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:52.558819+00:00", "epoch": 0, "step": 13709, "train_loss": 3.411379337310791, "perplexity": 30.30701904416362, "lr": 0.0026291804804649314, "grad_norm": 0.166151, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:52.863655+00:00", "epoch": 0, "step": 13710, "train_loss": 3.4535927772521973, "perplexity": 31.613769749658633, "lr": 0.0026291804804649314, "grad_norm": 0.18222, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:53.168487+00:00", "epoch": 0, "step": 13711, "train_loss": 3.5005271434783936, "perplexity": 33.132913155099956, "lr": 0.0026291804804649314, "grad_norm": 0.159948, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:53.474071+00:00", "epoch": 0, "step": 13712, "train_loss": 3.542520523071289, "perplexity": 34.553903432717256, "lr": 0.0026291804804649314, "grad_norm": 0.177867, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:53.778733+00:00", "epoch": 0, "step": 13713, "train_loss": 3.553802728652954, "perplexity": 34.94595512014283, "lr": 0.0026291804804649314, "grad_norm": 0.18924, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:54.082817+00:00", "epoch": 0, "step": 13714, "train_loss": 3.5789356231689453, "perplexity": 35.83537819466163, "lr": 0.0026291804804649314, "grad_norm": 0.192502, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:54.387628+00:00", "epoch": 0, "step": 13715, "train_loss": 3.6114044189453125, "perplexity": 37.018005112585485, "lr": 0.0026291804804649314, "grad_norm": 0.186575, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:54.692995+00:00", "epoch": 0, "step": 13716, "train_loss": 3.6162519454956055, "perplexity": 37.19788651289425, "lr": 0.0026291804804649314, "grad_norm": 0.186176, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:54.998031+00:00", "epoch": 0, "step": 13717, "train_loss": 3.5112648010253906, "perplexity": 33.490599952577426, "lr": 0.0026291804804649314, "grad_norm": 0.190671, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:55.303187+00:00", "epoch": 0, "step": 13718, "train_loss": 3.54321026802063, "perplexity": 34.57774503445695, "lr": 0.0026291804804649314, "grad_norm": 0.205368, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:55.607869+00:00", "epoch": 0, "step": 13719, "train_loss": 3.4874367713928223, "perplexity": 32.70201743934119, "lr": 0.0026291804804649314, "grad_norm": 0.166431, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:55.913417+00:00", "epoch": 0, "step": 13720, "train_loss": 3.5182018280029297, "perplexity": 33.72373283759112, "lr": 0.0026291804804649314, "grad_norm": 0.186254, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:56.217898+00:00", "epoch": 0, "step": 13721, "train_loss": 3.4380857944488525, "perplexity": 31.127317020779458, "lr": 0.0026291804804649314, "grad_norm": 0.168731, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:56.523138+00:00", "epoch": 0, "step": 13722, "train_loss": 3.5074589252471924, "perplexity": 33.36338113249393, "lr": 0.0026291804804649314, "grad_norm": 0.180818, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:56.829642+00:00", "epoch": 0, "step": 13723, "train_loss": 3.5096662044525146, "perplexity": 33.43710476427431, "lr": 0.0026291804804649314, "grad_norm": 0.18099, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:57.135519+00:00", "epoch": 0, "step": 13724, "train_loss": 3.626113176345825, "perplexity": 37.56651805203283, "lr": 0.0026291804804649314, "grad_norm": 0.175702, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:57.439773+00:00", "epoch": 0, "step": 13725, "train_loss": 3.5354156494140625, "perplexity": 34.309272379890004, "lr": 0.0026291804804649314, "grad_norm": 0.179586, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:57.743980+00:00", "epoch": 0, "step": 13726, "train_loss": 3.495739698410034, "perplexity": 32.97467024527329, "lr": 0.0026291804804649314, "grad_norm": 0.187465, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:58.048746+00:00", "epoch": 0, "step": 13727, "train_loss": 3.4741690158843994, "perplexity": 32.27100070242419, "lr": 0.0026291804804649314, "grad_norm": 0.176741, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:58.352979+00:00", "epoch": 0, "step": 13728, "train_loss": 3.544586658477783, "perplexity": 34.62537028070085, "lr": 0.0026291804804649314, "grad_norm": 0.179245, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:58.658764+00:00", "epoch": 0, "step": 13729, "train_loss": 3.525970220565796, "perplexity": 33.98673225168396, "lr": 0.0026291804804649314, "grad_norm": 0.156332, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:58.963476+00:00", "epoch": 0, "step": 13730, "train_loss": 3.6022794246673584, "perplexity": 36.68175251246198, "lr": 0.0026291804804649314, "grad_norm": 0.171414, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:59.267454+00:00", "epoch": 0, "step": 13731, "train_loss": 3.578164577484131, "perplexity": 35.807758130474966, "lr": 0.0026291804804649314, "grad_norm": 0.169392, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:59.572922+00:00", "epoch": 0, "step": 13732, "train_loss": 3.499547243118286, "perplexity": 33.100462103563046, "lr": 0.0026291804804649314, "grad_norm": 0.14525, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:09:59.877700+00:00", "epoch": 0, "step": 13733, "train_loss": 3.4375603199005127, "perplexity": 31.110964704666166, "lr": 0.0026291804804649314, "grad_norm": 0.188493, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:00.182553+00:00", "epoch": 0, "step": 13734, "train_loss": 3.433567523956299, "perplexity": 30.986992633265924, "lr": 0.0026291804804649314, "grad_norm": 0.187623, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:00.488055+00:00", "epoch": 0, "step": 13735, "train_loss": 3.500232219696045, "perplexity": 33.12314291184139, "lr": 0.0026291804804649314, "grad_norm": 0.149946, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:00.792725+00:00", "epoch": 0, "step": 13736, "train_loss": 3.4817025661468506, "perplexity": 32.51503397252654, "lr": 0.0026291804804649314, "grad_norm": 0.196154, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:01.096720+00:00", "epoch": 0, "step": 13737, "train_loss": 3.5875205993652344, "perplexity": 36.144348416812534, "lr": 0.0026291804804649314, "grad_norm": 0.155704, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:01.401768+00:00", "epoch": 0, "step": 13738, "train_loss": 3.5126535892486572, "perplexity": 33.53714361554634, "lr": 0.0026291804804649314, "grad_norm": 0.174122, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:01.706339+00:00", "epoch": 0, "step": 13739, "train_loss": 3.5237250328063965, "perplexity": 33.91051125369467, "lr": 0.0026291804804649314, "grad_norm": 0.188087, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:02.011796+00:00", "epoch": 0, "step": 13740, "train_loss": 3.5153229236602783, "perplexity": 33.62678505532957, "lr": 0.0026291804804649314, "grad_norm": 0.141821, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:02.316501+00:00", "epoch": 0, "step": 13741, "train_loss": 3.497986316680908, "perplexity": 33.048835020697844, "lr": 0.0026291804804649314, "grad_norm": 0.186102, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:02.620741+00:00", "epoch": 0, "step": 13742, "train_loss": 3.3606173992156982, "perplexity": 28.806970791199177, "lr": 0.0026291804804649314, "grad_norm": 0.165969, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:02.925593+00:00", "epoch": 0, "step": 13743, "train_loss": 3.4895901679992676, "perplexity": 32.77251372871251, "lr": 0.0026291804804649314, "grad_norm": 0.167259, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:03.230852+00:00", "epoch": 0, "step": 13744, "train_loss": 3.4794583320617676, "perplexity": 32.44214444618036, "lr": 0.0026291804804649314, "grad_norm": 0.173397, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:03.536209+00:00", "epoch": 0, "step": 13745, "train_loss": 3.5038013458251953, "perplexity": 33.24157481015707, "lr": 0.0026291804804649314, "grad_norm": 0.186259, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:03.840561+00:00", "epoch": 0, "step": 13746, "train_loss": 3.405367136001587, "perplexity": 30.1253537958244, "lr": 0.0026291804804649314, "grad_norm": 0.176171, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:04.145166+00:00", "epoch": 0, "step": 13747, "train_loss": 3.4819555282592773, "perplexity": 32.52326008460933, "lr": 0.0026291804804649314, "grad_norm": 0.175115, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:04.450065+00:00", "epoch": 0, "step": 13748, "train_loss": 3.5681161880493164, "perplexity": 35.449749549158035, "lr": 0.0026291804804649314, "grad_norm": 0.176184, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:04.754815+00:00", "epoch": 0, "step": 13749, "train_loss": 3.5503780841827393, "perplexity": 34.82648234084733, "lr": 0.0026291804804649314, "grad_norm": 0.183166, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:05.059412+00:00", "epoch": 0, "step": 13750, "train_loss": 3.564732551574707, "perplexity": 35.33000318689521, "lr": 0.0026291804804649314, "grad_norm": 0.185249, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:05.364805+00:00", "epoch": 0, "step": 13751, "train_loss": 3.6919407844543457, "perplexity": 40.122640839898594, "lr": 0.0026291804804649314, "grad_norm": 0.180035, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:05.670522+00:00", "epoch": 0, "step": 13752, "train_loss": 3.450031042098999, "perplexity": 31.501370162221722, "lr": 0.0026291804804649314, "grad_norm": 0.181384, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:05.975990+00:00", "epoch": 0, "step": 13753, "train_loss": 3.539185047149658, "perplexity": 34.43884171928201, "lr": 0.0026291804804649314, "grad_norm": 0.177552, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:06.281512+00:00", "epoch": 0, "step": 13754, "train_loss": 3.5309267044067383, "perplexity": 34.15560510295408, "lr": 0.0026291804804649314, "grad_norm": 0.198755, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:06.586636+00:00", "epoch": 0, "step": 13755, "train_loss": 3.515784978866577, "perplexity": 33.64232607656387, "lr": 0.0026291804804649314, "grad_norm": 0.23283, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:06.890683+00:00", "epoch": 0, "step": 13756, "train_loss": 3.4824953079223633, "perplexity": 32.5408202178553, "lr": 0.0026291804804649314, "grad_norm": 0.191586, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:07.195222+00:00", "epoch": 0, "step": 13757, "train_loss": 3.417017936706543, "perplexity": 30.478390877945667, "lr": 0.0026291804804649314, "grad_norm": 0.172992, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:07.501196+00:00", "epoch": 0, "step": 13758, "train_loss": 3.5477302074432373, "perplexity": 34.734388089259625, "lr": 0.0026291804804649314, "grad_norm": 0.155589, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:07.806105+00:00", "epoch": 0, "step": 13759, "train_loss": 3.438974142074585, "perplexity": 31.154981184828223, "lr": 0.0026291804804649314, "grad_norm": 0.175423, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:08.110922+00:00", "epoch": 0, "step": 13760, "train_loss": 3.534275531768799, "perplexity": 34.27017806334838, "lr": 0.0026291804804649314, "grad_norm": 0.157099, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:08.415529+00:00", "epoch": 0, "step": 13761, "train_loss": 3.501967430114746, "perplexity": 33.18066842952707, "lr": 0.0026291804804649314, "grad_norm": 0.160321, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:08.720448+00:00", "epoch": 0, "step": 13762, "train_loss": 3.520188808441162, "perplexity": 33.79080785134552, "lr": 0.0026291804804649314, "grad_norm": 0.186551, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:09.026265+00:00", "epoch": 0, "step": 13763, "train_loss": 3.5631778240203857, "perplexity": 35.27511733477534, "lr": 0.0026291804804649314, "grad_norm": 0.168962, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:09.331739+00:00", "epoch": 0, "step": 13764, "train_loss": 3.5482003688812256, "perplexity": 34.75072269876216, "lr": 0.0026291804804649314, "grad_norm": 0.155529, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:09.636636+00:00", "epoch": 0, "step": 13765, "train_loss": 3.5344536304473877, "perplexity": 34.276282080319014, "lr": 0.0026291804804649314, "grad_norm": 0.159584, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:09.942453+00:00", "epoch": 0, "step": 13766, "train_loss": 3.482713222503662, "perplexity": 32.547912109753455, "lr": 0.0026291804804649314, "grad_norm": 0.158192, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:10.247781+00:00", "epoch": 0, "step": 13767, "train_loss": 3.5951709747314453, "perplexity": 36.42192668469362, "lr": 0.0026291804804649314, "grad_norm": 0.17801, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:10.551973+00:00", "epoch": 0, "step": 13768, "train_loss": 3.5942330360412598, "perplexity": 36.38778116618962, "lr": 0.0026291804804649314, "grad_norm": 0.164991, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:10.857241+00:00", "epoch": 0, "step": 13769, "train_loss": 3.4942948818206787, "perplexity": 32.927062295334096, "lr": 0.0026291804804649314, "grad_norm": 0.15685, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:11.162195+00:00", "epoch": 0, "step": 13770, "train_loss": 3.525059223175049, "perplexity": 33.95578452607163, "lr": 0.0026291804804649314, "grad_norm": 0.168363, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:11.467182+00:00", "epoch": 0, "step": 13771, "train_loss": 3.485663890838623, "perplexity": 32.644092031126945, "lr": 0.0026291804804649314, "grad_norm": 0.157409, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:11.772499+00:00", "epoch": 0, "step": 13772, "train_loss": 3.546157121658325, "perplexity": 34.679790871421275, "lr": 0.0026291804804649314, "grad_norm": 0.162693, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:12.077054+00:00", "epoch": 0, "step": 13773, "train_loss": 3.5948092937469482, "perplexity": 36.408755948339035, "lr": 0.0026291804804649314, "grad_norm": 0.168137, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:12.382131+00:00", "epoch": 0, "step": 13774, "train_loss": 3.5577831268310547, "perplexity": 35.08533113824153, "lr": 0.0026291804804649314, "grad_norm": 0.186254, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:12.687647+00:00", "epoch": 0, "step": 13775, "train_loss": 3.466813087463379, "perplexity": 32.03448848113352, "lr": 0.0026291804804649314, "grad_norm": 0.163333, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:12.993202+00:00", "epoch": 0, "step": 13776, "train_loss": 3.4388012886047363, "perplexity": 31.149596403629822, "lr": 0.0026291804804649314, "grad_norm": 0.16631, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:13.298529+00:00", "epoch": 0, "step": 13777, "train_loss": 3.5465691089630127, "perplexity": 34.694081448555565, "lr": 0.0026291804804649314, "grad_norm": 0.173052, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:13.602505+00:00", "epoch": 0, "step": 13778, "train_loss": 3.45824933052063, "perplexity": 31.76132423354643, "lr": 0.0026291804804649314, "grad_norm": 0.163917, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:13.906987+00:00", "epoch": 0, "step": 13779, "train_loss": 3.4967215061187744, "perplexity": 33.00706092882275, "lr": 0.0026291804804649314, "grad_norm": 0.165097, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:14.211961+00:00", "epoch": 0, "step": 13780, "train_loss": 3.5318520069122314, "perplexity": 34.187223996197076, "lr": 0.0026291804804649314, "grad_norm": 0.166317, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:14.516531+00:00", "epoch": 0, "step": 13781, "train_loss": 3.453545570373535, "perplexity": 31.61227739749093, "lr": 0.0026291804804649314, "grad_norm": 0.169452, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:14.821472+00:00", "epoch": 0, "step": 13782, "train_loss": 3.5141069889068604, "perplexity": 33.58592192721855, "lr": 0.0026291804804649314, "grad_norm": 0.181065, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:15.125915+00:00", "epoch": 0, "step": 13783, "train_loss": 3.5284371376037598, "perplexity": 34.07067820181533, "lr": 0.0026291804804649314, "grad_norm": 0.181589, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:15.430031+00:00", "epoch": 0, "step": 13784, "train_loss": 3.4811320304870605, "perplexity": 32.496488277159266, "lr": 0.0026291804804649314, "grad_norm": 0.157228, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:15.735328+00:00", "epoch": 0, "step": 13785, "train_loss": 3.4688756465911865, "perplexity": 32.10062969438861, "lr": 0.0026291804804649314, "grad_norm": 0.152284, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:16.039513+00:00", "epoch": 0, "step": 13786, "train_loss": 3.4434280395507812, "perplexity": 31.294051750020426, "lr": 0.0026291804804649314, "grad_norm": 0.171622, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:16.345041+00:00", "epoch": 0, "step": 13787, "train_loss": 3.473217725753784, "perplexity": 32.240316215189836, "lr": 0.0026291804804649314, "grad_norm": 0.180768, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:16.649773+00:00", "epoch": 0, "step": 13788, "train_loss": 3.4907286167144775, "perplexity": 32.80984480058461, "lr": 0.0026291804804649314, "grad_norm": 0.182769, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:16.954067+00:00", "epoch": 0, "step": 13789, "train_loss": 3.556570529937744, "perplexity": 35.04281255886107, "lr": 0.0026291804804649314, "grad_norm": 0.169831, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:17.258569+00:00", "epoch": 0, "step": 13790, "train_loss": 3.4714128971099854, "perplexity": 32.18218044733004, "lr": 0.0026291804804649314, "grad_norm": 0.168538, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:17.563719+00:00", "epoch": 0, "step": 13791, "train_loss": 3.5137875080108643, "perplexity": 33.57519358063053, "lr": 0.0026291804804649314, "grad_norm": 0.169682, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:17.868792+00:00", "epoch": 0, "step": 13792, "train_loss": 3.4858150482177734, "perplexity": 32.64902679947705, "lr": 0.0026291804804649314, "grad_norm": 0.172058, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:18.174350+00:00", "epoch": 0, "step": 13793, "train_loss": 3.494685173034668, "perplexity": 32.939915946620864, "lr": 0.0026291804804649314, "grad_norm": 0.183732, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:18.478199+00:00", "epoch": 0, "step": 13794, "train_loss": 3.5412774085998535, "perplexity": 34.51097566290954, "lr": 0.0026291804804649314, "grad_norm": 0.184165, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:18.781891+00:00", "epoch": 0, "step": 13795, "train_loss": 3.471604585647583, "perplexity": 32.18834999373347, "lr": 0.0026291804804649314, "grad_norm": 0.165051, "tokens_per_sec": 107898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:19.085945+00:00", "epoch": 0, "step": 13796, "train_loss": 3.5501983165740967, "perplexity": 34.82022223009891, "lr": 0.0026291804804649314, "grad_norm": 0.173059, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:19.390383+00:00", "epoch": 0, "step": 13797, "train_loss": 3.4276208877563477, "perplexity": 30.803271065212872, "lr": 0.0026291804804649314, "grad_norm": 0.20408, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:19.696752+00:00", "epoch": 0, "step": 13798, "train_loss": 3.542057991027832, "perplexity": 34.537924840743365, "lr": 0.0026291804804649314, "grad_norm": 0.180101, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:20.002067+00:00", "epoch": 0, "step": 13799, "train_loss": 3.3892674446105957, "perplexity": 29.644228275137767, "lr": 0.0026291804804649314, "grad_norm": 0.162665, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:20.306999+00:00", "epoch": 0, "step": 13800, "train_loss": 3.516322612762451, "perplexity": 33.66041819443079, "lr": 0.0026291804804649314, "grad_norm": 0.185561, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:20.612476+00:00", "epoch": 0, "step": 13801, "train_loss": 3.5143868923187256, "perplexity": 33.59532405713936, "lr": 0.0026291804804649314, "grad_norm": 0.15083, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:20.917400+00:00", "epoch": 0, "step": 13802, "train_loss": 3.459404468536377, "perplexity": 31.79803414502425, "lr": 0.0026291804804649314, "grad_norm": 0.148084, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:21.223715+00:00", "epoch": 0, "step": 13803, "train_loss": 3.4768528938293457, "perplexity": 32.35772846076991, "lr": 0.0026291804804649314, "grad_norm": 0.167318, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:21.529871+00:00", "epoch": 0, "step": 13804, "train_loss": 3.5836236476898193, "perplexity": 36.00376972977285, "lr": 0.0026291804804649314, "grad_norm": 0.159238, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:21.834425+00:00", "epoch": 0, "step": 13805, "train_loss": 3.520963668823242, "perplexity": 33.81700115639463, "lr": 0.0026291804804649314, "grad_norm": 0.151054, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:22.139016+00:00", "epoch": 0, "step": 13806, "train_loss": 3.499629497528076, "perplexity": 33.10318487451558, "lr": 0.0026291804804649314, "grad_norm": 0.155974, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:22.442931+00:00", "epoch": 0, "step": 13807, "train_loss": 3.4116175174713135, "perplexity": 30.314238434548166, "lr": 0.0026291804804649314, "grad_norm": 0.164205, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:22.747807+00:00", "epoch": 0, "step": 13808, "train_loss": 3.5482797622680664, "perplexity": 34.753481785857566, "lr": 0.0026291804804649314, "grad_norm": 0.154565, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:23.053554+00:00", "epoch": 0, "step": 13809, "train_loss": 3.564075231552124, "perplexity": 35.30678769924087, "lr": 0.0026291804804649314, "grad_norm": 0.138794, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:23.359014+00:00", "epoch": 0, "step": 13810, "train_loss": 3.4777450561523438, "perplexity": 32.38660968841631, "lr": 0.0026291804804649314, "grad_norm": 0.148427, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:23.663228+00:00", "epoch": 0, "step": 13811, "train_loss": 3.463242530822754, "perplexity": 31.920311484577145, "lr": 0.0026291804804649314, "grad_norm": 0.148397, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:23.968171+00:00", "epoch": 0, "step": 13812, "train_loss": 3.5699210166931152, "perplexity": 35.513788044431855, "lr": 0.0026291804804649314, "grad_norm": 0.159609, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:24.274048+00:00", "epoch": 0, "step": 13813, "train_loss": 3.5250117778778076, "perplexity": 33.954173521999316, "lr": 0.0026291804804649314, "grad_norm": 0.183571, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:24.579027+00:00", "epoch": 0, "step": 13814, "train_loss": 3.5073623657226562, "perplexity": 33.360159735805624, "lr": 0.0026291804804649314, "grad_norm": 0.165553, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:24.884016+00:00", "epoch": 0, "step": 13815, "train_loss": 3.5674073696136475, "perplexity": 35.42463101642896, "lr": 0.0026291804804649314, "grad_norm": 0.16686, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:25.189131+00:00", "epoch": 0, "step": 13816, "train_loss": 3.570631265640259, "perplexity": 35.53902063464909, "lr": 0.0026291804804649314, "grad_norm": 0.17291, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:25.493496+00:00", "epoch": 0, "step": 13817, "train_loss": 3.51869797706604, "perplexity": 33.740468987512315, "lr": 0.0026291804804649314, "grad_norm": 0.17121, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:25.799521+00:00", "epoch": 0, "step": 13818, "train_loss": 3.5116703510284424, "perplexity": 33.50418481997571, "lr": 0.0026291804804649314, "grad_norm": 0.19107, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:26.104901+00:00", "epoch": 0, "step": 13819, "train_loss": 3.51255464553833, "perplexity": 33.53382549027976, "lr": 0.0026291804804649314, "grad_norm": 0.159308, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:26.409760+00:00", "epoch": 0, "step": 13820, "train_loss": 3.4151666164398193, "perplexity": 30.422017813615472, "lr": 0.0026291804804649314, "grad_norm": 0.155988, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:26.714952+00:00", "epoch": 0, "step": 13821, "train_loss": 3.505413055419922, "perplexity": 33.29519377271387, "lr": 0.0026291804804649314, "grad_norm": 0.172494, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:27.019372+00:00", "epoch": 0, "step": 13822, "train_loss": 3.5153725147247314, "perplexity": 33.62845268474402, "lr": 0.0026291804804649314, "grad_norm": 0.15757, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:27.324239+00:00", "epoch": 0, "step": 13823, "train_loss": 3.495410919189453, "perplexity": 32.963830640907105, "lr": 0.0026291804804649314, "grad_norm": 0.177037, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:27.628599+00:00", "epoch": 0, "step": 13824, "train_loss": 3.5033202171325684, "perplexity": 33.22558518157079, "lr": 0.0026291804804649314, "grad_norm": 0.19952, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:27.934008+00:00", "epoch": 0, "step": 13825, "train_loss": 3.5170938968658447, "perplexity": 33.68638995441304, "lr": 0.0026291804804649314, "grad_norm": 0.180781, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:28.238946+00:00", "epoch": 0, "step": 13826, "train_loss": 3.5860133171081543, "perplexity": 36.089909719298184, "lr": 0.0026291804804649314, "grad_norm": 0.184458, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:28.544209+00:00", "epoch": 0, "step": 13827, "train_loss": 3.3958137035369873, "perplexity": 29.83892363694388, "lr": 0.0026291804804649314, "grad_norm": 0.165231, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:28.849341+00:00", "epoch": 0, "step": 13828, "train_loss": 3.5797300338745117, "perplexity": 35.86385751337758, "lr": 0.0026291804804649314, "grad_norm": 0.176425, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:29.153961+00:00", "epoch": 0, "step": 13829, "train_loss": 3.51658034324646, "perplexity": 33.669094628346706, "lr": 0.0026291804804649314, "grad_norm": 0.153739, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:29.458528+00:00", "epoch": 0, "step": 13830, "train_loss": 3.624286651611328, "perplexity": 37.49796450405636, "lr": 0.0026291804804649314, "grad_norm": 0.173941, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:29.763383+00:00", "epoch": 0, "step": 13831, "train_loss": 3.5551865100860596, "perplexity": 34.99434615758799, "lr": 0.0026291804804649314, "grad_norm": 0.155263, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:30.068030+00:00", "epoch": 0, "step": 13832, "train_loss": 3.5103421211242676, "perplexity": 33.45971310065463, "lr": 0.0026291804804649314, "grad_norm": 0.175154, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:30.373091+00:00", "epoch": 0, "step": 13833, "train_loss": 3.490739107131958, "perplexity": 32.81018899135938, "lr": 0.0026291804804649314, "grad_norm": 0.170542, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:30.679196+00:00", "epoch": 0, "step": 13834, "train_loss": 3.502652168273926, "perplexity": 33.203396279772996, "lr": 0.0026291804804649314, "grad_norm": 0.159102, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:30.985055+00:00", "epoch": 0, "step": 13835, "train_loss": 3.6157515048980713, "perplexity": 37.1792758374979, "lr": 0.0026291804804649314, "grad_norm": 0.154428, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:31.289963+00:00", "epoch": 0, "step": 13836, "train_loss": 3.4697043895721436, "perplexity": 32.12724389257385, "lr": 0.0026291804804649314, "grad_norm": 0.16286, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:31.594663+00:00", "epoch": 0, "step": 13837, "train_loss": 3.5824427604675293, "perplexity": 35.96127843179793, "lr": 0.0026291804804649314, "grad_norm": 0.163313, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:31.898819+00:00", "epoch": 0, "step": 13838, "train_loss": 3.6056032180786133, "perplexity": 36.803877927167825, "lr": 0.0026291804804649314, "grad_norm": 0.178062, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:32.202946+00:00", "epoch": 0, "step": 13839, "train_loss": 3.5837910175323486, "perplexity": 36.00979617935187, "lr": 0.0026291804804649314, "grad_norm": 0.161823, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:32.507071+00:00", "epoch": 0, "step": 13840, "train_loss": 3.60935378074646, "perplexity": 36.94217235660317, "lr": 0.0026291804804649314, "grad_norm": 0.163762, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:32.812689+00:00", "epoch": 0, "step": 13841, "train_loss": 3.5382094383239746, "perplexity": 34.4052592656871, "lr": 0.0026291804804649314, "grad_norm": 0.18162, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:33.118325+00:00", "epoch": 0, "step": 13842, "train_loss": 3.6517412662506104, "perplexity": 38.54171904814689, "lr": 0.0026291804804649314, "grad_norm": 0.170851, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:33.422720+00:00", "epoch": 0, "step": 13843, "train_loss": 3.469712495803833, "perplexity": 32.127504324511946, "lr": 0.0026291804804649314, "grad_norm": 0.159766, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:33.726889+00:00", "epoch": 0, "step": 13844, "train_loss": 3.4293057918548584, "perplexity": 30.855215371176463, "lr": 0.0026291804804649314, "grad_norm": 0.169228, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:34.031435+00:00", "epoch": 0, "step": 13845, "train_loss": 3.5072479248046875, "perplexity": 33.35634218694794, "lr": 0.0026291804804649314, "grad_norm": 0.161654, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:34.336702+00:00", "epoch": 0, "step": 13846, "train_loss": 3.52203369140625, "perplexity": 33.85320547758787, "lr": 0.0026291804804649314, "grad_norm": 0.163758, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:34.641436+00:00", "epoch": 0, "step": 13847, "train_loss": 3.4942002296447754, "perplexity": 32.923945824734425, "lr": 0.0026291804804649314, "grad_norm": 0.158365, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:34.945922+00:00", "epoch": 0, "step": 13848, "train_loss": 3.599815845489502, "perplexity": 36.59149533426666, "lr": 0.0026291804804649314, "grad_norm": 0.15027, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:35.249997+00:00", "epoch": 0, "step": 13849, "train_loss": 3.472440481185913, "perplexity": 32.2152673403571, "lr": 0.0026291804804649314, "grad_norm": 0.146297, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:35.554568+00:00", "epoch": 0, "step": 13850, "train_loss": 3.5211265087127686, "perplexity": 33.82250836151122, "lr": 0.0026291804804649314, "grad_norm": 0.172266, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:35.860374+00:00", "epoch": 0, "step": 13851, "train_loss": 3.5117273330688477, "perplexity": 33.50609401118316, "lr": 0.0026291804804649314, "grad_norm": 0.15105, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:36.165732+00:00", "epoch": 0, "step": 13852, "train_loss": 3.5986218452453613, "perplexity": 36.54783115261329, "lr": 0.0026291804804649314, "grad_norm": 0.165543, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:36.469710+00:00", "epoch": 0, "step": 13853, "train_loss": 3.454087495803833, "perplexity": 31.629413537357866, "lr": 0.0026291804804649314, "grad_norm": 0.201915, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:36.774326+00:00", "epoch": 0, "step": 13854, "train_loss": 3.5534369945526123, "perplexity": 34.93317652961198, "lr": 0.0026291804804649314, "grad_norm": 0.18265, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:37.079095+00:00", "epoch": 0, "step": 13855, "train_loss": 3.555166006088257, "perplexity": 34.99362864094727, "lr": 0.0026291804804649314, "grad_norm": 0.157252, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:37.384375+00:00", "epoch": 0, "step": 13856, "train_loss": 3.4903125762939453, "perplexity": 32.796197418084404, "lr": 0.0026291804804649314, "grad_norm": 0.160213, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:37.689892+00:00", "epoch": 0, "step": 13857, "train_loss": 3.479387044906616, "perplexity": 32.43983182042702, "lr": 0.0026291804804649314, "grad_norm": 0.158733, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:37.995366+00:00", "epoch": 0, "step": 13858, "train_loss": 3.5880441665649414, "perplexity": 36.16327736695467, "lr": 0.0026291804804649314, "grad_norm": 0.154245, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:38.300784+00:00", "epoch": 0, "step": 13859, "train_loss": 3.5747792720794678, "perplexity": 35.68674288570058, "lr": 0.0026291804804649314, "grad_norm": 0.172829, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:38.604977+00:00", "epoch": 0, "step": 13860, "train_loss": 3.4279518127441406, "perplexity": 30.813466324154042, "lr": 0.0026291804804649314, "grad_norm": 0.17451, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:38.909988+00:00", "epoch": 0, "step": 13861, "train_loss": 3.502192497253418, "perplexity": 33.18813714807975, "lr": 0.0026291804804649314, "grad_norm": 0.157853, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:39.214999+00:00", "epoch": 0, "step": 13862, "train_loss": 3.5011439323425293, "perplexity": 33.153355470618614, "lr": 0.0026291804804649314, "grad_norm": 0.142409, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:39.520267+00:00", "epoch": 0, "step": 13863, "train_loss": 3.477750062942505, "perplexity": 32.386771841780984, "lr": 0.0026291804804649314, "grad_norm": 0.16547, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:39.824681+00:00", "epoch": 0, "step": 13864, "train_loss": 3.51572322845459, "perplexity": 33.640248713208116, "lr": 0.0026291804804649314, "grad_norm": 0.176172, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:40.128465+00:00", "epoch": 0, "step": 13865, "train_loss": 3.5875778198242188, "perplexity": 36.14641667219133, "lr": 0.0026291804804649314, "grad_norm": 0.168997, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:40.432712+00:00", "epoch": 0, "step": 13866, "train_loss": 3.510988712310791, "perplexity": 33.481354852178015, "lr": 0.0026291804804649314, "grad_norm": 0.161039, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:40.738218+00:00", "epoch": 0, "step": 13867, "train_loss": 3.46801495552063, "perplexity": 32.07301285553956, "lr": 0.0026291804804649314, "grad_norm": 0.148396, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:41.043924+00:00", "epoch": 0, "step": 13868, "train_loss": 3.5665283203125, "perplexity": 35.39350470207735, "lr": 0.0026291804804649314, "grad_norm": 0.169281, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:41.348360+00:00", "epoch": 0, "step": 13869, "train_loss": 3.53843092918396, "perplexity": 34.41288056014225, "lr": 0.0026291804804649314, "grad_norm": 0.194127, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:41.652987+00:00", "epoch": 0, "step": 13870, "train_loss": 3.3923838138580322, "perplexity": 29.736754734922453, "lr": 0.0026291804804649314, "grad_norm": 0.20267, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:41.957068+00:00", "epoch": 0, "step": 13871, "train_loss": 3.548180103302002, "perplexity": 34.750018462374136, "lr": 0.0026291804804649314, "grad_norm": 0.182605, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:42.261508+00:00", "epoch": 0, "step": 13872, "train_loss": 3.456064462661743, "perplexity": 31.69200569057634, "lr": 0.0026291804804649314, "grad_norm": 0.16903, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:42.566515+00:00", "epoch": 0, "step": 13873, "train_loss": 3.5233781337738037, "perplexity": 33.898749770287594, "lr": 0.0026291804804649314, "grad_norm": 0.159817, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:42.872687+00:00", "epoch": 0, "step": 13874, "train_loss": 3.5090231895446777, "perplexity": 33.41561111854387, "lr": 0.0026291804804649314, "grad_norm": 0.1684, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:43.177933+00:00", "epoch": 0, "step": 13875, "train_loss": 3.4742863178253174, "perplexity": 32.27478637547101, "lr": 0.0026291804804649314, "grad_norm": 0.171138, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:43.482210+00:00", "epoch": 0, "step": 13876, "train_loss": 3.546200752258301, "perplexity": 34.68130400451324, "lr": 0.0026291804804649314, "grad_norm": 0.1875, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:43.787190+00:00", "epoch": 0, "step": 13877, "train_loss": 3.613292694091797, "perplexity": 37.08797132855617, "lr": 0.0026291804804649314, "grad_norm": 0.186698, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:44.092895+00:00", "epoch": 0, "step": 13878, "train_loss": 3.5191969871520996, "perplexity": 33.757310023419386, "lr": 0.0026291804804649314, "grad_norm": 0.157057, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:44.398185+00:00", "epoch": 0, "step": 13879, "train_loss": 3.4918389320373535, "perplexity": 32.8462943054852, "lr": 0.0026291804804649314, "grad_norm": 0.180175, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:44.703310+00:00", "epoch": 0, "step": 13880, "train_loss": 3.4215431213378906, "perplexity": 30.616623752688714, "lr": 0.0026291804804649314, "grad_norm": 0.138696, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:45.008181+00:00", "epoch": 0, "step": 13881, "train_loss": 3.465420961380005, "perplexity": 31.98992346141188, "lr": 0.0026291804804649314, "grad_norm": 0.14938, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:45.313206+00:00", "epoch": 0, "step": 13882, "train_loss": 3.5502192974090576, "perplexity": 34.82095279509873, "lr": 0.0026291804804649314, "grad_norm": 0.146573, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:45.619193+00:00", "epoch": 0, "step": 13883, "train_loss": 3.4526891708374023, "perplexity": 31.585216247023904, "lr": 0.0026291804804649314, "grad_norm": 0.156545, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:45.925892+00:00", "epoch": 0, "step": 13884, "train_loss": 3.5648975372314453, "perplexity": 35.33583261154607, "lr": 0.0026291804804649314, "grad_norm": 0.149318, "tokens_per_sec": 106848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:46.230698+00:00", "epoch": 0, "step": 13885, "train_loss": 3.6165616512298584, "perplexity": 37.20940869580035, "lr": 0.0026291804804649314, "grad_norm": 0.141319, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:46.534978+00:00", "epoch": 0, "step": 13886, "train_loss": 3.547983169555664, "perplexity": 34.743175684862045, "lr": 0.0026291804804649314, "grad_norm": 0.180141, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:46.839810+00:00", "epoch": 0, "step": 13887, "train_loss": 3.5323243141174316, "perplexity": 34.20337468215872, "lr": 0.0026291804804649314, "grad_norm": 0.204079, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:47.144192+00:00", "epoch": 0, "step": 13888, "train_loss": 3.654491901397705, "perplexity": 38.64787919216794, "lr": 0.0026291804804649314, "grad_norm": 0.194197, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:47.450211+00:00", "epoch": 0, "step": 13889, "train_loss": 3.4607131481170654, "perplexity": 31.83967482422621, "lr": 0.0026291804804649314, "grad_norm": 0.179521, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:47.755094+00:00", "epoch": 0, "step": 13890, "train_loss": 3.568122625350952, "perplexity": 35.4499777506233, "lr": 0.0026291804804649314, "grad_norm": 0.180535, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:48.059969+00:00", "epoch": 0, "step": 13891, "train_loss": 3.411585569381714, "perplexity": 30.313269968012925, "lr": 0.0026291804804649314, "grad_norm": 0.176091, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:48.365221+00:00", "epoch": 0, "step": 13892, "train_loss": 3.5795493125915527, "perplexity": 35.85737673666068, "lr": 0.0026291804804649314, "grad_norm": 0.16737, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:48.670976+00:00", "epoch": 0, "step": 13893, "train_loss": 3.4611268043518066, "perplexity": 31.85284822867198, "lr": 0.0026291804804649314, "grad_norm": 0.167924, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:48.976191+00:00", "epoch": 0, "step": 13894, "train_loss": 3.484758138656616, "perplexity": 32.61453795989286, "lr": 0.0026291804804649314, "grad_norm": 0.190356, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:49.283060+00:00", "epoch": 0, "step": 13895, "train_loss": 3.4736359119415283, "perplexity": 32.253801489600754, "lr": 0.0026291804804649314, "grad_norm": 0.158817, "tokens_per_sec": 106788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:49.588527+00:00", "epoch": 0, "step": 13896, "train_loss": 3.506861686706543, "perplexity": 33.34346118450767, "lr": 0.0026291804804649314, "grad_norm": 0.169845, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:49.893527+00:00", "epoch": 0, "step": 13897, "train_loss": 3.5942089557647705, "perplexity": 36.38690494890813, "lr": 0.0026291804804649314, "grad_norm": 0.172941, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:50.199313+00:00", "epoch": 0, "step": 13898, "train_loss": 3.4968061447143555, "perplexity": 33.00985471833357, "lr": 0.0026291804804649314, "grad_norm": 0.205029, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:50.504454+00:00", "epoch": 0, "step": 13899, "train_loss": 3.493684768676758, "perplexity": 32.90697918894357, "lr": 0.0026291804804649314, "grad_norm": 0.173961, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:50.810316+00:00", "epoch": 0, "step": 13900, "train_loss": 3.5398976802825928, "perplexity": 34.46339272584241, "lr": 0.0026291804804649314, "grad_norm": 0.174849, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:51.116244+00:00", "epoch": 0, "step": 13901, "train_loss": 3.5429139137268066, "perplexity": 34.56749928950551, "lr": 0.0026291804804649314, "grad_norm": 0.165756, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:51.420653+00:00", "epoch": 0, "step": 13902, "train_loss": 3.5323336124420166, "perplexity": 34.20369271771701, "lr": 0.0026291804804649314, "grad_norm": 0.14794, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:51.726049+00:00", "epoch": 0, "step": 13903, "train_loss": 3.486689805984497, "perplexity": 32.67759928442544, "lr": 0.0026291804804649314, "grad_norm": 0.172273, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:52.031122+00:00", "epoch": 0, "step": 13904, "train_loss": 3.5221405029296875, "perplexity": 33.85682158315561, "lr": 0.0026291804804649314, "grad_norm": 0.192158, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:52.335986+00:00", "epoch": 0, "step": 13905, "train_loss": 3.5976450443267822, "perplexity": 36.512148627770046, "lr": 0.0026291804804649314, "grad_norm": 0.199713, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:52.640721+00:00", "epoch": 0, "step": 13906, "train_loss": 3.4940025806427, "perplexity": 32.9174390827445, "lr": 0.0026291804804649314, "grad_norm": 0.187011, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:52.945162+00:00", "epoch": 0, "step": 13907, "train_loss": 3.470139265060425, "perplexity": 32.1412182817882, "lr": 0.0026291804804649314, "grad_norm": 0.179632, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:53.248833+00:00", "epoch": 0, "step": 13908, "train_loss": 3.485921859741211, "perplexity": 32.65251427801638, "lr": 0.0026291804804649314, "grad_norm": 0.16918, "tokens_per_sec": 107906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:53.553795+00:00", "epoch": 0, "step": 13909, "train_loss": 3.5026638507843018, "perplexity": 33.20378418106038, "lr": 0.0026291804804649314, "grad_norm": 0.161533, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:53.859526+00:00", "epoch": 0, "step": 13910, "train_loss": 3.489274024963379, "perplexity": 32.762154564303735, "lr": 0.0026291804804649314, "grad_norm": 0.155192, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:54.165835+00:00", "epoch": 0, "step": 13911, "train_loss": 3.5977659225463867, "perplexity": 36.516562418050285, "lr": 0.0026291804804649314, "grad_norm": 0.158947, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:54.472313+00:00", "epoch": 0, "step": 13912, "train_loss": 3.5398850440979004, "perplexity": 34.462957242798225, "lr": 0.0026291804804649314, "grad_norm": 0.154921, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:54.776705+00:00", "epoch": 0, "step": 13913, "train_loss": 3.4681308269500732, "perplexity": 32.07672941670315, "lr": 0.0026291804804649314, "grad_norm": 0.146591, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:55.081870+00:00", "epoch": 0, "step": 13914, "train_loss": 3.535454750061035, "perplexity": 34.310613920864604, "lr": 0.0026291804804649314, "grad_norm": 0.146786, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:55.387456+00:00", "epoch": 0, "step": 13915, "train_loss": 3.4618964195251465, "perplexity": 31.87737209974403, "lr": 0.0026291804804649314, "grad_norm": 0.158598, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:55.693543+00:00", "epoch": 0, "step": 13916, "train_loss": 3.458906650543213, "perplexity": 31.78220845096407, "lr": 0.0026291804804649314, "grad_norm": 0.1672, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:55.999317+00:00", "epoch": 0, "step": 13917, "train_loss": 3.4783833026885986, "perplexity": 32.407286927757255, "lr": 0.0026291804804649314, "grad_norm": 0.152488, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:56.304751+00:00", "epoch": 0, "step": 13918, "train_loss": 3.4255990982055664, "perplexity": 30.741056247470162, "lr": 0.0026291804804649314, "grad_norm": 0.178943, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:56.609460+00:00", "epoch": 0, "step": 13919, "train_loss": 3.5609657764434814, "perplexity": 35.197173336648696, "lr": 0.0026291804804649314, "grad_norm": 0.174658, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:56.913844+00:00", "epoch": 0, "step": 13920, "train_loss": 3.4690608978271484, "perplexity": 32.1065769265637, "lr": 0.0026291804804649314, "grad_norm": 0.163135, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:57.218815+00:00", "epoch": 0, "step": 13921, "train_loss": 3.513050079345703, "perplexity": 33.55044339731476, "lr": 0.0026291804804649314, "grad_norm": 0.148335, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:57.525271+00:00", "epoch": 0, "step": 13922, "train_loss": 3.551561117172241, "perplexity": 34.86770759897692, "lr": 0.0026291804804649314, "grad_norm": 0.166094, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:57.830075+00:00", "epoch": 0, "step": 13923, "train_loss": 3.467149257659912, "perplexity": 32.04525933174011, "lr": 0.0026291804804649314, "grad_norm": 0.177842, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:58.135478+00:00", "epoch": 0, "step": 13924, "train_loss": 3.5158157348632812, "perplexity": 33.64336079574563, "lr": 0.0026291804804649314, "grad_norm": 0.167495, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:58.440279+00:00", "epoch": 0, "step": 13925, "train_loss": 3.4798073768615723, "perplexity": 32.45347018447385, "lr": 0.0026291804804649314, "grad_norm": 0.164799, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:58.744580+00:00", "epoch": 0, "step": 13926, "train_loss": 3.6349284648895264, "perplexity": 37.89914168088272, "lr": 0.0026291804804649314, "grad_norm": 0.189905, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:59.050381+00:00", "epoch": 0, "step": 13927, "train_loss": 3.5684943199157715, "perplexity": 35.46315676380822, "lr": 0.0026291804804649314, "grad_norm": 0.19993, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:59.354543+00:00", "epoch": 0, "step": 13928, "train_loss": 3.535430669784546, "perplexity": 34.30978772174247, "lr": 0.0026291804804649314, "grad_norm": 0.160811, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:59.659161+00:00", "epoch": 0, "step": 13929, "train_loss": 3.486285448074341, "perplexity": 32.664388509790484, "lr": 0.0026291804804649314, "grad_norm": 0.215481, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:10:59.963362+00:00", "epoch": 0, "step": 13930, "train_loss": 3.4471428394317627, "perplexity": 31.410519082217863, "lr": 0.0026291804804649314, "grad_norm": 0.18545, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:00.267637+00:00", "epoch": 0, "step": 13931, "train_loss": 3.486957550048828, "perplexity": 32.68634968905261, "lr": 0.0026291804804649314, "grad_norm": 0.184108, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:00.573531+00:00", "epoch": 0, "step": 13932, "train_loss": 3.5345404148101807, "perplexity": 34.279256854698346, "lr": 0.0026291804804649314, "grad_norm": 0.156834, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:00.879765+00:00", "epoch": 0, "step": 13933, "train_loss": 3.5055582523345947, "perplexity": 33.30002848310713, "lr": 0.0026291804804649314, "grad_norm": 0.1677, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:01.183985+00:00", "epoch": 0, "step": 13934, "train_loss": 3.595353603363037, "perplexity": 36.42857897875517, "lr": 0.0026291804804649314, "grad_norm": 0.186832, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:01.489334+00:00", "epoch": 0, "step": 13935, "train_loss": 3.4327495098114014, "perplexity": 30.961655199586524, "lr": 0.0026291804804649314, "grad_norm": 0.188148, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:01.794359+00:00", "epoch": 0, "step": 13936, "train_loss": 3.5892999172210693, "perplexity": 36.20871795128644, "lr": 0.0026291804804649314, "grad_norm": 0.15961, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:02.100903+00:00", "epoch": 0, "step": 13937, "train_loss": 3.596482753753662, "perplexity": 36.46973555454794, "lr": 0.0026291804804649314, "grad_norm": 0.164668, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:02.406496+00:00", "epoch": 0, "step": 13938, "train_loss": 3.436805248260498, "perplexity": 31.087482563989074, "lr": 0.0026291804804649314, "grad_norm": 0.163602, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:02.711396+00:00", "epoch": 0, "step": 13939, "train_loss": 3.529832363128662, "perplexity": 34.118247659036925, "lr": 0.0026291804804649314, "grad_norm": 0.162728, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:03.016315+00:00", "epoch": 0, "step": 13940, "train_loss": 3.526188611984253, "perplexity": 33.99415547290352, "lr": 0.0026291804804649314, "grad_norm": 0.162426, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:03.322289+00:00", "epoch": 0, "step": 13941, "train_loss": 3.5471887588500977, "perplexity": 34.71558629425719, "lr": 0.0026291804804649314, "grad_norm": 0.164488, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:03.628395+00:00", "epoch": 0, "step": 13942, "train_loss": 3.533482313156128, "perplexity": 34.243005098724396, "lr": 0.0026291804804649314, "grad_norm": 0.182537, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:03.935345+00:00", "epoch": 0, "step": 13943, "train_loss": 3.496811628341675, "perplexity": 33.01003573257102, "lr": 0.0026291804804649314, "grad_norm": 0.188473, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:04.240937+00:00", "epoch": 0, "step": 13944, "train_loss": 3.548676013946533, "perplexity": 34.767255640123985, "lr": 0.0026291804804649314, "grad_norm": 0.15716, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:04.545582+00:00", "epoch": 0, "step": 13945, "train_loss": 3.539259195327759, "perplexity": 34.44139539132528, "lr": 0.0026291804804649314, "grad_norm": 0.168128, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:04.850990+00:00", "epoch": 0, "step": 13946, "train_loss": 3.4972472190856934, "perplexity": 33.02441773070108, "lr": 0.0026291804804649314, "grad_norm": 0.148474, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:05.156708+00:00", "epoch": 0, "step": 13947, "train_loss": 3.424459934234619, "perplexity": 30.70605708244294, "lr": 0.0026291804804649314, "grad_norm": 0.151523, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:05.462468+00:00", "epoch": 0, "step": 13948, "train_loss": 3.4858691692352295, "perplexity": 32.650793845843125, "lr": 0.0026291804804649314, "grad_norm": 0.172595, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:05.768583+00:00", "epoch": 0, "step": 13949, "train_loss": 3.6516530513763428, "perplexity": 38.5383192452058, "lr": 0.0026291804804649314, "grad_norm": 0.207236, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:06.073508+00:00", "epoch": 0, "step": 13950, "train_loss": 3.4908480644226074, "perplexity": 32.813764095420865, "lr": 0.0026291804804649314, "grad_norm": 0.213618, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:06.377933+00:00", "epoch": 0, "step": 13951, "train_loss": 3.558666706085205, "perplexity": 35.11634550876969, "lr": 0.0026291804804649314, "grad_norm": 0.183899, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:06.683795+00:00", "epoch": 0, "step": 13952, "train_loss": 3.532595634460449, "perplexity": 34.212656012559776, "lr": 0.0026291804804649314, "grad_norm": 0.148567, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:06.989586+00:00", "epoch": 0, "step": 13953, "train_loss": 3.6355135440826416, "perplexity": 37.9213221681552, "lr": 0.0026291804804649314, "grad_norm": 0.164182, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:07.296091+00:00", "epoch": 0, "step": 13954, "train_loss": 3.4820008277893066, "perplexity": 32.52473340637633, "lr": 0.0026291804804649314, "grad_norm": 0.16453, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:07.601161+00:00", "epoch": 0, "step": 13955, "train_loss": 3.4822781085968018, "perplexity": 32.533753141160695, "lr": 0.0026291804804649314, "grad_norm": 0.169314, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:07.905970+00:00", "epoch": 0, "step": 13956, "train_loss": 3.501152515411377, "perplexity": 33.15364002937234, "lr": 0.0026291804804649314, "grad_norm": 0.147956, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:08.210664+00:00", "epoch": 0, "step": 13957, "train_loss": 3.4675350189208984, "perplexity": 32.057623536050755, "lr": 0.0026291804804649314, "grad_norm": 0.179892, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:08.516155+00:00", "epoch": 0, "step": 13958, "train_loss": 3.4390342235565186, "perplexity": 31.15685307849993, "lr": 0.0026291804804649314, "grad_norm": 0.16067, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:08.821245+00:00", "epoch": 0, "step": 13959, "train_loss": 3.591139554977417, "perplexity": 36.27539018351337, "lr": 0.0026291804804649314, "grad_norm": 0.149378, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:09.125837+00:00", "epoch": 0, "step": 13960, "train_loss": 3.5214593410491943, "perplexity": 33.83376745958498, "lr": 0.0026291804804649314, "grad_norm": 0.195973, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:09.429687+00:00", "epoch": 0, "step": 13961, "train_loss": 3.5226705074310303, "perplexity": 33.87477060710742, "lr": 0.0026291804804649314, "grad_norm": 0.175998, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:09.733010+00:00", "epoch": 0, "step": 13962, "train_loss": 3.503751039505005, "perplexity": 33.23990259091298, "lr": 0.0026291804804649314, "grad_norm": 0.173987, "tokens_per_sec": 108031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:10.037418+00:00", "epoch": 0, "step": 13963, "train_loss": 3.5366687774658203, "perplexity": 34.35229324123898, "lr": 0.0026291804804649314, "grad_norm": 0.16283, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:10.343574+00:00", "epoch": 0, "step": 13964, "train_loss": 3.4920971393585205, "perplexity": 32.854776554190316, "lr": 0.0026291804804649314, "grad_norm": 0.160776, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:10.648909+00:00", "epoch": 0, "step": 13965, "train_loss": 3.541353464126587, "perplexity": 34.513600513157584, "lr": 0.0026291804804649314, "grad_norm": 0.172282, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:10.953879+00:00", "epoch": 0, "step": 13966, "train_loss": 3.4769294261932373, "perplexity": 32.36020496898448, "lr": 0.0026291804804649314, "grad_norm": 0.158151, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:11.258240+00:00", "epoch": 0, "step": 13967, "train_loss": 3.4994871616363525, "perplexity": 33.098473438488746, "lr": 0.0026291804804649314, "grad_norm": 0.174558, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:11.563010+00:00", "epoch": 0, "step": 13968, "train_loss": 3.521026849746704, "perplexity": 33.81913781325367, "lr": 0.0026291804804649314, "grad_norm": 0.161211, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:11.867947+00:00", "epoch": 0, "step": 13969, "train_loss": 3.5263445377349854, "perplexity": 33.999456450384855, "lr": 0.0026291804804649314, "grad_norm": 0.15672, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:12.173216+00:00", "epoch": 0, "step": 13970, "train_loss": 3.5308785438537598, "perplexity": 34.15396018973529, "lr": 0.0026291804804649314, "grad_norm": 0.16729, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:12.478356+00:00", "epoch": 0, "step": 13971, "train_loss": 3.5303573608398438, "perplexity": 34.13616436367039, "lr": 0.0026291804804649314, "grad_norm": 0.15197, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:12.783104+00:00", "epoch": 0, "step": 13972, "train_loss": 3.5228331089019775, "perplexity": 33.88027914247198, "lr": 0.0026291804804649314, "grad_norm": 0.188601, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:13.088107+00:00", "epoch": 0, "step": 13973, "train_loss": 3.563703775405884, "perplexity": 35.29367521145385, "lr": 0.0026291804804649314, "grad_norm": 0.174502, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:13.394178+00:00", "epoch": 0, "step": 13974, "train_loss": 3.505553722381592, "perplexity": 33.29987763588477, "lr": 0.0026291804804649314, "grad_norm": 0.166758, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:13.700662+00:00", "epoch": 0, "step": 13975, "train_loss": 3.4970357418060303, "perplexity": 33.01743455509445, "lr": 0.0026291804804649314, "grad_norm": 0.14297, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:14.007617+00:00", "epoch": 0, "step": 13976, "train_loss": 3.525500535964966, "perplexity": 33.97077295512411, "lr": 0.0026291804804649314, "grad_norm": 0.163306, "tokens_per_sec": 106752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:14.313398+00:00", "epoch": 0, "step": 13977, "train_loss": 3.4789559841156006, "perplexity": 32.425851294306895, "lr": 0.0026291804804649314, "grad_norm": 0.146818, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:14.618220+00:00", "epoch": 0, "step": 13978, "train_loss": 3.5833122730255127, "perplexity": 35.992560813236345, "lr": 0.0026291804804649314, "grad_norm": 0.139205, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:14.924447+00:00", "epoch": 0, "step": 13979, "train_loss": 3.441596031188965, "perplexity": 31.236773268887564, "lr": 0.0026291804804649314, "grad_norm": 0.160295, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:15.228878+00:00", "epoch": 0, "step": 13980, "train_loss": 3.483242988586426, "perplexity": 32.5651594577806, "lr": 0.0026291804804649314, "grad_norm": 0.166129, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:15.533565+00:00", "epoch": 0, "step": 13981, "train_loss": 3.6232359409332275, "perplexity": 37.458585683841186, "lr": 0.0026291804804649314, "grad_norm": 0.179221, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:15.838807+00:00", "epoch": 0, "step": 13982, "train_loss": 3.565152168273926, "perplexity": 35.344831357072216, "lr": 0.0026291804804649314, "grad_norm": 0.16932, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:16.144692+00:00", "epoch": 0, "step": 13983, "train_loss": 3.4392919540405273, "perplexity": 31.164884184209967, "lr": 0.0026291804804649314, "grad_norm": 0.164348, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:16.449221+00:00", "epoch": 0, "step": 13984, "train_loss": 3.5127880573272705, "perplexity": 33.54165359402778, "lr": 0.0026291804804649314, "grad_norm": 0.170301, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:16.754079+00:00", "epoch": 0, "step": 13985, "train_loss": 3.5469183921813965, "perplexity": 34.70620162554678, "lr": 0.0026291804804649314, "grad_norm": 0.200059, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:17.059381+00:00", "epoch": 0, "step": 13986, "train_loss": 3.4909794330596924, "perplexity": 32.81807507804547, "lr": 0.0026291804804649314, "grad_norm": 0.171522, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:17.365679+00:00", "epoch": 0, "step": 13987, "train_loss": 3.556393623352051, "perplexity": 35.03661380285474, "lr": 0.0026291804804649314, "grad_norm": 0.189707, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:17.671232+00:00", "epoch": 0, "step": 13988, "train_loss": 3.580397844314575, "perplexity": 35.88781577074277, "lr": 0.0026291804804649314, "grad_norm": 0.163227, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:17.975186+00:00", "epoch": 0, "step": 13989, "train_loss": 3.4411251544952393, "perplexity": 31.22206806280917, "lr": 0.0026291804804649314, "grad_norm": 0.171036, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:18.279453+00:00", "epoch": 0, "step": 13990, "train_loss": 3.4791271686553955, "perplexity": 32.43140257387123, "lr": 0.0026291804804649314, "grad_norm": 0.170356, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:18.584314+00:00", "epoch": 0, "step": 13991, "train_loss": 3.4745640754699707, "perplexity": 32.28375218922036, "lr": 0.0026291804804649314, "grad_norm": 0.164273, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:18.890678+00:00", "epoch": 0, "step": 13992, "train_loss": 3.54754376411438, "perplexity": 34.727912697983946, "lr": 0.0026291804804649314, "grad_norm": 0.181222, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:19.196336+00:00", "epoch": 0, "step": 13993, "train_loss": 3.58372163772583, "perplexity": 36.007297913325786, "lr": 0.0026291804804649314, "grad_norm": 0.168805, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:19.500567+00:00", "epoch": 0, "step": 13994, "train_loss": 3.4926280975341797, "perplexity": 32.87222569838417, "lr": 0.0026291804804649314, "grad_norm": 0.179607, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:19.805334+00:00", "epoch": 0, "step": 13995, "train_loss": 3.423743724822998, "perplexity": 30.684072988913805, "lr": 0.0026291804804649314, "grad_norm": 0.202458, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:20.111143+00:00", "epoch": 0, "step": 13996, "train_loss": 3.4854843616485596, "perplexity": 32.63823198976422, "lr": 0.0026291804804649314, "grad_norm": 0.172004, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:20.416451+00:00", "epoch": 0, "step": 13997, "train_loss": 3.43076753616333, "perplexity": 30.900350786813686, "lr": 0.0026291804804649314, "grad_norm": 0.228918, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:20.722546+00:00", "epoch": 0, "step": 13998, "train_loss": 3.5450356006622314, "perplexity": 34.64091855995007, "lr": 0.0026291804804649314, "grad_norm": 0.213509, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:21.028905+00:00", "epoch": 0, "step": 13999, "train_loss": 3.662140130996704, "perplexity": 38.9446002949037, "lr": 0.0026291804804649314, "grad_norm": 0.19548, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:21.333773+00:00", "epoch": 0, "step": 14000, "train_loss": 3.560637950897217, "perplexity": 35.18563669517886, "lr": 0.0026291804804649314, "grad_norm": 0.187528, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:11:24.407917+00:00", "step": 14000, "epoch": 0, "val_loss": 3.4780593514442444, "val_ppl": 32.39679024712903, "eval_train_loss": 3.560637950897217, "eval_train_ppl": 35.18563669517886} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:11:25.354055+00:00", "step": 14000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4781_epoch_0000_step_0014000.pt", "val_loss": 3.4780593514442444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:26.590967+00:00", "epoch": 0, "step": 14001, "train_loss": 3.545789957046509, "perplexity": 34.66706001677077, "lr": 0.0026291804804649314, "grad_norm": 0.164836, "tokens_per_sec": 6233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:26.895093+00:00", "epoch": 0, "step": 14002, "train_loss": 3.5342721939086914, "perplexity": 34.270063674479054, "lr": 0.0026291804804649314, "grad_norm": 0.161835, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:27.198948+00:00", "epoch": 0, "step": 14003, "train_loss": 3.5307605266571045, "perplexity": 34.149929672938846, "lr": 0.0026291804804649314, "grad_norm": 0.168894, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:27.501850+00:00", "epoch": 0, "step": 14004, "train_loss": 3.4453070163726807, "perplexity": 31.352907825206376, "lr": 0.0026291804804649314, "grad_norm": 0.179193, "tokens_per_sec": 108180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:27.806239+00:00", "epoch": 0, "step": 14005, "train_loss": 3.384172201156616, "perplexity": 29.493567866747856, "lr": 0.0026291804804649314, "grad_norm": 0.160559, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:28.111591+00:00", "epoch": 0, "step": 14006, "train_loss": 3.4634599685668945, "perplexity": 31.92725291973627, "lr": 0.0026291804804649314, "grad_norm": 0.176199, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:28.417369+00:00", "epoch": 0, "step": 14007, "train_loss": 3.4313759803771973, "perplexity": 30.919157647338924, "lr": 0.0026291804804649314, "grad_norm": 0.183125, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:28.722685+00:00", "epoch": 0, "step": 14008, "train_loss": 3.5837090015411377, "perplexity": 36.006842921333764, "lr": 0.0026291804804649314, "grad_norm": 0.152962, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:29.026492+00:00", "epoch": 0, "step": 14009, "train_loss": 3.423616647720337, "perplexity": 30.680173993562292, "lr": 0.0026291804804649314, "grad_norm": 0.174825, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:29.330977+00:00", "epoch": 0, "step": 14010, "train_loss": 3.4983487129211426, "perplexity": 33.0608139646845, "lr": 0.0026291804804649314, "grad_norm": 0.164254, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:29.635838+00:00", "epoch": 0, "step": 14011, "train_loss": 3.486363410949707, "perplexity": 32.666935218713874, "lr": 0.0026291804804649314, "grad_norm": 0.17135, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:29.941148+00:00", "epoch": 0, "step": 14012, "train_loss": 3.4884297847747803, "perplexity": 32.73450710894406, "lr": 0.0026291804804649314, "grad_norm": 0.211502, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:30.246544+00:00", "epoch": 0, "step": 14013, "train_loss": 3.495396137237549, "perplexity": 32.96334337474937, "lr": 0.0026291804804649314, "grad_norm": 0.187478, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:30.551485+00:00", "epoch": 0, "step": 14014, "train_loss": 3.5295372009277344, "perplexity": 34.10817872802311, "lr": 0.0026291804804649314, "grad_norm": 0.189018, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:30.936242+00:00", "epoch": 0, "step": 14015, "train_loss": 3.5887839794158936, "perplexity": 36.19004132322197, "lr": 0.0026291804804649314, "grad_norm": 0.17832, "tokens_per_sec": 85165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:31.240718+00:00", "epoch": 0, "step": 14016, "train_loss": 3.6005969047546387, "perplexity": 36.62008662501545, "lr": 0.0026291804804649314, "grad_norm": 0.197826, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:31.544473+00:00", "epoch": 0, "step": 14017, "train_loss": 3.6604983806610107, "perplexity": 38.880715440136285, "lr": 0.0026291804804649314, "grad_norm": 0.151204, "tokens_per_sec": 107877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:31.849363+00:00", "epoch": 0, "step": 14018, "train_loss": 3.5277180671691895, "perplexity": 34.0461877906552, "lr": 0.0026291804804649314, "grad_norm": 0.182092, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:32.153679+00:00", "epoch": 0, "step": 14019, "train_loss": 3.5053486824035645, "perplexity": 33.293050529644766, "lr": 0.0026291804804649314, "grad_norm": 0.172255, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:32.458672+00:00", "epoch": 0, "step": 14020, "train_loss": 3.56010365486145, "perplexity": 35.16684217034386, "lr": 0.0026291804804649314, "grad_norm": 0.156934, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:32.763334+00:00", "epoch": 0, "step": 14021, "train_loss": 3.475656032562256, "perplexity": 32.319023915490035, "lr": 0.0026291804804649314, "grad_norm": 0.173255, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:33.068006+00:00", "epoch": 0, "step": 14022, "train_loss": 3.5120842456817627, "perplexity": 33.51805489311311, "lr": 0.0026291804804649314, "grad_norm": 0.160248, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:33.372179+00:00", "epoch": 0, "step": 14023, "train_loss": 3.514282464981079, "perplexity": 33.59181597006398, "lr": 0.0026291804804649314, "grad_norm": 0.154028, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:33.677357+00:00", "epoch": 0, "step": 14024, "train_loss": 3.486102819442749, "perplexity": 32.65842360191322, "lr": 0.0026291804804649314, "grad_norm": 0.166153, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:33.981511+00:00", "epoch": 0, "step": 14025, "train_loss": 3.5779430866241455, "perplexity": 35.79982791759976, "lr": 0.0026291804804649314, "grad_norm": 0.161439, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:34.285122+00:00", "epoch": 0, "step": 14026, "train_loss": 3.5719993114471436, "perplexity": 35.5876729144949, "lr": 0.0026291804804649314, "grad_norm": 0.15303, "tokens_per_sec": 107927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:34.589247+00:00", "epoch": 0, "step": 14027, "train_loss": 3.4583706855773926, "perplexity": 31.765178864736413, "lr": 0.0026291804804649314, "grad_norm": 0.191752, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:34.893193+00:00", "epoch": 0, "step": 14028, "train_loss": 3.591948986053467, "perplexity": 36.30476449826155, "lr": 0.0026291804804649314, "grad_norm": 0.170135, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:35.199729+00:00", "epoch": 0, "step": 14029, "train_loss": 3.6058502197265625, "perplexity": 36.812969668458074, "lr": 0.0026291804804649314, "grad_norm": 0.179232, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:35.504976+00:00", "epoch": 0, "step": 14030, "train_loss": 3.468736410140991, "perplexity": 32.09616042781056, "lr": 0.0026291804804649314, "grad_norm": 0.145479, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:35.809310+00:00", "epoch": 0, "step": 14031, "train_loss": 3.502535581588745, "perplexity": 33.19952543151307, "lr": 0.0026291804804649314, "grad_norm": 0.166392, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:36.113927+00:00", "epoch": 0, "step": 14032, "train_loss": 3.4529569149017334, "perplexity": 31.59367413341857, "lr": 0.0026291804804649314, "grad_norm": 0.146823, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:36.419226+00:00", "epoch": 0, "step": 14033, "train_loss": 3.482858419418335, "perplexity": 32.55263830927733, "lr": 0.0026291804804649314, "grad_norm": 0.154272, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:36.725289+00:00", "epoch": 0, "step": 14034, "train_loss": 3.485018730163574, "perplexity": 32.6230381389777, "lr": 0.0026291804804649314, "grad_norm": 0.164032, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:37.030516+00:00", "epoch": 0, "step": 14035, "train_loss": 3.4862422943115234, "perplexity": 32.6629789489303, "lr": 0.0026291804804649314, "grad_norm": 0.152925, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:37.335922+00:00", "epoch": 0, "step": 14036, "train_loss": 3.6669716835021973, "perplexity": 39.133218468414874, "lr": 0.0026291804804649314, "grad_norm": 0.161123, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:37.640706+00:00", "epoch": 0, "step": 14037, "train_loss": 3.474304437637329, "perplexity": 32.27537119383123, "lr": 0.0026291804804649314, "grad_norm": 0.156788, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:37.945353+00:00", "epoch": 0, "step": 14038, "train_loss": 3.520880699157715, "perplexity": 33.81419548751386, "lr": 0.0026291804804649314, "grad_norm": 0.160246, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:38.250263+00:00", "epoch": 0, "step": 14039, "train_loss": 3.537074327468872, "perplexity": 34.36622763922439, "lr": 0.0026291804804649314, "grad_norm": 0.134554, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:38.555876+00:00", "epoch": 0, "step": 14040, "train_loss": 3.636359930038452, "perplexity": 37.95343182933036, "lr": 0.0026291804804649314, "grad_norm": 0.161444, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:38.860636+00:00", "epoch": 0, "step": 14041, "train_loss": 3.450216770172119, "perplexity": 31.50722139435481, "lr": 0.0026291804804649314, "grad_norm": 0.156038, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:39.164750+00:00", "epoch": 0, "step": 14042, "train_loss": 3.5789313316345215, "perplexity": 35.83522440623251, "lr": 0.0026291804804649314, "grad_norm": 0.173552, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:39.468290+00:00", "epoch": 0, "step": 14043, "train_loss": 3.591918468475342, "perplexity": 36.303656581680215, "lr": 0.0026291804804649314, "grad_norm": 0.145629, "tokens_per_sec": 107952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:39.773340+00:00", "epoch": 0, "step": 14044, "train_loss": 3.5052900314331055, "perplexity": 33.291097917183535, "lr": 0.0026291804804649314, "grad_norm": 0.173965, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:40.078000+00:00", "epoch": 0, "step": 14045, "train_loss": 3.5460150241851807, "perplexity": 34.674863310874514, "lr": 0.0026291804804649314, "grad_norm": 0.185957, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:40.383754+00:00", "epoch": 0, "step": 14046, "train_loss": 3.5900919437408447, "perplexity": 36.23740757612302, "lr": 0.0026291804804649314, "grad_norm": 0.20247, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:40.687772+00:00", "epoch": 0, "step": 14047, "train_loss": 3.507113218307495, "perplexity": 33.35184917355863, "lr": 0.0026291804804649314, "grad_norm": 0.201313, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:40.991900+00:00", "epoch": 0, "step": 14048, "train_loss": 3.4231173992156982, "perplexity": 30.664860785440148, "lr": 0.0026291804804649314, "grad_norm": 0.159245, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:41.296610+00:00", "epoch": 0, "step": 14049, "train_loss": 3.608048439025879, "perplexity": 36.89398165728513, "lr": 0.0026291804804649314, "grad_norm": 0.161715, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:41.601391+00:00", "epoch": 0, "step": 14050, "train_loss": 3.4674248695373535, "perplexity": 32.054092603048964, "lr": 0.0026291804804649314, "grad_norm": 0.163764, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:41.905426+00:00", "epoch": 0, "step": 14051, "train_loss": 3.5900871753692627, "perplexity": 36.237234783110495, "lr": 0.0026291804804649314, "grad_norm": 0.16989, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:42.209985+00:00", "epoch": 0, "step": 14052, "train_loss": 3.485811710357666, "perplexity": 32.64891782177482, "lr": 0.0026291804804649314, "grad_norm": 0.162675, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:42.513855+00:00", "epoch": 0, "step": 14053, "train_loss": 3.5470879077911377, "perplexity": 34.71208536715607, "lr": 0.0026291804804649314, "grad_norm": 0.16062, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:42.817700+00:00", "epoch": 0, "step": 14054, "train_loss": 3.5684478282928467, "perplexity": 35.461508062421935, "lr": 0.0026291804804649314, "grad_norm": 0.168635, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:43.122041+00:00", "epoch": 0, "step": 14055, "train_loss": 3.5563766956329346, "perplexity": 35.0360207179173, "lr": 0.0026291804804649314, "grad_norm": 0.194533, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:43.426660+00:00", "epoch": 0, "step": 14056, "train_loss": 3.4578893184661865, "perplexity": 31.749891831981934, "lr": 0.0026291804804649314, "grad_norm": 0.196937, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:43.731788+00:00", "epoch": 0, "step": 14057, "train_loss": 3.5108659267425537, "perplexity": 33.47724407737421, "lr": 0.0026291804804649314, "grad_norm": 0.162816, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:44.037137+00:00", "epoch": 0, "step": 14058, "train_loss": 3.5474534034729004, "perplexity": 34.72477480328854, "lr": 0.0026291804804649314, "grad_norm": 0.217164, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:44.340743+00:00", "epoch": 0, "step": 14059, "train_loss": 3.5531954765319824, "perplexity": 34.92474055672297, "lr": 0.0026291804804649314, "grad_norm": 0.182536, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:44.645506+00:00", "epoch": 0, "step": 14060, "train_loss": 3.53645658493042, "perplexity": 34.34500471435215, "lr": 0.0026291804804649314, "grad_norm": 0.157258, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:44.951163+00:00", "epoch": 0, "step": 14061, "train_loss": 3.5833661556243896, "perplexity": 35.99450023820336, "lr": 0.0026291804804649314, "grad_norm": 0.201722, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:45.256562+00:00", "epoch": 0, "step": 14062, "train_loss": 3.5395314693450928, "perplexity": 34.450774165151344, "lr": 0.0026291804804649314, "grad_norm": 0.175551, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:45.561410+00:00", "epoch": 0, "step": 14063, "train_loss": 3.5356922149658203, "perplexity": 34.318762454989596, "lr": 0.0026291804804649314, "grad_norm": 0.176687, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:45.865581+00:00", "epoch": 0, "step": 14064, "train_loss": 3.495807647705078, "perplexity": 32.97691092699627, "lr": 0.0026291804804649314, "grad_norm": 0.167048, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:46.169824+00:00", "epoch": 0, "step": 14065, "train_loss": 3.488384485244751, "perplexity": 32.73302428474215, "lr": 0.0026291804804649314, "grad_norm": 0.170093, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:46.475679+00:00", "epoch": 0, "step": 14066, "train_loss": 3.627321481704712, "perplexity": 37.61193731175126, "lr": 0.0026291804804649314, "grad_norm": 0.159875, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:46.781088+00:00", "epoch": 0, "step": 14067, "train_loss": 3.6397128105163574, "perplexity": 38.08089872119392, "lr": 0.0026291804804649314, "grad_norm": 0.188224, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:47.085512+00:00", "epoch": 0, "step": 14068, "train_loss": 3.5348823070526123, "perplexity": 34.29097867037803, "lr": 0.0026291804804649314, "grad_norm": 0.163237, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:47.389760+00:00", "epoch": 0, "step": 14069, "train_loss": 3.517197847366333, "perplexity": 33.689891853517366, "lr": 0.0026291804804649314, "grad_norm": 0.165691, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:47.694051+00:00", "epoch": 0, "step": 14070, "train_loss": 3.5019452571868896, "perplexity": 33.17993272511614, "lr": 0.0026291804804649314, "grad_norm": 0.177517, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:47.998584+00:00", "epoch": 0, "step": 14071, "train_loss": 3.511401653289795, "perplexity": 33.49518353064656, "lr": 0.0026291804804649314, "grad_norm": 0.166543, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:48.304732+00:00", "epoch": 0, "step": 14072, "train_loss": 3.423696994781494, "perplexity": 30.68263915441136, "lr": 0.0026291804804649314, "grad_norm": 0.165895, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:48.610078+00:00", "epoch": 0, "step": 14073, "train_loss": 3.532555103302002, "perplexity": 34.21126936207949, "lr": 0.0026291804804649314, "grad_norm": 0.153883, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:48.914468+00:00", "epoch": 0, "step": 14074, "train_loss": 3.520989179611206, "perplexity": 33.817863865744854, "lr": 0.0026291804804649314, "grad_norm": 0.164488, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:49.218364+00:00", "epoch": 0, "step": 14075, "train_loss": 3.5023369789123535, "perplexity": 33.192932571608914, "lr": 0.0026291804804649314, "grad_norm": 0.146762, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:49.523530+00:00", "epoch": 0, "step": 14076, "train_loss": 3.572828769683838, "perplexity": 35.61720364848079, "lr": 0.0026291804804649314, "grad_norm": 0.141359, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:49.828179+00:00", "epoch": 0, "step": 14077, "train_loss": 3.466513156890869, "perplexity": 32.02488179940411, "lr": 0.0026291804804649314, "grad_norm": 0.159746, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:50.134453+00:00", "epoch": 0, "step": 14078, "train_loss": 3.3871281147003174, "perplexity": 29.580877279423348, "lr": 0.0026291804804649314, "grad_norm": 0.151902, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:50.438790+00:00", "epoch": 0, "step": 14079, "train_loss": 3.497136354446411, "perplexity": 33.020756693485396, "lr": 0.0026291804804649314, "grad_norm": 0.171634, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:50.743293+00:00", "epoch": 0, "step": 14080, "train_loss": 3.47119140625, "perplexity": 32.17505317784822, "lr": 0.0026291804804649314, "grad_norm": 0.158728, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:51.048035+00:00", "epoch": 0, "step": 14081, "train_loss": 3.5677602291107178, "perplexity": 35.43713313953004, "lr": 0.0026291804804649314, "grad_norm": 0.169308, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:51.353511+00:00", "epoch": 0, "step": 14082, "train_loss": 3.5290982723236084, "perplexity": 34.09321095787627, "lr": 0.0026291804804649314, "grad_norm": 0.184333, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:51.659501+00:00", "epoch": 0, "step": 14083, "train_loss": 3.482747793197632, "perplexity": 32.54903733311212, "lr": 0.0026291804804649314, "grad_norm": 0.180594, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:51.964907+00:00", "epoch": 0, "step": 14084, "train_loss": 3.4979097843170166, "perplexity": 33.04630581201424, "lr": 0.0026291804804649314, "grad_norm": 0.178457, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:52.270021+00:00", "epoch": 0, "step": 14085, "train_loss": 3.4959864616394043, "perplexity": 32.98280818542133, "lr": 0.0026291804804649314, "grad_norm": 0.159747, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:52.575052+00:00", "epoch": 0, "step": 14086, "train_loss": 3.618241786956787, "perplexity": 37.27197810055552, "lr": 0.0026291804804649314, "grad_norm": 0.189881, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:52.880762+00:00", "epoch": 0, "step": 14087, "train_loss": 3.5427417755126953, "perplexity": 34.561549414026686, "lr": 0.0026291804804649314, "grad_norm": 0.196026, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:53.185654+00:00", "epoch": 0, "step": 14088, "train_loss": 3.5879430770874023, "perplexity": 36.15962182491107, "lr": 0.0026291804804649314, "grad_norm": 0.195719, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:53.489870+00:00", "epoch": 0, "step": 14089, "train_loss": 3.4943933486938477, "perplexity": 32.93030467983207, "lr": 0.0026291804804649314, "grad_norm": 0.170223, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:53.795193+00:00", "epoch": 0, "step": 14090, "train_loss": 3.5125434398651123, "perplexity": 33.53344972329494, "lr": 0.0026291804804649314, "grad_norm": 0.177635, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:54.099662+00:00", "epoch": 0, "step": 14091, "train_loss": 3.585919141769409, "perplexity": 36.08651109986067, "lr": 0.0026291804804649314, "grad_norm": 0.172879, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:54.404966+00:00", "epoch": 0, "step": 14092, "train_loss": 3.5394368171691895, "perplexity": 34.44751347873302, "lr": 0.0026291804804649314, "grad_norm": 0.181862, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:54.709596+00:00", "epoch": 0, "step": 14093, "train_loss": 3.512972593307495, "perplexity": 33.54784380709305, "lr": 0.0026291804804649314, "grad_norm": 0.205654, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:55.015271+00:00", "epoch": 0, "step": 14094, "train_loss": 3.46600079536438, "perplexity": 32.008477684857176, "lr": 0.0026291804804649314, "grad_norm": 0.177385, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:55.319522+00:00", "epoch": 0, "step": 14095, "train_loss": 3.5936880111694336, "perplexity": 36.36795432397535, "lr": 0.0026291804804649314, "grad_norm": 0.186527, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:55.624056+00:00", "epoch": 0, "step": 14096, "train_loss": 3.503293991088867, "perplexity": 33.22471381734809, "lr": 0.0026291804804649314, "grad_norm": 0.212164, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:55.929175+00:00", "epoch": 0, "step": 14097, "train_loss": 3.5371124744415283, "perplexity": 34.36753863177548, "lr": 0.0026291804804649314, "grad_norm": 0.183005, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:56.234341+00:00", "epoch": 0, "step": 14098, "train_loss": 3.3738369941711426, "perplexity": 29.19031552532173, "lr": 0.0026291804804649314, "grad_norm": 0.178096, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:56.539664+00:00", "epoch": 0, "step": 14099, "train_loss": 3.514431953430176, "perplexity": 33.59683793388913, "lr": 0.0026291804804649314, "grad_norm": 0.169058, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:56.844540+00:00", "epoch": 0, "step": 14100, "train_loss": 3.549494504928589, "perplexity": 34.795723974284336, "lr": 0.0026291804804649314, "grad_norm": 0.169148, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:57.148830+00:00", "epoch": 0, "step": 14101, "train_loss": 3.5130178928375244, "perplexity": 33.54936354307242, "lr": 0.0026291804804649314, "grad_norm": 0.156124, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:57.452393+00:00", "epoch": 0, "step": 14102, "train_loss": 3.49434757232666, "perplexity": 32.92879728461525, "lr": 0.0026291804804649314, "grad_norm": 0.17605, "tokens_per_sec": 107946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:57.757652+00:00", "epoch": 0, "step": 14103, "train_loss": 3.532071590423584, "perplexity": 34.19473177114719, "lr": 0.0026291804804649314, "grad_norm": 0.167759, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:58.062991+00:00", "epoch": 0, "step": 14104, "train_loss": 3.397212505340576, "perplexity": 29.880691582873013, "lr": 0.0026291804804649314, "grad_norm": 0.167747, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:58.368516+00:00", "epoch": 0, "step": 14105, "train_loss": 3.5390448570251465, "perplexity": 34.43401407217547, "lr": 0.0026291804804649314, "grad_norm": 0.184091, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:58.672963+00:00", "epoch": 0, "step": 14106, "train_loss": 3.5834600925445557, "perplexity": 35.99788160951397, "lr": 0.0026291804804649314, "grad_norm": 0.170956, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:58.977166+00:00", "epoch": 0, "step": 14107, "train_loss": 3.5325801372528076, "perplexity": 34.212125816033875, "lr": 0.0026291804804649314, "grad_norm": 0.166201, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:59.282361+00:00", "epoch": 0, "step": 14108, "train_loss": 3.4533588886260986, "perplexity": 31.60637651311676, "lr": 0.0026291804804649314, "grad_norm": 0.165075, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:59.587062+00:00", "epoch": 0, "step": 14109, "train_loss": 3.5736594200134277, "perplexity": 35.646801381411954, "lr": 0.0026291804804649314, "grad_norm": 0.209176, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:11:59.892009+00:00", "epoch": 0, "step": 14110, "train_loss": 3.556969404220581, "perplexity": 35.05679302362765, "lr": 0.0026291804804649314, "grad_norm": 0.189127, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:00.196425+00:00", "epoch": 0, "step": 14111, "train_loss": 3.4338269233703613, "perplexity": 30.99503168361595, "lr": 0.0026291804804649314, "grad_norm": 0.179695, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:00.501092+00:00", "epoch": 0, "step": 14112, "train_loss": 3.6062493324279785, "perplexity": 36.82766512460613, "lr": 0.0026291804804649314, "grad_norm": 0.179855, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:00.805609+00:00", "epoch": 0, "step": 14113, "train_loss": 3.590512752532959, "perplexity": 36.252659804745356, "lr": 0.0026291804804649314, "grad_norm": 0.188699, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:01.111631+00:00", "epoch": 0, "step": 14114, "train_loss": 3.572216749191284, "perplexity": 35.59541185915149, "lr": 0.0026291804804649314, "grad_norm": 0.175493, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:01.416729+00:00", "epoch": 0, "step": 14115, "train_loss": 3.6273326873779297, "perplexity": 37.61235878119128, "lr": 0.0026291804804649314, "grad_norm": 0.164954, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:01.722729+00:00", "epoch": 0, "step": 14116, "train_loss": 3.4797098636627197, "perplexity": 32.450305697074455, "lr": 0.0026291804804649314, "grad_norm": 0.19817, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:02.027545+00:00", "epoch": 0, "step": 14117, "train_loss": 3.577054500579834, "perplexity": 35.76803081944447, "lr": 0.0026291804804649314, "grad_norm": 0.179518, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:02.331226+00:00", "epoch": 0, "step": 14118, "train_loss": 3.5100717544555664, "perplexity": 33.45066793229903, "lr": 0.0026291804804649314, "grad_norm": 0.153199, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:02.636349+00:00", "epoch": 0, "step": 14119, "train_loss": 3.4827215671539307, "perplexity": 32.5481837118302, "lr": 0.0026291804804649314, "grad_norm": 0.16983, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:02.943262+00:00", "epoch": 0, "step": 14120, "train_loss": 3.510244607925415, "perplexity": 33.456450496073586, "lr": 0.0026291804804649314, "grad_norm": 0.181086, "tokens_per_sec": 106825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:03.247789+00:00", "epoch": 0, "step": 14121, "train_loss": 3.603452444076538, "perplexity": 36.72480616657106, "lr": 0.0026291804804649314, "grad_norm": 0.165065, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:03.552570+00:00", "epoch": 0, "step": 14122, "train_loss": 3.5401363372802734, "perplexity": 34.47161863722686, "lr": 0.0026291804804649314, "grad_norm": 0.184245, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:03.857741+00:00", "epoch": 0, "step": 14123, "train_loss": 3.5452115535736084, "perplexity": 34.64701426668642, "lr": 0.0026291804804649314, "grad_norm": 0.164648, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:04.162462+00:00", "epoch": 0, "step": 14124, "train_loss": 3.502474308013916, "perplexity": 33.19749124022897, "lr": 0.0026291804804649314, "grad_norm": 0.177945, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:04.467095+00:00", "epoch": 0, "step": 14125, "train_loss": 3.4653613567352295, "perplexity": 31.98801677021196, "lr": 0.0026291804804649314, "grad_norm": 0.15466, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:04.771699+00:00", "epoch": 0, "step": 14126, "train_loss": 3.627164125442505, "perplexity": 37.60601930351156, "lr": 0.0026291804804649314, "grad_norm": 0.180703, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:05.076281+00:00", "epoch": 0, "step": 14127, "train_loss": 3.612271785736084, "perplexity": 37.05012722969571, "lr": 0.0026291804804649314, "grad_norm": 0.163922, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:05.380584+00:00", "epoch": 0, "step": 14128, "train_loss": 3.6218817234039307, "perplexity": 37.407892942731905, "lr": 0.0026291804804649314, "grad_norm": 0.171404, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:05.684398+00:00", "epoch": 0, "step": 14129, "train_loss": 3.4777605533599854, "perplexity": 32.387111594320515, "lr": 0.0026291804804649314, "grad_norm": 0.168337, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:05.989874+00:00", "epoch": 0, "step": 14130, "train_loss": 3.5365891456604004, "perplexity": 34.34955781502278, "lr": 0.0026291804804649314, "grad_norm": 0.169867, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:06.294933+00:00", "epoch": 0, "step": 14131, "train_loss": 3.4321913719177246, "perplexity": 30.94437914822662, "lr": 0.0026291804804649314, "grad_norm": 0.186009, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:06.600651+00:00", "epoch": 0, "step": 14132, "train_loss": 3.451085090637207, "perplexity": 31.534591640842443, "lr": 0.0026291804804649314, "grad_norm": 0.134338, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:06.905770+00:00", "epoch": 0, "step": 14133, "train_loss": 3.492246627807617, "perplexity": 32.85968833090061, "lr": 0.0026291804804649314, "grad_norm": 0.181757, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:07.210030+00:00", "epoch": 0, "step": 14134, "train_loss": 3.5092525482177734, "perplexity": 33.423276157758714, "lr": 0.0026291804804649314, "grad_norm": 0.157995, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:07.513789+00:00", "epoch": 0, "step": 14135, "train_loss": 3.424246311187744, "perplexity": 30.699498261554062, "lr": 0.0026291804804649314, "grad_norm": 0.17285, "tokens_per_sec": 107875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:07.819098+00:00", "epoch": 0, "step": 14136, "train_loss": 3.5047290325164795, "perplexity": 33.27242688502279, "lr": 0.0026291804804649314, "grad_norm": 0.177658, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:08.124737+00:00", "epoch": 0, "step": 14137, "train_loss": 3.5907809734344482, "perplexity": 36.26238483000879, "lr": 0.0026291804804649314, "grad_norm": 0.179826, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:08.429997+00:00", "epoch": 0, "step": 14138, "train_loss": 3.5532310009002686, "perplexity": 34.92598125810624, "lr": 0.0026291804804649314, "grad_norm": 0.203011, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:08.735115+00:00", "epoch": 0, "step": 14139, "train_loss": 3.490427017211914, "perplexity": 32.799950859792375, "lr": 0.0026291804804649314, "grad_norm": 0.183145, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:09.040812+00:00", "epoch": 0, "step": 14140, "train_loss": 3.510077953338623, "perplexity": 33.4508752897204, "lr": 0.0026291804804649314, "grad_norm": 0.175437, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:09.346911+00:00", "epoch": 0, "step": 14141, "train_loss": 3.5816519260406494, "perplexity": 35.93285025725064, "lr": 0.0026291804804649314, "grad_norm": 0.164854, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:09.652904+00:00", "epoch": 0, "step": 14142, "train_loss": 3.5321807861328125, "perplexity": 34.198465893006144, "lr": 0.0026291804804649314, "grad_norm": 0.171174, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:09.959002+00:00", "epoch": 0, "step": 14143, "train_loss": 3.4657552242279053, "perplexity": 32.00061829167484, "lr": 0.0026291804804649314, "grad_norm": 0.149549, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:10.264314+00:00", "epoch": 0, "step": 14144, "train_loss": 3.4708399772644043, "perplexity": 32.16374791816276, "lr": 0.0026291804804649314, "grad_norm": 0.159273, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:10.568465+00:00", "epoch": 0, "step": 14145, "train_loss": 3.49476957321167, "perplexity": 32.94269619868242, "lr": 0.0026291804804649314, "grad_norm": 0.13659, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:10.874847+00:00", "epoch": 0, "step": 14146, "train_loss": 3.5182316303253174, "perplexity": 33.72473789812576, "lr": 0.0026291804804649314, "grad_norm": 0.153075, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:11.181599+00:00", "epoch": 0, "step": 14147, "train_loss": 3.517143726348877, "perplexity": 33.68806857163165, "lr": 0.0026291804804649314, "grad_norm": 0.135911, "tokens_per_sec": 106823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:11.487411+00:00", "epoch": 0, "step": 14148, "train_loss": 3.4799294471740723, "perplexity": 32.457432031528015, "lr": 0.0026291804804649314, "grad_norm": 0.152434, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:11.792059+00:00", "epoch": 0, "step": 14149, "train_loss": 3.503420114517212, "perplexity": 33.22890449642636, "lr": 0.0026291804804649314, "grad_norm": 0.156502, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:12.097543+00:00", "epoch": 0, "step": 14150, "train_loss": 3.550403594970703, "perplexity": 34.8273708031865, "lr": 0.0026291804804649314, "grad_norm": 0.16285, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:12.402133+00:00", "epoch": 0, "step": 14151, "train_loss": 3.370803117752075, "perplexity": 29.101889919390082, "lr": 0.0026291804804649314, "grad_norm": 0.163558, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:12.707207+00:00", "epoch": 0, "step": 14152, "train_loss": 3.564460277557373, "perplexity": 35.32038505443829, "lr": 0.0026291804804649314, "grad_norm": 0.175095, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:13.012827+00:00", "epoch": 0, "step": 14153, "train_loss": 3.602928400039673, "perplexity": 36.70556579273634, "lr": 0.0026291804804649314, "grad_norm": 0.171359, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:13.317263+00:00", "epoch": 0, "step": 14154, "train_loss": 3.482325553894043, "perplexity": 32.53529675136708, "lr": 0.0026291804804649314, "grad_norm": 0.223642, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:13.622398+00:00", "epoch": 0, "step": 14155, "train_loss": 3.5006964206695557, "perplexity": 33.13852227630861, "lr": 0.0026291804804649314, "grad_norm": 0.184368, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:13.927037+00:00", "epoch": 0, "step": 14156, "train_loss": 3.5402841567993164, "perplexity": 34.47671459194596, "lr": 0.0026291804804649314, "grad_norm": 0.168956, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:14.232376+00:00", "epoch": 0, "step": 14157, "train_loss": 3.50285267829895, "perplexity": 33.21005456109572, "lr": 0.0026291804804649314, "grad_norm": 0.181761, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:14.537875+00:00", "epoch": 0, "step": 14158, "train_loss": 3.5063843727111816, "perplexity": 33.32754968152414, "lr": 0.0026291804804649314, "grad_norm": 0.19111, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:14.842202+00:00", "epoch": 0, "step": 14159, "train_loss": 3.5810556411743164, "perplexity": 35.91143042923234, "lr": 0.0026291804804649314, "grad_norm": 0.147147, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:15.146859+00:00", "epoch": 0, "step": 14160, "train_loss": 3.5501952171325684, "perplexity": 34.82011430702336, "lr": 0.0026291804804649314, "grad_norm": 0.204148, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:15.451733+00:00", "epoch": 0, "step": 14161, "train_loss": 3.5813302993774414, "perplexity": 35.921295152837295, "lr": 0.0026291804804649314, "grad_norm": 0.180747, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:15.756546+00:00", "epoch": 0, "step": 14162, "train_loss": 3.417003870010376, "perplexity": 30.477962150696925, "lr": 0.0026291804804649314, "grad_norm": 0.197302, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:16.061480+00:00", "epoch": 0, "step": 14163, "train_loss": 3.6156530380249023, "perplexity": 37.1756150906936, "lr": 0.0026291804804649314, "grad_norm": 0.163611, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:16.368225+00:00", "epoch": 0, "step": 14164, "train_loss": 3.647033214569092, "perplexity": 38.360689126475755, "lr": 0.0026291804804649314, "grad_norm": 0.172806, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:16.673342+00:00", "epoch": 0, "step": 14165, "train_loss": 3.5024843215942383, "perplexity": 33.19782366763839, "lr": 0.0026291804804649314, "grad_norm": 0.172207, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:16.977639+00:00", "epoch": 0, "step": 14166, "train_loss": 3.484018564224243, "perplexity": 32.590425998888044, "lr": 0.0026291804804649314, "grad_norm": 0.15966, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:17.283417+00:00", "epoch": 0, "step": 14167, "train_loss": 3.4818897247314453, "perplexity": 32.521120009772176, "lr": 0.0026291804804649314, "grad_norm": 0.162709, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:17.589423+00:00", "epoch": 0, "step": 14168, "train_loss": 3.5331268310546875, "perplexity": 34.230834486657805, "lr": 0.0026291804804649314, "grad_norm": 0.151335, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:17.894880+00:00", "epoch": 0, "step": 14169, "train_loss": 3.4335758686065674, "perplexity": 30.987251209961187, "lr": 0.0026291804804649314, "grad_norm": 0.147956, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:18.199900+00:00", "epoch": 0, "step": 14170, "train_loss": 3.5830914974212646, "perplexity": 35.98461541098213, "lr": 0.0026291804804649314, "grad_norm": 0.157433, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:18.504409+00:00", "epoch": 0, "step": 14171, "train_loss": 3.4354710578918457, "perplexity": 31.04603360071883, "lr": 0.0026291804804649314, "grad_norm": 0.163663, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:18.810051+00:00", "epoch": 0, "step": 14172, "train_loss": 3.519918203353882, "perplexity": 33.7816651239277, "lr": 0.0026291804804649314, "grad_norm": 0.154514, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:19.115539+00:00", "epoch": 0, "step": 14173, "train_loss": 3.5624146461486816, "perplexity": 35.24820641601869, "lr": 0.0026291804804649314, "grad_norm": 0.156305, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:19.419977+00:00", "epoch": 0, "step": 14174, "train_loss": 3.5308587551116943, "perplexity": 34.153284332513785, "lr": 0.0026291804804649314, "grad_norm": 0.174477, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:19.723560+00:00", "epoch": 0, "step": 14175, "train_loss": 3.535125970840454, "perplexity": 34.29933515817505, "lr": 0.0026291804804649314, "grad_norm": 0.19298, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:20.029003+00:00", "epoch": 0, "step": 14176, "train_loss": 3.6741809844970703, "perplexity": 39.41636102286452, "lr": 0.0026291804804649314, "grad_norm": 0.177334, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:20.334895+00:00", "epoch": 0, "step": 14177, "train_loss": 3.4824469089508057, "perplexity": 32.53924531373529, "lr": 0.0026291804804649314, "grad_norm": 0.158408, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:20.641315+00:00", "epoch": 0, "step": 14178, "train_loss": 3.4423959255218506, "perplexity": 31.26176938259506, "lr": 0.0026291804804649314, "grad_norm": 0.165378, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:20.946274+00:00", "epoch": 0, "step": 14179, "train_loss": 3.5686304569244385, "perplexity": 35.46798494052725, "lr": 0.0026291804804649314, "grad_norm": 0.172918, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:21.251348+00:00", "epoch": 0, "step": 14180, "train_loss": 3.577127695083618, "perplexity": 35.77064893852643, "lr": 0.0026291804804649314, "grad_norm": 0.175324, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:21.556168+00:00", "epoch": 0, "step": 14181, "train_loss": 3.4750685691833496, "perplexity": 32.300043248267016, "lr": 0.0026291804804649314, "grad_norm": 0.15801, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:21.861639+00:00", "epoch": 0, "step": 14182, "train_loss": 3.4117870330810547, "perplexity": 30.319377606732886, "lr": 0.0026291804804649314, "grad_norm": 0.163698, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:22.166181+00:00", "epoch": 0, "step": 14183, "train_loss": 3.656792640686035, "perplexity": 38.73690025409465, "lr": 0.0026291804804649314, "grad_norm": 0.205331, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:22.470859+00:00", "epoch": 0, "step": 14184, "train_loss": 3.5142886638641357, "perplexity": 33.59202420244824, "lr": 0.0026291804804649314, "grad_norm": 0.165103, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:22.774974+00:00", "epoch": 0, "step": 14185, "train_loss": 3.503160238265991, "perplexity": 33.22027021526456, "lr": 0.0026291804804649314, "grad_norm": 0.158669, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:23.080000+00:00", "epoch": 0, "step": 14186, "train_loss": 3.4309990406036377, "perplexity": 30.90750518333322, "lr": 0.0026291804804649314, "grad_norm": 0.17346, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:23.385717+00:00", "epoch": 0, "step": 14187, "train_loss": 3.658379554748535, "perplexity": 38.798421187151774, "lr": 0.0026291804804649314, "grad_norm": 0.199591, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:23.691148+00:00", "epoch": 0, "step": 14188, "train_loss": 3.6382884979248047, "perplexity": 38.02669822603443, "lr": 0.0026291804804649314, "grad_norm": 0.201484, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:23.996386+00:00", "epoch": 0, "step": 14189, "train_loss": 3.447366952896118, "perplexity": 31.417559391351148, "lr": 0.0026291804804649314, "grad_norm": 0.189013, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:24.300866+00:00", "epoch": 0, "step": 14190, "train_loss": 3.5734052658081055, "perplexity": 35.63774274812821, "lr": 0.0026291804804649314, "grad_norm": 0.165441, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:24.606118+00:00", "epoch": 0, "step": 14191, "train_loss": 3.461684465408325, "perplexity": 31.87061627548172, "lr": 0.0026291804804649314, "grad_norm": 0.189802, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:24.912962+00:00", "epoch": 0, "step": 14192, "train_loss": 3.5373761653900146, "perplexity": 34.37660223557706, "lr": 0.0026291804804649314, "grad_norm": 0.177676, "tokens_per_sec": 106792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:25.218421+00:00", "epoch": 0, "step": 14193, "train_loss": 3.4793286323547363, "perplexity": 32.43793698240953, "lr": 0.0026291804804649314, "grad_norm": 0.172323, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:25.524166+00:00", "epoch": 0, "step": 14194, "train_loss": 3.5742621421813965, "perplexity": 35.66829297490101, "lr": 0.0026291804804649314, "grad_norm": 0.194376, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:25.828449+00:00", "epoch": 0, "step": 14195, "train_loss": 3.6000125408172607, "perplexity": 36.598693418326164, "lr": 0.0026291804804649314, "grad_norm": 0.177226, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:26.132853+00:00", "epoch": 0, "step": 14196, "train_loss": 3.533721685409546, "perplexity": 34.25120290514324, "lr": 0.0026291804804649314, "grad_norm": 0.159848, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:26.437948+00:00", "epoch": 0, "step": 14197, "train_loss": 3.4834275245666504, "perplexity": 32.57116945591571, "lr": 0.0026291804804649314, "grad_norm": 0.165199, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:26.743594+00:00", "epoch": 0, "step": 14198, "train_loss": 3.4395902156829834, "perplexity": 31.174180860105693, "lr": 0.0026291804804649314, "grad_norm": 0.190554, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:27.049242+00:00", "epoch": 0, "step": 14199, "train_loss": 3.4666943550109863, "perplexity": 32.0306851735475, "lr": 0.0026291804804649314, "grad_norm": 0.187075, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:27.353941+00:00", "epoch": 0, "step": 14200, "train_loss": 3.4744198322296143, "perplexity": 32.279095812027265, "lr": 0.0026291804804649314, "grad_norm": 0.177933, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:27.659951+00:00", "epoch": 0, "step": 14201, "train_loss": 3.5516159534454346, "perplexity": 34.86961966654129, "lr": 0.0026291804804649314, "grad_norm": 0.153131, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:27.964503+00:00", "epoch": 0, "step": 14202, "train_loss": 3.600149154663086, "perplexity": 36.603693648128754, "lr": 0.0026291804804649314, "grad_norm": 0.154202, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:28.269696+00:00", "epoch": 0, "step": 14203, "train_loss": 3.6766068935394287, "perplexity": 39.51209760665782, "lr": 0.0026291804804649314, "grad_norm": 0.159943, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:28.574473+00:00", "epoch": 0, "step": 14204, "train_loss": 3.485147476196289, "perplexity": 32.62723849609706, "lr": 0.0026291804804649314, "grad_norm": 0.151429, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:28.878642+00:00", "epoch": 0, "step": 14205, "train_loss": 3.5029895305633545, "perplexity": 33.21459974326554, "lr": 0.0026291804804649314, "grad_norm": 0.150257, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:29.183363+00:00", "epoch": 0, "step": 14206, "train_loss": 3.511094808578491, "perplexity": 33.484907287412106, "lr": 0.0026291804804649314, "grad_norm": 0.162459, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:29.489504+00:00", "epoch": 0, "step": 14207, "train_loss": 3.601057291030884, "perplexity": 36.63694989184297, "lr": 0.0026291804804649314, "grad_norm": 0.156299, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:29.794738+00:00", "epoch": 0, "step": 14208, "train_loss": 3.608468532562256, "perplexity": 36.90948383646521, "lr": 0.0026291804804649314, "grad_norm": 0.162701, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:30.098718+00:00", "epoch": 0, "step": 14209, "train_loss": 3.5027565956115723, "perplexity": 33.206863803096276, "lr": 0.0026291804804649314, "grad_norm": 0.16255, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:30.405220+00:00", "epoch": 0, "step": 14210, "train_loss": 3.457671880722046, "perplexity": 31.74298895762519, "lr": 0.0026291804804649314, "grad_norm": 0.15822, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:30.709288+00:00", "epoch": 0, "step": 14211, "train_loss": 3.4769515991210938, "perplexity": 32.3609224974295, "lr": 0.0026291804804649314, "grad_norm": 0.186404, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:31.015254+00:00", "epoch": 0, "step": 14212, "train_loss": 3.6227991580963135, "perplexity": 37.44222798915961, "lr": 0.0026291804804649314, "grad_norm": 0.170118, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:31.320211+00:00", "epoch": 0, "step": 14213, "train_loss": 3.6553795337677, "perplexity": 38.68219953043878, "lr": 0.0026291804804649314, "grad_norm": 0.172064, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:31.624773+00:00", "epoch": 0, "step": 14214, "train_loss": 3.705890655517578, "perplexity": 40.6862686337748, "lr": 0.0026291804804649314, "grad_norm": 0.16314, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:31.929884+00:00", "epoch": 0, "step": 14215, "train_loss": 3.5687708854675293, "perplexity": 35.472966007712635, "lr": 0.0026291804804649314, "grad_norm": 0.173975, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:32.234770+00:00", "epoch": 0, "step": 14216, "train_loss": 3.4586620330810547, "perplexity": 31.774434918600654, "lr": 0.0026291804804649314, "grad_norm": 0.174276, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:32.539632+00:00", "epoch": 0, "step": 14217, "train_loss": 3.533243417739868, "perplexity": 34.234825578831156, "lr": 0.0026291804804649314, "grad_norm": 0.162807, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:32.845155+00:00", "epoch": 0, "step": 14218, "train_loss": 3.515456199645996, "perplexity": 33.6312669969153, "lr": 0.0026291804804649314, "grad_norm": 0.169038, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:33.150191+00:00", "epoch": 0, "step": 14219, "train_loss": 3.513305902481079, "perplexity": 33.559027474893924, "lr": 0.0026291804804649314, "grad_norm": 0.183334, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:33.455223+00:00", "epoch": 0, "step": 14220, "train_loss": 3.4213132858276367, "perplexity": 30.609587773936116, "lr": 0.0026291804804649314, "grad_norm": 0.202019, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:33.759315+00:00", "epoch": 0, "step": 14221, "train_loss": 3.5746922492980957, "perplexity": 35.68363746119986, "lr": 0.0026291804804649314, "grad_norm": 0.169169, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:34.064249+00:00", "epoch": 0, "step": 14222, "train_loss": 3.5086512565612793, "perplexity": 33.40318506158194, "lr": 0.0026291804804649314, "grad_norm": 0.175841, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:34.370035+00:00", "epoch": 0, "step": 14223, "train_loss": 3.5544633865356445, "perplexity": 34.96905006895107, "lr": 0.0026291804804649314, "grad_norm": 0.167878, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:34.675678+00:00", "epoch": 0, "step": 14224, "train_loss": 3.4833178520202637, "perplexity": 32.56759748869963, "lr": 0.0026291804804649314, "grad_norm": 0.1646, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:34.980344+00:00", "epoch": 0, "step": 14225, "train_loss": 3.4414167404174805, "perplexity": 31.23117330573536, "lr": 0.0026291804804649314, "grad_norm": 0.181302, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:35.284707+00:00", "epoch": 0, "step": 14226, "train_loss": 3.456183671951294, "perplexity": 31.69578389725335, "lr": 0.0026291804804649314, "grad_norm": 0.189699, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:35.589985+00:00", "epoch": 0, "step": 14227, "train_loss": 3.4448533058166504, "perplexity": 31.33868590652512, "lr": 0.0026291804804649314, "grad_norm": 0.142539, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:35.895762+00:00", "epoch": 0, "step": 14228, "train_loss": 3.4656529426574707, "perplexity": 31.99734538556293, "lr": 0.0026291804804649314, "grad_norm": 0.175676, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:36.201121+00:00", "epoch": 0, "step": 14229, "train_loss": 3.4639840126037598, "perplexity": 31.943988590973703, "lr": 0.0026291804804649314, "grad_norm": 0.172879, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:36.508032+00:00", "epoch": 0, "step": 14230, "train_loss": 3.421103000640869, "perplexity": 30.6031517077826, "lr": 0.0026291804804649314, "grad_norm": 0.163994, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:36.813545+00:00", "epoch": 0, "step": 14231, "train_loss": 3.484354257583618, "perplexity": 32.60136822498863, "lr": 0.0026291804804649314, "grad_norm": 0.187798, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:37.118475+00:00", "epoch": 0, "step": 14232, "train_loss": 3.5615110397338867, "perplexity": 35.216370296418305, "lr": 0.0026291804804649314, "grad_norm": 0.183177, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:37.423214+00:00", "epoch": 0, "step": 14233, "train_loss": 3.541584014892578, "perplexity": 34.52155856752552, "lr": 0.0026291804804649314, "grad_norm": 0.165661, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:37.729308+00:00", "epoch": 0, "step": 14234, "train_loss": 3.477818012237549, "perplexity": 32.38897257486467, "lr": 0.0026291804804649314, "grad_norm": 0.196382, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:38.036279+00:00", "epoch": 0, "step": 14235, "train_loss": 3.5971648693084717, "perplexity": 36.49462061472792, "lr": 0.0026291804804649314, "grad_norm": 0.160116, "tokens_per_sec": 106746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:38.341335+00:00", "epoch": 0, "step": 14236, "train_loss": 3.53871750831604, "perplexity": 34.422743986844175, "lr": 0.0026291804804649314, "grad_norm": 0.178042, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:38.645588+00:00", "epoch": 0, "step": 14237, "train_loss": 3.5610923767089844, "perplexity": 35.20162959021356, "lr": 0.0026291804804649314, "grad_norm": 0.207788, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:38.950121+00:00", "epoch": 0, "step": 14238, "train_loss": 3.391261339187622, "perplexity": 29.703394707347943, "lr": 0.0026291804804649314, "grad_norm": 0.156777, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:39.255101+00:00", "epoch": 0, "step": 14239, "train_loss": 3.5800812244415283, "perplexity": 35.87645477372314, "lr": 0.0026291804804649314, "grad_norm": 0.181056, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:39.561025+00:00", "epoch": 0, "step": 14240, "train_loss": 3.4750611782073975, "perplexity": 32.29980452030633, "lr": 0.0026291804804649314, "grad_norm": 0.186255, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:39.866041+00:00", "epoch": 0, "step": 14241, "train_loss": 3.4920003414154053, "perplexity": 32.85159643331542, "lr": 0.0026291804804649314, "grad_norm": 0.183409, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:40.171690+00:00", "epoch": 0, "step": 14242, "train_loss": 3.4484102725982666, "perplexity": 31.450355055263834, "lr": 0.0026291804804649314, "grad_norm": 0.172404, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:40.475848+00:00", "epoch": 0, "step": 14243, "train_loss": 3.519791841506958, "perplexity": 33.77739668002039, "lr": 0.0026291804804649314, "grad_norm": 0.161287, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:40.780372+00:00", "epoch": 0, "step": 14244, "train_loss": 3.547556161880493, "perplexity": 34.72834324919211, "lr": 0.0026291804804649314, "grad_norm": 0.194806, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:41.085598+00:00", "epoch": 0, "step": 14245, "train_loss": 3.55035400390625, "perplexity": 34.82564371962057, "lr": 0.0026291804804649314, "grad_norm": 0.196447, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:41.390935+00:00", "epoch": 0, "step": 14246, "train_loss": 3.4971518516540527, "perplexity": 33.02126842697356, "lr": 0.0026291804804649314, "grad_norm": 0.167582, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:41.695668+00:00", "epoch": 0, "step": 14247, "train_loss": 3.475677251815796, "perplexity": 32.319709708328645, "lr": 0.0026291804804649314, "grad_norm": 0.189105, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:42.000436+00:00", "epoch": 0, "step": 14248, "train_loss": 3.4420018196105957, "perplexity": 31.249451361946903, "lr": 0.0026291804804649314, "grad_norm": 0.199605, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:42.306066+00:00", "epoch": 0, "step": 14249, "train_loss": 3.514861822128296, "perplexity": 33.61128326744865, "lr": 0.0026291804804649314, "grad_norm": 0.150357, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:42.612018+00:00", "epoch": 0, "step": 14250, "train_loss": 3.5086638927459717, "perplexity": 33.40360715306449, "lr": 0.0026291804804649314, "grad_norm": 0.168782, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:42.917013+00:00", "epoch": 0, "step": 14251, "train_loss": 3.534926414489746, "perplexity": 34.29249119092045, "lr": 0.0026291804804649314, "grad_norm": 0.168383, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:43.222544+00:00", "epoch": 0, "step": 14252, "train_loss": 3.4479360580444336, "perplexity": 31.43544437488366, "lr": 0.0026291804804649314, "grad_norm": 0.217751, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:43.527776+00:00", "epoch": 0, "step": 14253, "train_loss": 3.4989335536956787, "perplexity": 33.08015493187224, "lr": 0.0026291804804649314, "grad_norm": 0.182281, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:43.832473+00:00", "epoch": 0, "step": 14254, "train_loss": 3.3451292514801025, "perplexity": 28.364241549940346, "lr": 0.0026291804804649314, "grad_norm": 0.173226, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:44.138092+00:00", "epoch": 0, "step": 14255, "train_loss": 3.5311813354492188, "perplexity": 34.16430328765492, "lr": 0.0026291804804649314, "grad_norm": 0.195998, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:44.444123+00:00", "epoch": 0, "step": 14256, "train_loss": 3.5096521377563477, "perplexity": 33.436634417989005, "lr": 0.0026291804804649314, "grad_norm": 0.162034, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:44.748928+00:00", "epoch": 0, "step": 14257, "train_loss": 3.533466339111328, "perplexity": 34.24245810379574, "lr": 0.0026291804804649314, "grad_norm": 0.1646, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:45.055508+00:00", "epoch": 0, "step": 14258, "train_loss": 3.5846314430236816, "perplexity": 36.04007245069177, "lr": 0.0026291804804649314, "grad_norm": 0.153111, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:45.360038+00:00", "epoch": 0, "step": 14259, "train_loss": 3.495767116546631, "perplexity": 32.97557436168084, "lr": 0.0026291804804649314, "grad_norm": 0.160861, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:45.664651+00:00", "epoch": 0, "step": 14260, "train_loss": 3.4901509284973145, "perplexity": 32.7908964134933, "lr": 0.0026291804804649314, "grad_norm": 0.146109, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:45.970392+00:00", "epoch": 0, "step": 14261, "train_loss": 3.513622999191284, "perplexity": 33.56967061946778, "lr": 0.0026291804804649314, "grad_norm": 0.174279, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:46.276658+00:00", "epoch": 0, "step": 14262, "train_loss": 3.6011178493499756, "perplexity": 36.639168631125955, "lr": 0.0026291804804649314, "grad_norm": 0.17392, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:46.581089+00:00", "epoch": 0, "step": 14263, "train_loss": 3.4319441318511963, "perplexity": 30.93672940356799, "lr": 0.0026291804804649314, "grad_norm": 0.163495, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:46.886839+00:00", "epoch": 0, "step": 14264, "train_loss": 3.51003098487854, "perplexity": 33.44930419051595, "lr": 0.0026291804804649314, "grad_norm": 0.169432, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:47.192370+00:00", "epoch": 0, "step": 14265, "train_loss": 3.3709118366241455, "perplexity": 29.105054016032625, "lr": 0.0026291804804649314, "grad_norm": 0.174357, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:47.498141+00:00", "epoch": 0, "step": 14266, "train_loss": 3.5234038829803467, "perplexity": 33.899622647434875, "lr": 0.0026291804804649314, "grad_norm": 0.184147, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:47.803046+00:00", "epoch": 0, "step": 14267, "train_loss": 3.5196707248687744, "perplexity": 33.77330592302228, "lr": 0.0026291804804649314, "grad_norm": 0.181042, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:48.107793+00:00", "epoch": 0, "step": 14268, "train_loss": 3.4932799339294434, "perplexity": 32.893659996556586, "lr": 0.0026291804804649314, "grad_norm": 0.155248, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:48.412775+00:00", "epoch": 0, "step": 14269, "train_loss": 3.45454478263855, "perplexity": 31.643880559295546, "lr": 0.0026291804804649314, "grad_norm": 0.189802, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:48.717231+00:00", "epoch": 0, "step": 14270, "train_loss": 3.5346784591674805, "perplexity": 34.28398923931155, "lr": 0.0026291804804649314, "grad_norm": 0.171416, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:49.022039+00:00", "epoch": 0, "step": 14271, "train_loss": 3.4998421669006348, "perplexity": 33.11022565672536, "lr": 0.0026291804804649314, "grad_norm": 0.146242, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:49.328532+00:00", "epoch": 0, "step": 14272, "train_loss": 3.508751392364502, "perplexity": 33.40653008382372, "lr": 0.0026291804804649314, "grad_norm": 0.157876, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:49.632825+00:00", "epoch": 0, "step": 14273, "train_loss": 3.4771618843078613, "perplexity": 32.36772823560873, "lr": 0.0026291804804649314, "grad_norm": 0.156773, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:49.937237+00:00", "epoch": 0, "step": 14274, "train_loss": 3.541900634765625, "perplexity": 34.532490509560276, "lr": 0.0026291804804649314, "grad_norm": 0.148588, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:50.241754+00:00", "epoch": 0, "step": 14275, "train_loss": 3.5839619636535645, "perplexity": 36.015952440514, "lr": 0.0026291804804649314, "grad_norm": 0.175621, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:50.547708+00:00", "epoch": 0, "step": 14276, "train_loss": 3.4181480407714844, "perplexity": 30.512854101216213, "lr": 0.0026291804804649314, "grad_norm": 0.165589, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:50.853231+00:00", "epoch": 0, "step": 14277, "train_loss": 3.4455599784851074, "perplexity": 31.360839926218777, "lr": 0.0026291804804649314, "grad_norm": 0.167897, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:51.159122+00:00", "epoch": 0, "step": 14278, "train_loss": 3.586947441101074, "perplexity": 36.12363792057776, "lr": 0.0026291804804649314, "grad_norm": 0.189116, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:51.464701+00:00", "epoch": 0, "step": 14279, "train_loss": 3.47920298576355, "perplexity": 32.43386152224177, "lr": 0.0026291804804649314, "grad_norm": 0.162205, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:51.770498+00:00", "epoch": 0, "step": 14280, "train_loss": 3.4123425483703613, "perplexity": 30.33622516366016, "lr": 0.0026291804804649314, "grad_norm": 0.153491, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:52.074659+00:00", "epoch": 0, "step": 14281, "train_loss": 3.457775115966797, "perplexity": 31.746266122015882, "lr": 0.0026291804804649314, "grad_norm": 0.149488, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:52.379172+00:00", "epoch": 0, "step": 14282, "train_loss": 3.4722483158111572, "perplexity": 32.209077276212824, "lr": 0.0026291804804649314, "grad_norm": 0.15743, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:52.685514+00:00", "epoch": 0, "step": 14283, "train_loss": 3.634291172027588, "perplexity": 37.87499652300148, "lr": 0.0026291804804649314, "grad_norm": 0.145203, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:52.990395+00:00", "epoch": 0, "step": 14284, "train_loss": 3.4011948108673096, "perplexity": 29.99992287625376, "lr": 0.0026291804804649314, "grad_norm": 0.163336, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:53.295404+00:00", "epoch": 0, "step": 14285, "train_loss": 3.534524440765381, "perplexity": 34.278709280687146, "lr": 0.0026291804804649314, "grad_norm": 0.155204, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:53.599730+00:00", "epoch": 0, "step": 14286, "train_loss": 3.4924449920654297, "perplexity": 32.8662071651188, "lr": 0.0026291804804649314, "grad_norm": 0.16029, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:53.903509+00:00", "epoch": 0, "step": 14287, "train_loss": 3.4634053707122803, "perplexity": 31.925509807808645, "lr": 0.0026291804804649314, "grad_norm": 0.187771, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:54.209179+00:00", "epoch": 0, "step": 14288, "train_loss": 3.592231273651123, "perplexity": 36.31501432964731, "lr": 0.0026291804804649314, "grad_norm": 0.158882, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:54.514530+00:00", "epoch": 0, "step": 14289, "train_loss": 3.4707934856414795, "perplexity": 32.162252608082674, "lr": 0.0026291804804649314, "grad_norm": 0.164231, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:54.819738+00:00", "epoch": 0, "step": 14290, "train_loss": 3.524961233139038, "perplexity": 33.95245736054034, "lr": 0.0026291804804649314, "grad_norm": 0.18778, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:55.124392+00:00", "epoch": 0, "step": 14291, "train_loss": 3.4982223510742188, "perplexity": 33.05663660310624, "lr": 0.0026291804804649314, "grad_norm": 0.187226, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:55.428954+00:00", "epoch": 0, "step": 14292, "train_loss": 3.5703721046447754, "perplexity": 35.5298115000587, "lr": 0.0026291804804649314, "grad_norm": 0.230879, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:55.734454+00:00", "epoch": 0, "step": 14293, "train_loss": 3.6380791664123535, "perplexity": 38.01873887288203, "lr": 0.0026291804804649314, "grad_norm": 0.191569, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:56.040800+00:00", "epoch": 0, "step": 14294, "train_loss": 3.5306825637817383, "perplexity": 34.14726735001051, "lr": 0.0026291804804649314, "grad_norm": 0.207648, "tokens_per_sec": 106964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:56.345945+00:00", "epoch": 0, "step": 14295, "train_loss": 3.438328504562378, "perplexity": 31.134872852328588, "lr": 0.0026291804804649314, "grad_norm": 0.171221, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:56.651037+00:00", "epoch": 0, "step": 14296, "train_loss": 3.4563546180725098, "perplexity": 31.7012026317121, "lr": 0.0026291804804649314, "grad_norm": 0.168174, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:56.955356+00:00", "epoch": 0, "step": 14297, "train_loss": 3.564052104949951, "perplexity": 35.305971182649415, "lr": 0.0026291804804649314, "grad_norm": 0.185174, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:57.261286+00:00", "epoch": 0, "step": 14298, "train_loss": 3.5457205772399902, "perplexity": 34.664654906288256, "lr": 0.0026291804804649314, "grad_norm": 0.173436, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:57.566319+00:00", "epoch": 0, "step": 14299, "train_loss": 3.4826366901397705, "perplexity": 32.54542123641739, "lr": 0.0026291804804649314, "grad_norm": 0.153983, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:57.872001+00:00", "epoch": 0, "step": 14300, "train_loss": 3.5593862533569336, "perplexity": 35.141622472268075, "lr": 0.0026291804804649314, "grad_norm": 0.157503, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:58.177734+00:00", "epoch": 0, "step": 14301, "train_loss": 3.539414405822754, "perplexity": 34.44674147222549, "lr": 0.0026291804804649314, "grad_norm": 0.151079, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:58.482756+00:00", "epoch": 0, "step": 14302, "train_loss": 3.5002989768981934, "perplexity": 33.12535419399723, "lr": 0.0026291804804649314, "grad_norm": 0.165342, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:58.787650+00:00", "epoch": 0, "step": 14303, "train_loss": 3.423865795135498, "perplexity": 30.687818831915795, "lr": 0.0026291804804649314, "grad_norm": 0.182404, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:59.093561+00:00", "epoch": 0, "step": 14304, "train_loss": 3.5359573364257812, "perplexity": 34.32786230162473, "lr": 0.0026291804804649314, "grad_norm": 0.158786, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:59.399869+00:00", "epoch": 0, "step": 14305, "train_loss": 3.64009428024292, "perplexity": 38.09542820231875, "lr": 0.0026291804804649314, "grad_norm": 0.163872, "tokens_per_sec": 106978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:12:59.704964+00:00", "epoch": 0, "step": 14306, "train_loss": 3.541353464126587, "perplexity": 34.513600513157584, "lr": 0.0026291804804649314, "grad_norm": 0.186161, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:00.009497+00:00", "epoch": 0, "step": 14307, "train_loss": 3.4390833377838135, "perplexity": 31.15838336084283, "lr": 0.0026291804804649314, "grad_norm": 0.15256, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:00.312899+00:00", "epoch": 0, "step": 14308, "train_loss": 3.4647815227508545, "perplexity": 31.969474407261437, "lr": 0.0026291804804649314, "grad_norm": 0.183322, "tokens_per_sec": 108002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:00.618050+00:00", "epoch": 0, "step": 14309, "train_loss": 3.5688700675964355, "perplexity": 35.476484466481125, "lr": 0.0026291804804649314, "grad_norm": 0.220618, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:00.924915+00:00", "epoch": 0, "step": 14310, "train_loss": 3.563579797744751, "perplexity": 35.289299855377166, "lr": 0.0026291804804649314, "grad_norm": 0.148081, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:01.230869+00:00", "epoch": 0, "step": 14311, "train_loss": 3.3923745155334473, "perplexity": 29.736478234210324, "lr": 0.0026291804804649314, "grad_norm": 0.198659, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:01.536209+00:00", "epoch": 0, "step": 14312, "train_loss": 3.621415853500366, "perplexity": 37.39046979003048, "lr": 0.0026291804804649314, "grad_norm": 0.173576, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:01.840887+00:00", "epoch": 0, "step": 14313, "train_loss": 3.404954671859741, "perplexity": 30.11293072983396, "lr": 0.0026291804804649314, "grad_norm": 0.178028, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:02.145974+00:00", "epoch": 0, "step": 14314, "train_loss": 3.5141751766204834, "perplexity": 33.58821215252642, "lr": 0.0026291804804649314, "grad_norm": 0.155906, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:02.451145+00:00", "epoch": 0, "step": 14315, "train_loss": 3.5418541431427, "perplexity": 34.53088507535276, "lr": 0.0026291804804649314, "grad_norm": 0.172448, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:02.756684+00:00", "epoch": 0, "step": 14316, "train_loss": 3.5854084491729736, "perplexity": 36.06808669081587, "lr": 0.0026291804804649314, "grad_norm": 0.163253, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:03.062773+00:00", "epoch": 0, "step": 14317, "train_loss": 3.4686474800109863, "perplexity": 32.093306239004626, "lr": 0.0026291804804649314, "grad_norm": 0.145051, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:03.367395+00:00", "epoch": 0, "step": 14318, "train_loss": 3.4156086444854736, "perplexity": 30.43546817119312, "lr": 0.0026291804804649314, "grad_norm": 0.161913, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:03.672149+00:00", "epoch": 0, "step": 14319, "train_loss": 3.529384136199951, "perplexity": 34.102958368467576, "lr": 0.0026291804804649314, "grad_norm": 0.139583, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:03.977729+00:00", "epoch": 0, "step": 14320, "train_loss": 3.4879469871520996, "perplexity": 32.71870678121948, "lr": 0.0026291804804649314, "grad_norm": 0.155091, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:04.283734+00:00", "epoch": 0, "step": 14321, "train_loss": 3.475653648376465, "perplexity": 32.318946861024294, "lr": 0.0026291804804649314, "grad_norm": 0.166785, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:04.588798+00:00", "epoch": 0, "step": 14322, "train_loss": 3.5903007984161377, "perplexity": 36.24497671851366, "lr": 0.0026291804804649314, "grad_norm": 0.147446, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:04.893495+00:00", "epoch": 0, "step": 14323, "train_loss": 3.4784278869628906, "perplexity": 32.40873181533609, "lr": 0.0026291804804649314, "grad_norm": 0.155553, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:05.197948+00:00", "epoch": 0, "step": 14324, "train_loss": 3.4395647048950195, "perplexity": 31.173385592331822, "lr": 0.0026291804804649314, "grad_norm": 0.17621, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:05.503497+00:00", "epoch": 0, "step": 14325, "train_loss": 3.4562180042266846, "perplexity": 31.696872104315037, "lr": 0.0026291804804649314, "grad_norm": 0.153692, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:05.808999+00:00", "epoch": 0, "step": 14326, "train_loss": 3.506692409515381, "perplexity": 33.33781737475234, "lr": 0.0026291804804649314, "grad_norm": 0.158551, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:06.113604+00:00", "epoch": 0, "step": 14327, "train_loss": 3.45040225982666, "perplexity": 31.51306620002553, "lr": 0.0026291804804649314, "grad_norm": 0.15462, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:06.417542+00:00", "epoch": 0, "step": 14328, "train_loss": 3.5522756576538086, "perplexity": 34.89263089084591, "lr": 0.0026291804804649314, "grad_norm": 0.153716, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:06.722771+00:00", "epoch": 0, "step": 14329, "train_loss": 3.5554556846618652, "perplexity": 35.003767013741154, "lr": 0.0026291804804649314, "grad_norm": 0.149544, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:07.026798+00:00", "epoch": 0, "step": 14330, "train_loss": 3.539175271987915, "perplexity": 34.43850507567933, "lr": 0.0026291804804649314, "grad_norm": 0.151681, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:07.331829+00:00", "epoch": 0, "step": 14331, "train_loss": 3.5150046348571777, "perplexity": 33.61608372930866, "lr": 0.0026291804804649314, "grad_norm": 0.14531, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:07.636775+00:00", "epoch": 0, "step": 14332, "train_loss": 3.512849807739258, "perplexity": 33.543724868906324, "lr": 0.0026291804804649314, "grad_norm": 0.166506, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:07.943407+00:00", "epoch": 0, "step": 14333, "train_loss": 3.6612837314605713, "perplexity": 38.91126243457788, "lr": 0.0026291804804649314, "grad_norm": 0.171499, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:08.248287+00:00", "epoch": 0, "step": 14334, "train_loss": 3.5859272480010986, "perplexity": 36.08680362666615, "lr": 0.0026291804804649314, "grad_norm": 0.158362, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:08.553617+00:00", "epoch": 0, "step": 14335, "train_loss": 3.5269250869750977, "perplexity": 34.01920053964483, "lr": 0.0026291804804649314, "grad_norm": 0.15829, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:08.858146+00:00", "epoch": 0, "step": 14336, "train_loss": 3.4865942001342773, "perplexity": 32.674475264102405, "lr": 0.0026291804804649314, "grad_norm": 0.17706, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:09.163363+00:00", "epoch": 0, "step": 14337, "train_loss": 3.5460734367370605, "perplexity": 34.67688881728351, "lr": 0.0026291804804649314, "grad_norm": 0.199096, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:09.468128+00:00", "epoch": 0, "step": 14338, "train_loss": 3.5711586475372314, "perplexity": 35.557768213900374, "lr": 0.0026291804804649314, "grad_norm": 0.153934, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:09.772316+00:00", "epoch": 0, "step": 14339, "train_loss": 3.476426124572754, "perplexity": 32.343922123319444, "lr": 0.0026291804804649314, "grad_norm": 0.169331, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:10.076473+00:00", "epoch": 0, "step": 14340, "train_loss": 3.6070611476898193, "perplexity": 36.85757452402307, "lr": 0.0026291804804649314, "grad_norm": 0.155466, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:10.381268+00:00", "epoch": 0, "step": 14341, "train_loss": 3.6176910400390625, "perplexity": 37.25145632517014, "lr": 0.0026291804804649314, "grad_norm": 0.16371, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:10.686468+00:00", "epoch": 0, "step": 14342, "train_loss": 3.5155160427093506, "perplexity": 33.63327965517811, "lr": 0.0026291804804649314, "grad_norm": 0.171745, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:10.991660+00:00", "epoch": 0, "step": 14343, "train_loss": 3.5905356407165527, "perplexity": 36.25348957177462, "lr": 0.0026291804804649314, "grad_norm": 0.16168, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:11.297658+00:00", "epoch": 0, "step": 14344, "train_loss": 3.5578715801239014, "perplexity": 35.088434688568924, "lr": 0.0026291804804649314, "grad_norm": 0.142373, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:11.602399+00:00", "epoch": 0, "step": 14345, "train_loss": 3.4800515174865723, "perplexity": 32.46139436223561, "lr": 0.0026291804804649314, "grad_norm": 0.165531, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:11.907084+00:00", "epoch": 0, "step": 14346, "train_loss": 3.5361363887786865, "perplexity": 34.33400933644414, "lr": 0.0026291804804649314, "grad_norm": 0.167166, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:12.211529+00:00", "epoch": 0, "step": 14347, "train_loss": 3.5057942867279053, "perplexity": 33.30788936280985, "lr": 0.0026291804804649314, "grad_norm": 0.166552, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:12.517453+00:00", "epoch": 0, "step": 14348, "train_loss": 3.501491069793701, "perplexity": 33.16486623972841, "lr": 0.0026291804804649314, "grad_norm": 0.209107, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:12.822996+00:00", "epoch": 0, "step": 14349, "train_loss": 3.4511284828186035, "perplexity": 31.535960025251565, "lr": 0.0026291804804649314, "grad_norm": 0.180985, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:13.127963+00:00", "epoch": 0, "step": 14350, "train_loss": 3.509549379348755, "perplexity": 33.43319869920847, "lr": 0.0026291804804649314, "grad_norm": 0.179505, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:13.433318+00:00", "epoch": 0, "step": 14351, "train_loss": 3.5952839851379395, "perplexity": 36.42604297402094, "lr": 0.0026291804804649314, "grad_norm": 0.218318, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:13.738696+00:00", "epoch": 0, "step": 14352, "train_loss": 3.5436251163482666, "perplexity": 34.59209252996944, "lr": 0.0026291804804649314, "grad_norm": 0.215063, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:14.044744+00:00", "epoch": 0, "step": 14353, "train_loss": 3.511901378631592, "perplexity": 33.511926105681056, "lr": 0.0026291804804649314, "grad_norm": 0.196237, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:14.350925+00:00", "epoch": 0, "step": 14354, "train_loss": 3.503225326538086, "perplexity": 33.22243253562147, "lr": 0.0026291804804649314, "grad_norm": 0.17786, "tokens_per_sec": 107019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:14.657693+00:00", "epoch": 0, "step": 14355, "train_loss": 3.4805397987365723, "perplexity": 32.47724852278121, "lr": 0.0026291804804649314, "grad_norm": 0.218164, "tokens_per_sec": 106818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:14.963203+00:00", "epoch": 0, "step": 14356, "train_loss": 3.494943618774414, "perplexity": 32.948430227757314, "lr": 0.0026291804804649314, "grad_norm": 0.170868, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:15.268168+00:00", "epoch": 0, "step": 14357, "train_loss": 3.4920098781585693, "perplexity": 32.851909732047055, "lr": 0.0026291804804649314, "grad_norm": 0.159776, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:15.574151+00:00", "epoch": 0, "step": 14358, "train_loss": 3.543687343597412, "perplexity": 34.59424516770544, "lr": 0.0026291804804649314, "grad_norm": 0.172689, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:15.880383+00:00", "epoch": 0, "step": 14359, "train_loss": 3.51582932472229, "perplexity": 33.64381800738214, "lr": 0.0026291804804649314, "grad_norm": 0.176983, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:16.185420+00:00", "epoch": 0, "step": 14360, "train_loss": 3.5010552406311035, "perplexity": 33.15041517317435, "lr": 0.0026291804804649314, "grad_norm": 0.165876, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:16.489673+00:00", "epoch": 0, "step": 14361, "train_loss": 3.5510964393615723, "perplexity": 34.8515091127768, "lr": 0.0026291804804649314, "grad_norm": 0.172297, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:16.794679+00:00", "epoch": 0, "step": 14362, "train_loss": 3.490175485610962, "perplexity": 32.79170167315051, "lr": 0.0026291804804649314, "grad_norm": 0.174014, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:17.099389+00:00", "epoch": 0, "step": 14363, "train_loss": 3.5086681842803955, "perplexity": 33.40375050610207, "lr": 0.0026291804804649314, "grad_norm": 0.182387, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:17.405898+00:00", "epoch": 0, "step": 14364, "train_loss": 3.545064926147461, "perplexity": 34.64193443659111, "lr": 0.0026291804804649314, "grad_norm": 0.178374, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:17.712624+00:00", "epoch": 0, "step": 14365, "train_loss": 3.568526029586792, "perplexity": 35.4642813066719, "lr": 0.0026291804804649314, "grad_norm": 0.16229, "tokens_per_sec": 106833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:18.017727+00:00", "epoch": 0, "step": 14366, "train_loss": 3.43601393699646, "perplexity": 31.062892419377953, "lr": 0.0026291804804649314, "grad_norm": 0.15977, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:18.322307+00:00", "epoch": 0, "step": 14367, "train_loss": 3.511927604675293, "perplexity": 33.51280500244456, "lr": 0.0026291804804649314, "grad_norm": 0.175049, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:18.628628+00:00", "epoch": 0, "step": 14368, "train_loss": 3.4806511402130127, "perplexity": 32.48086478889893, "lr": 0.0026291804804649314, "grad_norm": 0.161097, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:18.933723+00:00", "epoch": 0, "step": 14369, "train_loss": 3.4382612705230713, "perplexity": 31.13277959943314, "lr": 0.0026291804804649314, "grad_norm": 0.153363, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:19.239999+00:00", "epoch": 0, "step": 14370, "train_loss": 3.4813904762268066, "perplexity": 32.50488794149318, "lr": 0.0026291804804649314, "grad_norm": 0.153943, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:19.546710+00:00", "epoch": 0, "step": 14371, "train_loss": 3.4181296825408936, "perplexity": 30.5122939443464, "lr": 0.0026291804804649314, "grad_norm": 0.167019, "tokens_per_sec": 106774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:19.852387+00:00", "epoch": 0, "step": 14372, "train_loss": 3.53644061088562, "perplexity": 34.344456090090084, "lr": 0.0026291804804649314, "grad_norm": 0.18056, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:20.157384+00:00", "epoch": 0, "step": 14373, "train_loss": 3.5258350372314453, "perplexity": 33.98213812242635, "lr": 0.0026291804804649314, "grad_norm": 0.189911, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:20.462855+00:00", "epoch": 0, "step": 14374, "train_loss": 3.5625643730163574, "perplexity": 35.253484414675796, "lr": 0.0026291804804649314, "grad_norm": 0.189711, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:21.074164+00:00", "epoch": 0, "step": 14375, "train_loss": 3.495544672012329, "perplexity": 32.96823994117974, "lr": 0.0026291804804649314, "grad_norm": 0.19565, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:21.380034+00:00", "epoch": 0, "step": 14376, "train_loss": 3.5430052280426025, "perplexity": 34.570655941173456, "lr": 0.0026291804804649314, "grad_norm": 0.179067, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:21.685321+00:00", "epoch": 0, "step": 14377, "train_loss": 3.510993242263794, "perplexity": 33.4815065214855, "lr": 0.0026291804804649314, "grad_norm": 0.199182, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:21.990278+00:00", "epoch": 0, "step": 14378, "train_loss": 3.461840867996216, "perplexity": 31.87560131217103, "lr": 0.0026291804804649314, "grad_norm": 0.223619, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:22.295559+00:00", "epoch": 0, "step": 14379, "train_loss": 3.4589881896972656, "perplexity": 31.78480005101209, "lr": 0.0026291804804649314, "grad_norm": 0.193725, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:22.600882+00:00", "epoch": 0, "step": 14380, "train_loss": 3.480163335800171, "perplexity": 32.46502434355566, "lr": 0.0026291804804649314, "grad_norm": 0.176109, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:22.905074+00:00", "epoch": 0, "step": 14381, "train_loss": 3.4887266159057617, "perplexity": 32.74422517195179, "lr": 0.0026291804804649314, "grad_norm": 0.1699, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:23.210464+00:00", "epoch": 0, "step": 14382, "train_loss": 3.571956157684326, "perplexity": 35.58613720563478, "lr": 0.0026291804804649314, "grad_norm": 0.164956, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:23.516627+00:00", "epoch": 0, "step": 14383, "train_loss": 3.401463508605957, "perplexity": 30.00798487076155, "lr": 0.0026291804804649314, "grad_norm": 0.166918, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:23.822938+00:00", "epoch": 0, "step": 14384, "train_loss": 3.5471396446228027, "perplexity": 34.71388130693117, "lr": 0.0026291804804649314, "grad_norm": 0.179763, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:24.128098+00:00", "epoch": 0, "step": 14385, "train_loss": 3.431563377380371, "perplexity": 30.92495234776048, "lr": 0.0026291804804649314, "grad_norm": 0.181196, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:24.432696+00:00", "epoch": 0, "step": 14386, "train_loss": 3.493520736694336, "perplexity": 32.901581834593166, "lr": 0.0026291804804649314, "grad_norm": 0.20371, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:24.737686+00:00", "epoch": 0, "step": 14387, "train_loss": 3.517033576965332, "perplexity": 33.684358056004925, "lr": 0.0026291804804649314, "grad_norm": 0.188391, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:25.042170+00:00", "epoch": 0, "step": 14388, "train_loss": 3.4800446033477783, "perplexity": 32.461169920425455, "lr": 0.0026291804804649314, "grad_norm": 0.148043, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:25.347246+00:00", "epoch": 0, "step": 14389, "train_loss": 3.534320592880249, "perplexity": 34.271722350454894, "lr": 0.0026291804804649314, "grad_norm": 0.181638, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:25.652641+00:00", "epoch": 0, "step": 14390, "train_loss": 3.5102789402008057, "perplexity": 33.457599151863484, "lr": 0.0026291804804649314, "grad_norm": 0.169574, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:25.957882+00:00", "epoch": 0, "step": 14391, "train_loss": 3.4919886589050293, "perplexity": 32.85121264644102, "lr": 0.0026291804804649314, "grad_norm": 0.184115, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:26.263015+00:00", "epoch": 0, "step": 14392, "train_loss": 3.5258803367614746, "perplexity": 33.98367753217969, "lr": 0.0026291804804649314, "grad_norm": 0.166026, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:26.568588+00:00", "epoch": 0, "step": 14393, "train_loss": 3.4741508960723877, "perplexity": 32.27041596325573, "lr": 0.0026291804804649314, "grad_norm": 0.178344, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:26.874428+00:00", "epoch": 0, "step": 14394, "train_loss": 3.349421501159668, "perplexity": 28.486249613863873, "lr": 0.0026291804804649314, "grad_norm": 0.160866, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:27.180520+00:00", "epoch": 0, "step": 14395, "train_loss": 3.422492504119873, "perplexity": 30.645704450294478, "lr": 0.0026291804804649314, "grad_norm": 0.184888, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:27.485975+00:00", "epoch": 0, "step": 14396, "train_loss": 3.501915454864502, "perplexity": 33.17894390079897, "lr": 0.0026291804804649314, "grad_norm": 0.198033, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:27.791558+00:00", "epoch": 0, "step": 14397, "train_loss": 3.520691156387329, "perplexity": 33.807786858596614, "lr": 0.0026291804804649314, "grad_norm": 0.237618, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:28.096763+00:00", "epoch": 0, "step": 14398, "train_loss": 3.527573823928833, "perplexity": 34.04127721237387, "lr": 0.0026291804804649314, "grad_norm": 0.185113, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:28.402513+00:00", "epoch": 0, "step": 14399, "train_loss": 3.524887800216675, "perplexity": 33.949964223915224, "lr": 0.0026291804804649314, "grad_norm": 0.178201, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:28.709795+00:00", "epoch": 0, "step": 14400, "train_loss": 3.4679348468780518, "perplexity": 32.07044363292615, "lr": 0.0026291804804649314, "grad_norm": 0.21307, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:29.013794+00:00", "epoch": 0, "step": 14401, "train_loss": 3.562169313430786, "perplexity": 35.239559938413, "lr": 0.0026291804804649314, "grad_norm": 0.183065, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:29.318344+00:00", "epoch": 0, "step": 14402, "train_loss": 3.5054194927215576, "perplexity": 33.295408104609066, "lr": 0.0026291804804649314, "grad_norm": 0.171587, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:29.624746+00:00", "epoch": 0, "step": 14403, "train_loss": 3.530766487121582, "perplexity": 34.150133222988195, "lr": 0.0026291804804649314, "grad_norm": 0.19205, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:29.930813+00:00", "epoch": 0, "step": 14404, "train_loss": 3.5751326084136963, "perplexity": 35.6993545365584, "lr": 0.0026291804804649314, "grad_norm": 0.19374, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:30.236135+00:00", "epoch": 0, "step": 14405, "train_loss": 3.5848886966705322, "perplexity": 36.04934508342062, "lr": 0.0026291804804649314, "grad_norm": 0.193102, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:30.542836+00:00", "epoch": 0, "step": 14406, "train_loss": 3.4548964500427246, "perplexity": 31.65501063755801, "lr": 0.0026291804804649314, "grad_norm": 0.178861, "tokens_per_sec": 106842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:30.848133+00:00", "epoch": 0, "step": 14407, "train_loss": 3.426380157470703, "perplexity": 30.765076213559748, "lr": 0.0026291804804649314, "grad_norm": 0.204194, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:31.153799+00:00", "epoch": 0, "step": 14408, "train_loss": 3.432878255844116, "perplexity": 30.96564164647386, "lr": 0.0026291804804649314, "grad_norm": 0.182504, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:31.458611+00:00", "epoch": 0, "step": 14409, "train_loss": 3.5486838817596436, "perplexity": 34.76752918346982, "lr": 0.0026291804804649314, "grad_norm": 0.176045, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:31.762672+00:00", "epoch": 0, "step": 14410, "train_loss": 3.4826958179473877, "perplexity": 32.547345632715185, "lr": 0.0026291804804649314, "grad_norm": 0.173082, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:32.066902+00:00", "epoch": 0, "step": 14411, "train_loss": 3.5232737064361572, "perplexity": 33.8952099989272, "lr": 0.0026291804804649314, "grad_norm": 0.164591, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:32.373239+00:00", "epoch": 0, "step": 14412, "train_loss": 3.606607675552368, "perplexity": 36.84086442998978, "lr": 0.0026291804804649314, "grad_norm": 0.158533, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:32.679505+00:00", "epoch": 0, "step": 14413, "train_loss": 3.562241554260254, "perplexity": 35.242105765408255, "lr": 0.0026291804804649314, "grad_norm": 0.180628, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:32.984019+00:00", "epoch": 0, "step": 14414, "train_loss": 3.5564329624176025, "perplexity": 35.03799213761286, "lr": 0.0026291804804649314, "grad_norm": 0.159412, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:33.289346+00:00", "epoch": 0, "step": 14415, "train_loss": 3.4208996295928955, "perplexity": 30.59692854557545, "lr": 0.0026291804804649314, "grad_norm": 0.181271, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:33.594037+00:00", "epoch": 0, "step": 14416, "train_loss": 3.4118170738220215, "perplexity": 30.320288436982786, "lr": 0.0026291804804649314, "grad_norm": 0.143892, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:33.898982+00:00", "epoch": 0, "step": 14417, "train_loss": 3.5366735458374023, "perplexity": 34.35245704612839, "lr": 0.0026291804804649314, "grad_norm": 0.169361, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:34.204667+00:00", "epoch": 0, "step": 14418, "train_loss": 3.5641095638275146, "perplexity": 35.3079998824077, "lr": 0.0026291804804649314, "grad_norm": 0.172057, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:34.510571+00:00", "epoch": 0, "step": 14419, "train_loss": 3.535848617553711, "perplexity": 34.32413041802147, "lr": 0.0026291804804649314, "grad_norm": 0.192843, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:34.815049+00:00", "epoch": 0, "step": 14420, "train_loss": 3.5313687324523926, "perplexity": 34.170706175628816, "lr": 0.0026291804804649314, "grad_norm": 0.16168, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:35.119338+00:00", "epoch": 0, "step": 14421, "train_loss": 3.5858421325683594, "perplexity": 36.08373221347348, "lr": 0.0026291804804649314, "grad_norm": 0.15562, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:35.424439+00:00", "epoch": 0, "step": 14422, "train_loss": 3.5149385929107666, "perplexity": 33.61386373101577, "lr": 0.0026291804804649314, "grad_norm": 0.146913, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:35.731004+00:00", "epoch": 0, "step": 14423, "train_loss": 3.525848150253296, "perplexity": 33.98258373386773, "lr": 0.0026291804804649314, "grad_norm": 0.153934, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:36.037067+00:00", "epoch": 0, "step": 14424, "train_loss": 3.6143527030944824, "perplexity": 37.12730575579621, "lr": 0.0026291804804649314, "grad_norm": 0.172585, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:36.342228+00:00", "epoch": 0, "step": 14425, "train_loss": 3.4766592979431152, "perplexity": 32.351464743986604, "lr": 0.0026291804804649314, "grad_norm": 0.16574, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:36.647563+00:00", "epoch": 0, "step": 14426, "train_loss": 3.539374351501465, "perplexity": 34.44536175900713, "lr": 0.0026291804804649314, "grad_norm": 0.170141, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:36.951702+00:00", "epoch": 0, "step": 14427, "train_loss": 3.4272053241729736, "perplexity": 30.790473006897038, "lr": 0.0026291804804649314, "grad_norm": 0.167195, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:37.256666+00:00", "epoch": 0, "step": 14428, "train_loss": 3.5202126502990723, "perplexity": 33.79161349658898, "lr": 0.0026291804804649314, "grad_norm": 0.156773, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:37.563083+00:00", "epoch": 0, "step": 14429, "train_loss": 3.4792051315307617, "perplexity": 32.43393111783305, "lr": 0.0026291804804649314, "grad_norm": 0.168462, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:37.868198+00:00", "epoch": 0, "step": 14430, "train_loss": 3.5152838230133057, "perplexity": 33.625470251983295, "lr": 0.0026291804804649314, "grad_norm": 0.186996, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:38.172449+00:00", "epoch": 0, "step": 14431, "train_loss": 3.6226918697357178, "perplexity": 37.43821108938875, "lr": 0.0026291804804649314, "grad_norm": 0.166592, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:38.477982+00:00", "epoch": 0, "step": 14432, "train_loss": 3.4681758880615234, "perplexity": 32.0781748623488, "lr": 0.0026291804804649314, "grad_norm": 0.160118, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:38.783362+00:00", "epoch": 0, "step": 14433, "train_loss": 3.5465247631073, "perplexity": 34.69254294393896, "lr": 0.0026291804804649314, "grad_norm": 0.163095, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:39.088787+00:00", "epoch": 0, "step": 14434, "train_loss": 3.5183229446411133, "perplexity": 33.727817590100166, "lr": 0.0026291804804649314, "grad_norm": 0.153416, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:39.394485+00:00", "epoch": 0, "step": 14435, "train_loss": 3.475559949874878, "perplexity": 32.31591876599675, "lr": 0.0026291804804649314, "grad_norm": 0.173579, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:39.700008+00:00", "epoch": 0, "step": 14436, "train_loss": 3.588557481765747, "perplexity": 36.18184529212949, "lr": 0.0026291804804649314, "grad_norm": 0.158188, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:40.004211+00:00", "epoch": 0, "step": 14437, "train_loss": 3.440845489501953, "perplexity": 31.213337564218325, "lr": 0.0026291804804649314, "grad_norm": 0.170389, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:40.308288+00:00", "epoch": 0, "step": 14438, "train_loss": 3.456724166870117, "perplexity": 31.712919937952226, "lr": 0.0026291804804649314, "grad_norm": 0.138428, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:40.613457+00:00", "epoch": 0, "step": 14439, "train_loss": 3.547288656234741, "perplexity": 34.719054463762006, "lr": 0.0026291804804649314, "grad_norm": 0.157467, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:40.919385+00:00", "epoch": 0, "step": 14440, "train_loss": 3.5263407230377197, "perplexity": 33.999326752998684, "lr": 0.0026291804804649314, "grad_norm": 0.148661, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:41.223619+00:00", "epoch": 0, "step": 14441, "train_loss": 3.5963480472564697, "perplexity": 36.46482317509, "lr": 0.0026291804804649314, "grad_norm": 0.160566, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:41.529963+00:00", "epoch": 0, "step": 14442, "train_loss": 3.482198476791382, "perplexity": 32.53116252281015, "lr": 0.0026291804804649314, "grad_norm": 0.145091, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:41.834350+00:00", "epoch": 0, "step": 14443, "train_loss": 3.668642282485962, "perplexity": 39.198649022310555, "lr": 0.0026291804804649314, "grad_norm": 0.172924, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:42.139353+00:00", "epoch": 0, "step": 14444, "train_loss": 3.484184503555298, "perplexity": 32.59583448110463, "lr": 0.0026291804804649314, "grad_norm": 0.170015, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:42.445029+00:00", "epoch": 0, "step": 14445, "train_loss": 3.540895938873291, "perplexity": 34.4978132811447, "lr": 0.0026291804804649314, "grad_norm": 0.188331, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:42.751077+00:00", "epoch": 0, "step": 14446, "train_loss": 3.496262788772583, "perplexity": 32.99192348959632, "lr": 0.0026291804804649314, "grad_norm": 0.16306, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:43.055660+00:00", "epoch": 0, "step": 14447, "train_loss": 3.557037115097046, "perplexity": 35.059166830174675, "lr": 0.0026291804804649314, "grad_norm": 0.165153, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:43.360133+00:00", "epoch": 0, "step": 14448, "train_loss": 3.463768482208252, "perplexity": 31.937104432377957, "lr": 0.0026291804804649314, "grad_norm": 0.146213, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:43.664378+00:00", "epoch": 0, "step": 14449, "train_loss": 3.5311291217803955, "perplexity": 34.16251949060718, "lr": 0.0026291804804649314, "grad_norm": 0.180195, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:43.968920+00:00", "epoch": 0, "step": 14450, "train_loss": 3.575920343399048, "perplexity": 35.727487246186364, "lr": 0.0026291804804649314, "grad_norm": 0.182767, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:44.274722+00:00", "epoch": 0, "step": 14451, "train_loss": 3.5717594623565674, "perplexity": 35.57913826706527, "lr": 0.0026291804804649314, "grad_norm": 0.189294, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:44.580271+00:00", "epoch": 0, "step": 14452, "train_loss": 3.523151159286499, "perplexity": 33.89105649206018, "lr": 0.0026291804804649314, "grad_norm": 0.195453, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:44.886083+00:00", "epoch": 0, "step": 14453, "train_loss": 3.469460964202881, "perplexity": 32.119424258153224, "lr": 0.0026291804804649314, "grad_norm": 0.156944, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:45.191447+00:00", "epoch": 0, "step": 14454, "train_loss": 3.5532424449920654, "perplexity": 34.926380956528945, "lr": 0.0026291804804649314, "grad_norm": 0.194887, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:45.497357+00:00", "epoch": 0, "step": 14455, "train_loss": 3.409318208694458, "perplexity": 30.244616711644316, "lr": 0.0026291804804649314, "grad_norm": 0.216939, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:45.803300+00:00", "epoch": 0, "step": 14456, "train_loss": 3.5999491214752197, "perplexity": 36.59637242686867, "lr": 0.0026291804804649314, "grad_norm": 0.151049, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:46.109369+00:00", "epoch": 0, "step": 14457, "train_loss": 3.50858211517334, "perplexity": 33.40087559884583, "lr": 0.0026291804804649314, "grad_norm": 0.209149, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:46.414937+00:00", "epoch": 0, "step": 14458, "train_loss": 3.5759997367858887, "perplexity": 35.73032388500584, "lr": 0.0026291804804649314, "grad_norm": 0.201825, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:46.720396+00:00", "epoch": 0, "step": 14459, "train_loss": 3.5430822372436523, "perplexity": 34.57331830227909, "lr": 0.0026291804804649314, "grad_norm": 0.167506, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:47.025712+00:00", "epoch": 0, "step": 14460, "train_loss": 3.5798146724700928, "perplexity": 35.86689310837207, "lr": 0.0026291804804649314, "grad_norm": 0.19695, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:47.330746+00:00", "epoch": 0, "step": 14461, "train_loss": 3.589536190032959, "perplexity": 36.217274097644285, "lr": 0.0026291804804649314, "grad_norm": 0.185364, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:47.636420+00:00", "epoch": 0, "step": 14462, "train_loss": 3.479159116744995, "perplexity": 32.43243871177773, "lr": 0.0026291804804649314, "grad_norm": 0.156133, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:47.941720+00:00", "epoch": 0, "step": 14463, "train_loss": 3.4544150829315186, "perplexity": 31.639776623403066, "lr": 0.0026291804804649314, "grad_norm": 0.17609, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:48.247224+00:00", "epoch": 0, "step": 14464, "train_loss": 3.45448637008667, "perplexity": 31.642032213464518, "lr": 0.0026291804804649314, "grad_norm": 0.16029, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:48.552276+00:00", "epoch": 0, "step": 14465, "train_loss": 3.536271810531616, "perplexity": 34.338659223014375, "lr": 0.0026291804804649314, "grad_norm": 0.16475, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:48.856628+00:00", "epoch": 0, "step": 14466, "train_loss": 3.577911138534546, "perplexity": 35.79868419975969, "lr": 0.0026291804804649314, "grad_norm": 0.159047, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:49.161698+00:00", "epoch": 0, "step": 14467, "train_loss": 3.5041592121124268, "perplexity": 33.25347297796569, "lr": 0.0026291804804649314, "grad_norm": 0.168896, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:49.466714+00:00", "epoch": 0, "step": 14468, "train_loss": 3.5166730880737305, "perplexity": 33.67221740752094, "lr": 0.0026291804804649314, "grad_norm": 0.155829, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:49.771624+00:00", "epoch": 0, "step": 14469, "train_loss": 3.5276176929473877, "perplexity": 34.04277060255207, "lr": 0.0026291804804649314, "grad_norm": 0.154266, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:50.075998+00:00", "epoch": 0, "step": 14470, "train_loss": 3.447066307067871, "perplexity": 31.40811525292798, "lr": 0.0026291804804649314, "grad_norm": 0.172927, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:50.380623+00:00", "epoch": 0, "step": 14471, "train_loss": 3.451974868774414, "perplexity": 31.562662917802463, "lr": 0.0026291804804649314, "grad_norm": 0.163718, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:50.686284+00:00", "epoch": 0, "step": 14472, "train_loss": 3.5391485691070557, "perplexity": 34.437585480659294, "lr": 0.0026291804804649314, "grad_norm": 0.163463, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:50.992402+00:00", "epoch": 0, "step": 14473, "train_loss": 3.494954824447632, "perplexity": 32.94879943916811, "lr": 0.0026291804804649314, "grad_norm": 0.158348, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:51.298532+00:00", "epoch": 0, "step": 14474, "train_loss": 3.4973983764648438, "perplexity": 33.02940999243231, "lr": 0.0026291804804649314, "grad_norm": 0.171913, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:51.604242+00:00", "epoch": 0, "step": 14475, "train_loss": 3.5247676372528076, "perplexity": 33.94588494068518, "lr": 0.0026291804804649314, "grad_norm": 0.171057, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:51.908226+00:00", "epoch": 0, "step": 14476, "train_loss": 3.5139377117156982, "perplexity": 33.580237077863366, "lr": 0.0026291804804649314, "grad_norm": 0.178635, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:52.213050+00:00", "epoch": 0, "step": 14477, "train_loss": 3.5088233947753906, "perplexity": 33.408935521126786, "lr": 0.0026291804804649314, "grad_norm": 0.16391, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:52.518012+00:00", "epoch": 0, "step": 14478, "train_loss": 3.5280723571777344, "perplexity": 34.05825215183359, "lr": 0.0026291804804649314, "grad_norm": 0.171752, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:52.822647+00:00", "epoch": 0, "step": 14479, "train_loss": 3.444141149520874, "perplexity": 31.3163758091365, "lr": 0.0026291804804649314, "grad_norm": 0.189775, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:53.126697+00:00", "epoch": 0, "step": 14480, "train_loss": 3.595780611038208, "perplexity": 36.44413758316001, "lr": 0.0026291804804649314, "grad_norm": 0.165853, "tokens_per_sec": 107824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:53.431874+00:00", "epoch": 0, "step": 14481, "train_loss": 3.5136709213256836, "perplexity": 33.571279388282456, "lr": 0.0026291804804649314, "grad_norm": 0.159115, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:53.737858+00:00", "epoch": 0, "step": 14482, "train_loss": 3.4682271480560303, "perplexity": 32.079819231560855, "lr": 0.0026291804804649314, "grad_norm": 0.173194, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:54.043049+00:00", "epoch": 0, "step": 14483, "train_loss": 3.5085599422454834, "perplexity": 33.400135011851354, "lr": 0.0026291804804649314, "grad_norm": 0.156805, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:54.349608+00:00", "epoch": 0, "step": 14484, "train_loss": 3.5185317993164062, "perplexity": 33.73486253815085, "lr": 0.0026291804804649314, "grad_norm": 0.182137, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:54.655866+00:00", "epoch": 0, "step": 14485, "train_loss": 3.5769078731536865, "perplexity": 35.76278662962747, "lr": 0.0026291804804649314, "grad_norm": 0.174872, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:54.961084+00:00", "epoch": 0, "step": 14486, "train_loss": 3.5927646160125732, "perplexity": 36.33438783104124, "lr": 0.0026291804804649314, "grad_norm": 0.192189, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:55.265423+00:00", "epoch": 0, "step": 14487, "train_loss": 3.5656957626342773, "perplexity": 35.36404983111931, "lr": 0.0026291804804649314, "grad_norm": 0.196741, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:55.570815+00:00", "epoch": 0, "step": 14488, "train_loss": 3.5066428184509277, "perplexity": 33.33616415789491, "lr": 0.0026291804804649314, "grad_norm": 0.178884, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:55.876298+00:00", "epoch": 0, "step": 14489, "train_loss": 3.5301549434661865, "perplexity": 34.12925531021298, "lr": 0.0026291804804649314, "grad_norm": 0.173412, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:56.182474+00:00", "epoch": 0, "step": 14490, "train_loss": 3.602433681488037, "perplexity": 36.68741135942813, "lr": 0.0026291804804649314, "grad_norm": 0.176032, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:56.487266+00:00", "epoch": 0, "step": 14491, "train_loss": 3.5126307010650635, "perplexity": 33.536376020030524, "lr": 0.0026291804804649314, "grad_norm": 0.174018, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:56.792292+00:00", "epoch": 0, "step": 14492, "train_loss": 3.534132719039917, "perplexity": 34.26528419516154, "lr": 0.0026291804804649314, "grad_norm": 0.21496, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:57.096809+00:00", "epoch": 0, "step": 14493, "train_loss": 3.527200937271118, "perplexity": 34.028586040621164, "lr": 0.0026291804804649314, "grad_norm": 0.174938, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:57.402354+00:00", "epoch": 0, "step": 14494, "train_loss": 3.5658605098724365, "perplexity": 35.36987644060484, "lr": 0.0026291804804649314, "grad_norm": 0.163812, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:57.707819+00:00", "epoch": 0, "step": 14495, "train_loss": 3.564117670059204, "perplexity": 35.30828609839531, "lr": 0.0026291804804649314, "grad_norm": 0.19161, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:58.013778+00:00", "epoch": 0, "step": 14496, "train_loss": 3.488107204437256, "perplexity": 32.72394930355392, "lr": 0.0026291804804649314, "grad_norm": 0.158572, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:58.318673+00:00", "epoch": 0, "step": 14497, "train_loss": 3.5703012943267822, "perplexity": 35.527295711881074, "lr": 0.0026291804804649314, "grad_norm": 0.155452, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:58.623642+00:00", "epoch": 0, "step": 14498, "train_loss": 3.5284078121185303, "perplexity": 34.06967907729494, "lr": 0.0026291804804649314, "grad_norm": 0.149861, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:58.928765+00:00", "epoch": 0, "step": 14499, "train_loss": 3.4438042640686035, "perplexity": 31.30582755458508, "lr": 0.0026291804804649314, "grad_norm": 0.157551, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:13:59.233796+00:00", "epoch": 0, "step": 14500, "train_loss": 3.5914535522460938, "perplexity": 36.28678234541144, "lr": 0.0026291804804649314, "grad_norm": 0.161375, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:14:02.308398+00:00", "step": 14500, "epoch": 0, "val_loss": 3.476642608642578, "val_ppl": 32.350924825174125, "eval_train_loss": 3.5914535522460938, "eval_train_ppl": 36.28678234541144} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:14:03.265446+00:00", "step": 14500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4766_epoch_0000_step_0014500.pt", "val_loss": 3.476642608642578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:04.507999+00:00", "epoch": 0, "step": 14501, "train_loss": 3.426077127456665, "perplexity": 30.75575488447544, "lr": 0.0026291804804649314, "grad_norm": 0.156369, "tokens_per_sec": 6213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:04.812694+00:00", "epoch": 0, "step": 14502, "train_loss": 3.376560926437378, "perplexity": 29.269936359291172, "lr": 0.0026291804804649314, "grad_norm": 0.139844, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:05.118031+00:00", "epoch": 0, "step": 14503, "train_loss": 3.5555942058563232, "perplexity": 35.008616113202194, "lr": 0.0026291804804649314, "grad_norm": 0.174632, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:05.422625+00:00", "epoch": 0, "step": 14504, "train_loss": 3.549189329147339, "perplexity": 34.78510678217367, "lr": 0.0026291804804649314, "grad_norm": 0.188361, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:05.726991+00:00", "epoch": 0, "step": 14505, "train_loss": 3.4591078758239746, "perplexity": 31.788604478282455, "lr": 0.0026291804804649314, "grad_norm": 0.178835, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:06.031062+00:00", "epoch": 0, "step": 14506, "train_loss": 3.613642930984497, "perplexity": 37.100963179370765, "lr": 0.0026291804804649314, "grad_norm": 0.180495, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:06.335485+00:00", "epoch": 0, "step": 14507, "train_loss": 3.5605528354644775, "perplexity": 35.182641981935305, "lr": 0.0026291804804649314, "grad_norm": 0.168072, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:06.640267+00:00", "epoch": 0, "step": 14508, "train_loss": 3.5396411418914795, "perplexity": 34.45455267647472, "lr": 0.0026291804804649314, "grad_norm": 0.174818, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:06.944453+00:00", "epoch": 0, "step": 14509, "train_loss": 3.4868381023406982, "perplexity": 32.68244561266626, "lr": 0.0026291804804649314, "grad_norm": 0.158415, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:07.248360+00:00", "epoch": 0, "step": 14510, "train_loss": 3.549152374267578, "perplexity": 34.78382132648715, "lr": 0.0026291804804649314, "grad_norm": 0.152403, "tokens_per_sec": 107880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:07.553016+00:00", "epoch": 0, "step": 14511, "train_loss": 3.4561192989349365, "perplexity": 31.693743609708505, "lr": 0.0026291804804649314, "grad_norm": 0.140952, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:07.857156+00:00", "epoch": 0, "step": 14512, "train_loss": 3.446650981903076, "perplexity": 31.395073380780982, "lr": 0.0026291804804649314, "grad_norm": 0.167652, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:08.160697+00:00", "epoch": 0, "step": 14513, "train_loss": 3.522209644317627, "perplexity": 33.85916257171976, "lr": 0.0026291804804649314, "grad_norm": 0.184248, "tokens_per_sec": 107955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:08.565980+00:00", "epoch": 0, "step": 14514, "train_loss": 3.555225133895874, "perplexity": 34.99569779866116, "lr": 0.0026291804804649314, "grad_norm": 0.145247, "tokens_per_sec": 80851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:08.870396+00:00", "epoch": 0, "step": 14515, "train_loss": 3.4861369132995605, "perplexity": 32.659537072512336, "lr": 0.0026291804804649314, "grad_norm": 0.152001, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:09.175414+00:00", "epoch": 0, "step": 14516, "train_loss": 3.5718536376953125, "perplexity": 35.58248910224437, "lr": 0.0026291804804649314, "grad_norm": 0.137062, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:09.481008+00:00", "epoch": 0, "step": 14517, "train_loss": 3.458167552947998, "perplexity": 31.758726975747205, "lr": 0.0026291804804649314, "grad_norm": 0.154829, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:09.784322+00:00", "epoch": 0, "step": 14518, "train_loss": 3.443946361541748, "perplexity": 31.31027634964907, "lr": 0.0026291804804649314, "grad_norm": 0.157423, "tokens_per_sec": 108028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:10.087748+00:00", "epoch": 0, "step": 14519, "train_loss": 3.4962778091430664, "perplexity": 32.992419044231795, "lr": 0.0026291804804649314, "grad_norm": 0.182908, "tokens_per_sec": 107993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:10.391126+00:00", "epoch": 0, "step": 14520, "train_loss": 3.415464162826538, "perplexity": 30.4310711219154, "lr": 0.0026291804804649314, "grad_norm": 0.176051, "tokens_per_sec": 108107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:10.696305+00:00", "epoch": 0, "step": 14521, "train_loss": 3.5234038829803467, "perplexity": 33.899622647434875, "lr": 0.0026291804804649314, "grad_norm": 0.187086, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:11.001354+00:00", "epoch": 0, "step": 14522, "train_loss": 3.5546371936798096, "perplexity": 34.975128467896994, "lr": 0.0026291804804649314, "grad_norm": 0.177954, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:11.307003+00:00", "epoch": 0, "step": 14523, "train_loss": 3.434274196624756, "perplexity": 31.0088980330996, "lr": 0.0026291804804649314, "grad_norm": 0.134167, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:11.611703+00:00", "epoch": 0, "step": 14524, "train_loss": 3.4421870708465576, "perplexity": 31.255240897677602, "lr": 0.0026291804804649314, "grad_norm": 0.169512, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:11.916306+00:00", "epoch": 0, "step": 14525, "train_loss": 3.571014404296875, "perplexity": 35.55263961608503, "lr": 0.0026291804804649314, "grad_norm": 0.152673, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:12.221278+00:00", "epoch": 0, "step": 14526, "train_loss": 3.5308103561401367, "perplexity": 34.15163138867749, "lr": 0.0026291804804649314, "grad_norm": 0.166371, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:12.527356+00:00", "epoch": 0, "step": 14527, "train_loss": 3.702512502670288, "perplexity": 40.549056092500656, "lr": 0.0026291804804649314, "grad_norm": 0.150272, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:12.831547+00:00", "epoch": 0, "step": 14528, "train_loss": 3.4440295696258545, "perplexity": 31.312881726149655, "lr": 0.0026291804804649314, "grad_norm": 0.165635, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:13.135336+00:00", "epoch": 0, "step": 14529, "train_loss": 3.4327480792999268, "perplexity": 30.961610908615167, "lr": 0.0026291804804649314, "grad_norm": 0.172714, "tokens_per_sec": 107864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:13.439544+00:00", "epoch": 0, "step": 14530, "train_loss": 3.4875059127807617, "perplexity": 32.704278580383686, "lr": 0.0026291804804649314, "grad_norm": 0.174943, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:13.744089+00:00", "epoch": 0, "step": 14531, "train_loss": 3.4764604568481445, "perplexity": 32.34503258282319, "lr": 0.0026291804804649314, "grad_norm": 0.146366, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:14.048732+00:00", "epoch": 0, "step": 14532, "train_loss": 3.548286199569702, "perplexity": 34.75370550522279, "lr": 0.0026291804804649314, "grad_norm": 0.206386, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:14.353393+00:00", "epoch": 0, "step": 14533, "train_loss": 3.329911470413208, "perplexity": 27.93586844286984, "lr": 0.0026291804804649314, "grad_norm": 0.229803, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:14.658618+00:00", "epoch": 0, "step": 14534, "train_loss": 3.6272878646850586, "perplexity": 37.610672931767915, "lr": 0.0026291804804649314, "grad_norm": 0.225126, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:14.964650+00:00", "epoch": 0, "step": 14535, "train_loss": 3.4153642654418945, "perplexity": 30.42803128933661, "lr": 0.0026291804804649314, "grad_norm": 0.184721, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:15.269161+00:00", "epoch": 0, "step": 14536, "train_loss": 3.543261766433716, "perplexity": 34.57952577930669, "lr": 0.0026291804804649314, "grad_norm": 0.175073, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:15.575195+00:00", "epoch": 0, "step": 14537, "train_loss": 3.41701602935791, "perplexity": 30.478332745083936, "lr": 0.0026291804804649314, "grad_norm": 0.199539, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:15.879889+00:00", "epoch": 0, "step": 14538, "train_loss": 3.540165424346924, "perplexity": 34.47262133007833, "lr": 0.0026291804804649314, "grad_norm": 0.221726, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:16.184132+00:00", "epoch": 0, "step": 14539, "train_loss": 3.6146912574768066, "perplexity": 37.13987749585249, "lr": 0.0026291804804649314, "grad_norm": 0.195311, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:16.488621+00:00", "epoch": 0, "step": 14540, "train_loss": 3.5567312240600586, "perplexity": 35.04844418534117, "lr": 0.0026291804804649314, "grad_norm": 0.160773, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:16.792430+00:00", "epoch": 0, "step": 14541, "train_loss": 3.523153781890869, "perplexity": 33.89114537500959, "lr": 0.0026291804804649314, "grad_norm": 0.170387, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:17.095885+00:00", "epoch": 0, "step": 14542, "train_loss": 3.4071149826049805, "perplexity": 30.178054335945948, "lr": 0.0026291804804649314, "grad_norm": 0.162764, "tokens_per_sec": 107983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:17.401412+00:00", "epoch": 0, "step": 14543, "train_loss": 3.5181386470794678, "perplexity": 33.721602208316135, "lr": 0.0026291804804649314, "grad_norm": 0.152383, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:17.707135+00:00", "epoch": 0, "step": 14544, "train_loss": 3.504106044769287, "perplexity": 33.25170502615636, "lr": 0.0026291804804649314, "grad_norm": 0.154215, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:18.011844+00:00", "epoch": 0, "step": 14545, "train_loss": 3.47599196434021, "perplexity": 32.32988272646343, "lr": 0.0026291804804649314, "grad_norm": 0.149443, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:18.317347+00:00", "epoch": 0, "step": 14546, "train_loss": 3.5215108394622803, "perplexity": 33.83550988978368, "lr": 0.0026291804804649314, "grad_norm": 0.159318, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:18.622961+00:00", "epoch": 0, "step": 14547, "train_loss": 3.459523916244507, "perplexity": 31.80183257417822, "lr": 0.0026291804804649314, "grad_norm": 0.154154, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:18.927783+00:00", "epoch": 0, "step": 14548, "train_loss": 3.5775327682495117, "perplexity": 35.78514160363473, "lr": 0.0026291804804649314, "grad_norm": 0.143962, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:19.233455+00:00", "epoch": 0, "step": 14549, "train_loss": 3.534125328063965, "perplexity": 34.265030942205954, "lr": 0.0026291804804649314, "grad_norm": 0.157552, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:19.538504+00:00", "epoch": 0, "step": 14550, "train_loss": 3.5335311889648438, "perplexity": 34.24467879419273, "lr": 0.0026291804804649314, "grad_norm": 0.17537, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:19.843660+00:00", "epoch": 0, "step": 14551, "train_loss": 3.4635770320892334, "perplexity": 31.93099065519363, "lr": 0.0026291804804649314, "grad_norm": 0.169585, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:20.148806+00:00", "epoch": 0, "step": 14552, "train_loss": 3.5301809310913086, "perplexity": 34.13014226003049, "lr": 0.0026291804804649314, "grad_norm": 0.174166, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:20.454090+00:00", "epoch": 0, "step": 14553, "train_loss": 3.5297555923461914, "perplexity": 34.115628475007306, "lr": 0.0026291804804649314, "grad_norm": 0.162518, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:20.759534+00:00", "epoch": 0, "step": 14554, "train_loss": 3.401458978652954, "perplexity": 30.00784893630826, "lr": 0.0026291804804649314, "grad_norm": 0.170857, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:21.065734+00:00", "epoch": 0, "step": 14555, "train_loss": 3.5523486137390137, "perplexity": 34.89517661346006, "lr": 0.0026291804804649314, "grad_norm": 0.191285, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:21.371214+00:00", "epoch": 0, "step": 14556, "train_loss": 3.413227081298828, "perplexity": 30.363070424751196, "lr": 0.0026291804804649314, "grad_norm": 0.172537, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:21.675402+00:00", "epoch": 0, "step": 14557, "train_loss": 3.4717507362365723, "perplexity": 32.19305468383188, "lr": 0.0026291804804649314, "grad_norm": 0.164742, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:21.979457+00:00", "epoch": 0, "step": 14558, "train_loss": 3.4939305782318115, "perplexity": 32.91506903309593, "lr": 0.0026291804804649314, "grad_norm": 0.167881, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:22.285533+00:00", "epoch": 0, "step": 14559, "train_loss": 3.421713352203369, "perplexity": 30.621836090685978, "lr": 0.0026291804804649314, "grad_norm": 0.190352, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:22.590373+00:00", "epoch": 0, "step": 14560, "train_loss": 3.5269834995269775, "perplexity": 34.02118774599959, "lr": 0.0026291804804649314, "grad_norm": 0.166683, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:22.895349+00:00", "epoch": 0, "step": 14561, "train_loss": 3.6298561096191406, "perplexity": 37.707390496028104, "lr": 0.0026291804804649314, "grad_norm": 0.180112, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:23.200761+00:00", "epoch": 0, "step": 14562, "train_loss": 3.577861785888672, "perplexity": 35.796917483572045, "lr": 0.0026291804804649314, "grad_norm": 0.181291, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:23.506276+00:00", "epoch": 0, "step": 14563, "train_loss": 3.5236010551452637, "perplexity": 33.90630736842093, "lr": 0.0026291804804649314, "grad_norm": 0.169811, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:23.811986+00:00", "epoch": 0, "step": 14564, "train_loss": 3.5516622066497803, "perplexity": 34.87123253548508, "lr": 0.0026291804804649314, "grad_norm": 0.161072, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:24.116806+00:00", "epoch": 0, "step": 14565, "train_loss": 3.492633104324341, "perplexity": 32.87239028313239, "lr": 0.0026291804804649314, "grad_norm": 0.26119, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:24.422299+00:00", "epoch": 0, "step": 14566, "train_loss": 3.5714035034179688, "perplexity": 35.566475808562814, "lr": 0.0026291804804649314, "grad_norm": 0.203859, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:24.726975+00:00", "epoch": 0, "step": 14567, "train_loss": 3.4646425247192383, "perplexity": 31.965031022065073, "lr": 0.0026291804804649314, "grad_norm": 0.173856, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:25.031426+00:00", "epoch": 0, "step": 14568, "train_loss": 3.5307533740997314, "perplexity": 34.14968541448111, "lr": 0.0026291804804649314, "grad_norm": 0.18595, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:25.335535+00:00", "epoch": 0, "step": 14569, "train_loss": 3.3802850246429443, "perplexity": 29.379143699979085, "lr": 0.0026291804804649314, "grad_norm": 0.185476, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:25.639899+00:00", "epoch": 0, "step": 14570, "train_loss": 3.458770990371704, "perplexity": 31.77789716355629, "lr": 0.0026291804804649314, "grad_norm": 0.167042, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:25.944713+00:00", "epoch": 0, "step": 14571, "train_loss": 3.5689971446990967, "perplexity": 35.48099300179947, "lr": 0.0026291804804649314, "grad_norm": 0.177719, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:26.250187+00:00", "epoch": 0, "step": 14572, "train_loss": 3.537019729614258, "perplexity": 34.36435136814475, "lr": 0.0026291804804649314, "grad_norm": 0.149203, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:26.557070+00:00", "epoch": 0, "step": 14573, "train_loss": 3.4867074489593506, "perplexity": 32.678175819573774, "lr": 0.0026291804804649314, "grad_norm": 0.156987, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:26.862801+00:00", "epoch": 0, "step": 14574, "train_loss": 3.5541164875030518, "perplexity": 34.956921443137375, "lr": 0.0026291804804649314, "grad_norm": 0.168453, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:27.169286+00:00", "epoch": 0, "step": 14575, "train_loss": 3.5485947132110596, "perplexity": 34.7644291515694, "lr": 0.0026291804804649314, "grad_norm": 0.168582, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:27.474927+00:00", "epoch": 0, "step": 14576, "train_loss": 3.57114315032959, "perplexity": 35.5572171720529, "lr": 0.0026291804804649314, "grad_norm": 0.144984, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:27.780047+00:00", "epoch": 0, "step": 14577, "train_loss": 3.4040708541870117, "perplexity": 30.086328147125304, "lr": 0.0026291804804649314, "grad_norm": 0.153499, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:28.083984+00:00", "epoch": 0, "step": 14578, "train_loss": 3.4893953800201416, "perplexity": 32.766130658685256, "lr": 0.0026291804804649314, "grad_norm": 0.148638, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:28.389171+00:00", "epoch": 0, "step": 14579, "train_loss": 3.5099005699157715, "perplexity": 33.44494218519688, "lr": 0.0026291804804649314, "grad_norm": 0.177712, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:28.693972+00:00", "epoch": 0, "step": 14580, "train_loss": 3.498990058898926, "perplexity": 33.08202418556089, "lr": 0.0026291804804649314, "grad_norm": 0.192239, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:28.999621+00:00", "epoch": 0, "step": 14581, "train_loss": 3.5937206745147705, "perplexity": 36.36914224242722, "lr": 0.0026291804804649314, "grad_norm": 0.182218, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:29.303473+00:00", "epoch": 0, "step": 14582, "train_loss": 3.4579737186431885, "perplexity": 31.752571641558955, "lr": 0.0026291804804649314, "grad_norm": 0.173994, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:29.608790+00:00", "epoch": 0, "step": 14583, "train_loss": 3.506164073944092, "perplexity": 33.320208472078555, "lr": 0.0026291804804649314, "grad_norm": 0.21982, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:29.914146+00:00", "epoch": 0, "step": 14584, "train_loss": 3.4587807655334473, "perplexity": 31.778207799159176, "lr": 0.0026291804804649314, "grad_norm": 0.18491, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:30.220089+00:00", "epoch": 0, "step": 14585, "train_loss": 3.5655386447906494, "perplexity": 35.3584939443438, "lr": 0.0026291804804649314, "grad_norm": 0.170622, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:30.526342+00:00", "epoch": 0, "step": 14586, "train_loss": 3.6320583820343018, "perplexity": 37.79052394962123, "lr": 0.0026291804804649314, "grad_norm": 0.179771, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:30.830111+00:00", "epoch": 0, "step": 14587, "train_loss": 3.4914145469665527, "perplexity": 32.8323577859854, "lr": 0.0026291804804649314, "grad_norm": 0.164905, "tokens_per_sec": 107874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:31.134196+00:00", "epoch": 0, "step": 14588, "train_loss": 3.540799856185913, "perplexity": 34.494498797770774, "lr": 0.0026291804804649314, "grad_norm": 0.185266, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:31.439876+00:00", "epoch": 0, "step": 14589, "train_loss": 3.524500608444214, "perplexity": 33.936821621610314, "lr": 0.0026291804804649314, "grad_norm": 0.19195, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:31.744384+00:00", "epoch": 0, "step": 14590, "train_loss": 3.5743556022644043, "perplexity": 35.671626692305516, "lr": 0.0026291804804649314, "grad_norm": 0.211314, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:32.050440+00:00", "epoch": 0, "step": 14591, "train_loss": 3.469160795211792, "perplexity": 32.10978444983745, "lr": 0.0026291804804649314, "grad_norm": 0.179578, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:32.355522+00:00", "epoch": 0, "step": 14592, "train_loss": 3.4660730361938477, "perplexity": 32.01079008735906, "lr": 0.0026291804804649314, "grad_norm": 0.195162, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:32.660149+00:00", "epoch": 0, "step": 14593, "train_loss": 3.487933874130249, "perplexity": 32.71827774291553, "lr": 0.0026291804804649314, "grad_norm": 0.172664, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:32.964920+00:00", "epoch": 0, "step": 14594, "train_loss": 3.511176824569702, "perplexity": 33.48765369789714, "lr": 0.0026291804804649314, "grad_norm": 0.192391, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:33.270628+00:00", "epoch": 0, "step": 14595, "train_loss": 3.563154935836792, "perplexity": 35.27430996065319, "lr": 0.0026291804804649314, "grad_norm": 0.169284, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:33.576129+00:00", "epoch": 0, "step": 14596, "train_loss": 3.5428550243377686, "perplexity": 34.56546369052996, "lr": 0.0026291804804649314, "grad_norm": 0.19404, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:33.881395+00:00", "epoch": 0, "step": 14597, "train_loss": 3.4914040565490723, "perplexity": 32.832013362651935, "lr": 0.0026291804804649314, "grad_norm": 0.177488, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:34.185691+00:00", "epoch": 0, "step": 14598, "train_loss": 3.495638608932495, "perplexity": 32.97133702156593, "lr": 0.0026291804804649314, "grad_norm": 0.182858, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:34.490551+00:00", "epoch": 0, "step": 14599, "train_loss": 3.563631057739258, "perplexity": 35.29110883105754, "lr": 0.0026291804804649314, "grad_norm": 0.16548, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:34.796099+00:00", "epoch": 0, "step": 14600, "train_loss": 3.5112974643707275, "perplexity": 33.491693885474874, "lr": 0.0026291804804649314, "grad_norm": 0.175138, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:35.101955+00:00", "epoch": 0, "step": 14601, "train_loss": 3.4897329807281494, "perplexity": 32.777194395051836, "lr": 0.0026291804804649314, "grad_norm": 0.172656, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:35.406810+00:00", "epoch": 0, "step": 14602, "train_loss": 3.493539571762085, "perplexity": 32.90220154395219, "lr": 0.0026291804804649314, "grad_norm": 0.184984, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:35.711004+00:00", "epoch": 0, "step": 14603, "train_loss": 3.5340518951416016, "perplexity": 34.26251485323201, "lr": 0.0026291804804649314, "grad_norm": 0.160889, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:36.015610+00:00", "epoch": 0, "step": 14604, "train_loss": 3.502516031265259, "perplexity": 33.19887637639592, "lr": 0.0026291804804649314, "grad_norm": 0.169193, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:36.320880+00:00", "epoch": 0, "step": 14605, "train_loss": 3.6094682216644287, "perplexity": 36.94640029463936, "lr": 0.0026291804804649314, "grad_norm": 0.158356, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:36.626147+00:00", "epoch": 0, "step": 14606, "train_loss": 3.5461604595184326, "perplexity": 34.67990662790495, "lr": 0.0026291804804649314, "grad_norm": 0.163722, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:36.931348+00:00", "epoch": 0, "step": 14607, "train_loss": 3.4919545650482178, "perplexity": 32.85009264099373, "lr": 0.0026291804804649314, "grad_norm": 0.148926, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:37.236751+00:00", "epoch": 0, "step": 14608, "train_loss": 3.42983078956604, "perplexity": 30.87141854156587, "lr": 0.0026291804804649314, "grad_norm": 0.161117, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:37.540739+00:00", "epoch": 0, "step": 14609, "train_loss": 3.5025129318237305, "perplexity": 33.198773478579255, "lr": 0.0026291804804649314, "grad_norm": 0.151741, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:37.846402+00:00", "epoch": 0, "step": 14610, "train_loss": 3.4959681034088135, "perplexity": 32.9822026849811, "lr": 0.0026291804804649314, "grad_norm": 0.165146, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:38.150452+00:00", "epoch": 0, "step": 14611, "train_loss": 3.540856122970581, "perplexity": 34.49643974691183, "lr": 0.0026291804804649314, "grad_norm": 0.191795, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:38.454472+00:00", "epoch": 0, "step": 14612, "train_loss": 3.51892352104187, "perplexity": 33.74807980528892, "lr": 0.0026291804804649314, "grad_norm": 0.165894, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:38.758790+00:00", "epoch": 0, "step": 14613, "train_loss": 3.407041311264038, "perplexity": 30.175831160109166, "lr": 0.0026291804804649314, "grad_norm": 0.175306, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:39.062952+00:00", "epoch": 0, "step": 14614, "train_loss": 3.5207653045654297, "perplexity": 33.81029373733689, "lr": 0.0026291804804649314, "grad_norm": 0.17761, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:39.367109+00:00", "epoch": 0, "step": 14615, "train_loss": 3.5460450649261475, "perplexity": 34.67590498510754, "lr": 0.0026291804804649314, "grad_norm": 0.171334, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:39.671475+00:00", "epoch": 0, "step": 14616, "train_loss": 3.49225115776062, "perplexity": 32.85983718408159, "lr": 0.0026291804804649314, "grad_norm": 0.168081, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:39.975609+00:00", "epoch": 0, "step": 14617, "train_loss": 3.567676544189453, "perplexity": 35.43416770991601, "lr": 0.0026291804804649314, "grad_norm": 0.196705, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:40.280932+00:00", "epoch": 0, "step": 14618, "train_loss": 3.5001635551452637, "perplexity": 33.120868604195934, "lr": 0.0026291804804649314, "grad_norm": 0.194018, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:40.586103+00:00", "epoch": 0, "step": 14619, "train_loss": 3.518066883087158, "perplexity": 33.71918229834683, "lr": 0.0026291804804649314, "grad_norm": 0.191622, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:40.890458+00:00", "epoch": 0, "step": 14620, "train_loss": 3.5177314281463623, "perplexity": 33.70787292904321, "lr": 0.0026291804804649314, "grad_norm": 0.182393, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:41.194715+00:00", "epoch": 0, "step": 14621, "train_loss": 3.479276418685913, "perplexity": 32.4362433229271, "lr": 0.0026291804804649314, "grad_norm": 0.17788, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:41.499149+00:00", "epoch": 0, "step": 14622, "train_loss": 3.5092873573303223, "perplexity": 33.42443961258954, "lr": 0.0026291804804649314, "grad_norm": 0.159242, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:41.804104+00:00", "epoch": 0, "step": 14623, "train_loss": 3.548144578933716, "perplexity": 34.74878401184699, "lr": 0.0026291804804649314, "grad_norm": 0.168844, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:42.108505+00:00", "epoch": 0, "step": 14624, "train_loss": 3.5599913597106934, "perplexity": 35.16289332622287, "lr": 0.0026291804804649314, "grad_norm": 0.164066, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:42.413886+00:00", "epoch": 0, "step": 14625, "train_loss": 3.5190303325653076, "perplexity": 33.75168468162374, "lr": 0.0026291804804649314, "grad_norm": 0.163116, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:42.719180+00:00", "epoch": 0, "step": 14626, "train_loss": 3.470932960510254, "perplexity": 32.16673874688921, "lr": 0.0026291804804649314, "grad_norm": 0.144916, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:43.022929+00:00", "epoch": 0, "step": 14627, "train_loss": 3.5289852619171143, "perplexity": 34.08935828794723, "lr": 0.0026291804804649314, "grad_norm": 0.16779, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:43.327166+00:00", "epoch": 0, "step": 14628, "train_loss": 3.6494274139404297, "perplexity": 38.452642297454105, "lr": 0.0026291804804649314, "grad_norm": 0.17729, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:43.632662+00:00", "epoch": 0, "step": 14629, "train_loss": 3.5031378269195557, "perplexity": 33.21952571262277, "lr": 0.0026291804804649314, "grad_norm": 0.199407, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:43.937197+00:00", "epoch": 0, "step": 14630, "train_loss": 3.5573630332946777, "perplexity": 35.07059511287993, "lr": 0.0026291804804649314, "grad_norm": 0.174829, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:44.241319+00:00", "epoch": 0, "step": 14631, "train_loss": 3.5338239669799805, "perplexity": 34.25470635113157, "lr": 0.0026291804804649314, "grad_norm": 0.182073, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:44.545165+00:00", "epoch": 0, "step": 14632, "train_loss": 3.505150556564331, "perplexity": 33.28645496946496, "lr": 0.0026291804804649314, "grad_norm": 0.184546, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:44.850301+00:00", "epoch": 0, "step": 14633, "train_loss": 3.5265023708343506, "perplexity": 34.00482311348055, "lr": 0.0026291804804649314, "grad_norm": 0.198729, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:45.154677+00:00", "epoch": 0, "step": 14634, "train_loss": 3.5196144580841064, "perplexity": 33.7714056611517, "lr": 0.0026291804804649314, "grad_norm": 0.177142, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:45.459408+00:00", "epoch": 0, "step": 14635, "train_loss": 3.4687860012054443, "perplexity": 32.0977521500383, "lr": 0.0026291804804649314, "grad_norm": 0.189093, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:45.763585+00:00", "epoch": 0, "step": 14636, "train_loss": 3.528336524963379, "perplexity": 34.06725043336319, "lr": 0.0026291804804649314, "grad_norm": 0.159784, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:46.067867+00:00", "epoch": 0, "step": 14637, "train_loss": 3.508486032485962, "perplexity": 33.39766650712906, "lr": 0.0026291804804649314, "grad_norm": 0.159347, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:46.371848+00:00", "epoch": 0, "step": 14638, "train_loss": 3.5380611419677734, "perplexity": 34.40015746940209, "lr": 0.0026291804804649314, "grad_norm": 0.168471, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:46.677121+00:00", "epoch": 0, "step": 14639, "train_loss": 3.6596741676330566, "perplexity": 38.84868265066894, "lr": 0.0026291804804649314, "grad_norm": 0.156908, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:46.983329+00:00", "epoch": 0, "step": 14640, "train_loss": 3.4924659729003906, "perplexity": 32.866896732820955, "lr": 0.0026291804804649314, "grad_norm": 0.161705, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:47.288430+00:00", "epoch": 0, "step": 14641, "train_loss": 3.5184426307678223, "perplexity": 33.73185458353111, "lr": 0.0026291804804649314, "grad_norm": 0.159001, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:47.593053+00:00", "epoch": 0, "step": 14642, "train_loss": 3.5074288845062256, "perplexity": 33.36237888685773, "lr": 0.0026291804804649314, "grad_norm": 0.176199, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:47.897236+00:00", "epoch": 0, "step": 14643, "train_loss": 3.5635640621185303, "perplexity": 35.28874456051402, "lr": 0.0026291804804649314, "grad_norm": 0.180655, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:48.201929+00:00", "epoch": 0, "step": 14644, "train_loss": 3.507338523864746, "perplexity": 33.35936437709879, "lr": 0.0026291804804649314, "grad_norm": 0.193422, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:48.506663+00:00", "epoch": 0, "step": 14645, "train_loss": 3.5632894039154053, "perplexity": 35.279053548261416, "lr": 0.0026291804804649314, "grad_norm": 0.209883, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:48.811205+00:00", "epoch": 0, "step": 14646, "train_loss": 3.559241533279419, "perplexity": 35.13653714192347, "lr": 0.0026291804804649314, "grad_norm": 0.150873, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:49.115677+00:00", "epoch": 0, "step": 14647, "train_loss": 3.5217232704162598, "perplexity": 33.84269836292802, "lr": 0.0026291804804649314, "grad_norm": 0.180359, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:49.419695+00:00", "epoch": 0, "step": 14648, "train_loss": 3.4582748413085938, "perplexity": 31.76213450028963, "lr": 0.0026291804804649314, "grad_norm": 0.161226, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:49.724469+00:00", "epoch": 0, "step": 14649, "train_loss": 3.544398069381714, "perplexity": 34.618840929120154, "lr": 0.0026291804804649314, "grad_norm": 0.181304, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:50.030388+00:00", "epoch": 0, "step": 14650, "train_loss": 3.636608123779297, "perplexity": 37.96285280261892, "lr": 0.0026291804804649314, "grad_norm": 0.167053, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:50.335886+00:00", "epoch": 0, "step": 14651, "train_loss": 3.5355312824249268, "perplexity": 34.313239893739215, "lr": 0.0026291804804649314, "grad_norm": 0.152326, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:50.639269+00:00", "epoch": 0, "step": 14652, "train_loss": 3.4938180446624756, "perplexity": 32.91136519129988, "lr": 0.0026291804804649314, "grad_norm": 0.163625, "tokens_per_sec": 108009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:50.943419+00:00", "epoch": 0, "step": 14653, "train_loss": 3.634127616882324, "perplexity": 37.86880237899917, "lr": 0.0026291804804649314, "grad_norm": 0.151875, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:51.248280+00:00", "epoch": 0, "step": 14654, "train_loss": 3.5676794052124023, "perplexity": 35.43426908802804, "lr": 0.0026291804804649314, "grad_norm": 0.153475, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:51.553536+00:00", "epoch": 0, "step": 14655, "train_loss": 3.50834584236145, "perplexity": 33.392984812274385, "lr": 0.0026291804804649314, "grad_norm": 0.153325, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:51.860246+00:00", "epoch": 0, "step": 14656, "train_loss": 3.56868577003479, "perplexity": 35.469946839351074, "lr": 0.0026291804804649314, "grad_norm": 0.171975, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:52.165355+00:00", "epoch": 0, "step": 14657, "train_loss": 3.569906234741211, "perplexity": 35.51326308520502, "lr": 0.0026291804804649314, "grad_norm": 0.143901, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:52.470091+00:00", "epoch": 0, "step": 14658, "train_loss": 3.5119760036468506, "perplexity": 33.51442702699253, "lr": 0.0026291804804649314, "grad_norm": 0.158959, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:52.774475+00:00", "epoch": 0, "step": 14659, "train_loss": 3.558220386505127, "perplexity": 35.10067589327656, "lr": 0.0026291804804649314, "grad_norm": 0.155048, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:53.079029+00:00", "epoch": 0, "step": 14660, "train_loss": 3.550654172897339, "perplexity": 34.83609886703693, "lr": 0.0026291804804649314, "grad_norm": 0.15677, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:53.383700+00:00", "epoch": 0, "step": 14661, "train_loss": 3.5119056701660156, "perplexity": 33.512069923574145, "lr": 0.0026291804804649314, "grad_norm": 0.181747, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:53.690765+00:00", "epoch": 0, "step": 14662, "train_loss": 3.5317389965057373, "perplexity": 34.18336070241668, "lr": 0.0026291804804649314, "grad_norm": 0.17473, "tokens_per_sec": 106714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:53.995291+00:00", "epoch": 0, "step": 14663, "train_loss": 3.459635019302368, "perplexity": 31.80536605130923, "lr": 0.0026291804804649314, "grad_norm": 0.174693, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:54.300060+00:00", "epoch": 0, "step": 14664, "train_loss": 3.6119163036346436, "perplexity": 37.03695891329853, "lr": 0.0026291804804649314, "grad_norm": 0.178511, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:54.604465+00:00", "epoch": 0, "step": 14665, "train_loss": 3.540025234222412, "perplexity": 34.46778894773588, "lr": 0.0026291804804649314, "grad_norm": 0.177829, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:54.908880+00:00", "epoch": 0, "step": 14666, "train_loss": 3.574946880340576, "perplexity": 35.69272477991372, "lr": 0.0026291804804649314, "grad_norm": 0.204437, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:55.213530+00:00", "epoch": 0, "step": 14667, "train_loss": 3.598102569580078, "perplexity": 36.52885767993453, "lr": 0.0026291804804649314, "grad_norm": 0.207872, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:55.517871+00:00", "epoch": 0, "step": 14668, "train_loss": 3.5280985832214355, "perplexity": 34.05914537675574, "lr": 0.0026291804804649314, "grad_norm": 0.212309, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:55.822745+00:00", "epoch": 0, "step": 14669, "train_loss": 3.4075000286102295, "perplexity": 30.189676512606948, "lr": 0.0026291804804649314, "grad_norm": 0.153469, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:56.127968+00:00", "epoch": 0, "step": 14670, "train_loss": 3.5131609439849854, "perplexity": 33.5541631613111, "lr": 0.0026291804804649314, "grad_norm": 0.159561, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:56.432534+00:00", "epoch": 0, "step": 14671, "train_loss": 3.510385751724243, "perplexity": 33.46117299986013, "lr": 0.0026291804804649314, "grad_norm": 0.165216, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:56.737278+00:00", "epoch": 0, "step": 14672, "train_loss": 3.4742400646209717, "perplexity": 32.273293597704715, "lr": 0.0026291804804649314, "grad_norm": 0.172711, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:57.042630+00:00", "epoch": 0, "step": 14673, "train_loss": 3.5498712062835693, "perplexity": 34.80883403978766, "lr": 0.0026291804804649314, "grad_norm": 0.169929, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:57.347288+00:00", "epoch": 0, "step": 14674, "train_loss": 3.562006711959839, "perplexity": 35.23383039995979, "lr": 0.0026291804804649314, "grad_norm": 0.163247, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:57.651835+00:00", "epoch": 0, "step": 14675, "train_loss": 3.515371561050415, "perplexity": 33.62842061416769, "lr": 0.0026291804804649314, "grad_norm": 0.163532, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:57.956577+00:00", "epoch": 0, "step": 14676, "train_loss": 3.5314364433288574, "perplexity": 34.17301998242745, "lr": 0.0026291804804649314, "grad_norm": 0.174804, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:58.261696+00:00", "epoch": 0, "step": 14677, "train_loss": 3.5208840370178223, "perplexity": 33.814308354756406, "lr": 0.0026291804804649314, "grad_norm": 0.181868, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:58.568008+00:00", "epoch": 0, "step": 14678, "train_loss": 3.5181491374969482, "perplexity": 33.72195596385693, "lr": 0.0026291804804649314, "grad_norm": 0.174498, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:58.872111+00:00", "epoch": 0, "step": 14679, "train_loss": 3.5124704837799072, "perplexity": 33.531003343319945, "lr": 0.0026291804804649314, "grad_norm": 0.176336, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:59.176763+00:00", "epoch": 0, "step": 14680, "train_loss": 3.6317954063415527, "perplexity": 37.78058726701649, "lr": 0.0026291804804649314, "grad_norm": 0.163189, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:59.480990+00:00", "epoch": 0, "step": 14681, "train_loss": 3.5527541637420654, "perplexity": 34.909331222449, "lr": 0.0026291804804649314, "grad_norm": 0.155819, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:14:59.785539+00:00", "epoch": 0, "step": 14682, "train_loss": 3.427412748336792, "perplexity": 30.796860357436586, "lr": 0.0026291804804649314, "grad_norm": 0.163705, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:00.090794+00:00", "epoch": 0, "step": 14683, "train_loss": 3.4822275638580322, "perplexity": 32.532108772664415, "lr": 0.0026291804804649314, "grad_norm": 0.154989, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:00.396699+00:00", "epoch": 0, "step": 14684, "train_loss": 3.5650527477264404, "perplexity": 35.34131752926416, "lr": 0.0026291804804649314, "grad_norm": 0.149223, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:00.701328+00:00", "epoch": 0, "step": 14685, "train_loss": 3.5579845905303955, "perplexity": 35.092400270908136, "lr": 0.0026291804804649314, "grad_norm": 0.166884, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:01.005774+00:00", "epoch": 0, "step": 14686, "train_loss": 3.555931806564331, "perplexity": 35.02043704205309, "lr": 0.0026291804804649314, "grad_norm": 0.143576, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:01.310524+00:00", "epoch": 0, "step": 14687, "train_loss": 3.4916090965270996, "perplexity": 32.83874592814936, "lr": 0.0026291804804649314, "grad_norm": 0.170328, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:01.615809+00:00", "epoch": 0, "step": 14688, "train_loss": 3.5592358112335205, "perplexity": 35.13633608962045, "lr": 0.0026291804804649314, "grad_norm": 0.183534, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:01.921483+00:00", "epoch": 0, "step": 14689, "train_loss": 3.499229669570923, "perplexity": 33.08995194135653, "lr": 0.0026291804804649314, "grad_norm": 0.179486, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:02.226549+00:00", "epoch": 0, "step": 14690, "train_loss": 3.5085480213165283, "perplexity": 33.399736853588, "lr": 0.0026291804804649314, "grad_norm": 0.168473, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:02.530811+00:00", "epoch": 0, "step": 14691, "train_loss": 3.4392285346984863, "perplexity": 31.162907790431678, "lr": 0.0026291804804649314, "grad_norm": 0.18166, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:02.834945+00:00", "epoch": 0, "step": 14692, "train_loss": 3.5064289569854736, "perplexity": 33.32903559926466, "lr": 0.0026291804804649314, "grad_norm": 0.159894, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:03.139873+00:00", "epoch": 0, "step": 14693, "train_loss": 3.4984893798828125, "perplexity": 33.06546485604096, "lr": 0.0026291804804649314, "grad_norm": 0.174434, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:03.445063+00:00", "epoch": 0, "step": 14694, "train_loss": 3.4805471897125244, "perplexity": 32.477488562231095, "lr": 0.0026291804804649314, "grad_norm": 0.165847, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:03.749798+00:00", "epoch": 0, "step": 14695, "train_loss": 3.4921681880950928, "perplexity": 32.85711092748101, "lr": 0.0026291804804649314, "grad_norm": 0.186244, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:04.054692+00:00", "epoch": 0, "step": 14696, "train_loss": 3.5062201023101807, "perplexity": 33.32207540121699, "lr": 0.0026291804804649314, "grad_norm": 0.172338, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:04.358962+00:00", "epoch": 0, "step": 14697, "train_loss": 3.510983943939209, "perplexity": 33.48119520101765, "lr": 0.0026291804804649314, "grad_norm": 0.176235, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:04.663223+00:00", "epoch": 0, "step": 14698, "train_loss": 3.422363519668579, "perplexity": 30.641751885836598, "lr": 0.0026291804804649314, "grad_norm": 0.187267, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:04.968709+00:00", "epoch": 0, "step": 14699, "train_loss": 3.477998971939087, "perplexity": 32.394834204018274, "lr": 0.0026291804804649314, "grad_norm": 0.153469, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:05.274518+00:00", "epoch": 0, "step": 14700, "train_loss": 3.5241408348083496, "perplexity": 33.92461424398382, "lr": 0.0026291804804649314, "grad_norm": 0.161833, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:05.578632+00:00", "epoch": 0, "step": 14701, "train_loss": 3.6195554733276367, "perplexity": 37.32097396575469, "lr": 0.0026291804804649314, "grad_norm": 0.154283, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:05.882792+00:00", "epoch": 0, "step": 14702, "train_loss": 3.5263614654541016, "perplexity": 34.000031988505015, "lr": 0.0026291804804649314, "grad_norm": 0.164367, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:06.187156+00:00", "epoch": 0, "step": 14703, "train_loss": 3.441512107849121, "perplexity": 31.234151884548098, "lr": 0.0026291804804649314, "grad_norm": 0.162672, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:06.493339+00:00", "epoch": 0, "step": 14704, "train_loss": 3.6145198345184326, "perplexity": 37.133511413840395, "lr": 0.0026291804804649314, "grad_norm": 0.169698, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:06.797855+00:00", "epoch": 0, "step": 14705, "train_loss": 3.438366413116455, "perplexity": 31.136053152711394, "lr": 0.0026291804804649314, "grad_norm": 0.183479, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:07.102857+00:00", "epoch": 0, "step": 14706, "train_loss": 3.57491135597229, "perplexity": 35.691456840935, "lr": 0.0026291804804649314, "grad_norm": 0.190606, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:07.406720+00:00", "epoch": 0, "step": 14707, "train_loss": 3.429260492324829, "perplexity": 30.8538176760789, "lr": 0.0026291804804649314, "grad_norm": 0.150013, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:07.710755+00:00", "epoch": 0, "step": 14708, "train_loss": 3.446871280670166, "perplexity": 31.401990438621144, "lr": 0.0026291804804649314, "grad_norm": 0.163888, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:08.014878+00:00", "epoch": 0, "step": 14709, "train_loss": 3.5248191356658936, "perplexity": 33.947633144904906, "lr": 0.0026291804804649314, "grad_norm": 0.177289, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:08.321412+00:00", "epoch": 0, "step": 14710, "train_loss": 3.45151424407959, "perplexity": 31.54812772371978, "lr": 0.0026291804804649314, "grad_norm": 0.156541, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:08.626130+00:00", "epoch": 0, "step": 14711, "train_loss": 3.49980092048645, "perplexity": 33.108860006808456, "lr": 0.0026291804804649314, "grad_norm": 0.173349, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:08.931681+00:00", "epoch": 0, "step": 14712, "train_loss": 3.4772326946258545, "perplexity": 32.37002028588726, "lr": 0.0026291804804649314, "grad_norm": 0.157427, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:09.236124+00:00", "epoch": 0, "step": 14713, "train_loss": 3.506894588470459, "perplexity": 33.344558261243485, "lr": 0.0026291804804649314, "grad_norm": 0.163436, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:09.539910+00:00", "epoch": 0, "step": 14714, "train_loss": 3.4701807498931885, "perplexity": 32.142551682511204, "lr": 0.0026291804804649314, "grad_norm": 0.180417, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:09.845418+00:00", "epoch": 0, "step": 14715, "train_loss": 3.4445719718933105, "perplexity": 31.329870511159747, "lr": 0.0026291804804649314, "grad_norm": 0.177663, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:10.151475+00:00", "epoch": 0, "step": 14716, "train_loss": 3.5833914279937744, "perplexity": 35.99540991600401, "lr": 0.0026291804804649314, "grad_norm": 0.164495, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:10.455338+00:00", "epoch": 0, "step": 14717, "train_loss": 3.487571954727173, "perplexity": 32.70643850591916, "lr": 0.0026291804804649314, "grad_norm": 0.179815, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:10.759731+00:00", "epoch": 0, "step": 14718, "train_loss": 3.53798770904541, "perplexity": 34.397631458056686, "lr": 0.0026291804804649314, "grad_norm": 0.171756, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:11.063924+00:00", "epoch": 0, "step": 14719, "train_loss": 3.49479341506958, "perplexity": 32.94348162312732, "lr": 0.0026291804804649314, "grad_norm": 0.17645, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:11.368996+00:00", "epoch": 0, "step": 14720, "train_loss": 3.4957895278930664, "perplexity": 32.97631339698312, "lr": 0.0026291804804649314, "grad_norm": 0.155076, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:11.673258+00:00", "epoch": 0, "step": 14721, "train_loss": 3.607847213745117, "perplexity": 36.88655840236605, "lr": 0.0026291804804649314, "grad_norm": 0.187515, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:11.977764+00:00", "epoch": 0, "step": 14722, "train_loss": 3.621504068374634, "perplexity": 37.39376833110987, "lr": 0.0026291804804649314, "grad_norm": 0.165803, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:12.282008+00:00", "epoch": 0, "step": 14723, "train_loss": 3.575479030609131, "perplexity": 35.71172372769003, "lr": 0.0026291804804649314, "grad_norm": 0.164598, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:12.586131+00:00", "epoch": 0, "step": 14724, "train_loss": 3.5182912349700928, "perplexity": 33.72674810905668, "lr": 0.0026291804804649314, "grad_norm": 0.162624, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:12.891051+00:00", "epoch": 0, "step": 14725, "train_loss": 3.4965929985046387, "perplexity": 33.0028195427042, "lr": 0.0026291804804649314, "grad_norm": 0.190985, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:13.195799+00:00", "epoch": 0, "step": 14726, "train_loss": 3.4677720069885254, "perplexity": 32.06522171060807, "lr": 0.0026291804804649314, "grad_norm": 0.188729, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:13.500352+00:00", "epoch": 0, "step": 14727, "train_loss": 3.5618348121643066, "perplexity": 35.22777423226013, "lr": 0.0026291804804649314, "grad_norm": 0.188002, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:13.804661+00:00", "epoch": 0, "step": 14728, "train_loss": 3.5358126163482666, "perplexity": 34.32289473019385, "lr": 0.0026291804804649314, "grad_norm": 0.183463, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:14.108625+00:00", "epoch": 0, "step": 14729, "train_loss": 3.589231014251709, "perplexity": 36.20622314905349, "lr": 0.0026291804804649314, "grad_norm": 0.174237, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:14.413802+00:00", "epoch": 0, "step": 14730, "train_loss": 3.573578357696533, "perplexity": 35.64391188621827, "lr": 0.0026291804804649314, "grad_norm": 0.163015, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:14.718528+00:00", "epoch": 0, "step": 14731, "train_loss": 3.5977578163146973, "perplexity": 36.51626640753459, "lr": 0.0026291804804649314, "grad_norm": 0.177696, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:15.023123+00:00", "epoch": 0, "step": 14732, "train_loss": 3.498037576675415, "perplexity": 33.05052914721955, "lr": 0.0026291804804649314, "grad_norm": 0.238132, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:15.327270+00:00", "epoch": 0, "step": 14733, "train_loss": 3.574965238571167, "perplexity": 35.69338004120034, "lr": 0.0026291804804649314, "grad_norm": 0.233687, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:15.631627+00:00", "epoch": 0, "step": 14734, "train_loss": 3.5635972023010254, "perplexity": 35.2899140553273, "lr": 0.0026291804804649314, "grad_norm": 0.17564, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:15.935721+00:00", "epoch": 0, "step": 14735, "train_loss": 3.4158010482788086, "perplexity": 30.441324634105953, "lr": 0.0026291804804649314, "grad_norm": 0.189919, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:16.240990+00:00", "epoch": 0, "step": 14736, "train_loss": 3.518555164337158, "perplexity": 33.73565076312253, "lr": 0.0026291804804649314, "grad_norm": 0.177913, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:16.545462+00:00", "epoch": 0, "step": 14737, "train_loss": 3.480004072189331, "perplexity": 32.45985425826686, "lr": 0.0026291804804649314, "grad_norm": 0.180125, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:16.849926+00:00", "epoch": 0, "step": 14738, "train_loss": 3.4799327850341797, "perplexity": 32.457540370076394, "lr": 0.0026291804804649314, "grad_norm": 0.166647, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:17.154025+00:00", "epoch": 0, "step": 14739, "train_loss": 3.557274341583252, "perplexity": 35.06748477971107, "lr": 0.0026291804804649314, "grad_norm": 0.160694, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:17.458969+00:00", "epoch": 0, "step": 14740, "train_loss": 3.5167012214660645, "perplexity": 33.673164734549715, "lr": 0.0026291804804649314, "grad_norm": 0.154702, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:17.764354+00:00", "epoch": 0, "step": 14741, "train_loss": 3.4896438121795654, "perplexity": 32.77427183050333, "lr": 0.0026291804804649314, "grad_norm": 0.181851, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:18.069494+00:00", "epoch": 0, "step": 14742, "train_loss": 3.52188777923584, "perplexity": 33.848266243257264, "lr": 0.0026291804804649314, "grad_norm": 0.171908, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:18.374343+00:00", "epoch": 0, "step": 14743, "train_loss": 3.45211124420166, "perplexity": 31.566967582961507, "lr": 0.0026291804804649314, "grad_norm": 0.165076, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:18.678322+00:00", "epoch": 0, "step": 14744, "train_loss": 3.5326757431030273, "perplexity": 34.21539685177292, "lr": 0.0026291804804649314, "grad_norm": 0.156162, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:18.982123+00:00", "epoch": 0, "step": 14745, "train_loss": 3.5942866802215576, "perplexity": 36.389733211240596, "lr": 0.0026291804804649314, "grad_norm": 0.171103, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:19.287347+00:00", "epoch": 0, "step": 14746, "train_loss": 3.476701259613037, "perplexity": 32.35282229395401, "lr": 0.0026291804804649314, "grad_norm": 0.164865, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:19.592031+00:00", "epoch": 0, "step": 14747, "train_loss": 3.5041282176971436, "perplexity": 33.252442321986976, "lr": 0.0026291804804649314, "grad_norm": 0.17017, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:19.897396+00:00", "epoch": 0, "step": 14748, "train_loss": 3.5044236183166504, "perplexity": 33.26226656502131, "lr": 0.0026291804804649314, "grad_norm": 0.156964, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:20.201616+00:00", "epoch": 0, "step": 14749, "train_loss": 3.43709397315979, "perplexity": 31.09645959014473, "lr": 0.0026291804804649314, "grad_norm": 0.149122, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:20.507177+00:00", "epoch": 0, "step": 14750, "train_loss": 3.507819890975952, "perplexity": 33.3754263435017, "lr": 0.0026291804804649314, "grad_norm": 0.16836, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:20.811325+00:00", "epoch": 0, "step": 14751, "train_loss": 3.4871175289154053, "perplexity": 32.69157923252548, "lr": 0.0026291804804649314, "grad_norm": 0.166715, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:21.116447+00:00", "epoch": 0, "step": 14752, "train_loss": 3.4582359790802, "perplexity": 31.760900176948848, "lr": 0.0026291804804649314, "grad_norm": 0.158409, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:21.421108+00:00", "epoch": 0, "step": 14753, "train_loss": 3.5881094932556152, "perplexity": 36.165639871355424, "lr": 0.0026291804804649314, "grad_norm": 0.165284, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:21.725378+00:00", "epoch": 0, "step": 14754, "train_loss": 3.5684947967529297, "perplexity": 35.46317367396315, "lr": 0.0026291804804649314, "grad_norm": 0.163812, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:22.029332+00:00", "epoch": 0, "step": 14755, "train_loss": 3.5355072021484375, "perplexity": 34.312413631383684, "lr": 0.0026291804804649314, "grad_norm": 0.169231, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:22.333630+00:00", "epoch": 0, "step": 14756, "train_loss": 3.4771809577941895, "perplexity": 32.368345606918396, "lr": 0.0026291804804649314, "grad_norm": 0.191379, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:22.638883+00:00", "epoch": 0, "step": 14757, "train_loss": 3.5246782302856445, "perplexity": 33.94285007773593, "lr": 0.0026291804804649314, "grad_norm": 0.194733, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:22.944315+00:00", "epoch": 0, "step": 14758, "train_loss": 3.5285654067993164, "perplexity": 34.075048700594536, "lr": 0.0026291804804649314, "grad_norm": 0.200151, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:23.248932+00:00", "epoch": 0, "step": 14759, "train_loss": 3.536194324493408, "perplexity": 34.33599855943728, "lr": 0.0026291804804649314, "grad_norm": 0.173921, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:23.553342+00:00", "epoch": 0, "step": 14760, "train_loss": 3.5202178955078125, "perplexity": 33.79179074112028, "lr": 0.0026291804804649314, "grad_norm": 0.167818, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:23.857572+00:00", "epoch": 0, "step": 14761, "train_loss": 3.5446550846099854, "perplexity": 34.62773964192741, "lr": 0.0026291804804649314, "grad_norm": 0.161365, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:24.161645+00:00", "epoch": 0, "step": 14762, "train_loss": 3.572431802749634, "perplexity": 35.60306760230062, "lr": 0.0026291804804649314, "grad_norm": 0.157433, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:24.466524+00:00", "epoch": 0, "step": 14763, "train_loss": 3.585278034210205, "perplexity": 36.06338317934375, "lr": 0.0026291804804649314, "grad_norm": 0.15019, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:24.771158+00:00", "epoch": 0, "step": 14764, "train_loss": 3.457235336303711, "perplexity": 31.729134757181424, "lr": 0.0026291804804649314, "grad_norm": 0.17997, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:25.075784+00:00", "epoch": 0, "step": 14765, "train_loss": 3.5753180980682373, "perplexity": 35.70597701168005, "lr": 0.0026291804804649314, "grad_norm": 0.144816, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:25.380590+00:00", "epoch": 0, "step": 14766, "train_loss": 3.4321584701538086, "perplexity": 30.943361040318223, "lr": 0.0026291804804649314, "grad_norm": 0.155169, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:25.684242+00:00", "epoch": 0, "step": 14767, "train_loss": 3.536290168762207, "perplexity": 34.33928962582509, "lr": 0.0026291804804649314, "grad_norm": 0.139338, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:25.989876+00:00", "epoch": 0, "step": 14768, "train_loss": 3.4329159259796143, "perplexity": 30.96680814836147, "lr": 0.0026291804804649314, "grad_norm": 0.156638, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:26.295335+00:00", "epoch": 0, "step": 14769, "train_loss": 3.5592079162597656, "perplexity": 35.13535597611757, "lr": 0.0026291804804649314, "grad_norm": 0.183662, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:26.600152+00:00", "epoch": 0, "step": 14770, "train_loss": 3.5605902671813965, "perplexity": 35.18395895327853, "lr": 0.0026291804804649314, "grad_norm": 0.194741, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:26.904382+00:00", "epoch": 0, "step": 14771, "train_loss": 3.4475455284118652, "perplexity": 31.423170299193444, "lr": 0.0026291804804649314, "grad_norm": 0.197246, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:27.209712+00:00", "epoch": 0, "step": 14772, "train_loss": 3.541921854019165, "perplexity": 34.533223271006065, "lr": 0.0026291804804649314, "grad_norm": 0.156521, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:27.515293+00:00", "epoch": 0, "step": 14773, "train_loss": 3.5378377437591553, "perplexity": 34.39247339418352, "lr": 0.0026291804804649314, "grad_norm": 0.18719, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:27.821340+00:00", "epoch": 0, "step": 14774, "train_loss": 3.4615001678466797, "perplexity": 31.86474313983292, "lr": 0.0026291804804649314, "grad_norm": 0.180203, "tokens_per_sec": 107067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:28.126454+00:00", "epoch": 0, "step": 14775, "train_loss": 3.43565034866333, "perplexity": 31.051600367054636, "lr": 0.0026291804804649314, "grad_norm": 0.178413, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:28.430974+00:00", "epoch": 0, "step": 14776, "train_loss": 3.421288013458252, "perplexity": 30.608814206902156, "lr": 0.0026291804804649314, "grad_norm": 0.219836, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:28.735651+00:00", "epoch": 0, "step": 14777, "train_loss": 3.4490532875061035, "perplexity": 31.470584605675356, "lr": 0.0026291804804649314, "grad_norm": 0.188921, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:29.039404+00:00", "epoch": 0, "step": 14778, "train_loss": 3.4659602642059326, "perplexity": 32.00718037046748, "lr": 0.0026291804804649314, "grad_norm": 0.186336, "tokens_per_sec": 107877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:29.344972+00:00", "epoch": 0, "step": 14779, "train_loss": 3.5785341262817383, "perplexity": 35.82099328980509, "lr": 0.0026291804804649314, "grad_norm": 0.164937, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:29.650467+00:00", "epoch": 0, "step": 14780, "train_loss": 3.488754987716675, "perplexity": 32.745154198095875, "lr": 0.0026291804804649314, "grad_norm": 0.186961, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:29.956343+00:00", "epoch": 0, "step": 14781, "train_loss": 3.5178213119506836, "perplexity": 33.71090285706634, "lr": 0.0026291804804649314, "grad_norm": 0.165977, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:30.260832+00:00", "epoch": 0, "step": 14782, "train_loss": 3.5319766998291016, "perplexity": 34.19148716666491, "lr": 0.0026291804804649314, "grad_norm": 0.175996, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:30.565722+00:00", "epoch": 0, "step": 14783, "train_loss": 3.6020727157592773, "perplexity": 36.674170851079104, "lr": 0.0026291804804649314, "grad_norm": 0.186666, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:30.871846+00:00", "epoch": 0, "step": 14784, "train_loss": 3.4715576171875, "perplexity": 32.186838192005425, "lr": 0.0026291804804649314, "grad_norm": 0.193923, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:31.177683+00:00", "epoch": 0, "step": 14785, "train_loss": 3.4831206798553467, "perplexity": 32.561176698018045, "lr": 0.0026291804804649314, "grad_norm": 0.158607, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:31.484113+00:00", "epoch": 0, "step": 14786, "train_loss": 3.5455408096313477, "perplexity": 34.65842388425608, "lr": 0.0026291804804649314, "grad_norm": 0.203001, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:31.788759+00:00", "epoch": 0, "step": 14787, "train_loss": 3.622816324234009, "perplexity": 37.442870733117594, "lr": 0.0026291804804649314, "grad_norm": 0.174642, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:32.093633+00:00", "epoch": 0, "step": 14788, "train_loss": 3.457341194152832, "perplexity": 31.73249371292407, "lr": 0.0026291804804649314, "grad_norm": 0.159031, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:32.398559+00:00", "epoch": 0, "step": 14789, "train_loss": 3.6537017822265625, "perplexity": 38.61735482242707, "lr": 0.0026291804804649314, "grad_norm": 0.175682, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:32.704651+00:00", "epoch": 0, "step": 14790, "train_loss": 3.5794289112091064, "perplexity": 35.85305971882337, "lr": 0.0026291804804649314, "grad_norm": 0.155782, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:33.010306+00:00", "epoch": 0, "step": 14791, "train_loss": 3.489488124847412, "perplexity": 32.769169688738494, "lr": 0.0026291804804649314, "grad_norm": 0.188617, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:33.314804+00:00", "epoch": 0, "step": 14792, "train_loss": 3.4772582054138184, "perplexity": 32.37084608114445, "lr": 0.0026291804804649314, "grad_norm": 0.18681, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:33.618915+00:00", "epoch": 0, "step": 14793, "train_loss": 3.547696352005005, "perplexity": 34.733212161235016, "lr": 0.0026291804804649314, "grad_norm": 0.170325, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:33.922775+00:00", "epoch": 0, "step": 14794, "train_loss": 3.5362765789031982, "perplexity": 34.33882296289156, "lr": 0.0026291804804649314, "grad_norm": 0.169317, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:34.228400+00:00", "epoch": 0, "step": 14795, "train_loss": 3.4973244667053223, "perplexity": 33.02696888689451, "lr": 0.0026291804804649314, "grad_norm": 0.169083, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:34.533235+00:00", "epoch": 0, "step": 14796, "train_loss": 3.4813485145568848, "perplexity": 32.50352401073114, "lr": 0.0026291804804649314, "grad_norm": 0.166849, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:34.839567+00:00", "epoch": 0, "step": 14797, "train_loss": 3.555635690689087, "perplexity": 35.01006846991212, "lr": 0.0026291804804649314, "grad_norm": 0.175436, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:35.145172+00:00", "epoch": 0, "step": 14798, "train_loss": 3.5386810302734375, "perplexity": 34.42148833542452, "lr": 0.0026291804804649314, "grad_norm": 0.172014, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:35.449045+00:00", "epoch": 0, "step": 14799, "train_loss": 3.532196283340454, "perplexity": 34.19899587783975, "lr": 0.0026291804804649314, "grad_norm": 0.169668, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:35.753854+00:00", "epoch": 0, "step": 14800, "train_loss": 3.590909719467163, "perplexity": 36.26705376873965, "lr": 0.0026291804804649314, "grad_norm": 0.154731, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:36.058328+00:00", "epoch": 0, "step": 14801, "train_loss": 3.3568027019500732, "perplexity": 28.697290250525477, "lr": 0.0026291804804649314, "grad_norm": 0.162825, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:36.364019+00:00", "epoch": 0, "step": 14802, "train_loss": 3.37945818901062, "perplexity": 29.354862016984487, "lr": 0.0026291804804649314, "grad_norm": 0.162656, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:36.669128+00:00", "epoch": 0, "step": 14803, "train_loss": 3.4352593421936035, "perplexity": 31.039461363784817, "lr": 0.0026291804804649314, "grad_norm": 0.14608, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:36.973488+00:00", "epoch": 0, "step": 14804, "train_loss": 3.590909004211426, "perplexity": 36.267027828530644, "lr": 0.0026291804804649314, "grad_norm": 0.162251, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:37.278901+00:00", "epoch": 0, "step": 14805, "train_loss": 3.478252649307251, "perplexity": 32.403053082728945, "lr": 0.0026291804804649314, "grad_norm": 0.173332, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:37.583604+00:00", "epoch": 0, "step": 14806, "train_loss": 3.5303170680999756, "perplexity": 34.13478895178933, "lr": 0.0026291804804649314, "grad_norm": 0.148262, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:37.889254+00:00", "epoch": 0, "step": 14807, "train_loss": 3.5439774990081787, "perplexity": 34.60428433151065, "lr": 0.0026291804804649314, "grad_norm": 0.171222, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:38.194731+00:00", "epoch": 0, "step": 14808, "train_loss": 3.519338607788086, "perplexity": 33.7620910936752, "lr": 0.0026291804804649314, "grad_norm": 0.171377, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:38.499971+00:00", "epoch": 0, "step": 14809, "train_loss": 3.4666149616241455, "perplexity": 32.02814224991574, "lr": 0.0026291804804649314, "grad_norm": 0.158181, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:38.804843+00:00", "epoch": 0, "step": 14810, "train_loss": 3.405139923095703, "perplexity": 30.11850970421009, "lr": 0.0026291804804649314, "grad_norm": 0.164841, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:39.110052+00:00", "epoch": 0, "step": 14811, "train_loss": 3.4747745990753174, "perplexity": 32.29054939658655, "lr": 0.0026291804804649314, "grad_norm": 0.197256, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:39.415249+00:00", "epoch": 0, "step": 14812, "train_loss": 3.4987237453460693, "perplexity": 33.073215167196906, "lr": 0.0026291804804649314, "grad_norm": 0.176065, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:39.720199+00:00", "epoch": 0, "step": 14813, "train_loss": 3.5445058345794678, "perplexity": 34.62257183638626, "lr": 0.0026291804804649314, "grad_norm": 0.144628, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:40.024714+00:00", "epoch": 0, "step": 14814, "train_loss": 3.416269302368164, "perplexity": 30.455582246683456, "lr": 0.0026291804804649314, "grad_norm": 0.158208, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:40.329874+00:00", "epoch": 0, "step": 14815, "train_loss": 3.5147106647491455, "perplexity": 33.606203057925306, "lr": 0.0026291804804649314, "grad_norm": 0.167608, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:40.634424+00:00", "epoch": 0, "step": 14816, "train_loss": 3.5904979705810547, "perplexity": 36.25212392363242, "lr": 0.0026291804804649314, "grad_norm": 0.168884, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:40.938953+00:00", "epoch": 0, "step": 14817, "train_loss": 3.4990482330322266, "perplexity": 33.08394875962539, "lr": 0.0026291804804649314, "grad_norm": 0.176159, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:41.243482+00:00", "epoch": 0, "step": 14818, "train_loss": 3.4896528720855713, "perplexity": 32.774568763670615, "lr": 0.0026291804804649314, "grad_norm": 0.160263, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:41.548589+00:00", "epoch": 0, "step": 14819, "train_loss": 3.5040383338928223, "perplexity": 33.24945360028895, "lr": 0.0026291804804649314, "grad_norm": 0.149667, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:41.854200+00:00", "epoch": 0, "step": 14820, "train_loss": 3.478553533554077, "perplexity": 32.41280411784347, "lr": 0.0026291804804649314, "grad_norm": 0.154758, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:42.159969+00:00", "epoch": 0, "step": 14821, "train_loss": 3.488741636276245, "perplexity": 32.74471700603882, "lr": 0.0026291804804649314, "grad_norm": 0.171262, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:42.465475+00:00", "epoch": 0, "step": 14822, "train_loss": 3.542941093444824, "perplexity": 34.56843883715703, "lr": 0.0026291804804649314, "grad_norm": 0.163356, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:42.769581+00:00", "epoch": 0, "step": 14823, "train_loss": 3.57895565032959, "perplexity": 35.836095882724095, "lr": 0.0026291804804649314, "grad_norm": 0.171468, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:43.073919+00:00", "epoch": 0, "step": 14824, "train_loss": 3.3900418281555176, "perplexity": 29.66719316838605, "lr": 0.0026291804804649314, "grad_norm": 0.178821, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:43.379232+00:00", "epoch": 0, "step": 14825, "train_loss": 3.548428773880005, "perplexity": 34.75866084405927, "lr": 0.0026291804804649314, "grad_norm": 0.200489, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:43.685785+00:00", "epoch": 0, "step": 14826, "train_loss": 3.5179715156555176, "perplexity": 33.715966739866225, "lr": 0.0026291804804649314, "grad_norm": 0.177059, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:43.990055+00:00", "epoch": 0, "step": 14827, "train_loss": 3.4760119915008545, "perplexity": 32.33053020870201, "lr": 0.0026291804804649314, "grad_norm": 0.180346, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:44.294536+00:00", "epoch": 0, "step": 14828, "train_loss": 3.4477415084838867, "perplexity": 31.429329217864773, "lr": 0.0026291804804649314, "grad_norm": 0.185388, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:44.599175+00:00", "epoch": 0, "step": 14829, "train_loss": 3.536125898361206, "perplexity": 34.33364916024163, "lr": 0.0026291804804649314, "grad_norm": 0.162928, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:44.905723+00:00", "epoch": 0, "step": 14830, "train_loss": 3.4937989711761475, "perplexity": 32.910737462812364, "lr": 0.0026291804804649314, "grad_norm": 0.176299, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:45.210208+00:00", "epoch": 0, "step": 14831, "train_loss": 3.497040033340454, "perplexity": 33.01757625085548, "lr": 0.0026291804804649314, "grad_norm": 0.183377, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:45.515192+00:00", "epoch": 0, "step": 14832, "train_loss": 3.510906219482422, "perplexity": 33.47859299443692, "lr": 0.0026291804804649314, "grad_norm": 0.157075, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:45.819711+00:00", "epoch": 0, "step": 14833, "train_loss": 3.549452543258667, "perplexity": 34.794263918233646, "lr": 0.0026291804804649314, "grad_norm": 0.201997, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:46.124875+00:00", "epoch": 0, "step": 14834, "train_loss": 3.464852809906006, "perplexity": 31.971753501377716, "lr": 0.0026291804804649314, "grad_norm": 0.164859, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:46.430227+00:00", "epoch": 0, "step": 14835, "train_loss": 3.5366201400756836, "perplexity": 34.35062247598167, "lr": 0.0026291804804649314, "grad_norm": 0.193165, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:46.735581+00:00", "epoch": 0, "step": 14836, "train_loss": 3.5638909339904785, "perplexity": 35.30028134392949, "lr": 0.0026291804804649314, "grad_norm": 0.166734, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:47.040084+00:00", "epoch": 0, "step": 14837, "train_loss": 3.5299737453460693, "perplexity": 34.12307171355474, "lr": 0.0026291804804649314, "grad_norm": 0.208075, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:47.344824+00:00", "epoch": 0, "step": 14838, "train_loss": 3.464289426803589, "perplexity": 31.953746228673275, "lr": 0.0026291804804649314, "grad_norm": 0.168738, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:47.649687+00:00", "epoch": 0, "step": 14839, "train_loss": 3.5043742656707764, "perplexity": 33.26062502466607, "lr": 0.0026291804804649314, "grad_norm": 0.192406, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:47.954914+00:00", "epoch": 0, "step": 14840, "train_loss": 3.493925094604492, "perplexity": 32.91488853961905, "lr": 0.0026291804804649314, "grad_norm": 0.17229, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:48.259043+00:00", "epoch": 0, "step": 14841, "train_loss": 3.4945547580718994, "perplexity": 32.93562036881898, "lr": 0.0026291804804649314, "grad_norm": 0.171478, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:48.564408+00:00", "epoch": 0, "step": 14842, "train_loss": 3.5360984802246094, "perplexity": 34.3327078084642, "lr": 0.0026291804804649314, "grad_norm": 0.159963, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:48.868817+00:00", "epoch": 0, "step": 14843, "train_loss": 3.492435932159424, "perplexity": 32.865909401719975, "lr": 0.0026291804804649314, "grad_norm": 0.169049, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:49.173201+00:00", "epoch": 0, "step": 14844, "train_loss": 3.5169677734375, "perplexity": 33.682141579338875, "lr": 0.0026291804804649314, "grad_norm": 0.16743, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:49.478192+00:00", "epoch": 0, "step": 14845, "train_loss": 3.512982130050659, "perplexity": 33.548163745788735, "lr": 0.0026291804804649314, "grad_norm": 0.174968, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:49.782216+00:00", "epoch": 0, "step": 14846, "train_loss": 3.492913246154785, "perplexity": 32.88160050474132, "lr": 0.0026291804804649314, "grad_norm": 0.161179, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:50.086717+00:00", "epoch": 0, "step": 14847, "train_loss": 3.3944263458251953, "perplexity": 29.797555079253744, "lr": 0.0026291804804649314, "grad_norm": 0.158687, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:50.391337+00:00", "epoch": 0, "step": 14848, "train_loss": 3.5517687797546387, "perplexity": 34.87494906904436, "lr": 0.0026291804804649314, "grad_norm": 0.19056, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:50.696354+00:00", "epoch": 0, "step": 14849, "train_loss": 3.5872368812561035, "perplexity": 36.13409506522349, "lr": 0.0026291804804649314, "grad_norm": 0.192103, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:51.000824+00:00", "epoch": 0, "step": 14850, "train_loss": 3.5177831649780273, "perplexity": 33.70961691270443, "lr": 0.0026291804804649314, "grad_norm": 0.162401, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:51.305036+00:00", "epoch": 0, "step": 14851, "train_loss": 3.501156806945801, "perplexity": 33.153782309665104, "lr": 0.0026291804804649314, "grad_norm": 0.15889, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:51.608753+00:00", "epoch": 0, "step": 14852, "train_loss": 3.568878650665283, "perplexity": 35.47678896489654, "lr": 0.0026291804804649314, "grad_norm": 0.192351, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:51.912849+00:00", "epoch": 0, "step": 14853, "train_loss": 3.4005849361419678, "perplexity": 29.981632259588665, "lr": 0.0026291804804649314, "grad_norm": 0.192725, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:52.217721+00:00", "epoch": 0, "step": 14854, "train_loss": 3.5723540782928467, "perplexity": 35.600300480749176, "lr": 0.0026291804804649314, "grad_norm": 0.15363, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:52.523516+00:00", "epoch": 0, "step": 14855, "train_loss": 3.5916144847869873, "perplexity": 36.29262253942116, "lr": 0.0026291804804649314, "grad_norm": 0.174916, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:52.828930+00:00", "epoch": 0, "step": 14856, "train_loss": 3.454526424407959, "perplexity": 31.6432996389718, "lr": 0.0026291804804649314, "grad_norm": 0.175219, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:53.132987+00:00", "epoch": 0, "step": 14857, "train_loss": 3.459773302078247, "perplexity": 31.809764489721776, "lr": 0.0026291804804649314, "grad_norm": 0.165766, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:53.436646+00:00", "epoch": 0, "step": 14858, "train_loss": 3.617940902709961, "perplexity": 37.26076523646869, "lr": 0.0026291804804649314, "grad_norm": 0.171597, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:53.740724+00:00", "epoch": 0, "step": 14859, "train_loss": 3.471674919128418, "perplexity": 32.19061399204737, "lr": 0.0026291804804649314, "grad_norm": 0.171271, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:54.045505+00:00", "epoch": 0, "step": 14860, "train_loss": 3.5889976024627686, "perplexity": 36.19777317593753, "lr": 0.0026291804804649314, "grad_norm": 0.154001, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:54.350114+00:00", "epoch": 0, "step": 14861, "train_loss": 3.418937921524048, "perplexity": 30.5369651385451, "lr": 0.0026291804804649314, "grad_norm": 0.158658, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:54.654140+00:00", "epoch": 0, "step": 14862, "train_loss": 3.5359082221984863, "perplexity": 34.326176356595376, "lr": 0.0026291804804649314, "grad_norm": 0.144849, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:54.958256+00:00", "epoch": 0, "step": 14863, "train_loss": 3.512516498565674, "perplexity": 33.53254630075448, "lr": 0.0026291804804649314, "grad_norm": 0.159135, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:55.263301+00:00", "epoch": 0, "step": 14864, "train_loss": 3.5067386627197266, "perplexity": 33.339359391293144, "lr": 0.0026291804804649314, "grad_norm": 0.175421, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:55.568232+00:00", "epoch": 0, "step": 14865, "train_loss": 3.524364471435547, "perplexity": 33.93220187869721, "lr": 0.0026291804804649314, "grad_norm": 0.163115, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:55.874384+00:00", "epoch": 0, "step": 14866, "train_loss": 3.563654661178589, "perplexity": 35.291941832434574, "lr": 0.0026291804804649314, "grad_norm": 0.167451, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:56.180956+00:00", "epoch": 0, "step": 14867, "train_loss": 3.4998843669891357, "perplexity": 33.11162294066091, "lr": 0.0026291804804649314, "grad_norm": 0.19581, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:56.485720+00:00", "epoch": 0, "step": 14868, "train_loss": 3.496305227279663, "perplexity": 32.99332364728501, "lr": 0.0026291804804649314, "grad_norm": 0.202883, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:56.789514+00:00", "epoch": 0, "step": 14869, "train_loss": 3.5298657417297363, "perplexity": 34.1193864974212, "lr": 0.0026291804804649314, "grad_norm": 0.168029, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:57.095032+00:00", "epoch": 0, "step": 14870, "train_loss": 3.5601329803466797, "perplexity": 35.167873470176104, "lr": 0.0026291804804649314, "grad_norm": 0.192889, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:57.399681+00:00", "epoch": 0, "step": 14871, "train_loss": 3.4700331687927246, "perplexity": 32.13780839938026, "lr": 0.0026291804804649314, "grad_norm": 0.15836, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:57.705044+00:00", "epoch": 0, "step": 14872, "train_loss": 3.3880155086517334, "perplexity": 29.607138821442113, "lr": 0.0026291804804649314, "grad_norm": 0.156611, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:58.009347+00:00", "epoch": 0, "step": 14873, "train_loss": 3.511136770248413, "perplexity": 33.486312399519285, "lr": 0.0026291804804649314, "grad_norm": 0.170045, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:58.314163+00:00", "epoch": 0, "step": 14874, "train_loss": 3.4491279125213623, "perplexity": 31.472933186162095, "lr": 0.0026291804804649314, "grad_norm": 0.170561, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:58.618848+00:00", "epoch": 0, "step": 14875, "train_loss": 3.491616725921631, "perplexity": 32.83899646885369, "lr": 0.0026291804804649314, "grad_norm": 0.17045, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:58.923483+00:00", "epoch": 0, "step": 14876, "train_loss": 3.4668195247650146, "perplexity": 32.03469469746236, "lr": 0.0026291804804649314, "grad_norm": 0.157642, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:59.228230+00:00", "epoch": 0, "step": 14877, "train_loss": 3.473939895629883, "perplexity": 32.26360760951575, "lr": 0.0026291804804649314, "grad_norm": 0.153491, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:59.533543+00:00", "epoch": 0, "step": 14878, "train_loss": 3.521455764770508, "perplexity": 33.83364646081989, "lr": 0.0026291804804649314, "grad_norm": 0.176424, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:15:59.837710+00:00", "epoch": 0, "step": 14879, "train_loss": 3.4494271278381348, "perplexity": 31.48235177885848, "lr": 0.0026291804804649314, "grad_norm": 0.167624, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:00.143098+00:00", "epoch": 0, "step": 14880, "train_loss": 3.507277488708496, "perplexity": 33.357328345217034, "lr": 0.0026291804804649314, "grad_norm": 0.182324, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:00.447991+00:00", "epoch": 0, "step": 14881, "train_loss": 3.469147205352783, "perplexity": 32.109348085359045, "lr": 0.0026291804804649314, "grad_norm": 0.195951, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:00.754107+00:00", "epoch": 0, "step": 14882, "train_loss": 3.5762038230895996, "perplexity": 35.73761669889433, "lr": 0.0026291804804649314, "grad_norm": 0.169637, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:01.058959+00:00", "epoch": 0, "step": 14883, "train_loss": 3.4615509510040283, "perplexity": 31.866361373186816, "lr": 0.0026291804804649314, "grad_norm": 0.168146, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:01.365105+00:00", "epoch": 0, "step": 14884, "train_loss": 3.480558395385742, "perplexity": 32.47785249639392, "lr": 0.0026291804804649314, "grad_norm": 0.180322, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:01.669036+00:00", "epoch": 0, "step": 14885, "train_loss": 3.6411995887756348, "perplexity": 38.137558683469095, "lr": 0.0026291804804649314, "grad_norm": 0.167419, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:01.973420+00:00", "epoch": 0, "step": 14886, "train_loss": 3.4947032928466797, "perplexity": 32.940512817113074, "lr": 0.0026291804804649314, "grad_norm": 0.16581, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:02.279116+00:00", "epoch": 0, "step": 14887, "train_loss": 3.590235471725464, "perplexity": 36.24260903146851, "lr": 0.0026291804804649314, "grad_norm": 0.172736, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:02.584794+00:00", "epoch": 0, "step": 14888, "train_loss": 3.567898988723755, "perplexity": 35.442050723584764, "lr": 0.0026291804804649314, "grad_norm": 0.171069, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:02.889184+00:00", "epoch": 0, "step": 14889, "train_loss": 3.556353807449341, "perplexity": 35.03521881621979, "lr": 0.0026291804804649314, "grad_norm": 0.168598, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:03.194370+00:00", "epoch": 0, "step": 14890, "train_loss": 3.538508653640747, "perplexity": 34.415555386538635, "lr": 0.0026291804804649314, "grad_norm": 0.150757, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:03.498004+00:00", "epoch": 0, "step": 14891, "train_loss": 3.433778762817383, "perplexity": 30.99353898169544, "lr": 0.0026291804804649314, "grad_norm": 0.166458, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:03.802826+00:00", "epoch": 0, "step": 14892, "train_loss": 3.5267257690429688, "perplexity": 34.01242057864694, "lr": 0.0026291804804649314, "grad_norm": 0.164648, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:04.108488+00:00", "epoch": 0, "step": 14893, "train_loss": 3.5949392318725586, "perplexity": 36.413487141217274, "lr": 0.0026291804804649314, "grad_norm": 0.162408, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:04.413549+00:00", "epoch": 0, "step": 14894, "train_loss": 3.495786428451538, "perplexity": 32.976211188986326, "lr": 0.0026291804804649314, "grad_norm": 0.15632, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:04.717033+00:00", "epoch": 0, "step": 14895, "train_loss": 3.461862802505493, "perplexity": 31.876300495511828, "lr": 0.0026291804804649314, "grad_norm": 0.16568, "tokens_per_sec": 107973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:05.022134+00:00", "epoch": 0, "step": 14896, "train_loss": 3.529852867126465, "perplexity": 34.1189472266839, "lr": 0.0026291804804649314, "grad_norm": 0.180319, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:05.326815+00:00", "epoch": 0, "step": 14897, "train_loss": 3.5105247497558594, "perplexity": 33.46582436030017, "lr": 0.0026291804804649314, "grad_norm": 0.167364, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:05.632401+00:00", "epoch": 0, "step": 14898, "train_loss": 3.4666800498962402, "perplexity": 32.03022697419801, "lr": 0.0026291804804649314, "grad_norm": 0.164121, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:05.938200+00:00", "epoch": 0, "step": 14899, "train_loss": 3.580507278442383, "perplexity": 35.89174333746156, "lr": 0.0026291804804649314, "grad_norm": 0.159258, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:06.242427+00:00", "epoch": 0, "step": 14900, "train_loss": 3.5162363052368164, "perplexity": 33.65751317238903, "lr": 0.0026291804804649314, "grad_norm": 0.148511, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:06.546643+00:00", "epoch": 0, "step": 14901, "train_loss": 3.460683584213257, "perplexity": 31.838733533056597, "lr": 0.0026291804804649314, "grad_norm": 0.167036, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:06.851255+00:00", "epoch": 0, "step": 14902, "train_loss": 3.5383129119873047, "perplexity": 34.40881948809259, "lr": 0.0026291804804649314, "grad_norm": 0.157237, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:07.156263+00:00", "epoch": 0, "step": 14903, "train_loss": 3.541255474090576, "perplexity": 34.51021868989564, "lr": 0.0026291804804649314, "grad_norm": 0.177405, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:07.462331+00:00", "epoch": 0, "step": 14904, "train_loss": 3.519789695739746, "perplexity": 33.777324201667845, "lr": 0.0026291804804649314, "grad_norm": 0.179501, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:07.767440+00:00", "epoch": 0, "step": 14905, "train_loss": 3.4729297161102295, "perplexity": 32.23103203024018, "lr": 0.0026291804804649314, "grad_norm": 0.166152, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:08.072284+00:00", "epoch": 0, "step": 14906, "train_loss": 3.5917646884918213, "perplexity": 36.298074235206926, "lr": 0.0026291804804649314, "grad_norm": 0.157486, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:08.377432+00:00", "epoch": 0, "step": 14907, "train_loss": 3.470536470413208, "perplexity": 32.153987481564066, "lr": 0.0026291804804649314, "grad_norm": 0.162058, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:08.683023+00:00", "epoch": 0, "step": 14908, "train_loss": 3.5329291820526123, "perplexity": 34.22406946495324, "lr": 0.0026291804804649314, "grad_norm": 0.162853, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:08.988184+00:00", "epoch": 0, "step": 14909, "train_loss": 3.5164906978607178, "perplexity": 33.66607648465451, "lr": 0.0026291804804649314, "grad_norm": 0.164315, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:09.293379+00:00", "epoch": 0, "step": 14910, "train_loss": 3.41080379486084, "perplexity": 30.289581086795938, "lr": 0.0026291804804649314, "grad_norm": 0.181183, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:09.597552+00:00", "epoch": 0, "step": 14911, "train_loss": 3.4138336181640625, "perplexity": 30.381492332523848, "lr": 0.0026291804804649314, "grad_norm": 0.163773, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:09.902597+00:00", "epoch": 0, "step": 14912, "train_loss": 3.470797538757324, "perplexity": 32.1623829656825, "lr": 0.0026291804804649314, "grad_norm": 0.147109, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:10.207222+00:00", "epoch": 0, "step": 14913, "train_loss": 3.477034568786621, "perplexity": 32.36360758373409, "lr": 0.0026291804804649314, "grad_norm": 0.154737, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:10.513470+00:00", "epoch": 0, "step": 14914, "train_loss": 3.4210174083709717, "perplexity": 30.60053242665862, "lr": 0.0026291804804649314, "grad_norm": 0.171039, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:10.818689+00:00", "epoch": 0, "step": 14915, "train_loss": 3.5221221446990967, "perplexity": 33.85620003752317, "lr": 0.0026291804804649314, "grad_norm": 0.168988, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:11.123796+00:00", "epoch": 0, "step": 14916, "train_loss": 3.4954891204833984, "perplexity": 32.96640855591345, "lr": 0.0026291804804649314, "grad_norm": 0.154278, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:11.428340+00:00", "epoch": 0, "step": 14917, "train_loss": 3.5184974670410156, "perplexity": 33.733704363441426, "lr": 0.0026291804804649314, "grad_norm": 0.15518, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:11.732840+00:00", "epoch": 0, "step": 14918, "train_loss": 3.493725299835205, "perplexity": 32.90831297396085, "lr": 0.0026291804804649314, "grad_norm": 0.145692, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:12.037637+00:00", "epoch": 0, "step": 14919, "train_loss": 3.5102410316467285, "perplexity": 33.4563308466967, "lr": 0.0026291804804649314, "grad_norm": 0.160487, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:12.342871+00:00", "epoch": 0, "step": 14920, "train_loss": 3.4764273166656494, "perplexity": 32.343960680302196, "lr": 0.0026291804804649314, "grad_norm": 0.149605, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:12.647095+00:00", "epoch": 0, "step": 14921, "train_loss": 3.509557008743286, "perplexity": 33.43345377524483, "lr": 0.0026291804804649314, "grad_norm": 0.157355, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:12.950635+00:00", "epoch": 0, "step": 14922, "train_loss": 3.5362064838409424, "perplexity": 34.33641606531499, "lr": 0.0026291804804649314, "grad_norm": 0.157101, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:13.254980+00:00", "epoch": 0, "step": 14923, "train_loss": 3.532968521118164, "perplexity": 34.22541583434765, "lr": 0.0026291804804649314, "grad_norm": 0.157519, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:13.559299+00:00", "epoch": 0, "step": 14924, "train_loss": 3.5110647678375244, "perplexity": 33.483901391094996, "lr": 0.0026291804804649314, "grad_norm": 0.170647, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:13.864708+00:00", "epoch": 0, "step": 14925, "train_loss": 3.4178860187530518, "perplexity": 30.504860108943205, "lr": 0.0026291804804649314, "grad_norm": 0.152366, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:14.169485+00:00", "epoch": 0, "step": 14926, "train_loss": 3.4263792037963867, "perplexity": 30.76504687371071, "lr": 0.0026291804804649314, "grad_norm": 0.162049, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:14.473367+00:00", "epoch": 0, "step": 14927, "train_loss": 3.473667860031128, "perplexity": 32.25483195340126, "lr": 0.0026291804804649314, "grad_norm": 0.184787, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:14.777136+00:00", "epoch": 0, "step": 14928, "train_loss": 3.5180046558380127, "perplexity": 33.71708411167184, "lr": 0.0026291804804649314, "grad_norm": 0.168596, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:15.082100+00:00", "epoch": 0, "step": 14929, "train_loss": 3.4376049041748047, "perplexity": 31.112351795371033, "lr": 0.0026291804804649314, "grad_norm": 0.160689, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:15.387594+00:00", "epoch": 0, "step": 14930, "train_loss": 3.5585103034973145, "perplexity": 35.11085365093644, "lr": 0.0026291804804649314, "grad_norm": 0.177478, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:15.692256+00:00", "epoch": 0, "step": 14931, "train_loss": 3.5819153785705566, "perplexity": 35.94231810466711, "lr": 0.0026291804804649314, "grad_norm": 0.181087, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:15.996455+00:00", "epoch": 0, "step": 14932, "train_loss": 3.4905877113342285, "perplexity": 32.80522204262045, "lr": 0.0026291804804649314, "grad_norm": 0.163709, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:16.300307+00:00", "epoch": 0, "step": 14933, "train_loss": 3.6353354454040527, "perplexity": 37.91456903216685, "lr": 0.0026291804804649314, "grad_norm": 0.183476, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:16.604694+00:00", "epoch": 0, "step": 14934, "train_loss": 3.4465644359588623, "perplexity": 31.392356382085882, "lr": 0.0026291804804649314, "grad_norm": 0.205551, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:16.910895+00:00", "epoch": 0, "step": 14935, "train_loss": 3.5058274269104004, "perplexity": 33.30899321063262, "lr": 0.0026291804804649314, "grad_norm": 0.209765, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:17.216772+00:00", "epoch": 0, "step": 14936, "train_loss": 3.599719524383545, "perplexity": 36.58797097070547, "lr": 0.0026291804804649314, "grad_norm": 0.182409, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:17.521816+00:00", "epoch": 0, "step": 14937, "train_loss": 3.549342155456543, "perplexity": 34.79042326789756, "lr": 0.0026291804804649314, "grad_norm": 0.159607, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:17.825763+00:00", "epoch": 0, "step": 14938, "train_loss": 3.555054187774658, "perplexity": 34.98971593116632, "lr": 0.0026291804804649314, "grad_norm": 0.174022, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:18.130781+00:00", "epoch": 0, "step": 14939, "train_loss": 3.4305591583251953, "perplexity": 30.89391250933959, "lr": 0.0026291804804649314, "grad_norm": 0.157724, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:18.436151+00:00", "epoch": 0, "step": 14940, "train_loss": 3.5733635425567627, "perplexity": 35.636255856649434, "lr": 0.0026291804804649314, "grad_norm": 0.183476, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:18.741700+00:00", "epoch": 0, "step": 14941, "train_loss": 3.5231740474700928, "perplexity": 33.89183220566066, "lr": 0.0026291804804649314, "grad_norm": 0.200408, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:19.046243+00:00", "epoch": 0, "step": 14942, "train_loss": 3.6259374618530273, "perplexity": 37.55991765027725, "lr": 0.0026291804804649314, "grad_norm": 0.161318, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:19.350213+00:00", "epoch": 0, "step": 14943, "train_loss": 3.6103129386901855, "perplexity": 36.97762273321484, "lr": 0.0026291804804649314, "grad_norm": 0.175062, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:19.653561+00:00", "epoch": 0, "step": 14944, "train_loss": 3.5214028358459473, "perplexity": 33.83185572968991, "lr": 0.0026291804804649314, "grad_norm": 0.184844, "tokens_per_sec": 108022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:19.958245+00:00", "epoch": 0, "step": 14945, "train_loss": 3.536792755126953, "perplexity": 34.35655242222577, "lr": 0.0026291804804649314, "grad_norm": 0.172842, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:20.263408+00:00", "epoch": 0, "step": 14946, "train_loss": 3.4643397331237793, "perplexity": 31.9553537444961, "lr": 0.0026291804804649314, "grad_norm": 0.222044, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:20.568495+00:00", "epoch": 0, "step": 14947, "train_loss": 3.4542057514190674, "perplexity": 31.63315411428298, "lr": 0.0026291804804649314, "grad_norm": 0.171959, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:20.873178+00:00", "epoch": 0, "step": 14948, "train_loss": 3.4860517978668213, "perplexity": 32.65675736018123, "lr": 0.0026291804804649314, "grad_norm": 0.168894, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:21.177956+00:00", "epoch": 0, "step": 14949, "train_loss": 3.446810007095337, "perplexity": 31.400066385357633, "lr": 0.0026291804804649314, "grad_norm": 0.175367, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:21.482757+00:00", "epoch": 0, "step": 14950, "train_loss": 3.41314435005188, "perplexity": 30.36055855398031, "lr": 0.0026291804804649314, "grad_norm": 0.204176, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:21.787473+00:00", "epoch": 0, "step": 14951, "train_loss": 3.4601898193359375, "perplexity": 31.823016565262304, "lr": 0.0026291804804649314, "grad_norm": 0.174919, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:22.092917+00:00", "epoch": 0, "step": 14952, "train_loss": 3.4105727672576904, "perplexity": 30.282584165749125, "lr": 0.0026291804804649314, "grad_norm": 0.155856, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:22.397030+00:00", "epoch": 0, "step": 14953, "train_loss": 3.5665478706359863, "perplexity": 35.3941966633076, "lr": 0.0026291804804649314, "grad_norm": 0.225252, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:22.700997+00:00", "epoch": 0, "step": 14954, "train_loss": 3.5517520904541016, "perplexity": 34.87436703539502, "lr": 0.0026291804804649314, "grad_norm": 0.175404, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:23.005533+00:00", "epoch": 0, "step": 14955, "train_loss": 3.5588912963867188, "perplexity": 35.124233185110214, "lr": 0.0026291804804649314, "grad_norm": 0.187334, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:23.311032+00:00", "epoch": 0, "step": 14956, "train_loss": 3.618283987045288, "perplexity": 37.27355101451829, "lr": 0.0026291804804649314, "grad_norm": 0.169875, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:23.615646+00:00", "epoch": 0, "step": 14957, "train_loss": 3.5019328594207764, "perplexity": 33.1795213706205, "lr": 0.0026291804804649314, "grad_norm": 0.187106, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:23.920597+00:00", "epoch": 0, "step": 14958, "train_loss": 3.34894061088562, "perplexity": 28.47255414675286, "lr": 0.0026291804804649314, "grad_norm": 0.150572, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:24.225203+00:00", "epoch": 0, "step": 14959, "train_loss": 3.448789596557617, "perplexity": 31.462287191395866, "lr": 0.0026291804804649314, "grad_norm": 0.17256, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:24.529419+00:00", "epoch": 0, "step": 14960, "train_loss": 3.472712993621826, "perplexity": 32.224047597643924, "lr": 0.0026291804804649314, "grad_norm": 0.181153, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:24.833295+00:00", "epoch": 0, "step": 14961, "train_loss": 3.5870249271392822, "perplexity": 36.1264371066134, "lr": 0.0026291804804649314, "grad_norm": 0.183677, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:25.137629+00:00", "epoch": 0, "step": 14962, "train_loss": 3.500466823577881, "perplexity": 33.13091464135042, "lr": 0.0026291804804649314, "grad_norm": 0.149105, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:25.442121+00:00", "epoch": 0, "step": 14963, "train_loss": 3.4320390224456787, "perplexity": 30.93966514749748, "lr": 0.0026291804804649314, "grad_norm": 0.18275, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:25.747166+00:00", "epoch": 0, "step": 14964, "train_loss": 3.4987895488739014, "perplexity": 33.07539157303846, "lr": 0.0026291804804649314, "grad_norm": 0.175611, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:26.052238+00:00", "epoch": 0, "step": 14965, "train_loss": 3.5282602310180664, "perplexity": 34.06465140756791, "lr": 0.0026291804804649314, "grad_norm": 0.175613, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:26.356602+00:00", "epoch": 0, "step": 14966, "train_loss": 3.482511281967163, "perplexity": 32.541340030527024, "lr": 0.0026291804804649314, "grad_norm": 0.174857, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:26.662649+00:00", "epoch": 0, "step": 14967, "train_loss": 3.459190845489502, "perplexity": 31.791242077582453, "lr": 0.0026291804804649314, "grad_norm": 0.163636, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:26.967852+00:00", "epoch": 0, "step": 14968, "train_loss": 3.537318229675293, "perplexity": 34.37461066024906, "lr": 0.0026291804804649314, "grad_norm": 0.182167, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:27.272094+00:00", "epoch": 0, "step": 14969, "train_loss": 3.508138418197632, "perplexity": 33.38605901863561, "lr": 0.0026291804804649314, "grad_norm": 0.161302, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:27.577185+00:00", "epoch": 0, "step": 14970, "train_loss": 3.5094165802001953, "perplexity": 33.42875909368203, "lr": 0.0026291804804649314, "grad_norm": 0.153298, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:27.881840+00:00", "epoch": 0, "step": 14971, "train_loss": 3.5286452770233154, "perplexity": 34.0777703910566, "lr": 0.0026291804804649314, "grad_norm": 0.157715, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:28.187790+00:00", "epoch": 0, "step": 14972, "train_loss": 3.531769037246704, "perplexity": 34.184387611325384, "lr": 0.0026291804804649314, "grad_norm": 0.167965, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:28.492794+00:00", "epoch": 0, "step": 14973, "train_loss": 3.6613802909851074, "perplexity": 38.91501986898278, "lr": 0.0026291804804649314, "grad_norm": 0.18176, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:28.798273+00:00", "epoch": 0, "step": 14974, "train_loss": 3.6290483474731445, "perplexity": 37.676944191702646, "lr": 0.0026291804804649314, "grad_norm": 0.161282, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:29.102858+00:00", "epoch": 0, "step": 14975, "train_loss": 3.3840181827545166, "perplexity": 29.48902566435316, "lr": 0.0026291804804649314, "grad_norm": 0.164515, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:29.407924+00:00", "epoch": 0, "step": 14976, "train_loss": 3.541779041290283, "perplexity": 34.52829183929764, "lr": 0.0026291804804649314, "grad_norm": 0.160099, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:29.712852+00:00", "epoch": 0, "step": 14977, "train_loss": 3.5385210514068604, "perplexity": 34.4159820651899, "lr": 0.0026291804804649314, "grad_norm": 0.173207, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:30.018421+00:00", "epoch": 0, "step": 14978, "train_loss": 3.5207414627075195, "perplexity": 33.80948764672709, "lr": 0.0026291804804649314, "grad_norm": 0.148897, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:30.323401+00:00", "epoch": 0, "step": 14979, "train_loss": 3.4902968406677246, "perplexity": 32.795681353440685, "lr": 0.0026291804804649314, "grad_norm": 0.1792, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:30.627881+00:00", "epoch": 0, "step": 14980, "train_loss": 3.5703442096710205, "perplexity": 35.52882041072266, "lr": 0.0026291804804649314, "grad_norm": 0.160651, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:30.931812+00:00", "epoch": 0, "step": 14981, "train_loss": 3.5399301052093506, "perplexity": 34.464510216944554, "lr": 0.0026291804804649314, "grad_norm": 0.174665, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:31.236376+00:00", "epoch": 0, "step": 14982, "train_loss": 3.5561256408691406, "perplexity": 35.02722586205311, "lr": 0.0026291804804649314, "grad_norm": 0.164073, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:31.541925+00:00", "epoch": 0, "step": 14983, "train_loss": 3.5253005027770996, "perplexity": 33.963978352711266, "lr": 0.0026291804804649314, "grad_norm": 0.186422, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:31.847121+00:00", "epoch": 0, "step": 14984, "train_loss": 3.481297731399536, "perplexity": 32.50187342106835, "lr": 0.0026291804804649314, "grad_norm": 0.209391, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:32.151074+00:00", "epoch": 0, "step": 14985, "train_loss": 3.4942591190338135, "perplexity": 32.925884752879355, "lr": 0.0026291804804649314, "grad_norm": 0.161502, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:32.455183+00:00", "epoch": 0, "step": 14986, "train_loss": 3.5220603942871094, "perplexity": 33.854109467769966, "lr": 0.0026291804804649314, "grad_norm": 0.177908, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:32.758963+00:00", "epoch": 0, "step": 14987, "train_loss": 3.5126051902770996, "perplexity": 33.53552049156545, "lr": 0.0026291804804649314, "grad_norm": 0.196754, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:33.064773+00:00", "epoch": 0, "step": 14988, "train_loss": 3.64805006980896, "perplexity": 38.399716233315196, "lr": 0.0026291804804649314, "grad_norm": 0.16991, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:33.370032+00:00", "epoch": 0, "step": 14989, "train_loss": 3.5696544647216797, "perplexity": 35.504323035729875, "lr": 0.0026291804804649314, "grad_norm": 0.147258, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:33.675123+00:00", "epoch": 0, "step": 14990, "train_loss": 3.518256902694702, "perplexity": 33.72559021292929, "lr": 0.0026291804804649314, "grad_norm": 0.165929, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:33.979708+00:00", "epoch": 0, "step": 14991, "train_loss": 3.5043463706970215, "perplexity": 33.259697233344355, "lr": 0.0026291804804649314, "grad_norm": 0.159435, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:34.284596+00:00", "epoch": 0, "step": 14992, "train_loss": 3.501432180404663, "perplexity": 33.16291323852411, "lr": 0.0026291804804649314, "grad_norm": 0.179351, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:34.589654+00:00", "epoch": 0, "step": 14993, "train_loss": 3.489753246307373, "perplexity": 32.777858650612316, "lr": 0.0026291804804649314, "grad_norm": 0.189617, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:34.896281+00:00", "epoch": 0, "step": 14994, "train_loss": 3.530229330062866, "perplexity": 34.13179416378989, "lr": 0.0026291804804649314, "grad_norm": 0.159577, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:35.201307+00:00", "epoch": 0, "step": 14995, "train_loss": 3.6050126552581787, "perplexity": 36.78214934189504, "lr": 0.0026291804804649314, "grad_norm": 0.191941, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:35.505751+00:00", "epoch": 0, "step": 14996, "train_loss": 3.4315598011016846, "perplexity": 30.92484175171028, "lr": 0.0026291804804649314, "grad_norm": 0.161867, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:35.810370+00:00", "epoch": 0, "step": 14997, "train_loss": 3.389268398284912, "perplexity": 29.644256546090382, "lr": 0.0026291804804649314, "grad_norm": 0.177003, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:36.115361+00:00", "epoch": 0, "step": 14998, "train_loss": 3.5312485694885254, "perplexity": 34.16660036898521, "lr": 0.0026291804804649314, "grad_norm": 0.162684, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:36.420374+00:00", "epoch": 0, "step": 14999, "train_loss": 3.5689046382904053, "perplexity": 35.47771093436854, "lr": 0.0026291804804649314, "grad_norm": 0.13826, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:36.725816+00:00", "epoch": 0, "step": 15000, "train_loss": 3.5401177406311035, "perplexity": 34.47097758658946, "lr": 0.0026291804804649314, "grad_norm": 0.163077, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:16:39.799010+00:00", "step": 15000, "epoch": 0, "val_loss": 3.4733362436294555, "val_ppl": 32.24413749541916, "eval_train_loss": 3.5401177406311035, "eval_train_ppl": 34.47097758658946} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:16:40.756761+00:00", "step": 15000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4733_epoch_0000_step_0015000.pt", "val_loss": 3.4733362436294555} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T22:16:42.651377+00:00", "step": 15000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0015000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:43.912384+00:00", "epoch": 0, "step": 15001, "train_loss": 3.4722323417663574, "perplexity": 32.20856277107883, "lr": 0.0026291804804649314, "grad_norm": 0.167624, "tokens_per_sec": 4560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:44.215997+00:00", "epoch": 0, "step": 15002, "train_loss": 3.594813108444214, "perplexity": 36.4088948369857, "lr": 0.0026291804804649314, "grad_norm": 0.190849, "tokens_per_sec": 107918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:44.519089+00:00", "epoch": 0, "step": 15003, "train_loss": 3.555746555328369, "perplexity": 35.01395006368605, "lr": 0.0026291804804649314, "grad_norm": 0.184124, "tokens_per_sec": 108113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:44.823217+00:00", "epoch": 0, "step": 15004, "train_loss": 3.4877612590789795, "perplexity": 32.71263056313362, "lr": 0.0026291804804649314, "grad_norm": 0.197942, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:45.127314+00:00", "epoch": 0, "step": 15005, "train_loss": 3.3680102825164795, "perplexity": 29.02072652654818, "lr": 0.0026291804804649314, "grad_norm": 0.189593, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:45.432270+00:00", "epoch": 0, "step": 15006, "train_loss": 3.581799268722534, "perplexity": 35.93814508984315, "lr": 0.0026291804804649314, "grad_norm": 0.186707, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:45.736627+00:00", "epoch": 0, "step": 15007, "train_loss": 3.404550075531006, "perplexity": 30.10074961299695, "lr": 0.0026291804804649314, "grad_norm": 0.189504, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:46.040873+00:00", "epoch": 0, "step": 15008, "train_loss": 3.584601879119873, "perplexity": 36.03900698120638, "lr": 0.0026291804804649314, "grad_norm": 0.181786, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:46.344595+00:00", "epoch": 0, "step": 15009, "train_loss": 3.379969596862793, "perplexity": 29.36987816327971, "lr": 0.0026291804804649314, "grad_norm": 0.187276, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:46.649581+00:00", "epoch": 0, "step": 15010, "train_loss": 3.558032989501953, "perplexity": 35.094098748092684, "lr": 0.0026291804804649314, "grad_norm": 0.155522, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:46.955283+00:00", "epoch": 0, "step": 15011, "train_loss": 3.4315361976623535, "perplexity": 30.924111827698564, "lr": 0.0026291804804649314, "grad_norm": 0.167429, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:47.259777+00:00", "epoch": 0, "step": 15012, "train_loss": 3.5959372520446777, "perplexity": 36.449846676678355, "lr": 0.0026291804804649314, "grad_norm": 0.146082, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:47.563144+00:00", "epoch": 0, "step": 15013, "train_loss": 3.4406425952911377, "perplexity": 31.207005201147943, "lr": 0.0026291804804649314, "grad_norm": 0.159632, "tokens_per_sec": 108014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:47.867077+00:00", "epoch": 0, "step": 15014, "train_loss": 3.4622912406921387, "perplexity": 31.88996044590633, "lr": 0.0026291804804649314, "grad_norm": 0.160613, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:48.175087+00:00", "epoch": 0, "step": 15015, "train_loss": 3.483118772506714, "perplexity": 32.56111459256142, "lr": 0.0026291804804649314, "grad_norm": 0.160105, "tokens_per_sec": 106386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:48.478556+00:00", "epoch": 0, "step": 15016, "train_loss": 3.5462214946746826, "perplexity": 34.682023386022394, "lr": 0.0026291804804649314, "grad_norm": 0.142202, "tokens_per_sec": 107978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:48.781835+00:00", "epoch": 0, "step": 15017, "train_loss": 3.5443060398101807, "perplexity": 34.61565511861911, "lr": 0.0026291804804649314, "grad_norm": 0.151623, "tokens_per_sec": 108045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:49.085080+00:00", "epoch": 0, "step": 15018, "train_loss": 3.559202194213867, "perplexity": 35.135154930573215, "lr": 0.0026291804804649314, "grad_norm": 0.18827, "tokens_per_sec": 108058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:49.389414+00:00", "epoch": 0, "step": 15019, "train_loss": 3.5528528690338135, "perplexity": 34.912777128233806, "lr": 0.0026291804804649314, "grad_norm": 0.212987, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:49.694871+00:00", "epoch": 0, "step": 15020, "train_loss": 3.5136046409606934, "perplexity": 33.56905434537055, "lr": 0.0026291804804649314, "grad_norm": 0.198649, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:49.998589+00:00", "epoch": 0, "step": 15021, "train_loss": 3.5177416801452637, "perplexity": 33.70821850389086, "lr": 0.0026291804804649314, "grad_norm": 0.178776, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:50.302219+00:00", "epoch": 0, "step": 15022, "train_loss": 3.5867881774902344, "perplexity": 36.11788519767772, "lr": 0.0026291804804649314, "grad_norm": 0.168243, "tokens_per_sec": 107922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:50.606711+00:00", "epoch": 0, "step": 15023, "train_loss": 3.42258358001709, "perplexity": 30.648495662427273, "lr": 0.0026291804804649314, "grad_norm": 0.192912, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:50.912006+00:00", "epoch": 0, "step": 15024, "train_loss": 3.5261080265045166, "perplexity": 33.9914161479529, "lr": 0.0026291804804649314, "grad_norm": 0.206539, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:51.216526+00:00", "epoch": 0, "step": 15025, "train_loss": 3.491478204727173, "perplexity": 32.8344478868828, "lr": 0.0026291804804649314, "grad_norm": 0.191759, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:51.520746+00:00", "epoch": 0, "step": 15026, "train_loss": 3.5137431621551514, "perplexity": 33.57370469295372, "lr": 0.0026291804804649314, "grad_norm": 0.183731, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:51.824798+00:00", "epoch": 0, "step": 15027, "train_loss": 3.4995169639587402, "perplexity": 33.09945986456352, "lr": 0.0026291804804649314, "grad_norm": 0.181889, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:52.128819+00:00", "epoch": 0, "step": 15028, "train_loss": 3.541121006011963, "perplexity": 34.50557847908299, "lr": 0.0026291804804649314, "grad_norm": 0.166512, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:52.434442+00:00", "epoch": 0, "step": 15029, "train_loss": 3.5111300945281982, "perplexity": 33.48608885501284, "lr": 0.0026291804804649314, "grad_norm": 0.159713, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:52.739109+00:00", "epoch": 0, "step": 15030, "train_loss": 3.5654799938201904, "perplexity": 35.35642019517429, "lr": 0.0026291804804649314, "grad_norm": 0.189513, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:53.043418+00:00", "epoch": 0, "step": 15031, "train_loss": 3.5655789375305176, "perplexity": 35.35991866364515, "lr": 0.0026291804804649314, "grad_norm": 0.153981, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:53.348117+00:00", "epoch": 0, "step": 15032, "train_loss": 3.4900050163269043, "perplexity": 32.78611217167595, "lr": 0.0026291804804649314, "grad_norm": 0.175078, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:53.652237+00:00", "epoch": 0, "step": 15033, "train_loss": 3.4351344108581543, "perplexity": 31.03558380464438, "lr": 0.0026291804804649314, "grad_norm": 0.170626, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:53.955964+00:00", "epoch": 0, "step": 15034, "train_loss": 3.4846010208129883, "perplexity": 32.609414036558114, "lr": 0.0026291804804649314, "grad_norm": 0.174655, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:54.259893+00:00", "epoch": 0, "step": 15035, "train_loss": 3.5577986240386963, "perplexity": 35.08587486711648, "lr": 0.0026291804804649314, "grad_norm": 0.161835, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:54.563392+00:00", "epoch": 0, "step": 15036, "train_loss": 3.4388601779937744, "perplexity": 31.151430838344652, "lr": 0.0026291804804649314, "grad_norm": 0.168674, "tokens_per_sec": 107968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:54.867838+00:00", "epoch": 0, "step": 15037, "train_loss": 3.5323736667633057, "perplexity": 34.205062750852086, "lr": 0.0026291804804649314, "grad_norm": 0.154136, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:55.172176+00:00", "epoch": 0, "step": 15038, "train_loss": 3.41113543510437, "perplexity": 30.299627996731736, "lr": 0.0026291804804649314, "grad_norm": 0.179902, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:55.475969+00:00", "epoch": 0, "step": 15039, "train_loss": 3.473665237426758, "perplexity": 32.25474736184894, "lr": 0.0026291804804649314, "grad_norm": 0.197025, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:55.779220+00:00", "epoch": 0, "step": 15040, "train_loss": 3.5182254314422607, "perplexity": 33.72452884306737, "lr": 0.0026291804804649314, "grad_norm": 0.19717, "tokens_per_sec": 108113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:56.083405+00:00", "epoch": 0, "step": 15041, "train_loss": 3.530160903930664, "perplexity": 34.12945873703316, "lr": 0.0026291804804649314, "grad_norm": 0.174729, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:56.387510+00:00", "epoch": 0, "step": 15042, "train_loss": 3.538168430328369, "perplexity": 34.40384840389483, "lr": 0.0026291804804649314, "grad_norm": 0.160731, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:56.691655+00:00", "epoch": 0, "step": 15043, "train_loss": 3.475348711013794, "perplexity": 32.30909310906912, "lr": 0.0026291804804649314, "grad_norm": 0.184615, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:56.995746+00:00", "epoch": 0, "step": 15044, "train_loss": 3.4824609756469727, "perplexity": 32.53970303663194, "lr": 0.0026291804804649314, "grad_norm": 0.171723, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:57.300706+00:00", "epoch": 0, "step": 15045, "train_loss": 3.5005993843078613, "perplexity": 33.13530679068703, "lr": 0.0026291804804649314, "grad_norm": 0.170874, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:57.605806+00:00", "epoch": 0, "step": 15046, "train_loss": 3.489800453186035, "perplexity": 32.77940602753157, "lr": 0.0026291804804649314, "grad_norm": 0.152932, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:57.911142+00:00", "epoch": 0, "step": 15047, "train_loss": 3.4992733001708984, "perplexity": 33.091395707308855, "lr": 0.0026291804804649314, "grad_norm": 0.178408, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:58.215480+00:00", "epoch": 0, "step": 15048, "train_loss": 3.552314519882202, "perplexity": 34.893986922585874, "lr": 0.0026291804804649314, "grad_norm": 0.182255, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:58.519887+00:00", "epoch": 0, "step": 15049, "train_loss": 3.576110363006592, "perplexity": 35.73427681434652, "lr": 0.0026291804804649314, "grad_norm": 0.139496, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:58.825191+00:00", "epoch": 0, "step": 15050, "train_loss": 3.4552409648895264, "perplexity": 31.665918137488266, "lr": 0.0026291804804649314, "grad_norm": 0.17134, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:59.129836+00:00", "epoch": 0, "step": 15051, "train_loss": 3.4873206615448, "perplexity": 32.698220633493825, "lr": 0.0026291804804649314, "grad_norm": 0.166274, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:59.434706+00:00", "epoch": 0, "step": 15052, "train_loss": 3.471304416656494, "perplexity": 32.178689499154025, "lr": 0.0026291804804649314, "grad_norm": 0.134469, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:16:59.740144+00:00", "epoch": 0, "step": 15053, "train_loss": 3.562152862548828, "perplexity": 35.23898022134065, "lr": 0.0026291804804649314, "grad_norm": 0.163979, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:00.044879+00:00", "epoch": 0, "step": 15054, "train_loss": 3.3969619274139404, "perplexity": 29.873205079145198, "lr": 0.0026291804804649314, "grad_norm": 0.178399, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:00.348794+00:00", "epoch": 0, "step": 15055, "train_loss": 3.571589946746826, "perplexity": 35.573107558911886, "lr": 0.0026291804804649314, "grad_norm": 0.163356, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:00.653065+00:00", "epoch": 0, "step": 15056, "train_loss": 3.511063575744629, "perplexity": 33.48386147519783, "lr": 0.0026291804804649314, "grad_norm": 0.157666, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:00.957695+00:00", "epoch": 0, "step": 15057, "train_loss": 3.634411096572876, "perplexity": 37.879538937105316, "lr": 0.0026291804804649314, "grad_norm": 0.160686, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:01.261988+00:00", "epoch": 0, "step": 15058, "train_loss": 3.4946751594543457, "perplexity": 32.93958610177819, "lr": 0.0026291804804649314, "grad_norm": 0.153137, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:01.567492+00:00", "epoch": 0, "step": 15059, "train_loss": 3.5065371990203857, "perplexity": 33.332643397153745, "lr": 0.0026291804804649314, "grad_norm": 0.150707, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:01.871217+00:00", "epoch": 0, "step": 15060, "train_loss": 3.5578925609588623, "perplexity": 35.0891708809491, "lr": 0.0026291804804649314, "grad_norm": 0.153637, "tokens_per_sec": 107982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:02.174671+00:00", "epoch": 0, "step": 15061, "train_loss": 3.3865270614624023, "perplexity": 29.563102939551552, "lr": 0.0026291804804649314, "grad_norm": 0.159975, "tokens_per_sec": 107888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:02.478917+00:00", "epoch": 0, "step": 15062, "train_loss": 3.6097490787506104, "perplexity": 36.956778410286454, "lr": 0.0026291804804649314, "grad_norm": 0.179326, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:02.783298+00:00", "epoch": 0, "step": 15063, "train_loss": 3.4552524089813232, "perplexity": 31.66628052723587, "lr": 0.0026291804804649314, "grad_norm": 0.167087, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:03.088837+00:00", "epoch": 0, "step": 15064, "train_loss": 3.4758362770080566, "perplexity": 32.32484976506725, "lr": 0.0026291804804649314, "grad_norm": 0.154457, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:03.392711+00:00", "epoch": 0, "step": 15065, "train_loss": 3.5131852626800537, "perplexity": 33.554979164695325, "lr": 0.0026291804804649314, "grad_norm": 0.144538, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:03.696378+00:00", "epoch": 0, "step": 15066, "train_loss": 3.6204192638397217, "perplexity": 37.35322539620257, "lr": 0.0026291804804649314, "grad_norm": 0.186859, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:04.000404+00:00", "epoch": 0, "step": 15067, "train_loss": 3.474583387374878, "perplexity": 32.28437565599283, "lr": 0.0026291804804649314, "grad_norm": 0.162389, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:04.304585+00:00", "epoch": 0, "step": 15068, "train_loss": 3.503427267074585, "perplexity": 33.229142168922195, "lr": 0.0026291804804649314, "grad_norm": 0.159119, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:04.609038+00:00", "epoch": 0, "step": 15069, "train_loss": 3.5662858486175537, "perplexity": 35.38492381935478, "lr": 0.0026291804804649314, "grad_norm": 0.160557, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:04.914340+00:00", "epoch": 0, "step": 15070, "train_loss": 3.495039939880371, "perplexity": 32.95160400984504, "lr": 0.0026291804804649314, "grad_norm": 0.153612, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:05.218371+00:00", "epoch": 0, "step": 15071, "train_loss": 3.5674991607666016, "perplexity": 35.42788283339467, "lr": 0.0026291804804649314, "grad_norm": 0.141647, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:05.522317+00:00", "epoch": 0, "step": 15072, "train_loss": 3.520921230316162, "perplexity": 33.81556604380386, "lr": 0.0026291804804649314, "grad_norm": 0.156553, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:05.826744+00:00", "epoch": 0, "step": 15073, "train_loss": 3.523578643798828, "perplexity": 33.90554749093512, "lr": 0.0026291804804649314, "grad_norm": 0.173258, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:06.133889+00:00", "epoch": 0, "step": 15074, "train_loss": 3.443988561630249, "perplexity": 31.311597673961817, "lr": 0.0026291804804649314, "grad_norm": 0.166031, "tokens_per_sec": 106687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:06.437850+00:00", "epoch": 0, "step": 15075, "train_loss": 3.484478712081909, "perplexity": 32.60542586440518, "lr": 0.0026291804804649314, "grad_norm": 0.190986, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:06.741783+00:00", "epoch": 0, "step": 15076, "train_loss": 3.479903221130371, "perplexity": 32.45658081265923, "lr": 0.0026291804804649314, "grad_norm": 0.151673, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:07.045434+00:00", "epoch": 0, "step": 15077, "train_loss": 3.461453914642334, "perplexity": 31.863269327441586, "lr": 0.0026291804804649314, "grad_norm": 0.164331, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:07.350307+00:00", "epoch": 0, "step": 15078, "train_loss": 3.502692222595215, "perplexity": 33.20472624591074, "lr": 0.0026291804804649314, "grad_norm": 0.154342, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:07.655821+00:00", "epoch": 0, "step": 15079, "train_loss": 3.56669545173645, "perplexity": 35.399420563266176, "lr": 0.0026291804804649314, "grad_norm": 0.173756, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:07.961298+00:00", "epoch": 0, "step": 15080, "train_loss": 3.3710291385650635, "perplexity": 29.108468295606123, "lr": 0.0026291804804649314, "grad_norm": 0.143475, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:08.265488+00:00", "epoch": 0, "step": 15081, "train_loss": 3.486301898956299, "perplexity": 32.66492587221012, "lr": 0.0026291804804649314, "grad_norm": 0.160076, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:08.569776+00:00", "epoch": 0, "step": 15082, "train_loss": 3.488779067993164, "perplexity": 32.745942719956524, "lr": 0.0026291804804649314, "grad_norm": 0.182004, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:08.873726+00:00", "epoch": 0, "step": 15083, "train_loss": 3.580920934677124, "perplexity": 35.90659325203696, "lr": 0.0026291804804649314, "grad_norm": 0.2039, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:09.178722+00:00", "epoch": 0, "step": 15084, "train_loss": 3.517646074295044, "perplexity": 33.70499595505112, "lr": 0.0026291804804649314, "grad_norm": 0.188759, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:09.482938+00:00", "epoch": 0, "step": 15085, "train_loss": 3.5695362091064453, "perplexity": 35.50012469840918, "lr": 0.0026291804804649314, "grad_norm": 0.17975, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:09.787823+00:00", "epoch": 0, "step": 15086, "train_loss": 3.4153966903686523, "perplexity": 30.429017932018372, "lr": 0.0026291804804649314, "grad_norm": 0.177048, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:10.091826+00:00", "epoch": 0, "step": 15087, "train_loss": 3.53658390045166, "perplexity": 34.34937764489442, "lr": 0.0026291804804649314, "grad_norm": 0.158708, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:10.395520+00:00", "epoch": 0, "step": 15088, "train_loss": 3.496872901916504, "perplexity": 33.012058437434135, "lr": 0.0026291804804649314, "grad_norm": 0.177025, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:10.699992+00:00", "epoch": 0, "step": 15089, "train_loss": 3.556471347808838, "perplexity": 35.039337110462654, "lr": 0.0026291804804649314, "grad_norm": 0.166141, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:11.005469+00:00", "epoch": 0, "step": 15090, "train_loss": 3.431565046310425, "perplexity": 30.925003959385933, "lr": 0.0026291804804649314, "grad_norm": 0.159446, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:11.310330+00:00", "epoch": 0, "step": 15091, "train_loss": 3.6068551540374756, "perplexity": 36.849982879572465, "lr": 0.0026291804804649314, "grad_norm": 0.181151, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:11.615031+00:00", "epoch": 0, "step": 15092, "train_loss": 3.5178918838500977, "perplexity": 33.713281983460675, "lr": 0.0026291804804649314, "grad_norm": 0.180985, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:11.919520+00:00", "epoch": 0, "step": 15093, "train_loss": 3.519399881362915, "perplexity": 33.76415988107057, "lr": 0.0026291804804649314, "grad_norm": 0.151896, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:12.225254+00:00", "epoch": 0, "step": 15094, "train_loss": 3.539646625518799, "perplexity": 34.45474161291908, "lr": 0.0026291804804649314, "grad_norm": 0.153696, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:12.531126+00:00", "epoch": 0, "step": 15095, "train_loss": 3.56498384475708, "perplexity": 35.33888249143692, "lr": 0.0026291804804649314, "grad_norm": 0.17451, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:12.836199+00:00", "epoch": 0, "step": 15096, "train_loss": 3.5840721130371094, "perplexity": 36.01991979396986, "lr": 0.0026291804804649314, "grad_norm": 0.175071, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:13.141860+00:00", "epoch": 0, "step": 15097, "train_loss": 3.54740571975708, "perplexity": 34.723119036471765, "lr": 0.0026291804804649314, "grad_norm": 0.165657, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:13.446840+00:00", "epoch": 0, "step": 15098, "train_loss": 3.5143039226531982, "perplexity": 33.59253677997038, "lr": 0.0026291804804649314, "grad_norm": 0.187877, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:13.751486+00:00", "epoch": 0, "step": 15099, "train_loss": 3.504478693008423, "perplexity": 33.26409852454687, "lr": 0.0026291804804649314, "grad_norm": 0.191043, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:14.056291+00:00", "epoch": 0, "step": 15100, "train_loss": 3.4666855335235596, "perplexity": 32.03040261650727, "lr": 0.0026291804804649314, "grad_norm": 0.149477, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:14.360875+00:00", "epoch": 0, "step": 15101, "train_loss": 3.4875080585479736, "perplexity": 32.70434875622764, "lr": 0.0026291804804649314, "grad_norm": 0.164574, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:14.665460+00:00", "epoch": 0, "step": 15102, "train_loss": 3.509568214416504, "perplexity": 33.43382842170145, "lr": 0.0026291804804649314, "grad_norm": 0.17209, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:14.970319+00:00", "epoch": 0, "step": 15103, "train_loss": 3.395355701446533, "perplexity": 29.825260476658084, "lr": 0.0026291804804649314, "grad_norm": 0.172787, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:15.274725+00:00", "epoch": 0, "step": 15104, "train_loss": 3.5192503929138184, "perplexity": 33.7591129064165, "lr": 0.0026291804804649314, "grad_norm": 0.184579, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:15.578958+00:00", "epoch": 0, "step": 15105, "train_loss": 3.5766706466674805, "perplexity": 35.75430375563936, "lr": 0.0026291804804649314, "grad_norm": 0.178836, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:15.884197+00:00", "epoch": 0, "step": 15106, "train_loss": 3.5046350955963135, "perplexity": 33.269301522510524, "lr": 0.0026291804804649314, "grad_norm": 0.14716, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:16.189445+00:00", "epoch": 0, "step": 15107, "train_loss": 3.5119316577911377, "perplexity": 33.51294083400078, "lr": 0.0026291804804649314, "grad_norm": 0.159564, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:16.493459+00:00", "epoch": 0, "step": 15108, "train_loss": 3.5369434356689453, "perplexity": 34.36172967621166, "lr": 0.0026291804804649314, "grad_norm": 0.179071, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:16.798384+00:00", "epoch": 0, "step": 15109, "train_loss": 3.5403709411621094, "perplexity": 34.4797067614879, "lr": 0.0026291804804649314, "grad_norm": 0.187146, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:17.102552+00:00", "epoch": 0, "step": 15110, "train_loss": 3.52848482131958, "perplexity": 34.07230285708701, "lr": 0.0026291804804649314, "grad_norm": 0.16208, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:17.406222+00:00", "epoch": 0, "step": 15111, "train_loss": 3.459475040435791, "perplexity": 31.80027827187671, "lr": 0.0026291804804649314, "grad_norm": 0.178089, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:17.711871+00:00", "epoch": 0, "step": 15112, "train_loss": 3.475800037384033, "perplexity": 32.323678345891175, "lr": 0.0026291804804649314, "grad_norm": 0.200151, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:18.016495+00:00", "epoch": 0, "step": 15113, "train_loss": 3.476149320602417, "perplexity": 32.33497043624786, "lr": 0.0026291804804649314, "grad_norm": 0.16999, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:18.321275+00:00", "epoch": 0, "step": 15114, "train_loss": 3.488461494445801, "perplexity": 32.735545125853, "lr": 0.0026291804804649314, "grad_norm": 0.185802, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:18.625096+00:00", "epoch": 0, "step": 15115, "train_loss": 3.455930471420288, "perplexity": 31.68775952387059, "lr": 0.0026291804804649314, "grad_norm": 0.18187, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:18.929920+00:00", "epoch": 0, "step": 15116, "train_loss": 3.6055305004119873, "perplexity": 36.80120173234668, "lr": 0.0026291804804649314, "grad_norm": 0.174896, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:19.234944+00:00", "epoch": 0, "step": 15117, "train_loss": 3.4984219074249268, "perplexity": 33.063233923120094, "lr": 0.0026291804804649314, "grad_norm": 0.178505, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:19.539351+00:00", "epoch": 0, "step": 15118, "train_loss": 3.5324623584747314, "perplexity": 34.208096590943136, "lr": 0.0026291804804649314, "grad_norm": 0.160261, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:19.843535+00:00", "epoch": 0, "step": 15119, "train_loss": 3.5363903045654297, "perplexity": 34.34272839034272, "lr": 0.0026291804804649314, "grad_norm": 0.174814, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:20.148144+00:00", "epoch": 0, "step": 15120, "train_loss": 3.524456262588501, "perplexity": 33.93531669758414, "lr": 0.0026291804804649314, "grad_norm": 0.201282, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:20.452429+00:00", "epoch": 0, "step": 15121, "train_loss": 3.5248472690582275, "perplexity": 33.94858822042168, "lr": 0.0026291804804649314, "grad_norm": 0.201638, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:20.757613+00:00", "epoch": 0, "step": 15122, "train_loss": 3.4819650650024414, "perplexity": 32.52357025206661, "lr": 0.0026291804804649314, "grad_norm": 0.210545, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:21.061842+00:00", "epoch": 0, "step": 15123, "train_loss": 3.3756370544433594, "perplexity": 29.24290717249901, "lr": 0.0026291804804649314, "grad_norm": 0.144872, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:21.366004+00:00", "epoch": 0, "step": 15124, "train_loss": 3.505948781967163, "perplexity": 33.31303567067593, "lr": 0.0026291804804649314, "grad_norm": 0.169281, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:21.669859+00:00", "epoch": 0, "step": 15125, "train_loss": 3.4750821590423584, "perplexity": 32.300482204283405, "lr": 0.0026291804804649314, "grad_norm": 0.181664, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:21.973744+00:00", "epoch": 0, "step": 15126, "train_loss": 3.56878399848938, "perplexity": 35.47343116854082, "lr": 0.0026291804804649314, "grad_norm": 0.179407, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:22.278796+00:00", "epoch": 0, "step": 15127, "train_loss": 3.55008864402771, "perplexity": 34.81640361706352, "lr": 0.0026291804804649314, "grad_norm": 0.167495, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:22.584980+00:00", "epoch": 0, "step": 15128, "train_loss": 3.445448875427246, "perplexity": 31.357355834556078, "lr": 0.0026291804804649314, "grad_norm": 0.155759, "tokens_per_sec": 107021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:22.889685+00:00", "epoch": 0, "step": 15129, "train_loss": 3.63822603225708, "perplexity": 38.02432293712616, "lr": 0.0026291804804649314, "grad_norm": 0.176467, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:23.193492+00:00", "epoch": 0, "step": 15130, "train_loss": 3.54727840423584, "perplexity": 34.718698525878324, "lr": 0.0026291804804649314, "grad_norm": 0.163009, "tokens_per_sec": 107916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:23.497046+00:00", "epoch": 0, "step": 15131, "train_loss": 3.4995102882385254, "perplexity": 33.09923890256775, "lr": 0.0026291804804649314, "grad_norm": 0.159161, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:23.801979+00:00", "epoch": 0, "step": 15132, "train_loss": 3.490466356277466, "perplexity": 32.80124120458966, "lr": 0.0026291804804649314, "grad_norm": 0.166937, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:24.106805+00:00", "epoch": 0, "step": 15133, "train_loss": 3.5397939682006836, "perplexity": 34.459818640974255, "lr": 0.0026291804804649314, "grad_norm": 0.146381, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:24.412194+00:00", "epoch": 0, "step": 15134, "train_loss": 3.4888412952423096, "perplexity": 32.7479804732939, "lr": 0.0026291804804649314, "grad_norm": 0.17529, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:24.716091+00:00", "epoch": 0, "step": 15135, "train_loss": 3.53544020652771, "perplexity": 34.310114926936215, "lr": 0.0026291804804649314, "grad_norm": 0.152618, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:25.020048+00:00", "epoch": 0, "step": 15136, "train_loss": 3.611520528793335, "perplexity": 37.02230351707192, "lr": 0.0026291804804649314, "grad_norm": 0.14626, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:25.323881+00:00", "epoch": 0, "step": 15137, "train_loss": 3.370410442352295, "perplexity": 29.09046456650594, "lr": 0.0026291804804649314, "grad_norm": 0.156877, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:25.628987+00:00", "epoch": 0, "step": 15138, "train_loss": 3.412105083465576, "perplexity": 30.329022230095944, "lr": 0.0026291804804649314, "grad_norm": 0.173927, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:25.933739+00:00", "epoch": 0, "step": 15139, "train_loss": 3.4540727138519287, "perplexity": 31.628945996343788, "lr": 0.0026291804804649314, "grad_norm": 0.170246, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:26.238538+00:00", "epoch": 0, "step": 15140, "train_loss": 3.4473795890808105, "perplexity": 31.417956391942482, "lr": 0.0026291804804649314, "grad_norm": 0.166846, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:26.542919+00:00", "epoch": 0, "step": 15141, "train_loss": 3.510420799255371, "perplexity": 33.46234575191333, "lr": 0.0026291804804649314, "grad_norm": 0.201237, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:26.846491+00:00", "epoch": 0, "step": 15142, "train_loss": 3.4457545280456543, "perplexity": 31.366941757379845, "lr": 0.0026291804804649314, "grad_norm": 0.172267, "tokens_per_sec": 107941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:27.151464+00:00", "epoch": 0, "step": 15143, "train_loss": 3.5803062915802, "perplexity": 35.884530293477724, "lr": 0.0026291804804649314, "grad_norm": 0.160454, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:27.456995+00:00", "epoch": 0, "step": 15144, "train_loss": 3.3888649940490723, "perplexity": 29.632300339185765, "lr": 0.0026291804804649314, "grad_norm": 0.167954, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:27.762529+00:00", "epoch": 0, "step": 15145, "train_loss": 3.5982065200805664, "perplexity": 36.53265507033952, "lr": 0.0026291804804649314, "grad_norm": 0.192124, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:28.067159+00:00", "epoch": 0, "step": 15146, "train_loss": 3.417428731918335, "perplexity": 30.4909138269891, "lr": 0.0026291804804649314, "grad_norm": 0.192933, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:28.371264+00:00", "epoch": 0, "step": 15147, "train_loss": 3.589111328125, "perplexity": 36.201890025754565, "lr": 0.0026291804804649314, "grad_norm": 0.160547, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:28.676011+00:00", "epoch": 0, "step": 15148, "train_loss": 3.548663854598999, "perplexity": 34.76683289555, "lr": 0.0026291804804649314, "grad_norm": 0.185393, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:28.983235+00:00", "epoch": 0, "step": 15149, "train_loss": 3.5958023071289062, "perplexity": 36.44492828705199, "lr": 0.0026291804804649314, "grad_norm": 0.200604, "tokens_per_sec": 106662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:29.288791+00:00", "epoch": 0, "step": 15150, "train_loss": 3.587550401687622, "perplexity": 36.14542561838802, "lr": 0.0026291804804649314, "grad_norm": 0.164136, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:29.593442+00:00", "epoch": 0, "step": 15151, "train_loss": 3.575523853302002, "perplexity": 35.71332445918885, "lr": 0.0026291804804649314, "grad_norm": 0.204319, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:29.897547+00:00", "epoch": 0, "step": 15152, "train_loss": 3.4280824661254883, "perplexity": 30.81749247072944, "lr": 0.0026291804804649314, "grad_norm": 0.172867, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:30.201675+00:00", "epoch": 0, "step": 15153, "train_loss": 3.5226731300354004, "perplexity": 33.87485944734535, "lr": 0.0026291804804649314, "grad_norm": 0.156843, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:30.505684+00:00", "epoch": 0, "step": 15154, "train_loss": 3.513540506362915, "perplexity": 33.566901476609736, "lr": 0.0026291804804649314, "grad_norm": 0.170335, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:30.810921+00:00", "epoch": 0, "step": 15155, "train_loss": 3.518376588821411, "perplexity": 33.729626939758134, "lr": 0.0026291804804649314, "grad_norm": 0.169261, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:31.116204+00:00", "epoch": 0, "step": 15156, "train_loss": 3.540689468383789, "perplexity": 34.49069123602117, "lr": 0.0026291804804649314, "grad_norm": 0.149298, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:31.420936+00:00", "epoch": 0, "step": 15157, "train_loss": 3.519470453262329, "perplexity": 33.76654276604687, "lr": 0.0026291804804649314, "grad_norm": 0.174111, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:31.725816+00:00", "epoch": 0, "step": 15158, "train_loss": 3.4607319831848145, "perplexity": 31.840274532306385, "lr": 0.0026291804804649314, "grad_norm": 0.178216, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:32.030919+00:00", "epoch": 0, "step": 15159, "train_loss": 3.5423221588134766, "perplexity": 34.54704985308339, "lr": 0.0026291804804649314, "grad_norm": 0.17287, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:32.336656+00:00", "epoch": 0, "step": 15160, "train_loss": 3.6374638080596924, "perplexity": 37.995350921082725, "lr": 0.0026291804804649314, "grad_norm": 0.162571, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:32.640770+00:00", "epoch": 0, "step": 15161, "train_loss": 3.478349447250366, "perplexity": 32.406189783428644, "lr": 0.0026291804804649314, "grad_norm": 0.16447, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:32.944914+00:00", "epoch": 0, "step": 15162, "train_loss": 3.5906763076782227, "perplexity": 36.25858959869686, "lr": 0.0026291804804649314, "grad_norm": 0.171013, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:33.248628+00:00", "epoch": 0, "step": 15163, "train_loss": 3.528947114944458, "perplexity": 34.08805790693171, "lr": 0.0026291804804649314, "grad_norm": 0.157119, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:33.553830+00:00", "epoch": 0, "step": 15164, "train_loss": 3.5129847526550293, "perplexity": 33.54825172946496, "lr": 0.0026291804804649314, "grad_norm": 0.142662, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:33.858429+00:00", "epoch": 0, "step": 15165, "train_loss": 3.468877077102661, "perplexity": 32.100675614740574, "lr": 0.0026291804804649314, "grad_norm": 0.161534, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:34.163980+00:00", "epoch": 0, "step": 15166, "train_loss": 3.4470813274383545, "perplexity": 31.408587017998297, "lr": 0.0026291804804649314, "grad_norm": 0.174894, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:34.468006+00:00", "epoch": 0, "step": 15167, "train_loss": 3.462482213973999, "perplexity": 31.896051157874197, "lr": 0.0026291804804649314, "grad_norm": 0.172166, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:34.772010+00:00", "epoch": 0, "step": 15168, "train_loss": 3.4878153800964355, "perplexity": 32.71440105189326, "lr": 0.0026291804804649314, "grad_norm": 0.171492, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:35.076206+00:00", "epoch": 0, "step": 15169, "train_loss": 3.5115439891815186, "perplexity": 33.499951436776875, "lr": 0.0026291804804649314, "grad_norm": 0.176665, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:35.381507+00:00", "epoch": 0, "step": 15170, "train_loss": 3.4821197986602783, "perplexity": 32.52860313242558, "lr": 0.0026291804804649314, "grad_norm": 0.163638, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:35.687339+00:00", "epoch": 0, "step": 15171, "train_loss": 3.5058846473693848, "perplexity": 33.310899221043314, "lr": 0.0026291804804649314, "grad_norm": 0.157774, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:35.993912+00:00", "epoch": 0, "step": 15172, "train_loss": 3.545297384262085, "perplexity": 34.6499881713989, "lr": 0.0026291804804649314, "grad_norm": 0.165109, "tokens_per_sec": 106891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:36.299144+00:00", "epoch": 0, "step": 15173, "train_loss": 3.584834337234497, "perplexity": 36.04738551461346, "lr": 0.0026291804804649314, "grad_norm": 0.148869, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:36.603787+00:00", "epoch": 0, "step": 15174, "train_loss": 3.473729372024536, "perplexity": 32.25681607343472, "lr": 0.0026291804804649314, "grad_norm": 0.190039, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:36.908200+00:00", "epoch": 0, "step": 15175, "train_loss": 3.545950174331665, "perplexity": 34.67261472397918, "lr": 0.0026291804804649314, "grad_norm": 0.181564, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:37.213401+00:00", "epoch": 0, "step": 15176, "train_loss": 3.509289264678955, "perplexity": 33.42450336470953, "lr": 0.0026291804804649314, "grad_norm": 0.16563, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:37.518985+00:00", "epoch": 0, "step": 15177, "train_loss": 3.5192720890045166, "perplexity": 33.75984535513762, "lr": 0.0026291804804649314, "grad_norm": 0.146564, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:37.823556+00:00", "epoch": 0, "step": 15178, "train_loss": 3.5007128715515137, "perplexity": 33.13906743871104, "lr": 0.0026291804804649314, "grad_norm": 0.163205, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:38.127164+00:00", "epoch": 0, "step": 15179, "train_loss": 3.44930362701416, "perplexity": 31.478463922554802, "lr": 0.0026291804804649314, "grad_norm": 0.167406, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:38.432309+00:00", "epoch": 0, "step": 15180, "train_loss": 3.491105794906616, "perplexity": 32.82222229264414, "lr": 0.0026291804804649314, "grad_norm": 0.178045, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:38.736822+00:00", "epoch": 0, "step": 15181, "train_loss": 3.4382359981536865, "perplexity": 31.131992810269182, "lr": 0.0026291804804649314, "grad_norm": 0.163739, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:39.042775+00:00", "epoch": 0, "step": 15182, "train_loss": 3.595219373703003, "perplexity": 36.42368951114645, "lr": 0.0026291804804649314, "grad_norm": 0.179621, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:39.346823+00:00", "epoch": 0, "step": 15183, "train_loss": 3.4763808250427246, "perplexity": 32.34245699203308, "lr": 0.0026291804804649314, "grad_norm": 0.196639, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:39.650470+00:00", "epoch": 0, "step": 15184, "train_loss": 3.5461301803588867, "perplexity": 34.67885656537671, "lr": 0.0026291804804649314, "grad_norm": 0.170649, "tokens_per_sec": 107915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:39.954662+00:00", "epoch": 0, "step": 15185, "train_loss": 3.553222179412842, "perplexity": 34.92567316036065, "lr": 0.0026291804804649314, "grad_norm": 0.174178, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:40.259811+00:00", "epoch": 0, "step": 15186, "train_loss": 3.4965405464172363, "perplexity": 33.00108852132726, "lr": 0.0026291804804649314, "grad_norm": 0.189019, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:40.564697+00:00", "epoch": 0, "step": 15187, "train_loss": 3.4941351413726807, "perplexity": 32.921802931729516, "lr": 0.0026291804804649314, "grad_norm": 0.176669, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:40.870521+00:00", "epoch": 0, "step": 15188, "train_loss": 3.5477678775787354, "perplexity": 34.73569656301042, "lr": 0.0026291804804649314, "grad_norm": 0.187143, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:41.175725+00:00", "epoch": 0, "step": 15189, "train_loss": 3.5113720893859863, "perplexity": 33.49419329690027, "lr": 0.0026291804804649314, "grad_norm": 0.175731, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:41.479458+00:00", "epoch": 0, "step": 15190, "train_loss": 3.607631206512451, "perplexity": 36.878591499448284, "lr": 0.0026291804804649314, "grad_norm": 0.193367, "tokens_per_sec": 107942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:41.783954+00:00", "epoch": 0, "step": 15191, "train_loss": 3.527085065841675, "perplexity": 34.02464332814298, "lr": 0.0026291804804649314, "grad_norm": 0.167602, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:42.088923+00:00", "epoch": 0, "step": 15192, "train_loss": 3.5345888137817383, "perplexity": 34.280915975625426, "lr": 0.0026291804804649314, "grad_norm": 0.212584, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:42.393491+00:00", "epoch": 0, "step": 15193, "train_loss": 3.485139846801758, "perplexity": 32.62698957097168, "lr": 0.0026291804804649314, "grad_norm": 0.20856, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:42.697428+00:00", "epoch": 0, "step": 15194, "train_loss": 3.4964492321014404, "perplexity": 32.99807518709079, "lr": 0.0026291804804649314, "grad_norm": 0.189196, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:43.002172+00:00", "epoch": 0, "step": 15195, "train_loss": 3.5510077476501465, "perplexity": 34.84841820985858, "lr": 0.0026291804804649314, "grad_norm": 0.159878, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:43.306891+00:00", "epoch": 0, "step": 15196, "train_loss": 3.4777443408966064, "perplexity": 32.386586523716204, "lr": 0.0026291804804649314, "grad_norm": 0.159221, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:43.613344+00:00", "epoch": 0, "step": 15197, "train_loss": 3.5982511043548584, "perplexity": 36.53428388856336, "lr": 0.0026291804804649314, "grad_norm": 0.177608, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:43.919925+00:00", "epoch": 0, "step": 15198, "train_loss": 3.4940924644470215, "perplexity": 32.92039796037337, "lr": 0.0026291804804649314, "grad_norm": 0.17592, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:44.224542+00:00", "epoch": 0, "step": 15199, "train_loss": 3.444741725921631, "perplexity": 31.335189334318827, "lr": 0.0026291804804649314, "grad_norm": 0.192674, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:44.529846+00:00", "epoch": 0, "step": 15200, "train_loss": 3.536242723464966, "perplexity": 34.337660426670965, "lr": 0.0026291804804649314, "grad_norm": 0.16044, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:44.834264+00:00", "epoch": 0, "step": 15201, "train_loss": 3.4711179733276367, "perplexity": 32.17269055641434, "lr": 0.0026291804804649314, "grad_norm": 0.165913, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:45.140094+00:00", "epoch": 0, "step": 15202, "train_loss": 3.6372945308685303, "perplexity": 37.988919719144846, "lr": 0.0026291804804649314, "grad_norm": 0.168071, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:45.446000+00:00", "epoch": 0, "step": 15203, "train_loss": 3.574004650115967, "perplexity": 35.659109854813565, "lr": 0.0026291804804649314, "grad_norm": 0.185015, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:45.751294+00:00", "epoch": 0, "step": 15204, "train_loss": 3.597109079360962, "perplexity": 36.49258463855347, "lr": 0.0026291804804649314, "grad_norm": 0.192166, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:46.055517+00:00", "epoch": 0, "step": 15205, "train_loss": 3.3876900672912598, "perplexity": 29.597505001611076, "lr": 0.0026291804804649314, "grad_norm": 0.175436, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:46.360095+00:00", "epoch": 0, "step": 15206, "train_loss": 3.461132526397705, "perplexity": 31.853030492653, "lr": 0.0026291804804649314, "grad_norm": 0.172843, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:46.664534+00:00", "epoch": 0, "step": 15207, "train_loss": 3.4771041870117188, "perplexity": 32.36586075908195, "lr": 0.0026291804804649314, "grad_norm": 0.170155, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:46.969270+00:00", "epoch": 0, "step": 15208, "train_loss": 3.539250373840332, "perplexity": 34.441091568328964, "lr": 0.0026291804804649314, "grad_norm": 0.176092, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:47.274713+00:00", "epoch": 0, "step": 15209, "train_loss": 3.546353340148926, "perplexity": 34.686596355299464, "lr": 0.0026291804804649314, "grad_norm": 0.182174, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:47.579649+00:00", "epoch": 0, "step": 15210, "train_loss": 3.5417656898498535, "perplexity": 34.52783083994352, "lr": 0.0026291804804649314, "grad_norm": 0.199496, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:47.884611+00:00", "epoch": 0, "step": 15211, "train_loss": 3.5310592651367188, "perplexity": 34.16013309500979, "lr": 0.0026291804804649314, "grad_norm": 0.149427, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:48.189619+00:00", "epoch": 0, "step": 15212, "train_loss": 3.50044846534729, "perplexity": 33.130306421962686, "lr": 0.0026291804804649314, "grad_norm": 0.189274, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:48.494724+00:00", "epoch": 0, "step": 15213, "train_loss": 3.5536882877349854, "perplexity": 34.941956101789614, "lr": 0.0026291804804649314, "grad_norm": 0.187801, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:48.800217+00:00", "epoch": 0, "step": 15214, "train_loss": 3.419680118560791, "perplexity": 30.559637996405225, "lr": 0.0026291804804649314, "grad_norm": 0.146643, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:49.104742+00:00", "epoch": 0, "step": 15215, "train_loss": 3.4664504528045654, "perplexity": 32.022873771408335, "lr": 0.0026291804804649314, "grad_norm": 0.182406, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:49.409630+00:00", "epoch": 0, "step": 15216, "train_loss": 3.435436248779297, "perplexity": 31.04495293464965, "lr": 0.0026291804804649314, "grad_norm": 0.174356, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:49.713060+00:00", "epoch": 0, "step": 15217, "train_loss": 3.509464740753174, "perplexity": 33.43036907997398, "lr": 0.0026291804804649314, "grad_norm": 0.177507, "tokens_per_sec": 107984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:50.017863+00:00", "epoch": 0, "step": 15218, "train_loss": 3.546107053756714, "perplexity": 34.67805457053086, "lr": 0.0026291804804649314, "grad_norm": 0.159423, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:50.323279+00:00", "epoch": 0, "step": 15219, "train_loss": 3.45992374420166, "perplexity": 31.814550378227466, "lr": 0.0026291804804649314, "grad_norm": 0.164783, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:50.629149+00:00", "epoch": 0, "step": 15220, "train_loss": 3.461470127105713, "perplexity": 31.863785913716246, "lr": 0.0026291804804649314, "grad_norm": 0.175959, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:50.934383+00:00", "epoch": 0, "step": 15221, "train_loss": 3.510967969894409, "perplexity": 33.48066037517724, "lr": 0.0026291804804649314, "grad_norm": 0.16987, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:51.238716+00:00", "epoch": 0, "step": 15222, "train_loss": 3.631683111190796, "perplexity": 37.77634492847515, "lr": 0.0026291804804649314, "grad_norm": 0.164641, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:51.543302+00:00", "epoch": 0, "step": 15223, "train_loss": 3.5207712650299072, "perplexity": 33.81049526299228, "lr": 0.0026291804804649314, "grad_norm": 0.20432, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:51.849146+00:00", "epoch": 0, "step": 15224, "train_loss": 3.387392044067383, "perplexity": 29.58868557201448, "lr": 0.0026291804804649314, "grad_norm": 0.176854, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:52.155377+00:00", "epoch": 0, "step": 15225, "train_loss": 3.580388069152832, "perplexity": 35.887464963253606, "lr": 0.0026291804804649314, "grad_norm": 0.15289, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:52.460686+00:00", "epoch": 0, "step": 15226, "train_loss": 3.499262571334839, "perplexity": 33.09104067705386, "lr": 0.0026291804804649314, "grad_norm": 0.204323, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:52.765062+00:00", "epoch": 0, "step": 15227, "train_loss": 3.47153639793396, "perplexity": 32.186155218571294, "lr": 0.0026291804804649314, "grad_norm": 0.159215, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:53.069637+00:00", "epoch": 0, "step": 15228, "train_loss": 3.5465328693389893, "perplexity": 34.692824170869805, "lr": 0.0026291804804649314, "grad_norm": 0.16388, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:53.374908+00:00", "epoch": 0, "step": 15229, "train_loss": 3.586451530456543, "perplexity": 36.105728265175166, "lr": 0.0026291804804649314, "grad_norm": 0.181248, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:53.679895+00:00", "epoch": 0, "step": 15230, "train_loss": 3.4895877838134766, "perplexity": 32.77243559304409, "lr": 0.0026291804804649314, "grad_norm": 0.182934, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:53.984671+00:00", "epoch": 0, "step": 15231, "train_loss": 3.511631965637207, "perplexity": 33.502898773416234, "lr": 0.0026291804804649314, "grad_norm": 0.176054, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:54.289485+00:00", "epoch": 0, "step": 15232, "train_loss": 3.455110549926758, "perplexity": 31.661788697229586, "lr": 0.0026291804804649314, "grad_norm": 0.166677, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:54.593736+00:00", "epoch": 0, "step": 15233, "train_loss": 3.5426642894744873, "perplexity": 34.55887148024084, "lr": 0.0026291804804649314, "grad_norm": 0.176766, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:54.898486+00:00", "epoch": 0, "step": 15234, "train_loss": 3.5271246433258057, "perplexity": 34.025989964572425, "lr": 0.0026291804804649314, "grad_norm": 0.161409, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:55.203353+00:00", "epoch": 0, "step": 15235, "train_loss": 3.4441494941711426, "perplexity": 31.31663713443064, "lr": 0.0026291804804649314, "grad_norm": 0.167102, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:55.507583+00:00", "epoch": 0, "step": 15236, "train_loss": 3.4032509326934814, "perplexity": 30.061669830339454, "lr": 0.0026291804804649314, "grad_norm": 0.156231, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:55.811380+00:00", "epoch": 0, "step": 15237, "train_loss": 3.4841835498809814, "perplexity": 32.59580339530929, "lr": 0.0026291804804649314, "grad_norm": 0.165091, "tokens_per_sec": 107862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:56.116215+00:00", "epoch": 0, "step": 15238, "train_loss": 3.436691999435425, "perplexity": 31.083962142459782, "lr": 0.0026291804804649314, "grad_norm": 0.183156, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:56.422275+00:00", "epoch": 0, "step": 15239, "train_loss": 3.5278661251068115, "perplexity": 34.051228972187644, "lr": 0.0026291804804649314, "grad_norm": 0.159362, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:56.727745+00:00", "epoch": 0, "step": 15240, "train_loss": 3.4410548210144043, "perplexity": 31.21987218330628, "lr": 0.0026291804804649314, "grad_norm": 0.156557, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:57.031504+00:00", "epoch": 0, "step": 15241, "train_loss": 3.461132049560547, "perplexity": 31.853015303948084, "lr": 0.0026291804804649314, "grad_norm": 0.16018, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:57.337282+00:00", "epoch": 0, "step": 15242, "train_loss": 3.457533359527588, "perplexity": 31.738592185409214, "lr": 0.0026291804804649314, "grad_norm": 0.173705, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:57.642880+00:00", "epoch": 0, "step": 15243, "train_loss": 3.5003600120544434, "perplexity": 33.12737606686834, "lr": 0.0026291804804649314, "grad_norm": 0.171712, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:57.948235+00:00", "epoch": 0, "step": 15244, "train_loss": 3.4714269638061523, "perplexity": 32.18263314746838, "lr": 0.0026291804804649314, "grad_norm": 0.177585, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:58.252991+00:00", "epoch": 0, "step": 15245, "train_loss": 3.5610995292663574, "perplexity": 35.20188137278927, "lr": 0.0026291804804649314, "grad_norm": 0.145146, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:58.557258+00:00", "epoch": 0, "step": 15246, "train_loss": 3.5304667949676514, "perplexity": 34.1399002294561, "lr": 0.0026291804804649314, "grad_norm": 0.160409, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:58.862192+00:00", "epoch": 0, "step": 15247, "train_loss": 3.561458110809326, "perplexity": 35.214506381139536, "lr": 0.0026291804804649314, "grad_norm": 0.16004, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:59.166910+00:00", "epoch": 0, "step": 15248, "train_loss": 3.4680769443511963, "perplexity": 32.075001085722604, "lr": 0.0026291804804649314, "grad_norm": 0.165166, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:59.471812+00:00", "epoch": 0, "step": 15249, "train_loss": 3.5329763889312744, "perplexity": 34.225685114582376, "lr": 0.0026291804804649314, "grad_norm": 0.192449, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:17:59.776758+00:00", "epoch": 0, "step": 15250, "train_loss": 3.527697801589966, "perplexity": 34.0454978319305, "lr": 0.0026291804804649314, "grad_norm": 0.189188, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:00.081290+00:00", "epoch": 0, "step": 15251, "train_loss": 3.5611324310302734, "perplexity": 35.20303959583328, "lr": 0.0026291804804649314, "grad_norm": 0.164218, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:00.387345+00:00", "epoch": 0, "step": 15252, "train_loss": 3.4744625091552734, "perplexity": 32.280473413995274, "lr": 0.0026291804804649314, "grad_norm": 0.175626, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:00.693413+00:00", "epoch": 0, "step": 15253, "train_loss": 3.543102979660034, "perplexity": 34.57403544388062, "lr": 0.0026291804804649314, "grad_norm": 0.185723, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:00.998545+00:00", "epoch": 0, "step": 15254, "train_loss": 3.633419990539551, "perplexity": 37.8420148957487, "lr": 0.0026291804804649314, "grad_norm": 0.18051, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:01.303932+00:00", "epoch": 0, "step": 15255, "train_loss": 3.519101142883301, "perplexity": 33.754074733768036, "lr": 0.0026291804804649314, "grad_norm": 0.172399, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:01.608233+00:00", "epoch": 0, "step": 15256, "train_loss": 3.5831186771392822, "perplexity": 35.98559347597368, "lr": 0.0026291804804649314, "grad_norm": 0.168276, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:01.913620+00:00", "epoch": 0, "step": 15257, "train_loss": 3.5985867977142334, "perplexity": 36.546550263809436, "lr": 0.0026291804804649314, "grad_norm": 0.161622, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:02.219739+00:00", "epoch": 0, "step": 15258, "train_loss": 3.555809736251831, "perplexity": 35.016162347271454, "lr": 0.0026291804804649314, "grad_norm": 0.209547, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:02.523450+00:00", "epoch": 0, "step": 15259, "train_loss": 3.496722459793091, "perplexity": 33.007092406824036, "lr": 0.0026291804804649314, "grad_norm": 0.207679, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:02.827784+00:00", "epoch": 0, "step": 15260, "train_loss": 3.5670158863067627, "perplexity": 35.41076557895422, "lr": 0.0026291804804649314, "grad_norm": 0.152721, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:03.132011+00:00", "epoch": 0, "step": 15261, "train_loss": 3.487468957901001, "perplexity": 32.70307002003232, "lr": 0.0026291804804649314, "grad_norm": 0.179636, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:03.437301+00:00", "epoch": 0, "step": 15262, "train_loss": 3.618496894836426, "perplexity": 37.281487688792566, "lr": 0.0026291804804649314, "grad_norm": 0.158113, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:03.742874+00:00", "epoch": 0, "step": 15263, "train_loss": 3.4610486030578613, "perplexity": 31.850357392119246, "lr": 0.0026291804804649314, "grad_norm": 0.167733, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:04.046193+00:00", "epoch": 0, "step": 15264, "train_loss": 3.4524848461151123, "perplexity": 31.578763265760013, "lr": 0.0026291804804649314, "grad_norm": 0.155186, "tokens_per_sec": 108030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:04.350296+00:00", "epoch": 0, "step": 15265, "train_loss": 3.5848324298858643, "perplexity": 36.04731675974755, "lr": 0.0026291804804649314, "grad_norm": 0.148846, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:04.655541+00:00", "epoch": 0, "step": 15266, "train_loss": 3.4991185665130615, "perplexity": 33.08627575073316, "lr": 0.0026291804804649314, "grad_norm": 0.156972, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:04.960420+00:00", "epoch": 0, "step": 15267, "train_loss": 3.418083906173706, "perplexity": 30.510897234343467, "lr": 0.0026291804804649314, "grad_norm": 0.164628, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:05.264954+00:00", "epoch": 0, "step": 15268, "train_loss": 3.466869354248047, "perplexity": 32.036291009509604, "lr": 0.0026291804804649314, "grad_norm": 0.190464, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:05.570505+00:00", "epoch": 0, "step": 15269, "train_loss": 3.4581124782562256, "perplexity": 31.756977921812673, "lr": 0.0026291804804649314, "grad_norm": 0.181405, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:05.874758+00:00", "epoch": 0, "step": 15270, "train_loss": 3.5427098274230957, "perplexity": 34.560445256187265, "lr": 0.0026291804804649314, "grad_norm": 0.168427, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:06.179171+00:00", "epoch": 0, "step": 15271, "train_loss": 3.5081937313079834, "perplexity": 33.38790575647625, "lr": 0.0026291804804649314, "grad_norm": 0.153597, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:06.484126+00:00", "epoch": 0, "step": 15272, "train_loss": 3.4102656841278076, "perplexity": 30.273286322701004, "lr": 0.0026291804804649314, "grad_norm": 0.155412, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:06.789853+00:00", "epoch": 0, "step": 15273, "train_loss": 3.4057490825653076, "perplexity": 30.136862268858543, "lr": 0.0026291804804649314, "grad_norm": 0.150669, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:07.095298+00:00", "epoch": 0, "step": 15274, "train_loss": 3.5115420818328857, "perplexity": 33.49988754075124, "lr": 0.0026291804804649314, "grad_norm": 0.188732, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:07.399698+00:00", "epoch": 0, "step": 15275, "train_loss": 3.4442074298858643, "perplexity": 31.318451538744597, "lr": 0.0026291804804649314, "grad_norm": 0.168681, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:07.703431+00:00", "epoch": 0, "step": 15276, "train_loss": 3.550389051437378, "perplexity": 34.826864293841815, "lr": 0.0026291804804649314, "grad_norm": 0.150281, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:08.007390+00:00", "epoch": 0, "step": 15277, "train_loss": 3.4989540576934814, "perplexity": 33.08083321425, "lr": 0.0026291804804649314, "grad_norm": 0.165682, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:08.312138+00:00", "epoch": 0, "step": 15278, "train_loss": 3.5239973068237305, "perplexity": 33.919745461884, "lr": 0.0026291804804649314, "grad_norm": 0.183458, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:08.616206+00:00", "epoch": 0, "step": 15279, "train_loss": 3.475275993347168, "perplexity": 32.306743752628314, "lr": 0.0026291804804649314, "grad_norm": 0.165089, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:08.920151+00:00", "epoch": 0, "step": 15280, "train_loss": 3.4260129928588867, "perplexity": 30.753782439758204, "lr": 0.0026291804804649314, "grad_norm": 0.156451, "tokens_per_sec": 107874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:09.224519+00:00", "epoch": 0, "step": 15281, "train_loss": 3.65122652053833, "perplexity": 38.52188496871455, "lr": 0.0026291804804649314, "grad_norm": 0.153698, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:09.528563+00:00", "epoch": 0, "step": 15282, "train_loss": 3.6663143634796143, "perplexity": 39.10750387265268, "lr": 0.0026291804804649314, "grad_norm": 0.161332, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:09.832385+00:00", "epoch": 0, "step": 15283, "train_loss": 3.6088931560516357, "perplexity": 36.925159798238326, "lr": 0.0026291804804649314, "grad_norm": 0.162453, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:10.138712+00:00", "epoch": 0, "step": 15284, "train_loss": 3.4033782482147217, "perplexity": 30.065497391152036, "lr": 0.0026291804804649314, "grad_norm": 0.180243, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:10.444122+00:00", "epoch": 0, "step": 15285, "train_loss": 3.589906930923462, "perplexity": 36.23070381141303, "lr": 0.0026291804804649314, "grad_norm": 0.175511, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:10.748515+00:00", "epoch": 0, "step": 15286, "train_loss": 3.568397283554077, "perplexity": 35.4597157150578, "lr": 0.0026291804804649314, "grad_norm": 0.153683, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:11.052118+00:00", "epoch": 0, "step": 15287, "train_loss": 3.4859859943389893, "perplexity": 32.65460850104141, "lr": 0.0026291804804649314, "grad_norm": 0.169696, "tokens_per_sec": 107925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:11.357564+00:00", "epoch": 0, "step": 15288, "train_loss": 3.593832492828369, "perplexity": 36.37320920595606, "lr": 0.0026291804804649314, "grad_norm": 0.176345, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:11.662180+00:00", "epoch": 0, "step": 15289, "train_loss": 3.525739908218384, "perplexity": 33.978905588921336, "lr": 0.0026291804804649314, "grad_norm": 0.201514, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:11.967932+00:00", "epoch": 0, "step": 15290, "train_loss": 3.4951910972595215, "perplexity": 32.956585264413185, "lr": 0.0026291804804649314, "grad_norm": 0.161227, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:12.272415+00:00", "epoch": 0, "step": 15291, "train_loss": 3.4677042961120605, "perplexity": 32.06305061984606, "lr": 0.0026291804804649314, "grad_norm": 0.167915, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:12.577878+00:00", "epoch": 0, "step": 15292, "train_loss": 3.49157977104187, "perplexity": 32.83778293011095, "lr": 0.0026291804804649314, "grad_norm": 0.165235, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:12.882002+00:00", "epoch": 0, "step": 15293, "train_loss": 3.420260429382324, "perplexity": 30.57737723167395, "lr": 0.0026291804804649314, "grad_norm": 0.163367, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:13.187403+00:00", "epoch": 0, "step": 15294, "train_loss": 3.524474620819092, "perplexity": 33.935939695671806, "lr": 0.0026291804804649314, "grad_norm": 0.190397, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:13.492504+00:00", "epoch": 0, "step": 15295, "train_loss": 3.5140392780303955, "perplexity": 33.58364787199798, "lr": 0.0026291804804649314, "grad_norm": 0.152622, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:13.797612+00:00", "epoch": 0, "step": 15296, "train_loss": 3.556759834289551, "perplexity": 35.04944694371716, "lr": 0.0026291804804649314, "grad_norm": 0.185507, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:14.101518+00:00", "epoch": 0, "step": 15297, "train_loss": 3.581698179244995, "perplexity": 35.93451230515354, "lr": 0.0026291804804649314, "grad_norm": 0.191283, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:14.406611+00:00", "epoch": 0, "step": 15298, "train_loss": 3.445317029953003, "perplexity": 31.35322178163913, "lr": 0.0026291804804649314, "grad_norm": 0.169515, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:14.711722+00:00", "epoch": 0, "step": 15299, "train_loss": 3.549584150314331, "perplexity": 34.79884339020072, "lr": 0.0026291804804649314, "grad_norm": 0.18631, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:15.017241+00:00", "epoch": 0, "step": 15300, "train_loss": 3.464810609817505, "perplexity": 31.970404319018435, "lr": 0.0026291804804649314, "grad_norm": 0.172382, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:15.322887+00:00", "epoch": 0, "step": 15301, "train_loss": 3.5583810806274414, "perplexity": 35.106316818801524, "lr": 0.0026291804804649314, "grad_norm": 0.177037, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:15.627302+00:00", "epoch": 0, "step": 15302, "train_loss": 3.3966140747070312, "perplexity": 29.862815411036216, "lr": 0.0026291804804649314, "grad_norm": 0.156868, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:15.931332+00:00", "epoch": 0, "step": 15303, "train_loss": 3.416511297225952, "perplexity": 30.462953232812495, "lr": 0.0026291804804649314, "grad_norm": 0.157128, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:16.235393+00:00", "epoch": 0, "step": 15304, "train_loss": 3.5884249210357666, "perplexity": 36.17704931819144, "lr": 0.0026291804804649314, "grad_norm": 0.15947, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:16.540754+00:00", "epoch": 0, "step": 15305, "train_loss": 3.511187791824341, "perplexity": 33.488020967536464, "lr": 0.0026291804804649314, "grad_norm": 0.141686, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:16.845615+00:00", "epoch": 0, "step": 15306, "train_loss": 3.5851452350616455, "perplexity": 36.05859431074923, "lr": 0.0026291804804649314, "grad_norm": 0.179489, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:17.149346+00:00", "epoch": 0, "step": 15307, "train_loss": 3.5252010822296143, "perplexity": 33.96060180324064, "lr": 0.0026291804804649314, "grad_norm": 0.173464, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:17.454813+00:00", "epoch": 0, "step": 15308, "train_loss": 3.5496346950531006, "perplexity": 34.800602333101644, "lr": 0.0026291804804649314, "grad_norm": 0.178542, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:17.759359+00:00", "epoch": 0, "step": 15309, "train_loss": 3.54805850982666, "perplexity": 34.745793343739734, "lr": 0.0026291804804649314, "grad_norm": 0.205044, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:18.063758+00:00", "epoch": 0, "step": 15310, "train_loss": 3.498763084411621, "perplexity": 33.07451626216814, "lr": 0.0026291804804649314, "grad_norm": 0.174208, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:18.368700+00:00", "epoch": 0, "step": 15311, "train_loss": 3.4624314308166504, "perplexity": 31.894431416817575, "lr": 0.0026291804804649314, "grad_norm": 0.168611, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:18.673884+00:00", "epoch": 0, "step": 15312, "train_loss": 3.496142625808716, "perplexity": 32.98795932046406, "lr": 0.0026291804804649314, "grad_norm": 0.164008, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:18.978139+00:00", "epoch": 0, "step": 15313, "train_loss": 3.5955252647399902, "perplexity": 36.43483289554573, "lr": 0.0026291804804649314, "grad_norm": 0.159506, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:19.282638+00:00", "epoch": 0, "step": 15314, "train_loss": 3.4452567100524902, "perplexity": 31.35133061545856, "lr": 0.0026291804804649314, "grad_norm": 0.172869, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:19.588346+00:00", "epoch": 0, "step": 15315, "train_loss": 3.5304365158081055, "perplexity": 34.13886651762022, "lr": 0.0026291804804649314, "grad_norm": 0.156203, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:19.893497+00:00", "epoch": 0, "step": 15316, "train_loss": 3.4783804416656494, "perplexity": 32.407194209898265, "lr": 0.0026291804804649314, "grad_norm": 0.171336, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:20.198525+00:00", "epoch": 0, "step": 15317, "train_loss": 3.420971632003784, "perplexity": 30.599131677510975, "lr": 0.0026291804804649314, "grad_norm": 0.189761, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:20.503657+00:00", "epoch": 0, "step": 15318, "train_loss": 3.565925121307373, "perplexity": 35.37216181290499, "lr": 0.0026291804804649314, "grad_norm": 0.176377, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:20.808489+00:00", "epoch": 0, "step": 15319, "train_loss": 3.518044948577881, "perplexity": 33.718442692741355, "lr": 0.0026291804804649314, "grad_norm": 0.151815, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:21.113835+00:00", "epoch": 0, "step": 15320, "train_loss": 3.4423868656158447, "perplexity": 31.261486155185885, "lr": 0.0026291804804649314, "grad_norm": 0.194398, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:21.419275+00:00", "epoch": 0, "step": 15321, "train_loss": 3.621518850326538, "perplexity": 37.39432108808026, "lr": 0.0026291804804649314, "grad_norm": 0.242583, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:21.724007+00:00", "epoch": 0, "step": 15322, "train_loss": 3.5071115493774414, "perplexity": 33.35179351170164, "lr": 0.0026291804804649314, "grad_norm": 0.174538, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:22.027965+00:00", "epoch": 0, "step": 15323, "train_loss": 3.4974887371063232, "perplexity": 33.0323946859543, "lr": 0.0026291804804649314, "grad_norm": 0.198791, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:22.332132+00:00", "epoch": 0, "step": 15324, "train_loss": 3.5089457035064697, "perplexity": 33.41302197553651, "lr": 0.0026291804804649314, "grad_norm": 0.16989, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:22.636686+00:00", "epoch": 0, "step": 15325, "train_loss": 3.5371341705322266, "perplexity": 34.36828428109952, "lr": 0.0026291804804649314, "grad_norm": 0.167842, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:22.942249+00:00", "epoch": 0, "step": 15326, "train_loss": 3.5093793869018555, "perplexity": 33.42751579099334, "lr": 0.0026291804804649314, "grad_norm": 0.186315, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:23.246920+00:00", "epoch": 0, "step": 15327, "train_loss": 3.509883165359497, "perplexity": 33.444360095884036, "lr": 0.0026291804804649314, "grad_norm": 0.202587, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:23.552042+00:00", "epoch": 0, "step": 15328, "train_loss": 3.501331329345703, "perplexity": 33.15956889224908, "lr": 0.0026291804804649314, "grad_norm": 0.218109, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:23.855963+00:00", "epoch": 0, "step": 15329, "train_loss": 3.4879000186920166, "perplexity": 32.71717007003482, "lr": 0.0026291804804649314, "grad_norm": 0.163266, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:24.160665+00:00", "epoch": 0, "step": 15330, "train_loss": 3.4488635063171387, "perplexity": 31.464612647412064, "lr": 0.0026291804804649314, "grad_norm": 0.220332, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:24.465077+00:00", "epoch": 0, "step": 15331, "train_loss": 3.507932662963867, "perplexity": 33.3791903689123, "lr": 0.0026291804804649314, "grad_norm": 0.218073, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:24.771676+00:00", "epoch": 0, "step": 15332, "train_loss": 3.5265190601348877, "perplexity": 34.00539063492896, "lr": 0.0026291804804649314, "grad_norm": 0.2081, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:25.077576+00:00", "epoch": 0, "step": 15333, "train_loss": 3.5545074939727783, "perplexity": 34.97059249814466, "lr": 0.0026291804804649314, "grad_norm": 0.184886, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:25.382395+00:00", "epoch": 0, "step": 15334, "train_loss": 3.4832589626312256, "perplexity": 32.56567965925154, "lr": 0.0026291804804649314, "grad_norm": 0.16835, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:25.686703+00:00", "epoch": 0, "step": 15335, "train_loss": 3.5408692359924316, "perplexity": 34.49689210244586, "lr": 0.0026291804804649314, "grad_norm": 0.157641, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:25.990796+00:00", "epoch": 0, "step": 15336, "train_loss": 3.4688804149627686, "perplexity": 32.10078276248395, "lr": 0.0026291804804649314, "grad_norm": 0.160961, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:26.296558+00:00", "epoch": 0, "step": 15337, "train_loss": 3.4751579761505127, "perplexity": 32.302931226273834, "lr": 0.0026291804804649314, "grad_norm": 0.146641, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:26.602286+00:00", "epoch": 0, "step": 15338, "train_loss": 3.486438512802124, "perplexity": 32.669388658189355, "lr": 0.0026291804804649314, "grad_norm": 0.16147, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:26.906676+00:00", "epoch": 0, "step": 15339, "train_loss": 3.5646932125091553, "perplexity": 35.328613364921225, "lr": 0.0026291804804649314, "grad_norm": 0.152495, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:27.211100+00:00", "epoch": 0, "step": 15340, "train_loss": 3.532674789428711, "perplexity": 34.21536422144328, "lr": 0.0026291804804649314, "grad_norm": 0.157248, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:27.515088+00:00", "epoch": 0, "step": 15341, "train_loss": 3.5471701622009277, "perplexity": 34.71494070668105, "lr": 0.0026291804804649314, "grad_norm": 0.147061, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:27.819337+00:00", "epoch": 0, "step": 15342, "train_loss": 3.557199716567993, "perplexity": 35.0648679657654, "lr": 0.0026291804804649314, "grad_norm": 0.166279, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:28.125061+00:00", "epoch": 0, "step": 15343, "train_loss": 3.5202536582946777, "perplexity": 33.79299925134007, "lr": 0.0026291804804649314, "grad_norm": 0.170505, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:28.430833+00:00", "epoch": 0, "step": 15344, "train_loss": 3.5585834980010986, "perplexity": 35.11342366650122, "lr": 0.0026291804804649314, "grad_norm": 0.178438, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:28.735442+00:00", "epoch": 0, "step": 15345, "train_loss": 3.477250337600708, "perplexity": 32.370591394379176, "lr": 0.0026291804804649314, "grad_norm": 0.183487, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:29.040584+00:00", "epoch": 0, "step": 15346, "train_loss": 3.570361375808716, "perplexity": 35.52943030858076, "lr": 0.0026291804804649314, "grad_norm": 0.188968, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:29.344976+00:00", "epoch": 0, "step": 15347, "train_loss": 3.4724197387695312, "perplexity": 32.2145991247983, "lr": 0.0026291804804649314, "grad_norm": 0.208568, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:29.650084+00:00", "epoch": 0, "step": 15348, "train_loss": 3.494264841079712, "perplexity": 32.926073156842186, "lr": 0.0026291804804649314, "grad_norm": 0.16873, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:29.954440+00:00", "epoch": 0, "step": 15349, "train_loss": 3.5162599086761475, "perplexity": 33.65830761483498, "lr": 0.0026291804804649314, "grad_norm": 0.169178, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:30.259263+00:00", "epoch": 0, "step": 15350, "train_loss": 3.5252296924591064, "perplexity": 33.9615734377512, "lr": 0.0026291804804649314, "grad_norm": 0.162083, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:30.563039+00:00", "epoch": 0, "step": 15351, "train_loss": 3.3856937885284424, "perplexity": 29.538479066656343, "lr": 0.0026291804804649314, "grad_norm": 0.170723, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:30.867973+00:00", "epoch": 0, "step": 15352, "train_loss": 3.500375747680664, "perplexity": 33.12789735097715, "lr": 0.0026291804804649314, "grad_norm": 0.196061, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:31.174528+00:00", "epoch": 0, "step": 15353, "train_loss": 3.4686150550842285, "perplexity": 32.092265632771294, "lr": 0.0026291804804649314, "grad_norm": 0.159095, "tokens_per_sec": 106892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:31.480499+00:00", "epoch": 0, "step": 15354, "train_loss": 3.5585150718688965, "perplexity": 35.11102107293238, "lr": 0.0026291804804649314, "grad_norm": 0.20832, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:31.785478+00:00", "epoch": 0, "step": 15355, "train_loss": 3.547572612762451, "perplexity": 34.728914565766814, "lr": 0.0026291804804649314, "grad_norm": 0.164936, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:32.090263+00:00", "epoch": 0, "step": 15356, "train_loss": 3.4691193103790283, "perplexity": 32.108452408429415, "lr": 0.0026291804804649314, "grad_norm": 0.179616, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:32.394012+00:00", "epoch": 0, "step": 15357, "train_loss": 3.5011632442474365, "perplexity": 33.153995731249125, "lr": 0.0026291804804649314, "grad_norm": 0.17418, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:32.698959+00:00", "epoch": 0, "step": 15358, "train_loss": 3.465479850769043, "perplexity": 31.991807383930876, "lr": 0.0026291804804649314, "grad_norm": 0.177423, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:33.004316+00:00", "epoch": 0, "step": 15359, "train_loss": 3.5729942321777344, "perplexity": 35.623097447409975, "lr": 0.0026291804804649314, "grad_norm": 0.160861, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:33.308910+00:00", "epoch": 0, "step": 15360, "train_loss": 3.527963399887085, "perplexity": 34.05454145911206, "lr": 0.0026291804804649314, "grad_norm": 0.161623, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:33.613362+00:00", "epoch": 0, "step": 15361, "train_loss": 3.4538636207580566, "perplexity": 31.622333293528836, "lr": 0.0026291804804649314, "grad_norm": 0.171405, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:33.917241+00:00", "epoch": 0, "step": 15362, "train_loss": 3.6292355060577393, "perplexity": 37.68399641517092, "lr": 0.0026291804804649314, "grad_norm": 0.156151, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:34.221548+00:00", "epoch": 0, "step": 15363, "train_loss": 3.520573616027832, "perplexity": 33.80381331270604, "lr": 0.0026291804804649314, "grad_norm": 0.175911, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:34.526138+00:00", "epoch": 0, "step": 15364, "train_loss": 3.4440360069274902, "perplexity": 31.313083297263198, "lr": 0.0026291804804649314, "grad_norm": 0.147303, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:34.831676+00:00", "epoch": 0, "step": 15365, "train_loss": 3.4242684841156006, "perplexity": 30.700178966860832, "lr": 0.0026291804804649314, "grad_norm": 0.178006, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:35.135970+00:00", "epoch": 0, "step": 15366, "train_loss": 3.3674449920654297, "perplexity": 29.004326022920978, "lr": 0.0026291804804649314, "grad_norm": 0.157483, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:35.440413+00:00", "epoch": 0, "step": 15367, "train_loss": 3.602534055709839, "perplexity": 36.691094014611984, "lr": 0.0026291804804649314, "grad_norm": 0.168521, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:35.744875+00:00", "epoch": 0, "step": 15368, "train_loss": 3.4772608280181885, "perplexity": 32.370930977178176, "lr": 0.0026291804804649314, "grad_norm": 0.165087, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:36.050682+00:00", "epoch": 0, "step": 15369, "train_loss": 3.503387689590454, "perplexity": 33.22782706909967, "lr": 0.0026291804804649314, "grad_norm": 0.149161, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:36.356192+00:00", "epoch": 0, "step": 15370, "train_loss": 3.5440053939819336, "perplexity": 34.60524963057732, "lr": 0.0026291804804649314, "grad_norm": 0.16643, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:36.659916+00:00", "epoch": 0, "step": 15371, "train_loss": 3.547438383102417, "perplexity": 34.724253228223176, "lr": 0.0026291804804649314, "grad_norm": 0.153241, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:36.963868+00:00", "epoch": 0, "step": 15372, "train_loss": 3.5715725421905518, "perplexity": 35.57248843014737, "lr": 0.0026291804804649314, "grad_norm": 0.149222, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:37.268030+00:00", "epoch": 0, "step": 15373, "train_loss": 3.5444259643554688, "perplexity": 34.61980663424839, "lr": 0.0026291804804649314, "grad_norm": 0.162106, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:37.572675+00:00", "epoch": 0, "step": 15374, "train_loss": 3.636369228363037, "perplexity": 37.95378473429933, "lr": 0.0026291804804649314, "grad_norm": 0.187026, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:37.878359+00:00", "epoch": 0, "step": 15375, "train_loss": 3.5142722129821777, "perplexity": 33.591471588568865, "lr": 0.0026291804804649314, "grad_norm": 0.156302, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:38.183785+00:00", "epoch": 0, "step": 15376, "train_loss": 3.5533013343811035, "perplexity": 34.92843781032762, "lr": 0.0026291804804649314, "grad_norm": 0.163053, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:38.488195+00:00", "epoch": 0, "step": 15377, "train_loss": 3.4417245388031006, "perplexity": 31.240787690029965, "lr": 0.0026291804804649314, "grad_norm": 0.155689, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:38.792343+00:00", "epoch": 0, "step": 15378, "train_loss": 3.4154179096221924, "perplexity": 30.429663619915328, "lr": 0.0026291804804649314, "grad_norm": 0.155229, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:39.097444+00:00", "epoch": 0, "step": 15379, "train_loss": 3.53458309173584, "perplexity": 34.28071981921198, "lr": 0.0026291804804649314, "grad_norm": 0.1486, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:39.403544+00:00", "epoch": 0, "step": 15380, "train_loss": 3.5261335372924805, "perplexity": 33.992283306823744, "lr": 0.0026291804804649314, "grad_norm": 0.177135, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:39.708152+00:00", "epoch": 0, "step": 15381, "train_loss": 3.6267054080963135, "perplexity": 37.58877272609036, "lr": 0.0026291804804649314, "grad_norm": 0.155694, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:40.012012+00:00", "epoch": 0, "step": 15382, "train_loss": 3.403395414352417, "perplexity": 30.066013504049952, "lr": 0.0026291804804649314, "grad_norm": 0.156453, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:40.316138+00:00", "epoch": 0, "step": 15383, "train_loss": 3.473750352859497, "perplexity": 32.257492855468826, "lr": 0.0026291804804649314, "grad_norm": 0.172829, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:40.620620+00:00", "epoch": 0, "step": 15384, "train_loss": 3.5845718383789062, "perplexity": 36.037924358994424, "lr": 0.0026291804804649314, "grad_norm": 0.16374, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:40.926017+00:00", "epoch": 0, "step": 15385, "train_loss": 3.4547929763793945, "perplexity": 31.651735347100665, "lr": 0.0026291804804649314, "grad_norm": 0.170866, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:41.231563+00:00", "epoch": 0, "step": 15386, "train_loss": 3.5033042430877686, "perplexity": 33.22505443882367, "lr": 0.0026291804804649314, "grad_norm": 0.162433, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:41.535803+00:00", "epoch": 0, "step": 15387, "train_loss": 3.4551541805267334, "perplexity": 31.66317015020334, "lr": 0.0026291804804649314, "grad_norm": 0.176025, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:41.839594+00:00", "epoch": 0, "step": 15388, "train_loss": 3.4512360095977783, "perplexity": 31.539351167777316, "lr": 0.0026291804804649314, "grad_norm": 0.192103, "tokens_per_sec": 107864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:42.143205+00:00", "epoch": 0, "step": 15389, "train_loss": 3.547121286392212, "perplexity": 34.713244027343116, "lr": 0.0026291804804649314, "grad_norm": 0.220716, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:42.448614+00:00", "epoch": 0, "step": 15390, "train_loss": 3.4258248805999756, "perplexity": 30.74799782036937, "lr": 0.0026291804804649314, "grad_norm": 0.17911, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:42.752887+00:00", "epoch": 0, "step": 15391, "train_loss": 3.4989359378814697, "perplexity": 33.08023380120161, "lr": 0.0026291804804649314, "grad_norm": 0.17837, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:43.057730+00:00", "epoch": 0, "step": 15392, "train_loss": 3.471357822418213, "perplexity": 32.18040807246829, "lr": 0.0026291804804649314, "grad_norm": 0.189049, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:43.362231+00:00", "epoch": 0, "step": 15393, "train_loss": 3.442110538482666, "perplexity": 31.25284895173951, "lr": 0.0026291804804649314, "grad_norm": 0.17861, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:43.666591+00:00", "epoch": 0, "step": 15394, "train_loss": 3.440030336380005, "perplexity": 31.187904282074413, "lr": 0.0026291804804649314, "grad_norm": 0.18193, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:43.970764+00:00", "epoch": 0, "step": 15395, "train_loss": 3.558178424835205, "perplexity": 35.09920304120257, "lr": 0.0026291804804649314, "grad_norm": 0.170531, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:44.275446+00:00", "epoch": 0, "step": 15396, "train_loss": 3.4119153022766113, "perplexity": 30.3232668983411, "lr": 0.0026291804804649314, "grad_norm": 0.169348, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:44.580639+00:00", "epoch": 0, "step": 15397, "train_loss": 3.577782392501831, "perplexity": 35.7940755578711, "lr": 0.0026291804804649314, "grad_norm": 0.175928, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:44.885181+00:00", "epoch": 0, "step": 15398, "train_loss": 3.472787857055664, "perplexity": 32.22646009080187, "lr": 0.0026291804804649314, "grad_norm": 0.1648, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:45.189052+00:00", "epoch": 0, "step": 15399, "train_loss": 3.430178165435791, "perplexity": 30.882144390274846, "lr": 0.0026291804804649314, "grad_norm": 0.168114, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:45.493551+00:00", "epoch": 0, "step": 15400, "train_loss": 3.528733968734741, "perplexity": 34.080792940869735, "lr": 0.0026291804804649314, "grad_norm": 0.187212, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:45.798946+00:00", "epoch": 0, "step": 15401, "train_loss": 3.5249898433685303, "perplexity": 33.9534287620332, "lr": 0.0026291804804649314, "grad_norm": 0.17436, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:46.103763+00:00", "epoch": 0, "step": 15402, "train_loss": 3.5476558208465576, "perplexity": 34.731804412438564, "lr": 0.0026291804804649314, "grad_norm": 0.184938, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:46.407648+00:00", "epoch": 0, "step": 15403, "train_loss": 3.5164947509765625, "perplexity": 33.66621293743907, "lr": 0.0026291804804649314, "grad_norm": 0.187675, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:46.712205+00:00", "epoch": 0, "step": 15404, "train_loss": 3.563504457473755, "perplexity": 35.28664125011407, "lr": 0.0026291804804649314, "grad_norm": 0.19028, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:47.016336+00:00", "epoch": 0, "step": 15405, "train_loss": 3.54282546043396, "perplexity": 34.56444181559169, "lr": 0.0026291804804649314, "grad_norm": 0.192339, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:47.321491+00:00", "epoch": 0, "step": 15406, "train_loss": 3.5540432929992676, "perplexity": 34.95436288225596, "lr": 0.0026291804804649314, "grad_norm": 0.183198, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:47.627335+00:00", "epoch": 0, "step": 15407, "train_loss": 3.37878155708313, "perplexity": 29.335006298380847, "lr": 0.0026291804804649314, "grad_norm": 0.192352, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:47.932367+00:00", "epoch": 0, "step": 15408, "train_loss": 3.4793689250946045, "perplexity": 32.43924402209815, "lr": 0.0026291804804649314, "grad_norm": 0.162072, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:48.236612+00:00", "epoch": 0, "step": 15409, "train_loss": 3.5127789974212646, "perplexity": 33.541349711175506, "lr": 0.0026291804804649314, "grad_norm": 0.210074, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:48.541189+00:00", "epoch": 0, "step": 15410, "train_loss": 3.5453600883483887, "perplexity": 34.652160935367505, "lr": 0.0026291804804649314, "grad_norm": 0.185243, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:48.846480+00:00", "epoch": 0, "step": 15411, "train_loss": 3.5649638175964355, "perplexity": 35.3381747610472, "lr": 0.0026291804804649314, "grad_norm": 0.169306, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:49.151165+00:00", "epoch": 0, "step": 15412, "train_loss": 3.4492897987365723, "perplexity": 31.4780286326273, "lr": 0.0026291804804649314, "grad_norm": 0.16865, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:49.456974+00:00", "epoch": 0, "step": 15413, "train_loss": 3.595062255859375, "perplexity": 36.41796714914784, "lr": 0.0026291804804649314, "grad_norm": 0.173879, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:49.761043+00:00", "epoch": 0, "step": 15414, "train_loss": 3.5845158100128174, "perplexity": 36.03590526953902, "lr": 0.0026291804804649314, "grad_norm": 0.159573, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:50.065466+00:00", "epoch": 0, "step": 15415, "train_loss": 3.5492942333221436, "perplexity": 34.788756076505905, "lr": 0.0026291804804649314, "grad_norm": 0.151176, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:50.371117+00:00", "epoch": 0, "step": 15416, "train_loss": 3.496365785598755, "perplexity": 32.99532172800593, "lr": 0.0026291804804649314, "grad_norm": 0.154149, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:50.676189+00:00", "epoch": 0, "step": 15417, "train_loss": 3.3786516189575195, "perplexity": 29.33119481028184, "lr": 0.0026291804804649314, "grad_norm": 0.161141, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:50.981597+00:00", "epoch": 0, "step": 15418, "train_loss": 3.409842014312744, "perplexity": 30.260463161668245, "lr": 0.0026291804804649314, "grad_norm": 0.161167, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:51.287542+00:00", "epoch": 0, "step": 15419, "train_loss": 3.5641305446624756, "perplexity": 35.3087406814973, "lr": 0.0026291804804649314, "grad_norm": 0.165034, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:51.592407+00:00", "epoch": 0, "step": 15420, "train_loss": 3.448723554611206, "perplexity": 31.46020942932168, "lr": 0.0026291804804649314, "grad_norm": 0.154927, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:51.896557+00:00", "epoch": 0, "step": 15421, "train_loss": 3.6221959590911865, "perplexity": 37.41964968477678, "lr": 0.0026291804804649314, "grad_norm": 0.171809, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:52.201280+00:00", "epoch": 0, "step": 15422, "train_loss": 3.368229866027832, "perplexity": 29.027099699277098, "lr": 0.0026291804804649314, "grad_norm": 0.164482, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:52.506101+00:00", "epoch": 0, "step": 15423, "train_loss": 3.572324275970459, "perplexity": 35.5992395249267, "lr": 0.0026291804804649314, "grad_norm": 0.190137, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:52.811167+00:00", "epoch": 0, "step": 15424, "train_loss": 3.561232805252075, "perplexity": 35.20657325087875, "lr": 0.0026291804804649314, "grad_norm": 0.17311, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:53.116600+00:00", "epoch": 0, "step": 15425, "train_loss": 3.5024116039276123, "perplexity": 33.1954096871348, "lr": 0.0026291804804649314, "grad_norm": 0.152909, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:53.420657+00:00", "epoch": 0, "step": 15426, "train_loss": 3.4737727642059326, "perplexity": 32.25821579741737, "lr": 0.0026291804804649314, "grad_norm": 0.144584, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:53.724866+00:00", "epoch": 0, "step": 15427, "train_loss": 3.5169994831085205, "perplexity": 33.68320964590155, "lr": 0.0026291804804649314, "grad_norm": 0.169686, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:54.030911+00:00", "epoch": 0, "step": 15428, "train_loss": 3.4981849193573, "perplexity": 33.05539925960071, "lr": 0.0026291804804649314, "grad_norm": 0.168907, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:54.335347+00:00", "epoch": 0, "step": 15429, "train_loss": 3.5055606365203857, "perplexity": 33.30010787665652, "lr": 0.0026291804804649314, "grad_norm": 0.150779, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:54.639656+00:00", "epoch": 0, "step": 15430, "train_loss": 3.475822687149048, "perplexity": 32.324410477901395, "lr": 0.0026291804804649314, "grad_norm": 0.152553, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:54.943825+00:00", "epoch": 0, "step": 15431, "train_loss": 3.639171600341797, "perplexity": 38.06029452745217, "lr": 0.0026291804804649314, "grad_norm": 0.167577, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:55.248412+00:00", "epoch": 0, "step": 15432, "train_loss": 3.4840943813323975, "perplexity": 32.59289700441185, "lr": 0.0026291804804649314, "grad_norm": 0.143096, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:55.554784+00:00", "epoch": 0, "step": 15433, "train_loss": 3.471139669418335, "perplexity": 32.17338858559888, "lr": 0.0026291804804649314, "grad_norm": 0.164707, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:55.861303+00:00", "epoch": 0, "step": 15434, "train_loss": 3.5279910564422607, "perplexity": 34.05548330344094, "lr": 0.0026291804804649314, "grad_norm": 0.17903, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:56.166195+00:00", "epoch": 0, "step": 15435, "train_loss": 3.525331497192383, "perplexity": 33.96503106267498, "lr": 0.0026291804804649314, "grad_norm": 0.167876, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:56.470998+00:00", "epoch": 0, "step": 15436, "train_loss": 3.5226995944976807, "perplexity": 33.87575593914798, "lr": 0.0026291804804649314, "grad_norm": 0.176836, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:56.776604+00:00", "epoch": 0, "step": 15437, "train_loss": 3.527052402496338, "perplexity": 34.023531987618135, "lr": 0.0026291804804649314, "grad_norm": 0.157857, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:57.082552+00:00", "epoch": 0, "step": 15438, "train_loss": 3.4537558555603027, "perplexity": 31.618925690141843, "lr": 0.0026291804804649314, "grad_norm": 0.176398, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:57.388036+00:00", "epoch": 0, "step": 15439, "train_loss": 3.526258945465088, "perplexity": 33.99654648426907, "lr": 0.0026291804804649314, "grad_norm": 0.178418, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:57.693748+00:00", "epoch": 0, "step": 15440, "train_loss": 3.4331905841827393, "perplexity": 30.9753146043745, "lr": 0.0026291804804649314, "grad_norm": 0.163832, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:57.998152+00:00", "epoch": 0, "step": 15441, "train_loss": 3.434450149536133, "perplexity": 31.01435461902413, "lr": 0.0026291804804649314, "grad_norm": 0.173117, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:58.303112+00:00", "epoch": 0, "step": 15442, "train_loss": 3.495709180831909, "perplexity": 32.9736639535528, "lr": 0.0026291804804649314, "grad_norm": 0.198154, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:58.608835+00:00", "epoch": 0, "step": 15443, "train_loss": 3.541114568710327, "perplexity": 34.50535635698114, "lr": 0.0026291804804649314, "grad_norm": 0.175614, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:58.914294+00:00", "epoch": 0, "step": 15444, "train_loss": 3.520550012588501, "perplexity": 33.80301543586571, "lr": 0.0026291804804649314, "grad_norm": 0.219872, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:59.219931+00:00", "epoch": 0, "step": 15445, "train_loss": 3.5192110538482666, "perplexity": 33.757784880582726, "lr": 0.0026291804804649314, "grad_norm": 0.162156, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:59.525559+00:00", "epoch": 0, "step": 15446, "train_loss": 3.5350537300109863, "perplexity": 34.29685743525049, "lr": 0.0026291804804649314, "grad_norm": 0.208144, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:18:59.830907+00:00", "epoch": 0, "step": 15447, "train_loss": 3.4716641902923584, "perplexity": 32.190268626079884, "lr": 0.0026291804804649314, "grad_norm": 0.257475, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:00.135389+00:00", "epoch": 0, "step": 15448, "train_loss": 3.506287097930908, "perplexity": 33.32430790912572, "lr": 0.0026291804804649314, "grad_norm": 0.238463, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:00.441772+00:00", "epoch": 0, "step": 15449, "train_loss": 3.4114983081817627, "perplexity": 30.31062491110817, "lr": 0.0026291804804649314, "grad_norm": 0.174198, "tokens_per_sec": 106951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:00.747982+00:00", "epoch": 0, "step": 15450, "train_loss": 3.577115774154663, "perplexity": 35.7702225217034, "lr": 0.0026291804804649314, "grad_norm": 0.182893, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:01.052868+00:00", "epoch": 0, "step": 15451, "train_loss": 3.4792232513427734, "perplexity": 32.43451881989221, "lr": 0.0026291804804649314, "grad_norm": 0.181318, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:01.358747+00:00", "epoch": 0, "step": 15452, "train_loss": 3.548595905303955, "perplexity": 34.76447059402311, "lr": 0.0026291804804649314, "grad_norm": 0.179516, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:01.663071+00:00", "epoch": 0, "step": 15453, "train_loss": 3.528373956680298, "perplexity": 34.0685256529043, "lr": 0.0026291804804649314, "grad_norm": 0.178868, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:01.968587+00:00", "epoch": 0, "step": 15454, "train_loss": 3.513059616088867, "perplexity": 33.550763360802186, "lr": 0.0026291804804649314, "grad_norm": 0.154023, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:02.273834+00:00", "epoch": 0, "step": 15455, "train_loss": 3.5211586952209473, "perplexity": 33.82359700747298, "lr": 0.0026291804804649314, "grad_norm": 0.177135, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:02.579217+00:00", "epoch": 0, "step": 15456, "train_loss": 3.5081722736358643, "perplexity": 33.387189337428154, "lr": 0.0026291804804649314, "grad_norm": 0.151972, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:02.884935+00:00", "epoch": 0, "step": 15457, "train_loss": 3.5239272117614746, "perplexity": 33.91736793854144, "lr": 0.0026291804804649314, "grad_norm": 0.188791, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:03.191228+00:00", "epoch": 0, "step": 15458, "train_loss": 3.5810930728912354, "perplexity": 35.91277468088899, "lr": 0.0026291804804649314, "grad_norm": 0.18298, "tokens_per_sec": 106988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:03.497112+00:00", "epoch": 0, "step": 15459, "train_loss": 3.513941526412964, "perplexity": 33.58036517654625, "lr": 0.0026291804804649314, "grad_norm": 0.182056, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:03.802362+00:00", "epoch": 0, "step": 15460, "train_loss": 3.506535291671753, "perplexity": 33.332579820242564, "lr": 0.0026291804804649314, "grad_norm": 0.147117, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:04.107023+00:00", "epoch": 0, "step": 15461, "train_loss": 3.586047410964966, "perplexity": 36.09114018448803, "lr": 0.0026291804804649314, "grad_norm": 0.162537, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:04.412666+00:00", "epoch": 0, "step": 15462, "train_loss": 3.509070873260498, "perplexity": 33.41720453703817, "lr": 0.0026291804804649314, "grad_norm": 0.144857, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:04.718097+00:00", "epoch": 0, "step": 15463, "train_loss": 3.544100761413574, "perplexity": 34.60855000172732, "lr": 0.0026291804804649314, "grad_norm": 0.174278, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:05.023691+00:00", "epoch": 0, "step": 15464, "train_loss": 3.6162452697753906, "perplexity": 37.19763819104017, "lr": 0.0026291804804649314, "grad_norm": 0.156963, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:05.328386+00:00", "epoch": 0, "step": 15465, "train_loss": 3.3702118396759033, "perplexity": 29.08468769605551, "lr": 0.0026291804804649314, "grad_norm": 0.147248, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:05.632563+00:00", "epoch": 0, "step": 15466, "train_loss": 3.5497281551361084, "perplexity": 34.80385495227708, "lr": 0.0026291804804649314, "grad_norm": 0.161243, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:05.937283+00:00", "epoch": 0, "step": 15467, "train_loss": 3.4484472274780273, "perplexity": 31.45151732082889, "lr": 0.0026291804804649314, "grad_norm": 0.167016, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:06.242343+00:00", "epoch": 0, "step": 15468, "train_loss": 3.423224449157715, "perplexity": 30.668143632720348, "lr": 0.0026291804804649314, "grad_norm": 0.184137, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:06.549274+00:00", "epoch": 0, "step": 15469, "train_loss": 3.441253662109375, "perplexity": 31.226080594099198, "lr": 0.0026291804804649314, "grad_norm": 0.186581, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:06.854509+00:00", "epoch": 0, "step": 15470, "train_loss": 3.459878444671631, "perplexity": 31.813109226689228, "lr": 0.0026291804804649314, "grad_norm": 0.177884, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:07.159213+00:00", "epoch": 0, "step": 15471, "train_loss": 3.4119820594787598, "perplexity": 30.32529126236892, "lr": 0.0026291804804649314, "grad_norm": 0.19, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:07.464323+00:00", "epoch": 0, "step": 15472, "train_loss": 3.462515354156494, "perplexity": 31.8971082163459, "lr": 0.0026291804804649314, "grad_norm": 0.17363, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:07.769536+00:00", "epoch": 0, "step": 15473, "train_loss": 3.447183132171631, "perplexity": 31.411784723590273, "lr": 0.0026291804804649314, "grad_norm": 0.179071, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:08.074811+00:00", "epoch": 0, "step": 15474, "train_loss": 3.4800527095794678, "perplexity": 32.461433059256265, "lr": 0.0026291804804649314, "grad_norm": 0.200228, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:08.379664+00:00", "epoch": 0, "step": 15475, "train_loss": 3.4913458824157715, "perplexity": 32.830103444284504, "lr": 0.0026291804804649314, "grad_norm": 0.170366, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:08.683316+00:00", "epoch": 0, "step": 15476, "train_loss": 3.484154462814331, "perplexity": 32.59485529279223, "lr": 0.0026291804804649314, "grad_norm": 0.17774, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:08.988795+00:00", "epoch": 0, "step": 15477, "train_loss": 3.4813835620880127, "perplexity": 32.50466319896343, "lr": 0.0026291804804649314, "grad_norm": 0.175229, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:09.293352+00:00", "epoch": 0, "step": 15478, "train_loss": 3.49649977684021, "perplexity": 32.999743108332986, "lr": 0.0026291804804649314, "grad_norm": 0.171382, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:09.598958+00:00", "epoch": 0, "step": 15479, "train_loss": 3.523730516433716, "perplexity": 33.91069720681044, "lr": 0.0026291804804649314, "grad_norm": 0.188735, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:09.903224+00:00", "epoch": 0, "step": 15480, "train_loss": 3.5469632148742676, "perplexity": 34.70775728582714, "lr": 0.0026291804804649314, "grad_norm": 0.166403, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:10.207126+00:00", "epoch": 0, "step": 15481, "train_loss": 3.624718427658081, "perplexity": 37.51415872281749, "lr": 0.0026291804804649314, "grad_norm": 0.155978, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:10.511706+00:00", "epoch": 0, "step": 15482, "train_loss": 3.426766872406006, "perplexity": 30.77697582874842, "lr": 0.0026291804804649314, "grad_norm": 0.170964, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:10.816148+00:00", "epoch": 0, "step": 15483, "train_loss": 3.4770777225494385, "perplexity": 32.36500422531462, "lr": 0.0026291804804649314, "grad_norm": 0.192863, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:11.120957+00:00", "epoch": 0, "step": 15484, "train_loss": 3.5384469032287598, "perplexity": 34.413430277428596, "lr": 0.0026291804804649314, "grad_norm": 0.205113, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:11.425325+00:00", "epoch": 0, "step": 15485, "train_loss": 3.519758701324463, "perplexity": 33.776277309478374, "lr": 0.0026291804804649314, "grad_norm": 0.177321, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:11.730051+00:00", "epoch": 0, "step": 15486, "train_loss": 3.5302541255950928, "perplexity": 34.132640490284544, "lr": 0.0026291804804649314, "grad_norm": 0.156484, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:12.034690+00:00", "epoch": 0, "step": 15487, "train_loss": 3.438370943069458, "perplexity": 31.136194197888337, "lr": 0.0026291804804649314, "grad_norm": 0.178207, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:12.340396+00:00", "epoch": 0, "step": 15488, "train_loss": 3.5926575660705566, "perplexity": 36.3304984451137, "lr": 0.0026291804804649314, "grad_norm": 0.170813, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:12.646130+00:00", "epoch": 0, "step": 15489, "train_loss": 3.4752893447875977, "perplexity": 32.30717509707254, "lr": 0.0026291804804649314, "grad_norm": 0.153833, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:12.950659+00:00", "epoch": 0, "step": 15490, "train_loss": 3.6032297611236572, "perplexity": 36.71662908877161, "lr": 0.0026291804804649314, "grad_norm": 0.202279, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:13.254478+00:00", "epoch": 0, "step": 15491, "train_loss": 3.594019651412964, "perplexity": 36.380017401394454, "lr": 0.0026291804804649314, "grad_norm": 0.160131, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:13.558709+00:00", "epoch": 0, "step": 15492, "train_loss": 3.5105721950531006, "perplexity": 33.46741219395168, "lr": 0.0026291804804649314, "grad_norm": 0.184238, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:13.864046+00:00", "epoch": 0, "step": 15493, "train_loss": 3.509836435317993, "perplexity": 33.44279727606429, "lr": 0.0026291804804649314, "grad_norm": 0.173235, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:14.169849+00:00", "epoch": 0, "step": 15494, "train_loss": 3.438770055770874, "perplexity": 31.148623528653367, "lr": 0.0026291804804649314, "grad_norm": 0.168397, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:14.472947+00:00", "epoch": 0, "step": 15495, "train_loss": 3.5634965896606445, "perplexity": 35.28636362250759, "lr": 0.0026291804804649314, "grad_norm": 0.160521, "tokens_per_sec": 108116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:14.777454+00:00", "epoch": 0, "step": 15496, "train_loss": 3.5214810371398926, "perplexity": 33.834501528035624, "lr": 0.0026291804804649314, "grad_norm": 0.182968, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:15.082868+00:00", "epoch": 0, "step": 15497, "train_loss": 3.4046616554260254, "perplexity": 30.10410843886401, "lr": 0.0026291804804649314, "grad_norm": 0.189806, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:15.387911+00:00", "epoch": 0, "step": 15498, "train_loss": 3.5005133152008057, "perplexity": 33.1324549871474, "lr": 0.0026291804804649314, "grad_norm": 0.172861, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:15.693463+00:00", "epoch": 0, "step": 15499, "train_loss": 3.5824668407440186, "perplexity": 35.96214439975181, "lr": 0.0026291804804649314, "grad_norm": 0.17782, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:15.998708+00:00", "epoch": 0, "step": 15500, "train_loss": 3.6360626220703125, "perplexity": 37.94214964885344, "lr": 0.0026291804804649314, "grad_norm": 0.203491, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:19:19.072480+00:00", "step": 15500, "epoch": 0, "val_loss": 3.4744338154792787, "val_ppl": 32.279547181838744, "eval_train_loss": 3.6360626220703125, "eval_train_ppl": 37.94214964885344} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:19:20.049998+00:00", "step": 15500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4744_epoch_0000_step_0015500.pt", "val_loss": 3.4744338154792787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:21.321730+00:00", "epoch": 0, "step": 15501, "train_loss": 3.555478096008301, "perplexity": 35.00455150408096, "lr": 0.0026291804804649314, "grad_norm": 0.190278, "tokens_per_sec": 6156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:21.624834+00:00", "epoch": 0, "step": 15502, "train_loss": 3.584287643432617, "perplexity": 36.02768401821232, "lr": 0.0026291804804649314, "grad_norm": 0.211789, "tokens_per_sec": 108102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:21.929056+00:00", "epoch": 0, "step": 15503, "train_loss": 3.528169631958008, "perplexity": 34.061565321969496, "lr": 0.0026291804804649314, "grad_norm": 0.195537, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:22.234162+00:00", "epoch": 0, "step": 15504, "train_loss": 3.517728567123413, "perplexity": 33.70777649018315, "lr": 0.0026291804804649314, "grad_norm": 0.202698, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:22.539443+00:00", "epoch": 0, "step": 15505, "train_loss": 3.458430767059326, "perplexity": 31.767087421090363, "lr": 0.0026291804804649314, "grad_norm": 0.180117, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:22.844914+00:00", "epoch": 0, "step": 15506, "train_loss": 3.4906070232391357, "perplexity": 32.805855580066385, "lr": 0.0026291804804649314, "grad_norm": 0.182739, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:23.149514+00:00", "epoch": 0, "step": 15507, "train_loss": 3.5114669799804688, "perplexity": 33.49737173161331, "lr": 0.0026291804804649314, "grad_norm": 0.225433, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:23.453403+00:00", "epoch": 0, "step": 15508, "train_loss": 3.396986961364746, "perplexity": 29.873952932852383, "lr": 0.0026291804804649314, "grad_norm": 0.176925, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:23.758250+00:00", "epoch": 0, "step": 15509, "train_loss": 3.4664804935455322, "perplexity": 32.023835776713916, "lr": 0.0026291804804649314, "grad_norm": 0.182761, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:24.062122+00:00", "epoch": 0, "step": 15510, "train_loss": 3.4843130111694336, "perplexity": 32.60002356318327, "lr": 0.0026291804804649314, "grad_norm": 0.206178, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:24.365414+00:00", "epoch": 0, "step": 15511, "train_loss": 3.519376754760742, "perplexity": 33.76337903980644, "lr": 0.0026291804804649314, "grad_norm": 0.177899, "tokens_per_sec": 107986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:24.668898+00:00", "epoch": 0, "step": 15512, "train_loss": 3.5570902824401855, "perplexity": 35.061030882480644, "lr": 0.0026291804804649314, "grad_norm": 0.173622, "tokens_per_sec": 107972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:24.973918+00:00", "epoch": 0, "step": 15513, "train_loss": 3.6143174171447754, "perplexity": 37.12599570666585, "lr": 0.0026291804804649314, "grad_norm": 0.175113, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:25.353435+00:00", "epoch": 0, "step": 15514, "train_loss": 3.565242290496826, "perplexity": 35.3480168553821, "lr": 0.0026291804804649314, "grad_norm": 0.160077, "tokens_per_sec": 86341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:25.657991+00:00", "epoch": 0, "step": 15515, "train_loss": 3.6666300296783447, "perplexity": 39.119850738383555, "lr": 0.0026291804804649314, "grad_norm": 0.169478, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:25.962750+00:00", "epoch": 0, "step": 15516, "train_loss": 3.582214832305908, "perplexity": 35.95308277776109, "lr": 0.0026291804804649314, "grad_norm": 0.166951, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:26.266938+00:00", "epoch": 0, "step": 15517, "train_loss": 3.642017126083374, "perplexity": 38.16875030894039, "lr": 0.0026291804804649314, "grad_norm": 0.151171, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:26.571568+00:00", "epoch": 0, "step": 15518, "train_loss": 3.553603172302246, "perplexity": 34.93898212864256, "lr": 0.0026291804804649314, "grad_norm": 0.175868, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:26.876299+00:00", "epoch": 0, "step": 15519, "train_loss": 3.4888930320739746, "perplexity": 32.749674793876025, "lr": 0.0026291804804649314, "grad_norm": 0.171604, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:27.181422+00:00", "epoch": 0, "step": 15520, "train_loss": 3.5131115913391113, "perplexity": 33.55250721544199, "lr": 0.0026291804804649314, "grad_norm": 0.156478, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:27.486783+00:00", "epoch": 0, "step": 15521, "train_loss": 3.5567455291748047, "perplexity": 35.04894556094302, "lr": 0.0026291804804649314, "grad_norm": 0.160545, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:27.791758+00:00", "epoch": 0, "step": 15522, "train_loss": 3.4945549964904785, "perplexity": 32.93562822128373, "lr": 0.0026291804804649314, "grad_norm": 0.169348, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:28.096002+00:00", "epoch": 0, "step": 15523, "train_loss": 3.6558165550231934, "perplexity": 38.69910816829077, "lr": 0.0026291804804649314, "grad_norm": 0.176645, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:28.400284+00:00", "epoch": 0, "step": 15524, "train_loss": 3.5596823692321777, "perplexity": 35.15203000540579, "lr": 0.0026291804804649314, "grad_norm": 0.217378, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:28.703826+00:00", "epoch": 0, "step": 15525, "train_loss": 3.438025951385498, "perplexity": 31.12545432251027, "lr": 0.0026291804804649314, "grad_norm": 0.158049, "tokens_per_sec": 107952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:29.008475+00:00", "epoch": 0, "step": 15526, "train_loss": 3.4838144779205322, "perplexity": 32.583775417978885, "lr": 0.0026291804804649314, "grad_norm": 0.187479, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:29.313127+00:00", "epoch": 0, "step": 15527, "train_loss": 3.4201467037200928, "perplexity": 30.573899996928137, "lr": 0.0026291804804649314, "grad_norm": 0.177057, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:29.616879+00:00", "epoch": 0, "step": 15528, "train_loss": 3.5924456119537354, "perplexity": 36.32279886241004, "lr": 0.0026291804804649314, "grad_norm": 0.165997, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:29.921678+00:00", "epoch": 0, "step": 15529, "train_loss": 3.470554828643799, "perplexity": 32.154577777299046, "lr": 0.0026291804804649314, "grad_norm": 0.164728, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:30.226730+00:00", "epoch": 0, "step": 15530, "train_loss": 3.5190985202789307, "perplexity": 33.75398621030021, "lr": 0.0026291804804649314, "grad_norm": 0.159585, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:30.532096+00:00", "epoch": 0, "step": 15531, "train_loss": 3.465940475463867, "perplexity": 32.00654699489776, "lr": 0.0026291804804649314, "grad_norm": 0.164409, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:30.836300+00:00", "epoch": 0, "step": 15532, "train_loss": 3.454580068588257, "perplexity": 31.6449971633736, "lr": 0.0026291804804649314, "grad_norm": 0.158821, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:31.139742+00:00", "epoch": 0, "step": 15533, "train_loss": 3.5350682735443115, "perplexity": 34.2973562363667, "lr": 0.0026291804804649314, "grad_norm": 0.187752, "tokens_per_sec": 107988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:31.443020+00:00", "epoch": 0, "step": 15534, "train_loss": 3.5400609970092773, "perplexity": 34.46902163396775, "lr": 0.0026291804804649314, "grad_norm": 0.161451, "tokens_per_sec": 108047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:31.747046+00:00", "epoch": 0, "step": 15535, "train_loss": 3.553638458251953, "perplexity": 34.940215005560255, "lr": 0.0026291804804649314, "grad_norm": 0.161531, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:32.051309+00:00", "epoch": 0, "step": 15536, "train_loss": 3.4743525981903076, "perplexity": 32.27692563098649, "lr": 0.0026291804804649314, "grad_norm": 0.17836, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:32.355562+00:00", "epoch": 0, "step": 15537, "train_loss": 3.554949998855591, "perplexity": 34.98607058039086, "lr": 0.0026291804804649314, "grad_norm": 0.151978, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:32.660716+00:00", "epoch": 0, "step": 15538, "train_loss": 3.567918062210083, "perplexity": 35.442726733501594, "lr": 0.0026291804804649314, "grad_norm": 0.161476, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:32.966090+00:00", "epoch": 0, "step": 15539, "train_loss": 3.41756272315979, "perplexity": 30.49499961610974, "lr": 0.0026291804804649314, "grad_norm": 0.154658, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:33.271031+00:00", "epoch": 0, "step": 15540, "train_loss": 3.4757869243621826, "perplexity": 32.32325448756977, "lr": 0.0026291804804649314, "grad_norm": 0.163812, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:33.574956+00:00", "epoch": 0, "step": 15541, "train_loss": 3.4831528663635254, "perplexity": 32.56222474546455, "lr": 0.0026291804804649314, "grad_norm": 0.203369, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:33.880394+00:00", "epoch": 0, "step": 15542, "train_loss": 3.5496926307678223, "perplexity": 34.80261858927661, "lr": 0.0026291804804649314, "grad_norm": 0.152802, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:34.185659+00:00", "epoch": 0, "step": 15543, "train_loss": 3.44795298576355, "perplexity": 31.43597650976023, "lr": 0.0026291804804649314, "grad_norm": 0.166032, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:34.490465+00:00", "epoch": 0, "step": 15544, "train_loss": 3.5368638038635254, "perplexity": 34.358993498585015, "lr": 0.0026291804804649314, "grad_norm": 0.166998, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:34.794393+00:00", "epoch": 0, "step": 15545, "train_loss": 3.491321563720703, "perplexity": 32.829305068717545, "lr": 0.0026291804804649314, "grad_norm": 0.160865, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:35.099381+00:00", "epoch": 0, "step": 15546, "train_loss": 3.559197187423706, "perplexity": 35.13497901666558, "lr": 0.0026291804804649314, "grad_norm": 0.168981, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:35.404876+00:00", "epoch": 0, "step": 15547, "train_loss": 3.517223834991455, "perplexity": 33.69076738517371, "lr": 0.0026291804804649314, "grad_norm": 0.209269, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:35.710734+00:00", "epoch": 0, "step": 15548, "train_loss": 3.567356586456299, "perplexity": 35.42283208749607, "lr": 0.0026291804804649314, "grad_norm": 0.181672, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:36.015941+00:00", "epoch": 0, "step": 15549, "train_loss": 3.5715761184692383, "perplexity": 35.57261564750705, "lr": 0.0026291804804649314, "grad_norm": 0.173346, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:36.319091+00:00", "epoch": 0, "step": 15550, "train_loss": 3.450803756713867, "perplexity": 31.525721138300355, "lr": 0.0026291804804649314, "grad_norm": 0.182464, "tokens_per_sec": 108150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:36.623998+00:00", "epoch": 0, "step": 15551, "train_loss": 3.425201416015625, "perplexity": 30.72883350744465, "lr": 0.0026291804804649314, "grad_norm": 0.141621, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:36.928210+00:00", "epoch": 0, "step": 15552, "train_loss": 3.520899772644043, "perplexity": 33.81484044825999, "lr": 0.0026291804804649314, "grad_norm": 0.184082, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:37.232782+00:00", "epoch": 0, "step": 15553, "train_loss": 3.6072139739990234, "perplexity": 36.86320776154628, "lr": 0.0026291804804649314, "grad_norm": 0.156281, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:37.538071+00:00", "epoch": 0, "step": 15554, "train_loss": 3.514831066131592, "perplexity": 33.61024953482807, "lr": 0.0026291804804649314, "grad_norm": 0.168839, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:37.842570+00:00", "epoch": 0, "step": 15555, "train_loss": 3.491520881652832, "perplexity": 32.83584919007577, "lr": 0.0026291804804649314, "grad_norm": 0.168397, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:38.147354+00:00", "epoch": 0, "step": 15556, "train_loss": 3.560666799545288, "perplexity": 35.18665176787071, "lr": 0.0026291804804649314, "grad_norm": 0.163265, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:38.452490+00:00", "epoch": 0, "step": 15557, "train_loss": 3.4800784587860107, "perplexity": 32.4622689261622, "lr": 0.0026291804804649314, "grad_norm": 0.157027, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:38.758448+00:00", "epoch": 0, "step": 15558, "train_loss": 3.3985707759857178, "perplexity": 29.921305225021513, "lr": 0.0026291804804649314, "grad_norm": 0.153402, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:39.062666+00:00", "epoch": 0, "step": 15559, "train_loss": 3.462907552719116, "perplexity": 31.909620669862903, "lr": 0.0026291804804649314, "grad_norm": 0.152693, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:39.367082+00:00", "epoch": 0, "step": 15560, "train_loss": 3.4830479621887207, "perplexity": 32.558809011313336, "lr": 0.0026291804804649314, "grad_norm": 0.195074, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:39.673806+00:00", "epoch": 0, "step": 15561, "train_loss": 3.577531337738037, "perplexity": 35.78509041261565, "lr": 0.0026291804804649314, "grad_norm": 0.191081, "tokens_per_sec": 106776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:39.978502+00:00", "epoch": 0, "step": 15562, "train_loss": 3.599245071411133, "perplexity": 36.57061581655169, "lr": 0.0026291804804649314, "grad_norm": 0.186943, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:40.284327+00:00", "epoch": 0, "step": 15563, "train_loss": 3.3955304622650146, "perplexity": 29.830473219068708, "lr": 0.0026291804804649314, "grad_norm": 0.159362, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:40.589227+00:00", "epoch": 0, "step": 15564, "train_loss": 3.5236005783081055, "perplexity": 33.90629120063754, "lr": 0.0026291804804649314, "grad_norm": 0.138798, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:40.893627+00:00", "epoch": 0, "step": 15565, "train_loss": 3.5172882080078125, "perplexity": 33.69293623130052, "lr": 0.0026291804804649314, "grad_norm": 0.171547, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:41.197469+00:00", "epoch": 0, "step": 15566, "train_loss": 3.512279748916626, "perplexity": 33.524608421868194, "lr": 0.0026291804804649314, "grad_norm": 0.190239, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:41.501705+00:00", "epoch": 0, "step": 15567, "train_loss": 3.4920105934143066, "perplexity": 32.85193322957238, "lr": 0.0026291804804649314, "grad_norm": 0.153301, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:41.805995+00:00", "epoch": 0, "step": 15568, "train_loss": 3.479480266571045, "perplexity": 32.442856056503075, "lr": 0.0026291804804649314, "grad_norm": 0.198357, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:42.112212+00:00", "epoch": 0, "step": 15569, "train_loss": 3.5467467308044434, "perplexity": 34.70024442251239, "lr": 0.0026291804804649314, "grad_norm": 0.186846, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:42.417886+00:00", "epoch": 0, "step": 15570, "train_loss": 3.5979387760162354, "perplexity": 36.52287497812943, "lr": 0.0026291804804649314, "grad_norm": 0.163325, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:42.721697+00:00", "epoch": 0, "step": 15571, "train_loss": 3.4890756607055664, "perplexity": 32.75565636835546, "lr": 0.0026291804804649314, "grad_norm": 0.175604, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:43.025126+00:00", "epoch": 0, "step": 15572, "train_loss": 3.494218111038208, "perplexity": 32.92453455602673, "lr": 0.0026291804804649314, "grad_norm": 0.173353, "tokens_per_sec": 107993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:43.328799+00:00", "epoch": 0, "step": 15573, "train_loss": 3.462613344192505, "perplexity": 31.900233968272428, "lr": 0.0026291804804649314, "grad_norm": 0.157191, "tokens_per_sec": 107906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:43.633346+00:00", "epoch": 0, "step": 15574, "train_loss": 3.519782304763794, "perplexity": 33.77707455519951, "lr": 0.0026291804804649314, "grad_norm": 0.175741, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:43.937773+00:00", "epoch": 0, "step": 15575, "train_loss": 3.5362672805786133, "perplexity": 34.33850367085423, "lr": 0.0026291804804649314, "grad_norm": 0.150578, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:44.241811+00:00", "epoch": 0, "step": 15576, "train_loss": 3.5298562049865723, "perplexity": 34.11906111114683, "lr": 0.0026291804804649314, "grad_norm": 0.174451, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:44.545707+00:00", "epoch": 0, "step": 15577, "train_loss": 3.4422178268432617, "perplexity": 31.256202198546447, "lr": 0.0026291804804649314, "grad_norm": 0.150441, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:44.850076+00:00", "epoch": 0, "step": 15578, "train_loss": 3.4795217514038086, "perplexity": 32.444201970878275, "lr": 0.0026291804804649314, "grad_norm": 0.17364, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:45.154507+00:00", "epoch": 0, "step": 15579, "train_loss": 3.6411564350128174, "perplexity": 38.1359129398175, "lr": 0.0026291804804649314, "grad_norm": 0.203093, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:45.458969+00:00", "epoch": 0, "step": 15580, "train_loss": 3.5672104358673096, "perplexity": 35.41765539802014, "lr": 0.0026291804804649314, "grad_norm": 0.19623, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:45.763216+00:00", "epoch": 0, "step": 15581, "train_loss": 3.53334379196167, "perplexity": 34.23826204527059, "lr": 0.0026291804804649314, "grad_norm": 0.182737, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:46.066922+00:00", "epoch": 0, "step": 15582, "train_loss": 3.5129458904266357, "perplexity": 33.546947994977224, "lr": 0.0026291804804649314, "grad_norm": 0.176, "tokens_per_sec": 107892} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:46.370844+00:00", "epoch": 0, "step": 15583, "train_loss": 3.492388963699341, "perplexity": 32.86436577681727, "lr": 0.0026291804804649314, "grad_norm": 0.169586, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:46.675496+00:00", "epoch": 0, "step": 15584, "train_loss": 3.609694719314575, "perplexity": 36.95476951525609, "lr": 0.0026291804804649314, "grad_norm": 0.173508, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:46.981503+00:00", "epoch": 0, "step": 15585, "train_loss": 3.682734727859497, "perplexity": 39.75496455867383, "lr": 0.0026291804804649314, "grad_norm": 0.158956, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:47.285671+00:00", "epoch": 0, "step": 15586, "train_loss": 3.5806822776794434, "perplexity": 35.89802491478198, "lr": 0.0026291804804649314, "grad_norm": 0.172377, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:47.589670+00:00", "epoch": 0, "step": 15587, "train_loss": 3.5904266834259033, "perplexity": 36.2495397049616, "lr": 0.0026291804804649314, "grad_norm": 0.191501, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:47.893003+00:00", "epoch": 0, "step": 15588, "train_loss": 3.499447822570801, "perplexity": 33.09717140108312, "lr": 0.0026291804804649314, "grad_norm": 0.182152, "tokens_per_sec": 108017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:48.197979+00:00", "epoch": 0, "step": 15589, "train_loss": 3.491973638534546, "perplexity": 32.85071921276203, "lr": 0.0026291804804649314, "grad_norm": 0.181224, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:48.503729+00:00", "epoch": 0, "step": 15590, "train_loss": 3.533722400665283, "perplexity": 34.25122740352138, "lr": 0.0026291804804649314, "grad_norm": 0.164416, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:48.807806+00:00", "epoch": 0, "step": 15591, "train_loss": 3.602997064590454, "perplexity": 36.70808625045473, "lr": 0.0026291804804649314, "grad_norm": 0.202588, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:49.111808+00:00", "epoch": 0, "step": 15592, "train_loss": 3.4910452365875244, "perplexity": 32.82023469421666, "lr": 0.0026291804804649314, "grad_norm": 0.223463, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:49.415661+00:00", "epoch": 0, "step": 15593, "train_loss": 3.547764539718628, "perplexity": 34.73558062030806, "lr": 0.0026291804804649314, "grad_norm": 0.213273, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:49.720113+00:00", "epoch": 0, "step": 15594, "train_loss": 3.5321249961853027, "perplexity": 34.196558015609746, "lr": 0.0026291804804649314, "grad_norm": 0.168091, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:50.025007+00:00", "epoch": 0, "step": 15595, "train_loss": 3.588597059249878, "perplexity": 36.18327730687494, "lr": 0.0026291804804649314, "grad_norm": 0.21707, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:50.330174+00:00", "epoch": 0, "step": 15596, "train_loss": 3.4603073596954346, "perplexity": 31.82675727390748, "lr": 0.0026291804804649314, "grad_norm": 0.187325, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:50.634931+00:00", "epoch": 0, "step": 15597, "train_loss": 3.3995094299316406, "perplexity": 29.94940416176397, "lr": 0.0026291804804649314, "grad_norm": 0.177757, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:50.939026+00:00", "epoch": 0, "step": 15598, "train_loss": 3.538843870162964, "perplexity": 34.427093983181535, "lr": 0.0026291804804649314, "grad_norm": 0.175268, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:51.242819+00:00", "epoch": 0, "step": 15599, "train_loss": 3.4984519481658936, "perplexity": 33.06422718208495, "lr": 0.0026291804804649314, "grad_norm": 0.182061, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:51.548121+00:00", "epoch": 0, "step": 15600, "train_loss": 3.480001926422119, "perplexity": 32.459784607050615, "lr": 0.0026291804804649314, "grad_norm": 0.19914, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:51.853793+00:00", "epoch": 0, "step": 15601, "train_loss": 3.5364861488342285, "perplexity": 34.346020101777164, "lr": 0.0026291804804649314, "grad_norm": 0.176842, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:52.158957+00:00", "epoch": 0, "step": 15602, "train_loss": 3.4272961616516113, "perplexity": 30.79327006286788, "lr": 0.0026291804804649314, "grad_norm": 0.17402, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:52.463042+00:00", "epoch": 0, "step": 15603, "train_loss": 3.4595329761505127, "perplexity": 31.802120697097337, "lr": 0.0026291804804649314, "grad_norm": 0.171626, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:52.767053+00:00", "epoch": 0, "step": 15604, "train_loss": 3.4630045890808105, "perplexity": 31.912717213591993, "lr": 0.0026291804804649314, "grad_norm": 0.187731, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:53.070969+00:00", "epoch": 0, "step": 15605, "train_loss": 3.5430264472961426, "perplexity": 34.57138951246981, "lr": 0.0026291804804649314, "grad_norm": 0.16229, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:53.375714+00:00", "epoch": 0, "step": 15606, "train_loss": 3.537168264389038, "perplexity": 34.36945604843758, "lr": 0.0026291804804649314, "grad_norm": 0.158728, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:53.680314+00:00", "epoch": 0, "step": 15607, "train_loss": 3.3522379398345947, "perplexity": 28.566592476243592, "lr": 0.0026291804804649314, "grad_norm": 0.154133, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:53.985504+00:00", "epoch": 0, "step": 15608, "train_loss": 3.4826226234436035, "perplexity": 32.54496343308513, "lr": 0.0026291804804649314, "grad_norm": 0.170712, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:54.289623+00:00", "epoch": 0, "step": 15609, "train_loss": 3.4432506561279297, "perplexity": 31.288501196308776, "lr": 0.0026291804804649314, "grad_norm": 0.163787, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:54.595140+00:00", "epoch": 0, "step": 15610, "train_loss": 3.4892914295196533, "perplexity": 32.762724780028684, "lr": 0.0026291804804649314, "grad_norm": 0.170209, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:54.900481+00:00", "epoch": 0, "step": 15611, "train_loss": 3.6953325271606445, "perplexity": 40.25895755926275, "lr": 0.0026291804804649314, "grad_norm": 0.151848, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:55.205203+00:00", "epoch": 0, "step": 15612, "train_loss": 3.4553728103637695, "perplexity": 31.670093420722498, "lr": 0.0026291804804649314, "grad_norm": 0.148883, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:55.509587+00:00", "epoch": 0, "step": 15613, "train_loss": 3.4372808933258057, "perplexity": 31.102272688809574, "lr": 0.0026291804804649314, "grad_norm": 0.149817, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:55.813693+00:00", "epoch": 0, "step": 15614, "train_loss": 3.457127571105957, "perplexity": 31.725715644933718, "lr": 0.0026291804804649314, "grad_norm": 0.181652, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:56.118300+00:00", "epoch": 0, "step": 15615, "train_loss": 3.5948867797851562, "perplexity": 36.41157722789703, "lr": 0.0026291804804649314, "grad_norm": 0.17183, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:56.423851+00:00", "epoch": 0, "step": 15616, "train_loss": 3.6413984298706055, "perplexity": 38.14514275138442, "lr": 0.0026291804804649314, "grad_norm": 0.173991, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:56.728526+00:00", "epoch": 0, "step": 15617, "train_loss": 3.4016411304473877, "perplexity": 30.013315417688684, "lr": 0.0026291804804649314, "grad_norm": 0.2086, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:57.032700+00:00", "epoch": 0, "step": 15618, "train_loss": 3.5015010833740234, "perplexity": 33.16519834044313, "lr": 0.0026291804804649314, "grad_norm": 0.178023, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:57.336535+00:00", "epoch": 0, "step": 15619, "train_loss": 3.512730598449707, "perplexity": 33.539726383628846, "lr": 0.0026291804804649314, "grad_norm": 0.176617, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:57.641156+00:00", "epoch": 0, "step": 15620, "train_loss": 3.4865963459014893, "perplexity": 32.674545375995315, "lr": 0.0026291804804649314, "grad_norm": 0.188825, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:57.946678+00:00", "epoch": 0, "step": 15621, "train_loss": 3.5408225059509277, "perplexity": 34.49528009891095, "lr": 0.0026291804804649314, "grad_norm": 0.180129, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:58.250594+00:00", "epoch": 0, "step": 15622, "train_loss": 3.524333953857422, "perplexity": 33.93116636587617, "lr": 0.0026291804804649314, "grad_norm": 0.18317, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:58.555359+00:00", "epoch": 0, "step": 15623, "train_loss": 3.4496195316314697, "perplexity": 31.48840968552726, "lr": 0.0026291804804649314, "grad_norm": 0.168653, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:58.861194+00:00", "epoch": 0, "step": 15624, "train_loss": 3.5850908756256104, "perplexity": 36.05663423917296, "lr": 0.0026291804804649314, "grad_norm": 0.17201, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:59.166637+00:00", "epoch": 0, "step": 15625, "train_loss": 3.597236156463623, "perplexity": 36.49722230514233, "lr": 0.0026291804804649314, "grad_norm": 0.203786, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:59.471533+00:00", "epoch": 0, "step": 15626, "train_loss": 3.6004092693328857, "perplexity": 36.61321604421926, "lr": 0.0026291804804649314, "grad_norm": 0.201265, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:19:59.778706+00:00", "epoch": 0, "step": 15627, "train_loss": 3.4912688732147217, "perplexity": 32.82757532159348, "lr": 0.0026291804804649314, "grad_norm": 0.179276, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:00.082806+00:00", "epoch": 0, "step": 15628, "train_loss": 3.4028775691986084, "perplexity": 30.05044799527202, "lr": 0.0026291804804649314, "grad_norm": 0.195954, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:00.387850+00:00", "epoch": 0, "step": 15629, "train_loss": 3.457455635070801, "perplexity": 31.736125416437794, "lr": 0.0026291804804649314, "grad_norm": 0.194191, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:00.692032+00:00", "epoch": 0, "step": 15630, "train_loss": 3.455254554748535, "perplexity": 31.66634847577525, "lr": 0.0026291804804649314, "grad_norm": 0.180358, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:00.996149+00:00", "epoch": 0, "step": 15631, "train_loss": 3.4513351917266846, "perplexity": 31.542479462903376, "lr": 0.0026291804804649314, "grad_norm": 0.157905, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:01.301432+00:00", "epoch": 0, "step": 15632, "train_loss": 3.5226693153381348, "perplexity": 33.87473022525811, "lr": 0.0026291804804649314, "grad_norm": 0.164395, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:01.607638+00:00", "epoch": 0, "step": 15633, "train_loss": 3.443793535232544, "perplexity": 31.305491681295305, "lr": 0.0026291804804649314, "grad_norm": 0.170865, "tokens_per_sec": 107015} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:01.912090+00:00", "epoch": 0, "step": 15634, "train_loss": 3.44246768951416, "perplexity": 31.2640129324747, "lr": 0.0026291804804649314, "grad_norm": 0.164759, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:02.216186+00:00", "epoch": 0, "step": 15635, "train_loss": 3.4151225090026855, "perplexity": 30.42067600596934, "lr": 0.0026291804804649314, "grad_norm": 0.158837, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:02.521492+00:00", "epoch": 0, "step": 15636, "train_loss": 3.4871437549591064, "perplexity": 32.69243661455391, "lr": 0.0026291804804649314, "grad_norm": 0.155081, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:02.827054+00:00", "epoch": 0, "step": 15637, "train_loss": 3.5620670318603516, "perplexity": 35.235955765204466, "lr": 0.0026291804804649314, "grad_norm": 0.175448, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:03.131756+00:00", "epoch": 0, "step": 15638, "train_loss": 3.589360237121582, "perplexity": 36.210902123424816, "lr": 0.0026291804804649314, "grad_norm": 0.217272, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:03.435851+00:00", "epoch": 0, "step": 15639, "train_loss": 3.603320837020874, "perplexity": 36.71997324099217, "lr": 0.0026291804804649314, "grad_norm": 0.171807, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:03.740467+00:00", "epoch": 0, "step": 15640, "train_loss": 3.557922124862671, "perplexity": 35.09020826915629, "lr": 0.0026291804804649314, "grad_norm": 0.174361, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:04.044942+00:00", "epoch": 0, "step": 15641, "train_loss": 3.494479179382324, "perplexity": 32.9331312318552, "lr": 0.0026291804804649314, "grad_norm": 0.167845, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:04.350397+00:00", "epoch": 0, "step": 15642, "train_loss": 3.544461727142334, "perplexity": 34.6210447571536, "lr": 0.0026291804804649314, "grad_norm": 0.177467, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:04.655751+00:00", "epoch": 0, "step": 15643, "train_loss": 3.3675484657287598, "perplexity": 29.007327362064093, "lr": 0.0026291804804649314, "grad_norm": 0.17763, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:04.961654+00:00", "epoch": 0, "step": 15644, "train_loss": 3.505030632019043, "perplexity": 33.28246334584011, "lr": 0.0026291804804649314, "grad_norm": 0.158007, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:05.265374+00:00", "epoch": 0, "step": 15645, "train_loss": 3.4857404232025146, "perplexity": 32.64659045626116, "lr": 0.0026291804804649314, "grad_norm": 0.150784, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:05.570439+00:00", "epoch": 0, "step": 15646, "train_loss": 3.557532548904419, "perplexity": 35.07654063010923, "lr": 0.0026291804804649314, "grad_norm": 0.183358, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:05.874145+00:00", "epoch": 0, "step": 15647, "train_loss": 3.479875087738037, "perplexity": 32.45566771178178, "lr": 0.0026291804804649314, "grad_norm": 0.156784, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:06.180870+00:00", "epoch": 0, "step": 15648, "train_loss": 3.6012356281280518, "perplexity": 36.64348420177338, "lr": 0.0026291804804649314, "grad_norm": 0.159858, "tokens_per_sec": 106832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:06.487168+00:00", "epoch": 0, "step": 15649, "train_loss": 3.5117785930633545, "perplexity": 33.50781157739896, "lr": 0.0026291804804649314, "grad_norm": 0.186479, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:06.791672+00:00", "epoch": 0, "step": 15650, "train_loss": 3.4038453102111816, "perplexity": 30.079543122246378, "lr": 0.0026291804804649314, "grad_norm": 0.163195, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:07.095884+00:00", "epoch": 0, "step": 15651, "train_loss": 3.505802631378174, "perplexity": 33.30816730665744, "lr": 0.0026291804804649314, "grad_norm": 0.147152, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:07.401199+00:00", "epoch": 0, "step": 15652, "train_loss": 3.5008158683776855, "perplexity": 33.142480833260905, "lr": 0.0026291804804649314, "grad_norm": 0.164151, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:07.706449+00:00", "epoch": 0, "step": 15653, "train_loss": 3.6015665531158447, "perplexity": 36.65561245299556, "lr": 0.0026291804804649314, "grad_norm": 0.162806, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:08.011354+00:00", "epoch": 0, "step": 15654, "train_loss": 3.523430347442627, "perplexity": 33.90051979459095, "lr": 0.0026291804804649314, "grad_norm": 0.163469, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:08.315239+00:00", "epoch": 0, "step": 15655, "train_loss": 3.432559013366699, "perplexity": 30.95575767609537, "lr": 0.0026291804804649314, "grad_norm": 0.155375, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:08.620428+00:00", "epoch": 0, "step": 15656, "train_loss": 3.432893753051758, "perplexity": 30.966121531170636, "lr": 0.0026291804804649314, "grad_norm": 0.168407, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:08.924508+00:00", "epoch": 0, "step": 15657, "train_loss": 3.492401361465454, "perplexity": 32.86477322406335, "lr": 0.0026291804804649314, "grad_norm": 0.161852, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:09.230780+00:00", "epoch": 0, "step": 15658, "train_loss": 3.5323240756988525, "perplexity": 34.2033665274397, "lr": 0.0026291804804649314, "grad_norm": 0.148729, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:09.536626+00:00", "epoch": 0, "step": 15659, "train_loss": 3.5153627395629883, "perplexity": 33.62812396278651, "lr": 0.0026291804804649314, "grad_norm": 0.18605, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:09.842223+00:00", "epoch": 0, "step": 15660, "train_loss": 3.507380485534668, "perplexity": 33.36076422110528, "lr": 0.0026291804804649314, "grad_norm": 0.151981, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:10.147322+00:00", "epoch": 0, "step": 15661, "train_loss": 3.6330432891845703, "perplexity": 37.82776244209012, "lr": 0.0026291804804649314, "grad_norm": 0.160458, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:10.452389+00:00", "epoch": 0, "step": 15662, "train_loss": 3.458329677581787, "perplexity": 31.763876265129802, "lr": 0.0026291804804649314, "grad_norm": 0.168421, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:10.757383+00:00", "epoch": 0, "step": 15663, "train_loss": 3.4986629486083984, "perplexity": 33.07120448473254, "lr": 0.0026291804804649314, "grad_norm": 0.1664, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:11.062647+00:00", "epoch": 0, "step": 15664, "train_loss": 3.5664122104644775, "perplexity": 35.389395406194886, "lr": 0.0026291804804649314, "grad_norm": 0.191528, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:11.367319+00:00", "epoch": 0, "step": 15665, "train_loss": 3.4825854301452637, "perplexity": 32.54375300106082, "lr": 0.0026291804804649314, "grad_norm": 0.164999, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:11.672713+00:00", "epoch": 0, "step": 15666, "train_loss": 3.5526134967803955, "perplexity": 34.90442097825467, "lr": 0.0026291804804649314, "grad_norm": 0.181999, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:11.977037+00:00", "epoch": 0, "step": 15667, "train_loss": 3.5246150493621826, "perplexity": 33.9407056048687, "lr": 0.0026291804804649314, "grad_norm": 0.211258, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:12.281749+00:00", "epoch": 0, "step": 15668, "train_loss": 3.4696764945983887, "perplexity": 32.126347716448116, "lr": 0.0026291804804649314, "grad_norm": 0.179885, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:12.586483+00:00", "epoch": 0, "step": 15669, "train_loss": 3.507812261581421, "perplexity": 33.37517171017783, "lr": 0.0026291804804649314, "grad_norm": 0.172319, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:12.891287+00:00", "epoch": 0, "step": 15670, "train_loss": 3.484755516052246, "perplexity": 32.614452424975234, "lr": 0.0026291804804649314, "grad_norm": 0.174412, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:13.195553+00:00", "epoch": 0, "step": 15671, "train_loss": 3.4733376502990723, "perplexity": 32.244182852299595, "lr": 0.0026291804804649314, "grad_norm": 0.157254, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:13.501282+00:00", "epoch": 0, "step": 15672, "train_loss": 3.498702049255371, "perplexity": 33.072497615505014, "lr": 0.0026291804804649314, "grad_norm": 0.178333, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:13.807959+00:00", "epoch": 0, "step": 15673, "train_loss": 3.4915876388549805, "perplexity": 32.838041292666375, "lr": 0.0026291804804649314, "grad_norm": 0.149851, "tokens_per_sec": 106849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:14.112189+00:00", "epoch": 0, "step": 15674, "train_loss": 3.4781131744384766, "perplexity": 32.39853398630986, "lr": 0.0026291804804649314, "grad_norm": 0.170828, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:14.417041+00:00", "epoch": 0, "step": 15675, "train_loss": 3.523311138153076, "perplexity": 33.896478778578945, "lr": 0.0026291804804649314, "grad_norm": 0.171835, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:14.721967+00:00", "epoch": 0, "step": 15676, "train_loss": 3.4815750122070312, "perplexity": 32.51088681633866, "lr": 0.0026291804804649314, "grad_norm": 0.159183, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:15.028299+00:00", "epoch": 0, "step": 15677, "train_loss": 3.5675158500671387, "perplexity": 35.42847410491261, "lr": 0.0026291804804649314, "grad_norm": 0.164961, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:15.334670+00:00", "epoch": 0, "step": 15678, "train_loss": 3.602540969848633, "perplexity": 36.69134770280552, "lr": 0.0026291804804649314, "grad_norm": 0.180678, "tokens_per_sec": 106959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:15.640147+00:00", "epoch": 0, "step": 15679, "train_loss": 3.5060489177703857, "perplexity": 33.31637166528406, "lr": 0.0026291804804649314, "grad_norm": 0.1675, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:15.944569+00:00", "epoch": 0, "step": 15680, "train_loss": 3.4916815757751465, "perplexity": 32.84112614201805, "lr": 0.0026291804804649314, "grad_norm": 0.197232, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:16.249884+00:00", "epoch": 0, "step": 15681, "train_loss": 3.486830711364746, "perplexity": 32.68220405838934, "lr": 0.0026291804804649314, "grad_norm": 0.177943, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:16.554755+00:00", "epoch": 0, "step": 15682, "train_loss": 3.517432689666748, "perplexity": 33.69780459430774, "lr": 0.0026291804804649314, "grad_norm": 0.17388, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:16.858384+00:00", "epoch": 0, "step": 15683, "train_loss": 3.4588117599487305, "perplexity": 31.779192761392743, "lr": 0.0026291804804649314, "grad_norm": 0.196852, "tokens_per_sec": 107921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:17.162416+00:00", "epoch": 0, "step": 15684, "train_loss": 3.6117513179779053, "perplexity": 37.03084885035886, "lr": 0.0026291804804649314, "grad_norm": 0.217465, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:17.466646+00:00", "epoch": 0, "step": 15685, "train_loss": 3.5333263874053955, "perplexity": 34.237666148697755, "lr": 0.0026291804804649314, "grad_norm": 0.183129, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:17.771761+00:00", "epoch": 0, "step": 15686, "train_loss": 3.5619194507598877, "perplexity": 35.23075598778062, "lr": 0.0026291804804649314, "grad_norm": 0.184966, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:18.077332+00:00", "epoch": 0, "step": 15687, "train_loss": 3.458085298538208, "perplexity": 31.75611478783759, "lr": 0.0026291804804649314, "grad_norm": 0.166487, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:18.381610+00:00", "epoch": 0, "step": 15688, "train_loss": 3.4387028217315674, "perplexity": 31.146529351275486, "lr": 0.0026291804804649314, "grad_norm": 0.160033, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:18.685526+00:00", "epoch": 0, "step": 15689, "train_loss": 3.603243589401245, "perplexity": 36.71713682002126, "lr": 0.0026291804804649314, "grad_norm": 0.16338, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:18.990442+00:00", "epoch": 0, "step": 15690, "train_loss": 3.5285558700561523, "perplexity": 34.074723737156326, "lr": 0.0026291804804649314, "grad_norm": 0.165411, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:19.294319+00:00", "epoch": 0, "step": 15691, "train_loss": 3.5125207901000977, "perplexity": 33.53269020714004, "lr": 0.0026291804804649314, "grad_norm": 0.140885, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:19.599048+00:00", "epoch": 0, "step": 15692, "train_loss": 3.4778659343719482, "perplexity": 32.39052476075319, "lr": 0.0026291804804649314, "grad_norm": 0.167981, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:19.903325+00:00", "epoch": 0, "step": 15693, "train_loss": 3.457019329071045, "perplexity": 31.722281774761168, "lr": 0.0026291804804649314, "grad_norm": 0.158731, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:20.207683+00:00", "epoch": 0, "step": 15694, "train_loss": 3.4710352420806885, "perplexity": 32.170028979706224, "lr": 0.0026291804804649314, "grad_norm": 0.142398, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:20.512576+00:00", "epoch": 0, "step": 15695, "train_loss": 3.530763864517212, "perplexity": 34.15004366081701, "lr": 0.0026291804804649314, "grad_norm": 0.159655, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:20.817643+00:00", "epoch": 0, "step": 15696, "train_loss": 3.457094669342041, "perplexity": 31.724671830099272, "lr": 0.0026291804804649314, "grad_norm": 0.153778, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:21.123321+00:00", "epoch": 0, "step": 15697, "train_loss": 3.4751627445220947, "perplexity": 32.30308525902035, "lr": 0.0026291804804649314, "grad_norm": 0.153336, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:21.429016+00:00", "epoch": 0, "step": 15698, "train_loss": 3.462010622024536, "perplexity": 31.881012783195413, "lr": 0.0026291804804649314, "grad_norm": 0.152927, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:21.732941+00:00", "epoch": 0, "step": 15699, "train_loss": 3.5936310291290283, "perplexity": 36.365882062774, "lr": 0.0026291804804649314, "grad_norm": 0.164551, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:22.038989+00:00", "epoch": 0, "step": 15700, "train_loss": 3.5316002368927, "perplexity": 34.17861776158523, "lr": 0.0026291804804649314, "grad_norm": 0.183704, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:22.342805+00:00", "epoch": 0, "step": 15701, "train_loss": 3.656562566757202, "perplexity": 38.72798892843354, "lr": 0.0026291804804649314, "grad_norm": 0.202837, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:22.648655+00:00", "epoch": 0, "step": 15702, "train_loss": 3.4561679363250732, "perplexity": 31.695285148169244, "lr": 0.0026291804804649314, "grad_norm": 0.199494, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:22.953295+00:00", "epoch": 0, "step": 15703, "train_loss": 3.4944379329681396, "perplexity": 32.93177288629765, "lr": 0.0026291804804649314, "grad_norm": 0.173964, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:23.257120+00:00", "epoch": 0, "step": 15704, "train_loss": 3.5669596195220947, "perplexity": 35.408773185085785, "lr": 0.0026291804804649314, "grad_norm": 0.169001, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:23.561098+00:00", "epoch": 0, "step": 15705, "train_loss": 3.595893621444702, "perplexity": 36.448256382691824, "lr": 0.0026291804804649314, "grad_norm": 0.155375, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:23.865443+00:00", "epoch": 0, "step": 15706, "train_loss": 3.5524730682373047, "perplexity": 34.89951974541381, "lr": 0.0026291804804649314, "grad_norm": 0.151692, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:24.169868+00:00", "epoch": 0, "step": 15707, "train_loss": 3.4037325382232666, "perplexity": 30.07615118363433, "lr": 0.0026291804804649314, "grad_norm": 0.150112, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:24.475377+00:00", "epoch": 0, "step": 15708, "train_loss": 3.496612071990967, "perplexity": 33.00344902753476, "lr": 0.0026291804804649314, "grad_norm": 0.159535, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:24.779866+00:00", "epoch": 0, "step": 15709, "train_loss": 3.5067617893218994, "perplexity": 33.34013042631016, "lr": 0.0026291804804649314, "grad_norm": 0.159866, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:25.084256+00:00", "epoch": 0, "step": 15710, "train_loss": 3.54276180267334, "perplexity": 34.562241590660065, "lr": 0.0026291804804649314, "grad_norm": 0.161123, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:25.387940+00:00", "epoch": 0, "step": 15711, "train_loss": 3.533582925796509, "perplexity": 34.24645055120708, "lr": 0.0026291804804649314, "grad_norm": 0.166828, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:25.692512+00:00", "epoch": 0, "step": 15712, "train_loss": 3.5340123176574707, "perplexity": 34.26115885592778, "lr": 0.0026291804804649314, "grad_norm": 0.184854, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:25.998886+00:00", "epoch": 0, "step": 15713, "train_loss": 3.49275279045105, "perplexity": 32.87632488765514, "lr": 0.0026291804804649314, "grad_norm": 0.155296, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:26.304486+00:00", "epoch": 0, "step": 15714, "train_loss": 3.467629909515381, "perplexity": 32.06066564733738, "lr": 0.0026291804804649314, "grad_norm": 0.167825, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:26.608535+00:00", "epoch": 0, "step": 15715, "train_loss": 3.4399213790893555, "perplexity": 31.184506317642665, "lr": 0.0026291804804649314, "grad_norm": 0.183888, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:26.912913+00:00", "epoch": 0, "step": 15716, "train_loss": 3.4995276927948, "perplexity": 33.099814985147084, "lr": 0.0026291804804649314, "grad_norm": 0.188074, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:27.218103+00:00", "epoch": 0, "step": 15717, "train_loss": 3.478550910949707, "perplexity": 32.41271911199321, "lr": 0.0026291804804649314, "grad_norm": 0.187058, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:27.524265+00:00", "epoch": 0, "step": 15718, "train_loss": 3.5238049030303955, "perplexity": 33.913219801989165, "lr": 0.0026291804804649314, "grad_norm": 0.146565, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:27.828481+00:00", "epoch": 0, "step": 15719, "train_loss": 3.4963271617889404, "perplexity": 32.99404734758562, "lr": 0.0026291804804649314, "grad_norm": 0.177376, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:28.132896+00:00", "epoch": 0, "step": 15720, "train_loss": 3.423654794692993, "perplexity": 30.681344371643764, "lr": 0.0026291804804649314, "grad_norm": 0.174029, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:28.436393+00:00", "epoch": 0, "step": 15721, "train_loss": 3.513284921646118, "perplexity": 33.55832338586323, "lr": 0.0026291804804649314, "grad_norm": 0.22296, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:28.741225+00:00", "epoch": 0, "step": 15722, "train_loss": 3.4445173740386963, "perplexity": 31.328160014139655, "lr": 0.0026291804804649314, "grad_norm": 0.167519, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:29.045901+00:00", "epoch": 0, "step": 15723, "train_loss": 3.4870810508728027, "perplexity": 32.69038672945571, "lr": 0.0026291804804649314, "grad_norm": 0.177899, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:29.350753+00:00", "epoch": 0, "step": 15724, "train_loss": 3.5002031326293945, "perplexity": 33.12217947078775, "lr": 0.0026291804804649314, "grad_norm": 0.22568, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:29.654646+00:00", "epoch": 0, "step": 15725, "train_loss": 3.604189157485962, "perplexity": 36.75187179231302, "lr": 0.0026291804804649314, "grad_norm": 0.179957, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:29.959526+00:00", "epoch": 0, "step": 15726, "train_loss": 3.581214189529419, "perplexity": 35.91712457884337, "lr": 0.0026291804804649314, "grad_norm": 0.155451, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:30.264280+00:00", "epoch": 0, "step": 15727, "train_loss": 3.471519947052002, "perplexity": 32.18562573228638, "lr": 0.0026291804804649314, "grad_norm": 0.183499, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:30.570124+00:00", "epoch": 0, "step": 15728, "train_loss": 3.5719141960144043, "perplexity": 35.58464398322083, "lr": 0.0026291804804649314, "grad_norm": 0.172229, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:30.875485+00:00", "epoch": 0, "step": 15729, "train_loss": 3.5768275260925293, "perplexity": 35.75991331025592, "lr": 0.0026291804804649314, "grad_norm": 0.173686, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:31.180438+00:00", "epoch": 0, "step": 15730, "train_loss": 3.440065383911133, "perplexity": 31.188997360275287, "lr": 0.0026291804804649314, "grad_norm": 0.173865, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:31.484136+00:00", "epoch": 0, "step": 15731, "train_loss": 3.5580031871795654, "perplexity": 35.09305287803264, "lr": 0.0026291804804649314, "grad_norm": 0.163447, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:31.788759+00:00", "epoch": 0, "step": 15732, "train_loss": 3.513627052307129, "perplexity": 33.5698066815074, "lr": 0.0026291804804649314, "grad_norm": 0.157801, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:32.092865+00:00", "epoch": 0, "step": 15733, "train_loss": 3.635848045349121, "perplexity": 37.93400902021277, "lr": 0.0026291804804649314, "grad_norm": 0.158484, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:32.398081+00:00", "epoch": 0, "step": 15734, "train_loss": 3.45778489112854, "perplexity": 31.746576448418708, "lr": 0.0026291804804649314, "grad_norm": 0.178262, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:32.703955+00:00", "epoch": 0, "step": 15735, "train_loss": 3.4929189682006836, "perplexity": 32.88178865530693, "lr": 0.0026291804804649314, "grad_norm": 0.171699, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:33.009362+00:00", "epoch": 0, "step": 15736, "train_loss": 3.3910531997680664, "perplexity": 29.697212903375586, "lr": 0.0026291804804649314, "grad_norm": 0.160351, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:33.314227+00:00", "epoch": 0, "step": 15737, "train_loss": 3.5575952529907227, "perplexity": 35.07874014149859, "lr": 0.0026291804804649314, "grad_norm": 0.197824, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:33.619365+00:00", "epoch": 0, "step": 15738, "train_loss": 3.4701530933380127, "perplexity": 32.14166274254967, "lr": 0.0026291804804649314, "grad_norm": 0.213146, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:33.924718+00:00", "epoch": 0, "step": 15739, "train_loss": 3.524088144302368, "perplexity": 33.92282678598552, "lr": 0.0026291804804649314, "grad_norm": 0.215938, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:34.230669+00:00", "epoch": 0, "step": 15740, "train_loss": 3.5744848251342773, "perplexity": 35.67623658012479, "lr": 0.0026291804804649314, "grad_norm": 0.174158, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:34.534993+00:00", "epoch": 0, "step": 15741, "train_loss": 3.49402117729187, "perplexity": 32.91805124250274, "lr": 0.0026291804804649314, "grad_norm": 0.218972, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:34.839229+00:00", "epoch": 0, "step": 15742, "train_loss": 3.5238468647003174, "perplexity": 33.91464288718179, "lr": 0.0026291804804649314, "grad_norm": 0.21798, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:35.144500+00:00", "epoch": 0, "step": 15743, "train_loss": 3.5203187465667725, "perplexity": 33.795198850853524, "lr": 0.0026291804804649314, "grad_norm": 0.180805, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:35.450113+00:00", "epoch": 0, "step": 15744, "train_loss": 3.4668869972229004, "perplexity": 32.03685622997236, "lr": 0.0026291804804649314, "grad_norm": 0.172873, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:35.756006+00:00", "epoch": 0, "step": 15745, "train_loss": 3.5451323986053467, "perplexity": 34.644271891909504, "lr": 0.0026291804804649314, "grad_norm": 0.192313, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:36.061562+00:00", "epoch": 0, "step": 15746, "train_loss": 3.537128210067749, "perplexity": 34.36807943077241, "lr": 0.0026291804804649314, "grad_norm": 0.173926, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:36.366009+00:00", "epoch": 0, "step": 15747, "train_loss": 3.547619581222534, "perplexity": 34.73054576771155, "lr": 0.0026291804804649314, "grad_norm": 0.185992, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:36.670682+00:00", "epoch": 0, "step": 15748, "train_loss": 3.5612759590148926, "perplexity": 35.20809257977257, "lr": 0.0026291804804649314, "grad_norm": 0.174183, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:36.975205+00:00", "epoch": 0, "step": 15749, "train_loss": 3.526432752609253, "perplexity": 34.00245584045426, "lr": 0.0026291804804649314, "grad_norm": 0.15443, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:37.281204+00:00", "epoch": 0, "step": 15750, "train_loss": 3.5503194332122803, "perplexity": 34.82443979375964, "lr": 0.0026291804804649314, "grad_norm": 0.158917, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:37.585993+00:00", "epoch": 0, "step": 15751, "train_loss": 3.545233964920044, "perplexity": 34.64779076162722, "lr": 0.0026291804804649314, "grad_norm": 0.16989, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:37.890433+00:00", "epoch": 0, "step": 15752, "train_loss": 3.5272293090820312, "perplexity": 34.0295515069259, "lr": 0.0026291804804649314, "grad_norm": 0.175272, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:38.194099+00:00", "epoch": 0, "step": 15753, "train_loss": 3.4803075790405273, "perplexity": 32.46970754161694, "lr": 0.0026291804804649314, "grad_norm": 0.150988, "tokens_per_sec": 107906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:38.498205+00:00", "epoch": 0, "step": 15754, "train_loss": 3.3877460956573486, "perplexity": 29.599163347913407, "lr": 0.0026291804804649314, "grad_norm": 0.162258, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:38.802190+00:00", "epoch": 0, "step": 15755, "train_loss": 3.6346383094787598, "perplexity": 37.88814663507376, "lr": 0.0026291804804649314, "grad_norm": 0.174341, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:39.107528+00:00", "epoch": 0, "step": 15756, "train_loss": 3.4738011360168457, "perplexity": 32.25913103439977, "lr": 0.0026291804804649314, "grad_norm": 0.180788, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:39.412239+00:00", "epoch": 0, "step": 15757, "train_loss": 3.592402219772339, "perplexity": 36.32122277112824, "lr": 0.0026291804804649314, "grad_norm": 0.147337, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:39.717707+00:00", "epoch": 0, "step": 15758, "train_loss": 3.428598642349243, "perplexity": 30.833403833798943, "lr": 0.0026291804804649314, "grad_norm": 0.160829, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:40.022553+00:00", "epoch": 0, "step": 15759, "train_loss": 3.4819765090942383, "perplexity": 32.523942456919904, "lr": 0.0026291804804649314, "grad_norm": 0.156987, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:40.327702+00:00", "epoch": 0, "step": 15760, "train_loss": 3.5563910007476807, "perplexity": 35.03652191579876, "lr": 0.0026291804804649314, "grad_norm": 0.152309, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:40.633235+00:00", "epoch": 0, "step": 15761, "train_loss": 3.634049654006958, "perplexity": 37.8658501333633, "lr": 0.0026291804804649314, "grad_norm": 0.152052, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:40.939269+00:00", "epoch": 0, "step": 15762, "train_loss": 3.544682025909424, "perplexity": 34.628672570797086, "lr": 0.0026291804804649314, "grad_norm": 0.166915, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:41.243266+00:00", "epoch": 0, "step": 15763, "train_loss": 3.5032200813293457, "perplexity": 33.222258277484976, "lr": 0.0026291804804649314, "grad_norm": 0.1701, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:41.548177+00:00", "epoch": 0, "step": 15764, "train_loss": 3.4911487102508545, "perplexity": 32.82363089983771, "lr": 0.0026291804804649314, "grad_norm": 0.164959, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:41.852547+00:00", "epoch": 0, "step": 15765, "train_loss": 3.499403476715088, "perplexity": 33.09570371123889, "lr": 0.0026291804804649314, "grad_norm": 0.161764, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:42.157549+00:00", "epoch": 0, "step": 15766, "train_loss": 3.446969509124756, "perplexity": 31.40507515911416, "lr": 0.0026291804804649314, "grad_norm": 0.160882, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:42.462460+00:00", "epoch": 0, "step": 15767, "train_loss": 3.5062811374664307, "perplexity": 33.324109281364144, "lr": 0.0026291804804649314, "grad_norm": 0.151096, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:42.767297+00:00", "epoch": 0, "step": 15768, "train_loss": 3.45740008354187, "perplexity": 31.734362475116072, "lr": 0.0026291804804649314, "grad_norm": 0.159477, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:43.071878+00:00", "epoch": 0, "step": 15769, "train_loss": 3.5186996459960938, "perplexity": 33.74052529804202, "lr": 0.0026291804804649314, "grad_norm": 0.169077, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:43.377411+00:00", "epoch": 0, "step": 15770, "train_loss": 3.537388563156128, "perplexity": 34.377028431293276, "lr": 0.0026291804804649314, "grad_norm": 0.16011, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:43.682301+00:00", "epoch": 0, "step": 15771, "train_loss": 3.580277681350708, "perplexity": 35.88350364351722, "lr": 0.0026291804804649314, "grad_norm": 0.172559, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:43.988044+00:00", "epoch": 0, "step": 15772, "train_loss": 3.379897117614746, "perplexity": 29.367749533736877, "lr": 0.0026291804804649314, "grad_norm": 0.172877, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:44.293370+00:00", "epoch": 0, "step": 15773, "train_loss": 3.4860618114471436, "perplexity": 32.65708437288141, "lr": 0.0026291804804649314, "grad_norm": 0.192978, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:44.597923+00:00", "epoch": 0, "step": 15774, "train_loss": 3.463698387145996, "perplexity": 31.934865877511236, "lr": 0.0026291804804649314, "grad_norm": 0.150185, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:44.901714+00:00", "epoch": 0, "step": 15775, "train_loss": 3.527271032333374, "perplexity": 34.03097136007674, "lr": 0.0026291804804649314, "grad_norm": 0.195055, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:45.206888+00:00", "epoch": 0, "step": 15776, "train_loss": 3.4778475761413574, "perplexity": 32.389930133488846, "lr": 0.0026291804804649314, "grad_norm": 0.163066, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:45.512672+00:00", "epoch": 0, "step": 15777, "train_loss": 3.371448040008545, "perplexity": 29.120664429303346, "lr": 0.0026291804804649314, "grad_norm": 0.175184, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:45.817073+00:00", "epoch": 0, "step": 15778, "train_loss": 3.4883644580841064, "perplexity": 32.73236874177078, "lr": 0.0026291804804649314, "grad_norm": 0.154978, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:46.121143+00:00", "epoch": 0, "step": 15779, "train_loss": 3.4884958267211914, "perplexity": 32.73666903089632, "lr": 0.0026291804804649314, "grad_norm": 0.156967, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:46.425886+00:00", "epoch": 0, "step": 15780, "train_loss": 3.3843798637390137, "perplexity": 29.4996932131958, "lr": 0.0026291804804649314, "grad_norm": 0.140958, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:46.730033+00:00", "epoch": 0, "step": 15781, "train_loss": 3.481523036956787, "perplexity": 32.509197098772844, "lr": 0.0026291804804649314, "grad_norm": 0.160817, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:47.035548+00:00", "epoch": 0, "step": 15782, "train_loss": 3.4951205253601074, "perplexity": 32.95425953765931, "lr": 0.0026291804804649314, "grad_norm": 0.146809, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:47.340054+00:00", "epoch": 0, "step": 15783, "train_loss": 3.4427762031555176, "perplexity": 31.273659794965766, "lr": 0.0026291804804649314, "grad_norm": 0.157725, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:47.644297+00:00", "epoch": 0, "step": 15784, "train_loss": 3.5995700359344482, "perplexity": 36.58250190046106, "lr": 0.0026291804804649314, "grad_norm": 0.160494, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:47.947654+00:00", "epoch": 0, "step": 15785, "train_loss": 3.5488126277923584, "perplexity": 34.77200565307715, "lr": 0.0026291804804649314, "grad_norm": 0.179368, "tokens_per_sec": 108017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:48.251738+00:00", "epoch": 0, "step": 15786, "train_loss": 3.554818868637085, "perplexity": 34.981483150092934, "lr": 0.0026291804804649314, "grad_norm": 0.185901, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:48.554967+00:00", "epoch": 0, "step": 15787, "train_loss": 3.4261844158172607, "perplexity": 30.759054796013764, "lr": 0.0026291804804649314, "grad_norm": 0.168382, "tokens_per_sec": 108064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:48.860434+00:00", "epoch": 0, "step": 15788, "train_loss": 3.6208672523498535, "perplexity": 37.369962960834826, "lr": 0.0026291804804649314, "grad_norm": 0.177918, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:49.164956+00:00", "epoch": 0, "step": 15789, "train_loss": 3.4889800548553467, "perplexity": 32.75252488567528, "lr": 0.0026291804804649314, "grad_norm": 0.172795, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:49.469195+00:00", "epoch": 0, "step": 15790, "train_loss": 3.6276628971099854, "perplexity": 37.6247807989291, "lr": 0.0026291804804649314, "grad_norm": 0.160993, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:49.773381+00:00", "epoch": 0, "step": 15791, "train_loss": 3.549222707748413, "perplexity": 34.78626787975408, "lr": 0.0026291804804649314, "grad_norm": 0.175565, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:50.078852+00:00", "epoch": 0, "step": 15792, "train_loss": 3.5866587162017822, "perplexity": 36.11320963238275, "lr": 0.0026291804804649314, "grad_norm": 0.218967, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:50.384793+00:00", "epoch": 0, "step": 15793, "train_loss": 3.6216962337493896, "perplexity": 37.40095480908875, "lr": 0.0026291804804649314, "grad_norm": 0.185162, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:50.689711+00:00", "epoch": 0, "step": 15794, "train_loss": 3.4321911334991455, "perplexity": 30.944371770512593, "lr": 0.0026291804804649314, "grad_norm": 0.213156, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:50.994166+00:00", "epoch": 0, "step": 15795, "train_loss": 3.540632486343384, "perplexity": 34.488725942053314, "lr": 0.0026291804804649314, "grad_norm": 0.204799, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:51.298486+00:00", "epoch": 0, "step": 15796, "train_loss": 3.404944896697998, "perplexity": 30.112636372504213, "lr": 0.0026291804804649314, "grad_norm": 0.214648, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:51.604132+00:00", "epoch": 0, "step": 15797, "train_loss": 3.5410239696502686, "perplexity": 34.50223034573719, "lr": 0.0026291804804649314, "grad_norm": 0.199279, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:51.909569+00:00", "epoch": 0, "step": 15798, "train_loss": 3.5372824668884277, "perplexity": 34.373381350356354, "lr": 0.0026291804804649314, "grad_norm": 0.154353, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:52.215241+00:00", "epoch": 0, "step": 15799, "train_loss": 3.47293758392334, "perplexity": 32.23128561897414, "lr": 0.0026291804804649314, "grad_norm": 0.190927, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:52.519218+00:00", "epoch": 0, "step": 15800, "train_loss": 3.532426118850708, "perplexity": 34.206856924846775, "lr": 0.0026291804804649314, "grad_norm": 0.244296, "tokens_per_sec": 107856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:52.823327+00:00", "epoch": 0, "step": 15801, "train_loss": 3.4494826793670654, "perplexity": 31.48410072021186, "lr": 0.0026291804804649314, "grad_norm": 0.173118, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:53.127144+00:00", "epoch": 0, "step": 15802, "train_loss": 3.5070242881774902, "perplexity": 33.348883321154496, "lr": 0.0026291804804649314, "grad_norm": 0.198648, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:53.431480+00:00", "epoch": 0, "step": 15803, "train_loss": 3.5142581462860107, "perplexity": 33.59099907086762, "lr": 0.0026291804804649314, "grad_norm": 0.194888, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:53.737345+00:00", "epoch": 0, "step": 15804, "train_loss": 3.416078805923462, "perplexity": 30.44978111910876, "lr": 0.0026291804804649314, "grad_norm": 0.167659, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:54.042616+00:00", "epoch": 0, "step": 15805, "train_loss": 3.662346601486206, "perplexity": 38.95264203575256, "lr": 0.0026291804804649314, "grad_norm": 0.190621, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:54.347114+00:00", "epoch": 0, "step": 15806, "train_loss": 3.5367820262908936, "perplexity": 34.35618381838461, "lr": 0.0026291804804649314, "grad_norm": 0.162404, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:54.651071+00:00", "epoch": 0, "step": 15807, "train_loss": 3.5645251274108887, "perplexity": 35.322675650506795, "lr": 0.0026291804804649314, "grad_norm": 0.168251, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:54.955087+00:00", "epoch": 0, "step": 15808, "train_loss": 3.4457147121429443, "perplexity": 31.365692879141296, "lr": 0.0026291804804649314, "grad_norm": 0.159688, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:55.260374+00:00", "epoch": 0, "step": 15809, "train_loss": 3.529921293258667, "perplexity": 34.12128193415403, "lr": 0.0026291804804649314, "grad_norm": 0.174456, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:55.566242+00:00", "epoch": 0, "step": 15810, "train_loss": 3.6098408699035645, "perplexity": 36.96017087128278, "lr": 0.0026291804804649314, "grad_norm": 0.184745, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:55.870020+00:00", "epoch": 0, "step": 15811, "train_loss": 3.585916042327881, "perplexity": 36.08639925200289, "lr": 0.0026291804804649314, "grad_norm": 0.169571, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:56.173541+00:00", "epoch": 0, "step": 15812, "train_loss": 3.501009702682495, "perplexity": 33.148905605643414, "lr": 0.0026291804804649314, "grad_norm": 0.159018, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:56.478428+00:00", "epoch": 0, "step": 15813, "train_loss": 3.4564919471740723, "perplexity": 31.705556428332613, "lr": 0.0026291804804649314, "grad_norm": 0.166654, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:56.783380+00:00", "epoch": 0, "step": 15814, "train_loss": 3.508519172668457, "perplexity": 33.398773330232075, "lr": 0.0026291804804649314, "grad_norm": 0.168281, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:57.088348+00:00", "epoch": 0, "step": 15815, "train_loss": 3.532223701477051, "perplexity": 34.19993356343493, "lr": 0.0026291804804649314, "grad_norm": 0.174335, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:57.393757+00:00", "epoch": 0, "step": 15816, "train_loss": 3.5704164505004883, "perplexity": 35.531387134889165, "lr": 0.0026291804804649314, "grad_norm": 0.151538, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:57.698424+00:00", "epoch": 0, "step": 15817, "train_loss": 3.48315691947937, "perplexity": 32.56235672420107, "lr": 0.0026291804804649314, "grad_norm": 0.169043, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:58.002593+00:00", "epoch": 0, "step": 15818, "train_loss": 3.5378305912017822, "perplexity": 34.39222740092411, "lr": 0.0026291804804649314, "grad_norm": 0.158757, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:58.307818+00:00", "epoch": 0, "step": 15819, "train_loss": 3.581540584564209, "perplexity": 35.928849663370436, "lr": 0.0026291804804649314, "grad_norm": 0.177514, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:58.613734+00:00", "epoch": 0, "step": 15820, "train_loss": 3.623189687728882, "perplexity": 37.45685314429111, "lr": 0.0026291804804649314, "grad_norm": 0.181261, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:58.918426+00:00", "epoch": 0, "step": 15821, "train_loss": 3.417762041091919, "perplexity": 30.501078422160898, "lr": 0.0026291804804649314, "grad_norm": 0.196939, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:59.222565+00:00", "epoch": 0, "step": 15822, "train_loss": 3.45501708984375, "perplexity": 31.65882972210495, "lr": 0.0026291804804649314, "grad_norm": 0.17903, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:59.526050+00:00", "epoch": 0, "step": 15823, "train_loss": 3.4792139530181885, "perplexity": 32.43421723461059, "lr": 0.0026291804804649314, "grad_norm": 0.189802, "tokens_per_sec": 107967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:20:59.830999+00:00", "epoch": 0, "step": 15824, "train_loss": 3.439488410949707, "perplexity": 31.171007342480422, "lr": 0.0026291804804649314, "grad_norm": 0.17535, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:00.135600+00:00", "epoch": 0, "step": 15825, "train_loss": 3.563715696334839, "perplexity": 35.29409594735638, "lr": 0.0026291804804649314, "grad_norm": 0.200357, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:00.441165+00:00", "epoch": 0, "step": 15826, "train_loss": 3.412144184112549, "perplexity": 30.330208137671914, "lr": 0.0026291804804649314, "grad_norm": 0.152671, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:00.745240+00:00", "epoch": 0, "step": 15827, "train_loss": 3.574769973754883, "perplexity": 35.68641106032456, "lr": 0.0026291804804649314, "grad_norm": 0.189707, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:01.050112+00:00", "epoch": 0, "step": 15828, "train_loss": 3.496896505355835, "perplexity": 33.01283764474861, "lr": 0.0026291804804649314, "grad_norm": 0.1575, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:01.354266+00:00", "epoch": 0, "step": 15829, "train_loss": 3.5365986824035645, "perplexity": 34.3498853994955, "lr": 0.0026291804804649314, "grad_norm": 0.151051, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:01.659254+00:00", "epoch": 0, "step": 15830, "train_loss": 3.5912880897521973, "perplexity": 36.2807787406085, "lr": 0.0026291804804649314, "grad_norm": 0.1545, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:01.964512+00:00", "epoch": 0, "step": 15831, "train_loss": 3.586501359939575, "perplexity": 36.10752743977472, "lr": 0.0026291804804649314, "grad_norm": 0.156048, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:02.269229+00:00", "epoch": 0, "step": 15832, "train_loss": 3.530090570449829, "perplexity": 34.127058377814976, "lr": 0.0026291804804649314, "grad_norm": 0.144577, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:02.573155+00:00", "epoch": 0, "step": 15833, "train_loss": 3.519217014312744, "perplexity": 33.75798609326001, "lr": 0.0026291804804649314, "grad_norm": 0.167372, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:02.877968+00:00", "epoch": 0, "step": 15834, "train_loss": 3.508049726486206, "perplexity": 33.38309808323064, "lr": 0.0026291804804649314, "grad_norm": 0.170686, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:03.182489+00:00", "epoch": 0, "step": 15835, "train_loss": 3.4892144203186035, "perplexity": 32.76020184591497, "lr": 0.0026291804804649314, "grad_norm": 0.159833, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:03.488941+00:00", "epoch": 0, "step": 15836, "train_loss": 3.4622442722320557, "perplexity": 31.88846265874673, "lr": 0.0026291804804649314, "grad_norm": 0.156292, "tokens_per_sec": 106928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:03.793603+00:00", "epoch": 0, "step": 15837, "train_loss": 3.498270273208618, "perplexity": 33.058220785646725, "lr": 0.0026291804804649314, "grad_norm": 0.157253, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:04.097601+00:00", "epoch": 0, "step": 15838, "train_loss": 3.501567840576172, "perplexity": 33.16741243019544, "lr": 0.0026291804804649314, "grad_norm": 0.163879, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:04.401992+00:00", "epoch": 0, "step": 15839, "train_loss": 3.4764280319213867, "perplexity": 32.34398381451392, "lr": 0.0026291804804649314, "grad_norm": 0.181383, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:04.706173+00:00", "epoch": 0, "step": 15840, "train_loss": 3.5090131759643555, "perplexity": 33.415276510313234, "lr": 0.0026291804804649314, "grad_norm": 0.14611, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:05.010996+00:00", "epoch": 0, "step": 15841, "train_loss": 3.4597227573394775, "perplexity": 31.808156714117963, "lr": 0.0026291804804649314, "grad_norm": 0.172151, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:05.316686+00:00", "epoch": 0, "step": 15842, "train_loss": 3.4184112548828125, "perplexity": 30.520886572076012, "lr": 0.0026291804804649314, "grad_norm": 0.172834, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:05.620492+00:00", "epoch": 0, "step": 15843, "train_loss": 3.437079906463623, "perplexity": 31.09602216877235, "lr": 0.0026291804804649314, "grad_norm": 0.160396, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:05.925342+00:00", "epoch": 0, "step": 15844, "train_loss": 3.452744960784912, "perplexity": 31.586978433736107, "lr": 0.0026291804804649314, "grad_norm": 0.207173, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:06.230350+00:00", "epoch": 0, "step": 15845, "train_loss": 3.5460448265075684, "perplexity": 34.675896717728534, "lr": 0.0026291804804649314, "grad_norm": 0.190575, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:06.534962+00:00", "epoch": 0, "step": 15846, "train_loss": 3.6611626148223877, "perplexity": 38.90654991867213, "lr": 0.0026291804804649314, "grad_norm": 0.197437, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:06.839710+00:00", "epoch": 0, "step": 15847, "train_loss": 3.532546281814575, "perplexity": 34.210967569128094, "lr": 0.0026291804804649314, "grad_norm": 0.171419, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:07.144516+00:00", "epoch": 0, "step": 15848, "train_loss": 3.4217090606689453, "perplexity": 30.62170467630426, "lr": 0.0026291804804649314, "grad_norm": 0.172107, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:07.450298+00:00", "epoch": 0, "step": 15849, "train_loss": 3.5011658668518066, "perplexity": 33.15408268117723, "lr": 0.0026291804804649314, "grad_norm": 0.190022, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:07.755182+00:00", "epoch": 0, "step": 15850, "train_loss": 3.5173332691192627, "perplexity": 33.69445450666245, "lr": 0.0026291804804649314, "grad_norm": 0.175038, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:08.060899+00:00", "epoch": 0, "step": 15851, "train_loss": 3.526400566101074, "perplexity": 34.001361437743846, "lr": 0.0026291804804649314, "grad_norm": 0.158741, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:08.366554+00:00", "epoch": 0, "step": 15852, "train_loss": 3.509279727935791, "perplexity": 33.42418460532552, "lr": 0.0026291804804649314, "grad_norm": 0.165568, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:08.672471+00:00", "epoch": 0, "step": 15853, "train_loss": 3.511718273162842, "perplexity": 33.505790450495915, "lr": 0.0026291804804649314, "grad_norm": 0.188181, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:08.977960+00:00", "epoch": 0, "step": 15854, "train_loss": 3.5111961364746094, "perplexity": 33.48830041452556, "lr": 0.0026291804804649314, "grad_norm": 0.201692, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:09.282504+00:00", "epoch": 0, "step": 15855, "train_loss": 3.4429049491882324, "perplexity": 31.27768641379288, "lr": 0.0026291804804649314, "grad_norm": 0.144274, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:09.587108+00:00", "epoch": 0, "step": 15856, "train_loss": 3.4775969982147217, "perplexity": 32.38181494873766, "lr": 0.0026291804804649314, "grad_norm": 0.178604, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:09.892731+00:00", "epoch": 0, "step": 15857, "train_loss": 3.495605945587158, "perplexity": 32.97026008498685, "lr": 0.0026291804804649314, "grad_norm": 0.16208, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:10.197982+00:00", "epoch": 0, "step": 15858, "train_loss": 3.50555157661438, "perplexity": 33.299806182175836, "lr": 0.0026291804804649314, "grad_norm": 0.191529, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:10.502174+00:00", "epoch": 0, "step": 15859, "train_loss": 3.443232297897339, "perplexity": 31.287926800061438, "lr": 0.0026291804804649314, "grad_norm": 0.192635, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:10.806845+00:00", "epoch": 0, "step": 15860, "train_loss": 3.405125617980957, "perplexity": 30.118078858554444, "lr": 0.0026291804804649314, "grad_norm": 0.160846, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:11.110377+00:00", "epoch": 0, "step": 15861, "train_loss": 3.588822364807129, "perplexity": 36.19143051877922, "lr": 0.0026291804804649314, "grad_norm": 0.164067, "tokens_per_sec": 107895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:11.414627+00:00", "epoch": 0, "step": 15862, "train_loss": 3.397958755493164, "perplexity": 29.90299837571474, "lr": 0.0026291804804649314, "grad_norm": 0.199618, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:11.719783+00:00", "epoch": 0, "step": 15863, "train_loss": 3.4757795333862305, "perplexity": 32.32301558805601, "lr": 0.0026291804804649314, "grad_norm": 0.162367, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:12.025410+00:00", "epoch": 0, "step": 15864, "train_loss": 3.5037317276000977, "perplexity": 33.23926067127339, "lr": 0.0026291804804649314, "grad_norm": 0.167657, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:12.330433+00:00", "epoch": 0, "step": 15865, "train_loss": 3.530125617980957, "perplexity": 34.12825446791565, "lr": 0.0026291804804649314, "grad_norm": 0.165409, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:12.634083+00:00", "epoch": 0, "step": 15866, "train_loss": 3.593432664871216, "perplexity": 36.358669086990865, "lr": 0.0026291804804649314, "grad_norm": 0.16373, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:12.938075+00:00", "epoch": 0, "step": 15867, "train_loss": 3.45194149017334, "perplexity": 31.561609417850367, "lr": 0.0026291804804649314, "grad_norm": 0.182509, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:13.243721+00:00", "epoch": 0, "step": 15868, "train_loss": 3.4920847415924072, "perplexity": 32.85436923087985, "lr": 0.0026291804804649314, "grad_norm": 0.183471, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:13.548856+00:00", "epoch": 0, "step": 15869, "train_loss": 3.5181021690368652, "perplexity": 33.72037213270966, "lr": 0.0026291804804649314, "grad_norm": 0.190365, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:13.854202+00:00", "epoch": 0, "step": 15870, "train_loss": 3.5695767402648926, "perplexity": 35.50156358874798, "lr": 0.0026291804804649314, "grad_norm": 0.170366, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:14.158731+00:00", "epoch": 0, "step": 15871, "train_loss": 3.5859975814819336, "perplexity": 36.089341826436595, "lr": 0.0026291804804649314, "grad_norm": 0.15892, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:14.462919+00:00", "epoch": 0, "step": 15872, "train_loss": 3.518824338912964, "perplexity": 33.74473276487337, "lr": 0.0026291804804649314, "grad_norm": 0.16892, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:14.768441+00:00", "epoch": 0, "step": 15873, "train_loss": 3.5690908432006836, "perplexity": 35.4843176734345, "lr": 0.0026291804804649314, "grad_norm": 0.17856, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:15.074048+00:00", "epoch": 0, "step": 15874, "train_loss": 3.5200624465942383, "perplexity": 33.78653825221927, "lr": 0.0026291804804649314, "grad_norm": 0.190467, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:15.378877+00:00", "epoch": 0, "step": 15875, "train_loss": 3.549192190170288, "perplexity": 34.78520630330483, "lr": 0.0026291804804649314, "grad_norm": 0.18165, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:15.684600+00:00", "epoch": 0, "step": 15876, "train_loss": 3.543642520904541, "perplexity": 34.59269459522986, "lr": 0.0026291804804649314, "grad_norm": 0.168901, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:15.990207+00:00", "epoch": 0, "step": 15877, "train_loss": 3.523874044418335, "perplexity": 33.91556469013924, "lr": 0.0026291804804649314, "grad_norm": 0.174392, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:16.296467+00:00", "epoch": 0, "step": 15878, "train_loss": 3.4521827697753906, "perplexity": 31.56922550917758, "lr": 0.0026291804804649314, "grad_norm": 0.160424, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:16.601922+00:00", "epoch": 0, "step": 15879, "train_loss": 3.489065408706665, "perplexity": 32.755320559123724, "lr": 0.0026291804804649314, "grad_norm": 0.149725, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:16.907439+00:00", "epoch": 0, "step": 15880, "train_loss": 3.5272889137268066, "perplexity": 34.03157988670516, "lr": 0.0026291804804649314, "grad_norm": 0.164475, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:17.211981+00:00", "epoch": 0, "step": 15881, "train_loss": 3.480541229248047, "perplexity": 32.47729498189112, "lr": 0.0026291804804649314, "grad_norm": 0.169137, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:17.517293+00:00", "epoch": 0, "step": 15882, "train_loss": 3.5006046295166016, "perplexity": 33.13548059274363, "lr": 0.0026291804804649314, "grad_norm": 0.16368, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:17.823104+00:00", "epoch": 0, "step": 15883, "train_loss": 3.6326050758361816, "perplexity": 37.8111894431682, "lr": 0.0026291804804649314, "grad_norm": 0.188617, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:18.128018+00:00", "epoch": 0, "step": 15884, "train_loss": 3.53920578956604, "perplexity": 34.43955607148533, "lr": 0.0026291804804649314, "grad_norm": 0.177219, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:18.433307+00:00", "epoch": 0, "step": 15885, "train_loss": 3.537046194076538, "perplexity": 34.36526081425927, "lr": 0.0026291804804649314, "grad_norm": 0.188678, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:18.738921+00:00", "epoch": 0, "step": 15886, "train_loss": 3.6415820121765137, "perplexity": 38.15214616748188, "lr": 0.0026291804804649314, "grad_norm": 0.194194, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:19.044101+00:00", "epoch": 0, "step": 15887, "train_loss": 3.475362777709961, "perplexity": 32.30954759446186, "lr": 0.0026291804804649314, "grad_norm": 0.171778, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:19.350543+00:00", "epoch": 0, "step": 15888, "train_loss": 3.377732515335083, "perplexity": 29.304248787871614, "lr": 0.0026291804804649314, "grad_norm": 0.161686, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:19.655689+00:00", "epoch": 0, "step": 15889, "train_loss": 3.517843723297119, "perplexity": 33.71165837225495, "lr": 0.0026291804804649314, "grad_norm": 0.183672, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:19.960519+00:00", "epoch": 0, "step": 15890, "train_loss": 3.5779716968536377, "perplexity": 35.800852173544286, "lr": 0.0026291804804649314, "grad_norm": 0.173864, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:20.265746+00:00", "epoch": 0, "step": 15891, "train_loss": 3.5002894401550293, "perplexity": 33.125038287508424, "lr": 0.0026291804804649314, "grad_norm": 0.172785, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:20.572005+00:00", "epoch": 0, "step": 15892, "train_loss": 3.577147960662842, "perplexity": 35.77137385879181, "lr": 0.0026291804804649314, "grad_norm": 0.165099, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:20.877669+00:00", "epoch": 0, "step": 15893, "train_loss": 3.5117557048797607, "perplexity": 33.50704465323253, "lr": 0.0026291804804649314, "grad_norm": 0.191107, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:21.182690+00:00", "epoch": 0, "step": 15894, "train_loss": 3.4400458335876465, "perplexity": 31.188387611248096, "lr": 0.0026291804804649314, "grad_norm": 0.171267, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:21.487139+00:00", "epoch": 0, "step": 15895, "train_loss": 3.5372471809387207, "perplexity": 34.37216847434963, "lr": 0.0026291804804649314, "grad_norm": 0.178108, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:21.792039+00:00", "epoch": 0, "step": 15896, "train_loss": 3.4421448707580566, "perplexity": 31.25392195157562, "lr": 0.0026291804804649314, "grad_norm": 0.20689, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:22.098477+00:00", "epoch": 0, "step": 15897, "train_loss": 3.443042278289795, "perplexity": 31.281982045317896, "lr": 0.0026291804804649314, "grad_norm": 0.147167, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:22.404071+00:00", "epoch": 0, "step": 15898, "train_loss": 3.597562074661255, "perplexity": 36.509119352681424, "lr": 0.0026291804804649314, "grad_norm": 0.183752, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:22.708479+00:00", "epoch": 0, "step": 15899, "train_loss": 3.454108238220215, "perplexity": 31.63006961462766, "lr": 0.0026291804804649314, "grad_norm": 0.142118, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:23.013054+00:00", "epoch": 0, "step": 15900, "train_loss": 3.590430974960327, "perplexity": 36.2496952714429, "lr": 0.0026291804804649314, "grad_norm": 0.174381, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:23.318418+00:00", "epoch": 0, "step": 15901, "train_loss": 3.536781072616577, "perplexity": 34.356151053790114, "lr": 0.0026291804804649314, "grad_norm": 0.204117, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:23.623311+00:00", "epoch": 0, "step": 15902, "train_loss": 3.506805896759033, "perplexity": 33.34160100644849, "lr": 0.0026291804804649314, "grad_norm": 0.167862, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:23.928938+00:00", "epoch": 0, "step": 15903, "train_loss": 3.5224361419677734, "perplexity": 33.866832461051345, "lr": 0.0026291804804649314, "grad_norm": 0.166613, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:24.233824+00:00", "epoch": 0, "step": 15904, "train_loss": 3.6945245265960693, "perplexity": 40.22644143711723, "lr": 0.0026291804804649314, "grad_norm": 0.17823, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:24.537694+00:00", "epoch": 0, "step": 15905, "train_loss": 3.479297637939453, "perplexity": 32.43693160310042, "lr": 0.0026291804804649314, "grad_norm": 0.20523, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:24.843406+00:00", "epoch": 0, "step": 15906, "train_loss": 3.478971481323242, "perplexity": 32.42635380835113, "lr": 0.0026291804804649314, "grad_norm": 0.197887, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:25.148252+00:00", "epoch": 0, "step": 15907, "train_loss": 3.4247562885284424, "perplexity": 30.715158302831963, "lr": 0.0026291804804649314, "grad_norm": 0.165354, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:25.452644+00:00", "epoch": 0, "step": 15908, "train_loss": 3.42853045463562, "perplexity": 30.83130144616762, "lr": 0.0026291804804649314, "grad_norm": 0.208889, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:25.757125+00:00", "epoch": 0, "step": 15909, "train_loss": 3.4937875270843506, "perplexity": 32.910360831466846, "lr": 0.0026291804804649314, "grad_norm": 0.175357, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:26.061977+00:00", "epoch": 0, "step": 15910, "train_loss": 3.4461166858673096, "perplexity": 31.378303597944214, "lr": 0.0026291804804649314, "grad_norm": 0.172072, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:26.367329+00:00", "epoch": 0, "step": 15911, "train_loss": 3.4855151176452637, "perplexity": 32.63923582655665, "lr": 0.0026291804804649314, "grad_norm": 0.193688, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:26.671856+00:00", "epoch": 0, "step": 15912, "train_loss": 3.4685919284820557, "perplexity": 32.09152345629322, "lr": 0.0026291804804649314, "grad_norm": 0.156968, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:26.976642+00:00", "epoch": 0, "step": 15913, "train_loss": 3.3816585540771484, "perplexity": 29.419524544393425, "lr": 0.0026291804804649314, "grad_norm": 0.18305, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:27.281569+00:00", "epoch": 0, "step": 15914, "train_loss": 3.5303757190704346, "perplexity": 34.13679104899967, "lr": 0.0026291804804649314, "grad_norm": 0.189397, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:27.586112+00:00", "epoch": 0, "step": 15915, "train_loss": 3.5047707557678223, "perplexity": 33.273815147813714, "lr": 0.0026291804804649314, "grad_norm": 0.1624, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:27.890313+00:00", "epoch": 0, "step": 15916, "train_loss": 3.502613067626953, "perplexity": 33.20209803087812, "lr": 0.0026291804804649314, "grad_norm": 0.204387, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:28.196038+00:00", "epoch": 0, "step": 15917, "train_loss": 3.3563578128814697, "perplexity": 28.684525979347367, "lr": 0.0026291804804649314, "grad_norm": 0.202899, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:28.501071+00:00", "epoch": 0, "step": 15918, "train_loss": 3.5930328369140625, "perplexity": 36.34413478040968, "lr": 0.0026291804804649314, "grad_norm": 0.176019, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:28.806368+00:00", "epoch": 0, "step": 15919, "train_loss": 3.5101122856140137, "perplexity": 33.45202375409749, "lr": 0.0026291804804649314, "grad_norm": 0.182377, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:29.112488+00:00", "epoch": 0, "step": 15920, "train_loss": 3.586301565170288, "perplexity": 36.1003140652816, "lr": 0.0026291804804649314, "grad_norm": 0.159742, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:29.417211+00:00", "epoch": 0, "step": 15921, "train_loss": 3.4552841186523438, "perplexity": 31.667284670494276, "lr": 0.0026291804804649314, "grad_norm": 0.163612, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:29.722987+00:00", "epoch": 0, "step": 15922, "train_loss": 3.487013339996338, "perplexity": 32.688173309655426, "lr": 0.0026291804804649314, "grad_norm": 0.156164, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:30.028649+00:00", "epoch": 0, "step": 15923, "train_loss": 3.4880409240722656, "perplexity": 32.72178042012815, "lr": 0.0026291804804649314, "grad_norm": 0.158041, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:30.333763+00:00", "epoch": 0, "step": 15924, "train_loss": 3.4583332538604736, "perplexity": 31.763989861806618, "lr": 0.0026291804804649314, "grad_norm": 0.149726, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:30.638423+00:00", "epoch": 0, "step": 15925, "train_loss": 3.473574161529541, "perplexity": 32.25180986556305, "lr": 0.0026291804804649314, "grad_norm": 0.169926, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:30.944078+00:00", "epoch": 0, "step": 15926, "train_loss": 3.48821759223938, "perplexity": 32.727561827779994, "lr": 0.0026291804804649314, "grad_norm": 0.165902, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:31.249236+00:00", "epoch": 0, "step": 15927, "train_loss": 3.538064956665039, "perplexity": 34.400288695839016, "lr": 0.0026291804804649314, "grad_norm": 0.16444, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:31.554676+00:00", "epoch": 0, "step": 15928, "train_loss": 3.4711661338806152, "perplexity": 32.17424004829424, "lr": 0.0026291804804649314, "grad_norm": 0.171091, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:31.860676+00:00", "epoch": 0, "step": 15929, "train_loss": 3.518989324569702, "perplexity": 33.75030062106561, "lr": 0.0026291804804649314, "grad_norm": 0.168869, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:32.165739+00:00", "epoch": 0, "step": 15930, "train_loss": 3.4829118251800537, "perplexity": 32.554376854145936, "lr": 0.0026291804804649314, "grad_norm": 0.180357, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:32.471379+00:00", "epoch": 0, "step": 15931, "train_loss": 3.393843650817871, "perplexity": 29.780197250329852, "lr": 0.0026291804804649314, "grad_norm": 0.161481, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:32.775916+00:00", "epoch": 0, "step": 15932, "train_loss": 3.4938721656799316, "perplexity": 32.91314643607085, "lr": 0.0026291804804649314, "grad_norm": 0.20071, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:33.080426+00:00", "epoch": 0, "step": 15933, "train_loss": 3.451890468597412, "perplexity": 31.55999913587896, "lr": 0.0026291804804649314, "grad_norm": 0.168983, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:33.384886+00:00", "epoch": 0, "step": 15934, "train_loss": 3.7008869647979736, "perplexity": 40.48319560999232, "lr": 0.0026291804804649314, "grad_norm": 0.165618, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:33.690275+00:00", "epoch": 0, "step": 15935, "train_loss": 3.68882417678833, "perplexity": 39.99778896808628, "lr": 0.0026291804804649314, "grad_norm": 0.174361, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:33.994805+00:00", "epoch": 0, "step": 15936, "train_loss": 3.5208728313446045, "perplexity": 33.81392944478988, "lr": 0.0026291804804649314, "grad_norm": 0.152602, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:34.299630+00:00", "epoch": 0, "step": 15937, "train_loss": 3.5471506118774414, "perplexity": 34.714262024994675, "lr": 0.0026291804804649314, "grad_norm": 0.173346, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:34.605225+00:00", "epoch": 0, "step": 15938, "train_loss": 3.5773887634277344, "perplexity": 35.7799887417232, "lr": 0.0026291804804649314, "grad_norm": 0.168452, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:34.910419+00:00", "epoch": 0, "step": 15939, "train_loss": 3.5204527378082275, "perplexity": 33.79972741488996, "lr": 0.0026291804804649314, "grad_norm": 0.155864, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:35.215244+00:00", "epoch": 0, "step": 15940, "train_loss": 3.5630879402160645, "perplexity": 35.27194681552271, "lr": 0.0026291804804649314, "grad_norm": 0.176426, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:35.520583+00:00", "epoch": 0, "step": 15941, "train_loss": 3.531801462173462, "perplexity": 34.185496055560456, "lr": 0.0026291804804649314, "grad_norm": 0.154916, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:35.825021+00:00", "epoch": 0, "step": 15942, "train_loss": 3.4829154014587402, "perplexity": 32.55449327787821, "lr": 0.0026291804804649314, "grad_norm": 0.177649, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:36.128759+00:00", "epoch": 0, "step": 15943, "train_loss": 3.4900078773498535, "perplexity": 32.78620597362948, "lr": 0.0026291804804649314, "grad_norm": 0.191267, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:36.433830+00:00", "epoch": 0, "step": 15944, "train_loss": 3.5132853984832764, "perplexity": 33.558339387722604, "lr": 0.0026291804804649314, "grad_norm": 0.188334, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:36.737991+00:00", "epoch": 0, "step": 15945, "train_loss": 3.5482845306396484, "perplexity": 34.753647503767596, "lr": 0.0026291804804649314, "grad_norm": 0.189837, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:37.042629+00:00", "epoch": 0, "step": 15946, "train_loss": 3.4642367362976074, "perplexity": 31.952062613972128, "lr": 0.0026291804804649314, "grad_norm": 0.170989, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:37.347832+00:00", "epoch": 0, "step": 15947, "train_loss": 3.55071759223938, "perplexity": 34.83830821956346, "lr": 0.0026291804804649314, "grad_norm": 0.16581, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:37.652288+00:00", "epoch": 0, "step": 15948, "train_loss": 3.5951898097991943, "perplexity": 36.42261270061083, "lr": 0.0026291804804649314, "grad_norm": 0.155878, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:37.957566+00:00", "epoch": 0, "step": 15949, "train_loss": 3.5006344318389893, "perplexity": 33.136468121733984, "lr": 0.0026291804804649314, "grad_norm": 0.14493, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:38.263526+00:00", "epoch": 0, "step": 15950, "train_loss": 3.5356128215789795, "perplexity": 34.31603788036414, "lr": 0.0026291804804649314, "grad_norm": 0.16353, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:38.567195+00:00", "epoch": 0, "step": 15951, "train_loss": 3.524892807006836, "perplexity": 33.95013420468759, "lr": 0.0026291804804649314, "grad_norm": 0.173261, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:38.872060+00:00", "epoch": 0, "step": 15952, "train_loss": 3.484969139099121, "perplexity": 32.62142036790452, "lr": 0.0026291804804649314, "grad_norm": 0.15756, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:39.177234+00:00", "epoch": 0, "step": 15953, "train_loss": 3.5037484169006348, "perplexity": 33.2398154159135, "lr": 0.0026291804804649314, "grad_norm": 0.167731, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:39.483008+00:00", "epoch": 0, "step": 15954, "train_loss": 3.529646158218384, "perplexity": 34.11189526523454, "lr": 0.0026291804804649314, "grad_norm": 0.147377, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:39.788836+00:00", "epoch": 0, "step": 15955, "train_loss": 3.620574951171875, "perplexity": 37.35904127292924, "lr": 0.0026291804804649314, "grad_norm": 0.167255, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:40.094569+00:00", "epoch": 0, "step": 15956, "train_loss": 3.4852380752563477, "perplexity": 32.63019462714644, "lr": 0.0026291804804649314, "grad_norm": 0.153501, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:40.399502+00:00", "epoch": 0, "step": 15957, "train_loss": 3.518047571182251, "perplexity": 33.718531122992474, "lr": 0.0026291804804649314, "grad_norm": 0.160271, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:40.704511+00:00", "epoch": 0, "step": 15958, "train_loss": 3.4450929164886475, "perplexity": 31.34619588981483, "lr": 0.0026291804804649314, "grad_norm": 0.188633, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:41.009772+00:00", "epoch": 0, "step": 15959, "train_loss": 3.517167329788208, "perplexity": 33.688863735298625, "lr": 0.0026291804804649314, "grad_norm": 0.21056, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:41.315423+00:00", "epoch": 0, "step": 15960, "train_loss": 3.6130497455596924, "perplexity": 37.078961954814545, "lr": 0.0026291804804649314, "grad_norm": 0.178088, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:41.621181+00:00", "epoch": 0, "step": 15961, "train_loss": 3.546099901199341, "perplexity": 34.677806534643004, "lr": 0.0026291804804649314, "grad_norm": 0.163243, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:41.926190+00:00", "epoch": 0, "step": 15962, "train_loss": 3.477776288986206, "perplexity": 32.38762122981264, "lr": 0.0026291804804649314, "grad_norm": 0.173437, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:42.231105+00:00", "epoch": 0, "step": 15963, "train_loss": 3.425098180770874, "perplexity": 30.72566137253748, "lr": 0.0026291804804649314, "grad_norm": 0.164474, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:42.535750+00:00", "epoch": 0, "step": 15964, "train_loss": 3.4967379570007324, "perplexity": 33.00760392855228, "lr": 0.0026291804804649314, "grad_norm": 0.184881, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:42.840270+00:00", "epoch": 0, "step": 15965, "train_loss": 3.545478343963623, "perplexity": 34.656258990282346, "lr": 0.0026291804804649314, "grad_norm": 0.188578, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:43.145176+00:00", "epoch": 0, "step": 15966, "train_loss": 3.563837766647339, "perplexity": 35.29840457165029, "lr": 0.0026291804804649314, "grad_norm": 0.202162, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:43.450787+00:00", "epoch": 0, "step": 15967, "train_loss": 3.4434053897857666, "perplexity": 31.29334295512898, "lr": 0.0026291804804649314, "grad_norm": 0.218539, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:43.755762+00:00", "epoch": 0, "step": 15968, "train_loss": 3.507096767425537, "perplexity": 33.351300510737794, "lr": 0.0026291804804649314, "grad_norm": 0.227008, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:44.059927+00:00", "epoch": 0, "step": 15969, "train_loss": 3.5488226413726807, "perplexity": 34.772353847092056, "lr": 0.0026291804804649314, "grad_norm": 0.205349, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:44.364897+00:00", "epoch": 0, "step": 15970, "train_loss": 3.5066325664520264, "perplexity": 33.33582239732846, "lr": 0.0026291804804649314, "grad_norm": 0.187558, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:44.670398+00:00", "epoch": 0, "step": 15971, "train_loss": 3.5042483806610107, "perplexity": 33.25643827409013, "lr": 0.0026291804804649314, "grad_norm": 0.178412, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:44.974898+00:00", "epoch": 0, "step": 15972, "train_loss": 3.5415570735931396, "perplexity": 34.520628524407414, "lr": 0.0026291804804649314, "grad_norm": 0.220307, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:45.279039+00:00", "epoch": 0, "step": 15973, "train_loss": 3.3801610469818115, "perplexity": 29.37550156823323, "lr": 0.0026291804804649314, "grad_norm": 0.174578, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:45.582893+00:00", "epoch": 0, "step": 15974, "train_loss": 3.5740160942077637, "perplexity": 35.65951794327524, "lr": 0.0026291804804649314, "grad_norm": 0.171935, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:45.886992+00:00", "epoch": 0, "step": 15975, "train_loss": 3.4816577434539795, "perplexity": 32.51357659380716, "lr": 0.0026291804804649314, "grad_norm": 0.153512, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:46.191693+00:00", "epoch": 0, "step": 15976, "train_loss": 3.399925470352173, "perplexity": 29.961866916791276, "lr": 0.0026291804804649314, "grad_norm": 0.161835, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:46.497582+00:00", "epoch": 0, "step": 15977, "train_loss": 3.547457218170166, "perplexity": 34.72490726804468, "lr": 0.0026291804804649314, "grad_norm": 0.143399, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:46.801828+00:00", "epoch": 0, "step": 15978, "train_loss": 3.5469086170196533, "perplexity": 34.70586236847055, "lr": 0.0026291804804649314, "grad_norm": 0.166245, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:47.106069+00:00", "epoch": 0, "step": 15979, "train_loss": 3.443911075592041, "perplexity": 31.309171556304435, "lr": 0.0026291804804649314, "grad_norm": 0.168461, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:47.409977+00:00", "epoch": 0, "step": 15980, "train_loss": 3.519883394241333, "perplexity": 33.780489234610265, "lr": 0.0026291804804649314, "grad_norm": 0.138625, "tokens_per_sec": 107880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:47.714933+00:00", "epoch": 0, "step": 15981, "train_loss": 3.4936954975128174, "perplexity": 32.90733224442244, "lr": 0.0026291804804649314, "grad_norm": 0.171473, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:48.020788+00:00", "epoch": 0, "step": 15982, "train_loss": 3.4754350185394287, "perplexity": 32.31188174728936, "lr": 0.0026291804804649314, "grad_norm": 0.152789, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:48.327037+00:00", "epoch": 0, "step": 15983, "train_loss": 3.53541898727417, "perplexity": 34.30938689963272, "lr": 0.0026291804804649314, "grad_norm": 0.166942, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:48.631152+00:00", "epoch": 0, "step": 15984, "train_loss": 3.549982786178589, "perplexity": 34.812718222529895, "lr": 0.0026291804804649314, "grad_norm": 0.184111, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:48.935635+00:00", "epoch": 0, "step": 15985, "train_loss": 3.5494489669799805, "perplexity": 34.794139484471685, "lr": 0.0026291804804649314, "grad_norm": 0.178444, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:49.240545+00:00", "epoch": 0, "step": 15986, "train_loss": 3.374858856201172, "perplexity": 29.220159245887167, "lr": 0.0026291804804649314, "grad_norm": 0.183919, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:49.545147+00:00", "epoch": 0, "step": 15987, "train_loss": 3.405888557434082, "perplexity": 30.141065896912185, "lr": 0.0026291804804649314, "grad_norm": 0.168836, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:49.850808+00:00", "epoch": 0, "step": 15988, "train_loss": 3.4773004055023193, "perplexity": 32.372212162538105, "lr": 0.0026291804804649314, "grad_norm": 0.16689, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:50.156818+00:00", "epoch": 0, "step": 15989, "train_loss": 3.4527430534362793, "perplexity": 31.58691818641343, "lr": 0.0026291804804649314, "grad_norm": 0.200923, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:50.462232+00:00", "epoch": 0, "step": 15990, "train_loss": 3.5007004737854004, "perplexity": 33.13865659085052, "lr": 0.0026291804804649314, "grad_norm": 0.182255, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:50.766823+00:00", "epoch": 0, "step": 15991, "train_loss": 3.5070927143096924, "perplexity": 33.3511653343272, "lr": 0.0026291804804649314, "grad_norm": 0.182813, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:51.072995+00:00", "epoch": 0, "step": 15992, "train_loss": 3.538442611694336, "perplexity": 34.41328259132482, "lr": 0.0026291804804649314, "grad_norm": 0.196845, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:51.378633+00:00", "epoch": 0, "step": 15993, "train_loss": 3.4576425552368164, "perplexity": 31.742058092720473, "lr": 0.0026291804804649314, "grad_norm": 0.179877, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:51.684432+00:00", "epoch": 0, "step": 15994, "train_loss": 3.5333895683288574, "perplexity": 34.23982938439911, "lr": 0.0026291804804649314, "grad_norm": 0.162429, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:51.989628+00:00", "epoch": 0, "step": 15995, "train_loss": 3.5286672115325928, "perplexity": 34.078517878425245, "lr": 0.0026291804804649314, "grad_norm": 0.16053, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:52.294216+00:00", "epoch": 0, "step": 15996, "train_loss": 3.453568696975708, "perplexity": 31.613008490507898, "lr": 0.0026291804804649314, "grad_norm": 0.201444, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:52.599175+00:00", "epoch": 0, "step": 15997, "train_loss": 3.567150831222534, "perplexity": 35.41554440416451, "lr": 0.0026291804804649314, "grad_norm": 0.193177, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:52.904314+00:00", "epoch": 0, "step": 15998, "train_loss": 3.5332553386688232, "perplexity": 34.235233692187215, "lr": 0.0026291804804649314, "grad_norm": 0.163343, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:53.209372+00:00", "epoch": 0, "step": 15999, "train_loss": 3.39597749710083, "perplexity": 29.84381146087373, "lr": 0.0026291804804649314, "grad_norm": 0.174536, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:53.515342+00:00", "epoch": 0, "step": 16000, "train_loss": 3.3913707733154297, "perplexity": 29.706645450308535, "lr": 0.0026291804804649314, "grad_norm": 0.159046, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:21:56.588623+00:00", "step": 16000, "epoch": 0, "val_loss": 3.4681695699691772, "val_ppl": 32.077972190117976, "eval_train_loss": 3.3913707733154297, "eval_train_ppl": 29.706645450308535} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:21:57.579982+00:00", "step": 16000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4682_epoch_0000_step_0016000.pt", "val_loss": 3.4681695699691772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:58.856956+00:00", "epoch": 0, "step": 16001, "train_loss": 3.3882687091827393, "perplexity": 29.614636313857744, "lr": 0.0026291804804649314, "grad_norm": 0.157117, "tokens_per_sec": 6134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:59.161076+00:00", "epoch": 0, "step": 16002, "train_loss": 3.5864596366882324, "perplexity": 36.10602094776007, "lr": 0.0026291804804649314, "grad_norm": 0.182452, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:59.464994+00:00", "epoch": 0, "step": 16003, "train_loss": 3.4872119426727295, "perplexity": 32.69466591306394, "lr": 0.0026291804804649314, "grad_norm": 0.193964, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:21:59.770307+00:00", "epoch": 0, "step": 16004, "train_loss": 3.5304224491119385, "perplexity": 34.13838629993498, "lr": 0.0026291804804649314, "grad_norm": 0.174064, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:00.073744+00:00", "epoch": 0, "step": 16005, "train_loss": 3.496735095977783, "perplexity": 33.00750949317503, "lr": 0.0026291804804649314, "grad_norm": 0.16768, "tokens_per_sec": 107990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:00.376957+00:00", "epoch": 0, "step": 16006, "train_loss": 3.429194688796997, "perplexity": 30.851787452827388, "lr": 0.0026291804804649314, "grad_norm": 0.168375, "tokens_per_sec": 108073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:00.680538+00:00", "epoch": 0, "step": 16007, "train_loss": 3.4820709228515625, "perplexity": 32.52701330949334, "lr": 0.0026291804804649314, "grad_norm": 0.178572, "tokens_per_sec": 107935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:00.984578+00:00", "epoch": 0, "step": 16008, "train_loss": 3.5572752952575684, "perplexity": 35.067518222686594, "lr": 0.0026291804804649314, "grad_norm": 0.170645, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:01.288534+00:00", "epoch": 0, "step": 16009, "train_loss": 3.478860378265381, "perplexity": 32.422751341413985, "lr": 0.0026291804804649314, "grad_norm": 0.162762, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:01.594019+00:00", "epoch": 0, "step": 16010, "train_loss": 3.480095148086548, "perplexity": 32.46281070324535, "lr": 0.0026291804804649314, "grad_norm": 0.152376, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:01.898589+00:00", "epoch": 0, "step": 16011, "train_loss": 3.526273250579834, "perplexity": 33.997032812245976, "lr": 0.0026291804804649314, "grad_norm": 0.164731, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:02.201972+00:00", "epoch": 0, "step": 16012, "train_loss": 3.490997076034546, "perplexity": 32.818654091626556, "lr": 0.0026291804804649314, "grad_norm": 0.163469, "tokens_per_sec": 108018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:02.505307+00:00", "epoch": 0, "step": 16013, "train_loss": 3.6275081634521484, "perplexity": 37.618959429363294, "lr": 0.0026291804804649314, "grad_norm": 0.195102, "tokens_per_sec": 108019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:02.866103+00:00", "epoch": 0, "step": 16014, "train_loss": 3.4578118324279785, "perplexity": 31.74743175396242, "lr": 0.0026291804804649314, "grad_norm": 0.170167, "tokens_per_sec": 90821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:03.169506+00:00", "epoch": 0, "step": 16015, "train_loss": 3.6038005352020264, "perplexity": 36.73759197086632, "lr": 0.0026291804804649314, "grad_norm": 0.16395, "tokens_per_sec": 108001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:03.473593+00:00", "epoch": 0, "step": 16016, "train_loss": 3.4235727787017822, "perplexity": 30.678828113961533, "lr": 0.0026291804804649314, "grad_norm": 0.18117, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:03.776184+00:00", "epoch": 0, "step": 16017, "train_loss": 3.5746471881866455, "perplexity": 35.682029553062605, "lr": 0.0026291804804649314, "grad_norm": 0.177962, "tokens_per_sec": 108293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:04.079741+00:00", "epoch": 0, "step": 16018, "train_loss": 3.480587959289551, "perplexity": 32.47881268269439, "lr": 0.0026291804804649314, "grad_norm": 0.181502, "tokens_per_sec": 107951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:04.382294+00:00", "epoch": 0, "step": 16019, "train_loss": 3.4672329425811768, "perplexity": 32.04794114895646, "lr": 0.0026291804804649314, "grad_norm": 0.190537, "tokens_per_sec": 108300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:04.686610+00:00", "epoch": 0, "step": 16020, "train_loss": 3.5545198917388916, "perplexity": 34.97102605805888, "lr": 0.0026291804804649314, "grad_norm": 0.155753, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:04.990503+00:00", "epoch": 0, "step": 16021, "train_loss": 3.4789562225341797, "perplexity": 32.42585902523321, "lr": 0.0026291804804649314, "grad_norm": 0.190012, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:05.295651+00:00", "epoch": 0, "step": 16022, "train_loss": 3.618957996368408, "perplexity": 37.2986822037844, "lr": 0.0026291804804649314, "grad_norm": 0.172105, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:05.599616+00:00", "epoch": 0, "step": 16023, "train_loss": 3.4769480228424072, "perplexity": 32.36080676595904, "lr": 0.0026291804804649314, "grad_norm": 0.209134, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:05.903489+00:00", "epoch": 0, "step": 16024, "train_loss": 3.501108169555664, "perplexity": 33.15216983543399, "lr": 0.0026291804804649314, "grad_norm": 0.161452, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:06.207080+00:00", "epoch": 0, "step": 16025, "train_loss": 3.470781087875366, "perplexity": 32.161853870468896, "lr": 0.0026291804804649314, "grad_norm": 0.183041, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:06.512287+00:00", "epoch": 0, "step": 16026, "train_loss": 3.4335825443267822, "perplexity": 30.987458072870968, "lr": 0.0026291804804649314, "grad_norm": 0.187701, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:06.815441+00:00", "epoch": 0, "step": 16027, "train_loss": 3.562868118286133, "perplexity": 35.26419412023871, "lr": 0.0026291804804649314, "grad_norm": 0.163026, "tokens_per_sec": 108091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:07.119273+00:00", "epoch": 0, "step": 16028, "train_loss": 3.4976563453674316, "perplexity": 33.03793165219554, "lr": 0.0026291804804649314, "grad_norm": 0.187593, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:07.422678+00:00", "epoch": 0, "step": 16029, "train_loss": 3.4629456996917725, "perplexity": 31.91083794850767, "lr": 0.0026291804804649314, "grad_norm": 0.169184, "tokens_per_sec": 108002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:07.726254+00:00", "epoch": 0, "step": 16030, "train_loss": 3.497514247894287, "perplexity": 33.03323737911991, "lr": 0.0026291804804649314, "grad_norm": 0.178157, "tokens_per_sec": 108009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:08.030153+00:00", "epoch": 0, "step": 16031, "train_loss": 3.5226943492889404, "perplexity": 33.87557825420284, "lr": 0.0026291804804649314, "grad_norm": 0.231231, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:08.334823+00:00", "epoch": 0, "step": 16032, "train_loss": 3.4976134300231934, "perplexity": 33.03651384840875, "lr": 0.0026291804804649314, "grad_norm": 0.212623, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:08.638216+00:00", "epoch": 0, "step": 16033, "train_loss": 3.4628231525421143, "perplexity": 31.90692760587944, "lr": 0.0026291804804649314, "grad_norm": 0.151288, "tokens_per_sec": 108006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:08.941924+00:00", "epoch": 0, "step": 16034, "train_loss": 3.497528553009033, "perplexity": 33.033709926750966, "lr": 0.0026291804804649314, "grad_norm": 0.206443, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:09.245468+00:00", "epoch": 0, "step": 16035, "train_loss": 3.5549979209899902, "perplexity": 34.98774722774126, "lr": 0.0026291804804649314, "grad_norm": 0.209179, "tokens_per_sec": 107951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:09.550371+00:00", "epoch": 0, "step": 16036, "train_loss": 3.651074171066284, "perplexity": 38.5160166269081, "lr": 0.0026291804804649314, "grad_norm": 0.165423, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:09.855339+00:00", "epoch": 0, "step": 16037, "train_loss": 3.496702194213867, "perplexity": 33.00642350575578, "lr": 0.0026291804804649314, "grad_norm": 0.194753, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:10.159163+00:00", "epoch": 0, "step": 16038, "train_loss": 3.5985326766967773, "perplexity": 36.544572380847654, "lr": 0.0026291804804649314, "grad_norm": 0.248204, "tokens_per_sec": 107851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:10.461932+00:00", "epoch": 0, "step": 16039, "train_loss": 3.5036871433258057, "perplexity": 33.23777875599366, "lr": 0.0026291804804649314, "grad_norm": 0.14771, "tokens_per_sec": 108227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:10.765524+00:00", "epoch": 0, "step": 16040, "train_loss": 3.5799014568328857, "perplexity": 35.87000592890596, "lr": 0.0026291804804649314, "grad_norm": 0.185723, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:11.070218+00:00", "epoch": 0, "step": 16041, "train_loss": 3.485515832901001, "perplexity": 32.63925917196568, "lr": 0.0026291804804649314, "grad_norm": 0.151962, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:11.374792+00:00", "epoch": 0, "step": 16042, "train_loss": 3.4506125450134277, "perplexity": 31.519693627837547, "lr": 0.0026291804804649314, "grad_norm": 0.162978, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:11.678538+00:00", "epoch": 0, "step": 16043, "train_loss": 3.5958828926086426, "perplexity": 36.44786533742217, "lr": 0.0026291804804649314, "grad_norm": 0.169925, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:11.982015+00:00", "epoch": 0, "step": 16044, "train_loss": 3.466012716293335, "perplexity": 32.008859257919966, "lr": 0.0026291804804649314, "grad_norm": 0.144319, "tokens_per_sec": 107976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:12.286391+00:00", "epoch": 0, "step": 16045, "train_loss": 3.511471748352051, "perplexity": 33.497531459909574, "lr": 0.0026291804804649314, "grad_norm": 0.173028, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:12.590287+00:00", "epoch": 0, "step": 16046, "train_loss": 3.4868216514587402, "perplexity": 32.681907962033804, "lr": 0.0026291804804649314, "grad_norm": 0.170254, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:12.894436+00:00", "epoch": 0, "step": 16047, "train_loss": 3.5334694385528564, "perplexity": 34.24256423645689, "lr": 0.0026291804804649314, "grad_norm": 0.164573, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:13.199379+00:00", "epoch": 0, "step": 16048, "train_loss": 3.514401435852051, "perplexity": 33.595812655407315, "lr": 0.0026291804804649314, "grad_norm": 0.179201, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:13.503871+00:00", "epoch": 0, "step": 16049, "train_loss": 3.404845714569092, "perplexity": 30.109649885227338, "lr": 0.0026291804804649314, "grad_norm": 0.172732, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:13.808084+00:00", "epoch": 0, "step": 16050, "train_loss": 3.4280476570129395, "perplexity": 30.816419759835718, "lr": 0.0026291804804649314, "grad_norm": 0.182652, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:14.112754+00:00", "epoch": 0, "step": 16051, "train_loss": 3.507833957672119, "perplexity": 33.375895828785566, "lr": 0.0026291804804649314, "grad_norm": 0.158748, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:14.418600+00:00", "epoch": 0, "step": 16052, "train_loss": 3.4946231842041016, "perplexity": 32.93787410303878, "lr": 0.0026291804804649314, "grad_norm": 0.175281, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:14.723021+00:00", "epoch": 0, "step": 16053, "train_loss": 3.467712879180908, "perplexity": 32.06332582039803, "lr": 0.0026291804804649314, "grad_norm": 0.161637, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:15.026877+00:00", "epoch": 0, "step": 16054, "train_loss": 3.4147486686706543, "perplexity": 30.409305655830813, "lr": 0.0026291804804649314, "grad_norm": 0.163051, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:15.331559+00:00", "epoch": 0, "step": 16055, "train_loss": 3.440897226333618, "perplexity": 31.214952485184675, "lr": 0.0026291804804649314, "grad_norm": 0.156693, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:15.635387+00:00", "epoch": 0, "step": 16056, "train_loss": 3.5383925437927246, "perplexity": 34.411559633610715, "lr": 0.0026291804804649314, "grad_norm": 0.148443, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:15.940909+00:00", "epoch": 0, "step": 16057, "train_loss": 3.545757532119751, "perplexity": 34.66593595811268, "lr": 0.0026291804804649314, "grad_norm": 0.144813, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:16.247412+00:00", "epoch": 0, "step": 16058, "train_loss": 3.492180109024048, "perplexity": 32.85750261710069, "lr": 0.0026291804804649314, "grad_norm": 0.165999, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:16.552205+00:00", "epoch": 0, "step": 16059, "train_loss": 3.536646842956543, "perplexity": 34.35153974880796, "lr": 0.0026291804804649314, "grad_norm": 0.156199, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:16.856357+00:00", "epoch": 0, "step": 16060, "train_loss": 3.4189469814300537, "perplexity": 30.53724180183223, "lr": 0.0026291804804649314, "grad_norm": 0.167186, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:17.161208+00:00", "epoch": 0, "step": 16061, "train_loss": 3.5990636348724365, "perplexity": 36.56398117250156, "lr": 0.0026291804804649314, "grad_norm": 0.169659, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:17.466477+00:00", "epoch": 0, "step": 16062, "train_loss": 3.463581085205078, "perplexity": 31.93112007546007, "lr": 0.0026291804804649314, "grad_norm": 0.161065, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:17.770883+00:00", "epoch": 0, "step": 16063, "train_loss": 3.5469460487365723, "perplexity": 34.70716149280023, "lr": 0.0026291804804649314, "grad_norm": 0.172278, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:18.075415+00:00", "epoch": 0, "step": 16064, "train_loss": 3.4985876083374023, "perplexity": 33.0687129850808, "lr": 0.0026291804804649314, "grad_norm": 0.17313, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:18.379561+00:00", "epoch": 0, "step": 16065, "train_loss": 3.5583460330963135, "perplexity": 35.10508645063084, "lr": 0.0026291804804649314, "grad_norm": 0.141271, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:18.683219+00:00", "epoch": 0, "step": 16066, "train_loss": 3.4885027408599854, "perplexity": 32.736895377552145, "lr": 0.0026291804804649314, "grad_norm": 0.160537, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:18.988330+00:00", "epoch": 0, "step": 16067, "train_loss": 3.58152437210083, "perplexity": 35.928267172932834, "lr": 0.0026291804804649314, "grad_norm": 0.175709, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:19.292791+00:00", "epoch": 0, "step": 16068, "train_loss": 3.4715869426727295, "perplexity": 32.18778210049363, "lr": 0.0026291804804649314, "grad_norm": 0.183142, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:19.597441+00:00", "epoch": 0, "step": 16069, "train_loss": 3.4539456367492676, "perplexity": 31.62492693689697, "lr": 0.0026291804804649314, "grad_norm": 0.157153, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:19.902097+00:00", "epoch": 0, "step": 16070, "train_loss": 3.4375836849212646, "perplexity": 31.111691621494284, "lr": 0.0026291804804649314, "grad_norm": 0.217785, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:20.206407+00:00", "epoch": 0, "step": 16071, "train_loss": 3.501922845840454, "perplexity": 33.17918912648168, "lr": 0.0026291804804649314, "grad_norm": 0.174377, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:20.512312+00:00", "epoch": 0, "step": 16072, "train_loss": 3.544822931289673, "perplexity": 34.63355228085382, "lr": 0.0026291804804649314, "grad_norm": 0.157249, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:20.816711+00:00", "epoch": 0, "step": 16073, "train_loss": 3.5824577808380127, "perplexity": 35.96181858757969, "lr": 0.0026291804804649314, "grad_norm": 0.173883, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:21.122367+00:00", "epoch": 0, "step": 16074, "train_loss": 3.543159008026123, "perplexity": 34.575972624863674, "lr": 0.0026291804804649314, "grad_norm": 0.162227, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:21.427595+00:00", "epoch": 0, "step": 16075, "train_loss": 3.5468435287475586, "perplexity": 34.70360349737123, "lr": 0.0026291804804649314, "grad_norm": 0.195297, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:21.732444+00:00", "epoch": 0, "step": 16076, "train_loss": 3.542118549346924, "perplexity": 34.54001646274856, "lr": 0.0026291804804649314, "grad_norm": 0.181219, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:22.037877+00:00", "epoch": 0, "step": 16077, "train_loss": 3.5183510780334473, "perplexity": 33.7287664813727, "lr": 0.0026291804804649314, "grad_norm": 0.20697, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:22.342342+00:00", "epoch": 0, "step": 16078, "train_loss": 3.512970447540283, "perplexity": 33.547771821307016, "lr": 0.0026291804804649314, "grad_norm": 0.168747, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:22.646795+00:00", "epoch": 0, "step": 16079, "train_loss": 3.557943820953369, "perplexity": 35.09096959775642, "lr": 0.0026291804804649314, "grad_norm": 0.188978, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:22.952492+00:00", "epoch": 0, "step": 16080, "train_loss": 3.5351579189300537, "perplexity": 34.300430973912405, "lr": 0.0026291804804649314, "grad_norm": 0.181885, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:23.257882+00:00", "epoch": 0, "step": 16081, "train_loss": 3.549574375152588, "perplexity": 34.798503227540685, "lr": 0.0026291804804649314, "grad_norm": 0.160937, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:23.563043+00:00", "epoch": 0, "step": 16082, "train_loss": 3.4486260414123535, "perplexity": 31.45714179323359, "lr": 0.0026291804804649314, "grad_norm": 0.181945, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:23.867978+00:00", "epoch": 0, "step": 16083, "train_loss": 3.4499149322509766, "perplexity": 31.49771275525428, "lr": 0.0026291804804649314, "grad_norm": 0.152992, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:24.172001+00:00", "epoch": 0, "step": 16084, "train_loss": 3.4363718032836914, "perplexity": 31.074010770681635, "lr": 0.0026291804804649314, "grad_norm": 0.190942, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:24.476213+00:00", "epoch": 0, "step": 16085, "train_loss": 3.473552703857422, "perplexity": 32.251117824226526, "lr": 0.0026291804804649314, "grad_norm": 0.181335, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:24.780425+00:00", "epoch": 0, "step": 16086, "train_loss": 3.5397796630859375, "perplexity": 34.45932569284031, "lr": 0.0026291804804649314, "grad_norm": 0.18075, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:25.084832+00:00", "epoch": 0, "step": 16087, "train_loss": 3.4282891750335693, "perplexity": 30.82386337938696, "lr": 0.0026291804804649314, "grad_norm": 0.166669, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:25.390255+00:00", "epoch": 0, "step": 16088, "train_loss": 3.498812437057495, "perplexity": 33.07614861733687, "lr": 0.0026291804804649314, "grad_norm": 0.199708, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:25.694343+00:00", "epoch": 0, "step": 16089, "train_loss": 3.4309589862823486, "perplexity": 30.906267228983243, "lr": 0.0026291804804649314, "grad_norm": 0.184328, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:25.998634+00:00", "epoch": 0, "step": 16090, "train_loss": 3.562600612640381, "perplexity": 35.254762010846164, "lr": 0.0026291804804649314, "grad_norm": 0.172908, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:26.303819+00:00", "epoch": 0, "step": 16091, "train_loss": 3.691152572631836, "perplexity": 40.091028160418816, "lr": 0.0026291804804649314, "grad_norm": 0.178067, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:26.610284+00:00", "epoch": 0, "step": 16092, "train_loss": 3.6636571884155273, "perplexity": 39.00372632715743, "lr": 0.0026291804804649314, "grad_norm": 0.189523, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:26.915832+00:00", "epoch": 0, "step": 16093, "train_loss": 3.564154863357544, "perplexity": 35.309599354436045, "lr": 0.0026291804804649314, "grad_norm": 0.198186, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:27.220057+00:00", "epoch": 0, "step": 16094, "train_loss": 3.633514404296875, "perplexity": 37.84558787122609, "lr": 0.0026291804804649314, "grad_norm": 0.191351, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:27.525481+00:00", "epoch": 0, "step": 16095, "train_loss": 3.5170602798461914, "perplexity": 33.68525753741424, "lr": 0.0026291804804649314, "grad_norm": 0.204275, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:27.831145+00:00", "epoch": 0, "step": 16096, "train_loss": 3.5050597190856934, "perplexity": 33.28343144914931, "lr": 0.0026291804804649314, "grad_norm": 0.21371, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:28.137316+00:00", "epoch": 0, "step": 16097, "train_loss": 3.4030518531799316, "perplexity": 30.055685763405467, "lr": 0.0026291804804649314, "grad_norm": 0.183267, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:28.443185+00:00", "epoch": 0, "step": 16098, "train_loss": 3.5161263942718506, "perplexity": 33.65381404592891, "lr": 0.0026291804804649314, "grad_norm": 0.165474, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:28.746758+00:00", "epoch": 0, "step": 16099, "train_loss": 3.51249098777771, "perplexity": 33.53169086998732, "lr": 0.0026291804804649314, "grad_norm": 0.191067, "tokens_per_sec": 107941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:29.051963+00:00", "epoch": 0, "step": 16100, "train_loss": 3.5059471130371094, "perplexity": 33.31298007359591, "lr": 0.0026291804804649314, "grad_norm": 0.152076, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:29.357358+00:00", "epoch": 0, "step": 16101, "train_loss": 3.554171323776245, "perplexity": 34.958838402990615, "lr": 0.0026291804804649314, "grad_norm": 0.153827, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:29.662956+00:00", "epoch": 0, "step": 16102, "train_loss": 3.505565643310547, "perplexity": 33.300274603726386, "lr": 0.0026291804804649314, "grad_norm": 0.167428, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:29.967446+00:00", "epoch": 0, "step": 16103, "train_loss": 3.477354049682617, "perplexity": 32.37394878990355, "lr": 0.0026291804804649314, "grad_norm": 0.142567, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:30.272841+00:00", "epoch": 0, "step": 16104, "train_loss": 3.501042604446411, "perplexity": 33.149996281052196, "lr": 0.0026291804804649314, "grad_norm": 0.140483, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:30.578040+00:00", "epoch": 0, "step": 16105, "train_loss": 3.558950424194336, "perplexity": 35.12631006541277, "lr": 0.0026291804804649314, "grad_norm": 0.149447, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:30.882663+00:00", "epoch": 0, "step": 16106, "train_loss": 3.4269320964813232, "perplexity": 30.782061346234183, "lr": 0.0026291804804649314, "grad_norm": 0.149128, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:31.187796+00:00", "epoch": 0, "step": 16107, "train_loss": 3.585371494293213, "perplexity": 36.06675382363713, "lr": 0.0026291804804649314, "grad_norm": 0.159383, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:31.492029+00:00", "epoch": 0, "step": 16108, "train_loss": 3.539541721343994, "perplexity": 34.451127356260685, "lr": 0.0026291804804649314, "grad_norm": 0.20691, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:31.797749+00:00", "epoch": 0, "step": 16109, "train_loss": 3.4602901935577393, "perplexity": 31.826210936098988, "lr": 0.0026291804804649314, "grad_norm": 0.190269, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:32.103486+00:00", "epoch": 0, "step": 16110, "train_loss": 3.604769468307495, "perplexity": 36.77320549071503, "lr": 0.0026291804804649314, "grad_norm": 0.150215, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:32.408962+00:00", "epoch": 0, "step": 16111, "train_loss": 3.4650282859802246, "perplexity": 31.977364271431583, "lr": 0.0026291804804649314, "grad_norm": 0.175218, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:32.713076+00:00", "epoch": 0, "step": 16112, "train_loss": 3.5218329429626465, "perplexity": 33.84641018137265, "lr": 0.0026291804804649314, "grad_norm": 0.179743, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:33.017392+00:00", "epoch": 0, "step": 16113, "train_loss": 3.44972562789917, "perplexity": 31.491750665500327, "lr": 0.0026291804804649314, "grad_norm": 0.165235, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:33.321977+00:00", "epoch": 0, "step": 16114, "train_loss": 3.4782345294952393, "perplexity": 32.402465950817856, "lr": 0.0026291804804649314, "grad_norm": 0.177896, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:33.626590+00:00", "epoch": 0, "step": 16115, "train_loss": 3.588280200958252, "perplexity": 36.17181415163586, "lr": 0.0026291804804649314, "grad_norm": 0.180311, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:33.933593+00:00", "epoch": 0, "step": 16116, "train_loss": 3.510098457336426, "perplexity": 33.451561173425496, "lr": 0.0026291804804649314, "grad_norm": 0.207569, "tokens_per_sec": 106734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:34.239552+00:00", "epoch": 0, "step": 16117, "train_loss": 3.5366039276123047, "perplexity": 34.35006557228714, "lr": 0.0026291804804649314, "grad_norm": 0.193222, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:34.543614+00:00", "epoch": 0, "step": 16118, "train_loss": 3.53005313873291, "perplexity": 34.125780967334485, "lr": 0.0026291804804649314, "grad_norm": 0.171569, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:34.848044+00:00", "epoch": 0, "step": 16119, "train_loss": 3.554018020629883, "perplexity": 34.953479513848045, "lr": 0.0026291804804649314, "grad_norm": 0.198819, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:35.152694+00:00", "epoch": 0, "step": 16120, "train_loss": 3.472400188446045, "perplexity": 32.21396932512084, "lr": 0.0026291804804649314, "grad_norm": 0.171829, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:35.458255+00:00", "epoch": 0, "step": 16121, "train_loss": 3.523606538772583, "perplexity": 33.9064932984841, "lr": 0.0026291804804649314, "grad_norm": 0.168437, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:35.763264+00:00", "epoch": 0, "step": 16122, "train_loss": 3.46809458732605, "perplexity": 32.075566989152286, "lr": 0.0026291804804649314, "grad_norm": 0.178499, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:36.067865+00:00", "epoch": 0, "step": 16123, "train_loss": 3.550422430038452, "perplexity": 34.8280267852528, "lr": 0.0026291804804649314, "grad_norm": 0.167758, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:36.372501+00:00", "epoch": 0, "step": 16124, "train_loss": 3.414503812789917, "perplexity": 30.40186067002332, "lr": 0.0026291804804649314, "grad_norm": 0.182425, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:36.676685+00:00", "epoch": 0, "step": 16125, "train_loss": 3.435184955596924, "perplexity": 31.03715252976541, "lr": 0.0026291804804649314, "grad_norm": 0.17167, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:36.982237+00:00", "epoch": 0, "step": 16126, "train_loss": 3.475801706314087, "perplexity": 32.32373229189442, "lr": 0.0026291804804649314, "grad_norm": 0.179793, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:37.287392+00:00", "epoch": 0, "step": 16127, "train_loss": 3.453768491744995, "perplexity": 31.619325235250997, "lr": 0.0026291804804649314, "grad_norm": 0.183558, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:37.591522+00:00", "epoch": 0, "step": 16128, "train_loss": 3.5791759490966797, "perplexity": 35.84399140011882, "lr": 0.0026291804804649314, "grad_norm": 0.186313, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:37.895067+00:00", "epoch": 0, "step": 16129, "train_loss": 3.4970428943634033, "perplexity": 33.01767071503399, "lr": 0.0026291804804649314, "grad_norm": 0.167586, "tokens_per_sec": 107951} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:38.199621+00:00", "epoch": 0, "step": 16130, "train_loss": 3.5356829166412354, "perplexity": 34.31844334948051, "lr": 0.0026291804804649314, "grad_norm": 0.165058, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:38.504942+00:00", "epoch": 0, "step": 16131, "train_loss": 3.469823122024536, "perplexity": 32.131058665494024, "lr": 0.0026291804804649314, "grad_norm": 0.207995, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:38.810631+00:00", "epoch": 0, "step": 16132, "train_loss": 3.4670395851135254, "perplexity": 32.04174503926381, "lr": 0.0026291804804649314, "grad_norm": 0.172363, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:39.114824+00:00", "epoch": 0, "step": 16133, "train_loss": 3.417562246322632, "perplexity": 30.49498507496425, "lr": 0.0026291804804649314, "grad_norm": 0.177176, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:39.418326+00:00", "epoch": 0, "step": 16134, "train_loss": 3.4709370136260986, "perplexity": 32.16686912267191, "lr": 0.0026291804804649314, "grad_norm": 0.164127, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:39.722347+00:00", "epoch": 0, "step": 16135, "train_loss": 3.4689419269561768, "perplexity": 32.102757406353156, "lr": 0.0026291804804649314, "grad_norm": 0.164907, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:40.027036+00:00", "epoch": 0, "step": 16136, "train_loss": 3.5137343406677246, "perplexity": 33.573408524246226, "lr": 0.0026291804804649314, "grad_norm": 0.158163, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:40.331615+00:00", "epoch": 0, "step": 16137, "train_loss": 3.5373172760009766, "perplexity": 34.374577878081375, "lr": 0.0026291804804649314, "grad_norm": 0.161248, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:40.637502+00:00", "epoch": 0, "step": 16138, "train_loss": 3.479670763015747, "perplexity": 32.449036893932906, "lr": 0.0026291804804649314, "grad_norm": 0.164233, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:40.941108+00:00", "epoch": 0, "step": 16139, "train_loss": 3.429340124130249, "perplexity": 30.856274719112633, "lr": 0.0026291804804649314, "grad_norm": 0.162641, "tokens_per_sec": 107923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:41.245525+00:00", "epoch": 0, "step": 16140, "train_loss": 3.5953917503356934, "perplexity": 36.42996864526699, "lr": 0.0026291804804649314, "grad_norm": 0.160158, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:41.549741+00:00", "epoch": 0, "step": 16141, "train_loss": 3.4797229766845703, "perplexity": 32.450731221432065, "lr": 0.0026291804804649314, "grad_norm": 0.148033, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:41.854630+00:00", "epoch": 0, "step": 16142, "train_loss": 3.4782886505126953, "perplexity": 32.404219652698835, "lr": 0.0026291804804649314, "grad_norm": 0.141292, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:42.160659+00:00", "epoch": 0, "step": 16143, "train_loss": 3.5795185565948486, "perplexity": 35.85627392425909, "lr": 0.0026291804804649314, "grad_norm": 0.154546, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:42.465800+00:00", "epoch": 0, "step": 16144, "train_loss": 3.5039165019989014, "perplexity": 33.24540300313597, "lr": 0.0026291804804649314, "grad_norm": 0.141766, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:42.770843+00:00", "epoch": 0, "step": 16145, "train_loss": 3.3891043663024902, "perplexity": 29.639394338711366, "lr": 0.0026291804804649314, "grad_norm": 0.174986, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:43.075037+00:00", "epoch": 0, "step": 16146, "train_loss": 3.51344895362854, "perplexity": 33.563828475668025, "lr": 0.0026291804804649314, "grad_norm": 0.181694, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:43.380086+00:00", "epoch": 0, "step": 16147, "train_loss": 3.6101720333099365, "perplexity": 36.97241275428856, "lr": 0.0026291804804649314, "grad_norm": 0.167454, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:43.685970+00:00", "epoch": 0, "step": 16148, "train_loss": 3.560112476348877, "perplexity": 35.16715239556823, "lr": 0.0026291804804649314, "grad_norm": 0.158923, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:43.991019+00:00", "epoch": 0, "step": 16149, "train_loss": 3.535780429840088, "perplexity": 34.32179001384048, "lr": 0.0026291804804649314, "grad_norm": 0.211465, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:44.295209+00:00", "epoch": 0, "step": 16150, "train_loss": 3.4904346466064453, "perplexity": 32.800201104512695, "lr": 0.0026291804804649314, "grad_norm": 0.194903, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:44.599735+00:00", "epoch": 0, "step": 16151, "train_loss": 3.5017848014831543, "perplexity": 33.174609242763815, "lr": 0.0026291804804649314, "grad_norm": 0.167819, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:44.904235+00:00", "epoch": 0, "step": 16152, "train_loss": 3.4816718101501465, "perplexity": 32.51403395562718, "lr": 0.0026291804804649314, "grad_norm": 0.167795, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:45.209368+00:00", "epoch": 0, "step": 16153, "train_loss": 3.5355777740478516, "perplexity": 34.314835209033795, "lr": 0.0026291804804649314, "grad_norm": 0.164643, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:45.514889+00:00", "epoch": 0, "step": 16154, "train_loss": 3.5360896587371826, "perplexity": 34.3324049442498, "lr": 0.0026291804804649314, "grad_norm": 0.158608, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:45.820794+00:00", "epoch": 0, "step": 16155, "train_loss": 3.4820683002471924, "perplexity": 32.52692800411795, "lr": 0.0026291804804649314, "grad_norm": 0.160187, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:46.124892+00:00", "epoch": 0, "step": 16156, "train_loss": 3.5604922771453857, "perplexity": 35.18051144478719, "lr": 0.0026291804804649314, "grad_norm": 0.159677, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:46.429141+00:00", "epoch": 0, "step": 16157, "train_loss": 3.439822196960449, "perplexity": 31.18141352529459, "lr": 0.0026291804804649314, "grad_norm": 0.169657, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:46.733895+00:00", "epoch": 0, "step": 16158, "train_loss": 3.477800130844116, "perplexity": 32.388393420081236, "lr": 0.0026291804804649314, "grad_norm": 0.189018, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:47.039244+00:00", "epoch": 0, "step": 16159, "train_loss": 3.537745237350464, "perplexity": 34.38929201713496, "lr": 0.0026291804804649314, "grad_norm": 0.188155, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:47.344128+00:00", "epoch": 0, "step": 16160, "train_loss": 3.5079784393310547, "perplexity": 33.38071838196022, "lr": 0.0026291804804649314, "grad_norm": 0.201815, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:47.647827+00:00", "epoch": 0, "step": 16161, "train_loss": 3.4658467769622803, "perplexity": 32.00354816989827, "lr": 0.0026291804804649314, "grad_norm": 0.176353, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:47.952730+00:00", "epoch": 0, "step": 16162, "train_loss": 3.489616632461548, "perplexity": 32.77338104714242, "lr": 0.0026291804804649314, "grad_norm": 0.184199, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:48.257424+00:00", "epoch": 0, "step": 16163, "train_loss": 3.469271421432495, "perplexity": 32.11333683042832, "lr": 0.0026291804804649314, "grad_norm": 0.198024, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:48.563001+00:00", "epoch": 0, "step": 16164, "train_loss": 3.475736141204834, "perplexity": 32.32161305233009, "lr": 0.0026291804804649314, "grad_norm": 0.175625, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:48.867524+00:00", "epoch": 0, "step": 16165, "train_loss": 3.48573637008667, "perplexity": 32.64645813611626, "lr": 0.0026291804804649314, "grad_norm": 0.190094, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:49.172466+00:00", "epoch": 0, "step": 16166, "train_loss": 3.446185350418091, "perplexity": 31.380458249038274, "lr": 0.0026291804804649314, "grad_norm": 0.15794, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:49.476904+00:00", "epoch": 0, "step": 16167, "train_loss": 3.557755470275879, "perplexity": 35.08436081226304, "lr": 0.0026291804804649314, "grad_norm": 0.162546, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:49.781916+00:00", "epoch": 0, "step": 16168, "train_loss": 3.447462320327759, "perplexity": 31.42055574617378, "lr": 0.0026291804804649314, "grad_norm": 0.159801, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:50.087764+00:00", "epoch": 0, "step": 16169, "train_loss": 3.4979608058929443, "perplexity": 33.047991929629184, "lr": 0.0026291804804649314, "grad_norm": 0.185648, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:50.393757+00:00", "epoch": 0, "step": 16170, "train_loss": 3.5693881511688232, "perplexity": 35.49486901224363, "lr": 0.0026291804804649314, "grad_norm": 0.179289, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:50.698274+00:00", "epoch": 0, "step": 16171, "train_loss": 3.4994616508483887, "perplexity": 33.097629081121084, "lr": 0.0026291804804649314, "grad_norm": 0.163922, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:51.002763+00:00", "epoch": 0, "step": 16172, "train_loss": 3.520406484603882, "perplexity": 33.79816410534532, "lr": 0.0026291804804649314, "grad_norm": 0.195174, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:51.307746+00:00", "epoch": 0, "step": 16173, "train_loss": 3.5001301765441895, "perplexity": 33.11976309438585, "lr": 0.0026291804804649314, "grad_norm": 0.18863, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:51.613606+00:00", "epoch": 0, "step": 16174, "train_loss": 3.4857966899871826, "perplexity": 32.64842742661622, "lr": 0.0026291804804649314, "grad_norm": 0.157681, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:51.918757+00:00", "epoch": 0, "step": 16175, "train_loss": 3.6284570693969727, "perplexity": 37.65467322544083, "lr": 0.0026291804804649314, "grad_norm": 0.177821, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:52.222491+00:00", "epoch": 0, "step": 16176, "train_loss": 3.5598978996276855, "perplexity": 35.15960715285922, "lr": 0.0026291804804649314, "grad_norm": 0.160913, "tokens_per_sec": 107884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:52.526361+00:00", "epoch": 0, "step": 16177, "train_loss": 3.557922840118408, "perplexity": 35.09023336763806, "lr": 0.0026291804804649314, "grad_norm": 0.163069, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:52.830817+00:00", "epoch": 0, "step": 16178, "train_loss": 3.4560420513153076, "perplexity": 31.69129543801646, "lr": 0.0026291804804649314, "grad_norm": 0.149443, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:53.135817+00:00", "epoch": 0, "step": 16179, "train_loss": 3.5710511207580566, "perplexity": 35.55394500716192, "lr": 0.0026291804804649314, "grad_norm": 0.163474, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:53.441429+00:00", "epoch": 0, "step": 16180, "train_loss": 3.4972591400146484, "perplexity": 33.02481141478516, "lr": 0.0026291804804649314, "grad_norm": 0.177526, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:53.745678+00:00", "epoch": 0, "step": 16181, "train_loss": 3.4250106811523438, "perplexity": 30.72297300650551, "lr": 0.0026291804804649314, "grad_norm": 0.17528, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:54.048967+00:00", "epoch": 0, "step": 16182, "train_loss": 3.4592227935791016, "perplexity": 31.792257763257354, "lr": 0.0026291804804649314, "grad_norm": 0.186383, "tokens_per_sec": 108042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:54.353048+00:00", "epoch": 0, "step": 16183, "train_loss": 3.502393960952759, "perplexity": 33.194824026522845, "lr": 0.0026291804804649314, "grad_norm": 0.163568, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:54.657151+00:00", "epoch": 0, "step": 16184, "train_loss": 3.4156720638275146, "perplexity": 30.437398429766464, "lr": 0.0026291804804649314, "grad_norm": 0.18124, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:54.962386+00:00", "epoch": 0, "step": 16185, "train_loss": 3.4285526275634766, "perplexity": 30.831985073969292, "lr": 0.0026291804804649314, "grad_norm": 0.219951, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:55.267574+00:00", "epoch": 0, "step": 16186, "train_loss": 3.5357649326324463, "perplexity": 34.3212581260554, "lr": 0.0026291804804649314, "grad_norm": 0.205215, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:55.573440+00:00", "epoch": 0, "step": 16187, "train_loss": 3.4495036602020264, "perplexity": 31.484761289862593, "lr": 0.0026291804804649314, "grad_norm": 0.175964, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:55.878304+00:00", "epoch": 0, "step": 16188, "train_loss": 3.584155321121216, "perplexity": 36.02291706718253, "lr": 0.0026291804804649314, "grad_norm": 0.19851, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:56.183415+00:00", "epoch": 0, "step": 16189, "train_loss": 3.4861536026000977, "perplexity": 32.66008214189034, "lr": 0.0026291804804649314, "grad_norm": 0.166649, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:56.489317+00:00", "epoch": 0, "step": 16190, "train_loss": 3.613631248474121, "perplexity": 37.100529749515246, "lr": 0.0026291804804649314, "grad_norm": 0.14879, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:56.794615+00:00", "epoch": 0, "step": 16191, "train_loss": 3.5296528339385986, "perplexity": 34.11212298746343, "lr": 0.0026291804804649314, "grad_norm": 0.161964, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:57.099257+00:00", "epoch": 0, "step": 16192, "train_loss": 3.4812891483306885, "perplexity": 32.50159445644829, "lr": 0.0026291804804649314, "grad_norm": 0.157312, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:57.404094+00:00", "epoch": 0, "step": 16193, "train_loss": 3.4775452613830566, "perplexity": 32.380139659566105, "lr": 0.0026291804804649314, "grad_norm": 0.143926, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:57.709947+00:00", "epoch": 0, "step": 16194, "train_loss": 3.564277410507202, "perplexity": 35.31392671033965, "lr": 0.0026291804804649314, "grad_norm": 0.186012, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:58.015557+00:00", "epoch": 0, "step": 16195, "train_loss": 3.584803819656372, "perplexity": 36.04628545249552, "lr": 0.0026291804804649314, "grad_norm": 0.184092, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:58.321587+00:00", "epoch": 0, "step": 16196, "train_loss": 3.5336503982543945, "perplexity": 34.24876132135543, "lr": 0.0026291804804649314, "grad_norm": 0.185807, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:58.628272+00:00", "epoch": 0, "step": 16197, "train_loss": 3.572242021560669, "perplexity": 35.596311450915756, "lr": 0.0026291804804649314, "grad_norm": 0.163664, "tokens_per_sec": 106845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:58.933435+00:00", "epoch": 0, "step": 16198, "train_loss": 3.409911632537842, "perplexity": 30.262569914737547, "lr": 0.0026291804804649314, "grad_norm": 0.146863, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:59.237394+00:00", "epoch": 0, "step": 16199, "train_loss": 3.508849620819092, "perplexity": 33.40981171681929, "lr": 0.0026291804804649314, "grad_norm": 0.167995, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:59.542166+00:00", "epoch": 0, "step": 16200, "train_loss": 3.487880229949951, "perplexity": 32.71652264480108, "lr": 0.0026291804804649314, "grad_norm": 0.194082, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:22:59.847789+00:00", "epoch": 0, "step": 16201, "train_loss": 3.557359457015991, "perplexity": 35.070469690882376, "lr": 0.0026291804804649314, "grad_norm": 0.19815, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:00.153590+00:00", "epoch": 0, "step": 16202, "train_loss": 3.5698859691619873, "perplexity": 35.51254339565096, "lr": 0.0026291804804649314, "grad_norm": 0.19443, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:00.457695+00:00", "epoch": 0, "step": 16203, "train_loss": 3.5429317951202393, "perplexity": 34.5681174100867, "lr": 0.0026291804804649314, "grad_norm": 0.149724, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:00.761561+00:00", "epoch": 0, "step": 16204, "train_loss": 3.447888135910034, "perplexity": 31.43393795738908, "lr": 0.0026291804804649314, "grad_norm": 0.192208, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:01.066080+00:00", "epoch": 0, "step": 16205, "train_loss": 3.6986501216888428, "perplexity": 40.39274225555283, "lr": 0.0026291804804649314, "grad_norm": 0.201691, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:01.371804+00:00", "epoch": 0, "step": 16206, "train_loss": 3.5387701988220215, "perplexity": 34.4245577864267, "lr": 0.0026291804804649314, "grad_norm": 0.180947, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:01.676618+00:00", "epoch": 0, "step": 16207, "train_loss": 3.4966931343078613, "perplexity": 33.006124472015834, "lr": 0.0026291804804649314, "grad_norm": 0.164484, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:01.981088+00:00", "epoch": 0, "step": 16208, "train_loss": 3.549159049987793, "perplexity": 34.7840535343214, "lr": 0.0026291804804649314, "grad_norm": 0.203144, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:02.285819+00:00", "epoch": 0, "step": 16209, "train_loss": 3.5097172260284424, "perplexity": 33.43881082157634, "lr": 0.0026291804804649314, "grad_norm": 0.213012, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:02.590998+00:00", "epoch": 0, "step": 16210, "train_loss": 3.5006728172302246, "perplexity": 33.13774010243958, "lr": 0.0026291804804649314, "grad_norm": 0.19016, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:02.896041+00:00", "epoch": 0, "step": 16211, "train_loss": 3.463859796524048, "perplexity": 31.9400208803724, "lr": 0.0026291804804649314, "grad_norm": 0.191849, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:03.201167+00:00", "epoch": 0, "step": 16212, "train_loss": 3.4512336254119873, "perplexity": 31.539275972194044, "lr": 0.0026291804804649314, "grad_norm": 0.196711, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:03.506783+00:00", "epoch": 0, "step": 16213, "train_loss": 3.61881685256958, "perplexity": 37.29341809759359, "lr": 0.0026291804804649314, "grad_norm": 0.175837, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:03.811536+00:00", "epoch": 0, "step": 16214, "train_loss": 3.468421459197998, "perplexity": 32.08605330352507, "lr": 0.0026291804804649314, "grad_norm": 0.189665, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:04.116197+00:00", "epoch": 0, "step": 16215, "train_loss": 3.5112900733947754, "perplexity": 33.49144635008553, "lr": 0.0026291804804649314, "grad_norm": 0.202409, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:04.420566+00:00", "epoch": 0, "step": 16216, "train_loss": 3.40613055229187, "perplexity": 30.14836076249187, "lr": 0.0026291804804649314, "grad_norm": 0.171524, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:04.726178+00:00", "epoch": 0, "step": 16217, "train_loss": 3.5348129272460938, "perplexity": 34.288599651441444, "lr": 0.0026291804804649314, "grad_norm": 0.176157, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:05.031415+00:00", "epoch": 0, "step": 16218, "train_loss": 3.6071360111236572, "perplexity": 36.86033391190222, "lr": 0.0026291804804649314, "grad_norm": 0.20565, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:05.335807+00:00", "epoch": 0, "step": 16219, "train_loss": 3.4595835208892822, "perplexity": 31.803728167604536, "lr": 0.0026291804804649314, "grad_norm": 0.175054, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:05.640122+00:00", "epoch": 0, "step": 16220, "train_loss": 3.468221664428711, "perplexity": 32.079643318270044, "lr": 0.0026291804804649314, "grad_norm": 0.189771, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:05.944660+00:00", "epoch": 0, "step": 16221, "train_loss": 3.599592447280884, "perplexity": 36.58332177277182, "lr": 0.0026291804804649314, "grad_norm": 0.200562, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:06.249759+00:00", "epoch": 0, "step": 16222, "train_loss": 3.5340793132781982, "perplexity": 34.26345428042302, "lr": 0.0026291804804649314, "grad_norm": 0.175139, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:06.554863+00:00", "epoch": 0, "step": 16223, "train_loss": 3.5688486099243164, "perplexity": 35.4757232318767, "lr": 0.0026291804804649314, "grad_norm": 0.18753, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:06.860055+00:00", "epoch": 0, "step": 16224, "train_loss": 3.6067514419555664, "perplexity": 36.84616128930568, "lr": 0.0026291804804649314, "grad_norm": 0.17903, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:07.164983+00:00", "epoch": 0, "step": 16225, "train_loss": 3.489462375640869, "perplexity": 32.76832591948318, "lr": 0.0026291804804649314, "grad_norm": 0.173143, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:07.469080+00:00", "epoch": 0, "step": 16226, "train_loss": 3.5693883895874023, "perplexity": 35.49487747488088, "lr": 0.0026291804804649314, "grad_norm": 0.173532, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:07.774158+00:00", "epoch": 0, "step": 16227, "train_loss": 3.3895838260650635, "perplexity": 29.653608643005, "lr": 0.0026291804804649314, "grad_norm": 0.155529, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:08.079611+00:00", "epoch": 0, "step": 16228, "train_loss": 3.498591661453247, "perplexity": 33.06884701667698, "lr": 0.0026291804804649314, "grad_norm": 0.170244, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:08.385077+00:00", "epoch": 0, "step": 16229, "train_loss": 3.4743003845214844, "perplexity": 32.27524037827796, "lr": 0.0026291804804649314, "grad_norm": 0.157308, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:08.689896+00:00", "epoch": 0, "step": 16230, "train_loss": 3.5696184635162354, "perplexity": 35.50304486031017, "lr": 0.0026291804804649314, "grad_norm": 0.161052, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:08.994061+00:00", "epoch": 0, "step": 16231, "train_loss": 3.555934429168701, "perplexity": 35.02052888692476, "lr": 0.0026291804804649314, "grad_norm": 0.159916, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:09.298823+00:00", "epoch": 0, "step": 16232, "train_loss": 3.4444408416748047, "perplexity": 31.32576248774275, "lr": 0.0026291804804649314, "grad_norm": 0.159724, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:09.604324+00:00", "epoch": 0, "step": 16233, "train_loss": 3.4202470779418945, "perplexity": 30.576968982368708, "lr": 0.0026291804804649314, "grad_norm": 0.162732, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:09.909762+00:00", "epoch": 0, "step": 16234, "train_loss": 3.697721242904663, "perplexity": 40.355239714590624, "lr": 0.0026291804804649314, "grad_norm": 0.153257, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:10.214463+00:00", "epoch": 0, "step": 16235, "train_loss": 3.6015796661376953, "perplexity": 36.656093121994104, "lr": 0.0026291804804649314, "grad_norm": 0.159776, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:10.519196+00:00", "epoch": 0, "step": 16236, "train_loss": 3.5136220455169678, "perplexity": 33.56963860495036, "lr": 0.0026291804804649314, "grad_norm": 0.17809, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:10.823597+00:00", "epoch": 0, "step": 16237, "train_loss": 3.4968221187591553, "perplexity": 33.01038202344326, "lr": 0.0026291804804649314, "grad_norm": 0.163838, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:11.129297+00:00", "epoch": 0, "step": 16238, "train_loss": 3.4763948917388916, "perplexity": 32.34291194674873, "lr": 0.0026291804804649314, "grad_norm": 0.171766, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:11.434954+00:00", "epoch": 0, "step": 16239, "train_loss": 3.5667359828948975, "perplexity": 35.400855371867, "lr": 0.0026291804804649314, "grad_norm": 0.161376, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:11.740224+00:00", "epoch": 0, "step": 16240, "train_loss": 3.4092118740081787, "perplexity": 30.241400830797517, "lr": 0.0026291804804649314, "grad_norm": 0.170132, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:12.044616+00:00", "epoch": 0, "step": 16241, "train_loss": 3.4566900730133057, "perplexity": 31.711838740631986, "lr": 0.0026291804804649314, "grad_norm": 0.195708, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:12.348232+00:00", "epoch": 0, "step": 16242, "train_loss": 3.57100772857666, "perplexity": 35.55240227740226, "lr": 0.0026291804804649314, "grad_norm": 0.180071, "tokens_per_sec": 107925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:12.652405+00:00", "epoch": 0, "step": 16243, "train_loss": 3.438840627670288, "perplexity": 31.150821823747936, "lr": 0.0026291804804649314, "grad_norm": 0.14301, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:12.958176+00:00", "epoch": 0, "step": 16244, "train_loss": 3.550522804260254, "perplexity": 34.83152279679005, "lr": 0.0026291804804649314, "grad_norm": 0.170942, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:13.264657+00:00", "epoch": 0, "step": 16245, "train_loss": 3.615352153778076, "perplexity": 37.16443121635691, "lr": 0.0026291804804649314, "grad_norm": 0.153051, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:13.569278+00:00", "epoch": 0, "step": 16246, "train_loss": 3.478271245956421, "perplexity": 32.40365567654225, "lr": 0.0026291804804649314, "grad_norm": 0.170707, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:13.873341+00:00", "epoch": 0, "step": 16247, "train_loss": 3.596036672592163, "perplexity": 36.45347072053993, "lr": 0.0026291804804649314, "grad_norm": 0.183941, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:14.178050+00:00", "epoch": 0, "step": 16248, "train_loss": 3.4969735145568848, "perplexity": 33.01538003489248, "lr": 0.0026291804804649314, "grad_norm": 0.156559, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:14.483481+00:00", "epoch": 0, "step": 16249, "train_loss": 3.5626208782196045, "perplexity": 35.25547647625821, "lr": 0.0026291804804649314, "grad_norm": 0.162034, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:14.787972+00:00", "epoch": 0, "step": 16250, "train_loss": 3.514068841934204, "perplexity": 33.584640750409825, "lr": 0.0026291804804649314, "grad_norm": 0.184113, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:15.092300+00:00", "epoch": 0, "step": 16251, "train_loss": 3.5452632904052734, "perplexity": 34.64880683980186, "lr": 0.0026291804804649314, "grad_norm": 0.173345, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:15.397125+00:00", "epoch": 0, "step": 16252, "train_loss": 3.481597661972046, "perplexity": 32.51162318862477, "lr": 0.0026291804804649314, "grad_norm": 0.18252, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:15.700975+00:00", "epoch": 0, "step": 16253, "train_loss": 3.522840976715088, "perplexity": 33.88054570722504, "lr": 0.0026291804804649314, "grad_norm": 0.204712, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:16.005675+00:00", "epoch": 0, "step": 16254, "train_loss": 3.5091679096221924, "perplexity": 33.42044737831865, "lr": 0.0026291804804649314, "grad_norm": 0.185299, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:16.310053+00:00", "epoch": 0, "step": 16255, "train_loss": 3.43394136428833, "perplexity": 30.99857898646872, "lr": 0.0026291804804649314, "grad_norm": 0.172039, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:16.614977+00:00", "epoch": 0, "step": 16256, "train_loss": 3.546609878540039, "perplexity": 34.69549594041546, "lr": 0.0026291804804649314, "grad_norm": 0.189269, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:16.919624+00:00", "epoch": 0, "step": 16257, "train_loss": 3.408195972442627, "perplexity": 30.210694144475305, "lr": 0.0026291804804649314, "grad_norm": 0.169647, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:17.223761+00:00", "epoch": 0, "step": 16258, "train_loss": 3.5211780071258545, "perplexity": 33.8242502118693, "lr": 0.0026291804804649314, "grad_norm": 0.16017, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:17.528909+00:00", "epoch": 0, "step": 16259, "train_loss": 3.5961475372314453, "perplexity": 36.45751234545446, "lr": 0.0026291804804649314, "grad_norm": 0.208956, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:17.834846+00:00", "epoch": 0, "step": 16260, "train_loss": 3.4230270385742188, "perplexity": 30.662090014134918, "lr": 0.0026291804804649314, "grad_norm": 0.204053, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:18.140402+00:00", "epoch": 0, "step": 16261, "train_loss": 3.6375622749328613, "perplexity": 37.99909238868517, "lr": 0.0026291804804649314, "grad_norm": 0.172171, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:18.445112+00:00", "epoch": 0, "step": 16262, "train_loss": 3.505513906478882, "perplexity": 33.298551797591465, "lr": 0.0026291804804649314, "grad_norm": 0.170664, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:18.749708+00:00", "epoch": 0, "step": 16263, "train_loss": 3.595186233520508, "perplexity": 36.42248244343023, "lr": 0.0026291804804649314, "grad_norm": 0.177986, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:19.054744+00:00", "epoch": 0, "step": 16264, "train_loss": 3.6547439098358154, "perplexity": 38.65762001117218, "lr": 0.0026291804804649314, "grad_norm": 0.210367, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:19.360321+00:00", "epoch": 0, "step": 16265, "train_loss": 3.596531629562378, "perplexity": 36.4715180859278, "lr": 0.0026291804804649314, "grad_norm": 0.186042, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:19.666282+00:00", "epoch": 0, "step": 16266, "train_loss": 3.51981782913208, "perplexity": 33.7782744857489, "lr": 0.0026291804804649314, "grad_norm": 0.163934, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:19.971100+00:00", "epoch": 0, "step": 16267, "train_loss": 3.4223673343658447, "perplexity": 30.64186877506668, "lr": 0.0026291804804649314, "grad_norm": 0.15223, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:20.275457+00:00", "epoch": 0, "step": 16268, "train_loss": 3.475306510925293, "perplexity": 32.30772969124891, "lr": 0.0026291804804649314, "grad_norm": 0.155702, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:20.579089+00:00", "epoch": 0, "step": 16269, "train_loss": 3.54613995552063, "perplexity": 34.67919555846556, "lr": 0.0026291804804649314, "grad_norm": 0.153656, "tokens_per_sec": 107921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:20.884938+00:00", "epoch": 0, "step": 16270, "train_loss": 3.4694533348083496, "perplexity": 32.11917920732824, "lr": 0.0026291804804649314, "grad_norm": 0.155048, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:21.189733+00:00", "epoch": 0, "step": 16271, "train_loss": 3.501878023147583, "perplexity": 33.17770197920698, "lr": 0.0026291804804649314, "grad_norm": 0.166697, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:21.495737+00:00", "epoch": 0, "step": 16272, "train_loss": 3.5489978790283203, "perplexity": 34.778447806791306, "lr": 0.0026291804804649314, "grad_norm": 0.16139, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:21.800726+00:00", "epoch": 0, "step": 16273, "train_loss": 3.4964048862457275, "perplexity": 32.996611891655526, "lr": 0.0026291804804649314, "grad_norm": 0.157209, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:22.104890+00:00", "epoch": 0, "step": 16274, "train_loss": 3.462409257888794, "perplexity": 31.89372423173096, "lr": 0.0026291804804649314, "grad_norm": 0.167446, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:22.410381+00:00", "epoch": 0, "step": 16275, "train_loss": 3.5184459686279297, "perplexity": 33.731967175930784, "lr": 0.0026291804804649314, "grad_norm": 0.150396, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:22.716019+00:00", "epoch": 0, "step": 16276, "train_loss": 3.4688494205474854, "perplexity": 32.099787832910806, "lr": 0.0026291804804649314, "grad_norm": 0.157556, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:23.021823+00:00", "epoch": 0, "step": 16277, "train_loss": 3.5158803462982178, "perplexity": 33.64553461178866, "lr": 0.0026291804804649314, "grad_norm": 0.177443, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:23.326357+00:00", "epoch": 0, "step": 16278, "train_loss": 3.444674015045166, "perplexity": 31.333067673015393, "lr": 0.0026291804804649314, "grad_norm": 0.178018, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:23.631037+00:00", "epoch": 0, "step": 16279, "train_loss": 3.526029348373413, "perplexity": 33.98874187206171, "lr": 0.0026291804804649314, "grad_norm": 0.172748, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:23.935385+00:00", "epoch": 0, "step": 16280, "train_loss": 3.5226640701293945, "perplexity": 33.87455254569305, "lr": 0.0026291804804649314, "grad_norm": 0.19094, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:24.240606+00:00", "epoch": 0, "step": 16281, "train_loss": 3.491595506668091, "perplexity": 32.83829965725455, "lr": 0.0026291804804649314, "grad_norm": 0.179769, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:24.546562+00:00", "epoch": 0, "step": 16282, "train_loss": 3.5619325637817383, "perplexity": 35.231217972482696, "lr": 0.0026291804804649314, "grad_norm": 0.153949, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:24.852780+00:00", "epoch": 0, "step": 16283, "train_loss": 3.546355724334717, "perplexity": 34.68667905468822, "lr": 0.0026291804804649314, "grad_norm": 0.176886, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:25.157985+00:00", "epoch": 0, "step": 16284, "train_loss": 3.428339719772339, "perplexity": 30.825421402883954, "lr": 0.0026291804804649314, "grad_norm": 0.155136, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:25.462440+00:00", "epoch": 0, "step": 16285, "train_loss": 3.489095449447632, "perplexity": 32.756304568004026, "lr": 0.0026291804804649314, "grad_norm": 0.179123, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:25.767306+00:00", "epoch": 0, "step": 16286, "train_loss": 3.497363328933716, "perplexity": 33.02825241344273, "lr": 0.0026291804804649314, "grad_norm": 0.193337, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:26.073058+00:00", "epoch": 0, "step": 16287, "train_loss": 3.641472101211548, "perplexity": 38.14795305871964, "lr": 0.0026291804804649314, "grad_norm": 0.194418, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:26.378046+00:00", "epoch": 0, "step": 16288, "train_loss": 3.4018213748931885, "perplexity": 30.01872563865928, "lr": 0.0026291804804649314, "grad_norm": 0.169891, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:26.682367+00:00", "epoch": 0, "step": 16289, "train_loss": 3.541572093963623, "perplexity": 34.52114704093131, "lr": 0.0026291804804649314, "grad_norm": 0.190389, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:26.988895+00:00", "epoch": 0, "step": 16290, "train_loss": 3.528768539428711, "perplexity": 34.08197115789851, "lr": 0.0026291804804649314, "grad_norm": 0.17231, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:27.293741+00:00", "epoch": 0, "step": 16291, "train_loss": 3.456237316131592, "perplexity": 31.69748423720568, "lr": 0.0026291804804649314, "grad_norm": 0.171248, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:27.598410+00:00", "epoch": 0, "step": 16292, "train_loss": 3.5529022216796875, "perplexity": 34.91450020867884, "lr": 0.0026291804804649314, "grad_norm": 0.179354, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:27.902652+00:00", "epoch": 0, "step": 16293, "train_loss": 3.606423854827881, "perplexity": 36.83409293798887, "lr": 0.0026291804804649314, "grad_norm": 0.175025, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:28.207647+00:00", "epoch": 0, "step": 16294, "train_loss": 3.5096616744995117, "perplexity": 33.43695329610425, "lr": 0.0026291804804649314, "grad_norm": 0.193609, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:28.512837+00:00", "epoch": 0, "step": 16295, "train_loss": 3.4599575996398926, "perplexity": 31.815627492005664, "lr": 0.0026291804804649314, "grad_norm": 0.181094, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:28.818482+00:00", "epoch": 0, "step": 16296, "train_loss": 3.56122088432312, "perplexity": 35.20615355832184, "lr": 0.0026291804804649314, "grad_norm": 0.202066, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:29.125547+00:00", "epoch": 0, "step": 16297, "train_loss": 3.45219087600708, "perplexity": 31.56948141767104, "lr": 0.0026291804804649314, "grad_norm": 0.165351, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:29.431118+00:00", "epoch": 0, "step": 16298, "train_loss": 3.4970974922180176, "perplexity": 33.0194734582319, "lr": 0.0026291804804649314, "grad_norm": 0.200705, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:29.736524+00:00", "epoch": 0, "step": 16299, "train_loss": 3.541304588317871, "perplexity": 34.51191367424395, "lr": 0.0026291804804649314, "grad_norm": 0.190146, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:30.041779+00:00", "epoch": 0, "step": 16300, "train_loss": 3.580622434616089, "perplexity": 35.89587673128028, "lr": 0.0026291804804649314, "grad_norm": 0.184548, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:30.346267+00:00", "epoch": 0, "step": 16301, "train_loss": 3.6552488803863525, "perplexity": 38.67714590041628, "lr": 0.0026291804804649314, "grad_norm": 0.178638, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:30.652795+00:00", "epoch": 0, "step": 16302, "train_loss": 3.4723050594329834, "perplexity": 32.21090498776815, "lr": 0.0026291804804649314, "grad_norm": 0.164364, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:30.958255+00:00", "epoch": 0, "step": 16303, "train_loss": 3.416684150695801, "perplexity": 30.46821931509781, "lr": 0.0026291804804649314, "grad_norm": 0.181807, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:31.263267+00:00", "epoch": 0, "step": 16304, "train_loss": 3.4447696208953857, "perplexity": 31.336063440794444, "lr": 0.0026291804804649314, "grad_norm": 0.186115, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:31.568876+00:00", "epoch": 0, "step": 16305, "train_loss": 3.5433428287506104, "perplexity": 34.582328989399294, "lr": 0.0026291804804649314, "grad_norm": 0.164434, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:31.873176+00:00", "epoch": 0, "step": 16306, "train_loss": 3.590095043182373, "perplexity": 36.237519892022995, "lr": 0.0026291804804649314, "grad_norm": 0.149245, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:32.180400+00:00", "epoch": 0, "step": 16307, "train_loss": 3.4685356616973877, "perplexity": 32.08971782025238, "lr": 0.0026291804804649314, "grad_norm": 0.174847, "tokens_per_sec": 106659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:32.487532+00:00", "epoch": 0, "step": 16308, "train_loss": 3.4627623558044434, "perplexity": 31.90498782773859, "lr": 0.0026291804804649314, "grad_norm": 0.150987, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:32.793752+00:00", "epoch": 0, "step": 16309, "train_loss": 3.4897408485412598, "perplexity": 32.77745228090611, "lr": 0.0026291804804649314, "grad_norm": 0.167078, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:33.100117+00:00", "epoch": 0, "step": 16310, "train_loss": 3.485980272293091, "perplexity": 32.65442165040736, "lr": 0.0026291804804649314, "grad_norm": 0.170052, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:33.405478+00:00", "epoch": 0, "step": 16311, "train_loss": 3.5159695148468018, "perplexity": 33.64853486903863, "lr": 0.0026291804804649314, "grad_norm": 0.159547, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:33.711163+00:00", "epoch": 0, "step": 16312, "train_loss": 3.5016815662384033, "perplexity": 33.171184630632496, "lr": 0.0026291804804649314, "grad_norm": 0.200665, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:34.018320+00:00", "epoch": 0, "step": 16313, "train_loss": 3.4778573513031006, "perplexity": 32.39024675184224, "lr": 0.0026291804804649314, "grad_norm": 0.190407, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:34.322964+00:00", "epoch": 0, "step": 16314, "train_loss": 3.5327043533325195, "perplexity": 34.21637577613258, "lr": 0.0026291804804649314, "grad_norm": 0.188964, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:34.627757+00:00", "epoch": 0, "step": 16315, "train_loss": 3.639259099960327, "perplexity": 38.06362493440702, "lr": 0.0026291804804649314, "grad_norm": 0.190524, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:34.933841+00:00", "epoch": 0, "step": 16316, "train_loss": 3.4864509105682373, "perplexity": 32.669793688139734, "lr": 0.0026291804804649314, "grad_norm": 0.164788, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:35.240793+00:00", "epoch": 0, "step": 16317, "train_loss": 3.4822065830230713, "perplexity": 32.53142622901952, "lr": 0.0026291804804649314, "grad_norm": 0.156513, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:35.547879+00:00", "epoch": 0, "step": 16318, "train_loss": 3.4099795818328857, "perplexity": 30.26462630489381, "lr": 0.0026291804804649314, "grad_norm": 0.172684, "tokens_per_sec": 106708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:35.852588+00:00", "epoch": 0, "step": 16319, "train_loss": 3.5046722888946533, "perplexity": 33.27053894057929, "lr": 0.0026291804804649314, "grad_norm": 0.157202, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:36.157866+00:00", "epoch": 0, "step": 16320, "train_loss": 3.4680001735687256, "perplexity": 32.072538757310156, "lr": 0.0026291804804649314, "grad_norm": 0.165053, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:36.463009+00:00", "epoch": 0, "step": 16321, "train_loss": 3.5179388523101807, "perplexity": 33.71486548158672, "lr": 0.0026291804804649314, "grad_norm": 0.164412, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:36.767660+00:00", "epoch": 0, "step": 16322, "train_loss": 3.435119152069092, "perplexity": 31.035110242830672, "lr": 0.0026291804804649314, "grad_norm": 0.165034, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:37.073073+00:00", "epoch": 0, "step": 16323, "train_loss": 3.5215978622436523, "perplexity": 33.8384544780847, "lr": 0.0026291804804649314, "grad_norm": 0.145934, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:37.378386+00:00", "epoch": 0, "step": 16324, "train_loss": 3.382438898086548, "perplexity": 29.442490853776576, "lr": 0.0026291804804649314, "grad_norm": 0.161639, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:37.683305+00:00", "epoch": 0, "step": 16325, "train_loss": 3.5566322803497314, "perplexity": 35.04497653378626, "lr": 0.0026291804804649314, "grad_norm": 0.170679, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:37.989299+00:00", "epoch": 0, "step": 16326, "train_loss": 3.5287108421325684, "perplexity": 34.08000477704346, "lr": 0.0026291804804649314, "grad_norm": 0.14948, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:38.294690+00:00", "epoch": 0, "step": 16327, "train_loss": 3.395780324935913, "perplexity": 29.83792767203737, "lr": 0.0026291804804649314, "grad_norm": 0.164286, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:38.599551+00:00", "epoch": 0, "step": 16328, "train_loss": 3.5921716690063477, "perplexity": 36.31284985062533, "lr": 0.0026291804804649314, "grad_norm": 0.16165, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:38.903933+00:00", "epoch": 0, "step": 16329, "train_loss": 3.5452637672424316, "perplexity": 34.64882336164438, "lr": 0.0026291804804649314, "grad_norm": 0.155146, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:39.209067+00:00", "epoch": 0, "step": 16330, "train_loss": 3.5436322689056396, "perplexity": 34.59233995278077, "lr": 0.0026291804804649314, "grad_norm": 0.227477, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:39.514362+00:00", "epoch": 0, "step": 16331, "train_loss": 3.4739267826080322, "perplexity": 32.26318453889807, "lr": 0.0026291804804649314, "grad_norm": 0.228086, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:39.819777+00:00", "epoch": 0, "step": 16332, "train_loss": 3.550661325454712, "perplexity": 34.83634803512383, "lr": 0.0026291804804649314, "grad_norm": 0.180247, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:40.124065+00:00", "epoch": 0, "step": 16333, "train_loss": 3.573948383331299, "perplexity": 35.657103487804356, "lr": 0.0026291804804649314, "grad_norm": 0.231452, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:40.428343+00:00", "epoch": 0, "step": 16334, "train_loss": 3.414196252822876, "perplexity": 30.39251171251389, "lr": 0.0026291804804649314, "grad_norm": 0.189332, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:40.733334+00:00", "epoch": 0, "step": 16335, "train_loss": 3.4651176929473877, "perplexity": 31.980223398399993, "lr": 0.0026291804804649314, "grad_norm": 0.174785, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:41.038578+00:00", "epoch": 0, "step": 16336, "train_loss": 3.35819673538208, "perplexity": 28.737323129645564, "lr": 0.0026291804804649314, "grad_norm": 0.204403, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:41.343017+00:00", "epoch": 0, "step": 16337, "train_loss": 3.5561161041259766, "perplexity": 35.02689181798916, "lr": 0.0026291804804649314, "grad_norm": 0.161586, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:41.647886+00:00", "epoch": 0, "step": 16338, "train_loss": 3.6105830669403076, "perplexity": 36.98761278297402, "lr": 0.0026291804804649314, "grad_norm": 0.162191, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:41.951861+00:00", "epoch": 0, "step": 16339, "train_loss": 3.4500539302825928, "perplexity": 31.50209117961681, "lr": 0.0026291804804649314, "grad_norm": 0.164938, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:42.256329+00:00", "epoch": 0, "step": 16340, "train_loss": 3.479557752609253, "perplexity": 32.445370022284415, "lr": 0.0026291804804649314, "grad_norm": 0.15829, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:42.561390+00:00", "epoch": 0, "step": 16341, "train_loss": 3.4976656436920166, "perplexity": 33.03823885103587, "lr": 0.0026291804804649314, "grad_norm": 0.157887, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:42.867132+00:00", "epoch": 0, "step": 16342, "train_loss": 3.565861463546753, "perplexity": 35.36991017196366, "lr": 0.0026291804804649314, "grad_norm": 0.151703, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:43.173118+00:00", "epoch": 0, "step": 16343, "train_loss": 3.5426199436187744, "perplexity": 34.55733897149303, "lr": 0.0026291804804649314, "grad_norm": 0.168538, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:43.477945+00:00", "epoch": 0, "step": 16344, "train_loss": 3.5922787189483643, "perplexity": 36.31673734717072, "lr": 0.0026291804804649314, "grad_norm": 0.207209, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:43.781761+00:00", "epoch": 0, "step": 16345, "train_loss": 3.467956781387329, "perplexity": 32.07114709008451, "lr": 0.0026291804804649314, "grad_norm": 0.196634, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:44.086758+00:00", "epoch": 0, "step": 16346, "train_loss": 3.5192317962646484, "perplexity": 33.758485105875, "lr": 0.0026291804804649314, "grad_norm": 0.180508, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:44.393710+00:00", "epoch": 0, "step": 16347, "train_loss": 3.531120777130127, "perplexity": 34.16223441751916, "lr": 0.0026291804804649314, "grad_norm": 0.163431, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:44.699706+00:00", "epoch": 0, "step": 16348, "train_loss": 3.466766357421875, "perplexity": 32.032991543133505, "lr": 0.0026291804804649314, "grad_norm": 0.185222, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:45.004706+00:00", "epoch": 0, "step": 16349, "train_loss": 3.594672918319702, "perplexity": 36.40379102724539, "lr": 0.0026291804804649314, "grad_norm": 0.19178, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:45.309220+00:00", "epoch": 0, "step": 16350, "train_loss": 3.5799190998077393, "perplexity": 35.8706387881013, "lr": 0.0026291804804649314, "grad_norm": 0.162663, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:45.613678+00:00", "epoch": 0, "step": 16351, "train_loss": 3.5852408409118652, "perplexity": 36.0620418881177, "lr": 0.0026291804804649314, "grad_norm": 0.161789, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:45.918280+00:00", "epoch": 0, "step": 16352, "train_loss": 3.475914478302002, "perplexity": 32.32737770898842, "lr": 0.0026291804804649314, "grad_norm": 0.152321, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:46.223767+00:00", "epoch": 0, "step": 16353, "train_loss": 3.493119716644287, "perplexity": 32.888390285813735, "lr": 0.0026291804804649314, "grad_norm": 0.158048, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:46.529471+00:00", "epoch": 0, "step": 16354, "train_loss": 3.5284957885742188, "perplexity": 34.07267653875769, "lr": 0.0026291804804649314, "grad_norm": 0.197566, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:46.833631+00:00", "epoch": 0, "step": 16355, "train_loss": 3.5159597396850586, "perplexity": 33.648205950775484, "lr": 0.0026291804804649314, "grad_norm": 0.186805, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:47.137666+00:00", "epoch": 0, "step": 16356, "train_loss": 3.5249006748199463, "perplexity": 33.95040131904939, "lr": 0.0026291804804649314, "grad_norm": 0.168188, "tokens_per_sec": 107778} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:47.442668+00:00", "epoch": 0, "step": 16357, "train_loss": 3.4381120204925537, "perplexity": 31.128133377860806, "lr": 0.0026291804804649314, "grad_norm": 0.163556, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:47.748321+00:00", "epoch": 0, "step": 16358, "train_loss": 3.475616216659546, "perplexity": 32.31773712999557, "lr": 0.0026291804804649314, "grad_norm": 0.146649, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:48.054039+00:00", "epoch": 0, "step": 16359, "train_loss": 3.5237839221954346, "perplexity": 33.912508281785676, "lr": 0.0026291804804649314, "grad_norm": 0.154959, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:48.359098+00:00", "epoch": 0, "step": 16360, "train_loss": 3.629067897796631, "perplexity": 37.677680795349964, "lr": 0.0026291804804649314, "grad_norm": 0.160589, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:48.663278+00:00", "epoch": 0, "step": 16361, "train_loss": 3.457192897796631, "perplexity": 31.727788248643495, "lr": 0.0026291804804649314, "grad_norm": 0.167614, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:48.967628+00:00", "epoch": 0, "step": 16362, "train_loss": 3.488407850265503, "perplexity": 32.73378910146879, "lr": 0.0026291804804649314, "grad_norm": 0.167916, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:49.272631+00:00", "epoch": 0, "step": 16363, "train_loss": 3.6316046714782715, "perplexity": 37.773381879050625, "lr": 0.0026291804804649314, "grad_norm": 0.157958, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:49.578229+00:00", "epoch": 0, "step": 16364, "train_loss": 3.490635871887207, "perplexity": 32.806801998300074, "lr": 0.0026291804804649314, "grad_norm": 0.154073, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:49.883627+00:00", "epoch": 0, "step": 16365, "train_loss": 3.4017767906188965, "perplexity": 30.017387305396042, "lr": 0.0026291804804649314, "grad_norm": 0.156736, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:50.188196+00:00", "epoch": 0, "step": 16366, "train_loss": 3.396771192550659, "perplexity": 29.867507760814544, "lr": 0.0026291804804649314, "grad_norm": 0.154573, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:50.492538+00:00", "epoch": 0, "step": 16367, "train_loss": 3.4534120559692383, "perplexity": 31.608056984854947, "lr": 0.0026291804804649314, "grad_norm": 0.15458, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:50.798072+00:00", "epoch": 0, "step": 16368, "train_loss": 3.492924690246582, "perplexity": 32.88197680694914, "lr": 0.0026291804804649314, "grad_norm": 0.169399, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:51.103427+00:00", "epoch": 0, "step": 16369, "train_loss": 3.326054096221924, "perplexity": 27.828316911535715, "lr": 0.0026291804804649314, "grad_norm": 0.178859, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:51.409284+00:00", "epoch": 0, "step": 16370, "train_loss": 3.439417600631714, "perplexity": 31.16880019168378, "lr": 0.0026291804804649314, "grad_norm": 0.160097, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:51.713426+00:00", "epoch": 0, "step": 16371, "train_loss": 3.465378522872925, "perplexity": 31.98856588562552, "lr": 0.0026291804804649314, "grad_norm": 0.160101, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:52.017918+00:00", "epoch": 0, "step": 16372, "train_loss": 3.53629469871521, "perplexity": 34.33944518154558, "lr": 0.0026291804804649314, "grad_norm": 0.15774, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:52.322831+00:00", "epoch": 0, "step": 16373, "train_loss": 3.6734118461608887, "perplexity": 39.386056044383395, "lr": 0.0026291804804649314, "grad_norm": 0.182539, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:52.628427+00:00", "epoch": 0, "step": 16374, "train_loss": 3.4357471466064453, "perplexity": 31.054606243579588, "lr": 0.0026291804804649314, "grad_norm": 0.165356, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:52.934596+00:00", "epoch": 0, "step": 16375, "train_loss": 3.438004732131958, "perplexity": 31.124793870610624, "lr": 0.0026291804804649314, "grad_norm": 0.169085, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:53.240320+00:00", "epoch": 0, "step": 16376, "train_loss": 3.547483205795288, "perplexity": 34.725809697643115, "lr": 0.0026291804804649314, "grad_norm": 0.154488, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:53.545333+00:00", "epoch": 0, "step": 16377, "train_loss": 3.494035243988037, "perplexity": 32.91851429398477, "lr": 0.0026291804804649314, "grad_norm": 0.16744, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:53.849246+00:00", "epoch": 0, "step": 16378, "train_loss": 3.526329755783081, "perplexity": 33.99895387576936, "lr": 0.0026291804804649314, "grad_norm": 0.183654, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:54.154302+00:00", "epoch": 0, "step": 16379, "train_loss": 3.6172587871551514, "perplexity": 37.235357755320976, "lr": 0.0026291804804649314, "grad_norm": 0.170777, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:54.459903+00:00", "epoch": 0, "step": 16380, "train_loss": 3.6952898502349854, "perplexity": 40.257239467385574, "lr": 0.0026291804804649314, "grad_norm": 0.17337, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:54.766112+00:00", "epoch": 0, "step": 16381, "train_loss": 3.472477912902832, "perplexity": 32.216473235693876, "lr": 0.0026291804804649314, "grad_norm": 0.161782, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:55.071116+00:00", "epoch": 0, "step": 16382, "train_loss": 3.474508285522461, "perplexity": 32.28195113062125, "lr": 0.0026291804804649314, "grad_norm": 0.184402, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:55.375976+00:00", "epoch": 0, "step": 16383, "train_loss": 3.511643648147583, "perplexity": 33.50329017366504, "lr": 0.0026291804804649314, "grad_norm": 0.196766, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:55.681437+00:00", "epoch": 0, "step": 16384, "train_loss": 3.5357563495635986, "perplexity": 34.32096354559817, "lr": 0.0026291804804649314, "grad_norm": 0.175242, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:55.985868+00:00", "epoch": 0, "step": 16385, "train_loss": 3.6008265018463135, "perplexity": 36.62849545568603, "lr": 0.0026291804804649314, "grad_norm": 0.17326, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:56.291241+00:00", "epoch": 0, "step": 16386, "train_loss": 3.559760570526123, "perplexity": 35.154779047124926, "lr": 0.0026291804804649314, "grad_norm": 0.163213, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:56.596741+00:00", "epoch": 0, "step": 16387, "train_loss": 3.5552237033843994, "perplexity": 34.9956477369497, "lr": 0.0026291804804649314, "grad_norm": 0.176484, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:56.901218+00:00", "epoch": 0, "step": 16388, "train_loss": 3.5469231605529785, "perplexity": 34.706367118006895, "lr": 0.0026291804804649314, "grad_norm": 0.166665, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:57.206047+00:00", "epoch": 0, "step": 16389, "train_loss": 3.5340561866760254, "perplexity": 34.26266189230947, "lr": 0.0026291804804649314, "grad_norm": 0.149212, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:57.511405+00:00", "epoch": 0, "step": 16390, "train_loss": 3.5484206676483154, "perplexity": 34.758379083443266, "lr": 0.0026291804804649314, "grad_norm": 0.148933, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:57.817172+00:00", "epoch": 0, "step": 16391, "train_loss": 3.430799961090088, "perplexity": 30.901352744668856, "lr": 0.0026291804804649314, "grad_norm": 0.165771, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:58.122804+00:00", "epoch": 0, "step": 16392, "train_loss": 3.455751657485962, "perplexity": 31.682093817488862, "lr": 0.0026291804804649314, "grad_norm": 0.163194, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:58.427918+00:00", "epoch": 0, "step": 16393, "train_loss": 3.5807254314422607, "perplexity": 35.89957408306074, "lr": 0.0026291804804649314, "grad_norm": 0.192894, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:58.733449+00:00", "epoch": 0, "step": 16394, "train_loss": 3.456326961517334, "perplexity": 31.700325897776153, "lr": 0.0026291804804649314, "grad_norm": 0.15222, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:59.039710+00:00", "epoch": 0, "step": 16395, "train_loss": 3.5374252796173096, "perplexity": 34.37829065729531, "lr": 0.0026291804804649314, "grad_norm": 0.219328, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:59.344810+00:00", "epoch": 0, "step": 16396, "train_loss": 3.6028809547424316, "perplexity": 36.703824327569386, "lr": 0.0026291804804649314, "grad_norm": 0.186148, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:59.651464+00:00", "epoch": 0, "step": 16397, "train_loss": 3.550780773162842, "perplexity": 34.84050940558437, "lr": 0.0026291804804649314, "grad_norm": 0.18283, "tokens_per_sec": 106857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:23:59.956295+00:00", "epoch": 0, "step": 16398, "train_loss": 3.473074436187744, "perplexity": 32.23569684523203, "lr": 0.0026291804804649314, "grad_norm": 0.232543, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:00.260731+00:00", "epoch": 0, "step": 16399, "train_loss": 3.5436835289001465, "perplexity": 34.594113201384694, "lr": 0.0026291804804649314, "grad_norm": 0.203856, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:00.564837+00:00", "epoch": 0, "step": 16400, "train_loss": 3.5293493270874023, "perplexity": 34.10177129541208, "lr": 0.0026291804804649314, "grad_norm": 0.192424, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:00.870060+00:00", "epoch": 0, "step": 16401, "train_loss": 3.512507200241089, "perplexity": 33.53223450570441, "lr": 0.0026291804804649314, "grad_norm": 0.203005, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:01.175651+00:00", "epoch": 0, "step": 16402, "train_loss": 3.459251642227173, "perplexity": 31.79317494014255, "lr": 0.0026291804804649314, "grad_norm": 0.230961, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:01.480624+00:00", "epoch": 0, "step": 16403, "train_loss": 3.5181260108947754, "perplexity": 33.72117609861471, "lr": 0.0026291804804649314, "grad_norm": 0.184594, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:01.785270+00:00", "epoch": 0, "step": 16404, "train_loss": 3.5075860023498535, "perplexity": 33.36762112370042, "lr": 0.0026291804804649314, "grad_norm": 0.167696, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:02.091029+00:00", "epoch": 0, "step": 16405, "train_loss": 3.5340800285339355, "perplexity": 34.26347878756404, "lr": 0.0026291804804649314, "grad_norm": 0.178103, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:02.395484+00:00", "epoch": 0, "step": 16406, "train_loss": 3.431347370147705, "perplexity": 30.918273055797172, "lr": 0.0026291804804649314, "grad_norm": 0.175299, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:02.700917+00:00", "epoch": 0, "step": 16407, "train_loss": 3.4798049926757812, "perplexity": 32.45339280946361, "lr": 0.0026291804804649314, "grad_norm": 0.167508, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:03.007397+00:00", "epoch": 0, "step": 16408, "train_loss": 3.609502077102661, "perplexity": 36.9476511523865, "lr": 0.0026291804804649314, "grad_norm": 0.165198, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:03.313632+00:00", "epoch": 0, "step": 16409, "train_loss": 3.5243823528289795, "perplexity": 33.93280863917388, "lr": 0.0026291804804649314, "grad_norm": 0.169647, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:03.618641+00:00", "epoch": 0, "step": 16410, "train_loss": 3.528115749359131, "perplexity": 34.059730045753305, "lr": 0.0026291804804649314, "grad_norm": 0.168981, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:03.924078+00:00", "epoch": 0, "step": 16411, "train_loss": 3.5215678215026855, "perplexity": 33.83743796110755, "lr": 0.0026291804804649314, "grad_norm": 0.168861, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:04.229209+00:00", "epoch": 0, "step": 16412, "train_loss": 3.5299160480499268, "perplexity": 34.12110296137717, "lr": 0.0026291804804649314, "grad_norm": 0.146483, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:04.535684+00:00", "epoch": 0, "step": 16413, "train_loss": 3.464346170425415, "perplexity": 31.95555945140913, "lr": 0.0026291804804649314, "grad_norm": 0.15147, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:04.840751+00:00", "epoch": 0, "step": 16414, "train_loss": 3.4983348846435547, "perplexity": 33.060356793732666, "lr": 0.0026291804804649314, "grad_norm": 0.153109, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:05.145495+00:00", "epoch": 0, "step": 16415, "train_loss": 3.4590978622436523, "perplexity": 31.788286162131925, "lr": 0.0026291804804649314, "grad_norm": 0.171351, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:05.450920+00:00", "epoch": 0, "step": 16416, "train_loss": 3.4982969760894775, "perplexity": 33.05910354716387, "lr": 0.0026291804804649314, "grad_norm": 0.16872, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:05.756692+00:00", "epoch": 0, "step": 16417, "train_loss": 3.5359292030334473, "perplexity": 34.32689655599152, "lr": 0.0026291804804649314, "grad_norm": 0.173361, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:06.063235+00:00", "epoch": 0, "step": 16418, "train_loss": 3.4324586391448975, "perplexity": 30.95265067194252, "lr": 0.0026291804804649314, "grad_norm": 0.178933, "tokens_per_sec": 106895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:06.368370+00:00", "epoch": 0, "step": 16419, "train_loss": 3.5308361053466797, "perplexity": 34.15251077740963, "lr": 0.0026291804804649314, "grad_norm": 0.164544, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:06.673499+00:00", "epoch": 0, "step": 16420, "train_loss": 3.4531424045562744, "perplexity": 31.59953497666535, "lr": 0.0026291804804649314, "grad_norm": 0.167886, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:06.978947+00:00", "epoch": 0, "step": 16421, "train_loss": 3.554042339324951, "perplexity": 34.95432954719372, "lr": 0.0026291804804649314, "grad_norm": 0.16458, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:07.284505+00:00", "epoch": 0, "step": 16422, "train_loss": 3.5045056343078613, "perplexity": 33.26499471465798, "lr": 0.0026291804804649314, "grad_norm": 0.158832, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:07.589972+00:00", "epoch": 0, "step": 16423, "train_loss": 3.6100449562072754, "perplexity": 36.967714705710875, "lr": 0.0026291804804649314, "grad_norm": 0.155121, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:07.895131+00:00", "epoch": 0, "step": 16424, "train_loss": 3.4824273586273193, "perplexity": 32.53860916718186, "lr": 0.0026291804804649314, "grad_norm": 0.167878, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:08.200523+00:00", "epoch": 0, "step": 16425, "train_loss": 3.567863941192627, "perplexity": 35.44080858897579, "lr": 0.0026291804804649314, "grad_norm": 0.161318, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:08.505180+00:00", "epoch": 0, "step": 16426, "train_loss": 3.34507417678833, "perplexity": 28.36267944109635, "lr": 0.0026291804804649314, "grad_norm": 0.204407, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:08.809800+00:00", "epoch": 0, "step": 16427, "train_loss": 3.5520312786102295, "perplexity": 34.88410490490942, "lr": 0.0026291804804649314, "grad_norm": 0.201382, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:09.115267+00:00", "epoch": 0, "step": 16428, "train_loss": 3.492917776107788, "perplexity": 32.881749457183645, "lr": 0.0026291804804649314, "grad_norm": 0.163941, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:09.422104+00:00", "epoch": 0, "step": 16429, "train_loss": 3.590263843536377, "perplexity": 36.24363731450601, "lr": 0.0026291804804649314, "grad_norm": 0.179286, "tokens_per_sec": 106793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:09.727781+00:00", "epoch": 0, "step": 16430, "train_loss": 3.489978790283203, "perplexity": 32.78525233294048, "lr": 0.0026291804804649314, "grad_norm": 0.183304, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:10.032988+00:00", "epoch": 0, "step": 16431, "train_loss": 3.5786378383636475, "perplexity": 35.8247085522507, "lr": 0.0026291804804649314, "grad_norm": 0.174722, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:10.337920+00:00", "epoch": 0, "step": 16432, "train_loss": 3.5423319339752197, "perplexity": 34.54738755773401, "lr": 0.0026291804804649314, "grad_norm": 0.1538, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:10.643772+00:00", "epoch": 0, "step": 16433, "train_loss": 3.450164794921875, "perplexity": 31.505583841194827, "lr": 0.0026291804804649314, "grad_norm": 0.170838, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:10.949047+00:00", "epoch": 0, "step": 16434, "train_loss": 3.534975528717041, "perplexity": 34.29417548148829, "lr": 0.0026291804804649314, "grad_norm": 0.169796, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:11.254241+00:00", "epoch": 0, "step": 16435, "train_loss": 3.5733296871185303, "perplexity": 35.63504939601319, "lr": 0.0026291804804649314, "grad_norm": 0.157271, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:11.559230+00:00", "epoch": 0, "step": 16436, "train_loss": 3.4275732040405273, "perplexity": 30.80180228580777, "lr": 0.0026291804804649314, "grad_norm": 0.180534, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:11.864005+00:00", "epoch": 0, "step": 16437, "train_loss": 3.4981024265289307, "perplexity": 33.05267253869192, "lr": 0.0026291804804649314, "grad_norm": 0.167107, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:12.169222+00:00", "epoch": 0, "step": 16438, "train_loss": 3.5724573135375977, "perplexity": 35.603975876194426, "lr": 0.0026291804804649314, "grad_norm": 0.16459, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:12.475152+00:00", "epoch": 0, "step": 16439, "train_loss": 3.408846855163574, "perplexity": 30.230364164025257, "lr": 0.0026291804804649314, "grad_norm": 0.181205, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:12.780610+00:00", "epoch": 0, "step": 16440, "train_loss": 3.408362627029419, "perplexity": 30.21572931478011, "lr": 0.0026291804804649314, "grad_norm": 0.152447, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:13.085841+00:00", "epoch": 0, "step": 16441, "train_loss": 3.5575239658355713, "perplexity": 35.07623956703809, "lr": 0.0026291804804649314, "grad_norm": 0.186342, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:13.390386+00:00", "epoch": 0, "step": 16442, "train_loss": 3.432877540588379, "perplexity": 30.965619498128934, "lr": 0.0026291804804649314, "grad_norm": 0.194959, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:13.695628+00:00", "epoch": 0, "step": 16443, "train_loss": 3.535097599029541, "perplexity": 34.298362037728154, "lr": 0.0026291804804649314, "grad_norm": 0.195572, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:14.000819+00:00", "epoch": 0, "step": 16444, "train_loss": 3.531994581222534, "perplexity": 34.19209856356529, "lr": 0.0026291804804649314, "grad_norm": 0.174866, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:14.306853+00:00", "epoch": 0, "step": 16445, "train_loss": 3.5257203578948975, "perplexity": 33.97824129681895, "lr": 0.0026291804804649314, "grad_norm": 0.187532, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:14.612546+00:00", "epoch": 0, "step": 16446, "train_loss": 3.4991116523742676, "perplexity": 33.08604698842129, "lr": 0.0026291804804649314, "grad_norm": 0.200383, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:14.917425+00:00", "epoch": 0, "step": 16447, "train_loss": 3.374575614929199, "perplexity": 29.211884062807133, "lr": 0.0026291804804649314, "grad_norm": 0.204196, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:15.222120+00:00", "epoch": 0, "step": 16448, "train_loss": 3.5591375827789307, "perplexity": 35.13288487113312, "lr": 0.0026291804804649314, "grad_norm": 0.207458, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:15.527828+00:00", "epoch": 0, "step": 16449, "train_loss": 3.4933903217315674, "perplexity": 32.8972912558071, "lr": 0.0026291804804649314, "grad_norm": 0.154104, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:15.835248+00:00", "epoch": 0, "step": 16450, "train_loss": 3.5681817531585693, "perplexity": 35.452073892057285, "lr": 0.0026291804804649314, "grad_norm": 0.163723, "tokens_per_sec": 106650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:16.140652+00:00", "epoch": 0, "step": 16451, "train_loss": 3.505653142929077, "perplexity": 33.30318849253138, "lr": 0.0026291804804649314, "grad_norm": 0.156042, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:16.445782+00:00", "epoch": 0, "step": 16452, "train_loss": 3.5060768127441406, "perplexity": 33.317301037559616, "lr": 0.0026291804804649314, "grad_norm": 0.159266, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:16.750342+00:00", "epoch": 0, "step": 16453, "train_loss": 3.486128330230713, "perplexity": 32.659256754660106, "lr": 0.0026291804804649314, "grad_norm": 0.154739, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:17.055176+00:00", "epoch": 0, "step": 16454, "train_loss": 3.506272554397583, "perplexity": 33.323823259467375, "lr": 0.0026291804804649314, "grad_norm": 0.162668, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:17.361655+00:00", "epoch": 0, "step": 16455, "train_loss": 3.486917495727539, "perplexity": 32.6850404857202, "lr": 0.0026291804804649314, "grad_norm": 0.149599, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:17.668093+00:00", "epoch": 0, "step": 16456, "train_loss": 3.3485963344573975, "perplexity": 28.46275340468788, "lr": 0.0026291804804649314, "grad_norm": 0.160628, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:17.973242+00:00", "epoch": 0, "step": 16457, "train_loss": 3.5239439010620117, "perplexity": 33.91793400041197, "lr": 0.0026291804804649314, "grad_norm": 0.160894, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:18.278528+00:00", "epoch": 0, "step": 16458, "train_loss": 3.5885584354400635, "perplexity": 36.18187979784251, "lr": 0.0026291804804649314, "grad_norm": 0.17244, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:18.583689+00:00", "epoch": 0, "step": 16459, "train_loss": 3.47656512260437, "perplexity": 32.34841817729342, "lr": 0.0026291804804649314, "grad_norm": 0.155022, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:18.889374+00:00", "epoch": 0, "step": 16460, "train_loss": 3.629506826400757, "perplexity": 37.694222237178586, "lr": 0.0026291804804649314, "grad_norm": 0.15334, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:19.194678+00:00", "epoch": 0, "step": 16461, "train_loss": 3.587862014770508, "perplexity": 36.15669076098893, "lr": 0.0026291804804649314, "grad_norm": 0.161678, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:19.500458+00:00", "epoch": 0, "step": 16462, "train_loss": 3.4508094787597656, "perplexity": 31.525901530439796, "lr": 0.0026291804804649314, "grad_norm": 0.15799, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:19.805187+00:00", "epoch": 0, "step": 16463, "train_loss": 3.4505367279052734, "perplexity": 31.51730398640577, "lr": 0.0026291804804649314, "grad_norm": 0.163078, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:20.109528+00:00", "epoch": 0, "step": 16464, "train_loss": 3.5308899879455566, "perplexity": 34.15435105302746, "lr": 0.0026291804804649314, "grad_norm": 0.164134, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:20.414428+00:00", "epoch": 0, "step": 16465, "train_loss": 3.4749298095703125, "perplexity": 32.29556161770665, "lr": 0.0026291804804649314, "grad_norm": 0.154941, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:20.720461+00:00", "epoch": 0, "step": 16466, "train_loss": 3.5135488510131836, "perplexity": 33.567181581831846, "lr": 0.0026291804804649314, "grad_norm": 0.209392, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:21.026049+00:00", "epoch": 0, "step": 16467, "train_loss": 3.5068886280059814, "perplexity": 33.34435951278076, "lr": 0.0026291804804649314, "grad_norm": 0.193321, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:21.332194+00:00", "epoch": 0, "step": 16468, "train_loss": 3.466736316680908, "perplexity": 32.03202926278605, "lr": 0.0026291804804649314, "grad_norm": 0.156469, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:21.637268+00:00", "epoch": 0, "step": 16469, "train_loss": 3.446544647216797, "perplexity": 31.391735172989105, "lr": 0.0026291804804649314, "grad_norm": 0.226035, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:21.942018+00:00", "epoch": 0, "step": 16470, "train_loss": 3.573017120361328, "perplexity": 35.62391280473552, "lr": 0.0026291804804649314, "grad_norm": 0.206156, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:22.247222+00:00", "epoch": 0, "step": 16471, "train_loss": 3.476076364517212, "perplexity": 32.33261148944042, "lr": 0.0026291804804649314, "grad_norm": 0.162623, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:22.553208+00:00", "epoch": 0, "step": 16472, "train_loss": 3.548430919647217, "perplexity": 34.75873542813406, "lr": 0.0026291804804649314, "grad_norm": 0.178241, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:22.858311+00:00", "epoch": 0, "step": 16473, "train_loss": 3.528946876525879, "perplexity": 34.088049779706346, "lr": 0.0026291804804649314, "grad_norm": 0.166563, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:23.163728+00:00", "epoch": 0, "step": 16474, "train_loss": 3.4384491443634033, "perplexity": 31.138629183771425, "lr": 0.0026291804804649314, "grad_norm": 0.159626, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:23.467959+00:00", "epoch": 0, "step": 16475, "train_loss": 3.5502994060516357, "perplexity": 34.8237423660933, "lr": 0.0026291804804649314, "grad_norm": 0.167934, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:23.773172+00:00", "epoch": 0, "step": 16476, "train_loss": 3.446028709411621, "perplexity": 31.375543167436376, "lr": 0.0026291804804649314, "grad_norm": 0.165541, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:24.078074+00:00", "epoch": 0, "step": 16477, "train_loss": 3.6148509979248047, "perplexity": 37.14581071039682, "lr": 0.0026291804804649314, "grad_norm": 0.198757, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:24.384022+00:00", "epoch": 0, "step": 16478, "train_loss": 3.518475294113159, "perplexity": 33.73295639674058, "lr": 0.0026291804804649314, "grad_norm": 0.165842, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:24.688768+00:00", "epoch": 0, "step": 16479, "train_loss": 3.5215044021606445, "perplexity": 33.83529208110157, "lr": 0.0026291804804649314, "grad_norm": 0.159363, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:24.993370+00:00", "epoch": 0, "step": 16480, "train_loss": 3.4831974506378174, "perplexity": 32.563676540987544, "lr": 0.0026291804804649314, "grad_norm": 0.195943, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:25.297537+00:00", "epoch": 0, "step": 16481, "train_loss": 3.5712172985076904, "perplexity": 35.559853772672895, "lr": 0.0026291804804649314, "grad_norm": 0.17509, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:25.602418+00:00", "epoch": 0, "step": 16482, "train_loss": 3.575134754180908, "perplexity": 35.699431139145034, "lr": 0.0026291804804649314, "grad_norm": 0.153352, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:25.907629+00:00", "epoch": 0, "step": 16483, "train_loss": 3.454867124557495, "perplexity": 31.65408235262239, "lr": 0.0026291804804649314, "grad_norm": 0.180104, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:26.214034+00:00", "epoch": 0, "step": 16484, "train_loss": 3.5262558460235596, "perplexity": 33.99644111412437, "lr": 0.0026291804804649314, "grad_norm": 0.192753, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:26.518827+00:00", "epoch": 0, "step": 16485, "train_loss": 3.6013355255126953, "perplexity": 36.64714497285705, "lr": 0.0026291804804649314, "grad_norm": 0.17625, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:26.822653+00:00", "epoch": 0, "step": 16486, "train_loss": 3.5551395416259766, "perplexity": 34.99270256563615, "lr": 0.0026291804804649314, "grad_norm": 0.154304, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:27.127488+00:00", "epoch": 0, "step": 16487, "train_loss": 3.3950395584106445, "perplexity": 29.81583291857176, "lr": 0.0026291804804649314, "grad_norm": 0.163237, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:27.433815+00:00", "epoch": 0, "step": 16488, "train_loss": 3.4737884998321533, "perplexity": 32.25872340463746, "lr": 0.0026291804804649314, "grad_norm": 0.161673, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:27.738841+00:00", "epoch": 0, "step": 16489, "train_loss": 3.523198366165161, "perplexity": 33.89265642081526, "lr": 0.0026291804804649314, "grad_norm": 0.148244, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:28.044341+00:00", "epoch": 0, "step": 16490, "train_loss": 3.44528865814209, "perplexity": 31.352332246578147, "lr": 0.0026291804804649314, "grad_norm": 0.156616, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:28.349152+00:00", "epoch": 0, "step": 16491, "train_loss": 3.467585802078247, "perplexity": 32.05925156472888, "lr": 0.0026291804804649314, "grad_norm": 0.14777, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:28.654545+00:00", "epoch": 0, "step": 16492, "train_loss": 3.4821953773498535, "perplexity": 32.53106169453032, "lr": 0.0026291804804649314, "grad_norm": 0.177582, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:28.960198+00:00", "epoch": 0, "step": 16493, "train_loss": 3.4942617416381836, "perplexity": 32.92597110456183, "lr": 0.0026291804804649314, "grad_norm": 0.187708, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:29.266582+00:00", "epoch": 0, "step": 16494, "train_loss": 3.525888681411743, "perplexity": 33.98396111526673, "lr": 0.0026291804804649314, "grad_norm": 0.22866, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:29.572362+00:00", "epoch": 0, "step": 16495, "train_loss": 3.4640583992004395, "perplexity": 31.94636488395044, "lr": 0.0026291804804649314, "grad_norm": 0.236662, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:29.878274+00:00", "epoch": 0, "step": 16496, "train_loss": 3.493112087249756, "perplexity": 32.888139368265925, "lr": 0.0026291804804649314, "grad_norm": 0.169506, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:30.183657+00:00", "epoch": 0, "step": 16497, "train_loss": 3.4615633487701416, "perplexity": 31.866756447331017, "lr": 0.0026291804804649314, "grad_norm": 0.185701, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:30.488542+00:00", "epoch": 0, "step": 16498, "train_loss": 3.566349983215332, "perplexity": 35.387193289986364, "lr": 0.0026291804804649314, "grad_norm": 0.172051, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:30.796065+00:00", "epoch": 0, "step": 16499, "train_loss": 3.5787949562072754, "perplexity": 35.83033769541486, "lr": 0.0026291804804649314, "grad_norm": 0.170331, "tokens_per_sec": 106557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:31.103005+00:00", "epoch": 0, "step": 16500, "train_loss": 3.6219472885131836, "perplexity": 37.410345675725594, "lr": 0.0026291804804649314, "grad_norm": 0.178799, "tokens_per_sec": 106813} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:24:34.177066+00:00", "step": 16500, "epoch": 0, "val_loss": 3.4702879786491394, "val_ppl": 32.14599847313544, "eval_train_loss": 3.6219472885131836, "eval_train_ppl": 37.410345675725594} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:24:35.149708+00:00", "step": 16500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4703_epoch_0000_step_0016500.pt", "val_loss": 3.4702879786491394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:36.404918+00:00", "epoch": 0, "step": 16501, "train_loss": 3.587589979171753, "perplexity": 36.14685619170589, "lr": 0.0026291804804649314, "grad_norm": 0.182676, "tokens_per_sec": 6180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:36.708673+00:00", "epoch": 0, "step": 16502, "train_loss": 3.575303554534912, "perplexity": 35.70545772438962, "lr": 0.0026291804804649314, "grad_norm": 0.165252, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:37.012752+00:00", "epoch": 0, "step": 16503, "train_loss": 3.4849064350128174, "perplexity": 32.61937493567557, "lr": 0.0026291804804649314, "grad_norm": 0.164569, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:37.317609+00:00", "epoch": 0, "step": 16504, "train_loss": 3.54803466796875, "perplexity": 34.74496494934713, "lr": 0.0026291804804649314, "grad_norm": 0.187383, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:37.620984+00:00", "epoch": 0, "step": 16505, "train_loss": 3.436626434326172, "perplexity": 31.08192418589605, "lr": 0.0026291804804649314, "grad_norm": 0.155356, "tokens_per_sec": 108012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:37.925244+00:00", "epoch": 0, "step": 16506, "train_loss": 3.4468183517456055, "perplexity": 31.400328409023274, "lr": 0.0026291804804649314, "grad_norm": 0.196495, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:38.230019+00:00", "epoch": 0, "step": 16507, "train_loss": 3.5039255619049072, "perplexity": 33.24570420472673, "lr": 0.0026291804804649314, "grad_norm": 0.171265, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:38.534587+00:00", "epoch": 0, "step": 16508, "train_loss": 3.4557697772979736, "perplexity": 31.682667896274058, "lr": 0.0026291804804649314, "grad_norm": 0.173799, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:38.839327+00:00", "epoch": 0, "step": 16509, "train_loss": 3.6237268447875977, "perplexity": 37.47697876217979, "lr": 0.0026291804804649314, "grad_norm": 0.161213, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:39.144058+00:00", "epoch": 0, "step": 16510, "train_loss": 3.5947391986846924, "perplexity": 36.40620396376598, "lr": 0.0026291804804649314, "grad_norm": 0.156305, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:39.447727+00:00", "epoch": 0, "step": 16511, "train_loss": 3.584674596786499, "perplexity": 36.04162774898847, "lr": 0.0026291804804649314, "grad_norm": 0.159436, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:39.753024+00:00", "epoch": 0, "step": 16512, "train_loss": 3.5112802982330322, "perplexity": 33.491118967380565, "lr": 0.0026291804804649314, "grad_norm": 0.191125, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:40.058252+00:00", "epoch": 0, "step": 16513, "train_loss": 3.5334134101867676, "perplexity": 34.24064573527778, "lr": 0.0026291804804649314, "grad_norm": 0.158239, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:40.411287+00:00", "epoch": 0, "step": 16514, "train_loss": 3.4590797424316406, "perplexity": 31.787710169580947, "lr": 0.0026291804804649314, "grad_norm": 0.174744, "tokens_per_sec": 92818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:40.716611+00:00", "epoch": 0, "step": 16515, "train_loss": 3.447213649749756, "perplexity": 31.412743349812022, "lr": 0.0026291804804649314, "grad_norm": 0.180189, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:41.021653+00:00", "epoch": 0, "step": 16516, "train_loss": 3.571362018585205, "perplexity": 35.565000369866276, "lr": 0.0026291804804649314, "grad_norm": 0.162323, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:41.326700+00:00", "epoch": 0, "step": 16517, "train_loss": 3.612672805786133, "perplexity": 37.0649880531166, "lr": 0.0026291804804649314, "grad_norm": 0.165708, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:41.632010+00:00", "epoch": 0, "step": 16518, "train_loss": 3.526405096054077, "perplexity": 34.00151546266206, "lr": 0.0026291804804649314, "grad_norm": 0.203159, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:41.938096+00:00", "epoch": 0, "step": 16519, "train_loss": 3.4092397689819336, "perplexity": 30.242244425645975, "lr": 0.0026291804804649314, "grad_norm": 0.153868, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:42.244140+00:00", "epoch": 0, "step": 16520, "train_loss": 3.479534864425659, "perplexity": 32.44462741519707, "lr": 0.0026291804804649314, "grad_norm": 0.174996, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:42.548405+00:00", "epoch": 0, "step": 16521, "train_loss": 3.5225653648376465, "perplexity": 33.87120911311115, "lr": 0.0026291804804649314, "grad_norm": 0.187678, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:42.853462+00:00", "epoch": 0, "step": 16522, "train_loss": 3.472773790359497, "perplexity": 32.22600677416758, "lr": 0.0026291804804649314, "grad_norm": 0.16255, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:43.159335+00:00", "epoch": 0, "step": 16523, "train_loss": 3.6059751510620117, "perplexity": 36.817569049218015, "lr": 0.0026291804804649314, "grad_norm": 0.186794, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:43.466411+00:00", "epoch": 0, "step": 16524, "train_loss": 3.598994016647339, "perplexity": 36.56143574163505, "lr": 0.0026291804804649314, "grad_norm": 0.174906, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:43.771119+00:00", "epoch": 0, "step": 16525, "train_loss": 3.5194437503814697, "perplexity": 33.765641114116754, "lr": 0.0026291804804649314, "grad_norm": 0.165732, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:44.075601+00:00", "epoch": 0, "step": 16526, "train_loss": 3.5448765754699707, "perplexity": 34.63541021921007, "lr": 0.0026291804804649314, "grad_norm": 0.168367, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:44.380489+00:00", "epoch": 0, "step": 16527, "train_loss": 3.530848979949951, "perplexity": 34.15295048026711, "lr": 0.0026291804804649314, "grad_norm": 0.170357, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:44.685384+00:00", "epoch": 0, "step": 16528, "train_loss": 3.510429859161377, "perplexity": 33.462648918993914, "lr": 0.0026291804804649314, "grad_norm": 0.167237, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:44.990161+00:00", "epoch": 0, "step": 16529, "train_loss": 3.5327465534210205, "perplexity": 34.21781974068602, "lr": 0.0026291804804649314, "grad_norm": 0.192306, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:45.294425+00:00", "epoch": 0, "step": 16530, "train_loss": 3.5205559730529785, "perplexity": 33.803216918138915, "lr": 0.0026291804804649314, "grad_norm": 0.174527, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:45.612711+00:00", "epoch": 0, "step": 16531, "train_loss": 3.528855800628662, "perplexity": 34.0849453213611, "lr": 0.0026291804804649314, "grad_norm": 0.214153, "tokens_per_sec": 102897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:45.917569+00:00", "epoch": 0, "step": 16532, "train_loss": 3.518547534942627, "perplexity": 33.735393381514925, "lr": 0.0026291804804649314, "grad_norm": 0.206571, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:46.221520+00:00", "epoch": 0, "step": 16533, "train_loss": 3.557542324066162, "perplexity": 35.076883510643135, "lr": 0.0026291804804649314, "grad_norm": 0.163656, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:46.527044+00:00", "epoch": 0, "step": 16534, "train_loss": 3.4302899837493896, "perplexity": 30.885597772652975, "lr": 0.0026291804804649314, "grad_norm": 0.153224, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:46.832110+00:00", "epoch": 0, "step": 16535, "train_loss": 3.4752514362335205, "perplexity": 32.305950401991645, "lr": 0.0026291804804649314, "grad_norm": 0.174927, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:47.137241+00:00", "epoch": 0, "step": 16536, "train_loss": 3.527702569961548, "perplexity": 34.04566017390191, "lr": 0.0026291804804649314, "grad_norm": 0.178355, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:47.442177+00:00", "epoch": 0, "step": 16537, "train_loss": 3.475511074066162, "perplexity": 32.31433933793089, "lr": 0.0026291804804649314, "grad_norm": 0.174007, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:47.746946+00:00", "epoch": 0, "step": 16538, "train_loss": 3.5210375785827637, "perplexity": 33.81950065518537, "lr": 0.0026291804804649314, "grad_norm": 0.160115, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:48.051381+00:00", "epoch": 0, "step": 16539, "train_loss": 3.5257766246795654, "perplexity": 33.98015319699312, "lr": 0.0026291804804649314, "grad_norm": 0.159239, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:48.355653+00:00", "epoch": 0, "step": 16540, "train_loss": 3.426812171936035, "perplexity": 30.778370042867568, "lr": 0.0026291804804649314, "grad_norm": 0.181022, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:48.660004+00:00", "epoch": 0, "step": 16541, "train_loss": 3.4392712116241455, "perplexity": 31.164237755909994, "lr": 0.0026291804804649314, "grad_norm": 0.189808, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:48.965113+00:00", "epoch": 0, "step": 16542, "train_loss": 3.4987857341766357, "perplexity": 33.075265400673324, "lr": 0.0026291804804649314, "grad_norm": 0.18765, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:49.269993+00:00", "epoch": 0, "step": 16543, "train_loss": 3.6298506259918213, "perplexity": 37.70718372331837, "lr": 0.0026291804804649314, "grad_norm": 0.184194, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:49.574736+00:00", "epoch": 0, "step": 16544, "train_loss": 3.5055456161499023, "perplexity": 33.2996077004555, "lr": 0.0026291804804649314, "grad_norm": 0.15852, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:49.878979+00:00", "epoch": 0, "step": 16545, "train_loss": 3.5077810287475586, "perplexity": 33.374129325263084, "lr": 0.0026291804804649314, "grad_norm": 0.181249, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:50.183721+00:00", "epoch": 0, "step": 16546, "train_loss": 3.542142629623413, "perplexity": 34.540848205909185, "lr": 0.0026291804804649314, "grad_norm": 0.190376, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:50.488975+00:00", "epoch": 0, "step": 16547, "train_loss": 3.5173397064208984, "perplexity": 33.6946714087277, "lr": 0.0026291804804649314, "grad_norm": 0.180563, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:50.794453+00:00", "epoch": 0, "step": 16548, "train_loss": 3.526294708251953, "perplexity": 33.9977623172558, "lr": 0.0026291804804649314, "grad_norm": 0.178856, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:51.100125+00:00", "epoch": 0, "step": 16549, "train_loss": 3.5272369384765625, "perplexity": 34.029811132790456, "lr": 0.0026291804804649314, "grad_norm": 0.19292, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:51.405456+00:00", "epoch": 0, "step": 16550, "train_loss": 3.6203832626342773, "perplexity": 37.35188065926729, "lr": 0.0026291804804649314, "grad_norm": 0.181645, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:51.717389+00:00", "epoch": 0, "step": 16551, "train_loss": 3.5424680709838867, "perplexity": 34.5520910558862, "lr": 0.0026291804804649314, "grad_norm": 0.162299, "tokens_per_sec": 104988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:52.022469+00:00", "epoch": 0, "step": 16552, "train_loss": 3.4947264194488525, "perplexity": 32.94127462805738, "lr": 0.0026291804804649314, "grad_norm": 0.192224, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:52.327126+00:00", "epoch": 0, "step": 16553, "train_loss": 3.4928066730499268, "perplexity": 32.87809639720795, "lr": 0.0026291804804649314, "grad_norm": 0.175983, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:52.631519+00:00", "epoch": 0, "step": 16554, "train_loss": 3.568035364151001, "perplexity": 35.44688447798984, "lr": 0.0026291804804649314, "grad_norm": 0.170788, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:52.936864+00:00", "epoch": 0, "step": 16555, "train_loss": 3.542855978012085, "perplexity": 34.565496654740635, "lr": 0.0026291804804649314, "grad_norm": 0.187651, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:53.241150+00:00", "epoch": 0, "step": 16556, "train_loss": 3.4435863494873047, "perplexity": 31.299006301533552, "lr": 0.0026291804804649314, "grad_norm": 0.208043, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:53.546323+00:00", "epoch": 0, "step": 16557, "train_loss": 3.363783359527588, "perplexity": 28.89831704043047, "lr": 0.0026291804804649314, "grad_norm": 0.164653, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:53.850639+00:00", "epoch": 0, "step": 16558, "train_loss": 3.5297446250915527, "perplexity": 34.11525432227438, "lr": 0.0026291804804649314, "grad_norm": 0.191689, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:54.156151+00:00", "epoch": 0, "step": 16559, "train_loss": 3.531381607055664, "perplexity": 34.171146112746335, "lr": 0.0026291804804649314, "grad_norm": 0.150582, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:54.461536+00:00", "epoch": 0, "step": 16560, "train_loss": 3.469031572341919, "perplexity": 32.10563539942171, "lr": 0.0026291804804649314, "grad_norm": 0.171721, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:54.766090+00:00", "epoch": 0, "step": 16561, "train_loss": 3.635465383529663, "perplexity": 37.919495900287295, "lr": 0.0026291804804649314, "grad_norm": 0.165762, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:55.070095+00:00", "epoch": 0, "step": 16562, "train_loss": 3.5785765647888184, "perplexity": 35.82251351154017, "lr": 0.0026291804804649314, "grad_norm": 0.193156, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:55.374091+00:00", "epoch": 0, "step": 16563, "train_loss": 3.5099971294403076, "perplexity": 33.44817176883344, "lr": 0.0026291804804649314, "grad_norm": 0.165903, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:55.678379+00:00", "epoch": 0, "step": 16564, "train_loss": 3.554537057876587, "perplexity": 34.97162638066013, "lr": 0.0026291804804649314, "grad_norm": 0.198919, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:55.983768+00:00", "epoch": 0, "step": 16565, "train_loss": 3.537879467010498, "perplexity": 34.39390838993138, "lr": 0.0026291804804649314, "grad_norm": 0.178079, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:56.288805+00:00", "epoch": 0, "step": 16566, "train_loss": 3.502647876739502, "perplexity": 33.20325378656063, "lr": 0.0026291804804649314, "grad_norm": 0.166931, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:56.593171+00:00", "epoch": 0, "step": 16567, "train_loss": 3.4299299716949463, "perplexity": 30.87448058642673, "lr": 0.0026291804804649314, "grad_norm": 0.182704, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:56.897109+00:00", "epoch": 0, "step": 16568, "train_loss": 3.633652687072754, "perplexity": 37.85082162603242, "lr": 0.0026291804804649314, "grad_norm": 0.17362, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:57.201073+00:00", "epoch": 0, "step": 16569, "train_loss": 3.5087876319885254, "perplexity": 33.407740745850724, "lr": 0.0026291804804649314, "grad_norm": 0.175116, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:57.506291+00:00", "epoch": 0, "step": 16570, "train_loss": 3.561439037322998, "perplexity": 35.21383472413897, "lr": 0.0026291804804649314, "grad_norm": 0.186104, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:57.811620+00:00", "epoch": 0, "step": 16571, "train_loss": 3.4819369316101074, "perplexity": 32.52265526657549, "lr": 0.0026291804804649314, "grad_norm": 0.180857, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:58.117869+00:00", "epoch": 0, "step": 16572, "train_loss": 3.480179786682129, "perplexity": 32.46555842623196, "lr": 0.0026291804804649314, "grad_norm": 0.151517, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:58.422903+00:00", "epoch": 0, "step": 16573, "train_loss": 3.4293856620788574, "perplexity": 30.857679882558926, "lr": 0.0026291804804649314, "grad_norm": 0.168332, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:58.728206+00:00", "epoch": 0, "step": 16574, "train_loss": 3.596264123916626, "perplexity": 36.46176305375193, "lr": 0.0026291804804649314, "grad_norm": 0.192427, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:59.033468+00:00", "epoch": 0, "step": 16575, "train_loss": 3.5205750465393066, "perplexity": 33.80386166948345, "lr": 0.0026291804804649314, "grad_norm": 0.157844, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:59.339347+00:00", "epoch": 0, "step": 16576, "train_loss": 3.552109718322754, "perplexity": 34.8868413113899, "lr": 0.0026291804804649314, "grad_norm": 0.165811, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:59.644017+00:00", "epoch": 0, "step": 16577, "train_loss": 3.5855212211608887, "perplexity": 36.07215439000923, "lr": 0.0026291804804649314, "grad_norm": 0.157535, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:24:59.947324+00:00", "epoch": 0, "step": 16578, "train_loss": 3.3855879306793213, "perplexity": 29.535352352292588, "lr": 0.0026291804804649314, "grad_norm": 0.153384, "tokens_per_sec": 108036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:00.251273+00:00", "epoch": 0, "step": 16579, "train_loss": 3.4923572540283203, "perplexity": 32.86332367511263, "lr": 0.0026291804804649314, "grad_norm": 0.147684, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:00.556844+00:00", "epoch": 0, "step": 16580, "train_loss": 3.627546787261963, "perplexity": 37.62041244495803, "lr": 0.0026291804804649314, "grad_norm": 0.161884, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:00.862230+00:00", "epoch": 0, "step": 16581, "train_loss": 3.551088809967041, "perplexity": 34.851243217878086, "lr": 0.0026291804804649314, "grad_norm": 0.173543, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:01.166609+00:00", "epoch": 0, "step": 16582, "train_loss": 3.479677200317383, "perplexity": 32.44924577884351, "lr": 0.0026291804804649314, "grad_norm": 0.163363, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:01.470698+00:00", "epoch": 0, "step": 16583, "train_loss": 3.44622802734375, "perplexity": 31.381797499099452, "lr": 0.0026291804804649314, "grad_norm": 0.17659, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:01.775078+00:00", "epoch": 0, "step": 16584, "train_loss": 3.5233359336853027, "perplexity": 33.89731927023104, "lr": 0.0026291804804649314, "grad_norm": 0.168535, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:02.079915+00:00", "epoch": 0, "step": 16585, "train_loss": 3.5964367389678955, "perplexity": 36.46805744608863, "lr": 0.0026291804804649314, "grad_norm": 0.178217, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:02.384364+00:00", "epoch": 0, "step": 16586, "train_loss": 3.4453859329223633, "perplexity": 31.35538218614719, "lr": 0.0026291804804649314, "grad_norm": 0.166129, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:02.689950+00:00", "epoch": 0, "step": 16587, "train_loss": 3.5343356132507324, "perplexity": 34.27223712828777, "lr": 0.0026291804804649314, "grad_norm": 0.186689, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:02.994340+00:00", "epoch": 0, "step": 16588, "train_loss": 3.5779967308044434, "perplexity": 35.80174842153467, "lr": 0.0026291804804649314, "grad_norm": 0.166071, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:03.298797+00:00", "epoch": 0, "step": 16589, "train_loss": 3.576131582260132, "perplexity": 35.735035077071174, "lr": 0.0026291804804649314, "grad_norm": 0.177552, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:03.603549+00:00", "epoch": 0, "step": 16590, "train_loss": 3.638558864593506, "perplexity": 38.03698076772115, "lr": 0.0026291804804649314, "grad_norm": 0.172351, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:03.908046+00:00", "epoch": 0, "step": 16591, "train_loss": 3.3840067386627197, "perplexity": 29.488688191167498, "lr": 0.0026291804804649314, "grad_norm": 0.189495, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:04.213880+00:00", "epoch": 0, "step": 16592, "train_loss": 3.524359703063965, "perplexity": 33.93204007773581, "lr": 0.0026291804804649314, "grad_norm": 0.175139, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:04.518460+00:00", "epoch": 0, "step": 16593, "train_loss": 3.3815877437591553, "perplexity": 29.417441412259723, "lr": 0.0026291804804649314, "grad_norm": 0.177724, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:04.822793+00:00", "epoch": 0, "step": 16594, "train_loss": 3.5278820991516113, "perplexity": 34.05177291238918, "lr": 0.0026291804804649314, "grad_norm": 0.196497, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:05.126934+00:00", "epoch": 0, "step": 16595, "train_loss": 3.5808157920837402, "perplexity": 35.902818138168975, "lr": 0.0026291804804649314, "grad_norm": 0.193345, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:05.431886+00:00", "epoch": 0, "step": 16596, "train_loss": 3.497812271118164, "perplexity": 33.0430835181349, "lr": 0.0026291804804649314, "grad_norm": 0.143225, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:05.737414+00:00", "epoch": 0, "step": 16597, "train_loss": 3.575244426727295, "perplexity": 35.703346601368054, "lr": 0.0026291804804649314, "grad_norm": 0.214832, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:06.043159+00:00", "epoch": 0, "step": 16598, "train_loss": 3.572599172592163, "perplexity": 35.60902698081498, "lr": 0.0026291804804649314, "grad_norm": 0.179889, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:06.347420+00:00", "epoch": 0, "step": 16599, "train_loss": 3.437446355819702, "perplexity": 31.10741937419447, "lr": 0.0026291804804649314, "grad_norm": 0.171492, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:06.651799+00:00", "epoch": 0, "step": 16600, "train_loss": 3.48521089553833, "perplexity": 32.629307759710066, "lr": 0.0026291804804649314, "grad_norm": 0.166022, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:06.955379+00:00", "epoch": 0, "step": 16601, "train_loss": 3.553208112716675, "perplexity": 34.92518187498327, "lr": 0.0026291804804649314, "grad_norm": 0.166328, "tokens_per_sec": 107880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:07.261323+00:00", "epoch": 0, "step": 16602, "train_loss": 3.4208672046661377, "perplexity": 30.59593645849261, "lr": 0.0026291804804649314, "grad_norm": 0.16462, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:07.566017+00:00", "epoch": 0, "step": 16603, "train_loss": 3.5095930099487305, "perplexity": 33.43465744154949, "lr": 0.0026291804804649314, "grad_norm": 0.164242, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:07.870460+00:00", "epoch": 0, "step": 16604, "train_loss": 3.4927799701690674, "perplexity": 32.877218469038624, "lr": 0.0026291804804649314, "grad_norm": 0.167177, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:08.174070+00:00", "epoch": 0, "step": 16605, "train_loss": 3.520972728729248, "perplexity": 33.817307536634395, "lr": 0.0026291804804649314, "grad_norm": 0.173095, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:08.477987+00:00", "epoch": 0, "step": 16606, "train_loss": 3.4665918350219727, "perplexity": 32.02740155637649, "lr": 0.0026291804804649314, "grad_norm": 0.181284, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:08.782676+00:00", "epoch": 0, "step": 16607, "train_loss": 3.49341082572937, "perplexity": 32.89796578871001, "lr": 0.0026291804804649314, "grad_norm": 0.16676, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:09.088186+00:00", "epoch": 0, "step": 16608, "train_loss": 3.5891191959381104, "perplexity": 36.20217485658003, "lr": 0.0026291804804649314, "grad_norm": 0.176571, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:09.394093+00:00", "epoch": 0, "step": 16609, "train_loss": 3.537227153778076, "perplexity": 34.37148010430296, "lr": 0.0026291804804649314, "grad_norm": 0.165932, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:09.698683+00:00", "epoch": 0, "step": 16610, "train_loss": 3.4744787216186523, "perplexity": 32.28099676423074, "lr": 0.0026291804804649314, "grad_norm": 0.187656, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:10.002875+00:00", "epoch": 0, "step": 16611, "train_loss": 3.5205180644989014, "perplexity": 33.801935511350685, "lr": 0.0026291804804649314, "grad_norm": 0.15963, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:10.306754+00:00", "epoch": 0, "step": 16612, "train_loss": 3.555518627166748, "perplexity": 35.005970307857034, "lr": 0.0026291804804649314, "grad_norm": 0.183886, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:10.611254+00:00", "epoch": 0, "step": 16613, "train_loss": 3.498823642730713, "perplexity": 33.07651925992622, "lr": 0.0026291804804649314, "grad_norm": 0.17513, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:10.916395+00:00", "epoch": 0, "step": 16614, "train_loss": 3.3927550315856934, "perplexity": 29.7477955945978, "lr": 0.0026291804804649314, "grad_norm": 0.179844, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:11.220931+00:00", "epoch": 0, "step": 16615, "train_loss": 3.578809976577759, "perplexity": 35.83087588440348, "lr": 0.0026291804804649314, "grad_norm": 0.200484, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:11.525853+00:00", "epoch": 0, "step": 16616, "train_loss": 3.6331772804260254, "perplexity": 37.832831370529625, "lr": 0.0026291804804649314, "grad_norm": 0.166178, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:11.831429+00:00", "epoch": 0, "step": 16617, "train_loss": 3.5140445232391357, "perplexity": 33.583824025703315, "lr": 0.0026291804804649314, "grad_norm": 0.192423, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:12.136978+00:00", "epoch": 0, "step": 16618, "train_loss": 3.4756200313568115, "perplexity": 32.31786041261417, "lr": 0.0026291804804649314, "grad_norm": 0.185857, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:12.442034+00:00", "epoch": 0, "step": 16619, "train_loss": 3.560786485671997, "perplexity": 35.19086337396304, "lr": 0.0026291804804649314, "grad_norm": 0.170544, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:12.746341+00:00", "epoch": 0, "step": 16620, "train_loss": 3.530810594558716, "perplexity": 34.15163953106189, "lr": 0.0026291804804649314, "grad_norm": 0.144789, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:13.050361+00:00", "epoch": 0, "step": 16621, "train_loss": 3.5341720581054688, "perplexity": 34.26663218593682, "lr": 0.0026291804804649314, "grad_norm": 0.147376, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:13.354366+00:00", "epoch": 0, "step": 16622, "train_loss": 3.5303587913513184, "perplexity": 34.13621319588014, "lr": 0.0026291804804649314, "grad_norm": 0.168021, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:13.659180+00:00", "epoch": 0, "step": 16623, "train_loss": 3.4409165382385254, "perplexity": 31.215555311199594, "lr": 0.0026291804804649314, "grad_norm": 0.153653, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:13.964600+00:00", "epoch": 0, "step": 16624, "train_loss": 3.5534701347351074, "perplexity": 34.934334240640574, "lr": 0.0026291804804649314, "grad_norm": 0.177768, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:14.270476+00:00", "epoch": 0, "step": 16625, "train_loss": 3.455369710922241, "perplexity": 31.669995261271865, "lr": 0.0026291804804649314, "grad_norm": 0.177782, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:14.574813+00:00", "epoch": 0, "step": 16626, "train_loss": 3.51436448097229, "perplexity": 33.59457114913017, "lr": 0.0026291804804649314, "grad_norm": 0.179011, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:14.879540+00:00", "epoch": 0, "step": 16627, "train_loss": 3.5375051498413086, "perplexity": 34.38103656872763, "lr": 0.0026291804804649314, "grad_norm": 0.182486, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:15.184233+00:00", "epoch": 0, "step": 16628, "train_loss": 3.595301866531372, "perplexity": 36.426694328250086, "lr": 0.0026291804804649314, "grad_norm": 0.187164, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:15.490035+00:00", "epoch": 0, "step": 16629, "train_loss": 3.5069947242736816, "perplexity": 33.347897412549585, "lr": 0.0026291804804649314, "grad_norm": 0.163248, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:15.794951+00:00", "epoch": 0, "step": 16630, "train_loss": 3.469153881072998, "perplexity": 32.109562439098625, "lr": 0.0026291804804649314, "grad_norm": 0.202136, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:16.099481+00:00", "epoch": 0, "step": 16631, "train_loss": 3.488835096359253, "perplexity": 32.74777747302179, "lr": 0.0026291804804649314, "grad_norm": 0.21649, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:16.404163+00:00", "epoch": 0, "step": 16632, "train_loss": 3.4687814712524414, "perplexity": 32.0976067490589, "lr": 0.0026291804804649314, "grad_norm": 0.187329, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:16.708929+00:00", "epoch": 0, "step": 16633, "train_loss": 3.5149052143096924, "perplexity": 33.612741765992645, "lr": 0.0026291804804649314, "grad_norm": 0.191595, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:17.013856+00:00", "epoch": 0, "step": 16634, "train_loss": 3.5922672748565674, "perplexity": 36.3163217374729, "lr": 0.0026291804804649314, "grad_norm": 0.18015, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:17.318669+00:00", "epoch": 0, "step": 16635, "train_loss": 3.4212183952331543, "perplexity": 30.606683349758836, "lr": 0.0026291804804649314, "grad_norm": 0.205303, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:17.624268+00:00", "epoch": 0, "step": 16636, "train_loss": 3.560081720352173, "perplexity": 35.16607081137774, "lr": 0.0026291804804649314, "grad_norm": 0.203775, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:17.928488+00:00", "epoch": 0, "step": 16637, "train_loss": 3.517026424407959, "perplexity": 33.684117127562985, "lr": 0.0026291804804649314, "grad_norm": 0.169204, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:18.232867+00:00", "epoch": 0, "step": 16638, "train_loss": 3.477241039276123, "perplexity": 32.37029040351274, "lr": 0.0026291804804649314, "grad_norm": 0.166264, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:18.538502+00:00", "epoch": 0, "step": 16639, "train_loss": 3.59134840965271, "perplexity": 36.28296725957761, "lr": 0.0026291804804649314, "grad_norm": 0.166767, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:18.844761+00:00", "epoch": 0, "step": 16640, "train_loss": 3.5131070613861084, "perplexity": 33.55235522450543, "lr": 0.0026291804804649314, "grad_norm": 0.167305, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:19.149985+00:00", "epoch": 0, "step": 16641, "train_loss": 3.569441318511963, "perplexity": 35.496756230292846, "lr": 0.0026291804804649314, "grad_norm": 0.16277, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:19.454044+00:00", "epoch": 0, "step": 16642, "train_loss": 3.45278000831604, "perplexity": 31.588085498745833, "lr": 0.0026291804804649314, "grad_norm": 0.148695, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:19.757781+00:00", "epoch": 0, "step": 16643, "train_loss": 3.543081283569336, "perplexity": 34.57328533060912, "lr": 0.0026291804804649314, "grad_norm": 0.163713, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:20.062053+00:00", "epoch": 0, "step": 16644, "train_loss": 3.464130163192749, "perplexity": 31.948657564899275, "lr": 0.0026291804804649314, "grad_norm": 0.148638, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:20.368664+00:00", "epoch": 0, "step": 16645, "train_loss": 3.452955722808838, "perplexity": 31.593636470846544, "lr": 0.0026291804804649314, "grad_norm": 0.152117, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:20.674416+00:00", "epoch": 0, "step": 16646, "train_loss": 3.6018192768096924, "perplexity": 36.664877365457066, "lr": 0.0026291804804649314, "grad_norm": 0.159056, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:20.978986+00:00", "epoch": 0, "step": 16647, "train_loss": 3.490690231323242, "perplexity": 32.8085854060269, "lr": 0.0026291804804649314, "grad_norm": 0.155138, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:21.284234+00:00", "epoch": 0, "step": 16648, "train_loss": 3.48602557182312, "perplexity": 32.65590091386622, "lr": 0.0026291804804649314, "grad_norm": 0.140216, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:21.588863+00:00", "epoch": 0, "step": 16649, "train_loss": 3.4969818592071533, "perplexity": 33.01565553784184, "lr": 0.0026291804804649314, "grad_norm": 0.150384, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:21.893900+00:00", "epoch": 0, "step": 16650, "train_loss": 3.5192577838897705, "perplexity": 33.75936242013023, "lr": 0.0026291804804649314, "grad_norm": 0.181229, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:22.200129+00:00", "epoch": 0, "step": 16651, "train_loss": 3.463641405105591, "perplexity": 31.933046215537985, "lr": 0.0026291804804649314, "grad_norm": 0.158668, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:22.506158+00:00", "epoch": 0, "step": 16652, "train_loss": 3.5331058502197266, "perplexity": 34.230116302702946, "lr": 0.0026291804804649314, "grad_norm": 0.162438, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:22.810550+00:00", "epoch": 0, "step": 16653, "train_loss": 3.4926788806915283, "perplexity": 32.873895096182494, "lr": 0.0026291804804649314, "grad_norm": 0.15337, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:23.115241+00:00", "epoch": 0, "step": 16654, "train_loss": 3.4654648303985596, "perplexity": 31.99132685874038, "lr": 0.0026291804804649314, "grad_norm": 0.160306, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:23.421037+00:00", "epoch": 0, "step": 16655, "train_loss": 3.589710235595703, "perplexity": 36.223578102071755, "lr": 0.0026291804804649314, "grad_norm": 0.181428, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:23.725581+00:00", "epoch": 0, "step": 16656, "train_loss": 3.4120776653289795, "perplexity": 30.328190676221478, "lr": 0.0026291804804649314, "grad_norm": 0.168278, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:24.030378+00:00", "epoch": 0, "step": 16657, "train_loss": 3.49745774269104, "perplexity": 33.031370882061786, "lr": 0.0026291804804649314, "grad_norm": 0.16141, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:24.334830+00:00", "epoch": 0, "step": 16658, "train_loss": 3.4669077396392822, "perplexity": 32.03752075867578, "lr": 0.0026291804804649314, "grad_norm": 0.163132, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:24.640325+00:00", "epoch": 0, "step": 16659, "train_loss": 3.4707634449005127, "perplexity": 32.16128644469537, "lr": 0.0026291804804649314, "grad_norm": 0.178189, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:24.944838+00:00", "epoch": 0, "step": 16660, "train_loss": 3.496701955795288, "perplexity": 33.00641563641212, "lr": 0.0026291804804649314, "grad_norm": 0.190989, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:25.249556+00:00", "epoch": 0, "step": 16661, "train_loss": 3.475768804550171, "perplexity": 32.32266880158112, "lr": 0.0026291804804649314, "grad_norm": 0.154674, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:25.554979+00:00", "epoch": 0, "step": 16662, "train_loss": 3.474496364593506, "perplexity": 32.28156630206906, "lr": 0.0026291804804649314, "grad_norm": 0.182754, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:25.859671+00:00", "epoch": 0, "step": 16663, "train_loss": 3.5112624168395996, "perplexity": 33.490520104860074, "lr": 0.0026291804804649314, "grad_norm": 0.203538, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:26.163569+00:00", "epoch": 0, "step": 16664, "train_loss": 3.4062113761901855, "perplexity": 30.15079756901129, "lr": 0.0026291804804649314, "grad_norm": 0.189042, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:26.467691+00:00", "epoch": 0, "step": 16665, "train_loss": 3.4468114376068115, "perplexity": 31.400111303545028, "lr": 0.0026291804804649314, "grad_norm": 0.172172, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:26.771640+00:00", "epoch": 0, "step": 16666, "train_loss": 3.493044137954712, "perplexity": 32.88590471830285, "lr": 0.0026291804804649314, "grad_norm": 0.174012, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:27.075777+00:00", "epoch": 0, "step": 16667, "train_loss": 3.507814407348633, "perplexity": 33.37524332560381, "lr": 0.0026291804804649314, "grad_norm": 0.190941, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:27.380863+00:00", "epoch": 0, "step": 16668, "train_loss": 3.548722267150879, "perplexity": 34.76886377429404, "lr": 0.0026291804804649314, "grad_norm": 0.180095, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:27.685519+00:00", "epoch": 0, "step": 16669, "train_loss": 3.481996536254883, "perplexity": 32.5245938256628, "lr": 0.0026291804804649314, "grad_norm": 0.17742, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:27.990103+00:00", "epoch": 0, "step": 16670, "train_loss": 3.554197311401367, "perplexity": 34.95974691198268, "lr": 0.0026291804804649314, "grad_norm": 0.156123, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:28.296609+00:00", "epoch": 0, "step": 16671, "train_loss": 3.4817769527435303, "perplexity": 32.517452745205695, "lr": 0.0026291804804649314, "grad_norm": 0.160701, "tokens_per_sec": 106799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:28.602319+00:00", "epoch": 0, "step": 16672, "train_loss": 3.4233429431915283, "perplexity": 30.67177784008065, "lr": 0.0026291804804649314, "grad_norm": 0.156668, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:28.907494+00:00", "epoch": 0, "step": 16673, "train_loss": 3.408304452896118, "perplexity": 30.213971592042665, "lr": 0.0026291804804649314, "grad_norm": 0.174294, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:29.211745+00:00", "epoch": 0, "step": 16674, "train_loss": 3.619168519973755, "perplexity": 37.30653528343694, "lr": 0.0026291804804649314, "grad_norm": 0.154167, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:29.516203+00:00", "epoch": 0, "step": 16675, "train_loss": 3.4781389236450195, "perplexity": 32.39936823359372, "lr": 0.0026291804804649314, "grad_norm": 0.149701, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:29.820347+00:00", "epoch": 0, "step": 16676, "train_loss": 3.4829349517822266, "perplexity": 32.55512973497418, "lr": 0.0026291804804649314, "grad_norm": 0.192136, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:30.125764+00:00", "epoch": 0, "step": 16677, "train_loss": 3.5333940982818604, "perplexity": 34.23998448956836, "lr": 0.0026291804804649314, "grad_norm": 0.14744, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:30.431477+00:00", "epoch": 0, "step": 16678, "train_loss": 3.5346851348876953, "perplexity": 34.2842181103955, "lr": 0.0026291804804649314, "grad_norm": 0.183048, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:30.735646+00:00", "epoch": 0, "step": 16679, "train_loss": 3.501201868057251, "perplexity": 33.1552762896047, "lr": 0.0026291804804649314, "grad_norm": 0.168008, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:31.040319+00:00", "epoch": 0, "step": 16680, "train_loss": 3.4794609546661377, "perplexity": 32.44222952920173, "lr": 0.0026291804804649314, "grad_norm": 0.16232, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:31.345053+00:00", "epoch": 0, "step": 16681, "train_loss": 3.474648952484131, "perplexity": 32.28649245400317, "lr": 0.0026291804804649314, "grad_norm": 0.160482, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:31.649102+00:00", "epoch": 0, "step": 16682, "train_loss": 3.644699811935425, "perplexity": 38.271282544781855, "lr": 0.0026291804804649314, "grad_norm": 0.16635, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:31.953477+00:00", "epoch": 0, "step": 16683, "train_loss": 3.4701197147369385, "perplexity": 32.14058991671594, "lr": 0.0026291804804649314, "grad_norm": 0.190231, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:32.259264+00:00", "epoch": 0, "step": 16684, "train_loss": 3.4574995040893555, "perplexity": 31.737517679650928, "lr": 0.0026291804804649314, "grad_norm": 0.167839, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:32.564437+00:00", "epoch": 0, "step": 16685, "train_loss": 3.534729242324829, "perplexity": 34.285730332740364, "lr": 0.0026291804804649314, "grad_norm": 0.16289, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:32.868109+00:00", "epoch": 0, "step": 16686, "train_loss": 3.521275043487549, "perplexity": 33.82753255329756, "lr": 0.0026291804804649314, "grad_norm": 0.157157, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:33.172462+00:00", "epoch": 0, "step": 16687, "train_loss": 3.530761480331421, "perplexity": 34.14996224086521, "lr": 0.0026291804804649314, "grad_norm": 0.169764, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:33.478442+00:00", "epoch": 0, "step": 16688, "train_loss": 3.651416540145874, "perplexity": 38.52920557768576, "lr": 0.0026291804804649314, "grad_norm": 0.154491, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:33.784844+00:00", "epoch": 0, "step": 16689, "train_loss": 3.5702667236328125, "perplexity": 35.52606752984312, "lr": 0.0026291804804649314, "grad_norm": 0.168502, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:34.090166+00:00", "epoch": 0, "step": 16690, "train_loss": 3.497251510620117, "perplexity": 33.02455945643071, "lr": 0.0026291804804649314, "grad_norm": 0.154864, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:34.394255+00:00", "epoch": 0, "step": 16691, "train_loss": 3.5656771659851074, "perplexity": 35.3633921844064, "lr": 0.0026291804804649314, "grad_norm": 0.179215, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:34.698160+00:00", "epoch": 0, "step": 16692, "train_loss": 3.454502820968628, "perplexity": 31.642552757083063, "lr": 0.0026291804804649314, "grad_norm": 0.167648, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:35.003456+00:00", "epoch": 0, "step": 16693, "train_loss": 3.5902907848358154, "perplexity": 36.24461377834518, "lr": 0.0026291804804649314, "grad_norm": 0.157592, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:35.308170+00:00", "epoch": 0, "step": 16694, "train_loss": 3.4913408756256104, "perplexity": 32.82993907125708, "lr": 0.0026291804804649314, "grad_norm": 0.182536, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:35.611947+00:00", "epoch": 0, "step": 16695, "train_loss": 3.5258872509002686, "perplexity": 33.983912500855176, "lr": 0.0026291804804649314, "grad_norm": 0.202701, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:35.916420+00:00", "epoch": 0, "step": 16696, "train_loss": 3.5152297019958496, "perplexity": 33.623650456565855, "lr": 0.0026291804804649314, "grad_norm": 0.200333, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:36.221393+00:00", "epoch": 0, "step": 16697, "train_loss": 3.467961311340332, "perplexity": 32.07129237120264, "lr": 0.0026291804804649314, "grad_norm": 0.179977, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:36.526952+00:00", "epoch": 0, "step": 16698, "train_loss": 3.4724674224853516, "perplexity": 32.21613527321257, "lr": 0.0026291804804649314, "grad_norm": 0.188935, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:36.832373+00:00", "epoch": 0, "step": 16699, "train_loss": 3.5174448490142822, "perplexity": 33.69821434011606, "lr": 0.0026291804804649314, "grad_norm": 0.188755, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:37.137034+00:00", "epoch": 0, "step": 16700, "train_loss": 3.477653980255127, "perplexity": 32.383660183197584, "lr": 0.0026291804804649314, "grad_norm": 0.170435, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:37.442359+00:00", "epoch": 0, "step": 16701, "train_loss": 3.6043484210968018, "perplexity": 36.75772549424826, "lr": 0.0026291804804649314, "grad_norm": 0.186488, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:37.747960+00:00", "epoch": 0, "step": 16702, "train_loss": 3.4798190593719482, "perplexity": 32.45384932469067, "lr": 0.0026291804804649314, "grad_norm": 0.153224, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:38.053377+00:00", "epoch": 0, "step": 16703, "train_loss": 3.4807381629943848, "perplexity": 32.48369148708601, "lr": 0.0026291804804649314, "grad_norm": 0.194633, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:38.358671+00:00", "epoch": 0, "step": 16704, "train_loss": 3.5214028358459473, "perplexity": 33.83185572968991, "lr": 0.0026291804804649314, "grad_norm": 0.188967, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:38.664623+00:00", "epoch": 0, "step": 16705, "train_loss": 3.5441927909851074, "perplexity": 34.61173515831742, "lr": 0.0026291804804649314, "grad_norm": 0.170662, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:38.969414+00:00", "epoch": 0, "step": 16706, "train_loss": 3.550696849822998, "perplexity": 34.83758559636283, "lr": 0.0026291804804649314, "grad_norm": 0.174823, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:39.274073+00:00", "epoch": 0, "step": 16707, "train_loss": 3.5134785175323486, "perplexity": 33.56482076813248, "lr": 0.0026291804804649314, "grad_norm": 0.163626, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:39.578105+00:00", "epoch": 0, "step": 16708, "train_loss": 3.4583122730255127, "perplexity": 31.763323433768758, "lr": 0.0026291804804649314, "grad_norm": 0.156001, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:39.882863+00:00", "epoch": 0, "step": 16709, "train_loss": 3.5167293548583984, "perplexity": 33.674112088230395, "lr": 0.0026291804804649314, "grad_norm": 0.199014, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:40.189008+00:00", "epoch": 0, "step": 16710, "train_loss": 3.4184746742248535, "perplexity": 30.522822247999922, "lr": 0.0026291804804649314, "grad_norm": 0.210551, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:40.493764+00:00", "epoch": 0, "step": 16711, "train_loss": 3.560380458831787, "perplexity": 35.17657783925457, "lr": 0.0026291804804649314, "grad_norm": 0.186138, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:40.798284+00:00", "epoch": 0, "step": 16712, "train_loss": 3.527632236480713, "perplexity": 34.043265708321094, "lr": 0.0026291804804649314, "grad_norm": 0.161991, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:41.102527+00:00", "epoch": 0, "step": 16713, "train_loss": 3.5752527713775635, "perplexity": 35.703644534551934, "lr": 0.0026291804804649314, "grad_norm": 0.219913, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:41.408361+00:00", "epoch": 0, "step": 16714, "train_loss": 3.5521953105926514, "perplexity": 34.889827483122076, "lr": 0.0026291804804649314, "grad_norm": 0.195251, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:41.715141+00:00", "epoch": 0, "step": 16715, "train_loss": 3.523172616958618, "perplexity": 33.89178372304047, "lr": 0.0026291804804649314, "grad_norm": 0.16932, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:42.019930+00:00", "epoch": 0, "step": 16716, "train_loss": 3.603485345840454, "perplexity": 36.72601449735141, "lr": 0.0026291804804649314, "grad_norm": 0.16829, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:42.324583+00:00", "epoch": 0, "step": 16717, "train_loss": 3.535289764404297, "perplexity": 34.304953628640135, "lr": 0.0026291804804649314, "grad_norm": 0.189123, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:42.629688+00:00", "epoch": 0, "step": 16718, "train_loss": 3.5768439769744873, "perplexity": 35.76050159720752, "lr": 0.0026291804804649314, "grad_norm": 0.202118, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:42.935548+00:00", "epoch": 0, "step": 16719, "train_loss": 3.5231175422668457, "perplexity": 33.88991719489801, "lr": 0.0026291804804649314, "grad_norm": 0.229537, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:43.240384+00:00", "epoch": 0, "step": 16720, "train_loss": 3.5775325298309326, "perplexity": 35.785133071793126, "lr": 0.0026291804804649314, "grad_norm": 0.179906, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:43.546635+00:00", "epoch": 0, "step": 16721, "train_loss": 3.4743025302886963, "perplexity": 32.27530963350482, "lr": 0.0026291804804649314, "grad_norm": 0.14666, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:43.851768+00:00", "epoch": 0, "step": 16722, "train_loss": 3.369434118270874, "perplexity": 29.062076705541042, "lr": 0.0026291804804649314, "grad_norm": 0.150152, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:44.156613+00:00", "epoch": 0, "step": 16723, "train_loss": 3.4981038570404053, "perplexity": 33.05271982095307, "lr": 0.0026291804804649314, "grad_norm": 0.146856, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:44.462038+00:00", "epoch": 0, "step": 16724, "train_loss": 3.5187909603118896, "perplexity": 33.74360643169787, "lr": 0.0026291804804649314, "grad_norm": 0.163742, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:44.767234+00:00", "epoch": 0, "step": 16725, "train_loss": 3.5200459957122803, "perplexity": 33.78598243843854, "lr": 0.0026291804804649314, "grad_norm": 0.164479, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:45.073925+00:00", "epoch": 0, "step": 16726, "train_loss": 3.477208375930786, "perplexity": 32.36923309880628, "lr": 0.0026291804804649314, "grad_norm": 0.158663, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:45.378666+00:00", "epoch": 0, "step": 16727, "train_loss": 3.549058675765991, "perplexity": 34.78056228723532, "lr": 0.0026291804804649314, "grad_norm": 0.161921, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:45.683218+00:00", "epoch": 0, "step": 16728, "train_loss": 3.545482873916626, "perplexity": 34.65641598186241, "lr": 0.0026291804804649314, "grad_norm": 0.167841, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:45.988710+00:00", "epoch": 0, "step": 16729, "train_loss": 3.607100248336792, "perplexity": 36.85901570720822, "lr": 0.0026291804804649314, "grad_norm": 0.161354, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:46.293741+00:00", "epoch": 0, "step": 16730, "train_loss": 3.4920156002044678, "perplexity": 32.852097712720216, "lr": 0.0026291804804649314, "grad_norm": 0.168607, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:46.599611+00:00", "epoch": 0, "step": 16731, "train_loss": 3.606154680252075, "perplexity": 36.8241794709336, "lr": 0.0026291804804649314, "grad_norm": 0.147768, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:46.903733+00:00", "epoch": 0, "step": 16732, "train_loss": 3.4908552169799805, "perplexity": 32.813998798590546, "lr": 0.0026291804804649314, "grad_norm": 0.152657, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:47.207886+00:00", "epoch": 0, "step": 16733, "train_loss": 3.569948673248291, "perplexity": 35.51477024705252, "lr": 0.0026291804804649314, "grad_norm": 0.145752, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:47.513201+00:00", "epoch": 0, "step": 16734, "train_loss": 3.5444138050079346, "perplexity": 34.61938568254722, "lr": 0.0026291804804649314, "grad_norm": 0.150275, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:47.819134+00:00", "epoch": 0, "step": 16735, "train_loss": 3.4654645919799805, "perplexity": 31.991319231414597, "lr": 0.0026291804804649314, "grad_norm": 0.144736, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:48.124285+00:00", "epoch": 0, "step": 16736, "train_loss": 3.5065388679504395, "perplexity": 33.3326990270505, "lr": 0.0026291804804649314, "grad_norm": 0.162521, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:48.429153+00:00", "epoch": 0, "step": 16737, "train_loss": 3.4726827144622803, "perplexity": 32.2230718953373, "lr": 0.0026291804804649314, "grad_norm": 0.162338, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:48.735170+00:00", "epoch": 0, "step": 16738, "train_loss": 3.5470407009124756, "perplexity": 34.710446756631185, "lr": 0.0026291804804649314, "grad_norm": 0.16133, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:49.041833+00:00", "epoch": 0, "step": 16739, "train_loss": 3.472400665283203, "perplexity": 32.213984685942094, "lr": 0.0026291804804649314, "grad_norm": 0.157011, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:49.348757+00:00", "epoch": 0, "step": 16740, "train_loss": 3.4895946979522705, "perplexity": 32.772662186995746, "lr": 0.0026291804804649314, "grad_norm": 0.17976, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:49.654059+00:00", "epoch": 0, "step": 16741, "train_loss": 3.550194025039673, "perplexity": 34.82007279823721, "lr": 0.0026291804804649314, "grad_norm": 0.171469, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:49.959364+00:00", "epoch": 0, "step": 16742, "train_loss": 3.4703238010406494, "perplexity": 32.14715004030405, "lr": 0.0026291804804649314, "grad_norm": 0.166828, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:50.265124+00:00", "epoch": 0, "step": 16743, "train_loss": 3.460186243057251, "perplexity": 31.822902757489928, "lr": 0.0026291804804649314, "grad_norm": 0.184838, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:50.570611+00:00", "epoch": 0, "step": 16744, "train_loss": 3.4370415210723877, "perplexity": 31.09482855870428, "lr": 0.0026291804804649314, "grad_norm": 0.169188, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:50.876719+00:00", "epoch": 0, "step": 16745, "train_loss": 3.5288984775543213, "perplexity": 34.08639999307891, "lr": 0.0026291804804649314, "grad_norm": 0.155626, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:51.182184+00:00", "epoch": 0, "step": 16746, "train_loss": 3.382075309753418, "perplexity": 29.431787853464787, "lr": 0.0026291804804649314, "grad_norm": 0.197733, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:51.487892+00:00", "epoch": 0, "step": 16747, "train_loss": 3.45933198928833, "perplexity": 31.795729530939404, "lr": 0.0026291804804649314, "grad_norm": 0.172812, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:51.793008+00:00", "epoch": 0, "step": 16748, "train_loss": 3.4443511962890625, "perplexity": 31.322954403548646, "lr": 0.0026291804804649314, "grad_norm": 0.169369, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:52.098317+00:00", "epoch": 0, "step": 16749, "train_loss": 3.397042989730835, "perplexity": 29.875626768514525, "lr": 0.0026291804804649314, "grad_norm": 0.179666, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:52.403746+00:00", "epoch": 0, "step": 16750, "train_loss": 3.491729259490967, "perplexity": 32.84269216628085, "lr": 0.0026291804804649314, "grad_norm": 0.182436, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:52.709149+00:00", "epoch": 0, "step": 16751, "train_loss": 3.6281161308288574, "perplexity": 37.64183748329261, "lr": 0.0026291804804649314, "grad_norm": 0.163907, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:53.013870+00:00", "epoch": 0, "step": 16752, "train_loss": 3.4211816787719727, "perplexity": 30.605559601287883, "lr": 0.0026291804804649314, "grad_norm": 0.174959, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:53.319042+00:00", "epoch": 0, "step": 16753, "train_loss": 3.5673065185546875, "perplexity": 35.421058585022465, "lr": 0.0026291804804649314, "grad_norm": 0.144941, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:53.624015+00:00", "epoch": 0, "step": 16754, "train_loss": 3.501596689224243, "perplexity": 33.1683692790059, "lr": 0.0026291804804649314, "grad_norm": 0.167274, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:53.928880+00:00", "epoch": 0, "step": 16755, "train_loss": 3.431856870651245, "perplexity": 30.934029945218374, "lr": 0.0026291804804649314, "grad_norm": 0.152221, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:54.233082+00:00", "epoch": 0, "step": 16756, "train_loss": 3.5624752044677734, "perplexity": 35.2503410527846, "lr": 0.0026291804804649314, "grad_norm": 0.153022, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:54.537029+00:00", "epoch": 0, "step": 16757, "train_loss": 3.5256617069244385, "perplexity": 33.97624849843275, "lr": 0.0026291804804649314, "grad_norm": 0.156165, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:54.841918+00:00", "epoch": 0, "step": 16758, "train_loss": 3.538710594177246, "perplexity": 34.42250598403738, "lr": 0.0026291804804649314, "grad_norm": 0.186127, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:55.146971+00:00", "epoch": 0, "step": 16759, "train_loss": 3.5786325931549072, "perplexity": 35.82452064466909, "lr": 0.0026291804804649314, "grad_norm": 0.171414, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:55.451488+00:00", "epoch": 0, "step": 16760, "train_loss": 3.4661052227020264, "perplexity": 32.011820419497326, "lr": 0.0026291804804649314, "grad_norm": 0.187259, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:55.755521+00:00", "epoch": 0, "step": 16761, "train_loss": 3.553305149078369, "perplexity": 34.92857105199797, "lr": 0.0026291804804649314, "grad_norm": 0.199693, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:56.059416+00:00", "epoch": 0, "step": 16762, "train_loss": 3.590998649597168, "perplexity": 36.27027914596068, "lr": 0.0026291804804649314, "grad_norm": 0.176661, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:56.363551+00:00", "epoch": 0, "step": 16763, "train_loss": 3.5799145698547363, "perplexity": 35.87047629616145, "lr": 0.0026291804804649314, "grad_norm": 0.173959, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:56.668871+00:00", "epoch": 0, "step": 16764, "train_loss": 3.534050941467285, "perplexity": 34.262482177967165, "lr": 0.0026291804804649314, "grad_norm": 0.173216, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:56.974591+00:00", "epoch": 0, "step": 16765, "train_loss": 3.4828901290893555, "perplexity": 32.55367055909503, "lr": 0.0026291804804649314, "grad_norm": 0.165015, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:57.279821+00:00", "epoch": 0, "step": 16766, "train_loss": 3.5544962882995605, "perplexity": 34.970200631308465, "lr": 0.0026291804804649314, "grad_norm": 0.17598, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:57.584121+00:00", "epoch": 0, "step": 16767, "train_loss": 3.655569076538086, "perplexity": 38.68953215660358, "lr": 0.0026291804804649314, "grad_norm": 0.170112, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:57.888766+00:00", "epoch": 0, "step": 16768, "train_loss": 3.4725089073181152, "perplexity": 32.2174717819189, "lr": 0.0026291804804649314, "grad_norm": 0.165874, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:58.194394+00:00", "epoch": 0, "step": 16769, "train_loss": 3.5231850147247314, "perplexity": 33.892203908052906, "lr": 0.0026291804804649314, "grad_norm": 0.180823, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:58.500918+00:00", "epoch": 0, "step": 16770, "train_loss": 3.4193286895751953, "perplexity": 30.548900340696214, "lr": 0.0026291804804649314, "grad_norm": 0.167047, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:58.806456+00:00", "epoch": 0, "step": 16771, "train_loss": 3.457909345626831, "perplexity": 31.75052769853338, "lr": 0.0026291804804649314, "grad_norm": 0.189502, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:59.112743+00:00", "epoch": 0, "step": 16772, "train_loss": 3.5323221683502197, "perplexity": 34.203301289757526, "lr": 0.0026291804804649314, "grad_norm": 0.174195, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:59.417475+00:00", "epoch": 0, "step": 16773, "train_loss": 3.5169661045074463, "perplexity": 33.68208536624743, "lr": 0.0026291804804649314, "grad_norm": 0.166047, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:25:59.721695+00:00", "epoch": 0, "step": 16774, "train_loss": 3.633897304534912, "perplexity": 37.86008173050464, "lr": 0.0026291804804649314, "grad_norm": 0.16025, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:00.026896+00:00", "epoch": 0, "step": 16775, "train_loss": 3.548306703567505, "perplexity": 34.754418102429625, "lr": 0.0026291804804649314, "grad_norm": 0.199408, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:00.332422+00:00", "epoch": 0, "step": 16776, "train_loss": 3.4945764541625977, "perplexity": 32.93633495077749, "lr": 0.0026291804804649314, "grad_norm": 0.193248, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:00.637371+00:00", "epoch": 0, "step": 16777, "train_loss": 3.4571728706359863, "perplexity": 31.727152837494106, "lr": 0.0026291804804649314, "grad_norm": 0.155306, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:00.941332+00:00", "epoch": 0, "step": 16778, "train_loss": 3.5204126834869385, "perplexity": 33.79837361686151, "lr": 0.0026291804804649314, "grad_norm": 0.193897, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:01.245278+00:00", "epoch": 0, "step": 16779, "train_loss": 3.564053535461426, "perplexity": 35.306021688282435, "lr": 0.0026291804804649314, "grad_norm": 0.192389, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:01.551521+00:00", "epoch": 0, "step": 16780, "train_loss": 3.5051567554473877, "perplexity": 33.28666130894623, "lr": 0.0026291804804649314, "grad_norm": 0.158354, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:01.857223+00:00", "epoch": 0, "step": 16781, "train_loss": 3.5584301948547363, "perplexity": 35.1080410807678, "lr": 0.0026291804804649314, "grad_norm": 0.16984, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:02.163014+00:00", "epoch": 0, "step": 16782, "train_loss": 3.5185117721557617, "perplexity": 33.73418693140474, "lr": 0.0026291804804649314, "grad_norm": 0.164131, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:02.468780+00:00", "epoch": 0, "step": 16783, "train_loss": 3.559929132461548, "perplexity": 35.16070530417719, "lr": 0.0026291804804649314, "grad_norm": 0.14777, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:02.773696+00:00", "epoch": 0, "step": 16784, "train_loss": 3.5468897819519043, "perplexity": 34.70520868735763, "lr": 0.0026291804804649314, "grad_norm": 0.179223, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:03.078184+00:00", "epoch": 0, "step": 16785, "train_loss": 3.546173334121704, "perplexity": 34.68035312081848, "lr": 0.0026291804804649314, "grad_norm": 0.160272, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:03.383383+00:00", "epoch": 0, "step": 16786, "train_loss": 3.554908037185669, "perplexity": 34.98460253724629, "lr": 0.0026291804804649314, "grad_norm": 0.179079, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:03.688749+00:00", "epoch": 0, "step": 16787, "train_loss": 3.537679433822632, "perplexity": 34.387029154853565, "lr": 0.0026291804804649314, "grad_norm": 0.174011, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:03.993076+00:00", "epoch": 0, "step": 16788, "train_loss": 3.483060598373413, "perplexity": 32.559220433036764, "lr": 0.0026291804804649314, "grad_norm": 0.155786, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:04.298268+00:00", "epoch": 0, "step": 16789, "train_loss": 3.4860856533050537, "perplexity": 32.657862987728564, "lr": 0.0026291804804649314, "grad_norm": 0.182609, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:04.603021+00:00", "epoch": 0, "step": 16790, "train_loss": 3.409932851791382, "perplexity": 30.26321207069435, "lr": 0.0026291804804649314, "grad_norm": 0.153558, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:04.908144+00:00", "epoch": 0, "step": 16791, "train_loss": 3.5991017818450928, "perplexity": 36.56537600429569, "lr": 0.0026291804804649314, "grad_norm": 0.164352, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:05.213503+00:00", "epoch": 0, "step": 16792, "train_loss": 3.507537364959717, "perplexity": 33.3659982491604, "lr": 0.0026291804804649314, "grad_norm": 0.162162, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:05.518798+00:00", "epoch": 0, "step": 16793, "train_loss": 3.4235987663269043, "perplexity": 30.67962539420541, "lr": 0.0026291804804649314, "grad_norm": 0.16638, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:05.823657+00:00", "epoch": 0, "step": 16794, "train_loss": 3.6361663341522217, "perplexity": 37.94608491224928, "lr": 0.0026291804804649314, "grad_norm": 0.155263, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:06.128300+00:00", "epoch": 0, "step": 16795, "train_loss": 3.5904386043548584, "perplexity": 36.24997183572477, "lr": 0.0026291804804649314, "grad_norm": 0.141949, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:06.433661+00:00", "epoch": 0, "step": 16796, "train_loss": 3.41860032081604, "perplexity": 30.52665757751177, "lr": 0.0026291804804649314, "grad_norm": 0.158376, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:06.739408+00:00", "epoch": 0, "step": 16797, "train_loss": 3.52009916305542, "perplexity": 33.78777879711354, "lr": 0.0026291804804649314, "grad_norm": 0.15777, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:07.044790+00:00", "epoch": 0, "step": 16798, "train_loss": 3.4547204971313477, "perplexity": 31.64944133625842, "lr": 0.0026291804804649314, "grad_norm": 0.174075, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:07.349415+00:00", "epoch": 0, "step": 16799, "train_loss": 3.588430881500244, "perplexity": 36.177264950851445, "lr": 0.0026291804804649314, "grad_norm": 0.157946, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:07.654791+00:00", "epoch": 0, "step": 16800, "train_loss": 3.4859511852264404, "perplexity": 32.653471842882006, "lr": 0.0026291804804649314, "grad_norm": 0.171694, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:07.959756+00:00", "epoch": 0, "step": 16801, "train_loss": 3.4785826206207275, "perplexity": 32.41374692494885, "lr": 0.0026291804804649314, "grad_norm": 0.168109, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:08.266082+00:00", "epoch": 0, "step": 16802, "train_loss": 3.483605146408081, "perplexity": 32.576955320844796, "lr": 0.0026291804804649314, "grad_norm": 0.171385, "tokens_per_sec": 106971} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:08.571594+00:00", "epoch": 0, "step": 16803, "train_loss": 3.494826078414917, "perplexity": 32.94455768501797, "lr": 0.0026291804804649314, "grad_norm": 0.172499, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:08.876692+00:00", "epoch": 0, "step": 16804, "train_loss": 3.5165114402770996, "perplexity": 33.66677480767333, "lr": 0.0026291804804649314, "grad_norm": 0.176959, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:09.181598+00:00", "epoch": 0, "step": 16805, "train_loss": 3.526038408279419, "perplexity": 33.98904980826326, "lr": 0.0026291804804649314, "grad_norm": 0.159, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:09.487249+00:00", "epoch": 0, "step": 16806, "train_loss": 3.418119192123413, "perplexity": 30.51197385932355, "lr": 0.0026291804804649314, "grad_norm": 0.162838, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:09.792262+00:00", "epoch": 0, "step": 16807, "train_loss": 3.586121082305908, "perplexity": 36.093799165125695, "lr": 0.0026291804804649314, "grad_norm": 0.172465, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:10.097721+00:00", "epoch": 0, "step": 16808, "train_loss": 3.498090982437134, "perplexity": 33.052294283037654, "lr": 0.0026291804804649314, "grad_norm": 0.185454, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:10.403777+00:00", "epoch": 0, "step": 16809, "train_loss": 3.541063070297241, "perplexity": 34.503579431640595, "lr": 0.0026291804804649314, "grad_norm": 0.164348, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:10.709016+00:00", "epoch": 0, "step": 16810, "train_loss": 3.4692299365997314, "perplexity": 32.11200464165343, "lr": 0.0026291804804649314, "grad_norm": 0.160158, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:11.014658+00:00", "epoch": 0, "step": 16811, "train_loss": 3.539375066757202, "perplexity": 34.445386396258556, "lr": 0.0026291804804649314, "grad_norm": 0.18933, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:11.319288+00:00", "epoch": 0, "step": 16812, "train_loss": 3.500638484954834, "perplexity": 33.13660242795015, "lr": 0.0026291804804649314, "grad_norm": 0.14563, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:11.624951+00:00", "epoch": 0, "step": 16813, "train_loss": 3.5168116092681885, "perplexity": 33.67688204636449, "lr": 0.0026291804804649314, "grad_norm": 0.194533, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:11.930255+00:00", "epoch": 0, "step": 16814, "train_loss": 3.598952293395996, "perplexity": 36.559910311485325, "lr": 0.0026291804804649314, "grad_norm": 0.229262, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:12.234829+00:00", "epoch": 0, "step": 16815, "train_loss": 3.4222636222839355, "perplexity": 30.638691007851705, "lr": 0.0026291804804649314, "grad_norm": 0.156522, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:12.539614+00:00", "epoch": 0, "step": 16816, "train_loss": 3.447056293487549, "perplexity": 31.40780074681779, "lr": 0.0026291804804649314, "grad_norm": 0.177475, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:12.844258+00:00", "epoch": 0, "step": 16817, "train_loss": 3.3989055156707764, "perplexity": 29.93132274984709, "lr": 0.0026291804804649314, "grad_norm": 0.192231, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:13.149525+00:00", "epoch": 0, "step": 16818, "train_loss": 3.526378870010376, "perplexity": 34.00062374912474, "lr": 0.0026291804804649314, "grad_norm": 0.155992, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:13.455176+00:00", "epoch": 0, "step": 16819, "train_loss": 3.4953043460845947, "perplexity": 32.96031777031977, "lr": 0.0026291804804649314, "grad_norm": 0.196823, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:13.759837+00:00", "epoch": 0, "step": 16820, "train_loss": 3.4776947498321533, "perplexity": 32.38498047823957, "lr": 0.0026291804804649314, "grad_norm": 0.165522, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:14.064390+00:00", "epoch": 0, "step": 16821, "train_loss": 3.5396342277526855, "perplexity": 34.45431445373898, "lr": 0.0026291804804649314, "grad_norm": 0.16492, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:14.368225+00:00", "epoch": 0, "step": 16822, "train_loss": 3.456080913543701, "perplexity": 31.692527056309423, "lr": 0.0026291804804649314, "grad_norm": 0.157269, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:14.672543+00:00", "epoch": 0, "step": 16823, "train_loss": 3.592783212661743, "perplexity": 36.33506353518744, "lr": 0.0026291804804649314, "grad_norm": 0.166352, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:14.978130+00:00", "epoch": 0, "step": 16824, "train_loss": 3.429384708404541, "perplexity": 30.85765045439619, "lr": 0.0026291804804649314, "grad_norm": 0.170309, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:15.283312+00:00", "epoch": 0, "step": 16825, "train_loss": 3.544229030609131, "perplexity": 34.61298949731461, "lr": 0.0026291804804649314, "grad_norm": 0.177461, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:15.587500+00:00", "epoch": 0, "step": 16826, "train_loss": 3.4579665660858154, "perplexity": 31.752344530280762, "lr": 0.0026291804804649314, "grad_norm": 0.154267, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:15.892177+00:00", "epoch": 0, "step": 16827, "train_loss": 3.535440444946289, "perplexity": 34.310123107106044, "lr": 0.0026291804804649314, "grad_norm": 0.161545, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:16.197519+00:00", "epoch": 0, "step": 16828, "train_loss": 3.4608154296875, "perplexity": 31.84293160272073, "lr": 0.0026291804804649314, "grad_norm": 0.177359, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:16.502875+00:00", "epoch": 0, "step": 16829, "train_loss": 3.5233442783355713, "perplexity": 33.897602132685584, "lr": 0.0026291804804649314, "grad_norm": 0.16013, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:16.807603+00:00", "epoch": 0, "step": 16830, "train_loss": 3.4599459171295166, "perplexity": 31.815255807778478, "lr": 0.0026291804804649314, "grad_norm": 0.167313, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:17.112587+00:00", "epoch": 0, "step": 16831, "train_loss": 3.5155599117279053, "perplexity": 33.634755146511296, "lr": 0.0026291804804649314, "grad_norm": 0.159621, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:17.416787+00:00", "epoch": 0, "step": 16832, "train_loss": 3.599344253540039, "perplexity": 36.57424314796404, "lr": 0.0026291804804649314, "grad_norm": 0.197104, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:17.721025+00:00", "epoch": 0, "step": 16833, "train_loss": 3.406632423400879, "perplexity": 30.163495151181092, "lr": 0.0026291804804649314, "grad_norm": 0.185038, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:18.026385+00:00", "epoch": 0, "step": 16834, "train_loss": 3.428349018096924, "perplexity": 30.825708028990196, "lr": 0.0026291804804649314, "grad_norm": 0.17744, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:18.331053+00:00", "epoch": 0, "step": 16835, "train_loss": 3.420445203781128, "perplexity": 30.58302767018104, "lr": 0.0026291804804649314, "grad_norm": 0.150447, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:18.635953+00:00", "epoch": 0, "step": 16836, "train_loss": 3.540745973587036, "perplexity": 34.492640194602224, "lr": 0.0026291804804649314, "grad_norm": 0.19034, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:18.940941+00:00", "epoch": 0, "step": 16837, "train_loss": 3.495898723602295, "perplexity": 32.97991446551929, "lr": 0.0026291804804649314, "grad_norm": 0.166985, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:19.245325+00:00", "epoch": 0, "step": 16838, "train_loss": 3.5434670448303223, "perplexity": 34.58662493754106, "lr": 0.0026291804804649314, "grad_norm": 0.149941, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:19.550387+00:00", "epoch": 0, "step": 16839, "train_loss": 3.4489612579345703, "perplexity": 31.467688514523253, "lr": 0.0026291804804649314, "grad_norm": 0.162643, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:19.857033+00:00", "epoch": 0, "step": 16840, "train_loss": 3.52683424949646, "perplexity": 34.01611046159211, "lr": 0.0026291804804649314, "grad_norm": 0.148793, "tokens_per_sec": 106960} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:20.162882+00:00", "epoch": 0, "step": 16841, "train_loss": 3.5163636207580566, "perplexity": 33.66179856901519, "lr": 0.0026291804804649314, "grad_norm": 0.16, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:20.467237+00:00", "epoch": 0, "step": 16842, "train_loss": 3.604046106338501, "perplexity": 36.74661477090223, "lr": 0.0026291804804649314, "grad_norm": 0.178268, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:20.772051+00:00", "epoch": 0, "step": 16843, "train_loss": 3.5022053718566895, "perplexity": 33.18856443492942, "lr": 0.0026291804804649314, "grad_norm": 0.174051, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:21.076624+00:00", "epoch": 0, "step": 16844, "train_loss": 3.5213255882263184, "perplexity": 33.829242400305205, "lr": 0.0026291804804649314, "grad_norm": 0.181132, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:21.381469+00:00", "epoch": 0, "step": 16845, "train_loss": 3.5334794521331787, "perplexity": 34.242907128841104, "lr": 0.0026291804804649314, "grad_norm": 0.167469, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:21.686616+00:00", "epoch": 0, "step": 16846, "train_loss": 3.5354132652282715, "perplexity": 34.309190580307806, "lr": 0.0026291804804649314, "grad_norm": 0.179795, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:21.991008+00:00", "epoch": 0, "step": 16847, "train_loss": 3.428082227706909, "perplexity": 30.81748512326755, "lr": 0.0026291804804649314, "grad_norm": 0.182632, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:22.295738+00:00", "epoch": 0, "step": 16848, "train_loss": 3.501176357269287, "perplexity": 33.15443048317003, "lr": 0.0026291804804649314, "grad_norm": 0.164192, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:22.599987+00:00", "epoch": 0, "step": 16849, "train_loss": 3.469029664993286, "perplexity": 32.10557416284032, "lr": 0.0026291804804649314, "grad_norm": 0.192142, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:22.904747+00:00", "epoch": 0, "step": 16850, "train_loss": 3.476841926574707, "perplexity": 32.35737358726835, "lr": 0.0026291804804649314, "grad_norm": 0.17617, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:23.210742+00:00", "epoch": 0, "step": 16851, "train_loss": 3.6404776573181152, "perplexity": 38.110035916113645, "lr": 0.0026291804804649314, "grad_norm": 0.168974, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:23.516299+00:00", "epoch": 0, "step": 16852, "train_loss": 3.5835936069488525, "perplexity": 36.00268816609814, "lr": 0.0026291804804649314, "grad_norm": 0.159927, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:23.820920+00:00", "epoch": 0, "step": 16853, "train_loss": 3.6050026416778564, "perplexity": 36.78178102273228, "lr": 0.0026291804804649314, "grad_norm": 0.162374, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:24.124881+00:00", "epoch": 0, "step": 16854, "train_loss": 3.644113063812256, "perplexity": 38.24883352818067, "lr": 0.0026291804804649314, "grad_norm": 0.176049, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:24.429427+00:00", "epoch": 0, "step": 16855, "train_loss": 3.508251190185547, "perplexity": 33.38982424318173, "lr": 0.0026291804804649314, "grad_norm": 0.165518, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:24.734575+00:00", "epoch": 0, "step": 16856, "train_loss": 3.474334716796875, "perplexity": 32.27634847974063, "lr": 0.0026291804804649314, "grad_norm": 0.171104, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:25.041027+00:00", "epoch": 0, "step": 16857, "train_loss": 3.5363690853118896, "perplexity": 34.34199967101321, "lr": 0.0026291804804649314, "grad_norm": 0.172243, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:25.345928+00:00", "epoch": 0, "step": 16858, "train_loss": 3.5306644439697266, "perplexity": 34.14664861355113, "lr": 0.0026291804804649314, "grad_norm": 0.164046, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:25.651018+00:00", "epoch": 0, "step": 16859, "train_loss": 3.6231324672698975, "perplexity": 37.45470990728117, "lr": 0.0026291804804649314, "grad_norm": 0.179598, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:25.956561+00:00", "epoch": 0, "step": 16860, "train_loss": 3.4909775257110596, "perplexity": 32.81801248259453, "lr": 0.0026291804804649314, "grad_norm": 0.196895, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:26.262590+00:00", "epoch": 0, "step": 16861, "train_loss": 3.40268874168396, "perplexity": 30.04477417956612, "lr": 0.0026291804804649314, "grad_norm": 0.209892, "tokens_per_sec": 107017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:26.567967+00:00", "epoch": 0, "step": 16862, "train_loss": 3.6832358837127686, "perplexity": 39.77489298506568, "lr": 0.0026291804804649314, "grad_norm": 0.196892, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:26.874220+00:00", "epoch": 0, "step": 16863, "train_loss": 3.4909400939941406, "perplexity": 32.81678407103236, "lr": 0.0026291804804649314, "grad_norm": 0.16732, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:27.179327+00:00", "epoch": 0, "step": 16864, "train_loss": 3.526632785797119, "perplexity": 34.0092581404115, "lr": 0.0026291804804649314, "grad_norm": 0.18525, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:27.483782+00:00", "epoch": 0, "step": 16865, "train_loss": 3.5462193489074707, "perplexity": 34.68194896655361, "lr": 0.0026291804804649314, "grad_norm": 0.173614, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:27.788965+00:00", "epoch": 0, "step": 16866, "train_loss": 3.4620070457458496, "perplexity": 31.880898768012766, "lr": 0.0026291804804649314, "grad_norm": 0.173027, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:28.093888+00:00", "epoch": 0, "step": 16867, "train_loss": 3.4957644939422607, "perplexity": 32.97548787990881, "lr": 0.0026291804804649314, "grad_norm": 0.168637, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:28.399576+00:00", "epoch": 0, "step": 16868, "train_loss": 3.4979119300842285, "perplexity": 33.0463767217698, "lr": 0.0026291804804649314, "grad_norm": 0.191394, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:28.704419+00:00", "epoch": 0, "step": 16869, "train_loss": 3.536226749420166, "perplexity": 34.33711191972594, "lr": 0.0026291804804649314, "grad_norm": 0.198382, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:29.008864+00:00", "epoch": 0, "step": 16870, "train_loss": 3.4699316024780273, "perplexity": 32.13454444637534, "lr": 0.0026291804804649314, "grad_norm": 0.166086, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:29.313606+00:00", "epoch": 0, "step": 16871, "train_loss": 3.500485897064209, "perplexity": 33.131546569424394, "lr": 0.0026291804804649314, "grad_norm": 0.185602, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:29.618098+00:00", "epoch": 0, "step": 16872, "train_loss": 3.4825806617736816, "perplexity": 32.54359782072382, "lr": 0.0026291804804649314, "grad_norm": 0.157702, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:29.922876+00:00", "epoch": 0, "step": 16873, "train_loss": 3.4049010276794434, "perplexity": 30.111315389675777, "lr": 0.0026291804804649314, "grad_norm": 0.177651, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:30.227469+00:00", "epoch": 0, "step": 16874, "train_loss": 3.492920398712158, "perplexity": 32.88183569311655, "lr": 0.0026291804804649314, "grad_norm": 0.183991, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:30.532031+00:00", "epoch": 0, "step": 16875, "train_loss": 3.462944507598877, "perplexity": 31.910799907847135, "lr": 0.0026291804804649314, "grad_norm": 0.165562, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:30.836713+00:00", "epoch": 0, "step": 16876, "train_loss": 3.500152111053467, "perplexity": 33.120489568104105, "lr": 0.0026291804804649314, "grad_norm": 0.155876, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:31.140870+00:00", "epoch": 0, "step": 16877, "train_loss": 3.5519521236419678, "perplexity": 34.88134376397329, "lr": 0.0026291804804649314, "grad_norm": 0.177065, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:31.445916+00:00", "epoch": 0, "step": 16878, "train_loss": 3.5504770278930664, "perplexity": 34.829928372706554, "lr": 0.0026291804804649314, "grad_norm": 0.161603, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:31.750262+00:00", "epoch": 0, "step": 16879, "train_loss": 3.505347490310669, "perplexity": 33.29301084125942, "lr": 0.0026291804804649314, "grad_norm": 0.159423, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:32.055168+00:00", "epoch": 0, "step": 16880, "train_loss": 3.3566677570343018, "perplexity": 28.693417958388697, "lr": 0.0026291804804649314, "grad_norm": 0.163642, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:32.359286+00:00", "epoch": 0, "step": 16881, "train_loss": 3.384613513946533, "perplexity": 29.506586627929355, "lr": 0.0026291804804649314, "grad_norm": 0.15843, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:32.663937+00:00", "epoch": 0, "step": 16882, "train_loss": 3.481882333755493, "perplexity": 32.5208796478445, "lr": 0.0026291804804649314, "grad_norm": 0.156305, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:32.969402+00:00", "epoch": 0, "step": 16883, "train_loss": 3.5345613956451416, "perplexity": 34.279976069673836, "lr": 0.0026291804804649314, "grad_norm": 0.166097, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:33.273908+00:00", "epoch": 0, "step": 16884, "train_loss": 3.5157310962677, "perplexity": 33.640513389439185, "lr": 0.0026291804804649314, "grad_norm": 0.171508, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:33.577915+00:00", "epoch": 0, "step": 16885, "train_loss": 3.5082449913024902, "perplexity": 33.38961726420748, "lr": 0.0026291804804649314, "grad_norm": 0.160155, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:33.882004+00:00", "epoch": 0, "step": 16886, "train_loss": 3.548734664916992, "perplexity": 34.76929483320722, "lr": 0.0026291804804649314, "grad_norm": 0.164021, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:34.186899+00:00", "epoch": 0, "step": 16887, "train_loss": 3.5223946571350098, "perplexity": 33.865427530312324, "lr": 0.0026291804804649314, "grad_norm": 0.161957, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:34.492148+00:00", "epoch": 0, "step": 16888, "train_loss": 3.5807130336761475, "perplexity": 35.89912901129665, "lr": 0.0026291804804649314, "grad_norm": 0.160438, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:34.797720+00:00", "epoch": 0, "step": 16889, "train_loss": 3.577535390853882, "perplexity": 35.78523545402655, "lr": 0.0026291804804649314, "grad_norm": 0.175669, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:35.104249+00:00", "epoch": 0, "step": 16890, "train_loss": 3.5237228870391846, "perplexity": 33.91043848970955, "lr": 0.0026291804804649314, "grad_norm": 0.169793, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:35.410558+00:00", "epoch": 0, "step": 16891, "train_loss": 3.5095298290252686, "perplexity": 33.43254507574802, "lr": 0.0026291804804649314, "grad_norm": 0.179292, "tokens_per_sec": 106882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:35.714597+00:00", "epoch": 0, "step": 16892, "train_loss": 3.5782485008239746, "perplexity": 35.81076336323242, "lr": 0.0026291804804649314, "grad_norm": 0.201801, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:36.019309+00:00", "epoch": 0, "step": 16893, "train_loss": 3.5025644302368164, "perplexity": 33.20048320675357, "lr": 0.0026291804804649314, "grad_norm": 0.188422, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:36.324793+00:00", "epoch": 0, "step": 16894, "train_loss": 3.6522092819213867, "perplexity": 38.55976139836206, "lr": 0.0026291804804649314, "grad_norm": 0.178158, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:36.629022+00:00", "epoch": 0, "step": 16895, "train_loss": 3.5229759216308594, "perplexity": 33.885118023110245, "lr": 0.0026291804804649314, "grad_norm": 0.172217, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:36.933366+00:00", "epoch": 0, "step": 16896, "train_loss": 3.4667370319366455, "perplexity": 32.032052173886946, "lr": 0.0026291804804649314, "grad_norm": 0.211417, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:37.237880+00:00", "epoch": 0, "step": 16897, "train_loss": 3.4688525199890137, "perplexity": 32.09988732448045, "lr": 0.0026291804804649314, "grad_norm": 0.203941, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:37.542484+00:00", "epoch": 0, "step": 16898, "train_loss": 3.5983762741088867, "perplexity": 36.538857162103085, "lr": 0.0026291804804649314, "grad_norm": 0.160687, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:37.847566+00:00", "epoch": 0, "step": 16899, "train_loss": 3.5209317207336426, "perplexity": 33.815920785069686, "lr": 0.0026291804804649314, "grad_norm": 0.184741, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:38.152653+00:00", "epoch": 0, "step": 16900, "train_loss": 3.5111100673675537, "perplexity": 33.48541823044736, "lr": 0.0026291804804649314, "grad_norm": 0.169518, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:38.456422+00:00", "epoch": 0, "step": 16901, "train_loss": 3.5015370845794678, "perplexity": 33.16639234905494, "lr": 0.0026291804804649314, "grad_norm": 0.15773, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:38.761344+00:00", "epoch": 0, "step": 16902, "train_loss": 3.516373872756958, "perplexity": 33.66214367150613, "lr": 0.0026291804804649314, "grad_norm": 0.151584, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:39.066137+00:00", "epoch": 0, "step": 16903, "train_loss": 3.5369603633880615, "perplexity": 34.36231134684313, "lr": 0.0026291804804649314, "grad_norm": 0.167934, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:39.371384+00:00", "epoch": 0, "step": 16904, "train_loss": 3.5256736278533936, "perplexity": 33.976653529291426, "lr": 0.0026291804804649314, "grad_norm": 0.157651, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:39.677562+00:00", "epoch": 0, "step": 16905, "train_loss": 3.4996626377105713, "perplexity": 33.10428193828184, "lr": 0.0026291804804649314, "grad_norm": 0.162544, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:39.982587+00:00", "epoch": 0, "step": 16906, "train_loss": 3.433727502822876, "perplexity": 30.991950293775908, "lr": 0.0026291804804649314, "grad_norm": 0.163985, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:40.286981+00:00", "epoch": 0, "step": 16907, "train_loss": 3.548992872238159, "perplexity": 34.77827367883692, "lr": 0.0026291804804649314, "grad_norm": 0.178849, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:40.591470+00:00", "epoch": 0, "step": 16908, "train_loss": 3.524470329284668, "perplexity": 33.9357940587309, "lr": 0.0026291804804649314, "grad_norm": 0.16882, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:40.896344+00:00", "epoch": 0, "step": 16909, "train_loss": 3.520777702331543, "perplexity": 33.810712912049276, "lr": 0.0026291804804649314, "grad_norm": 0.185075, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:41.201831+00:00", "epoch": 0, "step": 16910, "train_loss": 3.405895471572876, "perplexity": 30.141274297145646, "lr": 0.0026291804804649314, "grad_norm": 0.214024, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:41.506726+00:00", "epoch": 0, "step": 16911, "train_loss": 3.4940712451934814, "perplexity": 32.91969942151368, "lr": 0.0026291804804649314, "grad_norm": 0.160227, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:41.810561+00:00", "epoch": 0, "step": 16912, "train_loss": 3.4693989753723145, "perplexity": 32.11743327431501, "lr": 0.0026291804804649314, "grad_norm": 0.183402, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:42.114771+00:00", "epoch": 0, "step": 16913, "train_loss": 3.4724299907684326, "perplexity": 32.21492939052607, "lr": 0.0026291804804649314, "grad_norm": 0.193785, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:42.418857+00:00", "epoch": 0, "step": 16914, "train_loss": 3.5438101291656494, "perplexity": 34.59849310254331, "lr": 0.0026291804804649314, "grad_norm": 0.163747, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:42.724861+00:00", "epoch": 0, "step": 16915, "train_loss": 3.5345592498779297, "perplexity": 34.27990251290408, "lr": 0.0026291804804649314, "grad_norm": 0.182093, "tokens_per_sec": 107086} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:43.029688+00:00", "epoch": 0, "step": 16916, "train_loss": 3.4844417572021484, "perplexity": 32.60422095707655, "lr": 0.0026291804804649314, "grad_norm": 0.187672, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:43.334337+00:00", "epoch": 0, "step": 16917, "train_loss": 3.434980630874634, "perplexity": 31.03081152002868, "lr": 0.0026291804804649314, "grad_norm": 0.15375, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:43.638665+00:00", "epoch": 0, "step": 16918, "train_loss": 3.5289466381073, "perplexity": 34.08804165248292, "lr": 0.0026291804804649314, "grad_norm": 0.202281, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:43.942922+00:00", "epoch": 0, "step": 16919, "train_loss": 3.596170425415039, "perplexity": 36.45834680123995, "lr": 0.0026291804804649314, "grad_norm": 0.226815, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:44.249378+00:00", "epoch": 0, "step": 16920, "train_loss": 3.487900733947754, "perplexity": 32.71719347118679, "lr": 0.0026291804804649314, "grad_norm": 0.181396, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:44.554508+00:00", "epoch": 0, "step": 16921, "train_loss": 3.490286350250244, "perplexity": 32.79533731485629, "lr": 0.0026291804804649314, "grad_norm": 0.173397, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:44.860566+00:00", "epoch": 0, "step": 16922, "train_loss": 3.5616025924682617, "perplexity": 35.21959459900831, "lr": 0.0026291804804649314, "grad_norm": 0.226164, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:45.166363+00:00", "epoch": 0, "step": 16923, "train_loss": 3.549180269241333, "perplexity": 34.78479163380343, "lr": 0.0026291804804649314, "grad_norm": 0.177855, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:45.471631+00:00", "epoch": 0, "step": 16924, "train_loss": 3.5883805751800537, "perplexity": 36.175445051553815, "lr": 0.0026291804804649314, "grad_norm": 0.161035, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:45.778056+00:00", "epoch": 0, "step": 16925, "train_loss": 3.5554394721984863, "perplexity": 35.00319952105056, "lr": 0.0026291804804649314, "grad_norm": 0.192385, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:46.084597+00:00", "epoch": 0, "step": 16926, "train_loss": 3.541792392730713, "perplexity": 34.5287528448068, "lr": 0.0026291804804649314, "grad_norm": 0.17306, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:46.390676+00:00", "epoch": 0, "step": 16927, "train_loss": 3.503147840499878, "perplexity": 33.21985836067726, "lr": 0.0026291804804649314, "grad_norm": 0.16106, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:46.695799+00:00", "epoch": 0, "step": 16928, "train_loss": 3.5965046882629395, "perplexity": 36.470535509074075, "lr": 0.0026291804804649314, "grad_norm": 0.155483, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:47.000896+00:00", "epoch": 0, "step": 16929, "train_loss": 3.5258872509002686, "perplexity": 33.983912500855176, "lr": 0.0026291804804649314, "grad_norm": 0.179089, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:47.307089+00:00", "epoch": 0, "step": 16930, "train_loss": 3.513417959213257, "perplexity": 33.562788200551196, "lr": 0.0026291804804649314, "grad_norm": 0.189456, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:47.613359+00:00", "epoch": 0, "step": 16931, "train_loss": 3.4613037109375, "perplexity": 31.858483705758534, "lr": 0.0026291804804649314, "grad_norm": 0.179615, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:47.920224+00:00", "epoch": 0, "step": 16932, "train_loss": 3.4871809482574463, "perplexity": 32.69365257671505, "lr": 0.0026291804804649314, "grad_norm": 0.173795, "tokens_per_sec": 106782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:48.226754+00:00", "epoch": 0, "step": 16933, "train_loss": 3.506573438644409, "perplexity": 33.333851381506484, "lr": 0.0026291804804649314, "grad_norm": 0.170916, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:48.532006+00:00", "epoch": 0, "step": 16934, "train_loss": 3.5523319244384766, "perplexity": 34.89459424222996, "lr": 0.0026291804804649314, "grad_norm": 0.197078, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:48.836690+00:00", "epoch": 0, "step": 16935, "train_loss": 3.4915616512298584, "perplexity": 32.83718792104811, "lr": 0.0026291804804649314, "grad_norm": 0.157, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:49.142739+00:00", "epoch": 0, "step": 16936, "train_loss": 3.4912097454071045, "perplexity": 32.8256343564184, "lr": 0.0026291804804649314, "grad_norm": 0.177325, "tokens_per_sec": 107059} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:49.449563+00:00", "epoch": 0, "step": 16937, "train_loss": 3.6150856018066406, "perplexity": 37.154526284097166, "lr": 0.0026291804804649314, "grad_norm": 0.222011, "tokens_per_sec": 106797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:49.755381+00:00", "epoch": 0, "step": 16938, "train_loss": 3.477360248565674, "perplexity": 32.37414947284819, "lr": 0.0026291804804649314, "grad_norm": 0.203679, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:50.060817+00:00", "epoch": 0, "step": 16939, "train_loss": 3.468590497970581, "perplexity": 32.09147754903351, "lr": 0.0026291804804649314, "grad_norm": 0.19627, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:50.366023+00:00", "epoch": 0, "step": 16940, "train_loss": 3.4511308670043945, "perplexity": 31.536035212928994, "lr": 0.0026291804804649314, "grad_norm": 0.209941, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:50.671355+00:00", "epoch": 0, "step": 16941, "train_loss": 3.486912488937378, "perplexity": 32.68487683899075, "lr": 0.0026291804804649314, "grad_norm": 0.171015, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:50.977297+00:00", "epoch": 0, "step": 16942, "train_loss": 3.570768356323242, "perplexity": 35.543893037233325, "lr": 0.0026291804804649314, "grad_norm": 0.186038, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:51.284163+00:00", "epoch": 0, "step": 16943, "train_loss": 3.498018980026245, "perplexity": 33.04991452383911, "lr": 0.0026291804804649314, "grad_norm": 0.168709, "tokens_per_sec": 106784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:51.590759+00:00", "epoch": 0, "step": 16944, "train_loss": 3.4523544311523438, "perplexity": 31.57464519105955, "lr": 0.0026291804804649314, "grad_norm": 0.151376, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:51.896284+00:00", "epoch": 0, "step": 16945, "train_loss": 3.516587257385254, "perplexity": 33.66932742194482, "lr": 0.0026291804804649314, "grad_norm": 0.160364, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:52.201386+00:00", "epoch": 0, "step": 16946, "train_loss": 3.4266514778137207, "perplexity": 30.77342453707476, "lr": 0.0026291804804649314, "grad_norm": 0.187458, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:52.507792+00:00", "epoch": 0, "step": 16947, "train_loss": 3.4172427654266357, "perplexity": 30.48524406592512, "lr": 0.0026291804804649314, "grad_norm": 0.160821, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:52.813406+00:00", "epoch": 0, "step": 16948, "train_loss": 3.4884262084960938, "perplexity": 32.7343900414333, "lr": 0.0026291804804649314, "grad_norm": 0.1841, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:53.118405+00:00", "epoch": 0, "step": 16949, "train_loss": 3.465925931930542, "perplexity": 32.006081509999824, "lr": 0.0026291804804649314, "grad_norm": 0.160965, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:53.424323+00:00", "epoch": 0, "step": 16950, "train_loss": 3.498430013656616, "perplexity": 33.06350194244099, "lr": 0.0026291804804649314, "grad_norm": 0.160974, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:53.728864+00:00", "epoch": 0, "step": 16951, "train_loss": 3.4580373764038086, "perplexity": 31.75459300350059, "lr": 0.0026291804804649314, "grad_norm": 0.182152, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:54.034390+00:00", "epoch": 0, "step": 16952, "train_loss": 3.4652507305145264, "perplexity": 31.984478252538914, "lr": 0.0026291804804649314, "grad_norm": 0.149355, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:54.339292+00:00", "epoch": 0, "step": 16953, "train_loss": 3.506274461746216, "perplexity": 33.32388681967672, "lr": 0.0026291804804649314, "grad_norm": 0.190755, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:54.644323+00:00", "epoch": 0, "step": 16954, "train_loss": 3.5441343784332275, "perplexity": 34.60971345758876, "lr": 0.0026291804804649314, "grad_norm": 0.184461, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:54.949888+00:00", "epoch": 0, "step": 16955, "train_loss": 3.502835512161255, "perplexity": 33.209484477619334, "lr": 0.0026291804804649314, "grad_norm": 0.175024, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:55.255892+00:00", "epoch": 0, "step": 16956, "train_loss": 3.5078747272491455, "perplexity": 33.37725657767977, "lr": 0.0026291804804649314, "grad_norm": 0.202349, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:55.561414+00:00", "epoch": 0, "step": 16957, "train_loss": 3.430297613143921, "perplexity": 30.885833411962608, "lr": 0.0026291804804649314, "grad_norm": 0.201107, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:55.867830+00:00", "epoch": 0, "step": 16958, "train_loss": 3.450951337814331, "perplexity": 31.53037408225371, "lr": 0.0026291804804649314, "grad_norm": 0.167392, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:56.173398+00:00", "epoch": 0, "step": 16959, "train_loss": 3.5276262760162354, "perplexity": 34.04306279524987, "lr": 0.0026291804804649314, "grad_norm": 0.175701, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:56.479413+00:00", "epoch": 0, "step": 16960, "train_loss": 3.50227952003479, "perplexity": 33.19102539775287, "lr": 0.0026291804804649314, "grad_norm": 0.20341, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:56.784401+00:00", "epoch": 0, "step": 16961, "train_loss": 3.6487276554107666, "perplexity": 38.42574414522169, "lr": 0.0026291804804649314, "grad_norm": 0.158891, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:57.091400+00:00", "epoch": 0, "step": 16962, "train_loss": 3.514509439468384, "perplexity": 33.5994413206187, "lr": 0.0026291804804649314, "grad_norm": 0.169452, "tokens_per_sec": 106738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:57.397175+00:00", "epoch": 0, "step": 16963, "train_loss": 3.4852747917175293, "perplexity": 32.63139271441544, "lr": 0.0026291804804649314, "grad_norm": 0.16579, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:57.702223+00:00", "epoch": 0, "step": 16964, "train_loss": 3.570816993713379, "perplexity": 35.54562184146788, "lr": 0.0026291804804649314, "grad_norm": 0.152884, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:58.007309+00:00", "epoch": 0, "step": 16965, "train_loss": 3.581641912460327, "perplexity": 35.9324904425699, "lr": 0.0026291804804649314, "grad_norm": 0.187298, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:58.313642+00:00", "epoch": 0, "step": 16966, "train_loss": 3.5904018878936768, "perplexity": 36.248640889475, "lr": 0.0026291804804649314, "grad_norm": 0.172295, "tokens_per_sec": 106969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:58.619253+00:00", "epoch": 0, "step": 16967, "train_loss": 3.4835286140441895, "perplexity": 32.57446222484818, "lr": 0.0026291804804649314, "grad_norm": 0.184047, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:58.924743+00:00", "epoch": 0, "step": 16968, "train_loss": 3.494633674621582, "perplexity": 32.93821963690143, "lr": 0.0026291804804649314, "grad_norm": 0.199485, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:59.229729+00:00", "epoch": 0, "step": 16969, "train_loss": 3.5265207290649414, "perplexity": 34.005447387594735, "lr": 0.0026291804804649314, "grad_norm": 0.184514, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:59.535694+00:00", "epoch": 0, "step": 16970, "train_loss": 3.555222749710083, "perplexity": 34.995614362515184, "lr": 0.0026291804804649314, "grad_norm": 0.156403, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:26:59.841688+00:00", "epoch": 0, "step": 16971, "train_loss": 3.5232484340667725, "perplexity": 33.894353397483954, "lr": 0.0026291804804649314, "grad_norm": 0.168675, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:00.147627+00:00", "epoch": 0, "step": 16972, "train_loss": 3.4849066734313965, "perplexity": 32.61938271274152, "lr": 0.0026291804804649314, "grad_norm": 0.15766, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:00.452446+00:00", "epoch": 0, "step": 16973, "train_loss": 3.358410120010376, "perplexity": 28.7434558869546, "lr": 0.0026291804804649314, "grad_norm": 0.164687, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:00.756730+00:00", "epoch": 0, "step": 16974, "train_loss": 3.4377756118774414, "perplexity": 31.11766336681983, "lr": 0.0026291804804649314, "grad_norm": 0.171983, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:01.062058+00:00", "epoch": 0, "step": 16975, "train_loss": 3.6124279499053955, "perplexity": 37.05591358383626, "lr": 0.0026291804804649314, "grad_norm": 0.169976, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:01.367065+00:00", "epoch": 0, "step": 16976, "train_loss": 3.4866943359375, "perplexity": 32.67774731274973, "lr": 0.0026291804804649314, "grad_norm": 0.189096, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:01.672762+00:00", "epoch": 0, "step": 16977, "train_loss": 3.524949312210083, "perplexity": 33.95205261812076, "lr": 0.0026291804804649314, "grad_norm": 0.159723, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:01.977322+00:00", "epoch": 0, "step": 16978, "train_loss": 3.458402633666992, "perplexity": 31.76619371772815, "lr": 0.0026291804804649314, "grad_norm": 0.183824, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:02.282741+00:00", "epoch": 0, "step": 16979, "train_loss": 3.5586109161376953, "perplexity": 35.11438642434614, "lr": 0.0026291804804649314, "grad_norm": 0.162107, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:02.587813+00:00", "epoch": 0, "step": 16980, "train_loss": 3.588359832763672, "perplexity": 36.1746946931919, "lr": 0.0026291804804649314, "grad_norm": 0.171273, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:02.893171+00:00", "epoch": 0, "step": 16981, "train_loss": 3.443727970123291, "perplexity": 31.3034392005993, "lr": 0.0026291804804649314, "grad_norm": 0.188094, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:03.197412+00:00", "epoch": 0, "step": 16982, "train_loss": 3.4015052318573, "perplexity": 30.00923692757636, "lr": 0.0026291804804649314, "grad_norm": 0.157127, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:03.501872+00:00", "epoch": 0, "step": 16983, "train_loss": 3.466977119445801, "perplexity": 32.03974359277636, "lr": 0.0026291804804649314, "grad_norm": 0.192067, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:03.806665+00:00", "epoch": 0, "step": 16984, "train_loss": 3.513892889022827, "perplexity": 33.57873195494236, "lr": 0.0026291804804649314, "grad_norm": 0.201954, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:04.111667+00:00", "epoch": 0, "step": 16985, "train_loss": 3.5128138065338135, "perplexity": 33.54251727611348, "lr": 0.0026291804804649314, "grad_norm": 0.164522, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:04.417740+00:00", "epoch": 0, "step": 16986, "train_loss": 3.456329345703125, "perplexity": 31.70040147733283, "lr": 0.0026291804804649314, "grad_norm": 0.158182, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:04.722513+00:00", "epoch": 0, "step": 16987, "train_loss": 3.398075819015503, "perplexity": 29.906499130934467, "lr": 0.0026291804804649314, "grad_norm": 0.170879, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:05.027840+00:00", "epoch": 0, "step": 16988, "train_loss": 3.4623382091522217, "perplexity": 31.891458303416336, "lr": 0.0026291804804649314, "grad_norm": 0.176353, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:05.332751+00:00", "epoch": 0, "step": 16989, "train_loss": 3.4886515140533447, "perplexity": 32.74176611232654, "lr": 0.0026291804804649314, "grad_norm": 0.184559, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:05.637808+00:00", "epoch": 0, "step": 16990, "train_loss": 3.506946086883545, "perplexity": 33.34627549729608, "lr": 0.0026291804804649314, "grad_norm": 0.190068, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:05.941888+00:00", "epoch": 0, "step": 16991, "train_loss": 3.4567983150482178, "perplexity": 31.71527148036809, "lr": 0.0026291804804649314, "grad_norm": 0.160831, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:06.247882+00:00", "epoch": 0, "step": 16992, "train_loss": 3.601480007171631, "perplexity": 36.65244019568006, "lr": 0.0026291804804649314, "grad_norm": 0.161261, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:06.552425+00:00", "epoch": 0, "step": 16993, "train_loss": 3.5221774578094482, "perplexity": 33.85807278104509, "lr": 0.0026291804804649314, "grad_norm": 0.17057, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:06.857685+00:00", "epoch": 0, "step": 16994, "train_loss": 3.4619901180267334, "perplexity": 31.88035910168092, "lr": 0.0026291804804649314, "grad_norm": 0.148349, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:07.163250+00:00", "epoch": 0, "step": 16995, "train_loss": 3.4908041954040527, "perplexity": 32.81232461936935, "lr": 0.0026291804804649314, "grad_norm": 0.156242, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:07.468411+00:00", "epoch": 0, "step": 16996, "train_loss": 3.4837725162506104, "perplexity": 32.582408177036044, "lr": 0.0026291804804649314, "grad_norm": 0.190841, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:07.774213+00:00", "epoch": 0, "step": 16997, "train_loss": 3.5610811710357666, "perplexity": 35.20123513446571, "lr": 0.0026291804804649314, "grad_norm": 0.166675, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:08.080370+00:00", "epoch": 0, "step": 16998, "train_loss": 3.5890424251556396, "perplexity": 36.199395693969755, "lr": 0.0026291804804649314, "grad_norm": 0.194774, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:08.385044+00:00", "epoch": 0, "step": 16999, "train_loss": 3.4633235931396484, "perplexity": 31.92289912386068, "lr": 0.0026291804804649314, "grad_norm": 0.186219, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:08.689877+00:00", "epoch": 0, "step": 17000, "train_loss": 3.533609628677368, "perplexity": 34.24736504230573, "lr": 0.0026291804804649314, "grad_norm": 0.167926, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:27:11.763264+00:00", "step": 17000, "epoch": 0, "val_loss": 3.4679521203041075, "val_ppl": 32.070997604147294, "eval_train_loss": 3.533609628677368, "eval_train_ppl": 34.24736504230573} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:27:12.725472+00:00", "step": 17000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4680_epoch_0000_step_0017000.pt", "val_loss": 3.4679521203041075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:14.252153+00:00", "epoch": 0, "step": 17001, "train_loss": 3.5298328399658203, "perplexity": 34.11826392788906, "lr": 0.0026291804804649314, "grad_norm": 0.201562, "tokens_per_sec": 5891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:14.556073+00:00", "epoch": 0, "step": 17002, "train_loss": 3.499621629714966, "perplexity": 33.10292442586821, "lr": 0.0026291804804649314, "grad_norm": 0.165946, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:14.859545+00:00", "epoch": 0, "step": 17003, "train_loss": 3.3947856426239014, "perplexity": 29.80826316897897, "lr": 0.0026291804804649314, "grad_norm": 0.153728, "tokens_per_sec": 107977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:15.163751+00:00", "epoch": 0, "step": 17004, "train_loss": 3.517782211303711, "perplexity": 33.70958476472389, "lr": 0.0026291804804649314, "grad_norm": 0.183145, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:15.467656+00:00", "epoch": 0, "step": 17005, "train_loss": 3.47941517829895, "perplexity": 32.44074447578085, "lr": 0.0026291804804649314, "grad_norm": 0.165226, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:15.773234+00:00", "epoch": 0, "step": 17006, "train_loss": 3.487109899520874, "perplexity": 32.69132981652111, "lr": 0.0026291804804649314, "grad_norm": 0.185815, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:16.077165+00:00", "epoch": 0, "step": 17007, "train_loss": 3.4327187538146973, "perplexity": 30.9607029576649, "lr": 0.0026291804804649314, "grad_norm": 0.176181, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:16.381303+00:00", "epoch": 0, "step": 17008, "train_loss": 3.484328031539917, "perplexity": 32.600513231292446, "lr": 0.0026291804804649314, "grad_norm": 0.162507, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:16.687882+00:00", "epoch": 0, "step": 17009, "train_loss": 3.421119451522827, "perplexity": 30.60365516076, "lr": 0.0026291804804649314, "grad_norm": 0.180062, "tokens_per_sec": 106885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:16.993256+00:00", "epoch": 0, "step": 17010, "train_loss": 3.4454634189605713, "perplexity": 31.357811884621928, "lr": 0.0026291804804649314, "grad_norm": 0.172069, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:17.297051+00:00", "epoch": 0, "step": 17011, "train_loss": 3.5044991970062256, "perplexity": 33.26478057854232, "lr": 0.0026291804804649314, "grad_norm": 0.140844, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:17.601138+00:00", "epoch": 0, "step": 17012, "train_loss": 3.619403839111328, "perplexity": 37.315315258153205, "lr": 0.0026291804804649314, "grad_norm": 0.177125, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:17.905014+00:00", "epoch": 0, "step": 17013, "train_loss": 3.5475592613220215, "perplexity": 34.728450887828195, "lr": 0.0026291804804649314, "grad_norm": 0.174029, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:18.266967+00:00", "epoch": 0, "step": 17014, "train_loss": 3.544201135635376, "perplexity": 34.61202398234757, "lr": 0.0026291804804649314, "grad_norm": 0.169628, "tokens_per_sec": 90531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:18.570802+00:00", "epoch": 0, "step": 17015, "train_loss": 3.474318265914917, "perplexity": 32.275817509709235, "lr": 0.0026291804804649314, "grad_norm": 0.185486, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:18.875347+00:00", "epoch": 0, "step": 17016, "train_loss": 3.5669970512390137, "perplexity": 35.410098621066616, "lr": 0.0026291804804649314, "grad_norm": 0.158834, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:19.180041+00:00", "epoch": 0, "step": 17017, "train_loss": 3.3909573554992676, "perplexity": 29.69436673211628, "lr": 0.0026291804804649314, "grad_norm": 0.191834, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:19.484522+00:00", "epoch": 0, "step": 17018, "train_loss": 3.3806633949279785, "perplexity": 29.3902619982391, "lr": 0.0026291804804649314, "grad_norm": 0.166645, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:19.789172+00:00", "epoch": 0, "step": 17019, "train_loss": 3.5053799152374268, "perplexity": 33.294090382199414, "lr": 0.0026291804804649314, "grad_norm": 0.174097, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:20.093157+00:00", "epoch": 0, "step": 17020, "train_loss": 3.4610395431518555, "perplexity": 31.850068832182185, "lr": 0.0026291804804649314, "grad_norm": 0.178534, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:20.397665+00:00", "epoch": 0, "step": 17021, "train_loss": 3.627943992614746, "perplexity": 37.63535844227364, "lr": 0.0026291804804649314, "grad_norm": 0.173016, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:20.703203+00:00", "epoch": 0, "step": 17022, "train_loss": 3.5668716430664062, "perplexity": 35.405658183746084, "lr": 0.0026291804804649314, "grad_norm": 0.170841, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:21.009249+00:00", "epoch": 0, "step": 17023, "train_loss": 3.509884834289551, "perplexity": 33.44441591222831, "lr": 0.0026291804804649314, "grad_norm": 0.148899, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:21.313810+00:00", "epoch": 0, "step": 17024, "train_loss": 3.492863655090332, "perplexity": 32.879969911603126, "lr": 0.0026291804804649314, "grad_norm": 0.157478, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:21.619140+00:00", "epoch": 0, "step": 17025, "train_loss": 3.522770404815674, "perplexity": 33.87815477712853, "lr": 0.0026291804804649314, "grad_norm": 0.163854, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:21.924625+00:00", "epoch": 0, "step": 17026, "train_loss": 3.3579061031341553, "perplexity": 28.728972350386705, "lr": 0.0026291804804649314, "grad_norm": 0.166049, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:22.230498+00:00", "epoch": 0, "step": 17027, "train_loss": 3.5528533458709717, "perplexity": 34.91279377594721, "lr": 0.0026291804804649314, "grad_norm": 0.186112, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:22.536157+00:00", "epoch": 0, "step": 17028, "train_loss": 3.454326629638672, "perplexity": 31.636978104746714, "lr": 0.0026291804804649314, "grad_norm": 0.202627, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:22.842397+00:00", "epoch": 0, "step": 17029, "train_loss": 3.493070125579834, "perplexity": 32.88675935597142, "lr": 0.0026291804804649314, "grad_norm": 0.184531, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:23.147946+00:00", "epoch": 0, "step": 17030, "train_loss": 3.5408363342285156, "perplexity": 34.495757112517765, "lr": 0.0026291804804649314, "grad_norm": 0.169222, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:23.452685+00:00", "epoch": 0, "step": 17031, "train_loss": 3.5885355472564697, "perplexity": 36.18105166981214, "lr": 0.0026291804804649314, "grad_norm": 0.153315, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:23.757854+00:00", "epoch": 0, "step": 17032, "train_loss": 3.5522139072418213, "perplexity": 34.89047632303649, "lr": 0.0026291804804649314, "grad_norm": 0.177696, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:24.063579+00:00", "epoch": 0, "step": 17033, "train_loss": 3.485016107559204, "perplexity": 32.6229525817675, "lr": 0.0026291804804649314, "grad_norm": 0.184812, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:24.368196+00:00", "epoch": 0, "step": 17034, "train_loss": 3.5464510917663574, "perplexity": 34.68998719192358, "lr": 0.0026291804804649314, "grad_norm": 0.190903, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:24.672471+00:00", "epoch": 0, "step": 17035, "train_loss": 3.438399076461792, "perplexity": 31.137070176977566, "lr": 0.0026291804804649314, "grad_norm": 0.184555, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:24.976207+00:00", "epoch": 0, "step": 17036, "train_loss": 3.4801812171936035, "perplexity": 32.46560486861903, "lr": 0.0026291804804649314, "grad_norm": 0.244617, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:25.280658+00:00", "epoch": 0, "step": 17037, "train_loss": 3.4049999713897705, "perplexity": 30.114294862340873, "lr": 0.0026291804804649314, "grad_norm": 0.217871, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:25.585692+00:00", "epoch": 0, "step": 17038, "train_loss": 3.46077036857605, "perplexity": 31.84149675715899, "lr": 0.0026291804804649314, "grad_norm": 0.167316, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:25.891265+00:00", "epoch": 0, "step": 17039, "train_loss": 3.5243613719940186, "perplexity": 33.93209670798454, "lr": 0.0026291804804649314, "grad_norm": 0.190616, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:26.195928+00:00", "epoch": 0, "step": 17040, "train_loss": 3.556544303894043, "perplexity": 35.041893536578705, "lr": 0.0026291804804649314, "grad_norm": 0.18338, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:26.499952+00:00", "epoch": 0, "step": 17041, "train_loss": 3.545069932937622, "perplexity": 34.64210788192181, "lr": 0.0026291804804649314, "grad_norm": 0.148942, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:26.803534+00:00", "epoch": 0, "step": 17042, "train_loss": 3.543064594268799, "perplexity": 34.572708331474544, "lr": 0.0026291804804649314, "grad_norm": 0.182883, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:27.107914+00:00", "epoch": 0, "step": 17043, "train_loss": 3.45650053024292, "perplexity": 31.705828560474153, "lr": 0.0026291804804649314, "grad_norm": 0.178087, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:27.413406+00:00", "epoch": 0, "step": 17044, "train_loss": 3.541250705718994, "perplexity": 34.51005413274188, "lr": 0.0026291804804649314, "grad_norm": 0.181226, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:27.718117+00:00", "epoch": 0, "step": 17045, "train_loss": 3.4789581298828125, "perplexity": 32.42592087271007, "lr": 0.0026291804804649314, "grad_norm": 0.222752, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:28.023294+00:00", "epoch": 0, "step": 17046, "train_loss": 3.6206772327423096, "perplexity": 37.3628626097634, "lr": 0.0026291804804649314, "grad_norm": 0.164613, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:28.326821+00:00", "epoch": 0, "step": 17047, "train_loss": 3.5461161136627197, "perplexity": 34.67836875186896, "lr": 0.0026291804804649314, "grad_norm": 0.183031, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:28.631331+00:00", "epoch": 0, "step": 17048, "train_loss": 3.524186849594116, "perplexity": 33.92617531375634, "lr": 0.0026291804804649314, "grad_norm": 0.163971, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:28.937586+00:00", "epoch": 0, "step": 17049, "train_loss": 3.44138503074646, "perplexity": 31.23018299120561, "lr": 0.0026291804804649314, "grad_norm": 0.182497, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:29.241839+00:00", "epoch": 0, "step": 17050, "train_loss": 3.5146644115448, "perplexity": 33.604648699295296, "lr": 0.0026291804804649314, "grad_norm": 0.176199, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:29.546079+00:00", "epoch": 0, "step": 17051, "train_loss": 3.3988394737243652, "perplexity": 29.929346092305906, "lr": 0.0026291804804649314, "grad_norm": 0.153827, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:29.850589+00:00", "epoch": 0, "step": 17052, "train_loss": 3.5193116664886475, "perplexity": 33.7611815113221, "lr": 0.0026291804804649314, "grad_norm": 0.166665, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:30.155361+00:00", "epoch": 0, "step": 17053, "train_loss": 3.4815332889556885, "perplexity": 32.50953038473421, "lr": 0.0026291804804649314, "grad_norm": 0.15746, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:30.460209+00:00", "epoch": 0, "step": 17054, "train_loss": 3.5173726081848145, "perplexity": 33.69578004108949, "lr": 0.0026291804804649314, "grad_norm": 0.177363, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:30.765769+00:00", "epoch": 0, "step": 17055, "train_loss": 3.6083943843841553, "perplexity": 36.906747166944896, "lr": 0.0026291804804649314, "grad_norm": 0.172919, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:31.070342+00:00", "epoch": 0, "step": 17056, "train_loss": 3.4875340461730957, "perplexity": 32.705198675626626, "lr": 0.0026291804804649314, "grad_norm": 0.186337, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:31.374747+00:00", "epoch": 0, "step": 17057, "train_loss": 3.5708818435668945, "perplexity": 35.54792704458265, "lr": 0.0026291804804649314, "grad_norm": 0.200021, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:31.678737+00:00", "epoch": 0, "step": 17058, "train_loss": 3.4692459106445312, "perplexity": 32.112517604351225, "lr": 0.0026291804804649314, "grad_norm": 0.156799, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:31.983664+00:00", "epoch": 0, "step": 17059, "train_loss": 3.4846580028533936, "perplexity": 32.61127224044796, "lr": 0.0026291804804649314, "grad_norm": 0.157425, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:32.290007+00:00", "epoch": 0, "step": 17060, "train_loss": 3.5110063552856445, "perplexity": 33.481945568090715, "lr": 0.0026291804804649314, "grad_norm": 0.173117, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:32.595170+00:00", "epoch": 0, "step": 17061, "train_loss": 3.4757070541381836, "perplexity": 32.32067292508983, "lr": 0.0026291804804649314, "grad_norm": 0.166254, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:32.899623+00:00", "epoch": 0, "step": 17062, "train_loss": 3.593226909637451, "perplexity": 36.351188870106775, "lr": 0.0026291804804649314, "grad_norm": 0.178063, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:33.203856+00:00", "epoch": 0, "step": 17063, "train_loss": 3.5461535453796387, "perplexity": 34.67966684704611, "lr": 0.0026291804804649314, "grad_norm": 0.209442, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:33.509238+00:00", "epoch": 0, "step": 17064, "train_loss": 3.4622910022735596, "perplexity": 31.889952842748176, "lr": 0.0026291804804649314, "grad_norm": 0.161899, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:33.814733+00:00", "epoch": 0, "step": 17065, "train_loss": 3.4944701194763184, "perplexity": 32.93283286213337, "lr": 0.0026291804804649314, "grad_norm": 0.190766, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:34.119565+00:00", "epoch": 0, "step": 17066, "train_loss": 3.5661914348602295, "perplexity": 35.381583153449256, "lr": 0.0026291804804649314, "grad_norm": 0.162984, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:34.423702+00:00", "epoch": 0, "step": 17067, "train_loss": 3.547088861465454, "perplexity": 34.71211847119614, "lr": 0.0026291804804649314, "grad_norm": 0.169438, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:34.728860+00:00", "epoch": 0, "step": 17068, "train_loss": 3.5035715103149414, "perplexity": 33.23393559376416, "lr": 0.0026291804804649314, "grad_norm": 0.167383, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:35.034306+00:00", "epoch": 0, "step": 17069, "train_loss": 3.4882564544677734, "perplexity": 32.72883371847661, "lr": 0.0026291804804649314, "grad_norm": 0.151036, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:35.340507+00:00", "epoch": 0, "step": 17070, "train_loss": 3.5373003482818604, "perplexity": 34.37399599980726, "lr": 0.0026291804804649314, "grad_norm": 0.144772, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:35.644833+00:00", "epoch": 0, "step": 17071, "train_loss": 3.4160454273223877, "perplexity": 30.448764764974324, "lr": 0.0026291804804649314, "grad_norm": 0.14683, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:35.949907+00:00", "epoch": 0, "step": 17072, "train_loss": 3.5246622562408447, "perplexity": 33.94230787745874, "lr": 0.0026291804804649314, "grad_norm": 0.138031, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:36.253873+00:00", "epoch": 0, "step": 17073, "train_loss": 3.4904720783233643, "perplexity": 32.80142889533434, "lr": 0.0026291804804649314, "grad_norm": 0.155569, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:36.557718+00:00", "epoch": 0, "step": 17074, "train_loss": 3.443934917449951, "perplexity": 31.309918034022637, "lr": 0.0026291804804649314, "grad_norm": 0.172715, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:36.861864+00:00", "epoch": 0, "step": 17075, "train_loss": 3.4476191997528076, "perplexity": 31.425485371562242, "lr": 0.0026291804804649314, "grad_norm": 0.15867, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:37.166481+00:00", "epoch": 0, "step": 17076, "train_loss": 3.523078203201294, "perplexity": 33.88858402344697, "lr": 0.0026291804804649314, "grad_norm": 0.173399, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:37.471748+00:00", "epoch": 0, "step": 17077, "train_loss": 3.5023560523986816, "perplexity": 33.19356568259231, "lr": 0.0026291804804649314, "grad_norm": 0.174061, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:37.775721+00:00", "epoch": 0, "step": 17078, "train_loss": 3.5387117862701416, "perplexity": 34.42254701888666, "lr": 0.0026291804804649314, "grad_norm": 0.16782, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:38.079425+00:00", "epoch": 0, "step": 17079, "train_loss": 3.5037784576416016, "perplexity": 33.240813979596915, "lr": 0.0026291804804649314, "grad_norm": 0.161883, "tokens_per_sec": 107895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:38.383848+00:00", "epoch": 0, "step": 17080, "train_loss": 3.540015459060669, "perplexity": 34.46745202117075, "lr": 0.0026291804804649314, "grad_norm": 0.161225, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:38.688935+00:00", "epoch": 0, "step": 17081, "train_loss": 3.564091444015503, "perplexity": 35.30736011388358, "lr": 0.0026291804804649314, "grad_norm": 0.164327, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:38.994193+00:00", "epoch": 0, "step": 17082, "train_loss": 3.5653202533721924, "perplexity": 35.35077279584376, "lr": 0.0026291804804649314, "grad_norm": 0.167897, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:39.298379+00:00", "epoch": 0, "step": 17083, "train_loss": 3.469006061553955, "perplexity": 32.10481636981168, "lr": 0.0026291804804649314, "grad_norm": 0.143886, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:39.602527+00:00", "epoch": 0, "step": 17084, "train_loss": 3.495006799697876, "perplexity": 32.95051200576936, "lr": 0.0026291804804649314, "grad_norm": 0.158004, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:39.906078+00:00", "epoch": 0, "step": 17085, "train_loss": 3.5940439701080322, "perplexity": 36.380902126701855, "lr": 0.0026291804804649314, "grad_norm": 0.155211, "tokens_per_sec": 107950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:40.211393+00:00", "epoch": 0, "step": 17086, "train_loss": 3.463365077972412, "perplexity": 31.924223467462063, "lr": 0.0026291804804649314, "grad_norm": 0.167631, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:40.516706+00:00", "epoch": 0, "step": 17087, "train_loss": 3.427022695541382, "perplexity": 30.78485029839513, "lr": 0.0026291804804649314, "grad_norm": 0.167618, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:40.822476+00:00", "epoch": 0, "step": 17088, "train_loss": 3.5272679328918457, "perplexity": 34.03086588323432, "lr": 0.0026291804804649314, "grad_norm": 0.155992, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:41.127987+00:00", "epoch": 0, "step": 17089, "train_loss": 3.4513282775878906, "perplexity": 31.542261374576412, "lr": 0.0026291804804649314, "grad_norm": 0.169251, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:41.432534+00:00", "epoch": 0, "step": 17090, "train_loss": 3.522721767425537, "perplexity": 33.876507072167875, "lr": 0.0026291804804649314, "grad_norm": 0.160578, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:41.737313+00:00", "epoch": 0, "step": 17091, "train_loss": 3.43184494972229, "perplexity": 30.93366118504309, "lr": 0.0026291804804649314, "grad_norm": 0.158496, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:42.042591+00:00", "epoch": 0, "step": 17092, "train_loss": 3.4610042572021484, "perplexity": 31.848944992083208, "lr": 0.0026291804804649314, "grad_norm": 0.165355, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:42.347389+00:00", "epoch": 0, "step": 17093, "train_loss": 3.500046730041504, "perplexity": 33.11699948129438, "lr": 0.0026291804804649314, "grad_norm": 0.155225, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:42.652466+00:00", "epoch": 0, "step": 17094, "train_loss": 3.5198514461517334, "perplexity": 33.77941002975284, "lr": 0.0026291804804649314, "grad_norm": 0.15847, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:42.957011+00:00", "epoch": 0, "step": 17095, "train_loss": 3.453312397003174, "perplexity": 31.604907115535497, "lr": 0.0026291804804649314, "grad_norm": 0.174935, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:43.261860+00:00", "epoch": 0, "step": 17096, "train_loss": 3.5173938274383545, "perplexity": 33.69649504797535, "lr": 0.0026291804804649314, "grad_norm": 0.167332, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:43.567667+00:00", "epoch": 0, "step": 17097, "train_loss": 3.532559633255005, "perplexity": 34.21142433787289, "lr": 0.0026291804804649314, "grad_norm": 0.156459, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:43.871854+00:00", "epoch": 0, "step": 17098, "train_loss": 3.3921544551849365, "perplexity": 29.729935134411438, "lr": 0.0026291804804649314, "grad_norm": 0.167242, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:44.177603+00:00", "epoch": 0, "step": 17099, "train_loss": 3.558570384979248, "perplexity": 35.11296322642832, "lr": 0.0026291804804649314, "grad_norm": 0.1513, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:44.484023+00:00", "epoch": 0, "step": 17100, "train_loss": 3.5179734230041504, "perplexity": 33.71603104803062, "lr": 0.0026291804804649314, "grad_norm": 0.18113, "tokens_per_sec": 106996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:44.788354+00:00", "epoch": 0, "step": 17101, "train_loss": 3.461864948272705, "perplexity": 31.876368894705653, "lr": 0.0026291804804649314, "grad_norm": 0.1879, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:45.093169+00:00", "epoch": 0, "step": 17102, "train_loss": 3.5843849182128906, "perplexity": 36.03118877371835, "lr": 0.0026291804804649314, "grad_norm": 0.214186, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:45.397881+00:00", "epoch": 0, "step": 17103, "train_loss": 3.5161397457122803, "perplexity": 33.65426337582197, "lr": 0.0026291804804649314, "grad_norm": 0.187635, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:45.701951+00:00", "epoch": 0, "step": 17104, "train_loss": 3.482893705368042, "perplexity": 32.553786980301396, "lr": 0.0026291804804649314, "grad_norm": 0.193866, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:46.006170+00:00", "epoch": 0, "step": 17105, "train_loss": 3.515868663787842, "perplexity": 33.645141549777435, "lr": 0.0026291804804649314, "grad_norm": 0.151309, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:46.310455+00:00", "epoch": 0, "step": 17106, "train_loss": 3.4309475421905518, "perplexity": 30.90591353684782, "lr": 0.0026291804804649314, "grad_norm": 0.17242, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:46.614905+00:00", "epoch": 0, "step": 17107, "train_loss": 3.4550461769104004, "perplexity": 31.659750597987873, "lr": 0.0026291804804649314, "grad_norm": 0.170859, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:46.920260+00:00", "epoch": 0, "step": 17108, "train_loss": 3.4996204376220703, "perplexity": 33.1028849641307, "lr": 0.0026291804804649314, "grad_norm": 0.189278, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:47.224813+00:00", "epoch": 0, "step": 17109, "train_loss": 3.587245464324951, "perplexity": 36.134405207980166, "lr": 0.0026291804804649314, "grad_norm": 0.191427, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:47.528961+00:00", "epoch": 0, "step": 17110, "train_loss": 3.5407674312591553, "perplexity": 34.493380334306856, "lr": 0.0026291804804649314, "grad_norm": 0.184386, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:47.832401+00:00", "epoch": 0, "step": 17111, "train_loss": 3.4860522747039795, "perplexity": 32.65677293214032, "lr": 0.0026291804804649314, "grad_norm": 0.184546, "tokens_per_sec": 107931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:48.136475+00:00", "epoch": 0, "step": 17112, "train_loss": 3.5033392906188965, "perplexity": 33.22621891535923, "lr": 0.0026291804804649314, "grad_norm": 0.181977, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:48.440359+00:00", "epoch": 0, "step": 17113, "train_loss": 3.540266513824463, "perplexity": 34.47610632550321, "lr": 0.0026291804804649314, "grad_norm": 0.195003, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:48.744935+00:00", "epoch": 0, "step": 17114, "train_loss": 3.505446672439575, "perplexity": 33.296313076711016, "lr": 0.0026291804804649314, "grad_norm": 0.177539, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:49.049379+00:00", "epoch": 0, "step": 17115, "train_loss": 3.515655279159546, "perplexity": 33.637962959681495, "lr": 0.0026291804804649314, "grad_norm": 0.185076, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:49.355177+00:00", "epoch": 0, "step": 17116, "train_loss": 3.366975784301758, "perplexity": 28.99072016020878, "lr": 0.0026291804804649314, "grad_norm": 0.177621, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:49.660260+00:00", "epoch": 0, "step": 17117, "train_loss": 3.4957363605499268, "perplexity": 32.97456018062061, "lr": 0.0026291804804649314, "grad_norm": 0.168131, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:49.966708+00:00", "epoch": 0, "step": 17118, "train_loss": 3.5537056922912598, "perplexity": 34.942564256323244, "lr": 0.0026291804804649314, "grad_norm": 0.154119, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:50.271478+00:00", "epoch": 0, "step": 17119, "train_loss": 3.5912904739379883, "perplexity": 36.280865240828774, "lr": 0.0026291804804649314, "grad_norm": 0.154908, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:50.577021+00:00", "epoch": 0, "step": 17120, "train_loss": 3.4358654022216797, "perplexity": 31.058278842295213, "lr": 0.0026291804804649314, "grad_norm": 0.156689, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:50.881948+00:00", "epoch": 0, "step": 17121, "train_loss": 3.4877569675445557, "perplexity": 32.71249017605471, "lr": 0.0026291804804649314, "grad_norm": 0.16182, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:51.186982+00:00", "epoch": 0, "step": 17122, "train_loss": 3.4954328536987305, "perplexity": 32.964553694285996, "lr": 0.0026291804804649314, "grad_norm": 0.148182, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:51.492495+00:00", "epoch": 0, "step": 17123, "train_loss": 3.5673930644989014, "perplexity": 35.424124266642, "lr": 0.0026291804804649314, "grad_norm": 0.158762, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:51.797922+00:00", "epoch": 0, "step": 17124, "train_loss": 3.6210389137268066, "perplexity": 37.37637849076696, "lr": 0.0026291804804649314, "grad_norm": 0.166698, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:52.102876+00:00", "epoch": 0, "step": 17125, "train_loss": 3.530763864517212, "perplexity": 34.15004366081701, "lr": 0.0026291804804649314, "grad_norm": 0.223943, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:52.407585+00:00", "epoch": 0, "step": 17126, "train_loss": 3.4950757026672363, "perplexity": 32.95278247210854, "lr": 0.0026291804804649314, "grad_norm": 0.20225, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:52.711841+00:00", "epoch": 0, "step": 17127, "train_loss": 3.4022107124328613, "perplexity": 30.0304153309139, "lr": 0.0026291804804649314, "grad_norm": 0.172848, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:53.016817+00:00", "epoch": 0, "step": 17128, "train_loss": 3.538360118865967, "perplexity": 34.41044385939952, "lr": 0.0026291804804649314, "grad_norm": 0.189479, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:53.322092+00:00", "epoch": 0, "step": 17129, "train_loss": 3.4735283851623535, "perplexity": 32.250333528663106, "lr": 0.0026291804804649314, "grad_norm": 0.179494, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:53.627651+00:00", "epoch": 0, "step": 17130, "train_loss": 3.4558727741241455, "perplexity": 31.685931278568155, "lr": 0.0026291804804649314, "grad_norm": 0.176086, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:53.933496+00:00", "epoch": 0, "step": 17131, "train_loss": 3.609527587890625, "perplexity": 36.94859372810368, "lr": 0.0026291804804649314, "grad_norm": 0.22766, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:54.239080+00:00", "epoch": 0, "step": 17132, "train_loss": 3.476114273071289, "perplexity": 32.33383719522374, "lr": 0.0026291804804649314, "grad_norm": 0.188454, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:54.544323+00:00", "epoch": 0, "step": 17133, "train_loss": 3.326348066329956, "perplexity": 27.83649880742204, "lr": 0.0026291804804649314, "grad_norm": 0.16088, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:54.848947+00:00", "epoch": 0, "step": 17134, "train_loss": 3.49521803855896, "perplexity": 32.95747316960587, "lr": 0.0026291804804649314, "grad_norm": 0.173874, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:55.154448+00:00", "epoch": 0, "step": 17135, "train_loss": 3.437366485595703, "perplexity": 31.104934916859424, "lr": 0.0026291804804649314, "grad_norm": 0.178076, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:55.459910+00:00", "epoch": 0, "step": 17136, "train_loss": 3.5535776615142822, "perplexity": 34.938090819046856, "lr": 0.0026291804804649314, "grad_norm": 0.160264, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:55.764876+00:00", "epoch": 0, "step": 17137, "train_loss": 3.6262733936309814, "perplexity": 37.572537339751996, "lr": 0.0026291804804649314, "grad_norm": 0.173616, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:56.070528+00:00", "epoch": 0, "step": 17138, "train_loss": 3.524003267288208, "perplexity": 33.919947639924445, "lr": 0.0026291804804649314, "grad_norm": 0.163049, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:56.375656+00:00", "epoch": 0, "step": 17139, "train_loss": 3.5938384532928467, "perplexity": 36.37342600782359, "lr": 0.0026291804804649314, "grad_norm": 0.183744, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:56.680954+00:00", "epoch": 0, "step": 17140, "train_loss": 3.4778149127960205, "perplexity": 32.38887218729358, "lr": 0.0026291804804649314, "grad_norm": 0.1795, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:56.986555+00:00", "epoch": 0, "step": 17141, "train_loss": 3.6046512126922607, "perplexity": 36.76885710979088, "lr": 0.0026291804804649314, "grad_norm": 0.179077, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:57.291480+00:00", "epoch": 0, "step": 17142, "train_loss": 3.5820932388305664, "perplexity": 35.9487113832488, "lr": 0.0026291804804649314, "grad_norm": 0.185973, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:57.595865+00:00", "epoch": 0, "step": 17143, "train_loss": 3.563856601715088, "perplexity": 35.299069425753096, "lr": 0.0026291804804649314, "grad_norm": 0.174754, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:57.900351+00:00", "epoch": 0, "step": 17144, "train_loss": 3.5101230144500732, "perplexity": 33.452382657301506, "lr": 0.0026291804804649314, "grad_norm": 0.163218, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:58.204654+00:00", "epoch": 0, "step": 17145, "train_loss": 3.4604873657226562, "perplexity": 31.832486797702643, "lr": 0.0026291804804649314, "grad_norm": 0.172401, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:58.509631+00:00", "epoch": 0, "step": 17146, "train_loss": 3.4785618782043457, "perplexity": 32.413074592486566, "lr": 0.0026291804804649314, "grad_norm": 0.191675, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:58.814648+00:00", "epoch": 0, "step": 17147, "train_loss": 3.452732563018799, "perplexity": 31.586586828192782, "lr": 0.0026291804804649314, "grad_norm": 0.169156, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:59.119284+00:00", "epoch": 0, "step": 17148, "train_loss": 3.458069324493408, "perplexity": 31.755607518288887, "lr": 0.0026291804804649314, "grad_norm": 0.166631, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:59.423727+00:00", "epoch": 0, "step": 17149, "train_loss": 3.5210816860198975, "perplexity": 33.82099237958225, "lr": 0.0026291804804649314, "grad_norm": 0.170313, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:27:59.728375+00:00", "epoch": 0, "step": 17150, "train_loss": 3.4985105991363525, "perplexity": 33.06616648796719, "lr": 0.0026291804804649314, "grad_norm": 0.188934, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:00.034344+00:00", "epoch": 0, "step": 17151, "train_loss": 3.488635540008545, "perplexity": 32.74124309806517, "lr": 0.0026291804804649314, "grad_norm": 0.190077, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:00.339888+00:00", "epoch": 0, "step": 17152, "train_loss": 3.542527914047241, "perplexity": 34.55415882073036, "lr": 0.0026291804804649314, "grad_norm": 0.160459, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:00.644692+00:00", "epoch": 0, "step": 17153, "train_loss": 3.6154165267944336, "perplexity": 37.166823679899736, "lr": 0.0026291804804649314, "grad_norm": 0.18063, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:00.949867+00:00", "epoch": 0, "step": 17154, "train_loss": 3.5001697540283203, "perplexity": 33.121073917223505, "lr": 0.0026291804804649314, "grad_norm": 0.166184, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:01.254901+00:00", "epoch": 0, "step": 17155, "train_loss": 3.5473756790161133, "perplexity": 34.722075943914945, "lr": 0.0026291804804649314, "grad_norm": 0.174556, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:01.560035+00:00", "epoch": 0, "step": 17156, "train_loss": 3.577521562576294, "perplexity": 35.78474060927857, "lr": 0.0026291804804649314, "grad_norm": 0.194124, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:01.864681+00:00", "epoch": 0, "step": 17157, "train_loss": 3.3851206302642822, "perplexity": 29.521553694190466, "lr": 0.0026291804804649314, "grad_norm": 0.2114, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:02.170985+00:00", "epoch": 0, "step": 17158, "train_loss": 3.53739333152771, "perplexity": 34.37719235412954, "lr": 0.0026291804804649314, "grad_norm": 0.177092, "tokens_per_sec": 106979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:02.474904+00:00", "epoch": 0, "step": 17159, "train_loss": 3.5660362243652344, "perplexity": 35.37609198656868, "lr": 0.0026291804804649314, "grad_norm": 0.151547, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:02.781258+00:00", "epoch": 0, "step": 17160, "train_loss": 3.5088958740234375, "perplexity": 33.41135706340612, "lr": 0.0026291804804649314, "grad_norm": 0.185846, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:03.086387+00:00", "epoch": 0, "step": 17161, "train_loss": 3.449249744415283, "perplexity": 31.476767826805432, "lr": 0.0026291804804649314, "grad_norm": 0.177374, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:03.390935+00:00", "epoch": 0, "step": 17162, "train_loss": 3.5137534141540527, "perplexity": 33.574048892301704, "lr": 0.0026291804804649314, "grad_norm": 0.158803, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:03.694996+00:00", "epoch": 0, "step": 17163, "train_loss": 3.6088335514068604, "perplexity": 36.92295895279623, "lr": 0.0026291804804649314, "grad_norm": 0.163927, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:04.000159+00:00", "epoch": 0, "step": 17164, "train_loss": 3.4493532180786133, "perplexity": 31.48002501179579, "lr": 0.0026291804804649314, "grad_norm": 0.170949, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:04.305795+00:00", "epoch": 0, "step": 17165, "train_loss": 3.4633827209472656, "perplexity": 31.924786710702545, "lr": 0.0026291804804649314, "grad_norm": 0.149058, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:04.610488+00:00", "epoch": 0, "step": 17166, "train_loss": 3.414950370788574, "perplexity": 30.415439895809868, "lr": 0.0026291804804649314, "grad_norm": 0.1485, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:04.915205+00:00", "epoch": 0, "step": 17167, "train_loss": 3.447826385498047, "perplexity": 31.431996958699177, "lr": 0.0026291804804649314, "grad_norm": 0.176384, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:05.219705+00:00", "epoch": 0, "step": 17168, "train_loss": 3.4209372997283936, "perplexity": 30.598081157728977, "lr": 0.0026291804804649314, "grad_norm": 0.168087, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:05.525193+00:00", "epoch": 0, "step": 17169, "train_loss": 3.3814001083374023, "perplexity": 29.411922176051743, "lr": 0.0026291804804649314, "grad_norm": 0.169468, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:05.831804+00:00", "epoch": 0, "step": 17170, "train_loss": 3.57716703414917, "perplexity": 35.77205615010886, "lr": 0.0026291804804649314, "grad_norm": 0.172216, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:06.441494+00:00", "epoch": 0, "step": 17171, "train_loss": 3.6334280967712402, "perplexity": 37.84232165313236, "lr": 0.0026291804804649314, "grad_norm": 0.184911, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:06.746283+00:00", "epoch": 0, "step": 17172, "train_loss": 3.5516903400421143, "perplexity": 34.872213595351376, "lr": 0.0026291804804649314, "grad_norm": 0.176957, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:07.051339+00:00", "epoch": 0, "step": 17173, "train_loss": 3.5158345699310303, "perplexity": 33.643994476693216, "lr": 0.0026291804804649314, "grad_norm": 0.17088, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:07.355694+00:00", "epoch": 0, "step": 17174, "train_loss": 3.4362034797668457, "perplexity": 31.06878072408848, "lr": 0.0026291804804649314, "grad_norm": 0.18437, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:07.660390+00:00", "epoch": 0, "step": 17175, "train_loss": 3.5560433864593506, "perplexity": 35.024344836753386, "lr": 0.0026291804804649314, "grad_norm": 0.167263, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:07.965728+00:00", "epoch": 0, "step": 17176, "train_loss": 3.4806265830993652, "perplexity": 32.48006716240468, "lr": 0.0026291804804649314, "grad_norm": 0.18846, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:08.270775+00:00", "epoch": 0, "step": 17177, "train_loss": 3.5280299186706543, "perplexity": 34.05680680112799, "lr": 0.0026291804804649314, "grad_norm": 0.188404, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:08.576221+00:00", "epoch": 0, "step": 17178, "train_loss": 3.4782981872558594, "perplexity": 32.404528684892675, "lr": 0.0026291804804649314, "grad_norm": 0.190673, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:08.880493+00:00", "epoch": 0, "step": 17179, "train_loss": 3.5707192420959473, "perplexity": 35.54214736926066, "lr": 0.0026291804804649314, "grad_norm": 0.170116, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:09.184934+00:00", "epoch": 0, "step": 17180, "train_loss": 3.4150986671447754, "perplexity": 30.419950729180485, "lr": 0.0026291804804649314, "grad_norm": 0.186234, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:09.488978+00:00", "epoch": 0, "step": 17181, "train_loss": 3.5591156482696533, "perplexity": 35.13211425699553, "lr": 0.0026291804804649314, "grad_norm": 0.184312, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:09.793037+00:00", "epoch": 0, "step": 17182, "train_loss": 3.6654436588287354, "perplexity": 39.07346760706419, "lr": 0.0026291804804649314, "grad_norm": 0.195566, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:10.099312+00:00", "epoch": 0, "step": 17183, "train_loss": 3.4879403114318848, "perplexity": 32.71848836101627, "lr": 0.0026291804804649314, "grad_norm": 0.160169, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:10.403744+00:00", "epoch": 0, "step": 17184, "train_loss": 3.517432928085327, "perplexity": 33.69781262849139, "lr": 0.0026291804804649314, "grad_norm": 0.158183, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:10.707754+00:00", "epoch": 0, "step": 17185, "train_loss": 3.5166256427764893, "perplexity": 33.6706198570557, "lr": 0.0026291804804649314, "grad_norm": 0.157684, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:11.011863+00:00", "epoch": 0, "step": 17186, "train_loss": 3.390411376953125, "perplexity": 29.678158669968983, "lr": 0.0026291804804649314, "grad_norm": 0.16384, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:11.315712+00:00", "epoch": 0, "step": 17187, "train_loss": 3.4551568031311035, "perplexity": 31.663253190280635, "lr": 0.0026291804804649314, "grad_norm": 0.164875, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:11.621092+00:00", "epoch": 0, "step": 17188, "train_loss": 3.5565459728240967, "perplexity": 35.04195201909677, "lr": 0.0026291804804649314, "grad_norm": 0.232765, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:11.925863+00:00", "epoch": 0, "step": 17189, "train_loss": 3.4955224990844727, "perplexity": 32.96750894687814, "lr": 0.0026291804804649314, "grad_norm": 0.200457, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:12.230206+00:00", "epoch": 0, "step": 17190, "train_loss": 3.521597146987915, "perplexity": 33.83843027494465, "lr": 0.0026291804804649314, "grad_norm": 0.154689, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:12.534484+00:00", "epoch": 0, "step": 17191, "train_loss": 3.5649538040161133, "perplexity": 35.3378209011675, "lr": 0.0026291804804649314, "grad_norm": 0.177267, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:12.838602+00:00", "epoch": 0, "step": 17192, "train_loss": 3.61507248878479, "perplexity": 37.15403907917652, "lr": 0.0026291804804649314, "grad_norm": 0.153318, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:13.144263+00:00", "epoch": 0, "step": 17193, "train_loss": 3.588052988052368, "perplexity": 36.16359638225836, "lr": 0.0026291804804649314, "grad_norm": 0.181309, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:13.449346+00:00", "epoch": 0, "step": 17194, "train_loss": 3.460216999053955, "perplexity": 31.823881517633545, "lr": 0.0026291804804649314, "grad_norm": 0.181193, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:13.753274+00:00", "epoch": 0, "step": 17195, "train_loss": 3.5753064155578613, "perplexity": 35.70555987866971, "lr": 0.0026291804804649314, "grad_norm": 0.173722, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:14.057629+00:00", "epoch": 0, "step": 17196, "train_loss": 3.662534475326538, "perplexity": 38.95996090569347, "lr": 0.0026291804804649314, "grad_norm": 0.179172, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:14.361702+00:00", "epoch": 0, "step": 17197, "train_loss": 3.6087088584899902, "perplexity": 36.91835520837806, "lr": 0.0026291804804649314, "grad_norm": 0.177336, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:14.666599+00:00", "epoch": 0, "step": 17198, "train_loss": 3.470958948135376, "perplexity": 32.16757469489927, "lr": 0.0026291804804649314, "grad_norm": 0.1816, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:14.971522+00:00", "epoch": 0, "step": 17199, "train_loss": 3.4414584636688232, "perplexity": 31.23247639901338, "lr": 0.0026291804804649314, "grad_norm": 0.171799, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:15.276674+00:00", "epoch": 0, "step": 17200, "train_loss": 3.4792802333831787, "perplexity": 32.43636705761181, "lr": 0.0026291804804649314, "grad_norm": 0.182604, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:15.580853+00:00", "epoch": 0, "step": 17201, "train_loss": 3.5426321029663086, "perplexity": 34.5577591687421, "lr": 0.0026291804804649314, "grad_norm": 0.161928, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:15.885758+00:00", "epoch": 0, "step": 17202, "train_loss": 3.613590955734253, "perplexity": 37.09903489763712, "lr": 0.0026291804804649314, "grad_norm": 0.19214, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:16.190425+00:00", "epoch": 0, "step": 17203, "train_loss": 3.5010175704956055, "perplexity": 33.14916641606353, "lr": 0.0026291804804649314, "grad_norm": 0.196385, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:16.495966+00:00", "epoch": 0, "step": 17204, "train_loss": 3.462893009185791, "perplexity": 31.909156594605953, "lr": 0.0026291804804649314, "grad_norm": 0.164191, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:16.800373+00:00", "epoch": 0, "step": 17205, "train_loss": 3.4888672828674316, "perplexity": 32.74883152659232, "lr": 0.0026291804804649314, "grad_norm": 0.203101, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:17.103911+00:00", "epoch": 0, "step": 17206, "train_loss": 3.533724308013916, "perplexity": 34.25129273261545, "lr": 0.0026291804804649314, "grad_norm": 0.188889, "tokens_per_sec": 107954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:17.408007+00:00", "epoch": 0, "step": 17207, "train_loss": 3.482158899307251, "perplexity": 32.52987504671935, "lr": 0.0026291804804649314, "grad_norm": 0.187361, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:17.712006+00:00", "epoch": 0, "step": 17208, "train_loss": 3.4541454315185547, "perplexity": 31.63124606312121, "lr": 0.0026291804804649314, "grad_norm": 0.178395, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:18.016661+00:00", "epoch": 0, "step": 17209, "train_loss": 3.486942768096924, "perplexity": 32.68586652457465, "lr": 0.0026291804804649314, "grad_norm": 0.186643, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:18.322677+00:00", "epoch": 0, "step": 17210, "train_loss": 3.4798998832702637, "perplexity": 32.45647247731372, "lr": 0.0026291804804649314, "grad_norm": 0.214189, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:18.626882+00:00", "epoch": 0, "step": 17211, "train_loss": 3.4878034591674805, "perplexity": 32.714011068167004, "lr": 0.0026291804804649314, "grad_norm": 0.154361, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:18.931176+00:00", "epoch": 0, "step": 17212, "train_loss": 3.4591426849365234, "perplexity": 31.78971103065245, "lr": 0.0026291804804649314, "grad_norm": 0.176554, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:19.235859+00:00", "epoch": 0, "step": 17213, "train_loss": 3.50986909866333, "perplexity": 33.44388964754091, "lr": 0.0026291804804649314, "grad_norm": 0.156878, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:19.540113+00:00", "epoch": 0, "step": 17214, "train_loss": 3.5959935188293457, "perplexity": 36.45189765005279, "lr": 0.0026291804804649314, "grad_norm": 0.195776, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:19.844713+00:00", "epoch": 0, "step": 17215, "train_loss": 3.376147747039795, "perplexity": 29.257845122715885, "lr": 0.0026291804804649314, "grad_norm": 0.204849, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:20.149515+00:00", "epoch": 0, "step": 17216, "train_loss": 3.543246269226074, "perplexity": 34.57898989736789, "lr": 0.0026291804804649314, "grad_norm": 0.156638, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:20.454931+00:00", "epoch": 0, "step": 17217, "train_loss": 3.496232748031616, "perplexity": 32.99093240265534, "lr": 0.0026291804804649314, "grad_norm": 0.163642, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:20.760187+00:00", "epoch": 0, "step": 17218, "train_loss": 3.508117914199829, "perplexity": 33.38537447797278, "lr": 0.0026291804804649314, "grad_norm": 0.156395, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:21.064800+00:00", "epoch": 0, "step": 17219, "train_loss": 3.4754297733306885, "perplexity": 32.3117122651693, "lr": 0.0026291804804649314, "grad_norm": 0.148847, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:21.370385+00:00", "epoch": 0, "step": 17220, "train_loss": 3.584468126296997, "perplexity": 36.03418698464025, "lr": 0.0026291804804649314, "grad_norm": 0.160572, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:21.675578+00:00", "epoch": 0, "step": 17221, "train_loss": 3.5528359413146973, "perplexity": 34.912186139551075, "lr": 0.0026291804804649314, "grad_norm": 0.178968, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:21.979444+00:00", "epoch": 0, "step": 17222, "train_loss": 3.5386862754821777, "perplexity": 34.42166888378949, "lr": 0.0026291804804649314, "grad_norm": 0.17974, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:22.283919+00:00", "epoch": 0, "step": 17223, "train_loss": 3.4886062145233154, "perplexity": 32.74028295930264, "lr": 0.0026291804804649314, "grad_norm": 0.161872, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:22.588040+00:00", "epoch": 0, "step": 17224, "train_loss": 3.5286645889282227, "perplexity": 34.07842850407253, "lr": 0.0026291804804649314, "grad_norm": 0.160014, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:22.893041+00:00", "epoch": 0, "step": 17225, "train_loss": 3.4602222442626953, "perplexity": 31.824048440972803, "lr": 0.0026291804804649314, "grad_norm": 0.157011, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:23.198108+00:00", "epoch": 0, "step": 17226, "train_loss": 3.5261213779449463, "perplexity": 33.991869985350405, "lr": 0.0026291804804649314, "grad_norm": 0.161834, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:23.501878+00:00", "epoch": 0, "step": 17227, "train_loss": 3.40287709236145, "perplexity": 30.050433666105214, "lr": 0.0026291804804649314, "grad_norm": 0.15854, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:23.805822+00:00", "epoch": 0, "step": 17228, "train_loss": 3.493563413619995, "perplexity": 32.90298600291777, "lr": 0.0026291804804649314, "grad_norm": 0.171908, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:24.110058+00:00", "epoch": 0, "step": 17229, "train_loss": 3.5669870376586914, "perplexity": 35.40974404097516, "lr": 0.0026291804804649314, "grad_norm": 0.168906, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:24.415357+00:00", "epoch": 0, "step": 17230, "train_loss": 3.511794328689575, "perplexity": 33.50833884794587, "lr": 0.0026291804804649314, "grad_norm": 0.171857, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:24.719671+00:00", "epoch": 0, "step": 17231, "train_loss": 3.408735752105713, "perplexity": 30.2270056646996, "lr": 0.0026291804804649314, "grad_norm": 0.163368, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:25.024798+00:00", "epoch": 0, "step": 17232, "train_loss": 3.5206496715545654, "perplexity": 33.80638437730373, "lr": 0.0026291804804649314, "grad_norm": 0.192488, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:25.329874+00:00", "epoch": 0, "step": 17233, "train_loss": 3.444570302963257, "perplexity": 31.329818223840906, "lr": 0.0026291804804649314, "grad_norm": 0.154285, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:25.635054+00:00", "epoch": 0, "step": 17234, "train_loss": 3.5654890537261963, "perplexity": 35.356740522469025, "lr": 0.0026291804804649314, "grad_norm": 0.167765, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:25.940065+00:00", "epoch": 0, "step": 17235, "train_loss": 3.4275617599487305, "perplexity": 30.80144978917191, "lr": 0.0026291804804649314, "grad_norm": 0.158456, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:26.245570+00:00", "epoch": 0, "step": 17236, "train_loss": 3.4255335330963135, "perplexity": 30.739040772831874, "lr": 0.0026291804804649314, "grad_norm": 0.157836, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:26.550826+00:00", "epoch": 0, "step": 17237, "train_loss": 3.6258935928344727, "perplexity": 37.558269969694265, "lr": 0.0026291804804649314, "grad_norm": 0.150761, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:26.855285+00:00", "epoch": 0, "step": 17238, "train_loss": 3.503173351287842, "perplexity": 33.22070583624992, "lr": 0.0026291804804649314, "grad_norm": 0.168991, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:27.159333+00:00", "epoch": 0, "step": 17239, "train_loss": 3.5067334175109863, "perplexity": 33.33918451985249, "lr": 0.0026291804804649314, "grad_norm": 0.196397, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:27.464238+00:00", "epoch": 0, "step": 17240, "train_loss": 3.472201108932495, "perplexity": 32.207556822098255, "lr": 0.0026291804804649314, "grad_norm": 0.181717, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:27.769262+00:00", "epoch": 0, "step": 17241, "train_loss": 3.4607508182525635, "perplexity": 31.84087425168221, "lr": 0.0026291804804649314, "grad_norm": 0.182851, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:28.073748+00:00", "epoch": 0, "step": 17242, "train_loss": 3.4043519496917725, "perplexity": 30.094786467464434, "lr": 0.0026291804804649314, "grad_norm": 0.165395, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:28.377815+00:00", "epoch": 0, "step": 17243, "train_loss": 3.5362296104431152, "perplexity": 34.337210159131686, "lr": 0.0026291804804649314, "grad_norm": 0.178838, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:28.682379+00:00", "epoch": 0, "step": 17244, "train_loss": 3.459232807159424, "perplexity": 31.79257611917803, "lr": 0.0026291804804649314, "grad_norm": 0.171054, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:28.986852+00:00", "epoch": 0, "step": 17245, "train_loss": 3.520934820175171, "perplexity": 33.81602559570131, "lr": 0.0026291804804649314, "grad_norm": 0.189323, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:29.291252+00:00", "epoch": 0, "step": 17246, "train_loss": 3.52421498298645, "perplexity": 33.927129785583034, "lr": 0.0026291804804649314, "grad_norm": 0.174292, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:29.595744+00:00", "epoch": 0, "step": 17247, "train_loss": 3.4835803508758545, "perplexity": 32.57614756791367, "lr": 0.0026291804804649314, "grad_norm": 0.174756, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:29.900330+00:00", "epoch": 0, "step": 17248, "train_loss": 3.5008718967437744, "perplexity": 33.144337804331165, "lr": 0.0026291804804649314, "grad_norm": 0.190055, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:30.204903+00:00", "epoch": 0, "step": 17249, "train_loss": 3.416036605834961, "perplexity": 30.448496162763526, "lr": 0.0026291804804649314, "grad_norm": 0.174963, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:30.510004+00:00", "epoch": 0, "step": 17250, "train_loss": 3.4398739337921143, "perplexity": 31.18302679456959, "lr": 0.0026291804804649314, "grad_norm": 0.196743, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:30.815361+00:00", "epoch": 0, "step": 17251, "train_loss": 3.54809308052063, "perplexity": 34.74699455069132, "lr": 0.0026291804804649314, "grad_norm": 0.208473, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:31.121214+00:00", "epoch": 0, "step": 17252, "train_loss": 3.5689361095428467, "perplexity": 35.47882747993485, "lr": 0.0026291804804649314, "grad_norm": 0.188032, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:31.425949+00:00", "epoch": 0, "step": 17253, "train_loss": 3.508246660232544, "perplexity": 33.38967298918972, "lr": 0.0026291804804649314, "grad_norm": 0.181811, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:31.730971+00:00", "epoch": 0, "step": 17254, "train_loss": 3.4987306594848633, "perplexity": 33.07344384078747, "lr": 0.0026291804804649314, "grad_norm": 0.188492, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:32.035006+00:00", "epoch": 0, "step": 17255, "train_loss": 3.4919872283935547, "perplexity": 32.85116565243799, "lr": 0.0026291804804649314, "grad_norm": 0.156497, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:32.339323+00:00", "epoch": 0, "step": 17256, "train_loss": 3.5147531032562256, "perplexity": 33.60762928527498, "lr": 0.0026291804804649314, "grad_norm": 0.175725, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:32.644560+00:00", "epoch": 0, "step": 17257, "train_loss": 3.526284694671631, "perplexity": 33.99742187963656, "lr": 0.0026291804804649314, "grad_norm": 0.158977, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:32.949333+00:00", "epoch": 0, "step": 17258, "train_loss": 3.4706599712371826, "perplexity": 32.1579587707358, "lr": 0.0026291804804649314, "grad_norm": 0.183499, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:33.254194+00:00", "epoch": 0, "step": 17259, "train_loss": 3.487639904022217, "perplexity": 32.7086609608653, "lr": 0.0026291804804649314, "grad_norm": 0.17497, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:33.558576+00:00", "epoch": 0, "step": 17260, "train_loss": 3.577693462371826, "perplexity": 35.79089252761408, "lr": 0.0026291804804649314, "grad_norm": 0.163584, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:33.862738+00:00", "epoch": 0, "step": 17261, "train_loss": 3.4814493656158447, "perplexity": 32.506802190848745, "lr": 0.0026291804804649314, "grad_norm": 0.163674, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:34.167854+00:00", "epoch": 0, "step": 17262, "train_loss": 3.5653440952301025, "perplexity": 35.35161563399315, "lr": 0.0026291804804649314, "grad_norm": 0.163023, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:34.473249+00:00", "epoch": 0, "step": 17263, "train_loss": 3.6230087280273438, "perplexity": 37.450075576577376, "lr": 0.0026291804804649314, "grad_norm": 0.151181, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:34.777898+00:00", "epoch": 0, "step": 17264, "train_loss": 3.5157039165496826, "perplexity": 33.63959906219693, "lr": 0.0026291804804649314, "grad_norm": 0.174358, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:35.082394+00:00", "epoch": 0, "step": 17265, "train_loss": 3.4460508823394775, "perplexity": 31.376238862804257, "lr": 0.0026291804804649314, "grad_norm": 0.164314, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:35.388294+00:00", "epoch": 0, "step": 17266, "train_loss": 3.5208542346954346, "perplexity": 33.81330062485392, "lr": 0.0026291804804649314, "grad_norm": 0.161499, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:35.693780+00:00", "epoch": 0, "step": 17267, "train_loss": 3.5611422061920166, "perplexity": 35.20338371292108, "lr": 0.0026291804804649314, "grad_norm": 0.167954, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:35.999352+00:00", "epoch": 0, "step": 17268, "train_loss": 3.4573240280151367, "perplexity": 31.73194899324296, "lr": 0.0026291804804649314, "grad_norm": 0.158125, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:36.304414+00:00", "epoch": 0, "step": 17269, "train_loss": 3.6250808238983154, "perplexity": 37.52775617657844, "lr": 0.0026291804804649314, "grad_norm": 0.160149, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:36.608369+00:00", "epoch": 0, "step": 17270, "train_loss": 3.396420955657959, "perplexity": 29.85704888935193, "lr": 0.0026291804804649314, "grad_norm": 0.160515, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:36.912374+00:00", "epoch": 0, "step": 17271, "train_loss": 3.482273578643799, "perplexity": 32.533605765121756, "lr": 0.0026291804804649314, "grad_norm": 0.181374, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:37.217325+00:00", "epoch": 0, "step": 17272, "train_loss": 3.4352903366088867, "perplexity": 31.040423428649735, "lr": 0.0026291804804649314, "grad_norm": 0.160705, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:37.522946+00:00", "epoch": 0, "step": 17273, "train_loss": 3.4975674152374268, "perplexity": 33.03499371527611, "lr": 0.0026291804804649314, "grad_norm": 0.162045, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:37.828807+00:00", "epoch": 0, "step": 17274, "train_loss": 3.552248001098633, "perplexity": 34.891665894218754, "lr": 0.0026291804804649314, "grad_norm": 0.151456, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:38.133805+00:00", "epoch": 0, "step": 17275, "train_loss": 3.562542676925659, "perplexity": 35.25271956017771, "lr": 0.0026291804804649314, "grad_norm": 0.152747, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:38.437976+00:00", "epoch": 0, "step": 17276, "train_loss": 3.4751884937286377, "perplexity": 32.303917048543575, "lr": 0.0026291804804649314, "grad_norm": 0.176563, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:38.742239+00:00", "epoch": 0, "step": 17277, "train_loss": 3.5314383506774902, "perplexity": 34.17308516235256, "lr": 0.0026291804804649314, "grad_norm": 0.166594, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:39.046326+00:00", "epoch": 0, "step": 17278, "train_loss": 3.507362127304077, "perplexity": 33.36015178212469, "lr": 0.0026291804804649314, "grad_norm": 0.18956, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:39.351059+00:00", "epoch": 0, "step": 17279, "train_loss": 3.436134099960327, "perplexity": 31.066625252867027, "lr": 0.0026291804804649314, "grad_norm": 0.175868, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:39.655817+00:00", "epoch": 0, "step": 17280, "train_loss": 3.505709409713745, "perplexity": 33.305062408586174, "lr": 0.0026291804804649314, "grad_norm": 0.186405, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:39.959635+00:00", "epoch": 0, "step": 17281, "train_loss": 3.3791067600250244, "perplexity": 29.344547680088187, "lr": 0.0026291804804649314, "grad_norm": 0.178554, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:40.263919+00:00", "epoch": 0, "step": 17282, "train_loss": 3.482557535171509, "perplexity": 32.542845206586485, "lr": 0.0026291804804649314, "grad_norm": 0.182916, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:40.568047+00:00", "epoch": 0, "step": 17283, "train_loss": 3.518317699432373, "perplexity": 33.72764068112052, "lr": 0.0026291804804649314, "grad_norm": 0.173981, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:40.873064+00:00", "epoch": 0, "step": 17284, "train_loss": 3.5635645389556885, "perplexity": 35.2887613875027, "lr": 0.0026291804804649314, "grad_norm": 0.175584, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:41.178935+00:00", "epoch": 0, "step": 17285, "train_loss": 3.576127290725708, "perplexity": 35.73488171926707, "lr": 0.0026291804804649314, "grad_norm": 0.186929, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:41.483285+00:00", "epoch": 0, "step": 17286, "train_loss": 3.479224443435669, "perplexity": 32.434557484874716, "lr": 0.0026291804804649314, "grad_norm": 0.18688, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:41.788092+00:00", "epoch": 0, "step": 17287, "train_loss": 3.5116090774536133, "perplexity": 33.502131961693685, "lr": 0.0026291804804649314, "grad_norm": 0.176864, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:42.093079+00:00", "epoch": 0, "step": 17288, "train_loss": 3.541849136352539, "perplexity": 34.530712186889915, "lr": 0.0026291804804649314, "grad_norm": 0.176354, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:42.399058+00:00", "epoch": 0, "step": 17289, "train_loss": 3.5819318294525146, "perplexity": 35.942909392363134, "lr": 0.0026291804804649314, "grad_norm": 0.165648, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:42.704641+00:00", "epoch": 0, "step": 17290, "train_loss": 3.466294050216675, "perplexity": 32.01786570272645, "lr": 0.0026291804804649314, "grad_norm": 0.167457, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:43.009452+00:00", "epoch": 0, "step": 17291, "train_loss": 3.483938455581665, "perplexity": 32.587815328670246, "lr": 0.0026291804804649314, "grad_norm": 0.157188, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:43.313966+00:00", "epoch": 0, "step": 17292, "train_loss": 3.3782060146331787, "perplexity": 29.318127614651956, "lr": 0.0026291804804649314, "grad_norm": 0.157039, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:43.619057+00:00", "epoch": 0, "step": 17293, "train_loss": 3.519956111907959, "perplexity": 33.78294576228029, "lr": 0.0026291804804649314, "grad_norm": 0.166587, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:43.923320+00:00", "epoch": 0, "step": 17294, "train_loss": 3.4437544345855713, "perplexity": 31.304267640247325, "lr": 0.0026291804804649314, "grad_norm": 0.178707, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:44.228525+00:00", "epoch": 0, "step": 17295, "train_loss": 3.509093999862671, "perplexity": 33.41797737236972, "lr": 0.0026291804804649314, "grad_norm": 0.167966, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:44.534319+00:00", "epoch": 0, "step": 17296, "train_loss": 3.440653085708618, "perplexity": 31.207332577377972, "lr": 0.0026291804804649314, "grad_norm": 0.198005, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:44.839003+00:00", "epoch": 0, "step": 17297, "train_loss": 3.458665609359741, "perplexity": 31.774548553038223, "lr": 0.0026291804804649314, "grad_norm": 0.177242, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:45.143001+00:00", "epoch": 0, "step": 17298, "train_loss": 3.536735773086548, "perplexity": 34.354594771543454, "lr": 0.0026291804804649314, "grad_norm": 0.183109, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:45.447190+00:00", "epoch": 0, "step": 17299, "train_loss": 3.5468454360961914, "perplexity": 34.703669689305045, "lr": 0.0026291804804649314, "grad_norm": 0.172992, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:45.753498+00:00", "epoch": 0, "step": 17300, "train_loss": 3.434229612350464, "perplexity": 31.00751555470283, "lr": 0.0026291804804649314, "grad_norm": 0.180661, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:46.058281+00:00", "epoch": 0, "step": 17301, "train_loss": 3.577578544616699, "perplexity": 35.78677975491065, "lr": 0.0026291804804649314, "grad_norm": 0.16892, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:46.363129+00:00", "epoch": 0, "step": 17302, "train_loss": 3.565488338470459, "perplexity": 35.35671523336656, "lr": 0.0026291804804649314, "grad_norm": 0.191629, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:46.667878+00:00", "epoch": 0, "step": 17303, "train_loss": 3.506943702697754, "perplexity": 33.34619599367463, "lr": 0.0026291804804649314, "grad_norm": 0.160561, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:46.972388+00:00", "epoch": 0, "step": 17304, "train_loss": 3.51108455657959, "perplexity": 33.48456400193907, "lr": 0.0026291804804649314, "grad_norm": 0.175874, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:47.277002+00:00", "epoch": 0, "step": 17305, "train_loss": 3.456167221069336, "perplexity": 31.695262477942805, "lr": 0.0026291804804649314, "grad_norm": 0.192156, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:47.582786+00:00", "epoch": 0, "step": 17306, "train_loss": 3.480184555053711, "perplexity": 32.46571323444724, "lr": 0.0026291804804649314, "grad_norm": 0.153476, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:47.887405+00:00", "epoch": 0, "step": 17307, "train_loss": 3.4774930477142334, "perplexity": 32.37844901781516, "lr": 0.0026291804804649314, "grad_norm": 0.169393, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:48.191996+00:00", "epoch": 0, "step": 17308, "train_loss": 3.5819404125213623, "perplexity": 35.94321789415298, "lr": 0.0026291804804649314, "grad_norm": 0.20276, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:48.496715+00:00", "epoch": 0, "step": 17309, "train_loss": 3.5038812160491943, "perplexity": 33.24422992821426, "lr": 0.0026291804804649314, "grad_norm": 0.214712, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:48.801273+00:00", "epoch": 0, "step": 17310, "train_loss": 3.500042676925659, "perplexity": 33.11686525453107, "lr": 0.0026291804804649314, "grad_norm": 0.156448, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:49.105679+00:00", "epoch": 0, "step": 17311, "train_loss": 3.5368154048919678, "perplexity": 34.357330598877574, "lr": 0.0026291804804649314, "grad_norm": 0.178293, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:49.410208+00:00", "epoch": 0, "step": 17312, "train_loss": 3.4683127403259277, "perplexity": 32.082565133619106, "lr": 0.0026291804804649314, "grad_norm": 0.168029, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:49.714311+00:00", "epoch": 0, "step": 17313, "train_loss": 3.5256130695343018, "perplexity": 33.974596022565535, "lr": 0.0026291804804649314, "grad_norm": 0.1709, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:50.017807+00:00", "epoch": 0, "step": 17314, "train_loss": 3.570761203765869, "perplexity": 35.5436388084083, "lr": 0.0026291804804649314, "grad_norm": 0.168813, "tokens_per_sec": 107967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:50.322932+00:00", "epoch": 0, "step": 17315, "train_loss": 3.425736427307129, "perplexity": 30.745278178996106, "lr": 0.0026291804804649314, "grad_norm": 0.188745, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:50.628401+00:00", "epoch": 0, "step": 17316, "train_loss": 3.6354880332946777, "perplexity": 37.92035477768555, "lr": 0.0026291804804649314, "grad_norm": 0.170324, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:50.933007+00:00", "epoch": 0, "step": 17317, "train_loss": 3.513610363006592, "perplexity": 33.56924642958984, "lr": 0.0026291804804649314, "grad_norm": 0.153335, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:51.237819+00:00", "epoch": 0, "step": 17318, "train_loss": 3.5659217834472656, "perplexity": 35.37204374577421, "lr": 0.0026291804804649314, "grad_norm": 0.187768, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:51.542119+00:00", "epoch": 0, "step": 17319, "train_loss": 3.534226179122925, "perplexity": 34.26848678112135, "lr": 0.0026291804804649314, "grad_norm": 0.196613, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:51.846070+00:00", "epoch": 0, "step": 17320, "train_loss": 3.404665946960449, "perplexity": 30.104237631958895, "lr": 0.0026291804804649314, "grad_norm": 0.179969, "tokens_per_sec": 107872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:52.151200+00:00", "epoch": 0, "step": 17321, "train_loss": 3.471278667449951, "perplexity": 32.17786093409932, "lr": 0.0026291804804649314, "grad_norm": 0.16993, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:52.456115+00:00", "epoch": 0, "step": 17322, "train_loss": 3.493473529815674, "perplexity": 32.90002869027155, "lr": 0.0026291804804649314, "grad_norm": 0.155334, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:52.761285+00:00", "epoch": 0, "step": 17323, "train_loss": 3.402564764022827, "perplexity": 30.04104952962554, "lr": 0.0026291804804649314, "grad_norm": 0.157305, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:53.065541+00:00", "epoch": 0, "step": 17324, "train_loss": 3.5536351203918457, "perplexity": 34.94009838020508, "lr": 0.0026291804804649314, "grad_norm": 0.171283, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:53.369932+00:00", "epoch": 0, "step": 17325, "train_loss": 3.5386962890625, "perplexity": 34.42201356966146, "lr": 0.0026291804804649314, "grad_norm": 0.184469, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:53.675665+00:00", "epoch": 0, "step": 17326, "train_loss": 3.5200088024139404, "perplexity": 33.78472584968249, "lr": 0.0026291804804649314, "grad_norm": 0.169026, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:53.981602+00:00", "epoch": 0, "step": 17327, "train_loss": 3.626865863800049, "perplexity": 37.59480454297745, "lr": 0.0026291804804649314, "grad_norm": 0.217666, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:54.287189+00:00", "epoch": 0, "step": 17328, "train_loss": 3.528188467025757, "perplexity": 34.06220687990185, "lr": 0.0026291804804649314, "grad_norm": 0.212141, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:54.592093+00:00", "epoch": 0, "step": 17329, "train_loss": 3.5761990547180176, "perplexity": 35.73744628906474, "lr": 0.0026291804804649314, "grad_norm": 0.169418, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:54.896544+00:00", "epoch": 0, "step": 17330, "train_loss": 3.4375174045562744, "perplexity": 31.109629595554818, "lr": 0.0026291804804649314, "grad_norm": 0.188383, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:55.200844+00:00", "epoch": 0, "step": 17331, "train_loss": 3.4407012462615967, "perplexity": 31.208835575964216, "lr": 0.0026291804804649314, "grad_norm": 0.177131, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:55.506281+00:00", "epoch": 0, "step": 17332, "train_loss": 3.5219552516937256, "perplexity": 33.85055014602526, "lr": 0.0026291804804649314, "grad_norm": 0.176818, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:55.811602+00:00", "epoch": 0, "step": 17333, "train_loss": 3.61031436920166, "perplexity": 36.977675630166296, "lr": 0.0026291804804649314, "grad_norm": 0.161423, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:56.116207+00:00", "epoch": 0, "step": 17334, "train_loss": 3.4924542903900146, "perplexity": 32.86651276720169, "lr": 0.0026291804804649314, "grad_norm": 0.168225, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:56.420790+00:00", "epoch": 0, "step": 17335, "train_loss": 3.5321762561798096, "perplexity": 34.19831097591376, "lr": 0.0026291804804649314, "grad_norm": 0.159649, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:56.724476+00:00", "epoch": 0, "step": 17336, "train_loss": 3.475649833679199, "perplexity": 32.31882357426123, "lr": 0.0026291804804649314, "grad_norm": 0.173687, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:57.029791+00:00", "epoch": 0, "step": 17337, "train_loss": 3.430457353591919, "perplexity": 30.890767522906682, "lr": 0.0026291804804649314, "grad_norm": 0.180873, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:57.334861+00:00", "epoch": 0, "step": 17338, "train_loss": 3.5054502487182617, "perplexity": 33.29643215381874, "lr": 0.0026291804804649314, "grad_norm": 0.170684, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:57.639929+00:00", "epoch": 0, "step": 17339, "train_loss": 3.50417423248291, "perplexity": 33.25397246120088, "lr": 0.0026291804804649314, "grad_norm": 0.179924, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:57.944879+00:00", "epoch": 0, "step": 17340, "train_loss": 3.4729156494140625, "perplexity": 32.23057864929426, "lr": 0.0026291804804649314, "grad_norm": 0.206924, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:58.249733+00:00", "epoch": 0, "step": 17341, "train_loss": 3.5278103351593018, "perplexity": 34.049329308902195, "lr": 0.0026291804804649314, "grad_norm": 0.175969, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:58.553616+00:00", "epoch": 0, "step": 17342, "train_loss": 3.5322253704071045, "perplexity": 34.19999064077952, "lr": 0.0026291804804649314, "grad_norm": 0.2278, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:58.858814+00:00", "epoch": 0, "step": 17343, "train_loss": 3.4221651554107666, "perplexity": 30.635674260277458, "lr": 0.0026291804804649314, "grad_norm": 0.179265, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:59.164450+00:00", "epoch": 0, "step": 17344, "train_loss": 3.4759023189544678, "perplexity": 32.32698463155778, "lr": 0.0026291804804649314, "grad_norm": 0.189323, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:59.469001+00:00", "epoch": 0, "step": 17345, "train_loss": 3.5278215408325195, "perplexity": 34.04971085669746, "lr": 0.0026291804804649314, "grad_norm": 0.177965, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:28:59.773318+00:00", "epoch": 0, "step": 17346, "train_loss": 3.5487780570983887, "perplexity": 34.77080358148935, "lr": 0.0026291804804649314, "grad_norm": 0.151419, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:00.078736+00:00", "epoch": 0, "step": 17347, "train_loss": 3.5137386322021484, "perplexity": 33.5735526059938, "lr": 0.0026291804804649314, "grad_norm": 0.165344, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:00.384459+00:00", "epoch": 0, "step": 17348, "train_loss": 3.454782724380493, "perplexity": 31.651410855208006, "lr": 0.0026291804804649314, "grad_norm": 0.162594, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:00.690279+00:00", "epoch": 0, "step": 17349, "train_loss": 3.595499277114868, "perplexity": 36.433886053070204, "lr": 0.0026291804804649314, "grad_norm": 0.165687, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:00.995399+00:00", "epoch": 0, "step": 17350, "train_loss": 3.6160991191864014, "perplexity": 37.19220213156087, "lr": 0.0026291804804649314, "grad_norm": 0.159449, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:01.299818+00:00", "epoch": 0, "step": 17351, "train_loss": 3.5109336376190186, "perplexity": 33.47951092765666, "lr": 0.0026291804804649314, "grad_norm": 0.153827, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:01.604292+00:00", "epoch": 0, "step": 17352, "train_loss": 3.508592367172241, "perplexity": 33.401218026341056, "lr": 0.0026291804804649314, "grad_norm": 0.165369, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:01.909298+00:00", "epoch": 0, "step": 17353, "train_loss": 3.4949069023132324, "perplexity": 32.94722050020645, "lr": 0.0026291804804649314, "grad_norm": 0.170054, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:02.215234+00:00", "epoch": 0, "step": 17354, "train_loss": 3.422558069229126, "perplexity": 30.647713805125957, "lr": 0.0026291804804649314, "grad_norm": 0.14315, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:02.522422+00:00", "epoch": 0, "step": 17355, "train_loss": 3.526202917098999, "perplexity": 33.994641766676494, "lr": 0.0026291804804649314, "grad_norm": 0.170758, "tokens_per_sec": 106671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:02.827772+00:00", "epoch": 0, "step": 17356, "train_loss": 3.456794023513794, "perplexity": 31.715135373480827, "lr": 0.0026291804804649314, "grad_norm": 0.212187, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:03.132743+00:00", "epoch": 0, "step": 17357, "train_loss": 3.4966087341308594, "perplexity": 33.00333886682269, "lr": 0.0026291804804649314, "grad_norm": 0.180281, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:03.438607+00:00", "epoch": 0, "step": 17358, "train_loss": 3.531451940536499, "perplexity": 34.17354957291744, "lr": 0.0026291804804649314, "grad_norm": 0.150827, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:03.744992+00:00", "epoch": 0, "step": 17359, "train_loss": 3.438857316970825, "perplexity": 31.151341713513617, "lr": 0.0026291804804649314, "grad_norm": 0.190187, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:04.050196+00:00", "epoch": 0, "step": 17360, "train_loss": 3.5636847019195557, "perplexity": 35.293002044442076, "lr": 0.0026291804804649314, "grad_norm": 0.181372, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:04.354563+00:00", "epoch": 0, "step": 17361, "train_loss": 3.4230732917785645, "perplexity": 30.66350826684912, "lr": 0.0026291804804649314, "grad_norm": 0.16589, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:04.659634+00:00", "epoch": 0, "step": 17362, "train_loss": 3.6480319499969482, "perplexity": 38.399020443979545, "lr": 0.0026291804804649314, "grad_norm": 0.193712, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:04.964479+00:00", "epoch": 0, "step": 17363, "train_loss": 3.468989372253418, "perplexity": 32.10428056735369, "lr": 0.0026291804804649314, "grad_norm": 0.170817, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:05.271180+00:00", "epoch": 0, "step": 17364, "train_loss": 3.5538601875305176, "perplexity": 34.947963133187955, "lr": 0.0026291804804649314, "grad_norm": 0.169415, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:05.578009+00:00", "epoch": 0, "step": 17365, "train_loss": 3.5043857097625732, "perplexity": 33.261005664490106, "lr": 0.0026291804804649314, "grad_norm": 0.169351, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:05.883559+00:00", "epoch": 0, "step": 17366, "train_loss": 3.5099117755889893, "perplexity": 33.44531696038959, "lr": 0.0026291804804649314, "grad_norm": 0.159778, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:06.188025+00:00", "epoch": 0, "step": 17367, "train_loss": 3.5547432899475098, "perplexity": 34.97883939534407, "lr": 0.0026291804804649314, "grad_norm": 0.168832, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:06.491989+00:00", "epoch": 0, "step": 17368, "train_loss": 3.4598498344421387, "perplexity": 31.812199059353503, "lr": 0.0026291804804649314, "grad_norm": 0.161767, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:06.797463+00:00", "epoch": 0, "step": 17369, "train_loss": 3.4071738719940186, "perplexity": 30.179831555457323, "lr": 0.0026291804804649314, "grad_norm": 0.162709, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:07.102763+00:00", "epoch": 0, "step": 17370, "train_loss": 3.489701271057129, "perplexity": 32.77615505747921, "lr": 0.0026291804804649314, "grad_norm": 0.172266, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:07.408207+00:00", "epoch": 0, "step": 17371, "train_loss": 3.4992289543151855, "perplexity": 33.08992827358703, "lr": 0.0026291804804649314, "grad_norm": 0.176346, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:07.712757+00:00", "epoch": 0, "step": 17372, "train_loss": 3.502932071685791, "perplexity": 33.212691324473894, "lr": 0.0026291804804649314, "grad_norm": 0.2001, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:08.018035+00:00", "epoch": 0, "step": 17373, "train_loss": 3.485071897506714, "perplexity": 32.62477266535036, "lr": 0.0026291804804649314, "grad_norm": 0.176512, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:08.323751+00:00", "epoch": 0, "step": 17374, "train_loss": 3.419480323791504, "perplexity": 30.553532950480665, "lr": 0.0026291804804649314, "grad_norm": 0.177341, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:08.630754+00:00", "epoch": 0, "step": 17375, "train_loss": 3.534468173980713, "perplexity": 34.27678058219469, "lr": 0.0026291804804649314, "grad_norm": 0.167114, "tokens_per_sec": 106736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:08.935912+00:00", "epoch": 0, "step": 17376, "train_loss": 3.4884262084960938, "perplexity": 32.7343900414333, "lr": 0.0026291804804649314, "grad_norm": 0.185763, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:09.241687+00:00", "epoch": 0, "step": 17377, "train_loss": 3.4665026664733887, "perplexity": 32.02454584678642, "lr": 0.0026291804804649314, "grad_norm": 0.165607, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:09.547047+00:00", "epoch": 0, "step": 17378, "train_loss": 3.465573787689209, "perplexity": 31.99481273694166, "lr": 0.0026291804804649314, "grad_norm": 0.159758, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:09.853281+00:00", "epoch": 0, "step": 17379, "train_loss": 3.4673829078674316, "perplexity": 32.052747588015244, "lr": 0.0026291804804649314, "grad_norm": 0.161869, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:10.158317+00:00", "epoch": 0, "step": 17380, "train_loss": 3.491888999938965, "perplexity": 32.84793889168693, "lr": 0.0026291804804649314, "grad_norm": 0.163246, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:10.463613+00:00", "epoch": 0, "step": 17381, "train_loss": 3.5094857215881348, "perplexity": 33.43107048438833, "lr": 0.0026291804804649314, "grad_norm": 0.179704, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:10.768191+00:00", "epoch": 0, "step": 17382, "train_loss": 3.408050775527954, "perplexity": 30.206307963333092, "lr": 0.0026291804804649314, "grad_norm": 0.1722, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:11.073899+00:00", "epoch": 0, "step": 17383, "train_loss": 3.4829063415527344, "perplexity": 32.55419833856511, "lr": 0.0026291804804649314, "grad_norm": 0.153401, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:11.379471+00:00", "epoch": 0, "step": 17384, "train_loss": 3.499326229095459, "perplexity": 33.093147245648964, "lr": 0.0026291804804649314, "grad_norm": 0.169748, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:11.684572+00:00", "epoch": 0, "step": 17385, "train_loss": 3.4733994007110596, "perplexity": 32.24617400535155, "lr": 0.0026291804804649314, "grad_norm": 0.153919, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:11.989680+00:00", "epoch": 0, "step": 17386, "train_loss": 3.5614449977874756, "perplexity": 35.21404461557548, "lr": 0.0026291804804649314, "grad_norm": 0.226029, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:12.294858+00:00", "epoch": 0, "step": 17387, "train_loss": 3.6223583221435547, "perplexity": 37.425725746568766, "lr": 0.0026291804804649314, "grad_norm": 0.193356, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:12.600447+00:00", "epoch": 0, "step": 17388, "train_loss": 3.6227200031280518, "perplexity": 37.439264368085695, "lr": 0.0026291804804649314, "grad_norm": 0.197396, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:12.905813+00:00", "epoch": 0, "step": 17389, "train_loss": 3.5000858306884766, "perplexity": 33.11829440271586, "lr": 0.0026291804804649314, "grad_norm": 0.18871, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:13.210522+00:00", "epoch": 0, "step": 17390, "train_loss": 3.7104551792144775, "perplexity": 40.87240656340748, "lr": 0.0026291804804649314, "grad_norm": 0.167322, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:13.514421+00:00", "epoch": 0, "step": 17391, "train_loss": 3.40442156791687, "perplexity": 30.096881686014843, "lr": 0.0026291804804649314, "grad_norm": 0.185106, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:13.819921+00:00", "epoch": 0, "step": 17392, "train_loss": 3.5007052421569824, "perplexity": 33.13881460865562, "lr": 0.0026291804804649314, "grad_norm": 0.19804, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:14.125502+00:00", "epoch": 0, "step": 17393, "train_loss": 3.532576084136963, "perplexity": 34.21198715060566, "lr": 0.0026291804804649314, "grad_norm": 0.156623, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:14.430415+00:00", "epoch": 0, "step": 17394, "train_loss": 3.4463677406311035, "perplexity": 31.386182259488628, "lr": 0.0026291804804649314, "grad_norm": 0.173904, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:14.735411+00:00", "epoch": 0, "step": 17395, "train_loss": 3.491029739379883, "perplexity": 32.81972607616585, "lr": 0.0026291804804649314, "grad_norm": 0.16683, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:15.040560+00:00", "epoch": 0, "step": 17396, "train_loss": 3.5102784633636475, "perplexity": 33.45758319804079, "lr": 0.0026291804804649314, "grad_norm": 0.147938, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:15.345671+00:00", "epoch": 0, "step": 17397, "train_loss": 3.5369479656219482, "perplexity": 34.36188533358475, "lr": 0.0026291804804649314, "grad_norm": 0.155959, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:15.651225+00:00", "epoch": 0, "step": 17398, "train_loss": 3.652067184448242, "perplexity": 38.554282542977276, "lr": 0.0026291804804649314, "grad_norm": 0.164743, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:15.954961+00:00", "epoch": 0, "step": 17399, "train_loss": 3.4373204708099365, "perplexity": 31.10350366287262, "lr": 0.0026291804804649314, "grad_norm": 0.18288, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:16.259866+00:00", "epoch": 0, "step": 17400, "train_loss": 3.4329981803894043, "perplexity": 30.969355409649094, "lr": 0.0026291804804649314, "grad_norm": 0.161366, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:16.564999+00:00", "epoch": 0, "step": 17401, "train_loss": 3.4417169094085693, "perplexity": 31.240549342644435, "lr": 0.0026291804804649314, "grad_norm": 0.159768, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:16.870391+00:00", "epoch": 0, "step": 17402, "train_loss": 3.5159881114959717, "perplexity": 33.649160624855135, "lr": 0.0026291804804649314, "grad_norm": 0.164481, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:17.176564+00:00", "epoch": 0, "step": 17403, "train_loss": 3.4529199600219727, "perplexity": 31.592506614562662, "lr": 0.0026291804804649314, "grad_norm": 0.154678, "tokens_per_sec": 107033} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:17.482082+00:00", "epoch": 0, "step": 17404, "train_loss": 3.5889148712158203, "perplexity": 36.19477861289961, "lr": 0.0026291804804649314, "grad_norm": 0.147514, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:17.786013+00:00", "epoch": 0, "step": 17405, "train_loss": 3.4559669494628906, "perplexity": 31.688915452395356, "lr": 0.0026291804804649314, "grad_norm": 0.156175, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:18.089798+00:00", "epoch": 0, "step": 17406, "train_loss": 3.4802005290985107, "perplexity": 32.46623184734707, "lr": 0.0026291804804649314, "grad_norm": 0.168972, "tokens_per_sec": 107866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:18.394295+00:00", "epoch": 0, "step": 17407, "train_loss": 3.4265151023864746, "perplexity": 30.76922808430871, "lr": 0.0026291804804649314, "grad_norm": 0.151919, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:18.700058+00:00", "epoch": 0, "step": 17408, "train_loss": 3.4967026710510254, "perplexity": 33.006439244448714, "lr": 0.0026291804804649314, "grad_norm": 0.148536, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:19.004925+00:00", "epoch": 0, "step": 17409, "train_loss": 3.5594916343688965, "perplexity": 35.1453259271397, "lr": 0.0026291804804649314, "grad_norm": 0.165287, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:19.309242+00:00", "epoch": 0, "step": 17410, "train_loss": 3.5772130489349365, "perplexity": 35.77370223148079, "lr": 0.0026291804804649314, "grad_norm": 0.167542, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:19.613781+00:00", "epoch": 0, "step": 17411, "train_loss": 3.4163758754730225, "perplexity": 30.458828165604512, "lr": 0.0026291804804649314, "grad_norm": 0.162584, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:19.918505+00:00", "epoch": 0, "step": 17412, "train_loss": 3.4348466396331787, "perplexity": 31.02665394161552, "lr": 0.0026291804804649314, "grad_norm": 0.171525, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:20.222466+00:00", "epoch": 0, "step": 17413, "train_loss": 3.5825870037078857, "perplexity": 35.96646597725148, "lr": 0.0026291804804649314, "grad_norm": 0.167739, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:20.527500+00:00", "epoch": 0, "step": 17414, "train_loss": 3.4577488899230957, "perplexity": 31.745433553970745, "lr": 0.0026291804804649314, "grad_norm": 0.174582, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:20.832691+00:00", "epoch": 0, "step": 17415, "train_loss": 3.4533069133758545, "perplexity": 31.604733806478595, "lr": 0.0026291804804649314, "grad_norm": 0.22223, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:21.137254+00:00", "epoch": 0, "step": 17416, "train_loss": 3.453986883163452, "perplexity": 31.626231378633552, "lr": 0.0026291804804649314, "grad_norm": 0.183463, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:21.442614+00:00", "epoch": 0, "step": 17417, "train_loss": 3.361328363418579, "perplexity": 28.82745879848203, "lr": 0.0026291804804649314, "grad_norm": 0.206917, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:21.747618+00:00", "epoch": 0, "step": 17418, "train_loss": 3.5657365322113037, "perplexity": 35.365491637863585, "lr": 0.0026291804804649314, "grad_norm": 0.215402, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:22.053236+00:00", "epoch": 0, "step": 17419, "train_loss": 3.498828649520874, "perplexity": 33.076684867532, "lr": 0.0026291804804649314, "grad_norm": 0.168818, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:22.357614+00:00", "epoch": 0, "step": 17420, "train_loss": 3.548543930053711, "perplexity": 34.7626637489204, "lr": 0.0026291804804649314, "grad_norm": 0.16618, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:22.661468+00:00", "epoch": 0, "step": 17421, "train_loss": 3.628032922744751, "perplexity": 37.638705508418, "lr": 0.0026291804804649314, "grad_norm": 0.176738, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:22.964703+00:00", "epoch": 0, "step": 17422, "train_loss": 3.506993293762207, "perplexity": 33.347849708033806, "lr": 0.0026291804804649314, "grad_norm": 0.155106, "tokens_per_sec": 108062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:23.269291+00:00", "epoch": 0, "step": 17423, "train_loss": 3.4807183742523193, "perplexity": 32.48304868205402, "lr": 0.0026291804804649314, "grad_norm": 0.191335, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:23.574210+00:00", "epoch": 0, "step": 17424, "train_loss": 3.5270402431488037, "perplexity": 34.02311828618353, "lr": 0.0026291804804649314, "grad_norm": 0.170559, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:23.880670+00:00", "epoch": 0, "step": 17425, "train_loss": 3.4154269695281982, "perplexity": 30.42993931105638, "lr": 0.0026291804804649314, "grad_norm": 0.165309, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:24.184225+00:00", "epoch": 0, "step": 17426, "train_loss": 3.509084463119507, "perplexity": 33.41765867522212, "lr": 0.0026291804804649314, "grad_norm": 0.177637, "tokens_per_sec": 107947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:24.488031+00:00", "epoch": 0, "step": 17427, "train_loss": 3.4882383346557617, "perplexity": 32.72824068353512, "lr": 0.0026291804804649314, "grad_norm": 0.185159, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:24.792427+00:00", "epoch": 0, "step": 17428, "train_loss": 3.491584539413452, "perplexity": 32.83793951323521, "lr": 0.0026291804804649314, "grad_norm": 0.157309, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:25.097486+00:00", "epoch": 0, "step": 17429, "train_loss": 3.463268280029297, "perplexity": 31.9211334178525, "lr": 0.0026291804804649314, "grad_norm": 0.153643, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:25.402866+00:00", "epoch": 0, "step": 17430, "train_loss": 3.5027449131011963, "perplexity": 33.206475865831386, "lr": 0.0026291804804649314, "grad_norm": 0.172315, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:25.706945+00:00", "epoch": 0, "step": 17431, "train_loss": 3.519056558609009, "perplexity": 33.75256986638859, "lr": 0.0026291804804649314, "grad_norm": 0.17684, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:26.012157+00:00", "epoch": 0, "step": 17432, "train_loss": 3.5708632469177246, "perplexity": 35.54726597840151, "lr": 0.0026291804804649314, "grad_norm": 0.172372, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:26.316604+00:00", "epoch": 0, "step": 17433, "train_loss": 3.433488368988037, "perplexity": 30.98453995591959, "lr": 0.0026291804804649314, "grad_norm": 0.1569, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:26.621029+00:00", "epoch": 0, "step": 17434, "train_loss": 3.541428804397583, "perplexity": 34.51620087512805, "lr": 0.0026291804804649314, "grad_norm": 0.176563, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:26.925813+00:00", "epoch": 0, "step": 17435, "train_loss": 3.6283838748931885, "perplexity": 37.65191721118272, "lr": 0.0026291804804649314, "grad_norm": 0.182114, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:27.230322+00:00", "epoch": 0, "step": 17436, "train_loss": 3.5143349170684814, "perplexity": 33.59357797714132, "lr": 0.0026291804804649314, "grad_norm": 0.167629, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:27.535021+00:00", "epoch": 0, "step": 17437, "train_loss": 3.408491611480713, "perplexity": 30.219626925406402, "lr": 0.0026291804804649314, "grad_norm": 0.175152, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:27.840001+00:00", "epoch": 0, "step": 17438, "train_loss": 3.5700693130493164, "perplexity": 35.51905500031928, "lr": 0.0026291804804649314, "grad_norm": 0.151951, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:28.145512+00:00", "epoch": 0, "step": 17439, "train_loss": 3.5713119506835938, "perplexity": 35.56321974950329, "lr": 0.0026291804804649314, "grad_norm": 0.171842, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:28.451318+00:00", "epoch": 0, "step": 17440, "train_loss": 3.4990808963775635, "perplexity": 33.08502940971756, "lr": 0.0026291804804649314, "grad_norm": 0.164645, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:28.756056+00:00", "epoch": 0, "step": 17441, "train_loss": 3.490381956100464, "perplexity": 32.79847289085083, "lr": 0.0026291804804649314, "grad_norm": 0.18248, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:29.060876+00:00", "epoch": 0, "step": 17442, "train_loss": 3.5175676345825195, "perplexity": 33.70235224854491, "lr": 0.0026291804804649314, "grad_norm": 0.157298, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:29.365837+00:00", "epoch": 0, "step": 17443, "train_loss": 3.4995388984680176, "perplexity": 33.10018589293551, "lr": 0.0026291804804649314, "grad_norm": 0.154594, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:29.670357+00:00", "epoch": 0, "step": 17444, "train_loss": 3.557518243789673, "perplexity": 35.07603885975957, "lr": 0.0026291804804649314, "grad_norm": 0.185718, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:29.974458+00:00", "epoch": 0, "step": 17445, "train_loss": 3.5167553424835205, "perplexity": 33.67498720980278, "lr": 0.0026291804804649314, "grad_norm": 0.188957, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:30.278848+00:00", "epoch": 0, "step": 17446, "train_loss": 3.5112252235412598, "perplexity": 33.48927450511838, "lr": 0.0026291804804649314, "grad_norm": 0.162065, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:30.582329+00:00", "epoch": 0, "step": 17447, "train_loss": 3.50136661529541, "perplexity": 33.160738979773015, "lr": 0.0026291804804649314, "grad_norm": 0.165717, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:30.885804+00:00", "epoch": 0, "step": 17448, "train_loss": 3.583364248275757, "perplexity": 35.99443158420801, "lr": 0.0026291804804649314, "grad_norm": 0.16284, "tokens_per_sec": 107976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:31.190785+00:00", "epoch": 0, "step": 17449, "train_loss": 3.397465705871582, "perplexity": 29.888258347762637, "lr": 0.0026291804804649314, "grad_norm": 0.1584, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:31.496409+00:00", "epoch": 0, "step": 17450, "train_loss": 3.4355051517486572, "perplexity": 31.047092097787, "lr": 0.0026291804804649314, "grad_norm": 0.183897, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:31.801014+00:00", "epoch": 0, "step": 17451, "train_loss": 3.419814348220825, "perplexity": 30.563740281542263, "lr": 0.0026291804804649314, "grad_norm": 0.183901, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:32.105227+00:00", "epoch": 0, "step": 17452, "train_loss": 3.4941906929016113, "perplexity": 32.92363183901635, "lr": 0.0026291804804649314, "grad_norm": 0.172751, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:32.408922+00:00", "epoch": 0, "step": 17453, "train_loss": 3.5014419555664062, "perplexity": 33.16323741294932, "lr": 0.0026291804804649314, "grad_norm": 0.158176, "tokens_per_sec": 107899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:32.713062+00:00", "epoch": 0, "step": 17454, "train_loss": 3.4941625595092773, "perplexity": 32.92270559859397, "lr": 0.0026291804804649314, "grad_norm": 0.168853, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:33.018405+00:00", "epoch": 0, "step": 17455, "train_loss": 3.6116416454315186, "perplexity": 37.02678780556721, "lr": 0.0026291804804649314, "grad_norm": 0.182118, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:33.324019+00:00", "epoch": 0, "step": 17456, "train_loss": 3.4593122005462646, "perplexity": 31.795100339674406, "lr": 0.0026291804804649314, "grad_norm": 0.210779, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:33.629075+00:00", "epoch": 0, "step": 17457, "train_loss": 3.4208552837371826, "perplexity": 30.595571728681737, "lr": 0.0026291804804649314, "grad_norm": 0.188767, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:33.932799+00:00", "epoch": 0, "step": 17458, "train_loss": 3.621581554412842, "perplexity": 37.39666593833212, "lr": 0.0026291804804649314, "grad_norm": 0.167319, "tokens_per_sec": 107888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:34.238037+00:00", "epoch": 0, "step": 17459, "train_loss": 3.41107177734375, "perplexity": 30.29769925165629, "lr": 0.0026291804804649314, "grad_norm": 0.201315, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:34.542973+00:00", "epoch": 0, "step": 17460, "train_loss": 3.360490322113037, "perplexity": 28.803310317400715, "lr": 0.0026291804804649314, "grad_norm": 0.17871, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:34.848298+00:00", "epoch": 0, "step": 17461, "train_loss": 3.4181041717529297, "perplexity": 30.511515561613916, "lr": 0.0026291804804649314, "grad_norm": 0.175733, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:35.155506+00:00", "epoch": 0, "step": 17462, "train_loss": 3.5551917552948, "perplexity": 34.9945297107197, "lr": 0.0026291804804649314, "grad_norm": 0.189037, "tokens_per_sec": 106658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:35.460847+00:00", "epoch": 0, "step": 17463, "train_loss": 3.455207586288452, "perplexity": 31.6648611910789, "lr": 0.0026291804804649314, "grad_norm": 0.167737, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:35.765209+00:00", "epoch": 0, "step": 17464, "train_loss": 3.4831316471099854, "perplexity": 32.56153380669247, "lr": 0.0026291804804649314, "grad_norm": 0.173878, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:36.070608+00:00", "epoch": 0, "step": 17465, "train_loss": 3.5532126426696777, "perplexity": 34.92534008477412, "lr": 0.0026291804804649314, "grad_norm": 0.186514, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:36.375966+00:00", "epoch": 0, "step": 17466, "train_loss": 3.514019012451172, "perplexity": 33.58296728681766, "lr": 0.0026291804804649314, "grad_norm": 0.177771, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:36.681186+00:00", "epoch": 0, "step": 17467, "train_loss": 3.4385530948638916, "perplexity": 31.141866228102874, "lr": 0.0026291804804649314, "grad_norm": 0.190067, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:36.986974+00:00", "epoch": 0, "step": 17468, "train_loss": 3.518120288848877, "perplexity": 33.720983145049374, "lr": 0.0026291804804649314, "grad_norm": 0.17961, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:37.292105+00:00", "epoch": 0, "step": 17469, "train_loss": 3.582967758178711, "perplexity": 35.98016297740365, "lr": 0.0026291804804649314, "grad_norm": 0.168961, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:37.596622+00:00", "epoch": 0, "step": 17470, "train_loss": 3.496155261993408, "perplexity": 32.988376165044315, "lr": 0.0026291804804649314, "grad_norm": 0.171494, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:37.901188+00:00", "epoch": 0, "step": 17471, "train_loss": 3.4937753677368164, "perplexity": 32.9099606653849, "lr": 0.0026291804804649314, "grad_norm": 0.178172, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:38.205647+00:00", "epoch": 0, "step": 17472, "train_loss": 3.6003904342651367, "perplexity": 36.61252643830896, "lr": 0.0026291804804649314, "grad_norm": 0.162306, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:38.510949+00:00", "epoch": 0, "step": 17473, "train_loss": 3.4583706855773926, "perplexity": 31.765178864736413, "lr": 0.0026291804804649314, "grad_norm": 0.179638, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:38.815569+00:00", "epoch": 0, "step": 17474, "train_loss": 3.423405647277832, "perplexity": 30.67370114618437, "lr": 0.0026291804804649314, "grad_norm": 0.170416, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:39.119790+00:00", "epoch": 0, "step": 17475, "train_loss": 3.571383476257324, "perplexity": 35.56576352017082, "lr": 0.0026291804804649314, "grad_norm": 0.186282, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:39.423634+00:00", "epoch": 0, "step": 17476, "train_loss": 3.561002016067505, "perplexity": 35.19844889208979, "lr": 0.0026291804804649314, "grad_norm": 0.167608, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:39.728187+00:00", "epoch": 0, "step": 17477, "train_loss": 3.5111680030822754, "perplexity": 33.48735828828407, "lr": 0.0026291804804649314, "grad_norm": 0.173782, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:40.032932+00:00", "epoch": 0, "step": 17478, "train_loss": 3.557898759841919, "perplexity": 35.08938839529012, "lr": 0.0026291804804649314, "grad_norm": 0.204223, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:40.336783+00:00", "epoch": 0, "step": 17479, "train_loss": 3.4443604946136475, "perplexity": 31.323245655899726, "lr": 0.0026291804804649314, "grad_norm": 0.171089, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:40.641941+00:00", "epoch": 0, "step": 17480, "train_loss": 3.612414598464966, "perplexity": 37.055418837316275, "lr": 0.0026291804804649314, "grad_norm": 0.180839, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:40.946414+00:00", "epoch": 0, "step": 17481, "train_loss": 3.4544053077697754, "perplexity": 31.6394673409807, "lr": 0.0026291804804649314, "grad_norm": 0.194525, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:41.251729+00:00", "epoch": 0, "step": 17482, "train_loss": 3.3705074787139893, "perplexity": 29.093287536310616, "lr": 0.0026291804804649314, "grad_norm": 0.174425, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:41.557062+00:00", "epoch": 0, "step": 17483, "train_loss": 3.547542095184326, "perplexity": 34.72785473957511, "lr": 0.0026291804804649314, "grad_norm": 0.156024, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:41.862048+00:00", "epoch": 0, "step": 17484, "train_loss": 3.5092411041259766, "perplexity": 33.42289366090688, "lr": 0.0026291804804649314, "grad_norm": 0.167022, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:42.166202+00:00", "epoch": 0, "step": 17485, "train_loss": 3.6108927726745605, "perplexity": 36.99906983281467, "lr": 0.0026291804804649314, "grad_norm": 0.157349, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:42.470542+00:00", "epoch": 0, "step": 17486, "train_loss": 3.5276715755462646, "perplexity": 34.044604964924766, "lr": 0.0026291804804649314, "grad_norm": 0.186996, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:42.775799+00:00", "epoch": 0, "step": 17487, "train_loss": 3.562098741531372, "perplexity": 35.237073103484995, "lr": 0.0026291804804649314, "grad_norm": 0.196948, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:43.081336+00:00", "epoch": 0, "step": 17488, "train_loss": 3.50325608253479, "perplexity": 33.22345434036028, "lr": 0.0026291804804649314, "grad_norm": 0.158297, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:43.385946+00:00", "epoch": 0, "step": 17489, "train_loss": 3.529601812362671, "perplexity": 34.110382577589974, "lr": 0.0026291804804649314, "grad_norm": 0.189911, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:43.690352+00:00", "epoch": 0, "step": 17490, "train_loss": 3.5004241466522217, "perplexity": 33.12950074593982, "lr": 0.0026291804804649314, "grad_norm": 0.166173, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:43.994028+00:00", "epoch": 0, "step": 17491, "train_loss": 3.531442880630493, "perplexity": 34.173239965172925, "lr": 0.0026291804804649314, "grad_norm": 0.190896, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:44.298216+00:00", "epoch": 0, "step": 17492, "train_loss": 3.5261383056640625, "perplexity": 33.99244539504792, "lr": 0.0026291804804649314, "grad_norm": 0.209575, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:44.603131+00:00", "epoch": 0, "step": 17493, "train_loss": 3.5899112224578857, "perplexity": 36.230859297059276, "lr": 0.0026291804804649314, "grad_norm": 0.199134, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:44.907896+00:00", "epoch": 0, "step": 17494, "train_loss": 3.4365036487579346, "perplexity": 31.078108008463527, "lr": 0.0026291804804649314, "grad_norm": 0.177266, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:45.211758+00:00", "epoch": 0, "step": 17495, "train_loss": 3.438849925994873, "perplexity": 31.15111147554698, "lr": 0.0026291804804649314, "grad_norm": 0.171943, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:45.516246+00:00", "epoch": 0, "step": 17496, "train_loss": 3.4799766540527344, "perplexity": 32.4589642817497, "lr": 0.0026291804804649314, "grad_norm": 0.171409, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:45.821335+00:00", "epoch": 0, "step": 17497, "train_loss": 3.5697925090789795, "perplexity": 35.50922454548981, "lr": 0.0026291804804649314, "grad_norm": 0.188724, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:46.127214+00:00", "epoch": 0, "step": 17498, "train_loss": 3.5117294788360596, "perplexity": 33.506165907538225, "lr": 0.0026291804804649314, "grad_norm": 0.15387, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:46.432905+00:00", "epoch": 0, "step": 17499, "train_loss": 3.541074514389038, "perplexity": 34.503974296030364, "lr": 0.0026291804804649314, "grad_norm": 0.17093, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:46.737835+00:00", "epoch": 0, "step": 17500, "train_loss": 3.5291025638580322, "perplexity": 34.09335727037867, "lr": 0.0026291804804649314, "grad_norm": 0.157166, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:29:49.812275+00:00", "step": 17500, "epoch": 0, "val_loss": 3.464469289779663, "val_ppl": 31.959494041460548, "eval_train_loss": 3.5291025638580322, "eval_train_ppl": 34.09335727037867} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:29:50.771008+00:00", "step": 17500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4645_epoch_0000_step_0017500.pt", "val_loss": 3.464469289779663} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T22:29:52.707203+00:00", "step": 17500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0017500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:54.142974+00:00", "epoch": 0, "step": 17501, "train_loss": 3.4968202114105225, "perplexity": 33.01031906119629, "lr": 0.0026291804804649314, "grad_norm": 0.158918, "tokens_per_sec": 4425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:54.446185+00:00", "epoch": 0, "step": 17502, "train_loss": 3.5602123737335205, "perplexity": 35.17066567759883, "lr": 0.0026291804804649314, "grad_norm": 0.147794, "tokens_per_sec": 108064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:54.749449+00:00", "epoch": 0, "step": 17503, "train_loss": 3.555887460708618, "perplexity": 35.01888406523932, "lr": 0.0026291804804649314, "grad_norm": 0.151087, "tokens_per_sec": 108052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:55.052318+00:00", "epoch": 0, "step": 17504, "train_loss": 3.576857089996338, "perplexity": 35.7609705285209, "lr": 0.0026291804804649314, "grad_norm": 0.173472, "tokens_per_sec": 108193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:55.356479+00:00", "epoch": 0, "step": 17505, "train_loss": 3.474963903427124, "perplexity": 32.296662716730346, "lr": 0.0026291804804649314, "grad_norm": 0.168201, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:55.660772+00:00", "epoch": 0, "step": 17506, "train_loss": 3.5242884159088135, "perplexity": 33.92962124534706, "lr": 0.0026291804804649314, "grad_norm": 0.169042, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:55.964273+00:00", "epoch": 0, "step": 17507, "train_loss": 3.505479097366333, "perplexity": 33.297392724727494, "lr": 0.0026291804804649314, "grad_norm": 0.179576, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:56.267148+00:00", "epoch": 0, "step": 17508, "train_loss": 3.5351791381835938, "perplexity": 34.30115881117583, "lr": 0.0026291804804649314, "grad_norm": 0.15863, "tokens_per_sec": 108190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:56.570824+00:00", "epoch": 0, "step": 17509, "train_loss": 3.5476951599121094, "perplexity": 34.733170756044245, "lr": 0.0026291804804649314, "grad_norm": 0.175553, "tokens_per_sec": 107907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:56.875442+00:00", "epoch": 0, "step": 17510, "train_loss": 3.476491928100586, "perplexity": 32.34605053752689, "lr": 0.0026291804804649314, "grad_norm": 0.173273, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:57.179534+00:00", "epoch": 0, "step": 17511, "train_loss": 3.529191732406616, "perplexity": 34.09639746110551, "lr": 0.0026291804804649314, "grad_norm": 0.161703, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:57.483822+00:00", "epoch": 0, "step": 17512, "train_loss": 3.4900705814361572, "perplexity": 32.78826186717421, "lr": 0.0026291804804649314, "grad_norm": 0.16009, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:57.788857+00:00", "epoch": 0, "step": 17513, "train_loss": 3.4786136150360107, "perplexity": 32.41475158565129, "lr": 0.0026291804804649314, "grad_norm": 0.167207, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:58.318806+00:00", "epoch": 0, "step": 17514, "train_loss": 3.5539538860321045, "perplexity": 34.951237858383074, "lr": 0.0026291804804649314, "grad_norm": 0.143698, "tokens_per_sec": 61832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:58.622165+00:00", "epoch": 0, "step": 17515, "train_loss": 3.3969898223876953, "perplexity": 29.874038403039574, "lr": 0.0026291804804649314, "grad_norm": 0.144248, "tokens_per_sec": 108026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:58.927087+00:00", "epoch": 0, "step": 17516, "train_loss": 3.5873513221740723, "perplexity": 36.13823052086088, "lr": 0.0026291804804649314, "grad_norm": 0.151161, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:59.232327+00:00", "epoch": 0, "step": 17517, "train_loss": 3.515132427215576, "perplexity": 33.62037988243047, "lr": 0.0026291804804649314, "grad_norm": 0.179176, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:59.536238+00:00", "epoch": 0, "step": 17518, "train_loss": 3.4964120388031006, "perplexity": 32.99684790265923, "lr": 0.0026291804804649314, "grad_norm": 0.149744, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:29:59.839896+00:00", "epoch": 0, "step": 17519, "train_loss": 3.6157941818237305, "perplexity": 37.180862568547035, "lr": 0.0026291804804649314, "grad_norm": 0.18087, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:00.144215+00:00", "epoch": 0, "step": 17520, "train_loss": 3.425727367401123, "perplexity": 30.74499963092749, "lr": 0.0026291804804649314, "grad_norm": 0.19401, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:00.449007+00:00", "epoch": 0, "step": 17521, "train_loss": 3.4563393592834473, "perplexity": 31.700718913438603, "lr": 0.0026291804804649314, "grad_norm": 0.203264, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:00.753887+00:00", "epoch": 0, "step": 17522, "train_loss": 3.364142656326294, "perplexity": 28.908701978756916, "lr": 0.0026291804804649314, "grad_norm": 0.146588, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:01.058781+00:00", "epoch": 0, "step": 17523, "train_loss": 3.4694933891296387, "perplexity": 32.12046574501727, "lr": 0.0026291804804649314, "grad_norm": 0.167096, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:01.362900+00:00", "epoch": 0, "step": 17524, "train_loss": 3.5965285301208496, "perplexity": 36.47140504476522, "lr": 0.0026291804804649314, "grad_norm": 0.157191, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:01.666083+00:00", "epoch": 0, "step": 17525, "train_loss": 3.466627359390259, "perplexity": 32.028539329793844, "lr": 0.0026291804804649314, "grad_norm": 0.173755, "tokens_per_sec": 108079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:01.969849+00:00", "epoch": 0, "step": 17526, "train_loss": 3.556769371032715, "perplexity": 35.049781202884574, "lr": 0.0026291804804649314, "grad_norm": 0.153321, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:02.274216+00:00", "epoch": 0, "step": 17527, "train_loss": 3.4218218326568604, "perplexity": 30.625158141537366, "lr": 0.0026291804804649314, "grad_norm": 0.170752, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:02.578680+00:00", "epoch": 0, "step": 17528, "train_loss": 3.48561954498291, "perplexity": 32.64264443302958, "lr": 0.0026291804804649314, "grad_norm": 0.162346, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:02.882926+00:00", "epoch": 0, "step": 17529, "train_loss": 3.5030174255371094, "perplexity": 33.215526276576405, "lr": 0.0026291804804649314, "grad_norm": 0.173196, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:03.187158+00:00", "epoch": 0, "step": 17530, "train_loss": 3.405224561691284, "perplexity": 30.121059000455354, "lr": 0.0026291804804649314, "grad_norm": 0.21627, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:03.490086+00:00", "epoch": 0, "step": 17531, "train_loss": 3.5413079261779785, "perplexity": 34.51202887037609, "lr": 0.0026291804804649314, "grad_norm": 0.172709, "tokens_per_sec": 108076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:03.793666+00:00", "epoch": 0, "step": 17532, "train_loss": 3.4587955474853516, "perplexity": 31.77867754657035, "lr": 0.0026291804804649314, "grad_norm": 0.186434, "tokens_per_sec": 107938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:04.098605+00:00", "epoch": 0, "step": 17533, "train_loss": 3.482631206512451, "perplexity": 32.5452427699457, "lr": 0.0026291804804649314, "grad_norm": 0.166517, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:04.402654+00:00", "epoch": 0, "step": 17534, "train_loss": 3.603303909301758, "perplexity": 36.71935166086017, "lr": 0.0026291804804649314, "grad_norm": 0.155694, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:04.706545+00:00", "epoch": 0, "step": 17535, "train_loss": 3.5110981464385986, "perplexity": 33.48501905553488, "lr": 0.0026291804804649314, "grad_norm": 0.169833, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:05.009454+00:00", "epoch": 0, "step": 17536, "train_loss": 3.4723289012908936, "perplexity": 32.211672964742995, "lr": 0.0026291804804649314, "grad_norm": 0.161111, "tokens_per_sec": 108178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:05.313638+00:00", "epoch": 0, "step": 17537, "train_loss": 3.568610429763794, "perplexity": 35.46727462460794, "lr": 0.0026291804804649314, "grad_norm": 0.169904, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:05.617841+00:00", "epoch": 0, "step": 17538, "train_loss": 3.63983154296875, "perplexity": 38.08542042811976, "lr": 0.0026291804804649314, "grad_norm": 0.165342, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:05.923421+00:00", "epoch": 0, "step": 17539, "train_loss": 3.3999650478363037, "perplexity": 29.96305275556981, "lr": 0.0026291804804649314, "grad_norm": 0.171684, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:06.227994+00:00", "epoch": 0, "step": 17540, "train_loss": 3.484788656234741, "perplexity": 32.61553329179054, "lr": 0.0026291804804649314, "grad_norm": 0.176993, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:06.531312+00:00", "epoch": 0, "step": 17541, "train_loss": 3.4246068000793457, "perplexity": 30.71056708462911, "lr": 0.0026291804804649314, "grad_norm": 0.167349, "tokens_per_sec": 107975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:06.835325+00:00", "epoch": 0, "step": 17542, "train_loss": 3.598011016845703, "perplexity": 36.525513516216336, "lr": 0.0026291804804649314, "grad_norm": 0.162524, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:07.141131+00:00", "epoch": 0, "step": 17543, "train_loss": 3.4338457584381104, "perplexity": 30.99561548263552, "lr": 0.0026291804804649314, "grad_norm": 0.168998, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:07.447496+00:00", "epoch": 0, "step": 17544, "train_loss": 3.4527575969696045, "perplexity": 31.587377575151276, "lr": 0.0026291804804649314, "grad_norm": 0.150398, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:07.752921+00:00", "epoch": 0, "step": 17545, "train_loss": 3.54364013671875, "perplexity": 34.59261211991725, "lr": 0.0026291804804649314, "grad_norm": 0.174788, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:08.057206+00:00", "epoch": 0, "step": 17546, "train_loss": 3.470945119857788, "perplexity": 32.16712987582261, "lr": 0.0026291804804649314, "grad_norm": 0.14946, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:08.361956+00:00", "epoch": 0, "step": 17547, "train_loss": 3.514707565307617, "perplexity": 33.60609889762536, "lr": 0.0026291804804649314, "grad_norm": 0.191902, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:08.665845+00:00", "epoch": 0, "step": 17548, "train_loss": 3.527886152267456, "perplexity": 34.05191092844921, "lr": 0.0026291804804649314, "grad_norm": 0.224655, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:08.970450+00:00", "epoch": 0, "step": 17549, "train_loss": 3.5394723415374756, "perplexity": 34.44873722662469, "lr": 0.0026291804804649314, "grad_norm": 0.217742, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:09.274880+00:00", "epoch": 0, "step": 17550, "train_loss": 3.5352678298950195, "perplexity": 34.304201174568895, "lr": 0.0026291804804649314, "grad_norm": 0.172825, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:09.579727+00:00", "epoch": 0, "step": 17551, "train_loss": 3.6193366050720215, "perplexity": 37.312806483118884, "lr": 0.0026291804804649314, "grad_norm": 0.237577, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:09.884274+00:00", "epoch": 0, "step": 17552, "train_loss": 3.525942087173462, "perplexity": 33.98577610306128, "lr": 0.0026291804804649314, "grad_norm": 0.210693, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:10.190481+00:00", "epoch": 0, "step": 17553, "train_loss": 3.510148525238037, "perplexity": 33.45323606482787, "lr": 0.0026291804804649314, "grad_norm": 0.185441, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:10.495416+00:00", "epoch": 0, "step": 17554, "train_loss": 3.421394109725952, "perplexity": 30.612061860127135, "lr": 0.0026291804804649314, "grad_norm": 0.16567, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:10.801079+00:00", "epoch": 0, "step": 17555, "train_loss": 3.430837631225586, "perplexity": 30.90251682473921, "lr": 0.0026291804804649314, "grad_norm": 0.177113, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:11.105790+00:00", "epoch": 0, "step": 17556, "train_loss": 3.5391247272491455, "perplexity": 34.43676443442717, "lr": 0.0026291804804649314, "grad_norm": 0.177285, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:11.410414+00:00", "epoch": 0, "step": 17557, "train_loss": 3.5647284984588623, "perplexity": 35.329859990589696, "lr": 0.0026291804804649314, "grad_norm": 0.153988, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:11.715007+00:00", "epoch": 0, "step": 17558, "train_loss": 3.419924259185791, "perplexity": 30.567099756347748, "lr": 0.0026291804804649314, "grad_norm": 0.190395, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:12.021251+00:00", "epoch": 0, "step": 17559, "train_loss": 3.495759963989258, "perplexity": 32.975338502836806, "lr": 0.0026291804804649314, "grad_norm": 0.179545, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:12.327250+00:00", "epoch": 0, "step": 17560, "train_loss": 3.6050026416778564, "perplexity": 36.78178102273228, "lr": 0.0026291804804649314, "grad_norm": 0.17429, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:12.631671+00:00", "epoch": 0, "step": 17561, "train_loss": 3.466092348098755, "perplexity": 32.01140828266248, "lr": 0.0026291804804649314, "grad_norm": 0.189738, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:12.936003+00:00", "epoch": 0, "step": 17562, "train_loss": 3.548051118850708, "perplexity": 34.745536539365716, "lr": 0.0026291804804649314, "grad_norm": 0.19496, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:13.241387+00:00", "epoch": 0, "step": 17563, "train_loss": 3.5607264041900635, "perplexity": 35.18874911825546, "lr": 0.0026291804804649314, "grad_norm": 0.205279, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:13.546677+00:00", "epoch": 0, "step": 17564, "train_loss": 3.5840346813201904, "perplexity": 36.01857153176273, "lr": 0.0026291804804649314, "grad_norm": 0.192166, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:13.850955+00:00", "epoch": 0, "step": 17565, "train_loss": 3.5201122760772705, "perplexity": 33.78822185990013, "lr": 0.0026291804804649314, "grad_norm": 0.210849, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:14.154850+00:00", "epoch": 0, "step": 17566, "train_loss": 3.456202507019043, "perplexity": 31.69638089511264, "lr": 0.0026291804804649314, "grad_norm": 0.194625, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:14.459437+00:00", "epoch": 0, "step": 17567, "train_loss": 3.496868848800659, "perplexity": 33.011924636008175, "lr": 0.0026291804804649314, "grad_norm": 0.192319, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:14.763938+00:00", "epoch": 0, "step": 17568, "train_loss": 3.4545013904571533, "perplexity": 31.642507492080632, "lr": 0.0026291804804649314, "grad_norm": 0.1656, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:15.069003+00:00", "epoch": 0, "step": 17569, "train_loss": 3.4691872596740723, "perplexity": 32.11063422926127, "lr": 0.0026291804804649314, "grad_norm": 0.157078, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:15.373230+00:00", "epoch": 0, "step": 17570, "train_loss": 3.4130849838256836, "perplexity": 30.358756215693294, "lr": 0.0026291804804649314, "grad_norm": 0.157452, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:15.677065+00:00", "epoch": 0, "step": 17571, "train_loss": 3.5186915397644043, "perplexity": 33.74025179063519, "lr": 0.0026291804804649314, "grad_norm": 0.169375, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:15.981391+00:00", "epoch": 0, "step": 17572, "train_loss": 3.5139682292938232, "perplexity": 33.58126188100902, "lr": 0.0026291804804649314, "grad_norm": 0.173677, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:16.287027+00:00", "epoch": 0, "step": 17573, "train_loss": 3.520819664001465, "perplexity": 33.81213169579138, "lr": 0.0026291804804649314, "grad_norm": 0.153525, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:16.592213+00:00", "epoch": 0, "step": 17574, "train_loss": 3.4885268211364746, "perplexity": 32.73768370053572, "lr": 0.0026291804804649314, "grad_norm": 0.148551, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:16.896183+00:00", "epoch": 0, "step": 17575, "train_loss": 3.617757797241211, "perplexity": 37.25394321117821, "lr": 0.0026291804804649314, "grad_norm": 0.169339, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:17.199594+00:00", "epoch": 0, "step": 17576, "train_loss": 3.380436420440674, "perplexity": 29.383591915588234, "lr": 0.0026291804804649314, "grad_norm": 0.163752, "tokens_per_sec": 108000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:17.503536+00:00", "epoch": 0, "step": 17577, "train_loss": 3.5097815990448, "perplexity": 33.44096344797712, "lr": 0.0026291804804649314, "grad_norm": 0.204933, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:17.808579+00:00", "epoch": 0, "step": 17578, "train_loss": 3.5477426052093506, "perplexity": 34.73481872074867, "lr": 0.0026291804804649314, "grad_norm": 0.150558, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:18.113566+00:00", "epoch": 0, "step": 17579, "train_loss": 3.531432867050171, "perplexity": 34.17289777040297, "lr": 0.0026291804804649314, "grad_norm": 0.158704, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:18.417520+00:00", "epoch": 0, "step": 17580, "train_loss": 3.4595091342926025, "perplexity": 31.80136248449307, "lr": 0.0026291804804649314, "grad_norm": 0.170717, "tokens_per_sec": 107904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:18.721649+00:00", "epoch": 0, "step": 17581, "train_loss": 3.590531587600708, "perplexity": 36.253342632479395, "lr": 0.0026291804804649314, "grad_norm": 0.161714, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:19.026593+00:00", "epoch": 0, "step": 17582, "train_loss": 3.6069753170013428, "perplexity": 36.85441114878538, "lr": 0.0026291804804649314, "grad_norm": 0.150091, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:19.331088+00:00", "epoch": 0, "step": 17583, "train_loss": 3.53204345703125, "perplexity": 34.193769770874745, "lr": 0.0026291804804649314, "grad_norm": 0.190416, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:19.636515+00:00", "epoch": 0, "step": 17584, "train_loss": 3.4631357192993164, "perplexity": 31.916902209557048, "lr": 0.0026291804804649314, "grad_norm": 0.183198, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:19.940740+00:00", "epoch": 0, "step": 17585, "train_loss": 3.463959217071533, "perplexity": 31.943196532594946, "lr": 0.0026291804804649314, "grad_norm": 0.170167, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:20.245210+00:00", "epoch": 0, "step": 17586, "train_loss": 3.5256943702697754, "perplexity": 33.97735829449544, "lr": 0.0026291804804649314, "grad_norm": 0.163738, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:20.550907+00:00", "epoch": 0, "step": 17587, "train_loss": 3.5042026042938232, "perplexity": 33.25491595000384, "lr": 0.0026291804804649314, "grad_norm": 0.170586, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:20.855384+00:00", "epoch": 0, "step": 17588, "train_loss": 3.596825122833252, "perplexity": 36.48222380201555, "lr": 0.0026291804804649314, "grad_norm": 0.180613, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:21.159633+00:00", "epoch": 0, "step": 17589, "train_loss": 3.5279271602630615, "perplexity": 34.05330735769511, "lr": 0.0026291804804649314, "grad_norm": 0.18554, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:21.464910+00:00", "epoch": 0, "step": 17590, "train_loss": 3.484053611755371, "perplexity": 32.59156823287383, "lr": 0.0026291804804649314, "grad_norm": 0.169369, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:21.769563+00:00", "epoch": 0, "step": 17591, "train_loss": 3.469787359237671, "perplexity": 32.12990958983841, "lr": 0.0026291804804649314, "grad_norm": 0.192849, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:22.075404+00:00", "epoch": 0, "step": 17592, "train_loss": 3.595046281814575, "perplexity": 36.41738541155545, "lr": 0.0026291804804649314, "grad_norm": 0.165808, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:22.380676+00:00", "epoch": 0, "step": 17593, "train_loss": 3.539616346359253, "perplexity": 34.45369836809504, "lr": 0.0026291804804649314, "grad_norm": 0.17338, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:22.685292+00:00", "epoch": 0, "step": 17594, "train_loss": 3.4739134311676025, "perplexity": 32.26275378178724, "lr": 0.0026291804804649314, "grad_norm": 0.167597, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:22.989104+00:00", "epoch": 0, "step": 17595, "train_loss": 3.4428813457489014, "perplexity": 31.276948161531873, "lr": 0.0026291804804649314, "grad_norm": 0.19194, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:23.294085+00:00", "epoch": 0, "step": 17596, "train_loss": 3.5084192752838135, "perplexity": 33.39543704677185, "lr": 0.0026291804804649314, "grad_norm": 0.190245, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:23.598716+00:00", "epoch": 0, "step": 17597, "train_loss": 3.5121147632598877, "perplexity": 33.51907779858013, "lr": 0.0026291804804649314, "grad_norm": 0.165786, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:23.904434+00:00", "epoch": 0, "step": 17598, "train_loss": 3.520325183868408, "perplexity": 33.79541640144259, "lr": 0.0026291804804649314, "grad_norm": 0.163488, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:24.208805+00:00", "epoch": 0, "step": 17599, "train_loss": 3.5076427459716797, "perplexity": 33.36951457709489, "lr": 0.0026291804804649314, "grad_norm": 0.178134, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:24.512634+00:00", "epoch": 0, "step": 17600, "train_loss": 3.5038578510284424, "perplexity": 33.24345318516645, "lr": 0.0026291804804649314, "grad_norm": 0.159692, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:24.817663+00:00", "epoch": 0, "step": 17601, "train_loss": 3.565075159072876, "perplexity": 35.342109584650274, "lr": 0.0026291804804649314, "grad_norm": 0.201095, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:25.123320+00:00", "epoch": 0, "step": 17602, "train_loss": 3.4272475242614746, "perplexity": 30.791772394999867, "lr": 0.0026291804804649314, "grad_norm": 0.19088, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:25.428066+00:00", "epoch": 0, "step": 17603, "train_loss": 3.525817394256592, "perplexity": 33.98153858170684, "lr": 0.0026291804804649314, "grad_norm": 0.166677, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:25.731917+00:00", "epoch": 0, "step": 17604, "train_loss": 3.50917387008667, "perplexity": 33.42064658030174, "lr": 0.0026291804804649314, "grad_norm": 0.189375, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:26.036019+00:00", "epoch": 0, "step": 17605, "train_loss": 3.557170867919922, "perplexity": 35.06385640632092, "lr": 0.0026291804804649314, "grad_norm": 0.149802, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:26.340245+00:00", "epoch": 0, "step": 17606, "train_loss": 3.581930637359619, "perplexity": 35.942866545101744, "lr": 0.0026291804804649314, "grad_norm": 0.182494, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:26.646167+00:00", "epoch": 0, "step": 17607, "train_loss": 3.4610350131988525, "perplexity": 31.849924553194025, "lr": 0.0026291804804649314, "grad_norm": 0.186868, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:26.951585+00:00", "epoch": 0, "step": 17608, "train_loss": 3.505852222442627, "perplexity": 33.309819135086784, "lr": 0.0026291804804649314, "grad_norm": 0.208494, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:27.256705+00:00", "epoch": 0, "step": 17609, "train_loss": 3.547685384750366, "perplexity": 34.732831235341685, "lr": 0.0026291804804649314, "grad_norm": 0.182597, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:27.560971+00:00", "epoch": 0, "step": 17610, "train_loss": 3.5584559440612793, "perplexity": 35.108945096607705, "lr": 0.0026291804804649314, "grad_norm": 0.19579, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:27.865775+00:00", "epoch": 0, "step": 17611, "train_loss": 3.52266001701355, "perplexity": 33.87441524848563, "lr": 0.0026291804804649314, "grad_norm": 0.212096, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:28.170040+00:00", "epoch": 0, "step": 17612, "train_loss": 3.564528703689575, "perplexity": 35.32280197446476, "lr": 0.0026291804804649314, "grad_norm": 0.177124, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:28.474584+00:00", "epoch": 0, "step": 17613, "train_loss": 3.450716018676758, "perplexity": 31.522955254747593, "lr": 0.0026291804804649314, "grad_norm": 0.222053, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:28.779562+00:00", "epoch": 0, "step": 17614, "train_loss": 3.52980899810791, "perplexity": 34.11745049478527, "lr": 0.0026291804804649314, "grad_norm": 0.187966, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:29.083851+00:00", "epoch": 0, "step": 17615, "train_loss": 3.4695541858673096, "perplexity": 32.12241862391077, "lr": 0.0026291804804649314, "grad_norm": 0.202832, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:29.387854+00:00", "epoch": 0, "step": 17616, "train_loss": 3.4499876499176025, "perplexity": 31.50000327870965, "lr": 0.0026291804804649314, "grad_norm": 0.159484, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:29.692316+00:00", "epoch": 0, "step": 17617, "train_loss": 3.5449838638305664, "perplexity": 34.639126394938685, "lr": 0.0026291804804649314, "grad_norm": 0.195071, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:29.997710+00:00", "epoch": 0, "step": 17618, "train_loss": 3.460176944732666, "perplexity": 31.822606859186532, "lr": 0.0026291804804649314, "grad_norm": 0.170147, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:30.303467+00:00", "epoch": 0, "step": 17619, "train_loss": 3.526444673538208, "perplexity": 34.00286118373066, "lr": 0.0026291804804649314, "grad_norm": 0.169087, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:30.607590+00:00", "epoch": 0, "step": 17620, "train_loss": 3.4474916458129883, "perplexity": 31.421477182727937, "lr": 0.0026291804804649314, "grad_norm": 0.158652, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:30.911693+00:00", "epoch": 0, "step": 17621, "train_loss": 3.551776170730591, "perplexity": 34.875206829906816, "lr": 0.0026291804804649314, "grad_norm": 0.169749, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:31.215454+00:00", "epoch": 0, "step": 17622, "train_loss": 3.4178466796875, "perplexity": 30.503660099855505, "lr": 0.0026291804804649314, "grad_norm": 0.171721, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:31.520018+00:00", "epoch": 0, "step": 17623, "train_loss": 3.498667001724243, "perplexity": 33.071338526427084, "lr": 0.0026291804804649314, "grad_norm": 0.153296, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:31.824933+00:00", "epoch": 0, "step": 17624, "train_loss": 3.443998098373413, "perplexity": 31.311896286050782, "lr": 0.0026291804804649314, "grad_norm": 0.174419, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:32.129956+00:00", "epoch": 0, "step": 17625, "train_loss": 3.5486836433410645, "perplexity": 34.7675208942459, "lr": 0.0026291804804649314, "grad_norm": 0.173422, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:32.434285+00:00", "epoch": 0, "step": 17626, "train_loss": 3.512671947479248, "perplexity": 33.53775930381365, "lr": 0.0026291804804649314, "grad_norm": 0.173658, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:32.739392+00:00", "epoch": 0, "step": 17627, "train_loss": 3.5581302642822266, "perplexity": 35.09751268487958, "lr": 0.0026291804804649314, "grad_norm": 0.168961, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:33.044044+00:00", "epoch": 0, "step": 17628, "train_loss": 3.567495346069336, "perplexity": 35.42774768700467, "lr": 0.0026291804804649314, "grad_norm": 0.154985, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:33.349825+00:00", "epoch": 0, "step": 17629, "train_loss": 3.5656614303588867, "perplexity": 35.36283572366323, "lr": 0.0026291804804649314, "grad_norm": 0.164983, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:33.655535+00:00", "epoch": 0, "step": 17630, "train_loss": 3.541813373565674, "perplexity": 34.529477294471405, "lr": 0.0026291804804649314, "grad_norm": 0.146937, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:33.959604+00:00", "epoch": 0, "step": 17631, "train_loss": 3.6566030979156494, "perplexity": 38.72955865050026, "lr": 0.0026291804804649314, "grad_norm": 0.146805, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:34.263132+00:00", "epoch": 0, "step": 17632, "train_loss": 3.519277334213257, "perplexity": 33.760022433037946, "lr": 0.0026291804804649314, "grad_norm": 0.14676, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:34.567984+00:00", "epoch": 0, "step": 17633, "train_loss": 3.483017683029175, "perplexity": 32.557823172865916, "lr": 0.0026291804804649314, "grad_norm": 0.181154, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:34.872546+00:00", "epoch": 0, "step": 17634, "train_loss": 3.5459933280944824, "perplexity": 34.674111010056194, "lr": 0.0026291804804649314, "grad_norm": 0.190784, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:35.176925+00:00", "epoch": 0, "step": 17635, "train_loss": 3.4172377586364746, "perplexity": 30.48509143308717, "lr": 0.0026291804804649314, "grad_norm": 0.170602, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:35.481256+00:00", "epoch": 0, "step": 17636, "train_loss": 3.6128640174865723, "perplexity": 37.07207599013565, "lr": 0.0026291804804649314, "grad_norm": 0.160209, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:35.785530+00:00", "epoch": 0, "step": 17637, "train_loss": 3.488548517227173, "perplexity": 32.738393987995735, "lr": 0.0026291804804649314, "grad_norm": 0.192916, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:36.089920+00:00", "epoch": 0, "step": 17638, "train_loss": 3.6358959674835205, "perplexity": 37.935826942450355, "lr": 0.0026291804804649314, "grad_norm": 0.176213, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:36.395617+00:00", "epoch": 0, "step": 17639, "train_loss": 3.61091685295105, "perplexity": 36.99996079137331, "lr": 0.0026291804804649314, "grad_norm": 0.169873, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:36.701362+00:00", "epoch": 0, "step": 17640, "train_loss": 3.5093624591827393, "perplexity": 33.426949944184535, "lr": 0.0026291804804649314, "grad_norm": 0.17103, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:37.006363+00:00", "epoch": 0, "step": 17641, "train_loss": 3.4232327938079834, "perplexity": 30.668399548721112, "lr": 0.0026291804804649314, "grad_norm": 0.190966, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:37.311278+00:00", "epoch": 0, "step": 17642, "train_loss": 3.5521063804626465, "perplexity": 34.88672486418835, "lr": 0.0026291804804649314, "grad_norm": 0.155545, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:37.615026+00:00", "epoch": 0, "step": 17643, "train_loss": 3.5681190490722656, "perplexity": 35.449850971850125, "lr": 0.0026291804804649314, "grad_norm": 0.173832, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:37.919371+00:00", "epoch": 0, "step": 17644, "train_loss": 3.4454376697540283, "perplexity": 31.35700445624234, "lr": 0.0026291804804649314, "grad_norm": 0.181444, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:38.225161+00:00", "epoch": 0, "step": 17645, "train_loss": 3.470088005065918, "perplexity": 32.13957076534184, "lr": 0.0026291804804649314, "grad_norm": 0.217836, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:38.530788+00:00", "epoch": 0, "step": 17646, "train_loss": 3.383246660232544, "perplexity": 29.466282991272454, "lr": 0.0026291804804649314, "grad_norm": 0.160118, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:38.834886+00:00", "epoch": 0, "step": 17647, "train_loss": 3.4479570388793945, "perplexity": 31.43610392367293, "lr": 0.0026291804804649314, "grad_norm": 0.217949, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:39.138813+00:00", "epoch": 0, "step": 17648, "train_loss": 3.4615609645843506, "perplexity": 31.86668047115366, "lr": 0.0026291804804649314, "grad_norm": 0.213085, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:39.443899+00:00", "epoch": 0, "step": 17649, "train_loss": 3.4937353134155273, "perplexity": 32.90864250564598, "lr": 0.0026291804804649314, "grad_norm": 0.18242, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:39.748773+00:00", "epoch": 0, "step": 17650, "train_loss": 3.509148597717285, "perplexity": 33.41980197204896, "lr": 0.0026291804804649314, "grad_norm": 0.202702, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:40.053674+00:00", "epoch": 0, "step": 17651, "train_loss": 3.4414851665496826, "perplexity": 31.23331040724477, "lr": 0.0026291804804649314, "grad_norm": 0.196926, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:40.357866+00:00", "epoch": 0, "step": 17652, "train_loss": 3.6413776874542236, "perplexity": 38.1443515371564, "lr": 0.0026291804804649314, "grad_norm": 0.181136, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:40.662989+00:00", "epoch": 0, "step": 17653, "train_loss": 3.4728918075561523, "perplexity": 32.22981022157815, "lr": 0.0026291804804649314, "grad_norm": 0.174533, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:40.967342+00:00", "epoch": 0, "step": 17654, "train_loss": 3.4910809993743896, "perplexity": 32.821408458263306, "lr": 0.0026291804804649314, "grad_norm": 0.178649, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:41.272156+00:00", "epoch": 0, "step": 17655, "train_loss": 3.480135917663574, "perplexity": 32.464134225286344, "lr": 0.0026291804804649314, "grad_norm": 0.174115, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:41.576275+00:00", "epoch": 0, "step": 17656, "train_loss": 3.492833137512207, "perplexity": 32.878966509863375, "lr": 0.0026291804804649314, "grad_norm": 0.164678, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:41.880966+00:00", "epoch": 0, "step": 17657, "train_loss": 3.4625887870788574, "perplexity": 31.89945060022016, "lr": 0.0026291804804649314, "grad_norm": 0.160503, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:42.186352+00:00", "epoch": 0, "step": 17658, "train_loss": 3.4991326332092285, "perplexity": 33.08674116859488, "lr": 0.0026291804804649314, "grad_norm": 0.197941, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:42.490694+00:00", "epoch": 0, "step": 17659, "train_loss": 3.396251916885376, "perplexity": 29.852002316999936, "lr": 0.0026291804804649314, "grad_norm": 0.174038, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:42.795510+00:00", "epoch": 0, "step": 17660, "train_loss": 3.437389612197876, "perplexity": 31.105654276633, "lr": 0.0026291804804649314, "grad_norm": 0.168627, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:43.100650+00:00", "epoch": 0, "step": 17661, "train_loss": 3.5375189781188965, "perplexity": 34.38151200253227, "lr": 0.0026291804804649314, "grad_norm": 0.190634, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:43.405745+00:00", "epoch": 0, "step": 17662, "train_loss": 3.4800209999084473, "perplexity": 32.46040373421297, "lr": 0.0026291804804649314, "grad_norm": 0.183482, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:43.711188+00:00", "epoch": 0, "step": 17663, "train_loss": 3.5779521465301514, "perplexity": 35.80015226214498, "lr": 0.0026291804804649314, "grad_norm": 0.193431, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:44.016040+00:00", "epoch": 0, "step": 17664, "train_loss": 3.472184419631958, "perplexity": 32.20701930498829, "lr": 0.0026291804804649314, "grad_norm": 0.203788, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:44.321081+00:00", "epoch": 0, "step": 17665, "train_loss": 3.5537922382354736, "perplexity": 34.94558852440724, "lr": 0.0026291804804649314, "grad_norm": 0.212833, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:44.625920+00:00", "epoch": 0, "step": 17666, "train_loss": 3.5067343711853027, "perplexity": 33.33921631459165, "lr": 0.0026291804804649314, "grad_norm": 0.178686, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:44.930797+00:00", "epoch": 0, "step": 17667, "train_loss": 3.4259045124053955, "perplexity": 30.75044643644142, "lr": 0.0026291804804649314, "grad_norm": 0.160933, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:45.235419+00:00", "epoch": 0, "step": 17668, "train_loss": 3.4506704807281494, "perplexity": 31.521519796715374, "lr": 0.0026291804804649314, "grad_norm": 0.159683, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:45.539881+00:00", "epoch": 0, "step": 17669, "train_loss": 3.4041476249694824, "perplexity": 30.08863798674179, "lr": 0.0026291804804649314, "grad_norm": 0.1607, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:45.844125+00:00", "epoch": 0, "step": 17670, "train_loss": 3.5011332035064697, "perplexity": 33.15299977561104, "lr": 0.0026291804804649314, "grad_norm": 0.154571, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:46.148501+00:00", "epoch": 0, "step": 17671, "train_loss": 3.5973012447357178, "perplexity": 36.49959792359004, "lr": 0.0026291804804649314, "grad_norm": 0.177219, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:46.453162+00:00", "epoch": 0, "step": 17672, "train_loss": 3.472780227661133, "perplexity": 32.2262142233614, "lr": 0.0026291804804649314, "grad_norm": 0.147344, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:46.757091+00:00", "epoch": 0, "step": 17673, "train_loss": 3.456399917602539, "perplexity": 31.702638713819358, "lr": 0.0026291804804649314, "grad_norm": 0.15252, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:47.061302+00:00", "epoch": 0, "step": 17674, "train_loss": 3.5261809825897217, "perplexity": 33.99389611906902, "lr": 0.0026291804804649314, "grad_norm": 0.146227, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:47.364994+00:00", "epoch": 0, "step": 17675, "train_loss": 3.545198678970337, "perplexity": 34.646568202994644, "lr": 0.0026291804804649314, "grad_norm": 0.149545, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:47.669851+00:00", "epoch": 0, "step": 17676, "train_loss": 3.4570608139038086, "perplexity": 31.72359779561274, "lr": 0.0026291804804649314, "grad_norm": 0.154826, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:47.974668+00:00", "epoch": 0, "step": 17677, "train_loss": 3.507331371307373, "perplexity": 33.35912577318447, "lr": 0.0026291804804649314, "grad_norm": 0.16311, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:48.279133+00:00", "epoch": 0, "step": 17678, "train_loss": 3.452632427215576, "perplexity": 31.583424038306575, "lr": 0.0026291804804649314, "grad_norm": 0.150128, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:48.583168+00:00", "epoch": 0, "step": 17679, "train_loss": 3.5610387325286865, "perplexity": 35.19974127829797, "lr": 0.0026291804804649314, "grad_norm": 0.152296, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:48.887403+00:00", "epoch": 0, "step": 17680, "train_loss": 3.40631103515625, "perplexity": 30.153802516055507, "lr": 0.0026291804804649314, "grad_norm": 0.147526, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:49.191558+00:00", "epoch": 0, "step": 17681, "train_loss": 3.446892738342285, "perplexity": 31.40266425946515, "lr": 0.0026291804804649314, "grad_norm": 0.164135, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:49.498708+00:00", "epoch": 0, "step": 17682, "train_loss": 3.5342283248901367, "perplexity": 34.268560313395575, "lr": 0.0026291804804649314, "grad_norm": 0.148412, "tokens_per_sec": 106681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:49.803972+00:00", "epoch": 0, "step": 17683, "train_loss": 3.466737985610962, "perplexity": 32.03208272204697, "lr": 0.0026291804804649314, "grad_norm": 0.151449, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:50.109845+00:00", "epoch": 0, "step": 17684, "train_loss": 3.5513412952423096, "perplexity": 34.86004375457161, "lr": 0.0026291804804649314, "grad_norm": 0.165779, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:50.415307+00:00", "epoch": 0, "step": 17685, "train_loss": 3.5971808433532715, "perplexity": 36.49520358608877, "lr": 0.0026291804804649314, "grad_norm": 0.159276, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:50.719756+00:00", "epoch": 0, "step": 17686, "train_loss": 3.5542197227478027, "perplexity": 34.960530415761674, "lr": 0.0026291804804649314, "grad_norm": 0.191557, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:51.024595+00:00", "epoch": 0, "step": 17687, "train_loss": 3.4456794261932373, "perplexity": 31.36458613040629, "lr": 0.0026291804804649314, "grad_norm": 0.176238, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:51.329364+00:00", "epoch": 0, "step": 17688, "train_loss": 3.5053353309631348, "perplexity": 33.292606022431315, "lr": 0.0026291804804649314, "grad_norm": 0.173647, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:51.635363+00:00", "epoch": 0, "step": 17689, "train_loss": 3.5195729732513428, "perplexity": 33.770004689095394, "lr": 0.0026291804804649314, "grad_norm": 0.170664, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:51.940701+00:00", "epoch": 0, "step": 17690, "train_loss": 3.403963088989258, "perplexity": 30.08308606271791, "lr": 0.0026291804804649314, "grad_norm": 0.174959, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:52.245703+00:00", "epoch": 0, "step": 17691, "train_loss": 3.486229419708252, "perplexity": 32.66255842874169, "lr": 0.0026291804804649314, "grad_norm": 0.185877, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:52.550564+00:00", "epoch": 0, "step": 17692, "train_loss": 3.5050764083862305, "perplexity": 33.28398693097496, "lr": 0.0026291804804649314, "grad_norm": 0.195636, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:52.856221+00:00", "epoch": 0, "step": 17693, "train_loss": 3.5632505416870117, "perplexity": 35.27768255226507, "lr": 0.0026291804804649314, "grad_norm": 0.182666, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:53.160825+00:00", "epoch": 0, "step": 17694, "train_loss": 3.4963419437408447, "perplexity": 32.99453506761136, "lr": 0.0026291804804649314, "grad_norm": 0.175973, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:53.465375+00:00", "epoch": 0, "step": 17695, "train_loss": 3.4850950241088867, "perplexity": 32.62552717421335, "lr": 0.0026291804804649314, "grad_norm": 0.173824, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:53.770273+00:00", "epoch": 0, "step": 17696, "train_loss": 3.528794050216675, "perplexity": 34.08284062692849, "lr": 0.0026291804804649314, "grad_norm": 0.186885, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:54.073822+00:00", "epoch": 0, "step": 17697, "train_loss": 3.361361265182495, "perplexity": 28.828407288329124, "lr": 0.0026291804804649314, "grad_norm": 0.169577, "tokens_per_sec": 107948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:54.379096+00:00", "epoch": 0, "step": 17698, "train_loss": 3.519484519958496, "perplexity": 33.767017753085305, "lr": 0.0026291804804649314, "grad_norm": 0.165521, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:54.684095+00:00", "epoch": 0, "step": 17699, "train_loss": 3.5198328495025635, "perplexity": 33.77878185175637, "lr": 0.0026291804804649314, "grad_norm": 0.183876, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:54.989610+00:00", "epoch": 0, "step": 17700, "train_loss": 3.4510231018066406, "perplexity": 31.532636908970638, "lr": 0.0026291804804649314, "grad_norm": 0.148811, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:55.294505+00:00", "epoch": 0, "step": 17701, "train_loss": 3.49350905418396, "perplexity": 32.90119746376721, "lr": 0.0026291804804649314, "grad_norm": 0.155333, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:55.598325+00:00", "epoch": 0, "step": 17702, "train_loss": 3.416841983795166, "perplexity": 30.4730285881057, "lr": 0.0026291804804649314, "grad_norm": 0.171208, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:55.902453+00:00", "epoch": 0, "step": 17703, "train_loss": 3.529588460922241, "perplexity": 34.10992715788922, "lr": 0.0026291804804649314, "grad_norm": 0.14669, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:56.207981+00:00", "epoch": 0, "step": 17704, "train_loss": 3.6112678050994873, "perplexity": 37.01294828596641, "lr": 0.0026291804804649314, "grad_norm": 0.173989, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:56.513818+00:00", "epoch": 0, "step": 17705, "train_loss": 3.5236120223999023, "perplexity": 33.90667922956685, "lr": 0.0026291804804649314, "grad_norm": 0.178994, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:56.818471+00:00", "epoch": 0, "step": 17706, "train_loss": 3.469278573989868, "perplexity": 32.11356652373389, "lr": 0.0026291804804649314, "grad_norm": 0.164372, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:57.123873+00:00", "epoch": 0, "step": 17707, "train_loss": 3.4592223167419434, "perplexity": 31.792242603531122, "lr": 0.0026291804804649314, "grad_norm": 0.167553, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:57.428815+00:00", "epoch": 0, "step": 17708, "train_loss": 3.550147533416748, "perplexity": 34.81845399417317, "lr": 0.0026291804804649314, "grad_norm": 0.18592, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:57.734219+00:00", "epoch": 0, "step": 17709, "train_loss": 3.4598336219787598, "perplexity": 31.81168330942205, "lr": 0.0026291804804649314, "grad_norm": 0.162711, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:58.038709+00:00", "epoch": 0, "step": 17710, "train_loss": 3.511948585510254, "perplexity": 33.513508136451534, "lr": 0.0026291804804649314, "grad_norm": 0.171956, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:58.343231+00:00", "epoch": 0, "step": 17711, "train_loss": 3.5253679752349854, "perplexity": 33.9662700631231, "lr": 0.0026291804804649314, "grad_norm": 0.16544, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:58.647167+00:00", "epoch": 0, "step": 17712, "train_loss": 3.5123350620269775, "perplexity": 33.52646282351922, "lr": 0.0026291804804649314, "grad_norm": 0.152356, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:58.952041+00:00", "epoch": 0, "step": 17713, "train_loss": 3.5284323692321777, "perplexity": 34.070515740548956, "lr": 0.0026291804804649314, "grad_norm": 0.184227, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:59.256676+00:00", "epoch": 0, "step": 17714, "train_loss": 3.572422742843628, "perplexity": 35.6027450433158, "lr": 0.0026291804804649314, "grad_norm": 0.178523, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:59.562051+00:00", "epoch": 0, "step": 17715, "train_loss": 3.4984028339385986, "perplexity": 33.06260329799403, "lr": 0.0026291804804649314, "grad_norm": 0.20419, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:30:59.867373+00:00", "epoch": 0, "step": 17716, "train_loss": 3.556011199951172, "perplexity": 35.02321754353376, "lr": 0.0026291804804649314, "grad_norm": 0.207075, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:00.172206+00:00", "epoch": 0, "step": 17717, "train_loss": 3.5831069946289062, "perplexity": 35.98517307636018, "lr": 0.0026291804804649314, "grad_norm": 0.178141, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:00.476367+00:00", "epoch": 0, "step": 17718, "train_loss": 3.4607386589050293, "perplexity": 31.84048708978021, "lr": 0.0026291804804649314, "grad_norm": 0.179009, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:00.781232+00:00", "epoch": 0, "step": 17719, "train_loss": 3.5395009517669678, "perplexity": 34.449722827001516, "lr": 0.0026291804804649314, "grad_norm": 0.178473, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:01.087480+00:00", "epoch": 0, "step": 17720, "train_loss": 3.4584274291992188, "perplexity": 31.766981387173495, "lr": 0.0026291804804649314, "grad_norm": 0.165114, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:01.392391+00:00", "epoch": 0, "step": 17721, "train_loss": 3.4114041328430176, "perplexity": 30.307770532147753, "lr": 0.0026291804804649314, "grad_norm": 0.15944, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:01.696927+00:00", "epoch": 0, "step": 17722, "train_loss": 3.461869478225708, "perplexity": 31.876513293485708, "lr": 0.0026291804804649314, "grad_norm": 0.162156, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:02.002729+00:00", "epoch": 0, "step": 17723, "train_loss": 3.436356782913208, "perplexity": 31.073544031032764, "lr": 0.0026291804804649314, "grad_norm": 0.212584, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:02.307445+00:00", "epoch": 0, "step": 17724, "train_loss": 3.4656386375427246, "perplexity": 31.996887663139514, "lr": 0.0026291804804649314, "grad_norm": 0.222716, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:02.613213+00:00", "epoch": 0, "step": 17725, "train_loss": 3.6564700603485107, "perplexity": 38.72440650696296, "lr": 0.0026291804804649314, "grad_norm": 0.197615, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:02.919387+00:00", "epoch": 0, "step": 17726, "train_loss": 3.5737686157226562, "perplexity": 35.65069407169919, "lr": 0.0026291804804649314, "grad_norm": 0.177454, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:03.224469+00:00", "epoch": 0, "step": 17727, "train_loss": 3.522460460662842, "perplexity": 33.86765606823738, "lr": 0.0026291804804649314, "grad_norm": 0.201058, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:03.530004+00:00", "epoch": 0, "step": 17728, "train_loss": 3.481698751449585, "perplexity": 32.51490993775192, "lr": 0.0026291804804649314, "grad_norm": 0.193105, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:03.834595+00:00", "epoch": 0, "step": 17729, "train_loss": 3.6189682483673096, "perplexity": 37.299064591793496, "lr": 0.0026291804804649314, "grad_norm": 0.182035, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:04.139851+00:00", "epoch": 0, "step": 17730, "train_loss": 3.4059741497039795, "perplexity": 30.143645849569857, "lr": 0.0026291804804649314, "grad_norm": 0.167856, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:04.445334+00:00", "epoch": 0, "step": 17731, "train_loss": 3.436941385269165, "perplexity": 31.091715008961987, "lr": 0.0026291804804649314, "grad_norm": 0.173793, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:04.749671+00:00", "epoch": 0, "step": 17732, "train_loss": 3.5834341049194336, "perplexity": 35.99694612221711, "lr": 0.0026291804804649314, "grad_norm": 0.181825, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:05.053774+00:00", "epoch": 0, "step": 17733, "train_loss": 3.542107582092285, "perplexity": 34.53963765567003, "lr": 0.0026291804804649314, "grad_norm": 0.17055, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:05.357494+00:00", "epoch": 0, "step": 17734, "train_loss": 3.5586910247802734, "perplexity": 35.117199502852, "lr": 0.0026291804804649314, "grad_norm": 0.201947, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:05.661791+00:00", "epoch": 0, "step": 17735, "train_loss": 3.3591513633728027, "perplexity": 28.764769681222386, "lr": 0.0026291804804649314, "grad_norm": 0.182845, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:05.966713+00:00", "epoch": 0, "step": 17736, "train_loss": 3.5353267192840576, "perplexity": 34.30622138750147, "lr": 0.0026291804804649314, "grad_norm": 0.174731, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:06.271154+00:00", "epoch": 0, "step": 17737, "train_loss": 3.486079216003418, "perplexity": 32.65765275989038, "lr": 0.0026291804804649314, "grad_norm": 0.173256, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:06.575089+00:00", "epoch": 0, "step": 17738, "train_loss": 3.5353407859802246, "perplexity": 34.306703966088506, "lr": 0.0026291804804649314, "grad_norm": 0.185971, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:06.879439+00:00", "epoch": 0, "step": 17739, "train_loss": 3.456129789352417, "perplexity": 31.694076092054427, "lr": 0.0026291804804649314, "grad_norm": 0.192398, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:07.186092+00:00", "epoch": 0, "step": 17740, "train_loss": 3.561685800552368, "perplexity": 35.22252527592421, "lr": 0.0026291804804649314, "grad_norm": 0.166112, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:07.491048+00:00", "epoch": 0, "step": 17741, "train_loss": 3.4801340103149414, "perplexity": 32.46407230492337, "lr": 0.0026291804804649314, "grad_norm": 0.159639, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:07.797080+00:00", "epoch": 0, "step": 17742, "train_loss": 3.584117889404297, "perplexity": 36.02156869278443, "lr": 0.0026291804804649314, "grad_norm": 0.180462, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:08.101142+00:00", "epoch": 0, "step": 17743, "train_loss": 3.4374189376831055, "perplexity": 31.10656647841336, "lr": 0.0026291804804649314, "grad_norm": 0.187176, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:08.405180+00:00", "epoch": 0, "step": 17744, "train_loss": 3.4894633293151855, "perplexity": 32.7683571698089, "lr": 0.0026291804804649314, "grad_norm": 0.153845, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:08.708870+00:00", "epoch": 0, "step": 17745, "train_loss": 3.5646896362304688, "perplexity": 35.32848702018015, "lr": 0.0026291804804649314, "grad_norm": 0.175062, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:09.012985+00:00", "epoch": 0, "step": 17746, "train_loss": 3.5028443336486816, "perplexity": 33.209777435961264, "lr": 0.0026291804804649314, "grad_norm": 0.179419, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:09.318337+00:00", "epoch": 0, "step": 17747, "train_loss": 3.5190253257751465, "perplexity": 33.751515694444, "lr": 0.0026291804804649314, "grad_norm": 0.204379, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:09.623394+00:00", "epoch": 0, "step": 17748, "train_loss": 3.4959373474121094, "perplexity": 32.981188300063316, "lr": 0.0026291804804649314, "grad_norm": 0.184561, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:09.928426+00:00", "epoch": 0, "step": 17749, "train_loss": 3.6111645698547363, "perplexity": 37.00912744241741, "lr": 0.0026291804804649314, "grad_norm": 0.15635, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:10.233207+00:00", "epoch": 0, "step": 17750, "train_loss": 3.4831814765930176, "perplexity": 32.563156371514246, "lr": 0.0026291804804649314, "grad_norm": 0.183484, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:10.537455+00:00", "epoch": 0, "step": 17751, "train_loss": 3.481443166732788, "perplexity": 32.50660068560797, "lr": 0.0026291804804649314, "grad_norm": 0.167844, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:10.843156+00:00", "epoch": 0, "step": 17752, "train_loss": 3.5368752479553223, "perplexity": 34.35938670831062, "lr": 0.0026291804804649314, "grad_norm": 0.172075, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:11.148910+00:00", "epoch": 0, "step": 17753, "train_loss": 3.586986780166626, "perplexity": 36.12505901869004, "lr": 0.0026291804804649314, "grad_norm": 0.160303, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:11.454010+00:00", "epoch": 0, "step": 17754, "train_loss": 3.5419485569000244, "perplexity": 34.53414541986472, "lr": 0.0026291804804649314, "grad_norm": 0.166726, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:11.757675+00:00", "epoch": 0, "step": 17755, "train_loss": 3.590362310409546, "perplexity": 36.247206287854574, "lr": 0.0026291804804649314, "grad_norm": 0.168488, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:12.061683+00:00", "epoch": 0, "step": 17756, "train_loss": 3.5411720275878906, "perplexity": 34.507339052988534, "lr": 0.0026291804804649314, "grad_norm": 0.18671, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:12.366584+00:00", "epoch": 0, "step": 17757, "train_loss": 3.5268681049346924, "perplexity": 34.017262111413444, "lr": 0.0026291804804649314, "grad_norm": 0.177983, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:12.672182+00:00", "epoch": 0, "step": 17758, "train_loss": 3.5582714080810547, "perplexity": 35.102466830764605, "lr": 0.0026291804804649314, "grad_norm": 0.161293, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:12.977147+00:00", "epoch": 0, "step": 17759, "train_loss": 3.5735156536102295, "perplexity": 35.6416769373621, "lr": 0.0026291804804649314, "grad_norm": 0.177949, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:13.281741+00:00", "epoch": 0, "step": 17760, "train_loss": 3.390249252319336, "perplexity": 29.673347499378256, "lr": 0.0026291804804649314, "grad_norm": 0.197091, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:13.585732+00:00", "epoch": 0, "step": 17761, "train_loss": 3.648101329803467, "perplexity": 38.40168465300854, "lr": 0.0026291804804649314, "grad_norm": 0.185227, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:13.889751+00:00", "epoch": 0, "step": 17762, "train_loss": 3.468822479248047, "perplexity": 32.098923034564336, "lr": 0.0026291804804649314, "grad_norm": 0.177156, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:14.195063+00:00", "epoch": 0, "step": 17763, "train_loss": 3.538257122039795, "perplexity": 34.40689987540752, "lr": 0.0026291804804649314, "grad_norm": 0.179593, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:14.499889+00:00", "epoch": 0, "step": 17764, "train_loss": 3.471893548965454, "perplexity": 32.19765259013428, "lr": 0.0026291804804649314, "grad_norm": 0.15455, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:14.803529+00:00", "epoch": 0, "step": 17765, "train_loss": 3.5983026027679443, "perplexity": 36.53616539465373, "lr": 0.0026291804804649314, "grad_norm": 0.160013, "tokens_per_sec": 107918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:15.107595+00:00", "epoch": 0, "step": 17766, "train_loss": 3.515920639038086, "perplexity": 33.646890309874706, "lr": 0.0026291804804649314, "grad_norm": 0.156417, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:15.411361+00:00", "epoch": 0, "step": 17767, "train_loss": 3.4732561111450195, "perplexity": 32.24155379609367, "lr": 0.0026291804804649314, "grad_norm": 0.178785, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:15.716950+00:00", "epoch": 0, "step": 17768, "train_loss": 3.4351019859313965, "perplexity": 31.034577494427484, "lr": 0.0026291804804649314, "grad_norm": 0.207788, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:16.021697+00:00", "epoch": 0, "step": 17769, "train_loss": 3.531677722930908, "perplexity": 34.18126622987533, "lr": 0.0026291804804649314, "grad_norm": 0.190995, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:16.327472+00:00", "epoch": 0, "step": 17770, "train_loss": 3.536961793899536, "perplexity": 34.362360502558964, "lr": 0.0026291804804649314, "grad_norm": 0.1843, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:16.631636+00:00", "epoch": 0, "step": 17771, "train_loss": 3.490837335586548, "perplexity": 32.81341204381395, "lr": 0.0026291804804649314, "grad_norm": 0.183792, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:16.935690+00:00", "epoch": 0, "step": 17772, "train_loss": 3.459672689437866, "perplexity": 31.806564186324756, "lr": 0.0026291804804649314, "grad_norm": 0.168665, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:17.241269+00:00", "epoch": 0, "step": 17773, "train_loss": 3.4791171550750732, "perplexity": 32.431077821042564, "lr": 0.0026291804804649314, "grad_norm": 0.17999, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:17.546324+00:00", "epoch": 0, "step": 17774, "train_loss": 3.4635918140411377, "perplexity": 31.931462661050325, "lr": 0.0026291804804649314, "grad_norm": 0.160637, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:17.852018+00:00", "epoch": 0, "step": 17775, "train_loss": 3.519005060195923, "perplexity": 33.7508317073595, "lr": 0.0026291804804649314, "grad_norm": 0.15366, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:18.157334+00:00", "epoch": 0, "step": 17776, "train_loss": 3.4050681591033936, "perplexity": 30.11634835726567, "lr": 0.0026291804804649314, "grad_norm": 0.156949, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:18.461780+00:00", "epoch": 0, "step": 17777, "train_loss": 3.5114550590515137, "perplexity": 33.49697241420484, "lr": 0.0026291804804649314, "grad_norm": 0.16449, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:18.766007+00:00", "epoch": 0, "step": 17778, "train_loss": 3.5912563800811768, "perplexity": 36.27962830729029, "lr": 0.0026291804804649314, "grad_norm": 0.18099, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:19.072043+00:00", "epoch": 0, "step": 17779, "train_loss": 3.5548155307769775, "perplexity": 34.9813663869907, "lr": 0.0026291804804649314, "grad_norm": 0.176368, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:19.376916+00:00", "epoch": 0, "step": 17780, "train_loss": 3.563920736312866, "perplexity": 35.301333389971106, "lr": 0.0026291804804649314, "grad_norm": 0.168097, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:19.681970+00:00", "epoch": 0, "step": 17781, "train_loss": 3.569819450378418, "perplexity": 35.51018122302813, "lr": 0.0026291804804649314, "grad_norm": 0.198845, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:19.986687+00:00", "epoch": 0, "step": 17782, "train_loss": 3.388727903366089, "perplexity": 29.628238305333756, "lr": 0.0026291804804649314, "grad_norm": 0.16721, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:20.291222+00:00", "epoch": 0, "step": 17783, "train_loss": 3.4800233840942383, "perplexity": 32.460481125938585, "lr": 0.0026291804804649314, "grad_norm": 0.146949, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:20.597750+00:00", "epoch": 0, "step": 17784, "train_loss": 3.560361385345459, "perplexity": 35.17590690567662, "lr": 0.0026291804804649314, "grad_norm": 0.185232, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:20.903631+00:00", "epoch": 0, "step": 17785, "train_loss": 3.4555118083953857, "perplexity": 31.674495807323527, "lr": 0.0026291804804649314, "grad_norm": 0.208927, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:21.208777+00:00", "epoch": 0, "step": 17786, "train_loss": 3.517596960067749, "perplexity": 33.70334060086986, "lr": 0.0026291804804649314, "grad_norm": 0.192135, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:21.513542+00:00", "epoch": 0, "step": 17787, "train_loss": 3.6370654106140137, "perplexity": 37.98021668524873, "lr": 0.0026291804804649314, "grad_norm": 0.148988, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:21.819180+00:00", "epoch": 0, "step": 17788, "train_loss": 3.545807123184204, "perplexity": 34.66765512140432, "lr": 0.0026291804804649314, "grad_norm": 0.164468, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:22.124961+00:00", "epoch": 0, "step": 17789, "train_loss": 3.486340045928955, "perplexity": 32.66617196401135, "lr": 0.0026291804804649314, "grad_norm": 0.167995, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:22.430857+00:00", "epoch": 0, "step": 17790, "train_loss": 3.4584531784057617, "perplexity": 31.767799372269668, "lr": 0.0026291804804649314, "grad_norm": 0.172206, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:22.736521+00:00", "epoch": 0, "step": 17791, "train_loss": 3.4637534618377686, "perplexity": 31.936624728839888, "lr": 0.0026291804804649314, "grad_norm": 0.189154, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:23.040959+00:00", "epoch": 0, "step": 17792, "train_loss": 3.5356333255767822, "perplexity": 34.31674150354296, "lr": 0.0026291804804649314, "grad_norm": 0.150226, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:23.346356+00:00", "epoch": 0, "step": 17793, "train_loss": 3.54964542388916, "perplexity": 34.80097570506177, "lr": 0.0026291804804649314, "grad_norm": 0.196237, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:23.651584+00:00", "epoch": 0, "step": 17794, "train_loss": 3.566953420639038, "perplexity": 35.40855369092194, "lr": 0.0026291804804649314, "grad_norm": 0.16595, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:23.956810+00:00", "epoch": 0, "step": 17795, "train_loss": 3.5814316272735596, "perplexity": 35.92493516651532, "lr": 0.0026291804804649314, "grad_norm": 0.21028, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:24.262524+00:00", "epoch": 0, "step": 17796, "train_loss": 3.5173349380493164, "perplexity": 33.69451074039715, "lr": 0.0026291804804649314, "grad_norm": 0.221629, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:24.566701+00:00", "epoch": 0, "step": 17797, "train_loss": 3.52957820892334, "perplexity": 34.109577464745996, "lr": 0.0026291804804649314, "grad_norm": 0.178574, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:24.871163+00:00", "epoch": 0, "step": 17798, "train_loss": 3.581677198410034, "perplexity": 35.93375837699051, "lr": 0.0026291804804649314, "grad_norm": 0.200984, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:25.177092+00:00", "epoch": 0, "step": 17799, "train_loss": 3.5023300647735596, "perplexity": 33.19270307185953, "lr": 0.0026291804804649314, "grad_norm": 0.16874, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:25.483451+00:00", "epoch": 0, "step": 17800, "train_loss": 3.5092382431030273, "perplexity": 33.42279803737788, "lr": 0.0026291804804649314, "grad_norm": 0.192548, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:25.787964+00:00", "epoch": 0, "step": 17801, "train_loss": 3.5296599864959717, "perplexity": 34.11236697725279, "lr": 0.0026291804804649314, "grad_norm": 0.158625, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:26.091960+00:00", "epoch": 0, "step": 17802, "train_loss": 3.4457478523254395, "perplexity": 31.366732361151612, "lr": 0.0026291804804649314, "grad_norm": 0.177145, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:26.396375+00:00", "epoch": 0, "step": 17803, "train_loss": 3.437136650085449, "perplexity": 31.097786719757625, "lr": 0.0026291804804649314, "grad_norm": 0.159408, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:26.701620+00:00", "epoch": 0, "step": 17804, "train_loss": 3.557460308074951, "perplexity": 35.074006763244675, "lr": 0.0026291804804649314, "grad_norm": 0.159949, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:27.005757+00:00", "epoch": 0, "step": 17805, "train_loss": 3.4887032508850098, "perplexity": 32.743460111389, "lr": 0.0026291804804649314, "grad_norm": 0.174392, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:27.310250+00:00", "epoch": 0, "step": 17806, "train_loss": 3.463007926940918, "perplexity": 31.912823733955474, "lr": 0.0026291804804649314, "grad_norm": 0.178819, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:27.615029+00:00", "epoch": 0, "step": 17807, "train_loss": 3.5187013149261475, "perplexity": 33.74058160866571, "lr": 0.0026291804804649314, "grad_norm": 0.172432, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:27.919946+00:00", "epoch": 0, "step": 17808, "train_loss": 3.49206280708313, "perplexity": 32.85364859431658, "lr": 0.0026291804804649314, "grad_norm": 0.165555, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:28.224990+00:00", "epoch": 0, "step": 17809, "train_loss": 3.5231897830963135, "perplexity": 33.892365519060185, "lr": 0.0026291804804649314, "grad_norm": 0.178554, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:28.529715+00:00", "epoch": 0, "step": 17810, "train_loss": 3.553102493286133, "perplexity": 34.921493291958505, "lr": 0.0026291804804649314, "grad_norm": 0.183374, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:28.833759+00:00", "epoch": 0, "step": 17811, "train_loss": 3.3673384189605713, "perplexity": 29.001235106549593, "lr": 0.0026291804804649314, "grad_norm": 0.176936, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:29.138256+00:00", "epoch": 0, "step": 17812, "train_loss": 3.495914936065674, "perplexity": 32.98044915550911, "lr": 0.0026291804804649314, "grad_norm": 0.208039, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:29.443953+00:00", "epoch": 0, "step": 17813, "train_loss": 3.542956590652466, "perplexity": 34.5689745555826, "lr": 0.0026291804804649314, "grad_norm": 0.224476, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:29.748688+00:00", "epoch": 0, "step": 17814, "train_loss": 3.551419496536255, "perplexity": 34.862769961695285, "lr": 0.0026291804804649314, "grad_norm": 0.195356, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:30.053416+00:00", "epoch": 0, "step": 17815, "train_loss": 3.474677085876465, "perplexity": 32.287400795339764, "lr": 0.0026291804804649314, "grad_norm": 0.163956, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:30.357637+00:00", "epoch": 0, "step": 17816, "train_loss": 3.475517749786377, "perplexity": 32.31455506013929, "lr": 0.0026291804804649314, "grad_norm": 0.204463, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:30.663375+00:00", "epoch": 0, "step": 17817, "train_loss": 3.525101900100708, "perplexity": 33.957233685486166, "lr": 0.0026291804804649314, "grad_norm": 0.205789, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:30.969603+00:00", "epoch": 0, "step": 17818, "train_loss": 3.548647403717041, "perplexity": 34.76626095519046, "lr": 0.0026291804804649314, "grad_norm": 0.176104, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:31.274695+00:00", "epoch": 0, "step": 17819, "train_loss": 3.592789888381958, "perplexity": 36.33530609871523, "lr": 0.0026291804804649314, "grad_norm": 0.184152, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:31.579586+00:00", "epoch": 0, "step": 17820, "train_loss": 3.5439839363098145, "perplexity": 34.60450709044376, "lr": 0.0026291804804649314, "grad_norm": 0.168222, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:31.883456+00:00", "epoch": 0, "step": 17821, "train_loss": 3.3777434825897217, "perplexity": 29.30457017679244, "lr": 0.0026291804804649314, "grad_norm": 0.157338, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:32.188185+00:00", "epoch": 0, "step": 17822, "train_loss": 3.412008762359619, "perplexity": 30.326101045820256, "lr": 0.0026291804804649314, "grad_norm": 0.180201, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:32.494418+00:00", "epoch": 0, "step": 17823, "train_loss": 3.646040678024292, "perplexity": 38.3226336294939, "lr": 0.0026291804804649314, "grad_norm": 0.161596, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:32.799526+00:00", "epoch": 0, "step": 17824, "train_loss": 3.4954123497009277, "perplexity": 32.96387779607881, "lr": 0.0026291804804649314, "grad_norm": 0.160974, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:33.104851+00:00", "epoch": 0, "step": 17825, "train_loss": 3.6266443729400635, "perplexity": 37.5864785594869, "lr": 0.0026291804804649314, "grad_norm": 0.191143, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:33.409186+00:00", "epoch": 0, "step": 17826, "train_loss": 3.4564120769500732, "perplexity": 31.703024199564855, "lr": 0.0026291804804649314, "grad_norm": 0.160578, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:33.715346+00:00", "epoch": 0, "step": 17827, "train_loss": 3.4507596492767334, "perplexity": 31.52433065020281, "lr": 0.0026291804804649314, "grad_norm": 0.182481, "tokens_per_sec": 107029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:34.019794+00:00", "epoch": 0, "step": 17828, "train_loss": 3.4606728553771973, "perplexity": 31.838391942336617, "lr": 0.0026291804804649314, "grad_norm": 0.166881, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:34.325401+00:00", "epoch": 0, "step": 17829, "train_loss": 3.579596519470215, "perplexity": 35.859069491447954, "lr": 0.0026291804804649314, "grad_norm": 0.165437, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:34.630755+00:00", "epoch": 0, "step": 17830, "train_loss": 3.564551591873169, "perplexity": 35.323610458493725, "lr": 0.0026291804804649314, "grad_norm": 0.16702, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:34.935331+00:00", "epoch": 0, "step": 17831, "train_loss": 3.619381904602051, "perplexity": 37.31449677400105, "lr": 0.0026291804804649314, "grad_norm": 0.179148, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:35.240738+00:00", "epoch": 0, "step": 17832, "train_loss": 3.5222606658935547, "perplexity": 33.86089016362561, "lr": 0.0026291804804649314, "grad_norm": 0.182291, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:35.546530+00:00", "epoch": 0, "step": 17833, "train_loss": 3.549877405166626, "perplexity": 34.809049816348, "lr": 0.0026291804804649314, "grad_norm": 0.172091, "tokens_per_sec": 107157} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:35.851052+00:00", "epoch": 0, "step": 17834, "train_loss": 3.47112774848938, "perplexity": 32.17300505120536, "lr": 0.0026291804804649314, "grad_norm": 0.167615, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:36.155647+00:00", "epoch": 0, "step": 17835, "train_loss": 3.5684940814971924, "perplexity": 35.463148308733786, "lr": 0.0026291804804649314, "grad_norm": 0.155302, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:36.461555+00:00", "epoch": 0, "step": 17836, "train_loss": 3.6682653427124023, "perplexity": 39.18387627681683, "lr": 0.0026291804804649314, "grad_norm": 0.174282, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:36.766678+00:00", "epoch": 0, "step": 17837, "train_loss": 3.504960298538208, "perplexity": 33.28012255666113, "lr": 0.0026291804804649314, "grad_norm": 0.14858, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:37.071586+00:00", "epoch": 0, "step": 17838, "train_loss": 3.5217952728271484, "perplexity": 33.84513520652938, "lr": 0.0026291804804649314, "grad_norm": 0.145512, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:37.378410+00:00", "epoch": 0, "step": 17839, "train_loss": 3.5693986415863037, "perplexity": 35.49524137019108, "lr": 0.0026291804804649314, "grad_norm": 0.190943, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:37.684139+00:00", "epoch": 0, "step": 17840, "train_loss": 3.4170663356781006, "perplexity": 30.47986603641668, "lr": 0.0026291804804649314, "grad_norm": 0.162122, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:37.989414+00:00", "epoch": 0, "step": 17841, "train_loss": 3.5447988510131836, "perplexity": 34.63271830538031, "lr": 0.0026291804804649314, "grad_norm": 0.173096, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:38.294456+00:00", "epoch": 0, "step": 17842, "train_loss": 3.4280624389648438, "perplexity": 30.816875290037274, "lr": 0.0026291804804649314, "grad_norm": 0.165896, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:38.599250+00:00", "epoch": 0, "step": 17843, "train_loss": 3.5052788257598877, "perplexity": 33.29072487010934, "lr": 0.0026291804804649314, "grad_norm": 0.148774, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:38.904760+00:00", "epoch": 0, "step": 17844, "train_loss": 3.604681968688965, "perplexity": 36.76998799002955, "lr": 0.0026291804804649314, "grad_norm": 0.167662, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:39.209771+00:00", "epoch": 0, "step": 17845, "train_loss": 3.440136432647705, "perplexity": 31.191213377854385, "lr": 0.0026291804804649314, "grad_norm": 0.161336, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:39.514267+00:00", "epoch": 0, "step": 17846, "train_loss": 3.612758159637451, "perplexity": 37.06815182761425, "lr": 0.0026291804804649314, "grad_norm": 0.159109, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:39.818959+00:00", "epoch": 0, "step": 17847, "train_loss": 3.386536121368408, "perplexity": 29.563370779698726, "lr": 0.0026291804804649314, "grad_norm": 0.160595, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:40.122819+00:00", "epoch": 0, "step": 17848, "train_loss": 3.574357509613037, "perplexity": 35.6716947305988, "lr": 0.0026291804804649314, "grad_norm": 0.174743, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:40.427587+00:00", "epoch": 0, "step": 17849, "train_loss": 3.509434700012207, "perplexity": 33.42936482200042, "lr": 0.0026291804804649314, "grad_norm": 0.174405, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:40.732687+00:00", "epoch": 0, "step": 17850, "train_loss": 3.431631088256836, "perplexity": 30.927046374281968, "lr": 0.0026291804804649314, "grad_norm": 0.158567, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:41.038098+00:00", "epoch": 0, "step": 17851, "train_loss": 3.483527898788452, "perplexity": 32.57443892578552, "lr": 0.0026291804804649314, "grad_norm": 0.1629, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:41.342845+00:00", "epoch": 0, "step": 17852, "train_loss": 3.4854469299316406, "perplexity": 32.637010307568616, "lr": 0.0026291804804649314, "grad_norm": 0.171099, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:41.647329+00:00", "epoch": 0, "step": 17853, "train_loss": 3.683934211730957, "perplexity": 39.80267860786788, "lr": 0.0026291804804649314, "grad_norm": 0.175091, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:41.951474+00:00", "epoch": 0, "step": 17854, "train_loss": 3.4490115642547607, "perplexity": 31.469271577956032, "lr": 0.0026291804804649314, "grad_norm": 0.173607, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:42.256560+00:00", "epoch": 0, "step": 17855, "train_loss": 3.5200517177581787, "perplexity": 33.78617576393388, "lr": 0.0026291804804649314, "grad_norm": 0.164092, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:42.561611+00:00", "epoch": 0, "step": 17856, "train_loss": 3.539402723312378, "perplexity": 34.44633905016148, "lr": 0.0026291804804649314, "grad_norm": 0.192196, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:42.865813+00:00", "epoch": 0, "step": 17857, "train_loss": 3.502091884613037, "perplexity": 33.184798169946475, "lr": 0.0026291804804649314, "grad_norm": 0.193657, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:43.169704+00:00", "epoch": 0, "step": 17858, "train_loss": 3.53059458732605, "perplexity": 34.14426332660121, "lr": 0.0026291804804649314, "grad_norm": 0.161776, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:43.473756+00:00", "epoch": 0, "step": 17859, "train_loss": 3.382575750350952, "perplexity": 29.44652040103966, "lr": 0.0026291804804649314, "grad_norm": 0.179731, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:43.779080+00:00", "epoch": 0, "step": 17860, "train_loss": 3.5648272037506104, "perplexity": 35.33334740683787, "lr": 0.0026291804804649314, "grad_norm": 0.168155, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:44.084069+00:00", "epoch": 0, "step": 17861, "train_loss": 3.5871400833129883, "perplexity": 36.13059752842476, "lr": 0.0026291804804649314, "grad_norm": 0.190776, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:44.388026+00:00", "epoch": 0, "step": 17862, "train_loss": 3.495070695877075, "perplexity": 32.9526174848545, "lr": 0.0026291804804649314, "grad_norm": 0.189183, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:44.692197+00:00", "epoch": 0, "step": 17863, "train_loss": 3.4794986248016357, "perplexity": 32.44345165540264, "lr": 0.0026291804804649314, "grad_norm": 0.159826, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:44.996512+00:00", "epoch": 0, "step": 17864, "train_loss": 3.500089406967163, "perplexity": 33.11841284317806, "lr": 0.0026291804804649314, "grad_norm": 0.179993, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:45.300573+00:00", "epoch": 0, "step": 17865, "train_loss": 3.6048102378845215, "perplexity": 36.77470474931065, "lr": 0.0026291804804649314, "grad_norm": 0.194327, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:45.605848+00:00", "epoch": 0, "step": 17866, "train_loss": 3.5637269020080566, "perplexity": 35.294491443677984, "lr": 0.0026291804804649314, "grad_norm": 0.173634, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:45.910528+00:00", "epoch": 0, "step": 17867, "train_loss": 3.520434856414795, "perplexity": 33.79912303406974, "lr": 0.0026291804804649314, "grad_norm": 0.183695, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:46.213978+00:00", "epoch": 0, "step": 17868, "train_loss": 3.6069624423980713, "perplexity": 36.85393666591743, "lr": 0.0026291804804649314, "grad_norm": 0.183016, "tokens_per_sec": 107979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:46.518296+00:00", "epoch": 0, "step": 17869, "train_loss": 3.470977783203125, "perplexity": 32.16818057905389, "lr": 0.0026291804804649314, "grad_norm": 0.163331, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:46.823227+00:00", "epoch": 0, "step": 17870, "train_loss": 3.571141004562378, "perplexity": 35.55714087462401, "lr": 0.0026291804804649314, "grad_norm": 0.191297, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:47.129386+00:00", "epoch": 0, "step": 17871, "train_loss": 3.494912624359131, "perplexity": 32.94740902625376, "lr": 0.0026291804804649314, "grad_norm": 0.164618, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:47.434813+00:00", "epoch": 0, "step": 17872, "train_loss": 3.5540361404418945, "perplexity": 34.954112870064115, "lr": 0.0026291804804649314, "grad_norm": 0.176037, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:47.738878+00:00", "epoch": 0, "step": 17873, "train_loss": 3.521502733230591, "perplexity": 33.83523561241286, "lr": 0.0026291804804649314, "grad_norm": 0.16627, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:48.043328+00:00", "epoch": 0, "step": 17874, "train_loss": 3.6522343158721924, "perplexity": 38.56072671361477, "lr": 0.0026291804804649314, "grad_norm": 0.18849, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:48.348377+00:00", "epoch": 0, "step": 17875, "train_loss": 3.5620458126068115, "perplexity": 35.23520809245792, "lr": 0.0026291804804649314, "grad_norm": 0.177071, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:48.653763+00:00", "epoch": 0, "step": 17876, "train_loss": 3.562032461166382, "perplexity": 35.23473765481656, "lr": 0.0026291804804649314, "grad_norm": 0.165511, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:48.958827+00:00", "epoch": 0, "step": 17877, "train_loss": 3.4822421073913574, "perplexity": 32.53258190791301, "lr": 0.0026291804804649314, "grad_norm": 0.167078, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:49.264131+00:00", "epoch": 0, "step": 17878, "train_loss": 3.590782642364502, "perplexity": 36.262445349443155, "lr": 0.0026291804804649314, "grad_norm": 0.153109, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:49.568224+00:00", "epoch": 0, "step": 17879, "train_loss": 3.5667829513549805, "perplexity": 35.40251813457783, "lr": 0.0026291804804649314, "grad_norm": 0.145717, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:49.873791+00:00", "epoch": 0, "step": 17880, "train_loss": 3.450406074523926, "perplexity": 31.513186413062282, "lr": 0.0026291804804649314, "grad_norm": 0.155097, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:50.179708+00:00", "epoch": 0, "step": 17881, "train_loss": 3.4414327144622803, "perplexity": 31.231672197881547, "lr": 0.0026291804804649314, "grad_norm": 0.152361, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:50.485461+00:00", "epoch": 0, "step": 17882, "train_loss": 3.5068161487579346, "perplexity": 33.341942826257544, "lr": 0.0026291804804649314, "grad_norm": 0.172455, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:50.790578+00:00", "epoch": 0, "step": 17883, "train_loss": 3.5624613761901855, "perplexity": 35.24985360465375, "lr": 0.0026291804804649314, "grad_norm": 0.17329, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:51.095918+00:00", "epoch": 0, "step": 17884, "train_loss": 3.5385329723358154, "perplexity": 34.41639233811243, "lr": 0.0026291804804649314, "grad_norm": 0.145928, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:51.400545+00:00", "epoch": 0, "step": 17885, "train_loss": 3.474382162094116, "perplexity": 32.27787987701663, "lr": 0.0026291804804649314, "grad_norm": 0.175818, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:51.705121+00:00", "epoch": 0, "step": 17886, "train_loss": 3.510967969894409, "perplexity": 33.48066037517724, "lr": 0.0026291804804649314, "grad_norm": 0.188904, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:52.010972+00:00", "epoch": 0, "step": 17887, "train_loss": 3.504526138305664, "perplexity": 33.2656767870291, "lr": 0.0026291804804649314, "grad_norm": 0.158508, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:52.316457+00:00", "epoch": 0, "step": 17888, "train_loss": 3.5158655643463135, "perplexity": 33.64503726879009, "lr": 0.0026291804804649314, "grad_norm": 0.200984, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:52.621275+00:00", "epoch": 0, "step": 17889, "train_loss": 3.5532822608947754, "perplexity": 34.92777160959999, "lr": 0.0026291804804649314, "grad_norm": 0.201985, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:52.926589+00:00", "epoch": 0, "step": 17890, "train_loss": 3.4967989921569824, "perplexity": 33.0096186142982, "lr": 0.0026291804804649314, "grad_norm": 0.193027, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:53.231059+00:00", "epoch": 0, "step": 17891, "train_loss": 3.3937437534332275, "perplexity": 29.77722243510098, "lr": 0.0026291804804649314, "grad_norm": 0.170476, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:53.536192+00:00", "epoch": 0, "step": 17892, "train_loss": 3.510812759399414, "perplexity": 33.47546422856632, "lr": 0.0026291804804649314, "grad_norm": 0.179964, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:53.842214+00:00", "epoch": 0, "step": 17893, "train_loss": 3.614992618560791, "perplexity": 37.15107169625715, "lr": 0.0026291804804649314, "grad_norm": 0.186109, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:54.146942+00:00", "epoch": 0, "step": 17894, "train_loss": 3.545816421508789, "perplexity": 34.66797747401291, "lr": 0.0026291804804649314, "grad_norm": 0.167738, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:54.450946+00:00", "epoch": 0, "step": 17895, "train_loss": 3.552927255630493, "perplexity": 34.91537426749999, "lr": 0.0026291804804649314, "grad_norm": 0.201854, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:54.755800+00:00", "epoch": 0, "step": 17896, "train_loss": 3.581411123275757, "perplexity": 35.92419856927523, "lr": 0.0026291804804649314, "grad_norm": 0.180717, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:55.059653+00:00", "epoch": 0, "step": 17897, "train_loss": 3.486682415008545, "perplexity": 32.677357765967486, "lr": 0.0026291804804649314, "grad_norm": 0.163396, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:55.365190+00:00", "epoch": 0, "step": 17898, "train_loss": 3.5417275428771973, "perplexity": 34.52651373284657, "lr": 0.0026291804804649314, "grad_norm": 0.174912, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:55.669804+00:00", "epoch": 0, "step": 17899, "train_loss": 3.4961583614349365, "perplexity": 32.9884784107458, "lr": 0.0026291804804649314, "grad_norm": 0.202675, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:55.974644+00:00", "epoch": 0, "step": 17900, "train_loss": 3.418513298034668, "perplexity": 30.524001178448675, "lr": 0.0026291804804649314, "grad_norm": 0.174906, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:56.279214+00:00", "epoch": 0, "step": 17901, "train_loss": 3.4761383533477783, "perplexity": 32.33461581233798, "lr": 0.0026291804804649314, "grad_norm": 0.17695, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:56.582837+00:00", "epoch": 0, "step": 17902, "train_loss": 3.5013744831085205, "perplexity": 33.160999883296284, "lr": 0.0026291804804649314, "grad_norm": 0.161551, "tokens_per_sec": 107923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:56.888080+00:00", "epoch": 0, "step": 17903, "train_loss": 3.527493715286255, "perplexity": 34.03855032109, "lr": 0.0026291804804649314, "grad_norm": 0.185272, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:57.193718+00:00", "epoch": 0, "step": 17904, "train_loss": 3.448113441467285, "perplexity": 31.441020996191202, "lr": 0.0026291804804649314, "grad_norm": 0.173531, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:57.498150+00:00", "epoch": 0, "step": 17905, "train_loss": 3.4752228260040283, "perplexity": 32.305026134558496, "lr": 0.0026291804804649314, "grad_norm": 0.162577, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:57.802302+00:00", "epoch": 0, "step": 17906, "train_loss": 3.449915647506714, "perplexity": 31.497735284182095, "lr": 0.0026291804804649314, "grad_norm": 0.193249, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:58.106776+00:00", "epoch": 0, "step": 17907, "train_loss": 3.450207233428955, "perplexity": 31.506920919509337, "lr": 0.0026291804804649314, "grad_norm": 0.184245, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:58.411879+00:00", "epoch": 0, "step": 17908, "train_loss": 3.5205166339874268, "perplexity": 33.80188715732866, "lr": 0.0026291804804649314, "grad_norm": 0.159961, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:58.717815+00:00", "epoch": 0, "step": 17909, "train_loss": 3.436206817626953, "perplexity": 31.068884427505317, "lr": 0.0026291804804649314, "grad_norm": 0.156765, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:59.022559+00:00", "epoch": 0, "step": 17910, "train_loss": 3.5104568004608154, "perplexity": 33.46355045838271, "lr": 0.0026291804804649314, "grad_norm": 0.193313, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:59.326384+00:00", "epoch": 0, "step": 17911, "train_loss": 3.5372495651245117, "perplexity": 34.372250424083, "lr": 0.0026291804804649314, "grad_norm": 0.162343, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:59.631221+00:00", "epoch": 0, "step": 17912, "train_loss": 3.38552188873291, "perplexity": 29.533401844543675, "lr": 0.0026291804804649314, "grad_norm": 0.177412, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:31:59.936174+00:00", "epoch": 0, "step": 17913, "train_loss": 3.4357872009277344, "perplexity": 31.055850139667115, "lr": 0.0026291804804649314, "grad_norm": 0.184894, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:00.242572+00:00", "epoch": 0, "step": 17914, "train_loss": 3.4825878143310547, "perplexity": 32.543830591506804, "lr": 0.0026291804804649314, "grad_norm": 0.173944, "tokens_per_sec": 106945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:00.547017+00:00", "epoch": 0, "step": 17915, "train_loss": 3.507146120071411, "perplexity": 33.352946526278615, "lr": 0.0026291804804649314, "grad_norm": 0.196847, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:00.850426+00:00", "epoch": 0, "step": 17916, "train_loss": 3.555596351623535, "perplexity": 35.00869123362338, "lr": 0.0026291804804649314, "grad_norm": 0.1922, "tokens_per_sec": 108002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:01.156874+00:00", "epoch": 0, "step": 17917, "train_loss": 3.514078140258789, "perplexity": 33.58495303275244, "lr": 0.0026291804804649314, "grad_norm": 0.19322, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:01.461893+00:00", "epoch": 0, "step": 17918, "train_loss": 3.449302911758423, "perplexity": 31.47844140741093, "lr": 0.0026291804804649314, "grad_norm": 0.195424, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:01.767186+00:00", "epoch": 0, "step": 17919, "train_loss": 3.4266819953918457, "perplexity": 30.77436368179238, "lr": 0.0026291804804649314, "grad_norm": 0.19078, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:02.072356+00:00", "epoch": 0, "step": 17920, "train_loss": 3.440216541290283, "perplexity": 31.193712163704284, "lr": 0.0026291804804649314, "grad_norm": 0.162637, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:02.376837+00:00", "epoch": 0, "step": 17921, "train_loss": 3.4327707290649414, "perplexity": 30.962312189768618, "lr": 0.0026291804804649314, "grad_norm": 0.222568, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:02.681394+00:00", "epoch": 0, "step": 17922, "train_loss": 3.4356887340545654, "perplexity": 31.05279231775981, "lr": 0.0026291804804649314, "grad_norm": 0.187501, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:02.985460+00:00", "epoch": 0, "step": 17923, "train_loss": 3.4891858100891113, "perplexity": 32.75926458242968, "lr": 0.0026291804804649314, "grad_norm": 0.157877, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:03.290673+00:00", "epoch": 0, "step": 17924, "train_loss": 3.4896602630615234, "perplexity": 32.77481100061537, "lr": 0.0026291804804649314, "grad_norm": 0.182817, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:03.595791+00:00", "epoch": 0, "step": 17925, "train_loss": 3.5384535789489746, "perplexity": 34.41366001262759, "lr": 0.0026291804804649314, "grad_norm": 0.178433, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:03.899897+00:00", "epoch": 0, "step": 17926, "train_loss": 3.4151675701141357, "perplexity": 30.422046826326348, "lr": 0.0026291804804649314, "grad_norm": 0.152725, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:04.204301+00:00", "epoch": 0, "step": 17927, "train_loss": 3.5516364574432373, "perplexity": 34.8703346404762, "lr": 0.0026291804804649314, "grad_norm": 0.174283, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:04.509303+00:00", "epoch": 0, "step": 17928, "train_loss": 3.4520788192749023, "perplexity": 31.56594404294386, "lr": 0.0026291804804649314, "grad_norm": 0.152706, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:04.814589+00:00", "epoch": 0, "step": 17929, "train_loss": 3.38307523727417, "perplexity": 29.461232226790667, "lr": 0.0026291804804649314, "grad_norm": 0.169582, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:05.119908+00:00", "epoch": 0, "step": 17930, "train_loss": 3.573526620864868, "perplexity": 35.642067830852326, "lr": 0.0026291804804649314, "grad_norm": 0.192271, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:05.424780+00:00", "epoch": 0, "step": 17931, "train_loss": 3.545147180557251, "perplexity": 34.644784005655374, "lr": 0.0026291804804649314, "grad_norm": 0.193133, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:05.729110+00:00", "epoch": 0, "step": 17932, "train_loss": 3.49946665763855, "perplexity": 33.09779479441957, "lr": 0.0026291804804649314, "grad_norm": 0.184043, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:06.033302+00:00", "epoch": 0, "step": 17933, "train_loss": 3.3688018321990967, "perplexity": 29.043706967300015, "lr": 0.0026291804804649314, "grad_norm": 0.194166, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:06.338230+00:00", "epoch": 0, "step": 17934, "train_loss": 3.6042850017547607, "perplexity": 36.75539441740096, "lr": 0.0026291804804649314, "grad_norm": 0.189505, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:06.643547+00:00", "epoch": 0, "step": 17935, "train_loss": 3.450840711593628, "perplexity": 31.526886189061415, "lr": 0.0026291804804649314, "grad_norm": 0.182615, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:06.950337+00:00", "epoch": 0, "step": 17936, "train_loss": 3.4343106746673584, "perplexity": 31.01002919763432, "lr": 0.0026291804804649314, "grad_norm": 0.206719, "tokens_per_sec": 106810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:07.255088+00:00", "epoch": 0, "step": 17937, "train_loss": 3.447073459625244, "perplexity": 31.408339902077714, "lr": 0.0026291804804649314, "grad_norm": 0.171804, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:07.558909+00:00", "epoch": 0, "step": 17938, "train_loss": 3.376875877380371, "perplexity": 29.279156405203747, "lr": 0.0026291804804649314, "grad_norm": 0.184001, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:07.863151+00:00", "epoch": 0, "step": 17939, "train_loss": 3.481910467147827, "perplexity": 32.52179458338074, "lr": 0.0026291804804649314, "grad_norm": 0.177528, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:08.168783+00:00", "epoch": 0, "step": 17940, "train_loss": 3.602851390838623, "perplexity": 36.70273923527742, "lr": 0.0026291804804649314, "grad_norm": 0.166748, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:08.474198+00:00", "epoch": 0, "step": 17941, "train_loss": 3.5218145847320557, "perplexity": 33.84578882687337, "lr": 0.0026291804804649314, "grad_norm": 0.152393, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:08.779007+00:00", "epoch": 0, "step": 17942, "train_loss": 3.5373826026916504, "perplexity": 34.376823528847126, "lr": 0.0026291804804649314, "grad_norm": 0.153626, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:09.083652+00:00", "epoch": 0, "step": 17943, "train_loss": 3.4700045585632324, "perplexity": 32.13688894245958, "lr": 0.0026291804804649314, "grad_norm": 0.149623, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:09.388805+00:00", "epoch": 0, "step": 17944, "train_loss": 3.542968273162842, "perplexity": 34.56937841034555, "lr": 0.0026291804804649314, "grad_norm": 0.160134, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:09.693809+00:00", "epoch": 0, "step": 17945, "train_loss": 3.49222731590271, "perplexity": 32.859053753851754, "lr": 0.0026291804804649314, "grad_norm": 0.15515, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:09.998429+00:00", "epoch": 0, "step": 17946, "train_loss": 3.5852084159851074, "perplexity": 36.06087259800791, "lr": 0.0026291804804649314, "grad_norm": 0.144081, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:10.303667+00:00", "epoch": 0, "step": 17947, "train_loss": 3.5154941082000732, "perplexity": 33.63254193378429, "lr": 0.0026291804804649314, "grad_norm": 0.169249, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:10.608251+00:00", "epoch": 0, "step": 17948, "train_loss": 3.4970219135284424, "perplexity": 33.01697798400099, "lr": 0.0026291804804649314, "grad_norm": 0.145893, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:10.913134+00:00", "epoch": 0, "step": 17949, "train_loss": 3.5276708602905273, "perplexity": 34.04458061433445, "lr": 0.0026291804804649314, "grad_norm": 0.160979, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:11.219577+00:00", "epoch": 0, "step": 17950, "train_loss": 3.471074342727661, "perplexity": 32.171286873244526, "lr": 0.0026291804804649314, "grad_norm": 0.158412, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:11.524573+00:00", "epoch": 0, "step": 17951, "train_loss": 3.527891159057617, "perplexity": 34.05208141964862, "lr": 0.0026291804804649314, "grad_norm": 0.16514, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:11.829386+00:00", "epoch": 0, "step": 17952, "train_loss": 3.4654085636138916, "perplexity": 31.989526860281313, "lr": 0.0026291804804649314, "grad_norm": 0.190259, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:12.133915+00:00", "epoch": 0, "step": 17953, "train_loss": 3.5593202114105225, "perplexity": 35.139301727754045, "lr": 0.0026291804804649314, "grad_norm": 0.154101, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:12.437658+00:00", "epoch": 0, "step": 17954, "train_loss": 3.496731996536255, "perplexity": 33.0074071884879, "lr": 0.0026291804804649314, "grad_norm": 0.166431, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:12.741633+00:00", "epoch": 0, "step": 17955, "train_loss": 3.585238456726074, "perplexity": 36.06195590961232, "lr": 0.0026291804804649314, "grad_norm": 0.176659, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:13.046359+00:00", "epoch": 0, "step": 17956, "train_loss": 3.355741500854492, "perplexity": 28.666852807651697, "lr": 0.0026291804804649314, "grad_norm": 0.202396, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:13.352324+00:00", "epoch": 0, "step": 17957, "train_loss": 3.4214117527008057, "perplexity": 30.61260195272915, "lr": 0.0026291804804649314, "grad_norm": 0.18002, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:13.658292+00:00", "epoch": 0, "step": 17958, "train_loss": 3.444624900817871, "perplexity": 31.33152881139816, "lr": 0.0026291804804649314, "grad_norm": 0.186905, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:13.962681+00:00", "epoch": 0, "step": 17959, "train_loss": 3.5070135593414307, "perplexity": 33.34852552837192, "lr": 0.0026291804804649314, "grad_norm": 0.159764, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:14.267225+00:00", "epoch": 0, "step": 17960, "train_loss": 3.5593600273132324, "perplexity": 35.14070085862658, "lr": 0.0026291804804649314, "grad_norm": 0.164943, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:14.571302+00:00", "epoch": 0, "step": 17961, "train_loss": 3.5819976329803467, "perplexity": 35.94527464042168, "lr": 0.0026291804804649314, "grad_norm": 0.202582, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:14.877214+00:00", "epoch": 0, "step": 17962, "train_loss": 3.4662721157073975, "perplexity": 32.01716341425636, "lr": 0.0026291804804649314, "grad_norm": 0.197722, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:15.181518+00:00", "epoch": 0, "step": 17963, "train_loss": 3.5614993572235107, "perplexity": 35.215958883210085, "lr": 0.0026291804804649314, "grad_norm": 0.18558, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:15.486386+00:00", "epoch": 0, "step": 17964, "train_loss": 3.5218729972839355, "perplexity": 33.847765903511615, "lr": 0.0026291804804649314, "grad_norm": 0.165935, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:15.791051+00:00", "epoch": 0, "step": 17965, "train_loss": 3.5325160026550293, "perplexity": 34.20993170546547, "lr": 0.0026291804804649314, "grad_norm": 0.169126, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:16.095497+00:00", "epoch": 0, "step": 17966, "train_loss": 3.404724359512329, "perplexity": 30.105996148660598, "lr": 0.0026291804804649314, "grad_norm": 0.182625, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:16.399959+00:00", "epoch": 0, "step": 17967, "train_loss": 3.483027219772339, "perplexity": 32.55813366994406, "lr": 0.0026291804804649314, "grad_norm": 0.211756, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:16.704946+00:00", "epoch": 0, "step": 17968, "train_loss": 3.5345261096954346, "perplexity": 34.278766489503, "lr": 0.0026291804804649314, "grad_norm": 0.203557, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:17.009593+00:00", "epoch": 0, "step": 17969, "train_loss": 3.520618200302124, "perplexity": 33.80532046478828, "lr": 0.0026291804804649314, "grad_norm": 0.168548, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:17.314233+00:00", "epoch": 0, "step": 17970, "train_loss": 3.435157060623169, "perplexity": 31.03628676128552, "lr": 0.0026291804804649314, "grad_norm": 0.188501, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:17.618457+00:00", "epoch": 0, "step": 17971, "train_loss": 3.5404164791107178, "perplexity": 34.481276932353346, "lr": 0.0026291804804649314, "grad_norm": 0.173369, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:17.922988+00:00", "epoch": 0, "step": 17972, "train_loss": 3.4799678325653076, "perplexity": 32.45867794666736, "lr": 0.0026291804804649314, "grad_norm": 0.163559, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:18.228171+00:00", "epoch": 0, "step": 17973, "train_loss": 3.5718486309051514, "perplexity": 35.58231094863401, "lr": 0.0026291804804649314, "grad_norm": 0.171168, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:18.533969+00:00", "epoch": 0, "step": 17974, "train_loss": 3.513373851776123, "perplexity": 33.56130786462776, "lr": 0.0026291804804649314, "grad_norm": 0.183434, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:18.838654+00:00", "epoch": 0, "step": 17975, "train_loss": 3.5678956508636475, "perplexity": 35.44193242317496, "lr": 0.0026291804804649314, "grad_norm": 0.16997, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:19.143980+00:00", "epoch": 0, "step": 17976, "train_loss": 3.5050010681152344, "perplexity": 33.28147940083994, "lr": 0.0026291804804649314, "grad_norm": 0.159804, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:19.448596+00:00", "epoch": 0, "step": 17977, "train_loss": 3.466827869415283, "perplexity": 32.03496201690142, "lr": 0.0026291804804649314, "grad_norm": 0.163563, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:19.755340+00:00", "epoch": 0, "step": 17978, "train_loss": 3.5312016010284424, "perplexity": 34.16499565406539, "lr": 0.0026291804804649314, "grad_norm": 0.159678, "tokens_per_sec": 106820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:20.059738+00:00", "epoch": 0, "step": 17979, "train_loss": 3.493189811706543, "perplexity": 32.89069568037576, "lr": 0.0026291804804649314, "grad_norm": 0.18656, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:20.364622+00:00", "epoch": 0, "step": 17980, "train_loss": 3.528276205062866, "perplexity": 34.06519556218175, "lr": 0.0026291804804649314, "grad_norm": 0.18257, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:20.669322+00:00", "epoch": 0, "step": 17981, "train_loss": 3.370497941970825, "perplexity": 29.093010082422587, "lr": 0.0026291804804649314, "grad_norm": 0.160884, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:20.975035+00:00", "epoch": 0, "step": 17982, "train_loss": 3.4710230827331543, "perplexity": 32.16963781552183, "lr": 0.0026291804804649314, "grad_norm": 0.185465, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:21.280740+00:00", "epoch": 0, "step": 17983, "train_loss": 3.6050055027008057, "perplexity": 36.78188625640244, "lr": 0.0026291804804649314, "grad_norm": 0.168878, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:21.586477+00:00", "epoch": 0, "step": 17984, "train_loss": 3.5156326293945312, "perplexity": 33.637201076353165, "lr": 0.0026291804804649314, "grad_norm": 0.153606, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:21.891603+00:00", "epoch": 0, "step": 17985, "train_loss": 3.524146318435669, "perplexity": 33.92480027443534, "lr": 0.0026291804804649314, "grad_norm": 0.144065, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:22.195913+00:00", "epoch": 0, "step": 17986, "train_loss": 3.569535255432129, "perplexity": 35.50009084286817, "lr": 0.0026291804804649314, "grad_norm": 0.182481, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:22.500889+00:00", "epoch": 0, "step": 17987, "train_loss": 3.4611852169036865, "perplexity": 31.854708889164094, "lr": 0.0026291804804649314, "grad_norm": 0.182726, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:22.805343+00:00", "epoch": 0, "step": 17988, "train_loss": 3.586805820465088, "perplexity": 36.11852243023935, "lr": 0.0026291804804649314, "grad_norm": 0.195436, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:23.110468+00:00", "epoch": 0, "step": 17989, "train_loss": 3.585822820663452, "perplexity": 36.08303537459694, "lr": 0.0026291804804649314, "grad_norm": 0.168206, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:23.416502+00:00", "epoch": 0, "step": 17990, "train_loss": 3.4295482635498047, "perplexity": 30.862697794646735, "lr": 0.0026291804804649314, "grad_norm": 0.205902, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:23.722309+00:00", "epoch": 0, "step": 17991, "train_loss": 3.470268726348877, "perplexity": 32.14537959467803, "lr": 0.0026291804804649314, "grad_norm": 0.171971, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:24.026442+00:00", "epoch": 0, "step": 17992, "train_loss": 3.523099660873413, "perplexity": 33.88931120137327, "lr": 0.0026291804804649314, "grad_norm": 0.194243, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:24.331506+00:00", "epoch": 0, "step": 17993, "train_loss": 3.5449888706207275, "perplexity": 34.63929982621008, "lr": 0.0026291804804649314, "grad_norm": 0.192368, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:24.636946+00:00", "epoch": 0, "step": 17994, "train_loss": 3.5902259349823, "perplexity": 36.2422633966627, "lr": 0.0026291804804649314, "grad_norm": 0.181862, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:24.942794+00:00", "epoch": 0, "step": 17995, "train_loss": 3.519406795501709, "perplexity": 33.7643933319653, "lr": 0.0026291804804649314, "grad_norm": 0.180579, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:25.248697+00:00", "epoch": 0, "step": 17996, "train_loss": 3.470177173614502, "perplexity": 32.14243673199424, "lr": 0.0026291804804649314, "grad_norm": 0.204059, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:25.552832+00:00", "epoch": 0, "step": 17997, "train_loss": 3.5691211223602295, "perplexity": 35.48539212501737, "lr": 0.0026291804804649314, "grad_norm": 0.178782, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:25.857495+00:00", "epoch": 0, "step": 17998, "train_loss": 3.511474132537842, "perplexity": 33.49761132434332, "lr": 0.0026291804804649314, "grad_norm": 0.16968, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:26.163628+00:00", "epoch": 0, "step": 17999, "train_loss": 3.4948713779449463, "perplexity": 32.94605009180053, "lr": 0.0026291804804649314, "grad_norm": 0.168727, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:26.470162+00:00", "epoch": 0, "step": 18000, "train_loss": 3.509410858154297, "perplexity": 33.428567813335434, "lr": 0.0026291804804649314, "grad_norm": 0.174955, "tokens_per_sec": 106956} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:32:29.543657+00:00", "step": 18000, "epoch": 0, "val_loss": 3.464576256275177, "val_ppl": 31.96291281938065, "eval_train_loss": 3.509410858154297, "eval_train_ppl": 33.428567813335434} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:32:30.499407+00:00", "step": 18000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4646_epoch_0000_step_0018000.pt", "val_loss": 3.464576256275177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:31.858132+00:00", "epoch": 0, "step": 18001, "train_loss": 3.452561140060425, "perplexity": 31.581172626106284, "lr": 0.0026291804804649314, "grad_norm": 0.183803, "tokens_per_sec": 6082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:32.161855+00:00", "epoch": 0, "step": 18002, "train_loss": 3.475813865661621, "perplexity": 32.324125329778504, "lr": 0.0026291804804649314, "grad_norm": 0.191945, "tokens_per_sec": 107880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:32.465623+00:00", "epoch": 0, "step": 18003, "train_loss": 3.495577812194824, "perplexity": 32.96933253277218, "lr": 0.0026291804804649314, "grad_norm": 0.159764, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:32.769904+00:00", "epoch": 0, "step": 18004, "train_loss": 3.5575389862060547, "perplexity": 35.07676642910837, "lr": 0.0026291804804649314, "grad_norm": 0.173403, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:33.074008+00:00", "epoch": 0, "step": 18005, "train_loss": 3.439878225326538, "perplexity": 31.18316061788967, "lr": 0.0026291804804649314, "grad_norm": 0.214252, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:33.377914+00:00", "epoch": 0, "step": 18006, "train_loss": 3.46291184425354, "perplexity": 31.909757611392305, "lr": 0.0026291804804649314, "grad_norm": 0.171439, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:33.682142+00:00", "epoch": 0, "step": 18007, "train_loss": 3.450090169906616, "perplexity": 31.50323282424337, "lr": 0.0026291804804649314, "grad_norm": 0.149825, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:33.985825+00:00", "epoch": 0, "step": 18008, "train_loss": 3.546370029449463, "perplexity": 34.68717525516135, "lr": 0.0026291804804649314, "grad_norm": 0.172274, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:34.289845+00:00", "epoch": 0, "step": 18009, "train_loss": 3.5540263652801514, "perplexity": 34.95377118962722, "lr": 0.0026291804804649314, "grad_norm": 0.194338, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:34.593136+00:00", "epoch": 0, "step": 18010, "train_loss": 3.428069591522217, "perplexity": 30.81709571029413, "lr": 0.0026291804804649314, "grad_norm": 0.182188, "tokens_per_sec": 108106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:34.897024+00:00", "epoch": 0, "step": 18011, "train_loss": 3.4142305850982666, "perplexity": 30.393555174507927, "lr": 0.0026291804804649314, "grad_norm": 0.162355, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:35.200999+00:00", "epoch": 0, "step": 18012, "train_loss": 3.5046393871307373, "perplexity": 33.26944429916963, "lr": 0.0026291804804649314, "grad_norm": 0.188218, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:35.506809+00:00", "epoch": 0, "step": 18013, "train_loss": 3.5117201805114746, "perplexity": 33.505854357780464, "lr": 0.0026291804804649314, "grad_norm": 0.162593, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:35.902554+00:00", "epoch": 0, "step": 18014, "train_loss": 3.4807233810424805, "perplexity": 32.4832113182697, "lr": 0.0026291804804649314, "grad_norm": 0.163244, "tokens_per_sec": 82800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:36.206330+00:00", "epoch": 0, "step": 18015, "train_loss": 3.5458717346191406, "perplexity": 34.66989512071161, "lr": 0.0026291804804649314, "grad_norm": 0.168932, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:36.510328+00:00", "epoch": 0, "step": 18016, "train_loss": 3.5361602306365967, "perplexity": 34.334827932774616, "lr": 0.0026291804804649314, "grad_norm": 0.176443, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:36.814053+00:00", "epoch": 0, "step": 18017, "train_loss": 3.493133544921875, "perplexity": 32.88884507874852, "lr": 0.0026291804804649314, "grad_norm": 0.179837, "tokens_per_sec": 107887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:37.118371+00:00", "epoch": 0, "step": 18018, "train_loss": 3.5004148483276367, "perplexity": 33.12919269852072, "lr": 0.0026291804804649314, "grad_norm": 0.151572, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:37.423225+00:00", "epoch": 0, "step": 18019, "train_loss": 3.4965474605560303, "perplexity": 33.00131669622247, "lr": 0.0026291804804649314, "grad_norm": 0.162788, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:37.729212+00:00", "epoch": 0, "step": 18020, "train_loss": 3.521477460861206, "perplexity": 33.83438052664531, "lr": 0.0026291804804649314, "grad_norm": 0.173164, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:38.034783+00:00", "epoch": 0, "step": 18021, "train_loss": 3.478299617767334, "perplexity": 32.404575039975946, "lr": 0.0026291804804649314, "grad_norm": 0.164163, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:38.340515+00:00", "epoch": 0, "step": 18022, "train_loss": 3.4303781986236572, "perplexity": 30.888322461954463, "lr": 0.0026291804804649314, "grad_norm": 0.171678, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:38.645496+00:00", "epoch": 0, "step": 18023, "train_loss": 3.5097129344940186, "perplexity": 33.43866731807653, "lr": 0.0026291804804649314, "grad_norm": 0.169476, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:38.950627+00:00", "epoch": 0, "step": 18024, "train_loss": 3.5703139305114746, "perplexity": 35.5277446441877, "lr": 0.0026291804804649314, "grad_norm": 0.161425, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:39.256395+00:00", "epoch": 0, "step": 18025, "train_loss": 3.500499963760376, "perplexity": 33.13201262410144, "lr": 0.0026291804804649314, "grad_norm": 0.164671, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:39.561192+00:00", "epoch": 0, "step": 18026, "train_loss": 3.532729148864746, "perplexity": 34.21722419989933, "lr": 0.0026291804804649314, "grad_norm": 0.16938, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:39.865894+00:00", "epoch": 0, "step": 18027, "train_loss": 3.4802370071411133, "perplexity": 32.46741617353636, "lr": 0.0026291804804649314, "grad_norm": 0.189099, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:40.170649+00:00", "epoch": 0, "step": 18028, "train_loss": 3.4403045177459717, "perplexity": 31.196456596661186, "lr": 0.0026291804804649314, "grad_norm": 0.170596, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:40.475276+00:00", "epoch": 0, "step": 18029, "train_loss": 3.466132164001465, "perplexity": 32.01268287115455, "lr": 0.0026291804804649314, "grad_norm": 0.14969, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:40.780316+00:00", "epoch": 0, "step": 18030, "train_loss": 3.407379627227783, "perplexity": 30.186041852632655, "lr": 0.0026291804804649314, "grad_norm": 0.167593, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:41.085886+00:00", "epoch": 0, "step": 18031, "train_loss": 3.569377899169922, "perplexity": 35.49450512075083, "lr": 0.0026291804804649314, "grad_norm": 0.167082, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:41.392203+00:00", "epoch": 0, "step": 18032, "train_loss": 3.4247801303863525, "perplexity": 30.71589061800175, "lr": 0.0026291804804649314, "grad_norm": 0.173599, "tokens_per_sec": 106973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:41.697594+00:00", "epoch": 0, "step": 18033, "train_loss": 3.4626810550689697, "perplexity": 31.902394034203006, "lr": 0.0026291804804649314, "grad_norm": 0.170844, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:42.001500+00:00", "epoch": 0, "step": 18034, "train_loss": 3.5735738277435303, "perplexity": 35.6437504213383, "lr": 0.0026291804804649314, "grad_norm": 0.178297, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:42.306858+00:00", "epoch": 0, "step": 18035, "train_loss": 3.4819247722625732, "perplexity": 32.52225981471159, "lr": 0.0026291804804649314, "grad_norm": 0.181274, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:42.611661+00:00", "epoch": 0, "step": 18036, "train_loss": 3.4067294597625732, "perplexity": 30.166422249021714, "lr": 0.0026291804804649314, "grad_norm": 0.180661, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:42.916441+00:00", "epoch": 0, "step": 18037, "train_loss": 3.4214181900024414, "perplexity": 30.612799015916053, "lr": 0.0026291804804649314, "grad_norm": 0.205192, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:43.220935+00:00", "epoch": 0, "step": 18038, "train_loss": 3.565416097640991, "perplexity": 35.354161127187325, "lr": 0.0026291804804649314, "grad_norm": 0.188818, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:43.526505+00:00", "epoch": 0, "step": 18039, "train_loss": 3.5048410892486572, "perplexity": 33.27615549335508, "lr": 0.0026291804804649314, "grad_norm": 0.177016, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:43.830318+00:00", "epoch": 0, "step": 18040, "train_loss": 3.4712445735931396, "perplexity": 32.17676388541755, "lr": 0.0026291804804649314, "grad_norm": 0.173575, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:44.135125+00:00", "epoch": 0, "step": 18041, "train_loss": 3.4900219440460205, "perplexity": 32.78666717047113, "lr": 0.0026291804804649314, "grad_norm": 0.160835, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:44.439808+00:00", "epoch": 0, "step": 18042, "train_loss": 3.428971767425537, "perplexity": 30.844910696572857, "lr": 0.0026291804804649314, "grad_norm": 0.169991, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:44.744536+00:00", "epoch": 0, "step": 18043, "train_loss": 3.595778465270996, "perplexity": 36.444059382608415, "lr": 0.0026291804804649314, "grad_norm": 0.180515, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:45.048944+00:00", "epoch": 0, "step": 18044, "train_loss": 3.5424587726593018, "perplexity": 34.551769780822134, "lr": 0.0026291804804649314, "grad_norm": 0.220025, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:45.352731+00:00", "epoch": 0, "step": 18045, "train_loss": 3.3934199810028076, "perplexity": 29.767582952005654, "lr": 0.0026291804804649314, "grad_norm": 0.16573, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:45.657585+00:00", "epoch": 0, "step": 18046, "train_loss": 3.446716070175171, "perplexity": 31.397116898363414, "lr": 0.0026291804804649314, "grad_norm": 0.211447, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:45.963006+00:00", "epoch": 0, "step": 18047, "train_loss": 3.4378786087036133, "perplexity": 31.120868552443635, "lr": 0.0026291804804649314, "grad_norm": 0.176339, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:46.268396+00:00", "epoch": 0, "step": 18048, "train_loss": 3.5157642364501953, "perplexity": 33.641628260665556, "lr": 0.0026291804804649314, "grad_norm": 0.177581, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:46.572700+00:00", "epoch": 0, "step": 18049, "train_loss": 3.5168139934539795, "perplexity": 33.67696233840387, "lr": 0.0026291804804649314, "grad_norm": 0.159352, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:46.876734+00:00", "epoch": 0, "step": 18050, "train_loss": 3.420758008956909, "perplexity": 30.592595695913303, "lr": 0.0026291804804649314, "grad_norm": 0.172494, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:47.181022+00:00", "epoch": 0, "step": 18051, "train_loss": 3.486929416656494, "perplexity": 32.68543012408815, "lr": 0.0026291804804649314, "grad_norm": 0.160554, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:47.485958+00:00", "epoch": 0, "step": 18052, "train_loss": 3.534515619277954, "perplexity": 34.278406892817976, "lr": 0.0026291804804649314, "grad_norm": 0.166151, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:47.791927+00:00", "epoch": 0, "step": 18053, "train_loss": 3.4409890174865723, "perplexity": 31.21781787316932, "lr": 0.0026291804804649314, "grad_norm": 0.180986, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:48.096703+00:00", "epoch": 0, "step": 18054, "train_loss": 3.38038969039917, "perplexity": 29.38221885120041, "lr": 0.0026291804804649314, "grad_norm": 0.177494, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:48.401595+00:00", "epoch": 0, "step": 18055, "train_loss": 3.5006985664367676, "perplexity": 33.13859338393946, "lr": 0.0026291804804649314, "grad_norm": 0.16292, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:48.705496+00:00", "epoch": 0, "step": 18056, "train_loss": 3.424189805984497, "perplexity": 30.69776362917355, "lr": 0.0026291804804649314, "grad_norm": 0.143879, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:49.009711+00:00", "epoch": 0, "step": 18057, "train_loss": 3.4969756603240967, "perplexity": 33.015450878288455, "lr": 0.0026291804804649314, "grad_norm": 0.163121, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:49.314814+00:00", "epoch": 0, "step": 18058, "train_loss": 3.6024250984191895, "perplexity": 36.68709647020195, "lr": 0.0026291804804649314, "grad_norm": 0.167344, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:49.619455+00:00", "epoch": 0, "step": 18059, "train_loss": 3.4250409603118896, "perplexity": 30.723903286390872, "lr": 0.0026291804804649314, "grad_norm": 0.220094, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:49.924922+00:00", "epoch": 0, "step": 18060, "train_loss": 3.4986186027526855, "perplexity": 33.069737946387896, "lr": 0.0026291804804649314, "grad_norm": 0.174196, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:50.228689+00:00", "epoch": 0, "step": 18061, "train_loss": 3.542574405670166, "perplexity": 34.55576533699722, "lr": 0.0026291804804649314, "grad_norm": 0.168047, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:50.533731+00:00", "epoch": 0, "step": 18062, "train_loss": 3.4367518424987793, "perplexity": 31.085822357635514, "lr": 0.0026291804804649314, "grad_norm": 0.173566, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:50.839698+00:00", "epoch": 0, "step": 18063, "train_loss": 3.530608892440796, "perplexity": 34.14475176769962, "lr": 0.0026291804804649314, "grad_norm": 0.179816, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:51.144630+00:00", "epoch": 0, "step": 18064, "train_loss": 3.5435471534729004, "perplexity": 34.58939573609714, "lr": 0.0026291804804649314, "grad_norm": 0.160263, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:51.447705+00:00", "epoch": 0, "step": 18065, "train_loss": 3.5125019550323486, "perplexity": 33.532058622596175, "lr": 0.0026291804804649314, "grad_norm": 0.173425, "tokens_per_sec": 108118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:51.752252+00:00", "epoch": 0, "step": 18066, "train_loss": 3.5022294521331787, "perplexity": 33.18936363435972, "lr": 0.0026291804804649314, "grad_norm": 0.16086, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:52.055735+00:00", "epoch": 0, "step": 18067, "train_loss": 3.5057578086853027, "perplexity": 33.306674378362935, "lr": 0.0026291804804649314, "grad_norm": 0.16292, "tokens_per_sec": 107973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:52.361372+00:00", "epoch": 0, "step": 18068, "train_loss": 3.5954389572143555, "perplexity": 36.43168843096903, "lr": 0.0026291804804649314, "grad_norm": 0.167481, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:52.666929+00:00", "epoch": 0, "step": 18069, "train_loss": 3.407578945159912, "perplexity": 30.192059071723794, "lr": 0.0026291804804649314, "grad_norm": 0.174812, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:52.970895+00:00", "epoch": 0, "step": 18070, "train_loss": 3.465714931488037, "perplexity": 31.999328925062624, "lr": 0.0026291804804649314, "grad_norm": 0.170644, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:53.276209+00:00", "epoch": 0, "step": 18071, "train_loss": 3.5208988189697266, "perplexity": 33.81480819993052, "lr": 0.0026291804804649314, "grad_norm": 0.187049, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:53.579980+00:00", "epoch": 0, "step": 18072, "train_loss": 3.459256410598755, "perplexity": 31.793326542175883, "lr": 0.0026291804804649314, "grad_norm": 0.1786, "tokens_per_sec": 107872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:53.884957+00:00", "epoch": 0, "step": 18073, "train_loss": 3.5283799171447754, "perplexity": 34.06872871774644, "lr": 0.0026291804804649314, "grad_norm": 0.191322, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:54.189732+00:00", "epoch": 0, "step": 18074, "train_loss": 3.491083860397339, "perplexity": 32.821502361200466, "lr": 0.0026291804804649314, "grad_norm": 0.187221, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:54.493741+00:00", "epoch": 0, "step": 18075, "train_loss": 3.543783187866211, "perplexity": 34.597560986736774, "lr": 0.0026291804804649314, "grad_norm": 0.195183, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:54.797827+00:00", "epoch": 0, "step": 18076, "train_loss": 3.4785075187683105, "perplexity": 32.411312683920166, "lr": 0.0026291804804649314, "grad_norm": 0.198902, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:55.102276+00:00", "epoch": 0, "step": 18077, "train_loss": 3.4598681926727295, "perplexity": 31.812783080400216, "lr": 0.0026291804804649314, "grad_norm": 0.214858, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:55.407045+00:00", "epoch": 0, "step": 18078, "train_loss": 3.404264211654663, "perplexity": 30.092146125803445, "lr": 0.0026291804804649314, "grad_norm": 0.17765, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:55.712612+00:00", "epoch": 0, "step": 18079, "train_loss": 3.5094540119171143, "perplexity": 33.43001041294875, "lr": 0.0026291804804649314, "grad_norm": 0.190847, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:56.017934+00:00", "epoch": 0, "step": 18080, "train_loss": 3.411630868911743, "perplexity": 30.314643175998732, "lr": 0.0026291804804649314, "grad_norm": 0.17949, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:56.321164+00:00", "epoch": 0, "step": 18081, "train_loss": 3.40816593170166, "perplexity": 30.209786606469706, "lr": 0.0026291804804649314, "grad_norm": 0.171402, "tokens_per_sec": 107998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:56.624917+00:00", "epoch": 0, "step": 18082, "train_loss": 3.377939224243164, "perplexity": 29.310306863248194, "lr": 0.0026291804804649314, "grad_norm": 0.193132, "tokens_per_sec": 107880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:56.929475+00:00", "epoch": 0, "step": 18083, "train_loss": 3.5355210304260254, "perplexity": 34.312888116244736, "lr": 0.0026291804804649314, "grad_norm": 0.186238, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:57.234217+00:00", "epoch": 0, "step": 18084, "train_loss": 3.493964433670044, "perplexity": 32.91618340604619, "lr": 0.0026291804804649314, "grad_norm": 0.155659, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:57.540323+00:00", "epoch": 0, "step": 18085, "train_loss": 3.457289457321167, "perplexity": 31.73085201670698, "lr": 0.0026291804804649314, "grad_norm": 0.168143, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:57.844313+00:00", "epoch": 0, "step": 18086, "train_loss": 3.4150116443634033, "perplexity": 30.41730361564009, "lr": 0.0026291804804649314, "grad_norm": 0.178601, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:58.148847+00:00", "epoch": 0, "step": 18087, "train_loss": 3.5427839756011963, "perplexity": 34.563007945245545, "lr": 0.0026291804804649314, "grad_norm": 0.172046, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:58.452591+00:00", "epoch": 0, "step": 18088, "train_loss": 3.527989387512207, "perplexity": 34.05542646726878, "lr": 0.0026291804804649314, "grad_norm": 0.18655, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:58.758052+00:00", "epoch": 0, "step": 18089, "train_loss": 3.6081736087799072, "perplexity": 36.89859995692398, "lr": 0.0026291804804649314, "grad_norm": 0.165585, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:59.063333+00:00", "epoch": 0, "step": 18090, "train_loss": 3.479257822036743, "perplexity": 32.435640123098395, "lr": 0.0026291804804649314, "grad_norm": 0.156788, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:59.368748+00:00", "epoch": 0, "step": 18091, "train_loss": 3.543602466583252, "perplexity": 34.59130903607529, "lr": 0.0026291804804649314, "grad_norm": 0.174807, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:59.673891+00:00", "epoch": 0, "step": 18092, "train_loss": 3.5317227840423584, "perplexity": 34.18280651042554, "lr": 0.0026291804804649314, "grad_norm": 0.159009, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:32:59.977955+00:00", "epoch": 0, "step": 18093, "train_loss": 3.47153377532959, "perplexity": 32.186070807130655, "lr": 0.0026291804804649314, "grad_norm": 0.145548, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:00.283169+00:00", "epoch": 0, "step": 18094, "train_loss": 3.435544490814209, "perplexity": 31.048313485402193, "lr": 0.0026291804804649314, "grad_norm": 0.156514, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:00.588868+00:00", "epoch": 0, "step": 18095, "train_loss": 3.495001792907715, "perplexity": 32.95034702988304, "lr": 0.0026291804804649314, "grad_norm": 0.158269, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:00.894342+00:00", "epoch": 0, "step": 18096, "train_loss": 3.5711278915405273, "perplexity": 35.556674616115814, "lr": 0.0026291804804649314, "grad_norm": 0.171683, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:01.199250+00:00", "epoch": 0, "step": 18097, "train_loss": 3.4333674907684326, "perplexity": 30.980794826251316, "lr": 0.0026291804804649314, "grad_norm": 0.180743, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:01.502706+00:00", "epoch": 0, "step": 18098, "train_loss": 3.6289496421813965, "perplexity": 37.673225461466224, "lr": 0.0026291804804649314, "grad_norm": 0.200711, "tokens_per_sec": 107984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:01.807842+00:00", "epoch": 0, "step": 18099, "train_loss": 3.555450439453125, "perplexity": 35.00358341215799, "lr": 0.0026291804804649314, "grad_norm": 0.180863, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:02.113562+00:00", "epoch": 0, "step": 18100, "train_loss": 3.4888241291046143, "perplexity": 32.74741832177685, "lr": 0.0026291804804649314, "grad_norm": 0.172253, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:02.418424+00:00", "epoch": 0, "step": 18101, "train_loss": 3.4563426971435547, "perplexity": 31.700824726180237, "lr": 0.0026291804804649314, "grad_norm": 0.21084, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:02.722868+00:00", "epoch": 0, "step": 18102, "train_loss": 3.638009548187256, "perplexity": 38.01609216789178, "lr": 0.0026291804804649314, "grad_norm": 0.220223, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:03.027840+00:00", "epoch": 0, "step": 18103, "train_loss": 3.4017601013183594, "perplexity": 30.016886340378353, "lr": 0.0026291804804649314, "grad_norm": 0.152044, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:03.332629+00:00", "epoch": 0, "step": 18104, "train_loss": 3.394993305206299, "perplexity": 29.814453872651935, "lr": 0.0026291804804649314, "grad_norm": 0.168384, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:03.637310+00:00", "epoch": 0, "step": 18105, "train_loss": 3.612454891204834, "perplexity": 37.056911931748424, "lr": 0.0026291804804649314, "grad_norm": 0.167993, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:03.942242+00:00", "epoch": 0, "step": 18106, "train_loss": 3.5734684467315674, "perplexity": 35.639994444756525, "lr": 0.0026291804804649314, "grad_norm": 0.160001, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:04.246673+00:00", "epoch": 0, "step": 18107, "train_loss": 3.4538204669952393, "perplexity": 31.620968700302033, "lr": 0.0026291804804649314, "grad_norm": 0.171007, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:04.549999+00:00", "epoch": 0, "step": 18108, "train_loss": 3.53344464302063, "perplexity": 34.24171518437824, "lr": 0.0026291804804649314, "grad_norm": 0.151922, "tokens_per_sec": 108029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:04.853984+00:00", "epoch": 0, "step": 18109, "train_loss": 3.5556280612945557, "perplexity": 35.009801365306124, "lr": 0.0026291804804649314, "grad_norm": 0.151678, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:05.158454+00:00", "epoch": 0, "step": 18110, "train_loss": 3.5074360370635986, "perplexity": 33.36261751404021, "lr": 0.0026291804804649314, "grad_norm": 0.149725, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:05.463637+00:00", "epoch": 0, "step": 18111, "train_loss": 3.525987148284912, "perplexity": 33.98730757441063, "lr": 0.0026291804804649314, "grad_norm": 0.156553, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:05.768874+00:00", "epoch": 0, "step": 18112, "train_loss": 3.5384278297424316, "perplexity": 34.412773899596424, "lr": 0.0026291804804649314, "grad_norm": 0.167816, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:06.072714+00:00", "epoch": 0, "step": 18113, "train_loss": 3.515535354614258, "perplexity": 33.63392918414832, "lr": 0.0026291804804649314, "grad_norm": 0.1756, "tokens_per_sec": 107846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:06.377343+00:00", "epoch": 0, "step": 18114, "train_loss": 3.5330023765563965, "perplexity": 34.22657457041406, "lr": 0.0026291804804649314, "grad_norm": 0.161599, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:06.682016+00:00", "epoch": 0, "step": 18115, "train_loss": 3.4995758533477783, "perplexity": 33.10140912892737, "lr": 0.0026291804804649314, "grad_norm": 0.180762, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:06.987118+00:00", "epoch": 0, "step": 18116, "train_loss": 3.5305190086364746, "perplexity": 34.1416828454384, "lr": 0.0026291804804649314, "grad_norm": 0.158074, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:07.291623+00:00", "epoch": 0, "step": 18117, "train_loss": 3.5162127017974854, "perplexity": 33.65671874869444, "lr": 0.0026291804804649314, "grad_norm": 0.150008, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:07.595957+00:00", "epoch": 0, "step": 18118, "train_loss": 3.4776298999786377, "perplexity": 32.382880385095554, "lr": 0.0026291804804649314, "grad_norm": 0.163385, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:07.900558+00:00", "epoch": 0, "step": 18119, "train_loss": 3.5831985473632812, "perplexity": 35.988467768168995, "lr": 0.0026291804804649314, "grad_norm": 0.142125, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:08.204787+00:00", "epoch": 0, "step": 18120, "train_loss": 3.4976675510406494, "perplexity": 33.03830186653567, "lr": 0.0026291804804649314, "grad_norm": 0.156191, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:08.510082+00:00", "epoch": 0, "step": 18121, "train_loss": 3.5953989028930664, "perplexity": 36.43022921363969, "lr": 0.0026291804804649314, "grad_norm": 0.173863, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:08.815492+00:00", "epoch": 0, "step": 18122, "train_loss": 3.6026575565338135, "perplexity": 36.69562567478144, "lr": 0.0026291804804649314, "grad_norm": 0.160423, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:09.120120+00:00", "epoch": 0, "step": 18123, "train_loss": 3.5430002212524414, "perplexity": 34.57048285358673, "lr": 0.0026291804804649314, "grad_norm": 0.183273, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:09.425157+00:00", "epoch": 0, "step": 18124, "train_loss": 3.489563465118408, "perplexity": 32.77163861986696, "lr": 0.0026291804804649314, "grad_norm": 0.171181, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:09.729231+00:00", "epoch": 0, "step": 18125, "train_loss": 3.5586354732513428, "perplexity": 35.11524874291221, "lr": 0.0026291804804649314, "grad_norm": 0.186472, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:10.034482+00:00", "epoch": 0, "step": 18126, "train_loss": 3.5175063610076904, "perplexity": 33.70028724820812, "lr": 0.0026291804804649314, "grad_norm": 0.176269, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:10.339821+00:00", "epoch": 0, "step": 18127, "train_loss": 3.5448012351989746, "perplexity": 34.63280087631363, "lr": 0.0026291804804649314, "grad_norm": 0.159393, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:10.645381+00:00", "epoch": 0, "step": 18128, "train_loss": 3.574610710144043, "perplexity": 35.68072796620824, "lr": 0.0026291804804649314, "grad_norm": 0.172162, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:10.949848+00:00", "epoch": 0, "step": 18129, "train_loss": 3.4772965908050537, "perplexity": 32.372088672584425, "lr": 0.0026291804804649314, "grad_norm": 0.181097, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:11.255058+00:00", "epoch": 0, "step": 18130, "train_loss": 3.6201138496398926, "perplexity": 37.341818932693776, "lr": 0.0026291804804649314, "grad_norm": 0.179041, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:11.561242+00:00", "epoch": 0, "step": 18131, "train_loss": 3.4014029502868652, "perplexity": 30.006167692661624, "lr": 0.0026291804804649314, "grad_norm": 0.164439, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:11.866309+00:00", "epoch": 0, "step": 18132, "train_loss": 3.497471809387207, "perplexity": 33.03183552758797, "lr": 0.0026291804804649314, "grad_norm": 0.155406, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:12.172163+00:00", "epoch": 0, "step": 18133, "train_loss": 3.403130292892456, "perplexity": 30.05804341522203, "lr": 0.0026291804804649314, "grad_norm": 0.173047, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:12.477464+00:00", "epoch": 0, "step": 18134, "train_loss": 3.542200803756714, "perplexity": 34.542857648265255, "lr": 0.0026291804804649314, "grad_norm": 0.1855, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:12.780831+00:00", "epoch": 0, "step": 18135, "train_loss": 3.4724855422973633, "perplexity": 32.21671902881622, "lr": 0.0026291804804649314, "grad_norm": 0.172052, "tokens_per_sec": 108014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:13.084579+00:00", "epoch": 0, "step": 18136, "train_loss": 3.532409191131592, "perplexity": 34.20627788568182, "lr": 0.0026291804804649314, "grad_norm": 0.174487, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:13.390206+00:00", "epoch": 0, "step": 18137, "train_loss": 3.493351936340332, "perplexity": 32.89602850464742, "lr": 0.0026291804804649314, "grad_norm": 0.210392, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:13.695428+00:00", "epoch": 0, "step": 18138, "train_loss": 3.493622064590454, "perplexity": 32.90491585157103, "lr": 0.0026291804804649314, "grad_norm": 0.16495, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:14.000418+00:00", "epoch": 0, "step": 18139, "train_loss": 3.6196346282958984, "perplexity": 37.32392822318498, "lr": 0.0026291804804649314, "grad_norm": 0.194653, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:14.304505+00:00", "epoch": 0, "step": 18140, "train_loss": 3.418523073196411, "perplexity": 30.52429955695559, "lr": 0.0026291804804649314, "grad_norm": 0.167232, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:14.608843+00:00", "epoch": 0, "step": 18141, "train_loss": 3.5346531867980957, "perplexity": 34.28312281261989, "lr": 0.0026291804804649314, "grad_norm": 0.227762, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:14.913883+00:00", "epoch": 0, "step": 18142, "train_loss": 3.4506421089172363, "perplexity": 31.520625486802665, "lr": 0.0026291804804649314, "grad_norm": 0.179227, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:15.219516+00:00", "epoch": 0, "step": 18143, "train_loss": 3.5194106101989746, "perplexity": 33.76452213314989, "lr": 0.0026291804804649314, "grad_norm": 0.178442, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:15.524126+00:00", "epoch": 0, "step": 18144, "train_loss": 3.5571706295013428, "perplexity": 35.0638480464471, "lr": 0.0026291804804649314, "grad_norm": 0.182813, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:15.828167+00:00", "epoch": 0, "step": 18145, "train_loss": 3.5233066082000732, "perplexity": 33.8963252294709, "lr": 0.0026291804804649314, "grad_norm": 0.181792, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:16.132741+00:00", "epoch": 0, "step": 18146, "train_loss": 3.4864308834075928, "perplexity": 32.66913941148499, "lr": 0.0026291804804649314, "grad_norm": 0.212976, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:16.437105+00:00", "epoch": 0, "step": 18147, "train_loss": 3.51299786567688, "perplexity": 33.54869165130728, "lr": 0.0026291804804649314, "grad_norm": 0.172064, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:16.742241+00:00", "epoch": 0, "step": 18148, "train_loss": 3.6339919567108154, "perplexity": 37.86366543922055, "lr": 0.0026291804804649314, "grad_norm": 0.178078, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:17.047255+00:00", "epoch": 0, "step": 18149, "train_loss": 3.522141933441162, "perplexity": 33.85687001576202, "lr": 0.0026291804804649314, "grad_norm": 0.173394, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:17.351932+00:00", "epoch": 0, "step": 18150, "train_loss": 3.4773199558258057, "perplexity": 32.37284505594447, "lr": 0.0026291804804649314, "grad_norm": 0.173393, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:17.655354+00:00", "epoch": 0, "step": 18151, "train_loss": 3.5263254642486572, "perplexity": 33.998807968401515, "lr": 0.0026291804804649314, "grad_norm": 0.166932, "tokens_per_sec": 107933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:17.959821+00:00", "epoch": 0, "step": 18152, "train_loss": 3.509230375289917, "perplexity": 33.42253507408377, "lr": 0.0026291804804649314, "grad_norm": 0.166481, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:18.263891+00:00", "epoch": 0, "step": 18153, "train_loss": 3.5251314640045166, "perplexity": 33.95823760871632, "lr": 0.0026291804804649314, "grad_norm": 0.153431, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:18.569221+00:00", "epoch": 0, "step": 18154, "train_loss": 3.5020039081573486, "perplexity": 33.181878817439774, "lr": 0.0026291804804649314, "grad_norm": 0.183876, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:18.874649+00:00", "epoch": 0, "step": 18155, "train_loss": 3.408088445663452, "perplexity": 30.2074458604792, "lr": 0.0026291804804649314, "grad_norm": 0.173219, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:19.178658+00:00", "epoch": 0, "step": 18156, "train_loss": 3.552393913269043, "perplexity": 34.896757384364754, "lr": 0.0026291804804649314, "grad_norm": 0.155147, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:19.483283+00:00", "epoch": 0, "step": 18157, "train_loss": 3.4633030891418457, "perplexity": 31.922244583517557, "lr": 0.0026291804804649314, "grad_norm": 0.182023, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:19.787156+00:00", "epoch": 0, "step": 18158, "train_loss": 3.5440337657928467, "perplexity": 34.60623145810448, "lr": 0.0026291804804649314, "grad_norm": 0.173112, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:20.091163+00:00", "epoch": 0, "step": 18159, "train_loss": 3.469752311706543, "perplexity": 32.12878353556475, "lr": 0.0026291804804649314, "grad_norm": 0.204406, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:20.396907+00:00", "epoch": 0, "step": 18160, "train_loss": 3.4686784744262695, "perplexity": 32.09430096768145, "lr": 0.0026291804804649314, "grad_norm": 0.190638, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:20.701285+00:00", "epoch": 0, "step": 18161, "train_loss": 3.6014647483825684, "perplexity": 36.65188092809337, "lr": 0.0026291804804649314, "grad_norm": 0.16287, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:21.004932+00:00", "epoch": 0, "step": 18162, "train_loss": 3.504162549972534, "perplexity": 33.25358397359182, "lr": 0.0026291804804649314, "grad_norm": 0.156654, "tokens_per_sec": 107911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:21.309512+00:00", "epoch": 0, "step": 18163, "train_loss": 3.4976022243499756, "perplexity": 33.03614365410445, "lr": 0.0026291804804649314, "grad_norm": 0.160185, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:21.615017+00:00", "epoch": 0, "step": 18164, "train_loss": 3.5206995010375977, "perplexity": 33.808068973931384, "lr": 0.0026291804804649314, "grad_norm": 0.162359, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:21.919108+00:00", "epoch": 0, "step": 18165, "train_loss": 3.501453161239624, "perplexity": 33.16360903143273, "lr": 0.0026291804804649314, "grad_norm": 0.177843, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:22.224575+00:00", "epoch": 0, "step": 18166, "train_loss": 3.4132087230682373, "perplexity": 30.362513017619413, "lr": 0.0026291804804649314, "grad_norm": 0.163014, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:22.529380+00:00", "epoch": 0, "step": 18167, "train_loss": 3.449401378631592, "perplexity": 31.48154114371672, "lr": 0.0026291804804649314, "grad_norm": 0.161605, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:22.834628+00:00", "epoch": 0, "step": 18168, "train_loss": 3.502610921859741, "perplexity": 33.20202678698123, "lr": 0.0026291804804649314, "grad_norm": 0.172561, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:23.139693+00:00", "epoch": 0, "step": 18169, "train_loss": 3.582062005996704, "perplexity": 35.94758862065222, "lr": 0.0026291804804649314, "grad_norm": 0.149721, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:23.445148+00:00", "epoch": 0, "step": 18170, "train_loss": 3.3985934257507324, "perplexity": 29.92198294322884, "lr": 0.0026291804804649314, "grad_norm": 0.162566, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:23.749699+00:00", "epoch": 0, "step": 18171, "train_loss": 3.585630416870117, "perplexity": 36.076093529555905, "lr": 0.0026291804804649314, "grad_norm": 0.158911, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:24.053874+00:00", "epoch": 0, "step": 18172, "train_loss": 3.4377384185791016, "perplexity": 31.1165060198055, "lr": 0.0026291804804649314, "grad_norm": 0.159588, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:24.357681+00:00", "epoch": 0, "step": 18173, "train_loss": 3.563051700592041, "perplexity": 35.27066859659277, "lr": 0.0026291804804649314, "grad_norm": 0.152463, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:24.662113+00:00", "epoch": 0, "step": 18174, "train_loss": 3.4506609439849854, "perplexity": 31.521219185510358, "lr": 0.0026291804804649314, "grad_norm": 0.143154, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:24.967544+00:00", "epoch": 0, "step": 18175, "train_loss": 3.5017290115356445, "perplexity": 33.17275848468283, "lr": 0.0026291804804649314, "grad_norm": 0.17363, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:25.272308+00:00", "epoch": 0, "step": 18176, "train_loss": 3.618868350982666, "perplexity": 37.2953386988977, "lr": 0.0026291804804649314, "grad_norm": 0.176482, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:25.576465+00:00", "epoch": 0, "step": 18177, "train_loss": 3.5740408897399902, "perplexity": 35.66040215096374, "lr": 0.0026291804804649314, "grad_norm": 0.182799, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:25.880325+00:00", "epoch": 0, "step": 18178, "train_loss": 3.5096917152404785, "perplexity": 33.43795778204461, "lr": 0.0026291804804649314, "grad_norm": 0.185505, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:26.185280+00:00", "epoch": 0, "step": 18179, "train_loss": 3.5964860916137695, "perplexity": 36.46985728562653, "lr": 0.0026291804804649314, "grad_norm": 0.152144, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:26.490408+00:00", "epoch": 0, "step": 18180, "train_loss": 3.5872812271118164, "perplexity": 36.13569749811993, "lr": 0.0026291804804649314, "grad_norm": 0.150254, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:26.795221+00:00", "epoch": 0, "step": 18181, "train_loss": 3.4831383228302, "perplexity": 32.56175117910749, "lr": 0.0026291804804649314, "grad_norm": 0.170606, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:27.099300+00:00", "epoch": 0, "step": 18182, "train_loss": 3.5258991718292236, "perplexity": 33.98431762307643, "lr": 0.0026291804804649314, "grad_norm": 0.176306, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:27.403005+00:00", "epoch": 0, "step": 18183, "train_loss": 3.5563571453094482, "perplexity": 35.0353357590742, "lr": 0.0026291804804649314, "grad_norm": 0.182468, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:27.708184+00:00", "epoch": 0, "step": 18184, "train_loss": 3.480926990509033, "perplexity": 32.48982588096908, "lr": 0.0026291804804649314, "grad_norm": 0.189819, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:28.013392+00:00", "epoch": 0, "step": 18185, "train_loss": 3.42948579788208, "perplexity": 30.86076999583246, "lr": 0.0026291804804649314, "grad_norm": 0.190295, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:28.319013+00:00", "epoch": 0, "step": 18186, "train_loss": 3.577878952026367, "perplexity": 35.79753198366092, "lr": 0.0026291804804649314, "grad_norm": 0.162735, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:28.623370+00:00", "epoch": 0, "step": 18187, "train_loss": 3.5719363689422607, "perplexity": 35.585433007712126, "lr": 0.0026291804804649314, "grad_norm": 0.182103, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:28.927392+00:00", "epoch": 0, "step": 18188, "train_loss": 3.474545478820801, "perplexity": 32.28315182518941, "lr": 0.0026291804804649314, "grad_norm": 0.175947, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:29.230979+00:00", "epoch": 0, "step": 18189, "train_loss": 3.522770404815674, "perplexity": 33.87815477712853, "lr": 0.0026291804804649314, "grad_norm": 0.168595, "tokens_per_sec": 107936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:29.535510+00:00", "epoch": 0, "step": 18190, "train_loss": 3.509012460708618, "perplexity": 33.415252609853546, "lr": 0.0026291804804649314, "grad_norm": 0.17222, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:29.840770+00:00", "epoch": 0, "step": 18191, "train_loss": 3.4899089336395264, "perplexity": 32.78296214524375, "lr": 0.0026291804804649314, "grad_norm": 0.186104, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:30.146306+00:00", "epoch": 0, "step": 18192, "train_loss": 3.5338714122772217, "perplexity": 34.25633161441156, "lr": 0.0026291804804649314, "grad_norm": 0.180074, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:30.450558+00:00", "epoch": 0, "step": 18193, "train_loss": 3.63197922706604, "perplexity": 37.78753276028272, "lr": 0.0026291804804649314, "grad_norm": 0.177576, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:30.754339+00:00", "epoch": 0, "step": 18194, "train_loss": 3.608457088470459, "perplexity": 36.909061443360976, "lr": 0.0026291804804649314, "grad_norm": 0.191959, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:31.059775+00:00", "epoch": 0, "step": 18195, "train_loss": 3.4718809127807617, "perplexity": 32.19724573722003, "lr": 0.0026291804804649314, "grad_norm": 0.175288, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:31.365478+00:00", "epoch": 0, "step": 18196, "train_loss": 3.5404109954833984, "perplexity": 34.48108785039958, "lr": 0.0026291804804649314, "grad_norm": 0.179295, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:31.670648+00:00", "epoch": 0, "step": 18197, "train_loss": 3.4103565216064453, "perplexity": 30.276036396603637, "lr": 0.0026291804804649314, "grad_norm": 0.181799, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:31.975319+00:00", "epoch": 0, "step": 18198, "train_loss": 3.4619507789611816, "perplexity": 31.8791049828125, "lr": 0.0026291804804649314, "grad_norm": 0.162506, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:32.280091+00:00", "epoch": 0, "step": 18199, "train_loss": 3.459806203842163, "perplexity": 31.81081110430088, "lr": 0.0026291804804649314, "grad_norm": 0.198296, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:32.584931+00:00", "epoch": 0, "step": 18200, "train_loss": 3.5020194053649902, "perplexity": 33.18239304789031, "lr": 0.0026291804804649314, "grad_norm": 0.169141, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:32.890074+00:00", "epoch": 0, "step": 18201, "train_loss": 3.469909191131592, "perplexity": 32.133824276037224, "lr": 0.0026291804804649314, "grad_norm": 0.194243, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:33.196456+00:00", "epoch": 0, "step": 18202, "train_loss": 3.5118088722229004, "perplexity": 33.50882618113234, "lr": 0.0026291804804649314, "grad_norm": 0.199964, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:33.501434+00:00", "epoch": 0, "step": 18203, "train_loss": 3.5173983573913574, "perplexity": 33.69664769186001, "lr": 0.0026291804804649314, "grad_norm": 0.17737, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:33.806372+00:00", "epoch": 0, "step": 18204, "train_loss": 3.3591294288635254, "perplexity": 28.764138747034593, "lr": 0.0026291804804649314, "grad_norm": 0.186249, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:34.110800+00:00", "epoch": 0, "step": 18205, "train_loss": 3.5241596698760986, "perplexity": 33.92525322240905, "lr": 0.0026291804804649314, "grad_norm": 0.18044, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:34.415742+00:00", "epoch": 0, "step": 18206, "train_loss": 3.479710578918457, "perplexity": 32.45032890735008, "lr": 0.0026291804804649314, "grad_norm": 0.17905, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:34.722134+00:00", "epoch": 0, "step": 18207, "train_loss": 3.5084424018859863, "perplexity": 33.3962093786895, "lr": 0.0026291804804649314, "grad_norm": 0.177416, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:35.027702+00:00", "epoch": 0, "step": 18208, "train_loss": 3.4631195068359375, "perplexity": 31.91638476214337, "lr": 0.0026291804804649314, "grad_norm": 0.160201, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:35.332167+00:00", "epoch": 0, "step": 18209, "train_loss": 3.5092689990997314, "perplexity": 33.423826004652156, "lr": 0.0026291804804649314, "grad_norm": 0.155212, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:35.636989+00:00", "epoch": 0, "step": 18210, "train_loss": 3.462728977203369, "perplexity": 31.90392290165058, "lr": 0.0026291804804649314, "grad_norm": 0.168792, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:35.941748+00:00", "epoch": 0, "step": 18211, "train_loss": 3.461176633834839, "perplexity": 31.854435479177926, "lr": 0.0026291804804649314, "grad_norm": 0.182542, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:36.247506+00:00", "epoch": 0, "step": 18212, "train_loss": 3.498108148574829, "perplexity": 33.052861668142356, "lr": 0.0026291804804649314, "grad_norm": 0.180643, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:36.553262+00:00", "epoch": 0, "step": 18213, "train_loss": 3.5876517295837402, "perplexity": 36.149088343885225, "lr": 0.0026291804804649314, "grad_norm": 0.185287, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:36.858315+00:00", "epoch": 0, "step": 18214, "train_loss": 3.512521266937256, "perplexity": 33.53270619677656, "lr": 0.0026291804804649314, "grad_norm": 0.144524, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:37.162737+00:00", "epoch": 0, "step": 18215, "train_loss": 3.4576926231384277, "perplexity": 31.743647390748077, "lr": 0.0026291804804649314, "grad_norm": 0.176757, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:37.467540+00:00", "epoch": 0, "step": 18216, "train_loss": 3.64180850982666, "perplexity": 38.160788517635915, "lr": 0.0026291804804649314, "grad_norm": 0.174049, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:37.773601+00:00", "epoch": 0, "step": 18217, "train_loss": 3.4274847507476807, "perplexity": 30.799077885462847, "lr": 0.0026291804804649314, "grad_norm": 0.177321, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:38.079968+00:00", "epoch": 0, "step": 18218, "train_loss": 3.527881622314453, "perplexity": 34.05175667524243, "lr": 0.0026291804804649314, "grad_norm": 0.177206, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:38.385428+00:00", "epoch": 0, "step": 18219, "train_loss": 3.4150753021240234, "perplexity": 30.419239974703853, "lr": 0.0026291804804649314, "grad_norm": 0.215384, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:38.691906+00:00", "epoch": 0, "step": 18220, "train_loss": 3.3716228008270264, "perplexity": 29.12575402517175, "lr": 0.0026291804804649314, "grad_norm": 0.210919, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:38.997314+00:00", "epoch": 0, "step": 18221, "train_loss": 3.46579909324646, "perplexity": 32.00202215818533, "lr": 0.0026291804804649314, "grad_norm": 0.185421, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:39.302236+00:00", "epoch": 0, "step": 18222, "train_loss": 3.5522117614746094, "perplexity": 34.890401456276706, "lr": 0.0026291804804649314, "grad_norm": 0.171514, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:39.607739+00:00", "epoch": 0, "step": 18223, "train_loss": 3.5334224700927734, "perplexity": 34.240955953714995, "lr": 0.0026291804804649314, "grad_norm": 0.209095, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:39.912719+00:00", "epoch": 0, "step": 18224, "train_loss": 3.6089894771575928, "perplexity": 36.92871664176453, "lr": 0.0026291804804649314, "grad_norm": 0.236191, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:40.216758+00:00", "epoch": 0, "step": 18225, "train_loss": 3.447120428085327, "perplexity": 31.409815138081186, "lr": 0.0026291804804649314, "grad_norm": 0.178288, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:40.522174+00:00", "epoch": 0, "step": 18226, "train_loss": 3.5112929344177246, "perplexity": 33.491542170019216, "lr": 0.0026291804804649314, "grad_norm": 0.166172, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:40.828395+00:00", "epoch": 0, "step": 18227, "train_loss": 3.4516377449035645, "perplexity": 31.552024184091596, "lr": 0.0026291804804649314, "grad_norm": 0.156534, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:41.133273+00:00", "epoch": 0, "step": 18228, "train_loss": 3.4925220012664795, "perplexity": 32.868738262931785, "lr": 0.0026291804804649314, "grad_norm": 0.165051, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:41.438100+00:00", "epoch": 0, "step": 18229, "train_loss": 3.4421591758728027, "perplexity": 31.25436904571326, "lr": 0.0026291804804649314, "grad_norm": 0.178533, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:41.743593+00:00", "epoch": 0, "step": 18230, "train_loss": 3.5133442878723145, "perplexity": 33.56031567601691, "lr": 0.0026291804804649314, "grad_norm": 0.174037, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:42.049333+00:00", "epoch": 0, "step": 18231, "train_loss": 3.5356647968292236, "perplexity": 34.317821511372294, "lr": 0.0026291804804649314, "grad_norm": 0.180653, "tokens_per_sec": 107098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:42.353797+00:00", "epoch": 0, "step": 18232, "train_loss": 3.4760122299194336, "perplexity": 32.330537916901996, "lr": 0.0026291804804649314, "grad_norm": 0.184032, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:42.658494+00:00", "epoch": 0, "step": 18233, "train_loss": 3.482362985610962, "perplexity": 32.53651462617838, "lr": 0.0026291804804649314, "grad_norm": 0.205469, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:42.962584+00:00", "epoch": 0, "step": 18234, "train_loss": 3.5258004665374756, "perplexity": 33.98096335663523, "lr": 0.0026291804804649314, "grad_norm": 0.16934, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:43.268257+00:00", "epoch": 0, "step": 18235, "train_loss": 3.54923939704895, "perplexity": 34.78684844307787, "lr": 0.0026291804804649314, "grad_norm": 0.186162, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:43.574215+00:00", "epoch": 0, "step": 18236, "train_loss": 3.474114179611206, "perplexity": 32.26923112953229, "lr": 0.0026291804804649314, "grad_norm": 0.185555, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:43.878285+00:00", "epoch": 0, "step": 18237, "train_loss": 3.490039587020874, "perplexity": 32.7872456299184, "lr": 0.0026291804804649314, "grad_norm": 0.176205, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:44.183038+00:00", "epoch": 0, "step": 18238, "train_loss": 3.4682860374450684, "perplexity": 32.08170844814271, "lr": 0.0026291804804649314, "grad_norm": 0.171389, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:44.487752+00:00", "epoch": 0, "step": 18239, "train_loss": 3.5514159202575684, "perplexity": 34.86264528293706, "lr": 0.0026291804804649314, "grad_norm": 0.173477, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:44.793452+00:00", "epoch": 0, "step": 18240, "train_loss": 3.5795645713806152, "perplexity": 35.857923880983016, "lr": 0.0026291804804649314, "grad_norm": 0.164731, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:45.098881+00:00", "epoch": 0, "step": 18241, "train_loss": 3.5978760719299316, "perplexity": 36.5205849164236, "lr": 0.0026291804804649314, "grad_norm": 0.161275, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:45.404996+00:00", "epoch": 0, "step": 18242, "train_loss": 3.5120906829833984, "perplexity": 33.518270659637174, "lr": 0.0026291804804649314, "grad_norm": 0.184776, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:45.709444+00:00", "epoch": 0, "step": 18243, "train_loss": 3.478538990020752, "perplexity": 32.41233272457449, "lr": 0.0026291804804649314, "grad_norm": 0.195262, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:46.013869+00:00", "epoch": 0, "step": 18244, "train_loss": 3.4188075065612793, "perplexity": 30.53298292104951, "lr": 0.0026291804804649314, "grad_norm": 0.212246, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:46.318492+00:00", "epoch": 0, "step": 18245, "train_loss": 3.4509196281433105, "perplexity": 31.529374280316187, "lr": 0.0026291804804649314, "grad_norm": 0.170946, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:46.624751+00:00", "epoch": 0, "step": 18246, "train_loss": 3.6495165824890137, "perplexity": 38.45607121663059, "lr": 0.0026291804804649314, "grad_norm": 0.207191, "tokens_per_sec": 106995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:46.929441+00:00", "epoch": 0, "step": 18247, "train_loss": 3.5343756675720215, "perplexity": 34.273609906977676, "lr": 0.0026291804804649314, "grad_norm": 0.188096, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:47.234191+00:00", "epoch": 0, "step": 18248, "train_loss": 3.470597743988037, "perplexity": 32.15595773168359, "lr": 0.0026291804804649314, "grad_norm": 0.179541, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:47.538249+00:00", "epoch": 0, "step": 18249, "train_loss": 3.4664876461029053, "perplexity": 32.024064829855774, "lr": 0.0026291804804649314, "grad_norm": 0.195543, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:47.842558+00:00", "epoch": 0, "step": 18250, "train_loss": 3.4739391803741455, "perplexity": 32.26358453279356, "lr": 0.0026291804804649314, "grad_norm": 0.185671, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:48.147961+00:00", "epoch": 0, "step": 18251, "train_loss": 3.468474864959717, "perplexity": 32.08776692940063, "lr": 0.0026291804804649314, "grad_norm": 0.187603, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:48.453958+00:00", "epoch": 0, "step": 18252, "train_loss": 3.4722087383270264, "perplexity": 32.20780254719351, "lr": 0.0026291804804649314, "grad_norm": 0.177498, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:48.758496+00:00", "epoch": 0, "step": 18253, "train_loss": 3.4770562648773193, "perplexity": 32.36430975511671, "lr": 0.0026291804804649314, "grad_norm": 0.196912, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:49.063831+00:00", "epoch": 0, "step": 18254, "train_loss": 3.52051043510437, "perplexity": 33.80167762403251, "lr": 0.0026291804804649314, "grad_norm": 0.152105, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:49.368265+00:00", "epoch": 0, "step": 18255, "train_loss": 3.419813632965088, "perplexity": 30.56371842065949, "lr": 0.0026291804804649314, "grad_norm": 0.176739, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:49.672942+00:00", "epoch": 0, "step": 18256, "train_loss": 3.4002127647399902, "perplexity": 29.970476029620688, "lr": 0.0026291804804649314, "grad_norm": 0.17139, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:49.978511+00:00", "epoch": 0, "step": 18257, "train_loss": 3.4210078716278076, "perplexity": 30.600240598631732, "lr": 0.0026291804804649314, "grad_norm": 0.161244, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:50.284125+00:00", "epoch": 0, "step": 18258, "train_loss": 3.4764535427093506, "perplexity": 32.34480894555174, "lr": 0.0026291804804649314, "grad_norm": 0.153878, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:50.589351+00:00", "epoch": 0, "step": 18259, "train_loss": 3.5644073486328125, "perplexity": 35.31851563391594, "lr": 0.0026291804804649314, "grad_norm": 0.163624, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:50.895200+00:00", "epoch": 0, "step": 18260, "train_loss": 3.5274932384490967, "perplexity": 34.03853409024826, "lr": 0.0026291804804649314, "grad_norm": 0.157052, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:51.200447+00:00", "epoch": 0, "step": 18261, "train_loss": 3.460636615753174, "perplexity": 31.837238151889707, "lr": 0.0026291804804649314, "grad_norm": 0.162932, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:51.505060+00:00", "epoch": 0, "step": 18262, "train_loss": 3.47629976272583, "perplexity": 32.33983534379512, "lr": 0.0026291804804649314, "grad_norm": 0.166459, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:51.810671+00:00", "epoch": 0, "step": 18263, "train_loss": 3.4281554222106934, "perplexity": 30.81974087635238, "lr": 0.0026291804804649314, "grad_norm": 0.180253, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:52.115568+00:00", "epoch": 0, "step": 18264, "train_loss": 3.525912046432495, "perplexity": 33.98475516049983, "lr": 0.0026291804804649314, "grad_norm": 0.164766, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:52.420564+00:00", "epoch": 0, "step": 18265, "train_loss": 3.4304447174072266, "perplexity": 30.890377183929175, "lr": 0.0026291804804649314, "grad_norm": 0.168746, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:52.725431+00:00", "epoch": 0, "step": 18266, "train_loss": 3.4626927375793457, "perplexity": 31.902766736429374, "lr": 0.0026291804804649314, "grad_norm": 0.171933, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:53.031075+00:00", "epoch": 0, "step": 18267, "train_loss": 3.5384819507598877, "perplexity": 34.41463640433323, "lr": 0.0026291804804649314, "grad_norm": 0.151524, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:53.335560+00:00", "epoch": 0, "step": 18268, "train_loss": 3.4426074028015137, "perplexity": 31.26838123564525, "lr": 0.0026291804804649314, "grad_norm": 0.164739, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:53.640678+00:00", "epoch": 0, "step": 18269, "train_loss": 3.4429211616516113, "perplexity": 31.278193506249035, "lr": 0.0026291804804649314, "grad_norm": 0.152901, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:53.945216+00:00", "epoch": 0, "step": 18270, "train_loss": 3.5693211555480957, "perplexity": 35.49249109111756, "lr": 0.0026291804804649314, "grad_norm": 0.154038, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:54.249660+00:00", "epoch": 0, "step": 18271, "train_loss": 3.3737337589263916, "perplexity": 29.18730221149689, "lr": 0.0026291804804649314, "grad_norm": 0.175512, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:54.554483+00:00", "epoch": 0, "step": 18272, "train_loss": 3.4252161979675293, "perplexity": 30.72928774294087, "lr": 0.0026291804804649314, "grad_norm": 0.185825, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:54.859177+00:00", "epoch": 0, "step": 18273, "train_loss": 3.59513258934021, "perplexity": 36.42052864162066, "lr": 0.0026291804804649314, "grad_norm": 0.165896, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:55.163618+00:00", "epoch": 0, "step": 18274, "train_loss": 3.597111463546753, "perplexity": 36.49267164375896, "lr": 0.0026291804804649314, "grad_norm": 0.142059, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:55.468931+00:00", "epoch": 0, "step": 18275, "train_loss": 3.5389769077301025, "perplexity": 34.4316743846847, "lr": 0.0026291804804649314, "grad_norm": 0.173502, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:55.773508+00:00", "epoch": 0, "step": 18276, "train_loss": 3.514927387237549, "perplexity": 33.6134870671536, "lr": 0.0026291804804649314, "grad_norm": 0.180774, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:56.078340+00:00", "epoch": 0, "step": 18277, "train_loss": 3.440192937850952, "perplexity": 31.192975893501007, "lr": 0.0026291804804649314, "grad_norm": 0.16398, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:56.384453+00:00", "epoch": 0, "step": 18278, "train_loss": 3.503253698348999, "perplexity": 33.22337512956693, "lr": 0.0026291804804649314, "grad_norm": 0.15655, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:56.689994+00:00", "epoch": 0, "step": 18279, "train_loss": 3.515186071395874, "perplexity": 33.62218346852608, "lr": 0.0026291804804649314, "grad_norm": 0.166213, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:56.995589+00:00", "epoch": 0, "step": 18280, "train_loss": 3.5600156784057617, "perplexity": 35.16374845230113, "lr": 0.0026291804804649314, "grad_norm": 0.157674, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:57.300212+00:00", "epoch": 0, "step": 18281, "train_loss": 3.436817169189453, "perplexity": 31.08785315786902, "lr": 0.0026291804804649314, "grad_norm": 0.163801, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:57.605380+00:00", "epoch": 0, "step": 18282, "train_loss": 3.552248477935791, "perplexity": 34.89168253186553, "lr": 0.0026291804804649314, "grad_norm": 0.164005, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:57.910696+00:00", "epoch": 0, "step": 18283, "train_loss": 3.5241658687591553, "perplexity": 33.925463521738244, "lr": 0.0026291804804649314, "grad_norm": 0.151097, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:58.216683+00:00", "epoch": 0, "step": 18284, "train_loss": 3.593804359436035, "perplexity": 36.372185918585366, "lr": 0.0026291804804649314, "grad_norm": 0.170763, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:58.521724+00:00", "epoch": 0, "step": 18285, "train_loss": 3.4826834201812744, "perplexity": 32.54694212083775, "lr": 0.0026291804804649314, "grad_norm": 0.166273, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:58.826343+00:00", "epoch": 0, "step": 18286, "train_loss": 3.561234712600708, "perplexity": 35.20664040215214, "lr": 0.0026291804804649314, "grad_norm": 0.160954, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:59.130270+00:00", "epoch": 0, "step": 18287, "train_loss": 3.474574089050293, "perplexity": 32.28407546678459, "lr": 0.0026291804804649314, "grad_norm": 0.161163, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:59.434882+00:00", "epoch": 0, "step": 18288, "train_loss": 3.485234260559082, "perplexity": 32.63007015306963, "lr": 0.0026291804804649314, "grad_norm": 0.178599, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:33:59.741403+00:00", "epoch": 0, "step": 18289, "train_loss": 3.495131254196167, "perplexity": 32.95461310040401, "lr": 0.0026291804804649314, "grad_norm": 0.18301, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:00.046238+00:00", "epoch": 0, "step": 18290, "train_loss": 3.4253904819488525, "perplexity": 30.734643832278667, "lr": 0.0026291804804649314, "grad_norm": 0.196933, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:00.350972+00:00", "epoch": 0, "step": 18291, "train_loss": 3.401651620864868, "perplexity": 30.01363027154886, "lr": 0.0026291804804649314, "grad_norm": 0.182928, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:00.655063+00:00", "epoch": 0, "step": 18292, "train_loss": 3.363083839416504, "perplexity": 28.878109155219285, "lr": 0.0026291804804649314, "grad_norm": 0.165649, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:00.959336+00:00", "epoch": 0, "step": 18293, "train_loss": 3.485428810119629, "perplexity": 32.63641893643501, "lr": 0.0026291804804649314, "grad_norm": 0.160066, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:01.265119+00:00", "epoch": 0, "step": 18294, "train_loss": 3.537672996520996, "perplexity": 34.386807795887016, "lr": 0.0026291804804649314, "grad_norm": 0.177637, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:01.570086+00:00", "epoch": 0, "step": 18295, "train_loss": 3.5238094329833984, "perplexity": 33.913373427629004, "lr": 0.0026291804804649314, "grad_norm": 0.177273, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:01.875389+00:00", "epoch": 0, "step": 18296, "train_loss": 3.490424633026123, "perplexity": 32.799872658708814, "lr": 0.0026291804804649314, "grad_norm": 0.183798, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:02.181160+00:00", "epoch": 0, "step": 18297, "train_loss": 3.547119379043579, "perplexity": 34.713177817147724, "lr": 0.0026291804804649314, "grad_norm": 0.180346, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:02.486052+00:00", "epoch": 0, "step": 18298, "train_loss": 3.4898929595947266, "perplexity": 32.78243847292037, "lr": 0.0026291804804649314, "grad_norm": 0.165485, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:02.792271+00:00", "epoch": 0, "step": 18299, "train_loss": 3.501664161682129, "perplexity": 33.170607305906955, "lr": 0.0026291804804649314, "grad_norm": 0.166247, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:03.097603+00:00", "epoch": 0, "step": 18300, "train_loss": 3.5904572010040283, "perplexity": 36.25064597000172, "lr": 0.0026291804804649314, "grad_norm": 0.179005, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:03.404028+00:00", "epoch": 0, "step": 18301, "train_loss": 3.477543354034424, "perplexity": 32.38007789940989, "lr": 0.0026291804804649314, "grad_norm": 0.161461, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:03.710225+00:00", "epoch": 0, "step": 18302, "train_loss": 3.5337202548980713, "perplexity": 34.251153908439505, "lr": 0.0026291804804649314, "grad_norm": 0.155791, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:04.015119+00:00", "epoch": 0, "step": 18303, "train_loss": 3.415928840637207, "perplexity": 30.445215051351223, "lr": 0.0026291804804649314, "grad_norm": 0.152757, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:04.320154+00:00", "epoch": 0, "step": 18304, "train_loss": 3.550916910171509, "perplexity": 34.84525281118448, "lr": 0.0026291804804649314, "grad_norm": 0.157831, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:04.626015+00:00", "epoch": 0, "step": 18305, "train_loss": 3.507533311843872, "perplexity": 33.365863013178284, "lr": 0.0026291804804649314, "grad_norm": 0.169362, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:04.931107+00:00", "epoch": 0, "step": 18306, "train_loss": 3.4740054607391357, "perplexity": 32.26572304582221, "lr": 0.0026291804804649314, "grad_norm": 0.152572, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:05.235670+00:00", "epoch": 0, "step": 18307, "train_loss": 3.5187740325927734, "perplexity": 33.74303523424078, "lr": 0.0026291804804649314, "grad_norm": 0.157955, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:05.540597+00:00", "epoch": 0, "step": 18308, "train_loss": 3.4967880249023438, "perplexity": 33.00925659139053, "lr": 0.0026291804804649314, "grad_norm": 0.169567, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:05.845877+00:00", "epoch": 0, "step": 18309, "train_loss": 3.4601194858551025, "perplexity": 31.82077842044578, "lr": 0.0026291804804649314, "grad_norm": 0.190892, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:06.151584+00:00", "epoch": 0, "step": 18310, "train_loss": 3.5058367252349854, "perplexity": 33.309302929903026, "lr": 0.0026291804804649314, "grad_norm": 0.171315, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:06.456883+00:00", "epoch": 0, "step": 18311, "train_loss": 3.4865000247955322, "perplexity": 32.67139827921641, "lr": 0.0026291804804649314, "grad_norm": 0.177831, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:06.762193+00:00", "epoch": 0, "step": 18312, "train_loss": 3.4377200603485107, "perplexity": 31.11593478105629, "lr": 0.0026291804804649314, "grad_norm": 0.183773, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:07.067310+00:00", "epoch": 0, "step": 18313, "train_loss": 3.4912126064300537, "perplexity": 32.82572827144596, "lr": 0.0026291804804649314, "grad_norm": 0.209359, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:07.372492+00:00", "epoch": 0, "step": 18314, "train_loss": 3.5135209560394287, "perplexity": 33.566245239242285, "lr": 0.0026291804804649314, "grad_norm": 0.175704, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:07.677846+00:00", "epoch": 0, "step": 18315, "train_loss": 3.528832197189331, "perplexity": 34.08414080891677, "lr": 0.0026291804804649314, "grad_norm": 0.182887, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:07.982936+00:00", "epoch": 0, "step": 18316, "train_loss": 3.5425655841827393, "perplexity": 34.55546050509232, "lr": 0.0026291804804649314, "grad_norm": 0.200662, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:08.287685+00:00", "epoch": 0, "step": 18317, "train_loss": 3.4347572326660156, "perplexity": 31.023880066589108, "lr": 0.0026291804804649314, "grad_norm": 0.175217, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:08.592767+00:00", "epoch": 0, "step": 18318, "train_loss": 3.4694631099700928, "perplexity": 32.119493179034606, "lr": 0.0026291804804649314, "grad_norm": 0.22436, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:08.897861+00:00", "epoch": 0, "step": 18319, "train_loss": 3.44728422164917, "perplexity": 31.41496028500177, "lr": 0.0026291804804649314, "grad_norm": 0.189785, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:09.203713+00:00", "epoch": 0, "step": 18320, "train_loss": 3.5606260299682617, "perplexity": 35.18521725220365, "lr": 0.0026291804804649314, "grad_norm": 0.194631, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:09.508707+00:00", "epoch": 0, "step": 18321, "train_loss": 3.489254951477051, "perplexity": 32.76152968175593, "lr": 0.0026291804804649314, "grad_norm": 0.184172, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:09.813760+00:00", "epoch": 0, "step": 18322, "train_loss": 3.4878852367401123, "perplexity": 32.71668644997483, "lr": 0.0026291804804649314, "grad_norm": 0.157719, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:10.118449+00:00", "epoch": 0, "step": 18323, "train_loss": 3.5733532905578613, "perplexity": 35.63589051566628, "lr": 0.0026291804804649314, "grad_norm": 0.174227, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:10.423432+00:00", "epoch": 0, "step": 18324, "train_loss": 3.3798038959503174, "perplexity": 29.365011950847816, "lr": 0.0026291804804649314, "grad_norm": 0.151401, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:10.728367+00:00", "epoch": 0, "step": 18325, "train_loss": 3.555562734603882, "perplexity": 35.007514365543656, "lr": 0.0026291804804649314, "grad_norm": 0.181161, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:11.034143+00:00", "epoch": 0, "step": 18326, "train_loss": 3.523061990737915, "perplexity": 33.888034610473206, "lr": 0.0026291804804649314, "grad_norm": 0.182091, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:11.339901+00:00", "epoch": 0, "step": 18327, "train_loss": 3.5087766647338867, "perplexity": 33.407374356660206, "lr": 0.0026291804804649314, "grad_norm": 0.146807, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:11.646072+00:00", "epoch": 0, "step": 18328, "train_loss": 3.6085047721862793, "perplexity": 36.91082144651944, "lr": 0.0026291804804649314, "grad_norm": 0.188053, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:11.951952+00:00", "epoch": 0, "step": 18329, "train_loss": 3.4827184677124023, "perplexity": 32.548082830794264, "lr": 0.0026291804804649314, "grad_norm": 0.190958, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:12.256764+00:00", "epoch": 0, "step": 18330, "train_loss": 3.4628448486328125, "perplexity": 31.907619868984355, "lr": 0.0026291804804649314, "grad_norm": 0.188579, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:12.562641+00:00", "epoch": 0, "step": 18331, "train_loss": 3.4340450763702393, "perplexity": 31.001794080350788, "lr": 0.0026291804804649314, "grad_norm": 0.176632, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:12.868391+00:00", "epoch": 0, "step": 18332, "train_loss": 3.409714698791504, "perplexity": 30.25661078026702, "lr": 0.0026291804804649314, "grad_norm": 0.185331, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:13.172883+00:00", "epoch": 0, "step": 18333, "train_loss": 3.555743455886841, "perplexity": 35.01384154016334, "lr": 0.0026291804804649314, "grad_norm": 0.154984, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:13.477176+00:00", "epoch": 0, "step": 18334, "train_loss": 3.502532958984375, "perplexity": 33.19943836240675, "lr": 0.0026291804804649314, "grad_norm": 0.158867, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:13.781901+00:00", "epoch": 0, "step": 18335, "train_loss": 3.435992956161499, "perplexity": 31.062240700795517, "lr": 0.0026291804804649314, "grad_norm": 0.180512, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:14.085695+00:00", "epoch": 0, "step": 18336, "train_loss": 3.4998884201049805, "perplexity": 33.111757146176465, "lr": 0.0026291804804649314, "grad_norm": 0.165306, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:14.391197+00:00", "epoch": 0, "step": 18337, "train_loss": 3.450603485107422, "perplexity": 31.51940806366954, "lr": 0.0026291804804649314, "grad_norm": 0.146534, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:14.696562+00:00", "epoch": 0, "step": 18338, "train_loss": 3.5420424938201904, "perplexity": 34.53738960349795, "lr": 0.0026291804804649314, "grad_norm": 0.178757, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:15.001262+00:00", "epoch": 0, "step": 18339, "train_loss": 3.4465582370758057, "perplexity": 31.39216178514294, "lr": 0.0026291804804649314, "grad_norm": 0.174014, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:15.306157+00:00", "epoch": 0, "step": 18340, "train_loss": 3.445753574371338, "perplexity": 31.36691184354737, "lr": 0.0026291804804649314, "grad_norm": 0.158108, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:15.611077+00:00", "epoch": 0, "step": 18341, "train_loss": 3.516098737716675, "perplexity": 33.6528833102344, "lr": 0.0026291804804649314, "grad_norm": 0.176031, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:15.916884+00:00", "epoch": 0, "step": 18342, "train_loss": 3.4973018169403076, "perplexity": 33.026220842281624, "lr": 0.0026291804804649314, "grad_norm": 0.223335, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:16.222138+00:00", "epoch": 0, "step": 18343, "train_loss": 3.648146390914917, "perplexity": 38.40341511458853, "lr": 0.0026291804804649314, "grad_norm": 0.209592, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:16.527189+00:00", "epoch": 0, "step": 18344, "train_loss": 3.601433515548706, "perplexity": 36.65073620386219, "lr": 0.0026291804804649314, "grad_norm": 0.201405, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:16.830973+00:00", "epoch": 0, "step": 18345, "train_loss": 3.4959359169006348, "perplexity": 32.981141120128754, "lr": 0.0026291804804649314, "grad_norm": 0.170818, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:17.135171+00:00", "epoch": 0, "step": 18346, "train_loss": 3.500338315963745, "perplexity": 33.1266573401094, "lr": 0.0026291804804649314, "grad_norm": 0.20429, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:17.440696+00:00", "epoch": 0, "step": 18347, "train_loss": 3.3794403076171875, "perplexity": 29.354337115840597, "lr": 0.0026291804804649314, "grad_norm": 0.179153, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:17.745964+00:00", "epoch": 0, "step": 18348, "train_loss": 3.541252613067627, "perplexity": 34.510119955509225, "lr": 0.0026291804804649314, "grad_norm": 0.183856, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:18.051002+00:00", "epoch": 0, "step": 18349, "train_loss": 3.5698986053466797, "perplexity": 35.512992141543414, "lr": 0.0026291804804649314, "grad_norm": 0.172937, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:18.355988+00:00", "epoch": 0, "step": 18350, "train_loss": 3.5305378437042236, "perplexity": 34.14232591240395, "lr": 0.0026291804804649314, "grad_norm": 0.162158, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:18.660137+00:00", "epoch": 0, "step": 18351, "train_loss": 3.483682632446289, "perplexity": 32.57947967784944, "lr": 0.0026291804804649314, "grad_norm": 0.168611, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:18.964394+00:00", "epoch": 0, "step": 18352, "train_loss": 3.612847328186035, "perplexity": 37.07145728828078, "lr": 0.0026291804804649314, "grad_norm": 0.173057, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:19.269509+00:00", "epoch": 0, "step": 18353, "train_loss": 3.4923911094665527, "perplexity": 32.864436296171455, "lr": 0.0026291804804649314, "grad_norm": 0.186571, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:19.575095+00:00", "epoch": 0, "step": 18354, "train_loss": 3.654355764389038, "perplexity": 38.64261814362323, "lr": 0.0026291804804649314, "grad_norm": 0.157082, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:19.880584+00:00", "epoch": 0, "step": 18355, "train_loss": 3.560525894165039, "perplexity": 35.181694128610886, "lr": 0.0026291804804649314, "grad_norm": 0.17871, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:20.185647+00:00", "epoch": 0, "step": 18356, "train_loss": 3.5435595512390137, "perplexity": 34.589824569993766, "lr": 0.0026291804804649314, "grad_norm": 0.180085, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:20.490134+00:00", "epoch": 0, "step": 18357, "train_loss": 3.4426257610321045, "perplexity": 31.268955273067316, "lr": 0.0026291804804649314, "grad_norm": 0.158978, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:20.795026+00:00", "epoch": 0, "step": 18358, "train_loss": 3.4775171279907227, "perplexity": 32.379228709207354, "lr": 0.0026291804804649314, "grad_norm": 0.211453, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:21.100890+00:00", "epoch": 0, "step": 18359, "train_loss": 3.4533910751342773, "perplexity": 31.607393828384726, "lr": 0.0026291804804649314, "grad_norm": 0.224768, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:21.407075+00:00", "epoch": 0, "step": 18360, "train_loss": 3.4900126457214355, "perplexity": 32.786362310815065, "lr": 0.0026291804804649314, "grad_norm": 0.185451, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:21.711433+00:00", "epoch": 0, "step": 18361, "train_loss": 3.5098915100097656, "perplexity": 33.44463917853692, "lr": 0.0026291804804649314, "grad_norm": 0.161603, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:22.016695+00:00", "epoch": 0, "step": 18362, "train_loss": 3.5069117546081543, "perplexity": 33.34513066343494, "lr": 0.0026291804804649314, "grad_norm": 0.170812, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:22.321174+00:00", "epoch": 0, "step": 18363, "train_loss": 3.4788575172424316, "perplexity": 32.42265857931101, "lr": 0.0026291804804649314, "grad_norm": 0.174681, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:22.625646+00:00", "epoch": 0, "step": 18364, "train_loss": 3.5242724418640137, "perplexity": 33.92907925638613, "lr": 0.0026291804804649314, "grad_norm": 0.171055, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:22.931011+00:00", "epoch": 0, "step": 18365, "train_loss": 3.590393304824829, "perplexity": 36.24832976622981, "lr": 0.0026291804804649314, "grad_norm": 0.176566, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:23.235965+00:00", "epoch": 0, "step": 18366, "train_loss": 3.5201282501220703, "perplexity": 33.788761598780724, "lr": 0.0026291804804649314, "grad_norm": 0.162066, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:23.540701+00:00", "epoch": 0, "step": 18367, "train_loss": 3.3406178951263428, "perplexity": 28.236568554331324, "lr": 0.0026291804804649314, "grad_norm": 0.174513, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:23.846121+00:00", "epoch": 0, "step": 18368, "train_loss": 3.504028081893921, "perplexity": 33.24911272867448, "lr": 0.0026291804804649314, "grad_norm": 0.166604, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:24.151445+00:00", "epoch": 0, "step": 18369, "train_loss": 3.505445957183838, "perplexity": 33.296289261340576, "lr": 0.0026291804804649314, "grad_norm": 0.168777, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:24.458111+00:00", "epoch": 0, "step": 18370, "train_loss": 3.459350109100342, "perplexity": 31.79630566880102, "lr": 0.0026291804804649314, "grad_norm": 0.159816, "tokens_per_sec": 106915} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:24.762853+00:00", "epoch": 0, "step": 18371, "train_loss": 3.6484127044677734, "perplexity": 38.413643826471365, "lr": 0.0026291804804649314, "grad_norm": 0.201434, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:25.067221+00:00", "epoch": 0, "step": 18372, "train_loss": 3.4052083492279053, "perplexity": 30.12057066784792, "lr": 0.0026291804804649314, "grad_norm": 0.18642, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:25.371949+00:00", "epoch": 0, "step": 18373, "train_loss": 3.537343978881836, "perplexity": 34.37549579059444, "lr": 0.0026291804804649314, "grad_norm": 0.168599, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:25.677302+00:00", "epoch": 0, "step": 18374, "train_loss": 3.4481818675994873, "perplexity": 31.443172457257695, "lr": 0.0026291804804649314, "grad_norm": 0.158814, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:25.982931+00:00", "epoch": 0, "step": 18375, "train_loss": 3.534853219985962, "perplexity": 34.28998126090187, "lr": 0.0026291804804649314, "grad_norm": 0.187531, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:26.287144+00:00", "epoch": 0, "step": 18376, "train_loss": 3.526937484741211, "perplexity": 34.019622304350946, "lr": 0.0026291804804649314, "grad_norm": 0.229211, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:26.592771+00:00", "epoch": 0, "step": 18377, "train_loss": 3.43400502204895, "perplexity": 31.000552349398667, "lr": 0.0026291804804649314, "grad_norm": 0.204715, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:26.896871+00:00", "epoch": 0, "step": 18378, "train_loss": 3.6004090309143066, "perplexity": 36.61320731494935, "lr": 0.0026291804804649314, "grad_norm": 0.182365, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:27.201341+00:00", "epoch": 0, "step": 18379, "train_loss": 3.618157148361206, "perplexity": 37.268823586173276, "lr": 0.0026291804804649314, "grad_norm": 0.211731, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:27.506601+00:00", "epoch": 0, "step": 18380, "train_loss": 3.5024573802948, "perplexity": 33.19692928717819, "lr": 0.0026291804804649314, "grad_norm": 0.188848, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:27.812848+00:00", "epoch": 0, "step": 18381, "train_loss": 3.5242154598236084, "perplexity": 33.92714596330305, "lr": 0.0026291804804649314, "grad_norm": 0.182345, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:28.117786+00:00", "epoch": 0, "step": 18382, "train_loss": 3.569550037384033, "perplexity": 35.500615607382116, "lr": 0.0026291804804649314, "grad_norm": 0.177744, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:28.421841+00:00", "epoch": 0, "step": 18383, "train_loss": 3.49177622795105, "perplexity": 32.84423477318354, "lr": 0.0026291804804649314, "grad_norm": 0.179374, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:28.725645+00:00", "epoch": 0, "step": 18384, "train_loss": 3.4704861640930176, "perplexity": 32.1523699734602, "lr": 0.0026291804804649314, "grad_norm": 0.187829, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:29.030540+00:00", "epoch": 0, "step": 18385, "train_loss": 3.5258140563964844, "perplexity": 33.98142515627412, "lr": 0.0026291804804649314, "grad_norm": 0.182715, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:29.335694+00:00", "epoch": 0, "step": 18386, "train_loss": 3.482267141342163, "perplexity": 32.53339633716222, "lr": 0.0026291804804649314, "grad_norm": 0.208706, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:29.640479+00:00", "epoch": 0, "step": 18387, "train_loss": 3.55122971534729, "perplexity": 34.856154291546076, "lr": 0.0026291804804649314, "grad_norm": 0.175618, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:29.944901+00:00", "epoch": 0, "step": 18388, "train_loss": 3.5391340255737305, "perplexity": 34.43708464012922, "lr": 0.0026291804804649314, "grad_norm": 0.215281, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:30.249701+00:00", "epoch": 0, "step": 18389, "train_loss": 3.556485176086426, "perplexity": 35.039821647492865, "lr": 0.0026291804804649314, "grad_norm": 0.173554, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:30.554079+00:00", "epoch": 0, "step": 18390, "train_loss": 3.3796653747558594, "perplexity": 29.360944556033807, "lr": 0.0026291804804649314, "grad_norm": 0.225013, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:30.858609+00:00", "epoch": 0, "step": 18391, "train_loss": 3.5069191455841064, "perplexity": 33.34537711740456, "lr": 0.0026291804804649314, "grad_norm": 0.204026, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:31.163721+00:00", "epoch": 0, "step": 18392, "train_loss": 3.6433050632476807, "perplexity": 38.217940931394104, "lr": 0.0026291804804649314, "grad_norm": 0.189942, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:31.469873+00:00", "epoch": 0, "step": 18393, "train_loss": 3.5261244773864746, "perplexity": 33.991975341327134, "lr": 0.0026291804804649314, "grad_norm": 0.162945, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:31.774788+00:00", "epoch": 0, "step": 18394, "train_loss": 3.553025484085083, "perplexity": 34.91880411920747, "lr": 0.0026291804804649314, "grad_norm": 0.192761, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:32.079135+00:00", "epoch": 0, "step": 18395, "train_loss": 3.5123984813690186, "perplexity": 33.52858911715581, "lr": 0.0026291804804649314, "grad_norm": 0.157678, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:32.385122+00:00", "epoch": 0, "step": 18396, "train_loss": 3.4880964756011963, "perplexity": 32.72359821555001, "lr": 0.0026291804804649314, "grad_norm": 0.187754, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:32.690288+00:00", "epoch": 0, "step": 18397, "train_loss": 3.5040123462677, "perplexity": 33.2485895371808, "lr": 0.0026291804804649314, "grad_norm": 0.197179, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:32.995448+00:00", "epoch": 0, "step": 18398, "train_loss": 3.4053776264190674, "perplexity": 30.1256698250201, "lr": 0.0026291804804649314, "grad_norm": 0.202459, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:33.298955+00:00", "epoch": 0, "step": 18399, "train_loss": 3.529139995574951, "perplexity": 34.0946334671618, "lr": 0.0026291804804649314, "grad_norm": 0.163354, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:33.603828+00:00", "epoch": 0, "step": 18400, "train_loss": 3.5203042030334473, "perplexity": 33.794707352826876, "lr": 0.0026291804804649314, "grad_norm": 0.190644, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:33.908089+00:00", "epoch": 0, "step": 18401, "train_loss": 3.499626636505127, "perplexity": 33.10309016567944, "lr": 0.0026291804804649314, "grad_norm": 0.178927, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:34.213264+00:00", "epoch": 0, "step": 18402, "train_loss": 3.543877363204956, "perplexity": 34.60081937719006, "lr": 0.0026291804804649314, "grad_norm": 0.154776, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:34.518250+00:00", "epoch": 0, "step": 18403, "train_loss": 3.621957302093506, "perplexity": 37.41072028910251, "lr": 0.0026291804804649314, "grad_norm": 0.177238, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:34.823629+00:00", "epoch": 0, "step": 18404, "train_loss": 3.496119976043701, "perplexity": 32.98721215939868, "lr": 0.0026291804804649314, "grad_norm": 0.154888, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:35.127997+00:00", "epoch": 0, "step": 18405, "train_loss": 3.534341335296631, "perplexity": 34.27243323616273, "lr": 0.0026291804804649314, "grad_norm": 0.162708, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:35.431896+00:00", "epoch": 0, "step": 18406, "train_loss": 3.4387903213500977, "perplexity": 31.149254779947512, "lr": 0.0026291804804649314, "grad_norm": 0.169071, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:35.737069+00:00", "epoch": 0, "step": 18407, "train_loss": 3.489509344100952, "perplexity": 32.76986503343574, "lr": 0.0026291804804649314, "grad_norm": 0.184025, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:36.041856+00:00", "epoch": 0, "step": 18408, "train_loss": 3.436770439147949, "perplexity": 31.08640045514338, "lr": 0.0026291804804649314, "grad_norm": 0.152163, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:36.347233+00:00", "epoch": 0, "step": 18409, "train_loss": 3.45107364654541, "perplexity": 31.53423075814592, "lr": 0.0026291804804649314, "grad_norm": 0.202552, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:36.652423+00:00", "epoch": 0, "step": 18410, "train_loss": 3.599313259124756, "perplexity": 36.57310956825065, "lr": 0.0026291804804649314, "grad_norm": 0.209592, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:36.956758+00:00", "epoch": 0, "step": 18411, "train_loss": 3.565016508102417, "perplexity": 35.340036796411184, "lr": 0.0026291804804649314, "grad_norm": 0.178622, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:37.261624+00:00", "epoch": 0, "step": 18412, "train_loss": 3.4579243659973145, "perplexity": 31.751004606804116, "lr": 0.0026291804804649314, "grad_norm": 0.178478, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:37.567090+00:00", "epoch": 0, "step": 18413, "train_loss": 3.4163143634796143, "perplexity": 30.456954639989906, "lr": 0.0026291804804649314, "grad_norm": 0.137088, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:37.872754+00:00", "epoch": 0, "step": 18414, "train_loss": 3.497382164001465, "perplexity": 33.028874508673155, "lr": 0.0026291804804649314, "grad_norm": 0.182497, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:38.177683+00:00", "epoch": 0, "step": 18415, "train_loss": 3.5245487689971924, "perplexity": 33.93845607706376, "lr": 0.0026291804804649314, "grad_norm": 0.180565, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:38.482435+00:00", "epoch": 0, "step": 18416, "train_loss": 3.5198824405670166, "perplexity": 33.78045701904065, "lr": 0.0026291804804649314, "grad_norm": 0.200317, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:38.786425+00:00", "epoch": 0, "step": 18417, "train_loss": 3.420938491821289, "perplexity": 30.598117633505883, "lr": 0.0026291804804649314, "grad_norm": 0.217404, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:39.093126+00:00", "epoch": 0, "step": 18418, "train_loss": 3.5265235900878906, "perplexity": 34.005544678099284, "lr": 0.0026291804804649314, "grad_norm": 0.174748, "tokens_per_sec": 106841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:39.400552+00:00", "epoch": 0, "step": 18419, "train_loss": 3.5068020820617676, "perplexity": 33.341473818576894, "lr": 0.0026291804804649314, "grad_norm": 0.194336, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:39.707830+00:00", "epoch": 0, "step": 18420, "train_loss": 3.5452353954315186, "perplexity": 34.64784032572493, "lr": 0.0026291804804649314, "grad_norm": 0.181547, "tokens_per_sec": 106699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:40.012572+00:00", "epoch": 0, "step": 18421, "train_loss": 3.470614433288574, "perplexity": 32.1564943966045, "lr": 0.0026291804804649314, "grad_norm": 0.1875, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:40.317986+00:00", "epoch": 0, "step": 18422, "train_loss": 3.516482353210449, "perplexity": 33.665795554192464, "lr": 0.0026291804804649314, "grad_norm": 0.195259, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:40.623621+00:00", "epoch": 0, "step": 18423, "train_loss": 3.5161540508270264, "perplexity": 33.654744807364715, "lr": 0.0026291804804649314, "grad_norm": 0.15548, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:40.930130+00:00", "epoch": 0, "step": 18424, "train_loss": 3.5642127990722656, "perplexity": 35.31164510057149, "lr": 0.0026291804804649314, "grad_norm": 0.187423, "tokens_per_sec": 106906} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:41.236134+00:00", "epoch": 0, "step": 18425, "train_loss": 3.596158504486084, "perplexity": 36.45791218646842, "lr": 0.0026291804804649314, "grad_norm": 0.186196, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:41.541414+00:00", "epoch": 0, "step": 18426, "train_loss": 3.53281569480896, "perplexity": 34.22018569002672, "lr": 0.0026291804804649314, "grad_norm": 0.190075, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:41.846419+00:00", "epoch": 0, "step": 18427, "train_loss": 3.597028970718384, "perplexity": 36.48966138422444, "lr": 0.0026291804804649314, "grad_norm": 0.196039, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:42.151766+00:00", "epoch": 0, "step": 18428, "train_loss": 3.4920504093170166, "perplexity": 32.85324128499021, "lr": 0.0026291804804649314, "grad_norm": 0.194801, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:42.458139+00:00", "epoch": 0, "step": 18429, "train_loss": 3.4898595809936523, "perplexity": 32.781344259246104, "lr": 0.0026291804804649314, "grad_norm": 0.181525, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:42.763608+00:00", "epoch": 0, "step": 18430, "train_loss": 3.532252073287964, "perplexity": 34.200903891248146, "lr": 0.0026291804804649314, "grad_norm": 0.210197, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:43.068794+00:00", "epoch": 0, "step": 18431, "train_loss": 3.5849156379699707, "perplexity": 36.050316312704105, "lr": 0.0026291804804649314, "grad_norm": 0.212024, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:43.374165+00:00", "epoch": 0, "step": 18432, "train_loss": 3.5688157081604004, "perplexity": 35.474556037207655, "lr": 0.0026291804804649314, "grad_norm": 0.178675, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:43.679352+00:00", "epoch": 0, "step": 18433, "train_loss": 3.5041799545288086, "perplexity": 33.25416274250201, "lr": 0.0026291804804649314, "grad_norm": 0.196015, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:43.985207+00:00", "epoch": 0, "step": 18434, "train_loss": 3.4794235229492188, "perplexity": 32.441015183577434, "lr": 0.0026291804804649314, "grad_norm": 0.195906, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:44.290565+00:00", "epoch": 0, "step": 18435, "train_loss": 3.557587146759033, "perplexity": 35.07845578625616, "lr": 0.0026291804804649314, "grad_norm": 0.171265, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:44.595108+00:00", "epoch": 0, "step": 18436, "train_loss": 3.517216444015503, "perplexity": 33.690518378442356, "lr": 0.0026291804804649314, "grad_norm": 0.16886, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:44.899079+00:00", "epoch": 0, "step": 18437, "train_loss": 3.4070565700531006, "perplexity": 30.17629161026457, "lr": 0.0026291804804649314, "grad_norm": 0.184372, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:45.203329+00:00", "epoch": 0, "step": 18438, "train_loss": 3.460486650466919, "perplexity": 31.83246402934197, "lr": 0.0026291804804649314, "grad_norm": 0.160233, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:45.509232+00:00", "epoch": 0, "step": 18439, "train_loss": 3.5501508712768555, "perplexity": 34.81857021349572, "lr": 0.0026291804804649314, "grad_norm": 0.176449, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:45.813936+00:00", "epoch": 0, "step": 18440, "train_loss": 3.5367157459259033, "perplexity": 34.35390675344463, "lr": 0.0026291804804649314, "grad_norm": 0.156516, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:46.118653+00:00", "epoch": 0, "step": 18441, "train_loss": 3.515889883041382, "perplexity": 33.645855482140895, "lr": 0.0026291804804649314, "grad_norm": 0.179309, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:46.423475+00:00", "epoch": 0, "step": 18442, "train_loss": 3.4401283264160156, "perplexity": 31.19096053567687, "lr": 0.0026291804804649314, "grad_norm": 0.171579, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:46.728265+00:00", "epoch": 0, "step": 18443, "train_loss": 3.473085403442383, "perplexity": 32.236050384266456, "lr": 0.0026291804804649314, "grad_norm": 0.173971, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:47.033105+00:00", "epoch": 0, "step": 18444, "train_loss": 3.492050886154175, "perplexity": 32.85325695064016, "lr": 0.0026291804804649314, "grad_norm": 0.143899, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:47.337219+00:00", "epoch": 0, "step": 18445, "train_loss": 3.3525235652923584, "perplexity": 28.574752987665793, "lr": 0.0026291804804649314, "grad_norm": 0.162313, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:47.641765+00:00", "epoch": 0, "step": 18446, "train_loss": 3.5228655338287354, "perplexity": 33.88137772585235, "lr": 0.0026291804804649314, "grad_norm": 0.147224, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:47.947419+00:00", "epoch": 0, "step": 18447, "train_loss": 3.586362361907959, "perplexity": 36.10250891332479, "lr": 0.0026291804804649314, "grad_norm": 0.169586, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:48.252272+00:00", "epoch": 0, "step": 18448, "train_loss": 3.5937418937683105, "perplexity": 36.36991397666528, "lr": 0.0026291804804649314, "grad_norm": 0.156633, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:48.558971+00:00", "epoch": 0, "step": 18449, "train_loss": 3.5952606201171875, "perplexity": 36.42519188871378, "lr": 0.0026291804804649314, "grad_norm": 0.169544, "tokens_per_sec": 106846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:48.863756+00:00", "epoch": 0, "step": 18450, "train_loss": 3.574618339538574, "perplexity": 35.681000189597505, "lr": 0.0026291804804649314, "grad_norm": 0.167919, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:49.168241+00:00", "epoch": 0, "step": 18451, "train_loss": 3.4614310264587402, "perplexity": 31.862540043429348, "lr": 0.0026291804804649314, "grad_norm": 0.174838, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:49.473856+00:00", "epoch": 0, "step": 18452, "train_loss": 3.4365286827087402, "perplexity": 31.07888602602893, "lr": 0.0026291804804649314, "grad_norm": 0.149071, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:49.779433+00:00", "epoch": 0, "step": 18453, "train_loss": 3.511000871658325, "perplexity": 33.48176196608269, "lr": 0.0026291804804649314, "grad_norm": 0.171792, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:50.085742+00:00", "epoch": 0, "step": 18454, "train_loss": 3.471062421798706, "perplexity": 32.17090336390521, "lr": 0.0026291804804649314, "grad_norm": 0.185915, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:50.390225+00:00", "epoch": 0, "step": 18455, "train_loss": 3.534888505935669, "perplexity": 34.29119123680354, "lr": 0.0026291804804649314, "grad_norm": 0.170259, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:50.694730+00:00", "epoch": 0, "step": 18456, "train_loss": 3.56685733795166, "perplexity": 35.405151705365725, "lr": 0.0026291804804649314, "grad_norm": 0.170003, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:50.998843+00:00", "epoch": 0, "step": 18457, "train_loss": 3.4093425273895264, "perplexity": 30.245352230198975, "lr": 0.0026291804804649314, "grad_norm": 0.161071, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:51.304754+00:00", "epoch": 0, "step": 18458, "train_loss": 3.6026179790496826, "perplexity": 36.694173382977844, "lr": 0.0026291804804649314, "grad_norm": 0.193497, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:51.609670+00:00", "epoch": 0, "step": 18459, "train_loss": 3.507262706756592, "perplexity": 33.356835262438146, "lr": 0.0026291804804649314, "grad_norm": 0.174815, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:51.914603+00:00", "epoch": 0, "step": 18460, "train_loss": 3.6099138259887695, "perplexity": 36.96286743902237, "lr": 0.0026291804804649314, "grad_norm": 0.152544, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:52.218676+00:00", "epoch": 0, "step": 18461, "train_loss": 3.5001771450042725, "perplexity": 33.121318715188984, "lr": 0.0026291804804649314, "grad_norm": 0.186916, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:52.523915+00:00", "epoch": 0, "step": 18462, "train_loss": 3.543158769607544, "perplexity": 34.57596438131039, "lr": 0.0026291804804649314, "grad_norm": 0.174199, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:52.829154+00:00", "epoch": 0, "step": 18463, "train_loss": 3.4988174438476562, "perplexity": 33.07631422308691, "lr": 0.0026291804804649314, "grad_norm": 0.183873, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:53.133830+00:00", "epoch": 0, "step": 18464, "train_loss": 3.4784669876098633, "perplexity": 32.409999042492174, "lr": 0.0026291804804649314, "grad_norm": 0.17639, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:53.439408+00:00", "epoch": 0, "step": 18465, "train_loss": 3.541656970977783, "perplexity": 34.52407721716808, "lr": 0.0026291804804649314, "grad_norm": 0.163845, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:53.745689+00:00", "epoch": 0, "step": 18466, "train_loss": 3.487360715866089, "perplexity": 32.69953036475868, "lr": 0.0026291804804649314, "grad_norm": 0.186349, "tokens_per_sec": 106986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:54.050454+00:00", "epoch": 0, "step": 18467, "train_loss": 3.5242247581481934, "perplexity": 33.92746143038511, "lr": 0.0026291804804649314, "grad_norm": 0.188065, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:54.355231+00:00", "epoch": 0, "step": 18468, "train_loss": 3.646012783050537, "perplexity": 38.32156463554443, "lr": 0.0026291804804649314, "grad_norm": 0.177679, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:54.660313+00:00", "epoch": 0, "step": 18469, "train_loss": 3.47055721282959, "perplexity": 32.15465443987789, "lr": 0.0026291804804649314, "grad_norm": 0.199336, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:54.964710+00:00", "epoch": 0, "step": 18470, "train_loss": 3.38077712059021, "perplexity": 29.393604615315077, "lr": 0.0026291804804649314, "grad_norm": 0.170993, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:55.269278+00:00", "epoch": 0, "step": 18471, "train_loss": 3.4649605751037598, "perplexity": 31.97519912937239, "lr": 0.0026291804804649314, "grad_norm": 0.17671, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:55.573212+00:00", "epoch": 0, "step": 18472, "train_loss": 3.465177059173584, "perplexity": 31.982121999931916, "lr": 0.0026291804804649314, "grad_norm": 0.185878, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:55.878560+00:00", "epoch": 0, "step": 18473, "train_loss": 3.467292070388794, "perplexity": 32.04983612947772, "lr": 0.0026291804804649314, "grad_norm": 0.184817, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:56.183688+00:00", "epoch": 0, "step": 18474, "train_loss": 3.4706642627716064, "perplexity": 32.158096778019, "lr": 0.0026291804804649314, "grad_norm": 0.179639, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:56.487864+00:00", "epoch": 0, "step": 18475, "train_loss": 3.491029977798462, "perplexity": 32.81973390099924, "lr": 0.0026291804804649314, "grad_norm": 0.169637, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:56.791919+00:00", "epoch": 0, "step": 18476, "train_loss": 3.436581611633301, "perplexity": 31.080531041576897, "lr": 0.0026291804804649314, "grad_norm": 0.17485, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:57.097177+00:00", "epoch": 0, "step": 18477, "train_loss": 3.4539053440093994, "perplexity": 31.623652707613818, "lr": 0.0026291804804649314, "grad_norm": 0.16385, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:57.402729+00:00", "epoch": 0, "step": 18478, "train_loss": 3.6935176849365234, "perplexity": 40.18596016259963, "lr": 0.0026291804804649314, "grad_norm": 0.18273, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:57.708071+00:00", "epoch": 0, "step": 18479, "train_loss": 3.4081003665924072, "perplexity": 30.207805963441594, "lr": 0.0026291804804649314, "grad_norm": 0.170042, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:58.012990+00:00", "epoch": 0, "step": 18480, "train_loss": 3.5186448097229004, "perplexity": 33.738675144107326, "lr": 0.0026291804804649314, "grad_norm": 0.159343, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:58.317078+00:00", "epoch": 0, "step": 18481, "train_loss": 3.401010513305664, "perplexity": 29.994394473069374, "lr": 0.0026291804804649314, "grad_norm": 0.154016, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:58.622344+00:00", "epoch": 0, "step": 18482, "train_loss": 3.493640899658203, "perplexity": 32.90553562372708, "lr": 0.0026291804804649314, "grad_norm": 0.160326, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:58.928163+00:00", "epoch": 0, "step": 18483, "train_loss": 3.471328020095825, "perplexity": 32.17944903586317, "lr": 0.0026291804804649314, "grad_norm": 0.188971, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:59.233047+00:00", "epoch": 0, "step": 18484, "train_loss": 3.397188663482666, "perplexity": 29.87997918016257, "lr": 0.0026291804804649314, "grad_norm": 0.199047, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:59.538001+00:00", "epoch": 0, "step": 18485, "train_loss": 3.5650086402893066, "perplexity": 35.33975874870017, "lr": 0.0026291804804649314, "grad_norm": 0.15083, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:34:59.843082+00:00", "epoch": 0, "step": 18486, "train_loss": 3.487921714782715, "perplexity": 32.71787991242442, "lr": 0.0026291804804649314, "grad_norm": 0.185141, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:00.147327+00:00", "epoch": 0, "step": 18487, "train_loss": 3.4362523555755615, "perplexity": 31.070299272982034, "lr": 0.0026291804804649314, "grad_norm": 0.166975, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:00.452117+00:00", "epoch": 0, "step": 18488, "train_loss": 3.499570608139038, "perplexity": 33.10123550558224, "lr": 0.0026291804804649314, "grad_norm": 0.184739, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:00.758249+00:00", "epoch": 0, "step": 18489, "train_loss": 3.585573673248291, "perplexity": 36.074046499426196, "lr": 0.0026291804804649314, "grad_norm": 0.189075, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:01.063073+00:00", "epoch": 0, "step": 18490, "train_loss": 3.479083299636841, "perplexity": 32.42997987127647, "lr": 0.0026291804804649314, "grad_norm": 0.176894, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:01.367230+00:00", "epoch": 0, "step": 18491, "train_loss": 3.577934741973877, "perplexity": 35.799529181802534, "lr": 0.0026291804804649314, "grad_norm": 0.211387, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:01.671319+00:00", "epoch": 0, "step": 18492, "train_loss": 3.61497163772583, "perplexity": 37.150292243930075, "lr": 0.0026291804804649314, "grad_norm": 0.211978, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:01.975511+00:00", "epoch": 0, "step": 18493, "train_loss": 3.5249407291412354, "perplexity": 33.951761206566225, "lr": 0.0026291804804649314, "grad_norm": 0.171539, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:02.280333+00:00", "epoch": 0, "step": 18494, "train_loss": 3.5420827865600586, "perplexity": 34.538781237589156, "lr": 0.0026291804804649314, "grad_norm": 0.18476, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:02.585123+00:00", "epoch": 0, "step": 18495, "train_loss": 3.5438458919525146, "perplexity": 34.59973046320359, "lr": 0.0026291804804649314, "grad_norm": 0.158927, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:02.890816+00:00", "epoch": 0, "step": 18496, "train_loss": 3.500473737716675, "perplexity": 33.13114371388455, "lr": 0.0026291804804649314, "grad_norm": 0.158651, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:03.196379+00:00", "epoch": 0, "step": 18497, "train_loss": 3.3802318572998047, "perplexity": 29.377581730488092, "lr": 0.0026291804804649314, "grad_norm": 0.172327, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:03.501265+00:00", "epoch": 0, "step": 18498, "train_loss": 3.4030442237854004, "perplexity": 30.055456457595604, "lr": 0.0026291804804649314, "grad_norm": 0.177419, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:03.806320+00:00", "epoch": 0, "step": 18499, "train_loss": 3.5472097396850586, "perplexity": 34.71631466388468, "lr": 0.0026291804804649314, "grad_norm": 0.173207, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:04.111063+00:00", "epoch": 0, "step": 18500, "train_loss": 3.5299782752990723, "perplexity": 34.123226289816024, "lr": 0.0026291804804649314, "grad_norm": 0.178154, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:35:07.184507+00:00", "step": 18500, "epoch": 0, "val_loss": 3.4646038174629212, "val_ppl": 31.963793767361643, "eval_train_loss": 3.5299782752990723, "eval_train_ppl": 34.123226289816024} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:35:08.145478+00:00", "step": 18500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4646_epoch_0000_step_0018500.pt", "val_loss": 3.4646038174629212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:09.533650+00:00", "epoch": 0, "step": 18501, "train_loss": 3.617020845413208, "perplexity": 37.22649896341477, "lr": 0.0026291804804649314, "grad_norm": 0.172239, "tokens_per_sec": 6043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:09.837574+00:00", "epoch": 0, "step": 18502, "train_loss": 3.4967219829559326, "perplexity": 33.00707666781964, "lr": 0.0026291804804649314, "grad_norm": 0.162455, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:10.141543+00:00", "epoch": 0, "step": 18503, "train_loss": 3.530743360519409, "perplexity": 34.149343455575355, "lr": 0.0026291804804649314, "grad_norm": 0.162057, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:10.445099+00:00", "epoch": 0, "step": 18504, "train_loss": 3.4980430603027344, "perplexity": 33.050710384501016, "lr": 0.0026291804804649314, "grad_norm": 0.160256, "tokens_per_sec": 107948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:10.748606+00:00", "epoch": 0, "step": 18505, "train_loss": 3.541144609451294, "perplexity": 34.50639293902319, "lr": 0.0026291804804649314, "grad_norm": 0.181944, "tokens_per_sec": 107965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:11.051964+00:00", "epoch": 0, "step": 18506, "train_loss": 3.5664408206939697, "perplexity": 35.390407919403096, "lr": 0.0026291804804649314, "grad_norm": 0.197189, "tokens_per_sec": 108016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:11.355298+00:00", "epoch": 0, "step": 18507, "train_loss": 3.4476890563964844, "perplexity": 31.427680727175407, "lr": 0.0026291804804649314, "grad_norm": 0.202623, "tokens_per_sec": 108029} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:11.658462+00:00", "epoch": 0, "step": 18508, "train_loss": 3.5029702186584473, "perplexity": 33.21395831226741, "lr": 0.0026291804804649314, "grad_norm": 0.172448, "tokens_per_sec": 108085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:11.962591+00:00", "epoch": 0, "step": 18509, "train_loss": 3.605896472930908, "perplexity": 36.814672425645405, "lr": 0.0026291804804649314, "grad_norm": 0.183648, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:12.266327+00:00", "epoch": 0, "step": 18510, "train_loss": 3.500441074371338, "perplexity": 33.13006155756953, "lr": 0.0026291804804649314, "grad_norm": 0.168433, "tokens_per_sec": 107948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:12.570223+00:00", "epoch": 0, "step": 18511, "train_loss": 3.575612783432007, "perplexity": 35.716500591000724, "lr": 0.0026291804804649314, "grad_norm": 0.166807, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:12.874996+00:00", "epoch": 0, "step": 18512, "train_loss": 3.5323989391326904, "perplexity": 34.20592720475611, "lr": 0.0026291804804649314, "grad_norm": 0.159875, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:13.179405+00:00", "epoch": 0, "step": 18513, "train_loss": 3.55924654006958, "perplexity": 35.136713063632335, "lr": 0.0026291804804649314, "grad_norm": 0.178478, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:13.564809+00:00", "epoch": 0, "step": 18514, "train_loss": 3.5804386138916016, "perplexity": 35.88927893163817, "lr": 0.0026291804804649314, "grad_norm": 0.160766, "tokens_per_sec": 85022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:13.868786+00:00", "epoch": 0, "step": 18515, "train_loss": 3.510120153427124, "perplexity": 33.452286949403934, "lr": 0.0026291804804649314, "grad_norm": 0.163666, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:14.173398+00:00", "epoch": 0, "step": 18516, "train_loss": 3.4341797828674316, "perplexity": 31.005970504727465, "lr": 0.0026291804804649314, "grad_norm": 0.187597, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:14.477946+00:00", "epoch": 0, "step": 18517, "train_loss": 3.5627853870391846, "perplexity": 35.26127679016536, "lr": 0.0026291804804649314, "grad_norm": 0.161745, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:14.783152+00:00", "epoch": 0, "step": 18518, "train_loss": 3.4681761264801025, "perplexity": 32.07818251038259, "lr": 0.0026291804804649314, "grad_norm": 0.200861, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:15.087528+00:00", "epoch": 0, "step": 18519, "train_loss": 3.471240520477295, "perplexity": 32.17663346953031, "lr": 0.0026291804804649314, "grad_norm": 0.159196, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:15.391885+00:00", "epoch": 0, "step": 18520, "train_loss": 3.435657024383545, "perplexity": 31.05180765954282, "lr": 0.0026291804804649314, "grad_norm": 0.199627, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:15.695966+00:00", "epoch": 0, "step": 18521, "train_loss": 3.4789764881134033, "perplexity": 32.426516160706775, "lr": 0.0026291804804649314, "grad_norm": 0.221839, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:16.000280+00:00", "epoch": 0, "step": 18522, "train_loss": 3.4425971508026123, "perplexity": 31.268060673878377, "lr": 0.0026291804804649314, "grad_norm": 0.20218, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:16.304296+00:00", "epoch": 0, "step": 18523, "train_loss": 3.5988216400146484, "perplexity": 36.5551339476122, "lr": 0.0026291804804649314, "grad_norm": 0.214829, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:16.608571+00:00", "epoch": 0, "step": 18524, "train_loss": 3.6005921363830566, "perplexity": 36.61991200725138, "lr": 0.0026291804804649314, "grad_norm": 0.192983, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:16.913026+00:00", "epoch": 0, "step": 18525, "train_loss": 3.5871496200561523, "perplexity": 36.130942098296785, "lr": 0.0026291804804649314, "grad_norm": 0.237381, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:17.217621+00:00", "epoch": 0, "step": 18526, "train_loss": 3.6369845867156982, "perplexity": 37.97714710012695, "lr": 0.0026291804804649314, "grad_norm": 0.188626, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:17.522083+00:00", "epoch": 0, "step": 18527, "train_loss": 3.482569694519043, "perplexity": 32.543240908756836, "lr": 0.0026291804804649314, "grad_norm": 0.186524, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:17.826014+00:00", "epoch": 0, "step": 18528, "train_loss": 3.3692705631256104, "perplexity": 29.057323842052035, "lr": 0.0026291804804649314, "grad_norm": 0.155867, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:18.130498+00:00", "epoch": 0, "step": 18529, "train_loss": 3.565183162689209, "perplexity": 35.345926866430666, "lr": 0.0026291804804649314, "grad_norm": 0.17648, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:18.435173+00:00", "epoch": 0, "step": 18530, "train_loss": 3.541515588760376, "perplexity": 34.51919647161062, "lr": 0.0026291804804649314, "grad_norm": 0.171175, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:18.739961+00:00", "epoch": 0, "step": 18531, "train_loss": 3.4124720096588135, "perplexity": 30.340152784688577, "lr": 0.0026291804804649314, "grad_norm": 0.152961, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:19.043973+00:00", "epoch": 0, "step": 18532, "train_loss": 3.6507766246795654, "perplexity": 38.504558030146534, "lr": 0.0026291804804649314, "grad_norm": 0.177044, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:19.348481+00:00", "epoch": 0, "step": 18533, "train_loss": 3.506469249725342, "perplexity": 33.3303785444814, "lr": 0.0026291804804649314, "grad_norm": 0.146372, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:19.652576+00:00", "epoch": 0, "step": 18534, "train_loss": 3.5154690742492676, "perplexity": 33.6316999889227, "lr": 0.0026291804804649314, "grad_norm": 0.158006, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:19.957881+00:00", "epoch": 0, "step": 18535, "train_loss": 3.5858006477355957, "perplexity": 36.08223531692658, "lr": 0.0026291804804649314, "grad_norm": 0.157189, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:20.262693+00:00", "epoch": 0, "step": 18536, "train_loss": 3.66711163520813, "perplexity": 39.13869561235705, "lr": 0.0026291804804649314, "grad_norm": 0.165077, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:20.568096+00:00", "epoch": 0, "step": 18537, "train_loss": 3.4591877460479736, "perplexity": 31.791143542639222, "lr": 0.0026291804804649314, "grad_norm": 0.167873, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:20.872526+00:00", "epoch": 0, "step": 18538, "train_loss": 3.562541961669922, "perplexity": 35.25269434547681, "lr": 0.0026291804804649314, "grad_norm": 0.175306, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:21.177074+00:00", "epoch": 0, "step": 18539, "train_loss": 3.520397424697876, "perplexity": 33.79785789854246, "lr": 0.0026291804804649314, "grad_norm": 0.198628, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:21.481058+00:00", "epoch": 0, "step": 18540, "train_loss": 3.5261969566345215, "perplexity": 33.99443914342569, "lr": 0.0026291804804649314, "grad_norm": 0.168154, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:21.787603+00:00", "epoch": 0, "step": 18541, "train_loss": 3.5039634704589844, "perplexity": 33.24696452519072, "lr": 0.0026291804804649314, "grad_norm": 0.215019, "tokens_per_sec": 106837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:22.093683+00:00", "epoch": 0, "step": 18542, "train_loss": 3.498110055923462, "perplexity": 33.052924711532995, "lr": 0.0026291804804649314, "grad_norm": 0.200683, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:22.398252+00:00", "epoch": 0, "step": 18543, "train_loss": 3.547607421875, "perplexity": 34.73012346950295, "lr": 0.0026291804804649314, "grad_norm": 0.179392, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:22.701945+00:00", "epoch": 0, "step": 18544, "train_loss": 3.5319032669067383, "perplexity": 34.18897647802705, "lr": 0.0026291804804649314, "grad_norm": 0.165681, "tokens_per_sec": 107898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:23.006320+00:00", "epoch": 0, "step": 18545, "train_loss": 3.3954033851623535, "perplexity": 29.826682689810855, "lr": 0.0026291804804649314, "grad_norm": 0.16916, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:23.311156+00:00", "epoch": 0, "step": 18546, "train_loss": 3.468229055404663, "perplexity": 32.079880419018565, "lr": 0.0026291804804649314, "grad_norm": 0.166357, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:23.615778+00:00", "epoch": 0, "step": 18547, "train_loss": 3.5207676887512207, "perplexity": 33.8103743474549, "lr": 0.0026291804804649314, "grad_norm": 0.154742, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:23.919328+00:00", "epoch": 0, "step": 18548, "train_loss": 3.498821496963501, "perplexity": 33.07644828549186, "lr": 0.0026291804804649314, "grad_norm": 0.160829, "tokens_per_sec": 107949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:24.223403+00:00", "epoch": 0, "step": 18549, "train_loss": 3.507905960083008, "perplexity": 33.378299060269, "lr": 0.0026291804804649314, "grad_norm": 0.156912, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:24.527854+00:00", "epoch": 0, "step": 18550, "train_loss": 3.5235886573791504, "perplexity": 33.90588700855818, "lr": 0.0026291804804649314, "grad_norm": 0.16393, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:24.833264+00:00", "epoch": 0, "step": 18551, "train_loss": 3.4679341316223145, "perplexity": 32.070420694365545, "lr": 0.0026291804804649314, "grad_norm": 0.160778, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:25.137978+00:00", "epoch": 0, "step": 18552, "train_loss": 3.5308339595794678, "perplexity": 34.15243749415042, "lr": 0.0026291804804649314, "grad_norm": 0.156779, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:25.442453+00:00", "epoch": 0, "step": 18553, "train_loss": 3.5276477336883545, "perplexity": 34.043793287966565, "lr": 0.0026291804804649314, "grad_norm": 0.160505, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:25.747212+00:00", "epoch": 0, "step": 18554, "train_loss": 3.515843152999878, "perplexity": 33.64428324665338, "lr": 0.0026291804804649314, "grad_norm": 0.1609, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:26.050666+00:00", "epoch": 0, "step": 18555, "train_loss": 3.3921329975128174, "perplexity": 29.729297206055453, "lr": 0.0026291804804649314, "grad_norm": 0.169224, "tokens_per_sec": 107986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:26.354737+00:00", "epoch": 0, "step": 18556, "train_loss": 3.505171537399292, "perplexity": 33.28715335440944, "lr": 0.0026291804804649314, "grad_norm": 0.18072, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:26.659951+00:00", "epoch": 0, "step": 18557, "train_loss": 3.5671231746673584, "perplexity": 35.4145649457509, "lr": 0.0026291804804649314, "grad_norm": 0.167754, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:26.964667+00:00", "epoch": 0, "step": 18558, "train_loss": 3.4836182594299316, "perplexity": 32.57738250597259, "lr": 0.0026291804804649314, "grad_norm": 0.17498, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:27.268981+00:00", "epoch": 0, "step": 18559, "train_loss": 3.441178321838379, "perplexity": 31.223728101344925, "lr": 0.0026291804804649314, "grad_norm": 0.181159, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:27.572711+00:00", "epoch": 0, "step": 18560, "train_loss": 3.381227731704712, "perplexity": 29.406852684889543, "lr": 0.0026291804804649314, "grad_norm": 0.170204, "tokens_per_sec": 107944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:27.877497+00:00", "epoch": 0, "step": 18561, "train_loss": 3.5681912899017334, "perplexity": 35.4524119909928, "lr": 0.0026291804804649314, "grad_norm": 0.1858, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:28.181680+00:00", "epoch": 0, "step": 18562, "train_loss": 3.40348482131958, "perplexity": 30.068701735303545, "lr": 0.0026291804804649314, "grad_norm": 0.192593, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:28.486209+00:00", "epoch": 0, "step": 18563, "train_loss": 3.349978446960449, "perplexity": 28.50211932984194, "lr": 0.0026291804804649314, "grad_norm": 0.163722, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:28.790135+00:00", "epoch": 0, "step": 18564, "train_loss": 3.4496662616729736, "perplexity": 31.489881174599862, "lr": 0.0026291804804649314, "grad_norm": 0.160628, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:29.093947+00:00", "epoch": 0, "step": 18565, "train_loss": 3.600933790206909, "perplexity": 36.6324254777336, "lr": 0.0026291804804649314, "grad_norm": 0.159681, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:29.398485+00:00", "epoch": 0, "step": 18566, "train_loss": 3.5631465911865234, "perplexity": 35.27401561010123, "lr": 0.0026291804804649314, "grad_norm": 0.188643, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:29.702459+00:00", "epoch": 0, "step": 18567, "train_loss": 3.530320882797241, "perplexity": 34.134919165923776, "lr": 0.0026291804804649314, "grad_norm": 0.175064, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:30.007811+00:00", "epoch": 0, "step": 18568, "train_loss": 3.542754650115967, "perplexity": 34.56199438312823, "lr": 0.0026291804804649314, "grad_norm": 0.163779, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:30.312273+00:00", "epoch": 0, "step": 18569, "train_loss": 3.492929220199585, "perplexity": 32.8821257610961, "lr": 0.0026291804804649314, "grad_norm": 0.163657, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:30.616899+00:00", "epoch": 0, "step": 18570, "train_loss": 3.487887144088745, "perplexity": 32.716748852161516, "lr": 0.0026291804804649314, "grad_norm": 0.144134, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:30.921115+00:00", "epoch": 0, "step": 18571, "train_loss": 3.4263076782226562, "perplexity": 30.76284646477592, "lr": 0.0026291804804649314, "grad_norm": 0.148274, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:31.226083+00:00", "epoch": 0, "step": 18572, "train_loss": 3.411292791366577, "perplexity": 30.304396208083684, "lr": 0.0026291804804649314, "grad_norm": 0.144581, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:31.531509+00:00", "epoch": 0, "step": 18573, "train_loss": 3.445772647857666, "perplexity": 31.367510125617216, "lr": 0.0026291804804649314, "grad_norm": 0.160184, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:31.836812+00:00", "epoch": 0, "step": 18574, "train_loss": 3.5195608139038086, "perplexity": 33.76959407036858, "lr": 0.0026291804804649314, "grad_norm": 0.167338, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:32.141269+00:00", "epoch": 0, "step": 18575, "train_loss": 3.5761754512786865, "perplexity": 35.7366027723744, "lr": 0.0026291804804649314, "grad_norm": 0.162638, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:32.444777+00:00", "epoch": 0, "step": 18576, "train_loss": 3.5674521923065186, "perplexity": 35.426218879370964, "lr": 0.0026291804804649314, "grad_norm": 0.1521, "tokens_per_sec": 107964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:32.749665+00:00", "epoch": 0, "step": 18577, "train_loss": 3.439645528793335, "perplexity": 31.175905248701937, "lr": 0.0026291804804649314, "grad_norm": 0.163418, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:33.055256+00:00", "epoch": 0, "step": 18578, "train_loss": 3.5403075218200684, "perplexity": 34.47752015050877, "lr": 0.0026291804804649314, "grad_norm": 0.166301, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:33.361453+00:00", "epoch": 0, "step": 18579, "train_loss": 3.46743106842041, "perplexity": 32.054291303236354, "lr": 0.0026291804804649314, "grad_norm": 0.174323, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:33.666433+00:00", "epoch": 0, "step": 18580, "train_loss": 3.5334432125091553, "perplexity": 34.2416662012468, "lr": 0.0026291804804649314, "grad_norm": 0.155487, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:33.970943+00:00", "epoch": 0, "step": 18581, "train_loss": 3.402165651321411, "perplexity": 30.029062157509756, "lr": 0.0026291804804649314, "grad_norm": 0.180645, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:34.275768+00:00", "epoch": 0, "step": 18582, "train_loss": 3.5533220767974854, "perplexity": 34.929162318042245, "lr": 0.0026291804804649314, "grad_norm": 0.158682, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:34.580893+00:00", "epoch": 0, "step": 18583, "train_loss": 3.555438280105591, "perplexity": 35.00315779400996, "lr": 0.0026291804804649314, "grad_norm": 0.194517, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:34.886758+00:00", "epoch": 0, "step": 18584, "train_loss": 3.6866416931152344, "perplexity": 39.91058963687441, "lr": 0.0026291804804649314, "grad_norm": 0.178071, "tokens_per_sec": 107132} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:35.191789+00:00", "epoch": 0, "step": 18585, "train_loss": 3.523397922515869, "perplexity": 33.89942059054046, "lr": 0.0026291804804649314, "grad_norm": 0.168896, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:35.495893+00:00", "epoch": 0, "step": 18586, "train_loss": 3.5256922245025635, "perplexity": 33.97728538707228, "lr": 0.0026291804804649314, "grad_norm": 0.146233, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:35.799370+00:00", "epoch": 0, "step": 18587, "train_loss": 3.6257126331329346, "perplexity": 37.55147405128249, "lr": 0.0026291804804649314, "grad_norm": 0.163159, "tokens_per_sec": 107975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:36.103843+00:00", "epoch": 0, "step": 18588, "train_loss": 3.568080425262451, "perplexity": 35.44848178998992, "lr": 0.0026291804804649314, "grad_norm": 0.170925, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:36.408617+00:00", "epoch": 0, "step": 18589, "train_loss": 3.4777305126190186, "perplexity": 32.38613867610412, "lr": 0.0026291804804649314, "grad_norm": 0.178029, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:36.713289+00:00", "epoch": 0, "step": 18590, "train_loss": 3.5167975425720215, "perplexity": 33.676408327228735, "lr": 0.0026291804804649314, "grad_norm": 0.151808, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:37.017378+00:00", "epoch": 0, "step": 18591, "train_loss": 3.4245779514312744, "perplexity": 30.70968113906644, "lr": 0.0026291804804649314, "grad_norm": 0.194499, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:37.321176+00:00", "epoch": 0, "step": 18592, "train_loss": 3.5361168384552, "perplexity": 34.33333810201648, "lr": 0.0026291804804649314, "grad_norm": 0.165298, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:37.626342+00:00", "epoch": 0, "step": 18593, "train_loss": 3.521710157394409, "perplexity": 33.842254585794535, "lr": 0.0026291804804649314, "grad_norm": 0.18478, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:37.930861+00:00", "epoch": 0, "step": 18594, "train_loss": 3.571380615234375, "perplexity": 35.56566176585074, "lr": 0.0026291804804649314, "grad_norm": 0.192401, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:38.235358+00:00", "epoch": 0, "step": 18595, "train_loss": 3.4452364444732666, "perplexity": 31.35069526902206, "lr": 0.0026291804804649314, "grad_norm": 0.153241, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:38.540349+00:00", "epoch": 0, "step": 18596, "train_loss": 3.4775843620300293, "perplexity": 32.38140576872854, "lr": 0.0026291804804649314, "grad_norm": 0.191131, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:38.845538+00:00", "epoch": 0, "step": 18597, "train_loss": 3.4832603931427, "perplexity": 32.5657262448633, "lr": 0.0026291804804649314, "grad_norm": 0.223708, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:39.149437+00:00", "epoch": 0, "step": 18598, "train_loss": 3.5622687339782715, "perplexity": 35.24306364892274, "lr": 0.0026291804804649314, "grad_norm": 0.201776, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:39.453448+00:00", "epoch": 0, "step": 18599, "train_loss": 3.5387091636657715, "perplexity": 34.422456742282804, "lr": 0.0026291804804649314, "grad_norm": 0.171359, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:39.758101+00:00", "epoch": 0, "step": 18600, "train_loss": 3.372089147567749, "perplexity": 29.139339893248906, "lr": 0.0026291804804649314, "grad_norm": 0.193419, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:40.062615+00:00", "epoch": 0, "step": 18601, "train_loss": 3.4546403884887695, "perplexity": 31.646906044025382, "lr": 0.0026291804804649314, "grad_norm": 0.18283, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:40.367605+00:00", "epoch": 0, "step": 18602, "train_loss": 3.5906453132629395, "perplexity": 36.257465802329044, "lr": 0.0026291804804649314, "grad_norm": 0.193799, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:40.671397+00:00", "epoch": 0, "step": 18603, "train_loss": 3.5092437267303467, "perplexity": 33.4229813160488, "lr": 0.0026291804804649314, "grad_norm": 0.185093, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:40.975706+00:00", "epoch": 0, "step": 18604, "train_loss": 3.568233013153076, "perplexity": 35.453891211747745, "lr": 0.0026291804804649314, "grad_norm": 0.234654, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:41.281810+00:00", "epoch": 0, "step": 18605, "train_loss": 3.5294008255004883, "perplexity": 34.10352752773841, "lr": 0.0026291804804649314, "grad_norm": 0.204952, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:41.586629+00:00", "epoch": 0, "step": 18606, "train_loss": 3.4591927528381348, "perplexity": 31.791302714622393, "lr": 0.0026291804804649314, "grad_norm": 0.17839, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:41.890679+00:00", "epoch": 0, "step": 18607, "train_loss": 3.4442343711853027, "perplexity": 31.319295309891547, "lr": 0.0026291804804649314, "grad_norm": 0.216749, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:42.195469+00:00", "epoch": 0, "step": 18608, "train_loss": 3.3990674018859863, "perplexity": 29.93616861063174, "lr": 0.0026291804804649314, "grad_norm": 0.164399, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:42.499722+00:00", "epoch": 0, "step": 18609, "train_loss": 3.4927046298980713, "perplexity": 32.87474158379534, "lr": 0.0026291804804649314, "grad_norm": 0.203489, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:42.805300+00:00", "epoch": 0, "step": 18610, "train_loss": 3.452258586883545, "perplexity": 31.5716190872983, "lr": 0.0026291804804649314, "grad_norm": 0.178052, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:43.110807+00:00", "epoch": 0, "step": 18611, "train_loss": 3.4727001190185547, "perplexity": 32.223632728485946, "lr": 0.0026291804804649314, "grad_norm": 0.181329, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:43.415161+00:00", "epoch": 0, "step": 18612, "train_loss": 3.5033445358276367, "perplexity": 33.22639319427016, "lr": 0.0026291804804649314, "grad_norm": 0.153219, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:43.720045+00:00", "epoch": 0, "step": 18613, "train_loss": 3.504615306854248, "perplexity": 33.26864317139799, "lr": 0.0026291804804649314, "grad_norm": 0.222887, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:44.025243+00:00", "epoch": 0, "step": 18614, "train_loss": 3.470174551010132, "perplexity": 32.14235243520974, "lr": 0.0026291804804649314, "grad_norm": 0.183884, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:44.329924+00:00", "epoch": 0, "step": 18615, "train_loss": 3.5709002017974854, "perplexity": 35.54857964761466, "lr": 0.0026291804804649314, "grad_norm": 0.193518, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:44.635495+00:00", "epoch": 0, "step": 18616, "train_loss": 3.558877944946289, "perplexity": 35.12376422913383, "lr": 0.0026291804804649314, "grad_norm": 0.177084, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:44.940436+00:00", "epoch": 0, "step": 18617, "train_loss": 3.454843521118164, "perplexity": 31.65333521622753, "lr": 0.0026291804804649314, "grad_norm": 0.197332, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:45.245125+00:00", "epoch": 0, "step": 18618, "train_loss": 3.45529842376709, "perplexity": 31.667737677875337, "lr": 0.0026291804804649314, "grad_norm": 0.188389, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:45.550175+00:00", "epoch": 0, "step": 18619, "train_loss": 3.4282631874084473, "perplexity": 30.823062350789105, "lr": 0.0026291804804649314, "grad_norm": 0.19889, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:45.855258+00:00", "epoch": 0, "step": 18620, "train_loss": 3.4975101947784424, "perplexity": 33.03310349185342, "lr": 0.0026291804804649314, "grad_norm": 0.201954, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:46.160720+00:00", "epoch": 0, "step": 18621, "train_loss": 3.3549516201019287, "perplexity": 28.644218352819294, "lr": 0.0026291804804649314, "grad_norm": 0.178232, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:46.466454+00:00", "epoch": 0, "step": 18622, "train_loss": 3.5427515506744385, "perplexity": 34.56188726041355, "lr": 0.0026291804804649314, "grad_norm": 0.20073, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:46.771574+00:00", "epoch": 0, "step": 18623, "train_loss": 3.5859975814819336, "perplexity": 36.089341826436595, "lr": 0.0026291804804649314, "grad_norm": 0.159777, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:47.075807+00:00", "epoch": 0, "step": 18624, "train_loss": 3.4543445110321045, "perplexity": 31.637543823057126, "lr": 0.0026291804804649314, "grad_norm": 0.182546, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:47.381561+00:00", "epoch": 0, "step": 18625, "train_loss": 3.4192380905151367, "perplexity": 30.546132764411322, "lr": 0.0026291804804649314, "grad_norm": 0.169596, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:47.687591+00:00", "epoch": 0, "step": 18626, "train_loss": 3.4498355388641357, "perplexity": 31.495212144428194, "lr": 0.0026291804804649314, "grad_norm": 0.156122, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:47.993342+00:00", "epoch": 0, "step": 18627, "train_loss": 3.534442901611328, "perplexity": 34.27591433768057, "lr": 0.0026291804804649314, "grad_norm": 0.18354, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:48.298081+00:00", "epoch": 0, "step": 18628, "train_loss": 3.411545753479004, "perplexity": 30.312063041832648, "lr": 0.0026291804804649314, "grad_norm": 0.192791, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:48.602812+00:00", "epoch": 0, "step": 18629, "train_loss": 3.5081348419189453, "perplexity": 33.38593962099782, "lr": 0.0026291804804649314, "grad_norm": 0.162806, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:48.906954+00:00", "epoch": 0, "step": 18630, "train_loss": 3.4333908557891846, "perplexity": 30.981518701621994, "lr": 0.0026291804804649314, "grad_norm": 0.156853, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:49.212493+00:00", "epoch": 0, "step": 18631, "train_loss": 3.472982168197632, "perplexity": 32.23272265948743, "lr": 0.0026291804804649314, "grad_norm": 0.162348, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:49.517678+00:00", "epoch": 0, "step": 18632, "train_loss": 3.5296854972839355, "perplexity": 34.11323722171395, "lr": 0.0026291804804649314, "grad_norm": 0.161213, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:49.823207+00:00", "epoch": 0, "step": 18633, "train_loss": 3.4382970333099365, "perplexity": 31.133893014303766, "lr": 0.0026291804804649314, "grad_norm": 0.170666, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:50.128326+00:00", "epoch": 0, "step": 18634, "train_loss": 3.4445788860321045, "perplexity": 31.330087130981727, "lr": 0.0026291804804649314, "grad_norm": 0.164781, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:50.432543+00:00", "epoch": 0, "step": 18635, "train_loss": 3.438732624053955, "perplexity": 31.147457604016445, "lr": 0.0026291804804649314, "grad_norm": 0.150755, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:50.738389+00:00", "epoch": 0, "step": 18636, "train_loss": 3.4366207122802734, "perplexity": 31.081746334208084, "lr": 0.0026291804804649314, "grad_norm": 0.156571, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:51.044560+00:00", "epoch": 0, "step": 18637, "train_loss": 3.5584304332733154, "perplexity": 35.10804945117807, "lr": 0.0026291804804649314, "grad_norm": 0.179054, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:51.350362+00:00", "epoch": 0, "step": 18638, "train_loss": 3.4224302768707275, "perplexity": 30.643797511740793, "lr": 0.0026291804804649314, "grad_norm": 0.178079, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:51.655719+00:00", "epoch": 0, "step": 18639, "train_loss": 3.549328088760376, "perplexity": 34.78993388502594, "lr": 0.0026291804804649314, "grad_norm": 0.17224, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:51.960680+00:00", "epoch": 0, "step": 18640, "train_loss": 3.504225015640259, "perplexity": 33.25566124579739, "lr": 0.0026291804804649314, "grad_norm": 0.165881, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:52.266886+00:00", "epoch": 0, "step": 18641, "train_loss": 3.528411626815796, "perplexity": 34.069809043054455, "lr": 0.0026291804804649314, "grad_norm": 0.171876, "tokens_per_sec": 106916} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:52.573031+00:00", "epoch": 0, "step": 18642, "train_loss": 3.551495313644409, "perplexity": 34.86541325629824, "lr": 0.0026291804804649314, "grad_norm": 0.200492, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:52.877931+00:00", "epoch": 0, "step": 18643, "train_loss": 3.445659637451172, "perplexity": 31.363965470842423, "lr": 0.0026291804804649314, "grad_norm": 0.172936, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:53.182792+00:00", "epoch": 0, "step": 18644, "train_loss": 3.3848824501037598, "perplexity": 29.514523083101988, "lr": 0.0026291804804649314, "grad_norm": 0.171574, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:53.487655+00:00", "epoch": 0, "step": 18645, "train_loss": 3.582906723022461, "perplexity": 35.977966989551334, "lr": 0.0026291804804649314, "grad_norm": 0.159907, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:53.793061+00:00", "epoch": 0, "step": 18646, "train_loss": 3.4496443271636963, "perplexity": 31.48919046708429, "lr": 0.0026291804804649314, "grad_norm": 0.209813, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:54.097674+00:00", "epoch": 0, "step": 18647, "train_loss": 3.500869035720825, "perplexity": 33.144242977755724, "lr": 0.0026291804804649314, "grad_norm": 0.211706, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:54.402482+00:00", "epoch": 0, "step": 18648, "train_loss": 3.4869320392608643, "perplexity": 32.68551584515244, "lr": 0.0026291804804649314, "grad_norm": 0.167299, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:54.706646+00:00", "epoch": 0, "step": 18649, "train_loss": 3.4914867877960205, "perplexity": 32.83472970841903, "lr": 0.0026291804804649314, "grad_norm": 0.193903, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:55.011887+00:00", "epoch": 0, "step": 18650, "train_loss": 3.5671987533569336, "perplexity": 35.41724163331037, "lr": 0.0026291804804649314, "grad_norm": 0.234911, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:55.316269+00:00", "epoch": 0, "step": 18651, "train_loss": 3.459838628768921, "perplexity": 31.81184258424378, "lr": 0.0026291804804649314, "grad_norm": 0.221187, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:55.620851+00:00", "epoch": 0, "step": 18652, "train_loss": 3.4577436447143555, "perplexity": 31.7452670429819, "lr": 0.0026291804804649314, "grad_norm": 0.147673, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:55.924795+00:00", "epoch": 0, "step": 18653, "train_loss": 3.5134918689727783, "perplexity": 33.56526890982916, "lr": 0.0026291804804649314, "grad_norm": 0.17389, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:56.229957+00:00", "epoch": 0, "step": 18654, "train_loss": 3.5645530223846436, "perplexity": 35.32366098935996, "lr": 0.0026291804804649314, "grad_norm": 0.175911, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:56.534772+00:00", "epoch": 0, "step": 18655, "train_loss": 3.41581392288208, "perplexity": 30.441716556606593, "lr": 0.0026291804804649314, "grad_norm": 0.17295, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:56.840389+00:00", "epoch": 0, "step": 18656, "train_loss": 3.4455020427703857, "perplexity": 31.35902306617443, "lr": 0.0026291804804649314, "grad_norm": 0.169849, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:57.144813+00:00", "epoch": 0, "step": 18657, "train_loss": 3.5636043548583984, "perplexity": 35.290166469364976, "lr": 0.0026291804804649314, "grad_norm": 0.183952, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:57.449466+00:00", "epoch": 0, "step": 18658, "train_loss": 3.509876012802124, "perplexity": 33.44412088403514, "lr": 0.0026291804804649314, "grad_norm": 0.167661, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:57.755229+00:00", "epoch": 0, "step": 18659, "train_loss": 3.4523630142211914, "perplexity": 31.574916199576105, "lr": 0.0026291804804649314, "grad_norm": 0.173533, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:58.060339+00:00", "epoch": 0, "step": 18660, "train_loss": 3.48786997795105, "perplexity": 32.716187236766174, "lr": 0.0026291804804649314, "grad_norm": 0.156977, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:58.365850+00:00", "epoch": 0, "step": 18661, "train_loss": 3.5319676399230957, "perplexity": 34.19117739640823, "lr": 0.0026291804804649314, "grad_norm": 0.184629, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:58.670691+00:00", "epoch": 0, "step": 18662, "train_loss": 3.577465534210205, "perplexity": 35.782735704897604, "lr": 0.0026291804804649314, "grad_norm": 0.166983, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:58.975000+00:00", "epoch": 0, "step": 18663, "train_loss": 3.4888949394226074, "perplexity": 32.74973725898304, "lr": 0.0026291804804649314, "grad_norm": 0.187322, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:59.279950+00:00", "epoch": 0, "step": 18664, "train_loss": 3.4541211128234863, "perplexity": 31.630476841846836, "lr": 0.0026291804804649314, "grad_norm": 0.157851, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:59.585787+00:00", "epoch": 0, "step": 18665, "train_loss": 3.6387157440185547, "perplexity": 38.042948455486055, "lr": 0.0026291804804649314, "grad_norm": 0.162144, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:35:59.890026+00:00", "epoch": 0, "step": 18666, "train_loss": 3.502779006958008, "perplexity": 33.20760802196445, "lr": 0.0026291804804649314, "grad_norm": 0.145257, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:00.194773+00:00", "epoch": 0, "step": 18667, "train_loss": 3.472500801086426, "perplexity": 32.21721062068671, "lr": 0.0026291804804649314, "grad_norm": 0.164909, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:00.499607+00:00", "epoch": 0, "step": 18668, "train_loss": 3.516531467437744, "perplexity": 33.66744906433248, "lr": 0.0026291804804649314, "grad_norm": 0.157059, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:00.804841+00:00", "epoch": 0, "step": 18669, "train_loss": 3.5339202880859375, "perplexity": 34.258005961240045, "lr": 0.0026291804804649314, "grad_norm": 0.15433, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:01.110025+00:00", "epoch": 0, "step": 18670, "train_loss": 3.531644821166992, "perplexity": 34.18014162442434, "lr": 0.0026291804804649314, "grad_norm": 0.184149, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:01.414665+00:00", "epoch": 0, "step": 18671, "train_loss": 3.4463536739349365, "perplexity": 31.38574076270415, "lr": 0.0026291804804649314, "grad_norm": 0.175024, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:01.718168+00:00", "epoch": 0, "step": 18672, "train_loss": 3.5352890491485596, "perplexity": 34.304929091834005, "lr": 0.0026291804804649314, "grad_norm": 0.164339, "tokens_per_sec": 107963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:02.022097+00:00", "epoch": 0, "step": 18673, "train_loss": 3.4571118354797363, "perplexity": 31.725216424858527, "lr": 0.0026291804804649314, "grad_norm": 0.198359, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:02.326459+00:00", "epoch": 0, "step": 18674, "train_loss": 3.480325698852539, "perplexity": 32.470295891944055, "lr": 0.0026291804804649314, "grad_norm": 0.171098, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:02.632033+00:00", "epoch": 0, "step": 18675, "train_loss": 3.530346393585205, "perplexity": 34.13578998571638, "lr": 0.0026291804804649314, "grad_norm": 0.193772, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:02.937146+00:00", "epoch": 0, "step": 18676, "train_loss": 3.4330642223358154, "perplexity": 30.971400753697953, "lr": 0.0026291804804649314, "grad_norm": 0.169602, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:03.240073+00:00", "epoch": 0, "step": 18677, "train_loss": 3.398646354675293, "perplexity": 29.92356672352027, "lr": 0.0026291804804649314, "grad_norm": 0.184409, "tokens_per_sec": 108168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:03.544464+00:00", "epoch": 0, "step": 18678, "train_loss": 3.439072847366333, "perplexity": 31.158056498107825, "lr": 0.0026291804804649314, "grad_norm": 0.159373, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:03.848563+00:00", "epoch": 0, "step": 18679, "train_loss": 3.5576775074005127, "perplexity": 35.081625641236016, "lr": 0.0026291804804649314, "grad_norm": 0.171887, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:04.154596+00:00", "epoch": 0, "step": 18680, "train_loss": 3.511615037918091, "perplexity": 33.50233165055628, "lr": 0.0026291804804649314, "grad_norm": 0.165486, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:04.459630+00:00", "epoch": 0, "step": 18681, "train_loss": 3.5679707527160645, "perplexity": 35.44459427790704, "lr": 0.0026291804804649314, "grad_norm": 0.176327, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:04.764014+00:00", "epoch": 0, "step": 18682, "train_loss": 3.4812121391296387, "perplexity": 32.49909163099788, "lr": 0.0026291804804649314, "grad_norm": 0.187106, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:05.068285+00:00", "epoch": 0, "step": 18683, "train_loss": 3.5234642028808594, "perplexity": 33.90166753097335, "lr": 0.0026291804804649314, "grad_norm": 0.174968, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:05.372696+00:00", "epoch": 0, "step": 18684, "train_loss": 3.458951950073242, "perplexity": 31.783648202679984, "lr": 0.0026291804804649314, "grad_norm": 0.180378, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:05.677040+00:00", "epoch": 0, "step": 18685, "train_loss": 3.475574493408203, "perplexity": 32.31638875705591, "lr": 0.0026291804804649314, "grad_norm": 0.173527, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:05.981717+00:00", "epoch": 0, "step": 18686, "train_loss": 3.5155301094055176, "perplexity": 33.63375276763167, "lr": 0.0026291804804649314, "grad_norm": 0.141916, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:06.286781+00:00", "epoch": 0, "step": 18687, "train_loss": 3.5939254760742188, "perplexity": 36.37659146225415, "lr": 0.0026291804804649314, "grad_norm": 0.182308, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:06.590803+00:00", "epoch": 0, "step": 18688, "train_loss": 3.533966064453125, "perplexity": 34.259574204193996, "lr": 0.0026291804804649314, "grad_norm": 0.185584, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:06.895252+00:00", "epoch": 0, "step": 18689, "train_loss": 3.493879556655884, "perplexity": 32.91338969724364, "lr": 0.0026291804804649314, "grad_norm": 0.15322, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:07.199743+00:00", "epoch": 0, "step": 18690, "train_loss": 3.520810604095459, "perplexity": 33.81182536244403, "lr": 0.0026291804804649314, "grad_norm": 0.175546, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:07.504098+00:00", "epoch": 0, "step": 18691, "train_loss": 3.511751651763916, "perplexity": 33.506908845574166, "lr": 0.0026291804804649314, "grad_norm": 0.190544, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:07.808619+00:00", "epoch": 0, "step": 18692, "train_loss": 3.498677968978882, "perplexity": 33.07170123020687, "lr": 0.0026291804804649314, "grad_norm": 0.150293, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:08.113099+00:00", "epoch": 0, "step": 18693, "train_loss": 3.5161526203155518, "perplexity": 33.65469666390053, "lr": 0.0026291804804649314, "grad_norm": 0.170366, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:08.417555+00:00", "epoch": 0, "step": 18694, "train_loss": 3.5574569702148438, "perplexity": 35.073889691312075, "lr": 0.0026291804804649314, "grad_norm": 0.202547, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:08.722696+00:00", "epoch": 0, "step": 18695, "train_loss": 3.4940319061279297, "perplexity": 32.918404416772496, "lr": 0.0026291804804649314, "grad_norm": 0.157781, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:09.027963+00:00", "epoch": 0, "step": 18696, "train_loss": 3.477963924407959, "perplexity": 32.393698864953656, "lr": 0.0026291804804649314, "grad_norm": 0.189574, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:09.332492+00:00", "epoch": 0, "step": 18697, "train_loss": 3.478104829788208, "perplexity": 32.39826363300253, "lr": 0.0026291804804649314, "grad_norm": 0.194292, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:09.636950+00:00", "epoch": 0, "step": 18698, "train_loss": 3.559699535369873, "perplexity": 35.1526334351724, "lr": 0.0026291804804649314, "grad_norm": 0.187094, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:09.940728+00:00", "epoch": 0, "step": 18699, "train_loss": 3.4948837757110596, "perplexity": 32.94645855175592, "lr": 0.0026291804804649314, "grad_norm": 0.165599, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:10.245181+00:00", "epoch": 0, "step": 18700, "train_loss": 3.4013240337371826, "perplexity": 30.003799802872198, "lr": 0.0026291804804649314, "grad_norm": 0.178437, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:10.550263+00:00", "epoch": 0, "step": 18701, "train_loss": 3.531557321548462, "perplexity": 34.17715100591179, "lr": 0.0026291804804649314, "grad_norm": 0.175726, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:10.855344+00:00", "epoch": 0, "step": 18702, "train_loss": 3.4863858222961426, "perplexity": 32.66766733691989, "lr": 0.0026291804804649314, "grad_norm": 0.180004, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:11.160070+00:00", "epoch": 0, "step": 18703, "train_loss": 3.5328710079193115, "perplexity": 34.22207856728402, "lr": 0.0026291804804649314, "grad_norm": 0.15479, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:11.463945+00:00", "epoch": 0, "step": 18704, "train_loss": 3.472522497177124, "perplexity": 32.21790961579308, "lr": 0.0026291804804649314, "grad_norm": 0.19358, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:11.768367+00:00", "epoch": 0, "step": 18705, "train_loss": 3.4687228202819824, "perplexity": 32.095724248479435, "lr": 0.0026291804804649314, "grad_norm": 0.163651, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:12.073024+00:00", "epoch": 0, "step": 18706, "train_loss": 3.592012882232666, "perplexity": 36.30708430811242, "lr": 0.0026291804804649314, "grad_norm": 0.151021, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:12.377752+00:00", "epoch": 0, "step": 18707, "train_loss": 3.5419394969940186, "perplexity": 34.53383254517053, "lr": 0.0026291804804649314, "grad_norm": 0.157891, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:12.682923+00:00", "epoch": 0, "step": 18708, "train_loss": 3.550664186477661, "perplexity": 34.8364477028576, "lr": 0.0026291804804649314, "grad_norm": 0.157565, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:12.986849+00:00", "epoch": 0, "step": 18709, "train_loss": 3.544175863265991, "perplexity": 34.611149265545464, "lr": 0.0026291804804649314, "grad_norm": 0.164018, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:13.292985+00:00", "epoch": 0, "step": 18710, "train_loss": 3.5143795013427734, "perplexity": 33.595075755824745, "lr": 0.0026291804804649314, "grad_norm": 0.179878, "tokens_per_sec": 107094} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:13.597169+00:00", "epoch": 0, "step": 18711, "train_loss": 3.489619731903076, "perplexity": 32.77348262647808, "lr": 0.0026291804804649314, "grad_norm": 0.178339, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:13.902631+00:00", "epoch": 0, "step": 18712, "train_loss": 3.5059080123901367, "perplexity": 33.31167753998756, "lr": 0.0026291804804649314, "grad_norm": 0.184234, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:14.207423+00:00", "epoch": 0, "step": 18713, "train_loss": 3.479794979095459, "perplexity": 32.45306783643505, "lr": 0.0026291804804649314, "grad_norm": 0.187777, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:14.511654+00:00", "epoch": 0, "step": 18714, "train_loss": 3.422412395477295, "perplexity": 30.64324956284028, "lr": 0.0026291804804649314, "grad_norm": 0.214509, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:14.816156+00:00", "epoch": 0, "step": 18715, "train_loss": 3.528388023376465, "perplexity": 34.06900488787414, "lr": 0.0026291804804649314, "grad_norm": 0.20659, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:15.121097+00:00", "epoch": 0, "step": 18716, "train_loss": 3.5016422271728516, "perplexity": 33.16987973289278, "lr": 0.0026291804804649314, "grad_norm": 0.178172, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:15.425520+00:00", "epoch": 0, "step": 18717, "train_loss": 3.4439525604248047, "perplexity": 31.310470438992198, "lr": 0.0026291804804649314, "grad_norm": 0.178385, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:15.729912+00:00", "epoch": 0, "step": 18718, "train_loss": 3.5153942108154297, "perplexity": 33.62918229861837, "lr": 0.0026291804804649314, "grad_norm": 0.189458, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:16.034691+00:00", "epoch": 0, "step": 18719, "train_loss": 3.4375264644622803, "perplexity": 31.109911447151603, "lr": 0.0026291804804649314, "grad_norm": 0.183874, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:16.338610+00:00", "epoch": 0, "step": 18720, "train_loss": 3.436725378036499, "perplexity": 31.084999698947936, "lr": 0.0026291804804649314, "grad_norm": 0.190174, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:16.642809+00:00", "epoch": 0, "step": 18721, "train_loss": 3.461092472076416, "perplexity": 31.85175466668696, "lr": 0.0026291804804649314, "grad_norm": 0.202745, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:16.947650+00:00", "epoch": 0, "step": 18722, "train_loss": 3.465092182159424, "perplexity": 31.97940756810812, "lr": 0.0026291804804649314, "grad_norm": 0.192608, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:17.253509+00:00", "epoch": 0, "step": 18723, "train_loss": 3.530757427215576, "perplexity": 34.14982382739266, "lr": 0.0026291804804649314, "grad_norm": 0.179522, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:17.558541+00:00", "epoch": 0, "step": 18724, "train_loss": 3.4862542152404785, "perplexity": 32.66336832430267, "lr": 0.0026291804804649314, "grad_norm": 0.18869, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:17.862982+00:00", "epoch": 0, "step": 18725, "train_loss": 3.5538532733917236, "perplexity": 34.94772149895563, "lr": 0.0026291804804649314, "grad_norm": 0.185196, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:18.167114+00:00", "epoch": 0, "step": 18726, "train_loss": 3.527712345123291, "perplexity": 34.04599297736336, "lr": 0.0026291804804649314, "grad_norm": 0.199129, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:18.471641+00:00", "epoch": 0, "step": 18727, "train_loss": 3.45176362991333, "perplexity": 31.555996360977655, "lr": 0.0026291804804649314, "grad_norm": 0.1915, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:18.777838+00:00", "epoch": 0, "step": 18728, "train_loss": 3.554227113723755, "perplexity": 34.96078880915614, "lr": 0.0026291804804649314, "grad_norm": 0.198146, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:19.082987+00:00", "epoch": 0, "step": 18729, "train_loss": 3.493260145187378, "perplexity": 32.89300907884377, "lr": 0.0026291804804649314, "grad_norm": 0.160102, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:19.387628+00:00", "epoch": 0, "step": 18730, "train_loss": 3.5271213054656982, "perplexity": 34.02587639076745, "lr": 0.0026291804804649314, "grad_norm": 0.18538, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:19.692181+00:00", "epoch": 0, "step": 18731, "train_loss": 3.502232313156128, "perplexity": 33.18945859002659, "lr": 0.0026291804804649314, "grad_norm": 0.178432, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:19.996768+00:00", "epoch": 0, "step": 18732, "train_loss": 3.3879971504211426, "perplexity": 29.606595291749635, "lr": 0.0026291804804649314, "grad_norm": 0.155558, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:20.302883+00:00", "epoch": 0, "step": 18733, "train_loss": 3.549391269683838, "perplexity": 34.79213201461517, "lr": 0.0026291804804649314, "grad_norm": 0.160409, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:20.607875+00:00", "epoch": 0, "step": 18734, "train_loss": 3.4673914909362793, "perplexity": 32.0530227001352, "lr": 0.0026291804804649314, "grad_norm": 0.154228, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:20.912696+00:00", "epoch": 0, "step": 18735, "train_loss": 3.5093066692352295, "perplexity": 33.42508510842177, "lr": 0.0026291804804649314, "grad_norm": 0.14292, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:21.231755+00:00", "epoch": 0, "step": 18736, "train_loss": 3.516096591949463, "perplexity": 33.65281109905828, "lr": 0.0026291804804649314, "grad_norm": 0.158903, "tokens_per_sec": 102702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:21.536248+00:00", "epoch": 0, "step": 18737, "train_loss": 3.6295864582061768, "perplexity": 37.69722401566617, "lr": 0.0026291804804649314, "grad_norm": 0.163453, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:21.842096+00:00", "epoch": 0, "step": 18738, "train_loss": 3.426987886428833, "perplexity": 30.783778723726687, "lr": 0.0026291804804649314, "grad_norm": 0.142914, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:22.147336+00:00", "epoch": 0, "step": 18739, "train_loss": 3.515580415725708, "perplexity": 33.635444800527225, "lr": 0.0026291804804649314, "grad_norm": 0.195859, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:22.453232+00:00", "epoch": 0, "step": 18740, "train_loss": 3.5158531665802, "perplexity": 33.64462014807285, "lr": 0.0026291804804649314, "grad_norm": 0.188311, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:22.757801+00:00", "epoch": 0, "step": 18741, "train_loss": 3.4966881275177, "perplexity": 33.00595921769027, "lr": 0.0026291804804649314, "grad_norm": 0.174872, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:23.062388+00:00", "epoch": 0, "step": 18742, "train_loss": 3.5095906257629395, "perplexity": 33.434577727209316, "lr": 0.0026291804804649314, "grad_norm": 0.183838, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:23.367435+00:00", "epoch": 0, "step": 18743, "train_loss": 3.5552642345428467, "perplexity": 34.99706617983846, "lr": 0.0026291804804649314, "grad_norm": 0.173508, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:23.671723+00:00", "epoch": 0, "step": 18744, "train_loss": 3.4802165031433105, "perplexity": 32.46675046853131, "lr": 0.0026291804804649314, "grad_norm": 0.163199, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:23.977314+00:00", "epoch": 0, "step": 18745, "train_loss": 3.6002132892608643, "perplexity": 36.606041286579696, "lr": 0.0026291804804649314, "grad_norm": 0.163514, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:24.281933+00:00", "epoch": 0, "step": 18746, "train_loss": 3.5221426486968994, "perplexity": 33.85689423209121, "lr": 0.0026291804804649314, "grad_norm": 0.160242, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:24.586572+00:00", "epoch": 0, "step": 18747, "train_loss": 3.497182846069336, "perplexity": 33.02229191774154, "lr": 0.0026291804804649314, "grad_norm": 0.17737, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:24.890913+00:00", "epoch": 0, "step": 18748, "train_loss": 3.3482160568237305, "perplexity": 28.451931713929387, "lr": 0.0026291804804649314, "grad_norm": 0.160225, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:25.195261+00:00", "epoch": 0, "step": 18749, "train_loss": 3.523002862930298, "perplexity": 33.886030944519, "lr": 0.0026291804804649314, "grad_norm": 0.152984, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:25.500823+00:00", "epoch": 0, "step": 18750, "train_loss": 3.4763360023498535, "perplexity": 32.341007348505336, "lr": 0.0026291804804649314, "grad_norm": 0.165025, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:25.805257+00:00", "epoch": 0, "step": 18751, "train_loss": 3.5007128715515137, "perplexity": 33.13906743871104, "lr": 0.0026291804804649314, "grad_norm": 0.160825, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:26.109816+00:00", "epoch": 0, "step": 18752, "train_loss": 3.4014627933502197, "perplexity": 30.00796340738588, "lr": 0.0026291804804649314, "grad_norm": 0.156594, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:26.414541+00:00", "epoch": 0, "step": 18753, "train_loss": 3.44464111328125, "perplexity": 31.332036776779297, "lr": 0.0026291804804649314, "grad_norm": 0.154284, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:26.719855+00:00", "epoch": 0, "step": 18754, "train_loss": 3.5263969898223877, "perplexity": 34.001239839617064, "lr": 0.0026291804804649314, "grad_norm": 0.171534, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:27.025636+00:00", "epoch": 0, "step": 18755, "train_loss": 3.596794843673706, "perplexity": 36.48111916766424, "lr": 0.0026291804804649314, "grad_norm": 0.155918, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:27.330495+00:00", "epoch": 0, "step": 18756, "train_loss": 3.621694326400757, "perplexity": 37.400883472496766, "lr": 0.0026291804804649314, "grad_norm": 0.17398, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:27.634504+00:00", "epoch": 0, "step": 18757, "train_loss": 3.6528468132019043, "perplexity": 38.584352290321334, "lr": 0.0026291804804649314, "grad_norm": 0.183488, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:27.938030+00:00", "epoch": 0, "step": 18758, "train_loss": 3.6380181312561035, "perplexity": 38.016418464028476, "lr": 0.0026291804804649314, "grad_norm": 0.192893, "tokens_per_sec": 107958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:28.241735+00:00", "epoch": 0, "step": 18759, "train_loss": 3.539973020553589, "perplexity": 34.465989305002076, "lr": 0.0026291804804649314, "grad_norm": 0.190817, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:28.546299+00:00", "epoch": 0, "step": 18760, "train_loss": 3.4473836421966553, "perplexity": 31.418083732817408, "lr": 0.0026291804804649314, "grad_norm": 0.210012, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:28.850591+00:00", "epoch": 0, "step": 18761, "train_loss": 3.5427932739257812, "perplexity": 34.56332932480619, "lr": 0.0026291804804649314, "grad_norm": 0.182277, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:29.156150+00:00", "epoch": 0, "step": 18762, "train_loss": 3.601710319519043, "perplexity": 36.66088267738648, "lr": 0.0026291804804649314, "grad_norm": 0.18413, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:29.460456+00:00", "epoch": 0, "step": 18763, "train_loss": 3.4945285320281982, "perplexity": 32.93475660912641, "lr": 0.0026291804804649314, "grad_norm": 0.167428, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:29.764278+00:00", "epoch": 0, "step": 18764, "train_loss": 3.5201661586761475, "perplexity": 33.79004250615551, "lr": 0.0026291804804649314, "grad_norm": 0.156924, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:30.069763+00:00", "epoch": 0, "step": 18765, "train_loss": 3.4143152236938477, "perplexity": 30.396127751200716, "lr": 0.0026291804804649314, "grad_norm": 0.166653, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:30.375321+00:00", "epoch": 0, "step": 18766, "train_loss": 3.4507765769958496, "perplexity": 31.52486428973402, "lr": 0.0026291804804649314, "grad_norm": 0.162262, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:30.679511+00:00", "epoch": 0, "step": 18767, "train_loss": 3.475543737411499, "perplexity": 32.3153948495942, "lr": 0.0026291804804649314, "grad_norm": 0.187575, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:30.983339+00:00", "epoch": 0, "step": 18768, "train_loss": 3.4763026237487793, "perplexity": 32.339927868938574, "lr": 0.0026291804804649314, "grad_norm": 0.178504, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:31.287205+00:00", "epoch": 0, "step": 18769, "train_loss": 3.5397586822509766, "perplexity": 34.45860271499945, "lr": 0.0026291804804649314, "grad_norm": 0.170987, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:31.591758+00:00", "epoch": 0, "step": 18770, "train_loss": 3.549928665161133, "perplexity": 34.81083417378306, "lr": 0.0026291804804649314, "grad_norm": 0.192201, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:31.896176+00:00", "epoch": 0, "step": 18771, "train_loss": 3.460038661956787, "perplexity": 31.818206645018268, "lr": 0.0026291804804649314, "grad_norm": 0.186041, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:32.201031+00:00", "epoch": 0, "step": 18772, "train_loss": 3.5304665565490723, "perplexity": 34.13989208987057, "lr": 0.0026291804804649314, "grad_norm": 0.159974, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:32.505608+00:00", "epoch": 0, "step": 18773, "train_loss": 3.516934394836426, "perplexity": 33.68101733533472, "lr": 0.0026291804804649314, "grad_norm": 0.1898, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:32.809945+00:00", "epoch": 0, "step": 18774, "train_loss": 3.586033582687378, "perplexity": 36.09064110963378, "lr": 0.0026291804804649314, "grad_norm": 0.150595, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:33.114306+00:00", "epoch": 0, "step": 18775, "train_loss": 3.4034011363983154, "perplexity": 30.066185543651414, "lr": 0.0026291804804649314, "grad_norm": 0.172908, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:33.419023+00:00", "epoch": 0, "step": 18776, "train_loss": 3.454026460647583, "perplexity": 31.627483090073692, "lr": 0.0026291804804649314, "grad_norm": 0.166552, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:33.725319+00:00", "epoch": 0, "step": 18777, "train_loss": 3.637756824493408, "perplexity": 38.006485814581126, "lr": 0.0026291804804649314, "grad_norm": 0.160392, "tokens_per_sec": 106983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:34.029997+00:00", "epoch": 0, "step": 18778, "train_loss": 3.474717855453491, "perplexity": 32.288717165847224, "lr": 0.0026291804804649314, "grad_norm": 0.174601, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:34.334360+00:00", "epoch": 0, "step": 18779, "train_loss": 3.5248990058898926, "perplexity": 33.95034465825157, "lr": 0.0026291804804649314, "grad_norm": 0.164079, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:34.640048+00:00", "epoch": 0, "step": 18780, "train_loss": 3.5954222679138184, "perplexity": 36.431080416645415, "lr": 0.0026291804804649314, "grad_norm": 0.196463, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:34.944345+00:00", "epoch": 0, "step": 18781, "train_loss": 3.4122114181518555, "perplexity": 30.332247428632137, "lr": 0.0026291804804649314, "grad_norm": 0.16959, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:35.250017+00:00", "epoch": 0, "step": 18782, "train_loss": 3.5130395889282227, "perplexity": 33.550091441002955, "lr": 0.0026291804804649314, "grad_norm": 0.178616, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:35.554563+00:00", "epoch": 0, "step": 18783, "train_loss": 3.613940477371216, "perplexity": 37.112004079417076, "lr": 0.0026291804804649314, "grad_norm": 0.244062, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:35.858457+00:00", "epoch": 0, "step": 18784, "train_loss": 3.442814588546753, "perplexity": 31.274860269672555, "lr": 0.0026291804804649314, "grad_norm": 0.222699, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:36.162543+00:00", "epoch": 0, "step": 18785, "train_loss": 3.5042190551757812, "perplexity": 33.2554630272006, "lr": 0.0026291804804649314, "grad_norm": 0.19024, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:36.467140+00:00", "epoch": 0, "step": 18786, "train_loss": 3.6396005153656006, "perplexity": 38.07662266102597, "lr": 0.0026291804804649314, "grad_norm": 0.17129, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:36.772518+00:00", "epoch": 0, "step": 18787, "train_loss": 3.5460944175720215, "perplexity": 34.6776163749971, "lr": 0.0026291804804649314, "grad_norm": 0.172884, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:37.077945+00:00", "epoch": 0, "step": 18788, "train_loss": 3.601445198059082, "perplexity": 36.65116437896925, "lr": 0.0026291804804649314, "grad_norm": 0.167838, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:37.382386+00:00", "epoch": 0, "step": 18789, "train_loss": 3.547712802886963, "perplexity": 34.73378355790828, "lr": 0.0026291804804649314, "grad_norm": 0.16598, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:37.686487+00:00", "epoch": 0, "step": 18790, "train_loss": 3.455780029296875, "perplexity": 31.682992708615508, "lr": 0.0026291804804649314, "grad_norm": 0.183001, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:37.990702+00:00", "epoch": 0, "step": 18791, "train_loss": 3.482445240020752, "perplexity": 32.53919100805618, "lr": 0.0026291804804649314, "grad_norm": 0.198135, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:38.296017+00:00", "epoch": 0, "step": 18792, "train_loss": 3.4867753982543945, "perplexity": 32.680396354025056, "lr": 0.0026291804804649314, "grad_norm": 0.212099, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:38.601333+00:00", "epoch": 0, "step": 18793, "train_loss": 3.4918673038482666, "perplexity": 32.847226227556526, "lr": 0.0026291804804649314, "grad_norm": 0.163881, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:38.906131+00:00", "epoch": 0, "step": 18794, "train_loss": 3.470167398452759, "perplexity": 32.142122536012025, "lr": 0.0026291804804649314, "grad_norm": 0.186175, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:39.210863+00:00", "epoch": 0, "step": 18795, "train_loss": 3.5294392108917236, "perplexity": 34.104836630110114, "lr": 0.0026291804804649314, "grad_norm": 0.190734, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:39.516049+00:00", "epoch": 0, "step": 18796, "train_loss": 3.5043506622314453, "perplexity": 33.25983996878623, "lr": 0.0026291804804649314, "grad_norm": 0.196714, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:39.821210+00:00", "epoch": 0, "step": 18797, "train_loss": 3.443608522415161, "perplexity": 31.299700299836214, "lr": 0.0026291804804649314, "grad_norm": 0.193498, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:40.126940+00:00", "epoch": 0, "step": 18798, "train_loss": 3.4371516704559326, "perplexity": 31.0982538235434, "lr": 0.0026291804804649314, "grad_norm": 0.16708, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:40.433369+00:00", "epoch": 0, "step": 18799, "train_loss": 3.525151252746582, "perplexity": 33.95890960617032, "lr": 0.0026291804804649314, "grad_norm": 0.188526, "tokens_per_sec": 106936} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:40.739471+00:00", "epoch": 0, "step": 18800, "train_loss": 3.435509204864502, "perplexity": 31.047217935502932, "lr": 0.0026291804804649314, "grad_norm": 0.178101, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:41.043536+00:00", "epoch": 0, "step": 18801, "train_loss": 3.4291625022888184, "perplexity": 30.85079445749882, "lr": 0.0026291804804649314, "grad_norm": 0.173556, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:41.348784+00:00", "epoch": 0, "step": 18802, "train_loss": 3.5386128425598145, "perplexity": 34.41914129285607, "lr": 0.0026291804804649314, "grad_norm": 0.15432, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:41.654736+00:00", "epoch": 0, "step": 18803, "train_loss": 3.5323426723480225, "perplexity": 34.20400260136184, "lr": 0.0026291804804649314, "grad_norm": 0.160533, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:41.959657+00:00", "epoch": 0, "step": 18804, "train_loss": 3.4885623455047607, "perplexity": 32.73884670672574, "lr": 0.0026291804804649314, "grad_norm": 0.167412, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:42.264418+00:00", "epoch": 0, "step": 18805, "train_loss": 3.592947244644165, "perplexity": 36.3410241365418, "lr": 0.0026291804804649314, "grad_norm": 0.174631, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:42.568481+00:00", "epoch": 0, "step": 18806, "train_loss": 3.3899331092834473, "perplexity": 29.663967959930982, "lr": 0.0026291804804649314, "grad_norm": 0.16627, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:42.872506+00:00", "epoch": 0, "step": 18807, "train_loss": 3.526345729827881, "perplexity": 33.9994969809195, "lr": 0.0026291804804649314, "grad_norm": 0.214977, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:43.176920+00:00", "epoch": 0, "step": 18808, "train_loss": 3.4801900386810303, "perplexity": 32.465891264807404, "lr": 0.0026291804804649314, "grad_norm": 0.190733, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:43.481994+00:00", "epoch": 0, "step": 18809, "train_loss": 3.4258878231048584, "perplexity": 30.74993323728167, "lr": 0.0026291804804649314, "grad_norm": 0.196909, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:43.787345+00:00", "epoch": 0, "step": 18810, "train_loss": 3.5319933891296387, "perplexity": 34.1920578034318, "lr": 0.0026291804804649314, "grad_norm": 0.222985, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:44.092345+00:00", "epoch": 0, "step": 18811, "train_loss": 3.518188238143921, "perplexity": 33.72327453993072, "lr": 0.0026291804804649314, "grad_norm": 0.165824, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:44.396522+00:00", "epoch": 0, "step": 18812, "train_loss": 3.5458884239196777, "perplexity": 34.670473741839245, "lr": 0.0026291804804649314, "grad_norm": 0.192463, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:44.700930+00:00", "epoch": 0, "step": 18813, "train_loss": 3.5655264854431152, "perplexity": 35.3580640107415, "lr": 0.0026291804804649314, "grad_norm": 0.17924, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:45.005359+00:00", "epoch": 0, "step": 18814, "train_loss": 3.445600986480713, "perplexity": 31.362125997774083, "lr": 0.0026291804804649314, "grad_norm": 0.173682, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:45.309881+00:00", "epoch": 0, "step": 18815, "train_loss": 3.5050153732299805, "perplexity": 33.281955499627, "lr": 0.0026291804804649314, "grad_norm": 0.188321, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:45.614347+00:00", "epoch": 0, "step": 18816, "train_loss": 3.4583523273468018, "perplexity": 31.764595717610845, "lr": 0.0026291804804649314, "grad_norm": 0.183356, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:45.920194+00:00", "epoch": 0, "step": 18817, "train_loss": 3.4782516956329346, "perplexity": 32.403022180784184, "lr": 0.0026291804804649314, "grad_norm": 0.207895, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:46.224597+00:00", "epoch": 0, "step": 18818, "train_loss": 3.391859769821167, "perplexity": 29.72117544840053, "lr": 0.0026291804804649314, "grad_norm": 0.170979, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:46.529353+00:00", "epoch": 0, "step": 18819, "train_loss": 3.4950215816497803, "perplexity": 32.95099908225301, "lr": 0.0026291804804649314, "grad_norm": 0.227344, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:46.835147+00:00", "epoch": 0, "step": 18820, "train_loss": 3.5293374061584473, "perplexity": 34.10136477304229, "lr": 0.0026291804804649314, "grad_norm": 0.174239, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:47.139159+00:00", "epoch": 0, "step": 18821, "train_loss": 3.3903958797454834, "perplexity": 29.67769874494544, "lr": 0.0026291804804649314, "grad_norm": 0.192731, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:47.442855+00:00", "epoch": 0, "step": 18822, "train_loss": 3.456526756286621, "perplexity": 31.706660089823377, "lr": 0.0026291804804649314, "grad_norm": 0.170276, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:47.747779+00:00", "epoch": 0, "step": 18823, "train_loss": 3.5339279174804688, "perplexity": 34.258267330080415, "lr": 0.0026291804804649314, "grad_norm": 0.197999, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:48.052479+00:00", "epoch": 0, "step": 18824, "train_loss": 3.568053960800171, "perplexity": 35.44754367739407, "lr": 0.0026291804804649314, "grad_norm": 0.150896, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:48.358522+00:00", "epoch": 0, "step": 18825, "train_loss": 3.402663230895996, "perplexity": 30.044007723479094, "lr": 0.0026291804804649314, "grad_norm": 0.19306, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:48.662992+00:00", "epoch": 0, "step": 18826, "train_loss": 3.5282766819000244, "perplexity": 34.065211805736666, "lr": 0.0026291804804649314, "grad_norm": 0.156101, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:48.966963+00:00", "epoch": 0, "step": 18827, "train_loss": 3.4881203174591064, "perplexity": 32.724378416229655, "lr": 0.0026291804804649314, "grad_norm": 0.175493, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:49.270884+00:00", "epoch": 0, "step": 18828, "train_loss": 3.5429253578186035, "perplexity": 34.56789488540418, "lr": 0.0026291804804649314, "grad_norm": 0.164099, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:49.576701+00:00", "epoch": 0, "step": 18829, "train_loss": 3.5490612983703613, "perplexity": 34.78065350300958, "lr": 0.0026291804804649314, "grad_norm": 0.154407, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:49.882245+00:00", "epoch": 0, "step": 18830, "train_loss": 3.4816741943359375, "perplexity": 32.514111475217355, "lr": 0.0026291804804649314, "grad_norm": 0.172031, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:50.187389+00:00", "epoch": 0, "step": 18831, "train_loss": 3.625220537185669, "perplexity": 37.532999669045125, "lr": 0.0026291804804649314, "grad_norm": 0.15719, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:50.491960+00:00", "epoch": 0, "step": 18832, "train_loss": 3.5093863010406494, "perplexity": 33.427746914276064, "lr": 0.0026291804804649314, "grad_norm": 0.152942, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:50.796648+00:00", "epoch": 0, "step": 18833, "train_loss": 3.483773946762085, "perplexity": 32.58245478657815, "lr": 0.0026291804804649314, "grad_norm": 0.155204, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:51.101119+00:00", "epoch": 0, "step": 18834, "train_loss": 3.5075182914733887, "perplexity": 33.36536184931817, "lr": 0.0026291804804649314, "grad_norm": 0.153176, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:51.406203+00:00", "epoch": 0, "step": 18835, "train_loss": 3.5362446308135986, "perplexity": 34.33772592062309, "lr": 0.0026291804804649314, "grad_norm": 0.167197, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:51.710730+00:00", "epoch": 0, "step": 18836, "train_loss": 3.5116350650787354, "perplexity": 33.503002613852935, "lr": 0.0026291804804649314, "grad_norm": 0.174946, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:52.015221+00:00", "epoch": 0, "step": 18837, "train_loss": 3.543431282043457, "perplexity": 34.58538804556252, "lr": 0.0026291804804649314, "grad_norm": 0.185486, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:52.319950+00:00", "epoch": 0, "step": 18838, "train_loss": 3.569486379623413, "perplexity": 35.49835578962015, "lr": 0.0026291804804649314, "grad_norm": 0.181556, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:52.623867+00:00", "epoch": 0, "step": 18839, "train_loss": 3.4059033393859863, "perplexity": 30.141511443991636, "lr": 0.0026291804804649314, "grad_norm": 0.179529, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:52.928712+00:00", "epoch": 0, "step": 18840, "train_loss": 3.569946527481079, "perplexity": 35.51469404070474, "lr": 0.0026291804804649314, "grad_norm": 0.178022, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:53.233203+00:00", "epoch": 0, "step": 18841, "train_loss": 3.5282976627349854, "perplexity": 34.0659265298212, "lr": 0.0026291804804649314, "grad_norm": 0.167831, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:53.538468+00:00", "epoch": 0, "step": 18842, "train_loss": 3.526564598083496, "perplexity": 34.0069392059192, "lr": 0.0026291804804649314, "grad_norm": 0.213565, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:53.843467+00:00", "epoch": 0, "step": 18843, "train_loss": 3.5239100456237793, "perplexity": 33.916785713330434, "lr": 0.0026291804804649314, "grad_norm": 0.203851, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:54.148143+00:00", "epoch": 0, "step": 18844, "train_loss": 3.5948352813720703, "perplexity": 36.409702137734335, "lr": 0.0026291804804649314, "grad_norm": 0.200134, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:54.453252+00:00", "epoch": 0, "step": 18845, "train_loss": 3.5912561416625977, "perplexity": 36.27961965755389, "lr": 0.0026291804804649314, "grad_norm": 0.171583, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:54.758964+00:00", "epoch": 0, "step": 18846, "train_loss": 3.4213790893554688, "perplexity": 30.611602059069934, "lr": 0.0026291804804649314, "grad_norm": 0.183149, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:55.064953+00:00", "epoch": 0, "step": 18847, "train_loss": 3.50771427154541, "perplexity": 33.37190143612984, "lr": 0.0026291804804649314, "grad_norm": 0.204871, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:55.370280+00:00", "epoch": 0, "step": 18848, "train_loss": 3.4562249183654785, "perplexity": 31.697091261645742, "lr": 0.0026291804804649314, "grad_norm": 0.165479, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:55.675325+00:00", "epoch": 0, "step": 18849, "train_loss": 3.403286933898926, "perplexity": 30.062752106172518, "lr": 0.0026291804804649314, "grad_norm": 0.161414, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:55.980270+00:00", "epoch": 0, "step": 18850, "train_loss": 3.473435401916504, "perplexity": 32.24733492738388, "lr": 0.0026291804804649314, "grad_norm": 0.171379, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:56.286754+00:00", "epoch": 0, "step": 18851, "train_loss": 3.565124034881592, "perplexity": 35.343837001052044, "lr": 0.0026291804804649314, "grad_norm": 0.17042, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:56.592510+00:00", "epoch": 0, "step": 18852, "train_loss": 3.5905842781066895, "perplexity": 36.25525288977199, "lr": 0.0026291804804649314, "grad_norm": 0.176991, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:56.898183+00:00", "epoch": 0, "step": 18853, "train_loss": 3.5825600624084473, "perplexity": 35.965497006974566, "lr": 0.0026291804804649314, "grad_norm": 0.176647, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:57.202142+00:00", "epoch": 0, "step": 18854, "train_loss": 3.6025235652923584, "perplexity": 36.69070911173686, "lr": 0.0026291804804649314, "grad_norm": 0.179129, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:57.506974+00:00", "epoch": 0, "step": 18855, "train_loss": 3.4843480587005615, "perplexity": 32.60116613354589, "lr": 0.0026291804804649314, "grad_norm": 0.164332, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:57.812972+00:00", "epoch": 0, "step": 18856, "train_loss": 3.556541919708252, "perplexity": 35.041809990293636, "lr": 0.0026291804804649314, "grad_norm": 0.188953, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:58.120033+00:00", "epoch": 0, "step": 18857, "train_loss": 3.4382238388061523, "perplexity": 31.13161426785059, "lr": 0.0026291804804649314, "grad_norm": 0.184339, "tokens_per_sec": 106721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:58.425568+00:00", "epoch": 0, "step": 18858, "train_loss": 3.540034770965576, "perplexity": 34.46811765975393, "lr": 0.0026291804804649314, "grad_norm": 0.158839, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:58.730831+00:00", "epoch": 0, "step": 18859, "train_loss": 3.5505409240722656, "perplexity": 34.83215394315332, "lr": 0.0026291804804649314, "grad_norm": 0.177853, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:59.035480+00:00", "epoch": 0, "step": 18860, "train_loss": 3.510357618331909, "perplexity": 33.4602316367941, "lr": 0.0026291804804649314, "grad_norm": 0.162953, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:59.341341+00:00", "epoch": 0, "step": 18861, "train_loss": 3.575413942337036, "perplexity": 35.70939938894348, "lr": 0.0026291804804649314, "grad_norm": 0.170808, "tokens_per_sec": 107039} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:59.647851+00:00", "epoch": 0, "step": 18862, "train_loss": 3.613152265548706, "perplexity": 37.08276348444985, "lr": 0.0026291804804649314, "grad_norm": 0.177143, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:36:59.952945+00:00", "epoch": 0, "step": 18863, "train_loss": 3.514286518096924, "perplexity": 33.59195212186146, "lr": 0.0026291804804649314, "grad_norm": 0.158144, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:00.258068+00:00", "epoch": 0, "step": 18864, "train_loss": 3.4962193965911865, "perplexity": 32.99049192912713, "lr": 0.0026291804804649314, "grad_norm": 0.169441, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:00.563503+00:00", "epoch": 0, "step": 18865, "train_loss": 3.5576975345611572, "perplexity": 35.08232823362385, "lr": 0.0026291804804649314, "grad_norm": 0.161731, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:00.869607+00:00", "epoch": 0, "step": 18866, "train_loss": 3.5220322608947754, "perplexity": 33.85315705022362, "lr": 0.0026291804804649314, "grad_norm": 0.18338, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:01.174743+00:00", "epoch": 0, "step": 18867, "train_loss": 3.532212495803833, "perplexity": 34.199550332302536, "lr": 0.0026291804804649314, "grad_norm": 0.181445, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:01.479578+00:00", "epoch": 0, "step": 18868, "train_loss": 3.5034494400024414, "perplexity": 33.229878964462664, "lr": 0.0026291804804649314, "grad_norm": 0.178566, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:01.784353+00:00", "epoch": 0, "step": 18869, "train_loss": 3.5194454193115234, "perplexity": 33.765697466657016, "lr": 0.0026291804804649314, "grad_norm": 0.183693, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:02.090468+00:00", "epoch": 0, "step": 18870, "train_loss": 3.430415391921997, "perplexity": 30.88947132191182, "lr": 0.0026291804804649314, "grad_norm": 0.181069, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:02.396408+00:00", "epoch": 0, "step": 18871, "train_loss": 3.5214450359344482, "perplexity": 33.83328346712097, "lr": 0.0026291804804649314, "grad_norm": 0.197212, "tokens_per_sec": 107047} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:02.701731+00:00", "epoch": 0, "step": 18872, "train_loss": 3.4768118858337402, "perplexity": 32.356401562390296, "lr": 0.0026291804804649314, "grad_norm": 0.18912, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:03.006859+00:00", "epoch": 0, "step": 18873, "train_loss": 3.460547685623169, "perplexity": 31.834406988051615, "lr": 0.0026291804804649314, "grad_norm": 0.177863, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:03.312835+00:00", "epoch": 0, "step": 18874, "train_loss": 3.4542598724365234, "perplexity": 31.634866179097916, "lr": 0.0026291804804649314, "grad_norm": 0.171485, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:03.618896+00:00", "epoch": 0, "step": 18875, "train_loss": 3.583007574081421, "perplexity": 35.98159558859241, "lr": 0.0026291804804649314, "grad_norm": 0.184483, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:03.925027+00:00", "epoch": 0, "step": 18876, "train_loss": 3.458512783050537, "perplexity": 31.76969293709862, "lr": 0.0026291804804649314, "grad_norm": 0.170967, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:04.230059+00:00", "epoch": 0, "step": 18877, "train_loss": 3.5215954780578613, "perplexity": 33.83837380101852, "lr": 0.0026291804804649314, "grad_norm": 0.182895, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:04.534521+00:00", "epoch": 0, "step": 18878, "train_loss": 3.449946641921997, "perplexity": 31.49871155319934, "lr": 0.0026291804804649314, "grad_norm": 0.182065, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:04.839522+00:00", "epoch": 0, "step": 18879, "train_loss": 3.438433885574341, "perplexity": 31.13815404962201, "lr": 0.0026291804804649314, "grad_norm": 0.1665, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:05.145098+00:00", "epoch": 0, "step": 18880, "train_loss": 3.5781023502349854, "perplexity": 35.80552998151495, "lr": 0.0026291804804649314, "grad_norm": 0.176297, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:05.449900+00:00", "epoch": 0, "step": 18881, "train_loss": 3.5225830078125, "perplexity": 33.87180670727344, "lr": 0.0026291804804649314, "grad_norm": 0.164537, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:05.754772+00:00", "epoch": 0, "step": 18882, "train_loss": 3.5342214107513428, "perplexity": 34.26832337663241, "lr": 0.0026291804804649314, "grad_norm": 0.173209, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:06.058698+00:00", "epoch": 0, "step": 18883, "train_loss": 3.473231554031372, "perplexity": 32.24076204631451, "lr": 0.0026291804804649314, "grad_norm": 0.194721, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:06.364035+00:00", "epoch": 0, "step": 18884, "train_loss": 3.4710886478424072, "perplexity": 32.1717470904865, "lr": 0.0026291804804649314, "grad_norm": 0.160546, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:06.669205+00:00", "epoch": 0, "step": 18885, "train_loss": 3.4332404136657715, "perplexity": 30.97685812674414, "lr": 0.0026291804804649314, "grad_norm": 0.163194, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:06.974064+00:00", "epoch": 0, "step": 18886, "train_loss": 3.467679023742676, "perplexity": 32.06224032082633, "lr": 0.0026291804804649314, "grad_norm": 0.172422, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:07.278228+00:00", "epoch": 0, "step": 18887, "train_loss": 3.6196436882019043, "perplexity": 37.32426637599826, "lr": 0.0026291804804649314, "grad_norm": 0.155919, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:07.582514+00:00", "epoch": 0, "step": 18888, "train_loss": 3.6011433601379395, "perplexity": 36.64010333711057, "lr": 0.0026291804804649314, "grad_norm": 0.174071, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:07.888163+00:00", "epoch": 0, "step": 18889, "train_loss": 3.3902840614318848, "perplexity": 29.674380420248486, "lr": 0.0026291804804649314, "grad_norm": 0.164091, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:08.193857+00:00", "epoch": 0, "step": 18890, "train_loss": 3.5209505558013916, "perplexity": 33.816557716226974, "lr": 0.0026291804804649314, "grad_norm": 0.181074, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:08.498441+00:00", "epoch": 0, "step": 18891, "train_loss": 3.4985241889953613, "perplexity": 33.06661585556113, "lr": 0.0026291804804649314, "grad_norm": 0.172257, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:08.802676+00:00", "epoch": 0, "step": 18892, "train_loss": 3.47251558303833, "perplexity": 32.21768685746444, "lr": 0.0026291804804649314, "grad_norm": 0.177521, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:09.106811+00:00", "epoch": 0, "step": 18893, "train_loss": 3.523770809173584, "perplexity": 33.9120635892392, "lr": 0.0026291804804649314, "grad_norm": 0.160729, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:09.410696+00:00", "epoch": 0, "step": 18894, "train_loss": 3.4732866287231445, "perplexity": 32.242537745244306, "lr": 0.0026291804804649314, "grad_norm": 0.156769, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:09.715136+00:00", "epoch": 0, "step": 18895, "train_loss": 3.499999523162842, "perplexity": 33.11543616801807, "lr": 0.0026291804804649314, "grad_norm": 0.176536, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:10.020646+00:00", "epoch": 0, "step": 18896, "train_loss": 3.42503023147583, "perplexity": 30.72357365643768, "lr": 0.0026291804804649314, "grad_norm": 0.152163, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:10.325855+00:00", "epoch": 0, "step": 18897, "train_loss": 3.4921131134033203, "perplexity": 32.85530138205468, "lr": 0.0026291804804649314, "grad_norm": 0.16504, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:10.631430+00:00", "epoch": 0, "step": 18898, "train_loss": 3.4493813514709473, "perplexity": 31.480910664148272, "lr": 0.0026291804804649314, "grad_norm": 0.157937, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:10.935802+00:00", "epoch": 0, "step": 18899, "train_loss": 3.4855194091796875, "perplexity": 32.639375899261324, "lr": 0.0026291804804649314, "grad_norm": 0.170144, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:11.240470+00:00", "epoch": 0, "step": 18900, "train_loss": 3.3971517086029053, "perplexity": 29.878874989527457, "lr": 0.0026291804804649314, "grad_norm": 0.189774, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:11.545732+00:00", "epoch": 0, "step": 18901, "train_loss": 3.4598138332366943, "perplexity": 31.81105380245497, "lr": 0.0026291804804649314, "grad_norm": 0.190579, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:11.850321+00:00", "epoch": 0, "step": 18902, "train_loss": 3.466311454772949, "perplexity": 32.018422964321296, "lr": 0.0026291804804649314, "grad_norm": 0.196602, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:12.155108+00:00", "epoch": 0, "step": 18903, "train_loss": 3.5447444915771484, "perplexity": 34.63083574151288, "lr": 0.0026291804804649314, "grad_norm": 0.160405, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:12.459984+00:00", "epoch": 0, "step": 18904, "train_loss": 3.5630414485931396, "perplexity": 35.270307003590595, "lr": 0.0026291804804649314, "grad_norm": 0.161209, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:12.765488+00:00", "epoch": 0, "step": 18905, "train_loss": 3.4925925731658936, "perplexity": 32.87105795407388, "lr": 0.0026291804804649314, "grad_norm": 0.18357, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:13.070885+00:00", "epoch": 0, "step": 18906, "train_loss": 3.531675100326538, "perplexity": 34.18117658605469, "lr": 0.0026291804804649314, "grad_norm": 0.155174, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:13.376389+00:00", "epoch": 0, "step": 18907, "train_loss": 3.4105727672576904, "perplexity": 30.282584165749125, "lr": 0.0026291804804649314, "grad_norm": 0.18874, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:13.681649+00:00", "epoch": 0, "step": 18908, "train_loss": 3.393266201019287, "perplexity": 29.763005645547928, "lr": 0.0026291804804649314, "grad_norm": 0.162329, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:13.986826+00:00", "epoch": 0, "step": 18909, "train_loss": 3.5680408477783203, "perplexity": 35.4470788560269, "lr": 0.0026291804804649314, "grad_norm": 0.212171, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:14.292035+00:00", "epoch": 0, "step": 18910, "train_loss": 3.481565237045288, "perplexity": 32.51056901871488, "lr": 0.0026291804804649314, "grad_norm": 0.182901, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:14.597794+00:00", "epoch": 0, "step": 18911, "train_loss": 3.6003081798553467, "perplexity": 36.60951502040876, "lr": 0.0026291804804649314, "grad_norm": 0.180346, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:14.903361+00:00", "epoch": 0, "step": 18912, "train_loss": 3.5166866779327393, "perplexity": 33.672675011317395, "lr": 0.0026291804804649314, "grad_norm": 0.195837, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:15.209813+00:00", "epoch": 0, "step": 18913, "train_loss": 3.446676015853882, "perplexity": 31.395859333341242, "lr": 0.0026291804804649314, "grad_norm": 0.163352, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:15.515180+00:00", "epoch": 0, "step": 18914, "train_loss": 3.5084726810455322, "perplexity": 33.39722060315094, "lr": 0.0026291804804649314, "grad_norm": 0.170863, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:15.819626+00:00", "epoch": 0, "step": 18915, "train_loss": 3.453192710876465, "perplexity": 31.601124672975306, "lr": 0.0026291804804649314, "grad_norm": 0.149328, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:16.124980+00:00", "epoch": 0, "step": 18916, "train_loss": 3.5545706748962402, "perplexity": 34.97280204227249, "lr": 0.0026291804804649314, "grad_norm": 0.163906, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:16.429882+00:00", "epoch": 0, "step": 18917, "train_loss": 3.6692092418670654, "perplexity": 39.22087936535563, "lr": 0.0026291804804649314, "grad_norm": 0.186569, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:16.734752+00:00", "epoch": 0, "step": 18918, "train_loss": 3.644233465194702, "perplexity": 38.25343901786252, "lr": 0.0026291804804649314, "grad_norm": 0.154476, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:17.038864+00:00", "epoch": 0, "step": 18919, "train_loss": 3.5177054405212402, "perplexity": 33.70699695286019, "lr": 0.0026291804804649314, "grad_norm": 0.17877, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:17.343492+00:00", "epoch": 0, "step": 18920, "train_loss": 3.5160560607910156, "perplexity": 33.65144713928105, "lr": 0.0026291804804649314, "grad_norm": 0.197282, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:17.648653+00:00", "epoch": 0, "step": 18921, "train_loss": 3.6420531272888184, "perplexity": 38.17012445469712, "lr": 0.0026291804804649314, "grad_norm": 0.174751, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:17.954400+00:00", "epoch": 0, "step": 18922, "train_loss": 3.5384247303009033, "perplexity": 34.41266723938119, "lr": 0.0026291804804649314, "grad_norm": 0.167902, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:18.258373+00:00", "epoch": 0, "step": 18923, "train_loss": 3.437840700149536, "perplexity": 31.119688827676157, "lr": 0.0026291804804649314, "grad_norm": 0.18841, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:18.563173+00:00", "epoch": 0, "step": 18924, "train_loss": 3.437920570373535, "perplexity": 31.122174463456428, "lr": 0.0026291804804649314, "grad_norm": 0.178913, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:18.868550+00:00", "epoch": 0, "step": 18925, "train_loss": 3.4848709106445312, "perplexity": 32.61821617356936, "lr": 0.0026291804804649314, "grad_norm": 0.166945, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:19.172728+00:00", "epoch": 0, "step": 18926, "train_loss": 3.4535810947418213, "perplexity": 31.613400423622842, "lr": 0.0026291804804649314, "grad_norm": 0.161464, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:19.477166+00:00", "epoch": 0, "step": 18927, "train_loss": 3.4451305866241455, "perplexity": 31.347376727502365, "lr": 0.0026291804804649314, "grad_norm": 0.149651, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:19.782317+00:00", "epoch": 0, "step": 18928, "train_loss": 3.5339810848236084, "perplexity": 34.26008879955584, "lr": 0.0026291804804649314, "grad_norm": 0.165152, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:20.086571+00:00", "epoch": 0, "step": 18929, "train_loss": 3.5082204341888428, "perplexity": 33.388797321649434, "lr": 0.0026291804804649314, "grad_norm": 0.21745, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:20.391891+00:00", "epoch": 0, "step": 18930, "train_loss": 3.4885878562927246, "perplexity": 32.73968191115557, "lr": 0.0026291804804649314, "grad_norm": 0.19371, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:20.696675+00:00", "epoch": 0, "step": 18931, "train_loss": 3.4991393089294434, "perplexity": 33.086962047159, "lr": 0.0026291804804649314, "grad_norm": 0.193448, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:21.002553+00:00", "epoch": 0, "step": 18932, "train_loss": 3.4478440284729004, "perplexity": 31.432551517523134, "lr": 0.0026291804804649314, "grad_norm": 0.185723, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:21.308043+00:00", "epoch": 0, "step": 18933, "train_loss": 3.495877504348755, "perplexity": 32.97921466377718, "lr": 0.0026291804804649314, "grad_norm": 0.171329, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:21.613213+00:00", "epoch": 0, "step": 18934, "train_loss": 3.4017879962921143, "perplexity": 30.017723672313643, "lr": 0.0026291804804649314, "grad_norm": 0.16807, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:21.917514+00:00", "epoch": 0, "step": 18935, "train_loss": 3.455271005630493, "perplexity": 31.666869419421047, "lr": 0.0026291804804649314, "grad_norm": 0.169944, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:22.222447+00:00", "epoch": 0, "step": 18936, "train_loss": 3.490260362625122, "perplexity": 32.794485052998574, "lr": 0.0026291804804649314, "grad_norm": 0.176128, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:22.526886+00:00", "epoch": 0, "step": 18937, "train_loss": 3.565772294998169, "perplexity": 35.36675642901951, "lr": 0.0026291804804649314, "grad_norm": 0.204333, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:22.831771+00:00", "epoch": 0, "step": 18938, "train_loss": 3.464231014251709, "perplexity": 31.951879783326383, "lr": 0.0026291804804649314, "grad_norm": 0.194613, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:23.136622+00:00", "epoch": 0, "step": 18939, "train_loss": 3.5005040168762207, "perplexity": 33.13214691225892, "lr": 0.0026291804804649314, "grad_norm": 0.18463, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:23.440403+00:00", "epoch": 0, "step": 18940, "train_loss": 3.494929075241089, "perplexity": 32.947951044648796, "lr": 0.0026291804804649314, "grad_norm": 0.203257, "tokens_per_sec": 107930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:23.744722+00:00", "epoch": 0, "step": 18941, "train_loss": 3.429428815841675, "perplexity": 30.8590115362904, "lr": 0.0026291804804649314, "grad_norm": 0.195725, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:24.050716+00:00", "epoch": 0, "step": 18942, "train_loss": 3.571432590484619, "perplexity": 35.567510348060964, "lr": 0.0026291804804649314, "grad_norm": 0.165371, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:24.355619+00:00", "epoch": 0, "step": 18943, "train_loss": 3.501149892807007, "perplexity": 33.15355308060513, "lr": 0.0026291804804649314, "grad_norm": 0.167207, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:24.661987+00:00", "epoch": 0, "step": 18944, "train_loss": 3.4898531436920166, "perplexity": 32.78113323652429, "lr": 0.0026291804804649314, "grad_norm": 0.156833, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:24.966810+00:00", "epoch": 0, "step": 18945, "train_loss": 3.5586225986480713, "perplexity": 35.114796650926124, "lr": 0.0026291804804649314, "grad_norm": 0.198095, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:25.271776+00:00", "epoch": 0, "step": 18946, "train_loss": 3.5923492908477783, "perplexity": 36.31930037874377, "lr": 0.0026291804804649314, "grad_norm": 0.186141, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:25.576498+00:00", "epoch": 0, "step": 18947, "train_loss": 3.4781746864318848, "perplexity": 32.40052694601369, "lr": 0.0026291804804649314, "grad_norm": 0.166217, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:25.881523+00:00", "epoch": 0, "step": 18948, "train_loss": 3.4897215366363525, "perplexity": 32.77681929197669, "lr": 0.0026291804804649314, "grad_norm": 0.169676, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:26.187603+00:00", "epoch": 0, "step": 18949, "train_loss": 3.5300722122192383, "perplexity": 34.12643187115868, "lr": 0.0026291804804649314, "grad_norm": 0.158807, "tokens_per_sec": 107057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:26.492447+00:00", "epoch": 0, "step": 18950, "train_loss": 3.5322775840759277, "perplexity": 34.20177639438457, "lr": 0.0026291804804649314, "grad_norm": 0.176658, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:26.797368+00:00", "epoch": 0, "step": 18951, "train_loss": 3.4938807487487793, "perplexity": 32.91342893308505, "lr": 0.0026291804804649314, "grad_norm": 0.169363, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:27.103141+00:00", "epoch": 0, "step": 18952, "train_loss": 3.558932065963745, "perplexity": 35.12566521443196, "lr": 0.0026291804804649314, "grad_norm": 0.177526, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:27.408104+00:00", "epoch": 0, "step": 18953, "train_loss": 3.562774658203125, "perplexity": 35.26089847973685, "lr": 0.0026291804804649314, "grad_norm": 0.149053, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:27.713926+00:00", "epoch": 0, "step": 18954, "train_loss": 3.6203086376190186, "perplexity": 37.34909337860487, "lr": 0.0026291804804649314, "grad_norm": 0.173781, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:28.018929+00:00", "epoch": 0, "step": 18955, "train_loss": 3.465346336364746, "perplexity": 31.98753630195746, "lr": 0.0026291804804649314, "grad_norm": 0.187433, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:28.322577+00:00", "epoch": 0, "step": 18956, "train_loss": 3.5039212703704834, "perplexity": 33.24556152994884, "lr": 0.0026291804804649314, "grad_norm": 0.223549, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:28.626584+00:00", "epoch": 0, "step": 18957, "train_loss": 3.5375421047210693, "perplexity": 34.382307139276826, "lr": 0.0026291804804649314, "grad_norm": 0.185359, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:28.930798+00:00", "epoch": 0, "step": 18958, "train_loss": 3.565216541290283, "perplexity": 35.34710668371335, "lr": 0.0026291804804649314, "grad_norm": 0.179952, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:29.235665+00:00", "epoch": 0, "step": 18959, "train_loss": 3.5830748081207275, "perplexity": 35.984014857932245, "lr": 0.0026291804804649314, "grad_norm": 0.149049, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:29.541512+00:00", "epoch": 0, "step": 18960, "train_loss": 3.4896035194396973, "perplexity": 32.77295129189833, "lr": 0.0026291804804649314, "grad_norm": 0.164261, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:29.846322+00:00", "epoch": 0, "step": 18961, "train_loss": 3.5521769523620605, "perplexity": 34.8891869735032, "lr": 0.0026291804804649314, "grad_norm": 0.160684, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:30.150894+00:00", "epoch": 0, "step": 18962, "train_loss": 3.5207931995391846, "perplexity": 33.81123688774786, "lr": 0.0026291804804649314, "grad_norm": 0.153935, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:30.455635+00:00", "epoch": 0, "step": 18963, "train_loss": 3.6823973655700684, "perplexity": 39.74155499488191, "lr": 0.0026291804804649314, "grad_norm": 0.155532, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:30.760155+00:00", "epoch": 0, "step": 18964, "train_loss": 3.6205008029937744, "perplexity": 37.35627127077984, "lr": 0.0026291804804649314, "grad_norm": 0.172387, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:31.064724+00:00", "epoch": 0, "step": 18965, "train_loss": 3.5752739906311035, "perplexity": 35.70440214727557, "lr": 0.0026291804804649314, "grad_norm": 0.178005, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:31.369131+00:00", "epoch": 0, "step": 18966, "train_loss": 3.512371778488159, "perplexity": 33.5276938191888, "lr": 0.0026291804804649314, "grad_norm": 0.162977, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:31.673384+00:00", "epoch": 0, "step": 18967, "train_loss": 3.546450614929199, "perplexity": 34.68997065045261, "lr": 0.0026291804804649314, "grad_norm": 0.14698, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:31.977791+00:00", "epoch": 0, "step": 18968, "train_loss": 3.445822238922119, "perplexity": 31.369065712404876, "lr": 0.0026291804804649314, "grad_norm": 0.186592, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:32.282409+00:00", "epoch": 0, "step": 18969, "train_loss": 3.479037284851074, "perplexity": 32.42848764703274, "lr": 0.0026291804804649314, "grad_norm": 0.160136, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:32.587901+00:00", "epoch": 0, "step": 18970, "train_loss": 3.511880874633789, "perplexity": 33.511238984266214, "lr": 0.0026291804804649314, "grad_norm": 0.158064, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:32.892347+00:00", "epoch": 0, "step": 18971, "train_loss": 3.5182132720947266, "perplexity": 33.72411877729381, "lr": 0.0026291804804649314, "grad_norm": 0.154614, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:33.197204+00:00", "epoch": 0, "step": 18972, "train_loss": 3.407423257827759, "perplexity": 30.18735891648151, "lr": 0.0026291804804649314, "grad_norm": 0.166187, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:33.501593+00:00", "epoch": 0, "step": 18973, "train_loss": 3.522397756576538, "perplexity": 33.865532494387445, "lr": 0.0026291804804649314, "grad_norm": 0.163211, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:33.807079+00:00", "epoch": 0, "step": 18974, "train_loss": 3.5021073818206787, "perplexity": 33.18531244563917, "lr": 0.0026291804804649314, "grad_norm": 0.154804, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:34.112392+00:00", "epoch": 0, "step": 18975, "train_loss": 3.555495023727417, "perplexity": 35.00514405631187, "lr": 0.0026291804804649314, "grad_norm": 0.173219, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:34.417656+00:00", "epoch": 0, "step": 18976, "train_loss": 3.564314126968384, "perplexity": 35.315223336562504, "lr": 0.0026291804804649314, "grad_norm": 0.185964, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:34.722295+00:00", "epoch": 0, "step": 18977, "train_loss": 3.4897637367248535, "perplexity": 32.778202505837264, "lr": 0.0026291804804649314, "grad_norm": 0.209819, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:35.027299+00:00", "epoch": 0, "step": 18978, "train_loss": 3.493744134902954, "perplexity": 32.90893281010253, "lr": 0.0026291804804649314, "grad_norm": 0.187444, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:35.332437+00:00", "epoch": 0, "step": 18979, "train_loss": 3.4948112964630127, "perplexity": 32.944070703750036, "lr": 0.0026291804804649314, "grad_norm": 0.186781, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:35.637918+00:00", "epoch": 0, "step": 18980, "train_loss": 3.6246235370635986, "perplexity": 37.51059915088239, "lr": 0.0026291804804649314, "grad_norm": 0.160585, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:35.943654+00:00", "epoch": 0, "step": 18981, "train_loss": 3.4624106884002686, "perplexity": 31.893769856102075, "lr": 0.0026291804804649314, "grad_norm": 0.181906, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:36.247810+00:00", "epoch": 0, "step": 18982, "train_loss": 3.4174773693084717, "perplexity": 30.492396861525698, "lr": 0.0026291804804649314, "grad_norm": 0.17128, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:36.551193+00:00", "epoch": 0, "step": 18983, "train_loss": 3.436408758163452, "perplexity": 31.075159128231927, "lr": 0.0026291804804649314, "grad_norm": 0.203926, "tokens_per_sec": 108009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:36.855629+00:00", "epoch": 0, "step": 18984, "train_loss": 3.5613341331481934, "perplexity": 35.21014083962086, "lr": 0.0026291804804649314, "grad_norm": 0.214382, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:37.160628+00:00", "epoch": 0, "step": 18985, "train_loss": 3.455728054046631, "perplexity": 31.681346019934892, "lr": 0.0026291804804649314, "grad_norm": 0.176658, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:37.466854+00:00", "epoch": 0, "step": 18986, "train_loss": 3.544703960418701, "perplexity": 34.62943214206723, "lr": 0.0026291804804649314, "grad_norm": 0.158464, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:37.771642+00:00", "epoch": 0, "step": 18987, "train_loss": 3.4871153831481934, "perplexity": 32.691509084081915, "lr": 0.0026291804804649314, "grad_norm": 0.190883, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:38.075226+00:00", "epoch": 0, "step": 18988, "train_loss": 3.363955020904541, "perplexity": 28.903278191132056, "lr": 0.0026291804804649314, "grad_norm": 0.175863, "tokens_per_sec": 107937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:38.379449+00:00", "epoch": 0, "step": 18989, "train_loss": 3.4514262676239014, "perplexity": 31.545352353344455, "lr": 0.0026291804804649314, "grad_norm": 0.168597, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:38.684566+00:00", "epoch": 0, "step": 18990, "train_loss": 3.449474334716797, "perplexity": 31.483837997498497, "lr": 0.0026291804804649314, "grad_norm": 0.157745, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:38.990752+00:00", "epoch": 0, "step": 18991, "train_loss": 3.48630690574646, "perplexity": 32.66508941904901, "lr": 0.0026291804804649314, "grad_norm": 0.171223, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:39.295718+00:00", "epoch": 0, "step": 18992, "train_loss": 3.4868874549865723, "perplexity": 32.684058617633575, "lr": 0.0026291804804649314, "grad_norm": 0.173504, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:39.600579+00:00", "epoch": 0, "step": 18993, "train_loss": 3.5740811824798584, "perplexity": 35.66183903521901, "lr": 0.0026291804804649314, "grad_norm": 0.191415, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:39.905428+00:00", "epoch": 0, "step": 18994, "train_loss": 3.427760601043701, "perplexity": 30.807574992125524, "lr": 0.0026291804804649314, "grad_norm": 0.177646, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:40.209679+00:00", "epoch": 0, "step": 18995, "train_loss": 3.461998224258423, "perplexity": 31.88061753230559, "lr": 0.0026291804804649314, "grad_norm": 0.170033, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:40.515580+00:00", "epoch": 0, "step": 18996, "train_loss": 3.5998573303222656, "perplexity": 36.593013357818435, "lr": 0.0026291804804649314, "grad_norm": 0.200527, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:40.820468+00:00", "epoch": 0, "step": 18997, "train_loss": 3.4351043701171875, "perplexity": 31.03465148671438, "lr": 0.0026291804804649314, "grad_norm": 0.14357, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:41.125220+00:00", "epoch": 0, "step": 18998, "train_loss": 3.490105628967285, "perplexity": 32.78941103494026, "lr": 0.0026291804804649314, "grad_norm": 0.187717, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:41.430139+00:00", "epoch": 0, "step": 18999, "train_loss": 3.518939733505249, "perplexity": 33.748626949232126, "lr": 0.0026291804804649314, "grad_norm": 0.177613, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:41.734964+00:00", "epoch": 0, "step": 19000, "train_loss": 3.550948143005371, "perplexity": 34.84634114417219, "lr": 0.0026291804804649314, "grad_norm": 0.168756, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:37:44.809242+00:00", "step": 19000, "epoch": 0, "val_loss": 3.460472512245178, "val_ppl": 31.832013978088433, "eval_train_loss": 3.550948143005371, "eval_train_ppl": 34.84634114417219} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:37:45.758323+00:00", "step": 19000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4605_epoch_0000_step_0019000.pt", "val_loss": 3.460472512245178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:47.155043+00:00", "epoch": 0, "step": 19001, "train_loss": 3.472207546234131, "perplexity": 32.207764152523794, "lr": 0.0026291804804649314, "grad_norm": 0.16017, "tokens_per_sec": 6045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:47.458492+00:00", "epoch": 0, "step": 19002, "train_loss": 3.4870266914367676, "perplexity": 32.68860974676765, "lr": 0.0026291804804649314, "grad_norm": 0.176238, "tokens_per_sec": 107979} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:47.761709+00:00", "epoch": 0, "step": 19003, "train_loss": 3.448394298553467, "perplexity": 31.449852669895783, "lr": 0.0026291804804649314, "grad_norm": 0.169573, "tokens_per_sec": 108068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:48.066082+00:00", "epoch": 0, "step": 19004, "train_loss": 3.4607255458831787, "perplexity": 31.840069567514767, "lr": 0.0026291804804649314, "grad_norm": 0.159317, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:48.371274+00:00", "epoch": 0, "step": 19005, "train_loss": 3.479322671890259, "perplexity": 32.437743637814634, "lr": 0.0026291804804649314, "grad_norm": 0.161327, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:48.676510+00:00", "epoch": 0, "step": 19006, "train_loss": 3.533997058868408, "perplexity": 34.26063607612027, "lr": 0.0026291804804649314, "grad_norm": 0.163369, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:48.979966+00:00", "epoch": 0, "step": 19007, "train_loss": 3.4425642490386963, "perplexity": 31.267031916452034, "lr": 0.0026291804804649314, "grad_norm": 0.1566, "tokens_per_sec": 107982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:49.284159+00:00", "epoch": 0, "step": 19008, "train_loss": 3.5418784618377686, "perplexity": 34.531724831628196, "lr": 0.0026291804804649314, "grad_norm": 0.176036, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:49.588247+00:00", "epoch": 0, "step": 19009, "train_loss": 3.5483224391937256, "perplexity": 34.7549649892652, "lr": 0.0026291804804649314, "grad_norm": 0.195374, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:49.893353+00:00", "epoch": 0, "step": 19010, "train_loss": 3.465944528579712, "perplexity": 32.00667672140342, "lr": 0.0026291804804649314, "grad_norm": 0.182144, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:50.198584+00:00", "epoch": 0, "step": 19011, "train_loss": 3.536374568939209, "perplexity": 34.342187990257145, "lr": 0.0026291804804649314, "grad_norm": 0.157398, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:50.503303+00:00", "epoch": 0, "step": 19012, "train_loss": 3.4415619373321533, "perplexity": 31.23570830496695, "lr": 0.0026291804804649314, "grad_norm": 0.200145, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:50.807942+00:00", "epoch": 0, "step": 19013, "train_loss": 3.5238916873931885, "perplexity": 33.91616306687277, "lr": 0.0026291804804649314, "grad_norm": 0.210847, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:51.163699+00:00", "epoch": 0, "step": 19014, "train_loss": 3.4854462146759033, "perplexity": 32.636986963768095, "lr": 0.0026291804804649314, "grad_norm": 0.189429, "tokens_per_sec": 92107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:51.467358+00:00", "epoch": 0, "step": 19015, "train_loss": 3.524810314178467, "perplexity": 33.94733367760683, "lr": 0.0026291804804649314, "grad_norm": 0.179131, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:51.771187+00:00", "epoch": 0, "step": 19016, "train_loss": 3.5314066410064697, "perplexity": 34.172001562244695, "lr": 0.0026291804804649314, "grad_norm": 0.153493, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:52.076414+00:00", "epoch": 0, "step": 19017, "train_loss": 3.5386526584625244, "perplexity": 34.42051174931994, "lr": 0.0026291804804649314, "grad_norm": 0.157668, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:52.381394+00:00", "epoch": 0, "step": 19018, "train_loss": 3.5104901790618896, "perplexity": 33.464667443525585, "lr": 0.0026291804804649314, "grad_norm": 0.166306, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:52.686036+00:00", "epoch": 0, "step": 19019, "train_loss": 3.4545645713806152, "perplexity": 31.644506758081704, "lr": 0.0026291804804649314, "grad_norm": 0.165473, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:52.991678+00:00", "epoch": 0, "step": 19020, "train_loss": 3.540921926498413, "perplexity": 34.49870980903305, "lr": 0.0026291804804649314, "grad_norm": 0.150691, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:53.296168+00:00", "epoch": 0, "step": 19021, "train_loss": 3.526263952255249, "perplexity": 33.99671669826963, "lr": 0.0026291804804649314, "grad_norm": 0.168175, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:53.601697+00:00", "epoch": 0, "step": 19022, "train_loss": 3.6003220081329346, "perplexity": 36.61002127044509, "lr": 0.0026291804804649314, "grad_norm": 0.170965, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:53.906006+00:00", "epoch": 0, "step": 19023, "train_loss": 3.5413966178894043, "perplexity": 34.51508993702499, "lr": 0.0026291804804649314, "grad_norm": 0.159057, "tokens_per_sec": 107680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:54.209835+00:00", "epoch": 0, "step": 19024, "train_loss": 3.466946601867676, "perplexity": 32.03876583231768, "lr": 0.0026291804804649314, "grad_norm": 0.193044, "tokens_per_sec": 107851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:54.514372+00:00", "epoch": 0, "step": 19025, "train_loss": 3.453462600708008, "perplexity": 31.609654646214608, "lr": 0.0026291804804649314, "grad_norm": 0.162594, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:54.817935+00:00", "epoch": 0, "step": 19026, "train_loss": 3.5447471141815186, "perplexity": 34.63092656461314, "lr": 0.0026291804804649314, "grad_norm": 0.170413, "tokens_per_sec": 107946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:55.122210+00:00", "epoch": 0, "step": 19027, "train_loss": 3.5535600185394287, "perplexity": 34.93747441262674, "lr": 0.0026291804804649314, "grad_norm": 0.168474, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:55.427582+00:00", "epoch": 0, "step": 19028, "train_loss": 3.524559259414673, "perplexity": 33.938812107504106, "lr": 0.0026291804804649314, "grad_norm": 0.156208, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:55.731554+00:00", "epoch": 0, "step": 19029, "train_loss": 3.5992162227630615, "perplexity": 36.56956081894395, "lr": 0.0026291804804649314, "grad_norm": 0.16449, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:56.035224+00:00", "epoch": 0, "step": 19030, "train_loss": 3.4290952682495117, "perplexity": 30.848720303699274, "lr": 0.0026291804804649314, "grad_norm": 0.180734, "tokens_per_sec": 107964} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:56.339243+00:00", "epoch": 0, "step": 19031, "train_loss": 3.421346426010132, "perplexity": 30.610602198070055, "lr": 0.0026291804804649314, "grad_norm": 0.156909, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:56.643712+00:00", "epoch": 0, "step": 19032, "train_loss": 3.5717196464538574, "perplexity": 35.57772167975906, "lr": 0.0026291804804649314, "grad_norm": 0.160307, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:56.948329+00:00", "epoch": 0, "step": 19033, "train_loss": 3.5313777923583984, "perplexity": 34.171015760417326, "lr": 0.0026291804804649314, "grad_norm": 0.155181, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:57.252293+00:00", "epoch": 0, "step": 19034, "train_loss": 3.4388036727905273, "perplexity": 31.149670670143493, "lr": 0.0026291804804649314, "grad_norm": 0.152628, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:57.555732+00:00", "epoch": 0, "step": 19035, "train_loss": 3.476043462753296, "perplexity": 32.331547706990655, "lr": 0.0026291804804649314, "grad_norm": 0.170384, "tokens_per_sec": 107989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:57.859759+00:00", "epoch": 0, "step": 19036, "train_loss": 3.4786922931671143, "perplexity": 32.417302018056546, "lr": 0.0026291804804649314, "grad_norm": 0.195738, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:58.165238+00:00", "epoch": 0, "step": 19037, "train_loss": 3.526747226715088, "perplexity": 34.01315041384594, "lr": 0.0026291804804649314, "grad_norm": 0.151739, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:58.470917+00:00", "epoch": 0, "step": 19038, "train_loss": 3.4256796836853027, "perplexity": 30.743533630054657, "lr": 0.0026291804804649314, "grad_norm": 0.195043, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:58.775399+00:00", "epoch": 0, "step": 19039, "train_loss": 3.5758461952209473, "perplexity": 35.72483821631053, "lr": 0.0026291804804649314, "grad_norm": 0.194136, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:59.079449+00:00", "epoch": 0, "step": 19040, "train_loss": 3.435772180557251, "perplexity": 31.0553836727956, "lr": 0.0026291804804649314, "grad_norm": 0.19688, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:59.383038+00:00", "epoch": 0, "step": 19041, "train_loss": 3.4988863468170166, "perplexity": 33.07859335787105, "lr": 0.0026291804804649314, "grad_norm": 0.170043, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:59.687664+00:00", "epoch": 0, "step": 19042, "train_loss": 3.5641543865203857, "perplexity": 35.309582517511046, "lr": 0.0026291804804649314, "grad_norm": 0.153082, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:37:59.992224+00:00", "epoch": 0, "step": 19043, "train_loss": 3.3877956867218018, "perplexity": 29.600631238327576, "lr": 0.0026291804804649314, "grad_norm": 0.153107, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:00.298459+00:00", "epoch": 0, "step": 19044, "train_loss": 3.4987306594848633, "perplexity": 33.07344384078747, "lr": 0.0026291804804649314, "grad_norm": 0.149876, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:00.602128+00:00", "epoch": 0, "step": 19045, "train_loss": 3.375945806503296, "perplexity": 29.251937374302017, "lr": 0.0026291804804649314, "grad_norm": 0.157068, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:00.905789+00:00", "epoch": 0, "step": 19046, "train_loss": 3.4544289112091064, "perplexity": 31.640214150042148, "lr": 0.0026291804804649314, "grad_norm": 0.171893, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:01.210010+00:00", "epoch": 0, "step": 19047, "train_loss": 3.6008691787719727, "perplexity": 36.63005868062018, "lr": 0.0026291804804649314, "grad_norm": 0.16402, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:01.514635+00:00", "epoch": 0, "step": 19048, "train_loss": 3.4902291297912598, "perplexity": 32.7934608042905, "lr": 0.0026291804804649314, "grad_norm": 0.203155, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:01.819014+00:00", "epoch": 0, "step": 19049, "train_loss": 3.497737169265747, "perplexity": 33.04060201453704, "lr": 0.0026291804804649314, "grad_norm": 0.190901, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:02.123017+00:00", "epoch": 0, "step": 19050, "train_loss": 3.4800565242767334, "perplexity": 32.461556890032384, "lr": 0.0026291804804649314, "grad_norm": 0.175888, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:02.426939+00:00", "epoch": 0, "step": 19051, "train_loss": 3.516857624053955, "perplexity": 33.67843171653074, "lr": 0.0026291804804649314, "grad_norm": 0.183148, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:02.731597+00:00", "epoch": 0, "step": 19052, "train_loss": 3.516672134399414, "perplexity": 33.672185295207335, "lr": 0.0026291804804649314, "grad_norm": 0.178765, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:03.037001+00:00", "epoch": 0, "step": 19053, "train_loss": 3.507876396179199, "perplexity": 33.37731228203287, "lr": 0.0026291804804649314, "grad_norm": 0.158465, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:03.341502+00:00", "epoch": 0, "step": 19054, "train_loss": 3.4833016395568848, "perplexity": 32.56706949199808, "lr": 0.0026291804804649314, "grad_norm": 0.168505, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:03.645918+00:00", "epoch": 0, "step": 19055, "train_loss": 3.6468091011047363, "perplexity": 38.35209294283679, "lr": 0.0026291804804649314, "grad_norm": 0.146354, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:03.950582+00:00", "epoch": 0, "step": 19056, "train_loss": 3.397193193435669, "perplexity": 29.880114535370563, "lr": 0.0026291804804649314, "grad_norm": 0.168245, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:04.254061+00:00", "epoch": 0, "step": 19057, "train_loss": 3.5368549823760986, "perplexity": 34.358690402492755, "lr": 0.0026291804804649314, "grad_norm": 0.165481, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:04.558336+00:00", "epoch": 0, "step": 19058, "train_loss": 3.516406774520874, "perplexity": 33.66325123363039, "lr": 0.0026291804804649314, "grad_norm": 0.168762, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:04.862880+00:00", "epoch": 0, "step": 19059, "train_loss": 3.5236823558807373, "perplexity": 33.90906408820734, "lr": 0.0026291804804649314, "grad_norm": 0.172627, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:05.167892+00:00", "epoch": 0, "step": 19060, "train_loss": 3.479337453842163, "perplexity": 32.43822313452492, "lr": 0.0026291804804649314, "grad_norm": 0.171995, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:05.473945+00:00", "epoch": 0, "step": 19061, "train_loss": 3.5735909938812256, "perplexity": 35.64436229211772, "lr": 0.0026291804804649314, "grad_norm": 0.170254, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:05.778714+00:00", "epoch": 0, "step": 19062, "train_loss": 3.518639326095581, "perplexity": 33.73849013429385, "lr": 0.0026291804804649314, "grad_norm": 0.155985, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:06.083290+00:00", "epoch": 0, "step": 19063, "train_loss": 3.4442055225372314, "perplexity": 31.31839180359584, "lr": 0.0026291804804649314, "grad_norm": 0.208248, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:06.387665+00:00", "epoch": 0, "step": 19064, "train_loss": 3.5467429161071777, "perplexity": 34.70011205183736, "lr": 0.0026291804804649314, "grad_norm": 0.207015, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:06.691222+00:00", "epoch": 0, "step": 19065, "train_loss": 3.5206074714660645, "perplexity": 33.804957774992694, "lr": 0.0026291804804649314, "grad_norm": 0.169409, "tokens_per_sec": 107946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:06.995108+00:00", "epoch": 0, "step": 19066, "train_loss": 3.4632139205932617, "perplexity": 31.919398250204093, "lr": 0.0026291804804649314, "grad_norm": 0.180779, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:07.299487+00:00", "epoch": 0, "step": 19067, "train_loss": 3.596167802810669, "perplexity": 36.45825118554568, "lr": 0.0026291804804649314, "grad_norm": 0.216564, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:07.605195+00:00", "epoch": 0, "step": 19068, "train_loss": 3.4429776668548584, "perplexity": 31.279960936864352, "lr": 0.0026291804804649314, "grad_norm": 0.215154, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:07.909634+00:00", "epoch": 0, "step": 19069, "train_loss": 3.4712841510772705, "perplexity": 32.17803738598042, "lr": 0.0026291804804649314, "grad_norm": 0.158932, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:08.213727+00:00", "epoch": 0, "step": 19070, "train_loss": 3.5721499919891357, "perplexity": 35.59303568836059, "lr": 0.0026291804804649314, "grad_norm": 0.188377, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:08.517406+00:00", "epoch": 0, "step": 19071, "train_loss": 3.4567971229553223, "perplexity": 31.715233672840817, "lr": 0.0026291804804649314, "grad_norm": 0.187269, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:08.821771+00:00", "epoch": 0, "step": 19072, "train_loss": 3.5182864665985107, "perplexity": 33.72658728777287, "lr": 0.0026291804804649314, "grad_norm": 0.156199, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:09.127344+00:00", "epoch": 0, "step": 19073, "train_loss": 3.501176595687866, "perplexity": 33.15443838780318, "lr": 0.0026291804804649314, "grad_norm": 0.177367, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:09.431085+00:00", "epoch": 0, "step": 19074, "train_loss": 3.5515825748443604, "perplexity": 34.86845578684128, "lr": 0.0026291804804649314, "grad_norm": 0.190644, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:09.735242+00:00", "epoch": 0, "step": 19075, "train_loss": 3.4485678672790527, "perplexity": 31.455311854501716, "lr": 0.0026291804804649314, "grad_norm": 0.170503, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:10.038988+00:00", "epoch": 0, "step": 19076, "train_loss": 3.4040844440460205, "perplexity": 30.08673701886116, "lr": 0.0026291804804649314, "grad_norm": 0.163222, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:10.343910+00:00", "epoch": 0, "step": 19077, "train_loss": 3.514633893966675, "perplexity": 33.60362318245148, "lr": 0.0026291804804649314, "grad_norm": 0.169467, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:10.649582+00:00", "epoch": 0, "step": 19078, "train_loss": 3.457561492919922, "perplexity": 31.739485112235766, "lr": 0.0026291804804649314, "grad_norm": 0.173975, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:10.953621+00:00", "epoch": 0, "step": 19079, "train_loss": 3.6044116020202637, "perplexity": 36.76004795465617, "lr": 0.0026291804804649314, "grad_norm": 0.160354, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:11.257570+00:00", "epoch": 0, "step": 19080, "train_loss": 3.453773260116577, "perplexity": 31.61947600830236, "lr": 0.0026291804804649314, "grad_norm": 0.165683, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:11.562300+00:00", "epoch": 0, "step": 19081, "train_loss": 3.5134925842285156, "perplexity": 33.565292917588906, "lr": 0.0026291804804649314, "grad_norm": 0.175335, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:11.867786+00:00", "epoch": 0, "step": 19082, "train_loss": 3.5127155780792236, "perplexity": 33.53922260829611, "lr": 0.0026291804804649314, "grad_norm": 0.185894, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:12.171456+00:00", "epoch": 0, "step": 19083, "train_loss": 3.5461440086364746, "perplexity": 34.679336117547415, "lr": 0.0026291804804649314, "grad_norm": 0.166401, "tokens_per_sec": 107909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:12.475164+00:00", "epoch": 0, "step": 19084, "train_loss": 3.4970405101776123, "perplexity": 33.01759199486646, "lr": 0.0026291804804649314, "grad_norm": 0.163134, "tokens_per_sec": 107895} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:12.779173+00:00", "epoch": 0, "step": 19085, "train_loss": 3.5416131019592285, "perplexity": 34.5225627130042, "lr": 0.0026291804804649314, "grad_norm": 0.156456, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:13.083296+00:00", "epoch": 0, "step": 19086, "train_loss": 3.4215445518493652, "perplexity": 30.616667550151632, "lr": 0.0026291804804649314, "grad_norm": 0.188547, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:13.388202+00:00", "epoch": 0, "step": 19087, "train_loss": 3.4936680793762207, "perplexity": 32.90642999906093, "lr": 0.0026291804804649314, "grad_norm": 0.154341, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:13.692251+00:00", "epoch": 0, "step": 19088, "train_loss": 3.5124154090881348, "perplexity": 33.529156684498545, "lr": 0.0026291804804649314, "grad_norm": 0.173589, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:13.996027+00:00", "epoch": 0, "step": 19089, "train_loss": 3.557630777359009, "perplexity": 35.079986313717, "lr": 0.0026291804804649314, "grad_norm": 0.164417, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:14.300697+00:00", "epoch": 0, "step": 19090, "train_loss": 3.445181369781494, "perplexity": 31.348968686689197, "lr": 0.0026291804804649314, "grad_norm": 0.160644, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:14.605058+00:00", "epoch": 0, "step": 19091, "train_loss": 3.480175495147705, "perplexity": 32.465419099469344, "lr": 0.0026291804804649314, "grad_norm": 0.159854, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:14.909096+00:00", "epoch": 0, "step": 19092, "train_loss": 3.523097276687622, "perplexity": 33.88923040305535, "lr": 0.0026291804804649314, "grad_norm": 0.162547, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:15.212879+00:00", "epoch": 0, "step": 19093, "train_loss": 3.550464153289795, "perplexity": 34.829479954083396, "lr": 0.0026291804804649314, "grad_norm": 0.190881, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:15.515803+00:00", "epoch": 0, "step": 19094, "train_loss": 3.476121187210083, "perplexity": 32.33406075663471, "lr": 0.0026291804804649314, "grad_norm": 0.188749, "tokens_per_sec": 108172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:15.819651+00:00", "epoch": 0, "step": 19095, "train_loss": 3.4632198810577393, "perplexity": 31.919588505210513, "lr": 0.0026291804804649314, "grad_norm": 0.190803, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:16.124758+00:00", "epoch": 0, "step": 19096, "train_loss": 3.485504150390625, "perplexity": 32.638877865709055, "lr": 0.0026291804804649314, "grad_norm": 0.207872, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:16.430520+00:00", "epoch": 0, "step": 19097, "train_loss": 3.540353536605835, "perplexity": 34.479106662713484, "lr": 0.0026291804804649314, "grad_norm": 0.180307, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:16.735084+00:00", "epoch": 0, "step": 19098, "train_loss": 3.56406307220459, "perplexity": 35.306358394348955, "lr": 0.0026291804804649314, "grad_norm": 0.217697, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:17.039126+00:00", "epoch": 0, "step": 19099, "train_loss": 3.4305622577667236, "perplexity": 30.894008263363386, "lr": 0.0026291804804649314, "grad_norm": 0.203444, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:17.343540+00:00", "epoch": 0, "step": 19100, "train_loss": 3.432237148284912, "perplexity": 30.945795701910992, "lr": 0.0026291804804649314, "grad_norm": 0.189213, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:17.649208+00:00", "epoch": 0, "step": 19101, "train_loss": 3.4185962677001953, "perplexity": 30.526533849682995, "lr": 0.0026291804804649314, "grad_norm": 0.164076, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:17.954144+00:00", "epoch": 0, "step": 19102, "train_loss": 3.4809160232543945, "perplexity": 32.48946955872942, "lr": 0.0026291804804649314, "grad_norm": 0.185539, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:18.257975+00:00", "epoch": 0, "step": 19103, "train_loss": 3.4464404582977295, "perplexity": 31.3884646724117, "lr": 0.0026291804804649314, "grad_norm": 0.155641, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:18.561355+00:00", "epoch": 0, "step": 19104, "train_loss": 3.5380160808563232, "perplexity": 34.39860739499676, "lr": 0.0026291804804649314, "grad_norm": 0.153624, "tokens_per_sec": 108010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:18.865219+00:00", "epoch": 0, "step": 19105, "train_loss": 3.40991473197937, "perplexity": 30.262663711948854, "lr": 0.0026291804804649314, "grad_norm": 0.161718, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:19.170057+00:00", "epoch": 0, "step": 19106, "train_loss": 3.3160550594329834, "perplexity": 27.55144706963084, "lr": 0.0026291804804649314, "grad_norm": 0.159116, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:19.475881+00:00", "epoch": 0, "step": 19107, "train_loss": 3.5378620624542236, "perplexity": 34.393309784426556, "lr": 0.0026291804804649314, "grad_norm": 0.16747, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:19.780127+00:00", "epoch": 0, "step": 19108, "train_loss": 3.512953042984009, "perplexity": 33.54718794230556, "lr": 0.0026291804804649314, "grad_norm": 0.161012, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:20.083746+00:00", "epoch": 0, "step": 19109, "train_loss": 3.532724618911743, "perplexity": 34.217069197832885, "lr": 0.0026291804804649314, "grad_norm": 0.149655, "tokens_per_sec": 107924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:20.387832+00:00", "epoch": 0, "step": 19110, "train_loss": 3.547351598739624, "perplexity": 34.72123983679282, "lr": 0.0026291804804649314, "grad_norm": 0.174191, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:20.692256+00:00", "epoch": 0, "step": 19111, "train_loss": 3.5505149364471436, "perplexity": 34.831248749956416, "lr": 0.0026291804804649314, "grad_norm": 0.156262, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:20.997704+00:00", "epoch": 0, "step": 19112, "train_loss": 3.5209128856658936, "perplexity": 33.81528386590893, "lr": 0.0026291804804649314, "grad_norm": 0.172429, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:21.302378+00:00", "epoch": 0, "step": 19113, "train_loss": 3.5655629634857178, "perplexity": 35.35935382723168, "lr": 0.0026291804804649314, "grad_norm": 0.182306, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:21.605949+00:00", "epoch": 0, "step": 19114, "train_loss": 3.4850432872772217, "perplexity": 32.623839276469575, "lr": 0.0026291804804649314, "grad_norm": 0.171187, "tokens_per_sec": 107942} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:21.910020+00:00", "epoch": 0, "step": 19115, "train_loss": 3.4959161281585693, "perplexity": 32.98048847129167, "lr": 0.0026291804804649314, "grad_norm": 0.19183, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:22.214522+00:00", "epoch": 0, "step": 19116, "train_loss": 3.5004892349243164, "perplexity": 33.13165715807655, "lr": 0.0026291804804649314, "grad_norm": 0.151567, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:22.519152+00:00", "epoch": 0, "step": 19117, "train_loss": 3.4513425827026367, "perplexity": 31.54271259347209, "lr": 0.0026291804804649314, "grad_norm": 0.177547, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:22.824757+00:00", "epoch": 0, "step": 19118, "train_loss": 3.3849427700042725, "perplexity": 29.516303449893275, "lr": 0.0026291804804649314, "grad_norm": 0.194623, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:23.128637+00:00", "epoch": 0, "step": 19119, "train_loss": 3.4941463470458984, "perplexity": 32.92217184476186, "lr": 0.0026291804804649314, "grad_norm": 0.165881, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:23.432794+00:00", "epoch": 0, "step": 19120, "train_loss": 3.469261407852173, "perplexity": 32.11301526256058, "lr": 0.0026291804804649314, "grad_norm": 0.154719, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:23.736715+00:00", "epoch": 0, "step": 19121, "train_loss": 3.480128526687622, "perplexity": 32.46389428453768, "lr": 0.0026291804804649314, "grad_norm": 0.183782, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:24.041538+00:00", "epoch": 0, "step": 19122, "train_loss": 3.5027458667755127, "perplexity": 33.20650753400966, "lr": 0.0026291804804649314, "grad_norm": 0.17385, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:24.346397+00:00", "epoch": 0, "step": 19123, "train_loss": 3.567174196243286, "perplexity": 35.416371898761625, "lr": 0.0026291804804649314, "grad_norm": 0.204464, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:24.651080+00:00", "epoch": 0, "step": 19124, "train_loss": 3.5266401767730713, "perplexity": 34.00950950294947, "lr": 0.0026291804804649314, "grad_norm": 0.166908, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:24.955136+00:00", "epoch": 0, "step": 19125, "train_loss": 3.436094045639038, "perplexity": 31.065380925198298, "lr": 0.0026291804804649314, "grad_norm": 0.20561, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:25.258800+00:00", "epoch": 0, "step": 19126, "train_loss": 3.5036847591400146, "perplexity": 33.2376995110483, "lr": 0.0026291804804649314, "grad_norm": 0.201522, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:25.562863+00:00", "epoch": 0, "step": 19127, "train_loss": 3.441305637359619, "perplexity": 31.227703619630432, "lr": 0.0026291804804649314, "grad_norm": 0.17854, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:25.868075+00:00", "epoch": 0, "step": 19128, "train_loss": 3.5097858905792236, "perplexity": 33.44110696133087, "lr": 0.0026291804804649314, "grad_norm": 0.182312, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:26.173052+00:00", "epoch": 0, "step": 19129, "train_loss": 3.510286569595337, "perplexity": 33.45785441406123, "lr": 0.0026291804804649314, "grad_norm": 0.168379, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:26.477853+00:00", "epoch": 0, "step": 19130, "train_loss": 3.5284550189971924, "perplexity": 34.07128743846376, "lr": 0.0026291804804649314, "grad_norm": 0.191397, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:26.781041+00:00", "epoch": 0, "step": 19131, "train_loss": 3.4752533435821533, "perplexity": 32.30601202076074, "lr": 0.0026291804804649314, "grad_norm": 0.155607, "tokens_per_sec": 108017} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:27.085068+00:00", "epoch": 0, "step": 19132, "train_loss": 3.5288124084472656, "perplexity": 34.083466333319315, "lr": 0.0026291804804649314, "grad_norm": 0.174894, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:27.389203+00:00", "epoch": 0, "step": 19133, "train_loss": 3.5471251010894775, "perplexity": 34.71337644811276, "lr": 0.0026291804804649314, "grad_norm": 0.165508, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:27.693688+00:00", "epoch": 0, "step": 19134, "train_loss": 3.562161445617676, "perplexity": 35.23928268123202, "lr": 0.0026291804804649314, "grad_norm": 0.175635, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:27.999820+00:00", "epoch": 0, "step": 19135, "train_loss": 3.570463180541992, "perplexity": 35.53304755688016, "lr": 0.0026291804804649314, "grad_norm": 0.195458, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:28.303789+00:00", "epoch": 0, "step": 19136, "train_loss": 3.6170408725738525, "perplexity": 37.227244511955334, "lr": 0.0026291804804649314, "grad_norm": 0.159234, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:28.607442+00:00", "epoch": 0, "step": 19137, "train_loss": 3.542802572250366, "perplexity": 34.56365070735515, "lr": 0.0026291804804649314, "grad_norm": 0.211232, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:28.911680+00:00", "epoch": 0, "step": 19138, "train_loss": 3.3826329708099365, "perplexity": 29.44820539266004, "lr": 0.0026291804804649314, "grad_norm": 0.196983, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:29.216558+00:00", "epoch": 0, "step": 19139, "train_loss": 3.4836366176605225, "perplexity": 32.577980574562396, "lr": 0.0026291804804649314, "grad_norm": 0.168295, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:29.522165+00:00", "epoch": 0, "step": 19140, "train_loss": 3.4676806926727295, "perplexity": 32.06229383050744, "lr": 0.0026291804804649314, "grad_norm": 0.197661, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:29.826513+00:00", "epoch": 0, "step": 19141, "train_loss": 3.5445058345794678, "perplexity": 34.62257183638626, "lr": 0.0026291804804649314, "grad_norm": 0.194789, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:30.130159+00:00", "epoch": 0, "step": 19142, "train_loss": 3.4109861850738525, "perplexity": 30.29510611378253, "lr": 0.0026291804804649314, "grad_norm": 0.187102, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:30.433839+00:00", "epoch": 0, "step": 19143, "train_loss": 3.4765567779541016, "perplexity": 32.34814824218325, "lr": 0.0026291804804649314, "grad_norm": 0.207765, "tokens_per_sec": 107903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:30.738757+00:00", "epoch": 0, "step": 19144, "train_loss": 3.510050058364868, "perplexity": 33.44994219144655, "lr": 0.0026291804804649314, "grad_norm": 0.196252, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:31.043092+00:00", "epoch": 0, "step": 19145, "train_loss": 3.3949363231658936, "perplexity": 29.812755032638847, "lr": 0.0026291804804649314, "grad_norm": 0.180345, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:31.347355+00:00", "epoch": 0, "step": 19146, "train_loss": 3.4400794506073, "perplexity": 31.189436089510636, "lr": 0.0026291804804649314, "grad_norm": 0.181986, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:31.651449+00:00", "epoch": 0, "step": 19147, "train_loss": 3.5933516025543213, "perplexity": 36.35572188849046, "lr": 0.0026291804804649314, "grad_norm": 0.198697, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:31.955624+00:00", "epoch": 0, "step": 19148, "train_loss": 3.489750623703003, "perplexity": 32.7777726873697, "lr": 0.0026291804804649314, "grad_norm": 0.181746, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:32.260455+00:00", "epoch": 0, "step": 19149, "train_loss": 3.5568952560424805, "perplexity": 35.0541937226628, "lr": 0.0026291804804649314, "grad_norm": 0.20579, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:32.566256+00:00", "epoch": 0, "step": 19150, "train_loss": 3.4819672107696533, "perplexity": 32.523640040152145, "lr": 0.0026291804804649314, "grad_norm": 0.177173, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:32.870670+00:00", "epoch": 0, "step": 19151, "train_loss": 3.476745843887329, "perplexity": 32.35426475321255, "lr": 0.0026291804804649314, "grad_norm": 0.207431, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:33.175002+00:00", "epoch": 0, "step": 19152, "train_loss": 3.514082908630371, "perplexity": 33.58511317866988, "lr": 0.0026291804804649314, "grad_norm": 0.20011, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:33.478585+00:00", "epoch": 0, "step": 19153, "train_loss": 3.4793009757995605, "perplexity": 32.437039873221124, "lr": 0.0026291804804649314, "grad_norm": 0.163829, "tokens_per_sec": 107931} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:33.784031+00:00", "epoch": 0, "step": 19154, "train_loss": 3.376909017562866, "perplexity": 29.280126737868727, "lr": 0.0026291804804649314, "grad_norm": 0.208189, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:34.088558+00:00", "epoch": 0, "step": 19155, "train_loss": 3.470700979232788, "perplexity": 32.15927753120743, "lr": 0.0026291804804649314, "grad_norm": 0.168421, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:34.393810+00:00", "epoch": 0, "step": 19156, "train_loss": 3.548565149307251, "perplexity": 34.76340139452233, "lr": 0.0026291804804649314, "grad_norm": 0.19275, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:34.698047+00:00", "epoch": 0, "step": 19157, "train_loss": 3.513063669204712, "perplexity": 33.55089934620835, "lr": 0.0026291804804649314, "grad_norm": 0.188466, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:35.001852+00:00", "epoch": 0, "step": 19158, "train_loss": 3.5288925170898438, "perplexity": 34.086196822908086, "lr": 0.0026291804804649314, "grad_norm": 0.161853, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:35.306574+00:00", "epoch": 0, "step": 19159, "train_loss": 3.5195517539978027, "perplexity": 33.76928812240638, "lr": 0.0026291804804649314, "grad_norm": 0.214332, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:35.611893+00:00", "epoch": 0, "step": 19160, "train_loss": 3.461176872253418, "perplexity": 31.85444307386808, "lr": 0.0026291804804649314, "grad_norm": 0.148313, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:35.917683+00:00", "epoch": 0, "step": 19161, "train_loss": 3.4919915199279785, "perplexity": 32.851306634648765, "lr": 0.0026291804804649314, "grad_norm": 0.163291, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:36.221861+00:00", "epoch": 0, "step": 19162, "train_loss": 3.562636613845825, "perplexity": 35.25603124762309, "lr": 0.0026291804804649314, "grad_norm": 0.1643, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:36.526603+00:00", "epoch": 0, "step": 19163, "train_loss": 3.473985195159912, "perplexity": 32.26506916888124, "lr": 0.0026291804804649314, "grad_norm": 0.16433, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:36.830339+00:00", "epoch": 0, "step": 19164, "train_loss": 3.461150884628296, "perplexity": 31.853615263299464, "lr": 0.0026291804804649314, "grad_norm": 0.163996, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:37.134474+00:00", "epoch": 0, "step": 19165, "train_loss": 3.4964962005615234, "perplexity": 32.99962509226607, "lr": 0.0026291804804649314, "grad_norm": 0.150002, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:37.438766+00:00", "epoch": 0, "step": 19166, "train_loss": 3.5756964683532715, "perplexity": 35.71948964860832, "lr": 0.0026291804804649314, "grad_norm": 0.167, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:37.744568+00:00", "epoch": 0, "step": 19167, "train_loss": 3.4386839866638184, "perplexity": 31.14594270980974, "lr": 0.0026291804804649314, "grad_norm": 0.165732, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:38.049503+00:00", "epoch": 0, "step": 19168, "train_loss": 3.516892194747925, "perplexity": 33.67959602341232, "lr": 0.0026291804804649314, "grad_norm": 0.182047, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:38.352963+00:00", "epoch": 0, "step": 19169, "train_loss": 3.4170093536376953, "perplexity": 30.47812928094105, "lr": 0.0026291804804649314, "grad_norm": 0.174975, "tokens_per_sec": 107982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:38.658281+00:00", "epoch": 0, "step": 19170, "train_loss": 3.55483078956604, "perplexity": 34.9819001643539, "lr": 0.0026291804804649314, "grad_norm": 0.173916, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:38.963516+00:00", "epoch": 0, "step": 19171, "train_loss": 3.533520221710205, "perplexity": 34.244303226139856, "lr": 0.0026291804804649314, "grad_norm": 0.177924, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:39.267963+00:00", "epoch": 0, "step": 19172, "train_loss": 3.553999185562134, "perplexity": 34.95282116889334, "lr": 0.0026291804804649314, "grad_norm": 0.199517, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:39.572501+00:00", "epoch": 0, "step": 19173, "train_loss": 3.5223398208618164, "perplexity": 33.86357052739248, "lr": 0.0026291804804649314, "grad_norm": 0.204853, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:39.877284+00:00", "epoch": 0, "step": 19174, "train_loss": 3.519794464111328, "perplexity": 33.777485264884696, "lr": 0.0026291804804649314, "grad_norm": 0.168576, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:40.182069+00:00", "epoch": 0, "step": 19175, "train_loss": 3.5568087100982666, "perplexity": 35.0511600556461, "lr": 0.0026291804804649314, "grad_norm": 0.199846, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:40.486793+00:00", "epoch": 0, "step": 19176, "train_loss": 3.4587197303771973, "perplexity": 31.776268270471128, "lr": 0.0026291804804649314, "grad_norm": 0.195209, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:40.791217+00:00", "epoch": 0, "step": 19177, "train_loss": 3.594785690307617, "perplexity": 36.407896586618875, "lr": 0.0026291804804649314, "grad_norm": 0.170153, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:41.095358+00:00", "epoch": 0, "step": 19178, "train_loss": 3.4912688732147217, "perplexity": 32.82757532159348, "lr": 0.0026291804804649314, "grad_norm": 0.162242, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:41.399349+00:00", "epoch": 0, "step": 19179, "train_loss": 3.549856424331665, "perplexity": 34.80831950107999, "lr": 0.0026291804804649314, "grad_norm": 0.184024, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:41.703934+00:00", "epoch": 0, "step": 19180, "train_loss": 3.5120561122894287, "perplexity": 33.51711192978897, "lr": 0.0026291804804649314, "grad_norm": 0.190542, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:42.010548+00:00", "epoch": 0, "step": 19181, "train_loss": 3.5026440620422363, "perplexity": 33.20312712644078, "lr": 0.0026291804804649314, "grad_norm": 0.180857, "tokens_per_sec": 106811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:42.316908+00:00", "epoch": 0, "step": 19182, "train_loss": 3.5141658782958984, "perplexity": 33.58789983987959, "lr": 0.0026291804804649314, "grad_norm": 0.171983, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:42.622290+00:00", "epoch": 0, "step": 19183, "train_loss": 3.5684683322906494, "perplexity": 35.46223517255964, "lr": 0.0026291804804649314, "grad_norm": 0.183902, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:42.927773+00:00", "epoch": 0, "step": 19184, "train_loss": 3.488530158996582, "perplexity": 32.73779297452652, "lr": 0.0026291804804649314, "grad_norm": 0.17152, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:43.232867+00:00", "epoch": 0, "step": 19185, "train_loss": 3.5179226398468018, "perplexity": 33.71431888499563, "lr": 0.0026291804804649314, "grad_norm": 0.160283, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:43.537704+00:00", "epoch": 0, "step": 19186, "train_loss": 3.535679340362549, "perplexity": 34.31832061738247, "lr": 0.0026291804804649314, "grad_norm": 0.150171, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:43.842341+00:00", "epoch": 0, "step": 19187, "train_loss": 3.4904565811157227, "perplexity": 32.80092056871864, "lr": 0.0026291804804649314, "grad_norm": 0.165496, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:44.147400+00:00", "epoch": 0, "step": 19188, "train_loss": 3.5020790100097656, "perplexity": 33.18437093158566, "lr": 0.0026291804804649314, "grad_norm": 0.17184, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:44.452302+00:00", "epoch": 0, "step": 19189, "train_loss": 3.3917365074157715, "perplexity": 29.717512170600433, "lr": 0.0026291804804649314, "grad_norm": 0.160651, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:44.756736+00:00", "epoch": 0, "step": 19190, "train_loss": 3.5143892765045166, "perplexity": 33.5954041547291, "lr": 0.0026291804804649314, "grad_norm": 0.155405, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:45.061267+00:00", "epoch": 0, "step": 19191, "train_loss": 3.444085121154785, "perplexity": 31.314621252921008, "lr": 0.0026291804804649314, "grad_norm": 0.170173, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:45.366281+00:00", "epoch": 0, "step": 19192, "train_loss": 3.5586416721343994, "perplexity": 35.115466418907346, "lr": 0.0026291804804649314, "grad_norm": 0.149646, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:45.672107+00:00", "epoch": 0, "step": 19193, "train_loss": 3.5429916381835938, "perplexity": 34.57018613402569, "lr": 0.0026291804804649314, "grad_norm": 0.176437, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:45.976510+00:00", "epoch": 0, "step": 19194, "train_loss": 3.5599186420440674, "perplexity": 35.16033645563433, "lr": 0.0026291804804649314, "grad_norm": 0.184201, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:46.280719+00:00", "epoch": 0, "step": 19195, "train_loss": 3.5380096435546875, "perplexity": 34.39838596149782, "lr": 0.0026291804804649314, "grad_norm": 0.159832, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:46.585420+00:00", "epoch": 0, "step": 19196, "train_loss": 3.4484381675720215, "perplexity": 31.45123237432902, "lr": 0.0026291804804649314, "grad_norm": 0.165818, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:46.889947+00:00", "epoch": 0, "step": 19197, "train_loss": 3.5733606815338135, "perplexity": 35.636153900649454, "lr": 0.0026291804804649314, "grad_norm": 0.156496, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:47.195268+00:00", "epoch": 0, "step": 19198, "train_loss": 3.539477586746216, "perplexity": 34.44891791791616, "lr": 0.0026291804804649314, "grad_norm": 0.160773, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:47.500631+00:00", "epoch": 0, "step": 19199, "train_loss": 3.4330272674560547, "perplexity": 30.970256230455067, "lr": 0.0026291804804649314, "grad_norm": 0.162605, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:47.805815+00:00", "epoch": 0, "step": 19200, "train_loss": 3.507848024368286, "perplexity": 33.376365320673585, "lr": 0.0026291804804649314, "grad_norm": 0.158208, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:48.110450+00:00", "epoch": 0, "step": 19201, "train_loss": 3.6107916831970215, "perplexity": 36.99532980521774, "lr": 0.0026291804804649314, "grad_norm": 0.159815, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:48.415047+00:00", "epoch": 0, "step": 19202, "train_loss": 3.390887498855591, "perplexity": 29.692292455772105, "lr": 0.0026291804804649314, "grad_norm": 0.157233, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:48.720310+00:00", "epoch": 0, "step": 19203, "train_loss": 3.6621179580688477, "perplexity": 38.94373678866423, "lr": 0.0026291804804649314, "grad_norm": 0.154955, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:49.025327+00:00", "epoch": 0, "step": 19204, "train_loss": 3.426985025405884, "perplexity": 30.783690650755283, "lr": 0.0026291804804649314, "grad_norm": 0.189621, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:49.329305+00:00", "epoch": 0, "step": 19205, "train_loss": 3.5057904720306396, "perplexity": 33.30776230353772, "lr": 0.0026291804804649314, "grad_norm": 0.171949, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:49.632930+00:00", "epoch": 0, "step": 19206, "train_loss": 3.5371205806732178, "perplexity": 34.36781722413539, "lr": 0.0026291804804649314, "grad_norm": 0.17922, "tokens_per_sec": 107922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:49.936478+00:00", "epoch": 0, "step": 19207, "train_loss": 3.547562837600708, "perplexity": 34.728575086669004, "lr": 0.0026291804804649314, "grad_norm": 0.195989, "tokens_per_sec": 107950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:50.241632+00:00", "epoch": 0, "step": 19208, "train_loss": 3.522632598876953, "perplexity": 33.87348648787372, "lr": 0.0026291804804649314, "grad_norm": 0.1763, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:50.548374+00:00", "epoch": 0, "step": 19209, "train_loss": 3.50813364982605, "perplexity": 33.3858998218801, "lr": 0.0026291804804649314, "grad_norm": 0.18434, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:50.853211+00:00", "epoch": 0, "step": 19210, "train_loss": 3.491309642791748, "perplexity": 32.82891371523683, "lr": 0.0026291804804649314, "grad_norm": 0.192561, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:51.158369+00:00", "epoch": 0, "step": 19211, "train_loss": 3.5121192932128906, "perplexity": 33.51922963877117, "lr": 0.0026291804804649314, "grad_norm": 0.181866, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:51.462941+00:00", "epoch": 0, "step": 19212, "train_loss": 3.4564900398254395, "perplexity": 31.70549595484058, "lr": 0.0026291804804649314, "grad_norm": 0.170049, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:51.768587+00:00", "epoch": 0, "step": 19213, "train_loss": 3.470874786376953, "perplexity": 32.16486752917026, "lr": 0.0026291804804649314, "grad_norm": 0.199748, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:52.075025+00:00", "epoch": 0, "step": 19214, "train_loss": 3.504849910736084, "perplexity": 33.27644903983713, "lr": 0.0026291804804649314, "grad_norm": 0.185555, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:52.379644+00:00", "epoch": 0, "step": 19215, "train_loss": 3.493572235107422, "perplexity": 32.90327625747533, "lr": 0.0026291804804649314, "grad_norm": 0.201473, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:52.683963+00:00", "epoch": 0, "step": 19216, "train_loss": 3.4422121047973633, "perplexity": 31.256023349634546, "lr": 0.0026291804804649314, "grad_norm": 0.169877, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:52.989280+00:00", "epoch": 0, "step": 19217, "train_loss": 3.504901647567749, "perplexity": 33.278170702415814, "lr": 0.0026291804804649314, "grad_norm": 0.210204, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:53.294645+00:00", "epoch": 0, "step": 19218, "train_loss": 3.4676389694213867, "perplexity": 32.06095611527043, "lr": 0.0026291804804649314, "grad_norm": 0.208295, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:53.600767+00:00", "epoch": 0, "step": 19219, "train_loss": 3.54068922996521, "perplexity": 34.49068301280055, "lr": 0.0026291804804649314, "grad_norm": 0.186046, "tokens_per_sec": 107040} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:53.905509+00:00", "epoch": 0, "step": 19220, "train_loss": 3.5247862339019775, "perplexity": 33.946516226268066, "lr": 0.0026291804804649314, "grad_norm": 0.207673, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:54.209811+00:00", "epoch": 0, "step": 19221, "train_loss": 3.597346544265747, "perplexity": 36.501251375672254, "lr": 0.0026291804804649314, "grad_norm": 0.165061, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:54.514446+00:00", "epoch": 0, "step": 19222, "train_loss": 3.3752615451812744, "perplexity": 29.23192825147772, "lr": 0.0026291804804649314, "grad_norm": 0.180601, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:54.817663+00:00", "epoch": 0, "step": 19223, "train_loss": 3.5633301734924316, "perplexity": 35.280491889672554, "lr": 0.0026291804804649314, "grad_norm": 0.176077, "tokens_per_sec": 108068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:55.122978+00:00", "epoch": 0, "step": 19224, "train_loss": 3.4940335750579834, "perplexity": 32.918459355332786, "lr": 0.0026291804804649314, "grad_norm": 0.166072, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:55.429723+00:00", "epoch": 0, "step": 19225, "train_loss": 3.4230589866638184, "perplexity": 30.663069624982263, "lr": 0.0026291804804649314, "grad_norm": 0.193333, "tokens_per_sec": 106826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:55.735493+00:00", "epoch": 0, "step": 19226, "train_loss": 3.487154722213745, "perplexity": 32.69279516279716, "lr": 0.0026291804804649314, "grad_norm": 0.170558, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:56.041987+00:00", "epoch": 0, "step": 19227, "train_loss": 3.5229854583740234, "perplexity": 33.885441178318835, "lr": 0.0026291804804649314, "grad_norm": 0.181335, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:56.346814+00:00", "epoch": 0, "step": 19228, "train_loss": 3.508826732635498, "perplexity": 33.409047035666, "lr": 0.0026291804804649314, "grad_norm": 0.173384, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:56.651691+00:00", "epoch": 0, "step": 19229, "train_loss": 3.5332987308502197, "perplexity": 34.236719265888645, "lr": 0.0026291804804649314, "grad_norm": 0.175297, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:56.957237+00:00", "epoch": 0, "step": 19230, "train_loss": 3.5302295684814453, "perplexity": 34.13180230144473, "lr": 0.0026291804804649314, "grad_norm": 0.157744, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:57.262806+00:00", "epoch": 0, "step": 19231, "train_loss": 3.612406015396118, "perplexity": 37.05510078947013, "lr": 0.0026291804804649314, "grad_norm": 0.173897, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:57.568433+00:00", "epoch": 0, "step": 19232, "train_loss": 3.497736930847168, "perplexity": 33.04059413704459, "lr": 0.0026291804804649314, "grad_norm": 0.186958, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:57.873633+00:00", "epoch": 0, "step": 19233, "train_loss": 3.5586655139923096, "perplexity": 35.11630364684865, "lr": 0.0026291804804649314, "grad_norm": 0.193842, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:58.178093+00:00", "epoch": 0, "step": 19234, "train_loss": 3.5267300605773926, "perplexity": 34.012566544433895, "lr": 0.0026291804804649314, "grad_norm": 0.187363, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:58.482837+00:00", "epoch": 0, "step": 19235, "train_loss": 3.412992238998413, "perplexity": 30.35594072865487, "lr": 0.0026291804804649314, "grad_norm": 0.220197, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:58.787741+00:00", "epoch": 0, "step": 19236, "train_loss": 3.400191307067871, "perplexity": 29.96983293987242, "lr": 0.0026291804804649314, "grad_norm": 0.223672, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:59.094011+00:00", "epoch": 0, "step": 19237, "train_loss": 3.542144536972046, "perplexity": 34.54091408741162, "lr": 0.0026291804804649314, "grad_norm": 0.183419, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:59.399474+00:00", "epoch": 0, "step": 19238, "train_loss": 3.4840052127838135, "perplexity": 32.589990872661524, "lr": 0.0026291804804649314, "grad_norm": 0.197156, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:38:59.704023+00:00", "epoch": 0, "step": 19239, "train_loss": 3.467651605606079, "perplexity": 32.06136124599297, "lr": 0.0026291804804649314, "grad_norm": 0.192907, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:00.008963+00:00", "epoch": 0, "step": 19240, "train_loss": 3.5183284282684326, "perplexity": 33.72800254138923, "lr": 0.0026291804804649314, "grad_norm": 0.200647, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:00.314680+00:00", "epoch": 0, "step": 19241, "train_loss": 3.4306039810180664, "perplexity": 30.89529728872612, "lr": 0.0026291804804649314, "grad_norm": 0.187855, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:00.620870+00:00", "epoch": 0, "step": 19242, "train_loss": 3.4810924530029297, "perplexity": 32.49520217336071, "lr": 0.0026291804804649314, "grad_norm": 0.166086, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:00.925963+00:00", "epoch": 0, "step": 19243, "train_loss": 3.5479798316955566, "perplexity": 34.743059717195464, "lr": 0.0026291804804649314, "grad_norm": 0.212637, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:01.230405+00:00", "epoch": 0, "step": 19244, "train_loss": 3.536264657974243, "perplexity": 34.338413614662535, "lr": 0.0026291804804649314, "grad_norm": 0.207048, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:01.535235+00:00", "epoch": 0, "step": 19245, "train_loss": 3.4309239387512207, "perplexity": 30.905184059601805, "lr": 0.0026291804804649314, "grad_norm": 0.172214, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:01.840473+00:00", "epoch": 0, "step": 19246, "train_loss": 3.470557451248169, "perplexity": 32.15466210614582, "lr": 0.0026291804804649314, "grad_norm": 0.171059, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:02.145691+00:00", "epoch": 0, "step": 19247, "train_loss": 3.4131019115448, "perplexity": 30.35927012454087, "lr": 0.0026291804804649314, "grad_norm": 0.198264, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:02.451401+00:00", "epoch": 0, "step": 19248, "train_loss": 3.468963861465454, "perplexity": 32.10346157230605, "lr": 0.0026291804804649314, "grad_norm": 0.169056, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:02.755803+00:00", "epoch": 0, "step": 19249, "train_loss": 3.4915215969085693, "perplexity": 32.835872676113695, "lr": 0.0026291804804649314, "grad_norm": 0.16612, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:03.060490+00:00", "epoch": 0, "step": 19250, "train_loss": 3.4982638359069824, "perplexity": 33.05800798059293, "lr": 0.0026291804804649314, "grad_norm": 0.201166, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:03.366058+00:00", "epoch": 0, "step": 19251, "train_loss": 3.5203497409820557, "perplexity": 33.7962463295142, "lr": 0.0026291804804649314, "grad_norm": 0.194084, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:03.671423+00:00", "epoch": 0, "step": 19252, "train_loss": 3.5110230445861816, "perplexity": 33.4825043630058, "lr": 0.0026291804804649314, "grad_norm": 0.168649, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:03.978218+00:00", "epoch": 0, "step": 19253, "train_loss": 3.5181853771209717, "perplexity": 33.72317805700636, "lr": 0.0026291804804649314, "grad_norm": 0.19631, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:04.283378+00:00", "epoch": 0, "step": 19254, "train_loss": 3.458869457244873, "perplexity": 31.781026387785804, "lr": 0.0026291804804649314, "grad_norm": 0.176051, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:04.587713+00:00", "epoch": 0, "step": 19255, "train_loss": 3.5756044387817383, "perplexity": 35.716202550538, "lr": 0.0026291804804649314, "grad_norm": 0.157956, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:04.893683+00:00", "epoch": 0, "step": 19256, "train_loss": 3.5388495922088623, "perplexity": 34.427290977157064, "lr": 0.0026291804804649314, "grad_norm": 0.172602, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:05.198902+00:00", "epoch": 0, "step": 19257, "train_loss": 3.526596784591675, "perplexity": 34.008033788161384, "lr": 0.0026291804804649314, "grad_norm": 0.181235, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:05.504912+00:00", "epoch": 0, "step": 19258, "train_loss": 3.522860050201416, "perplexity": 33.88119193351325, "lr": 0.0026291804804649314, "grad_norm": 0.153711, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:05.809924+00:00", "epoch": 0, "step": 19259, "train_loss": 3.470842123031616, "perplexity": 32.16381693415251, "lr": 0.0026291804804649314, "grad_norm": 0.161809, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:06.115012+00:00", "epoch": 0, "step": 19260, "train_loss": 3.547055244445801, "perplexity": 34.71095157284122, "lr": 0.0026291804804649314, "grad_norm": 0.161474, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:06.419793+00:00", "epoch": 0, "step": 19261, "train_loss": 3.4671645164489746, "perplexity": 32.04574830732329, "lr": 0.0026291804804649314, "grad_norm": 0.146876, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:06.724986+00:00", "epoch": 0, "step": 19262, "train_loss": 3.449732780456543, "perplexity": 31.491975912859285, "lr": 0.0026291804804649314, "grad_norm": 0.145343, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:07.030423+00:00", "epoch": 0, "step": 19263, "train_loss": 3.4249918460845947, "perplexity": 30.72239434267708, "lr": 0.0026291804804649314, "grad_norm": 0.171765, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:07.336796+00:00", "epoch": 0, "step": 19264, "train_loss": 3.5277717113494873, "perplexity": 34.0480142194797, "lr": 0.0026291804804649314, "grad_norm": 0.155092, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:07.641963+00:00", "epoch": 0, "step": 19265, "train_loss": 3.4771788120269775, "perplexity": 32.36827615205821, "lr": 0.0026291804804649314, "grad_norm": 0.159514, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:07.946993+00:00", "epoch": 0, "step": 19266, "train_loss": 3.4797847270965576, "perplexity": 32.45273512932471, "lr": 0.0026291804804649314, "grad_norm": 0.167515, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:08.252377+00:00", "epoch": 0, "step": 19267, "train_loss": 3.4587693214416504, "perplexity": 31.777844128512925, "lr": 0.0026291804804649314, "grad_norm": 0.169994, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:08.557539+00:00", "epoch": 0, "step": 19268, "train_loss": 3.6087217330932617, "perplexity": 36.91883052061452, "lr": 0.0026291804804649314, "grad_norm": 0.181802, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:08.862998+00:00", "epoch": 0, "step": 19269, "train_loss": 3.4703187942504883, "perplexity": 32.14698908667245, "lr": 0.0026291804804649314, "grad_norm": 0.162542, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:09.168524+00:00", "epoch": 0, "step": 19270, "train_loss": 3.546928644180298, "perplexity": 34.706557435311595, "lr": 0.0026291804804649314, "grad_norm": 0.146347, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:09.473044+00:00", "epoch": 0, "step": 19271, "train_loss": 3.4220733642578125, "perplexity": 30.632862305473843, "lr": 0.0026291804804649314, "grad_norm": 0.166719, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:09.778496+00:00", "epoch": 0, "step": 19272, "train_loss": 3.442612648010254, "perplexity": 31.268545245261933, "lr": 0.0026291804804649314, "grad_norm": 0.180461, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:10.084702+00:00", "epoch": 0, "step": 19273, "train_loss": 3.4734909534454346, "perplexity": 32.24912636590115, "lr": 0.0026291804804649314, "grad_norm": 0.168405, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:10.390837+00:00", "epoch": 0, "step": 19274, "train_loss": 3.5423905849456787, "perplexity": 34.54941385496266, "lr": 0.0026291804804649314, "grad_norm": 0.182568, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:10.696691+00:00", "epoch": 0, "step": 19275, "train_loss": 3.4109444618225098, "perplexity": 30.293842129824633, "lr": 0.0026291804804649314, "grad_norm": 0.200213, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:11.002409+00:00", "epoch": 0, "step": 19276, "train_loss": 3.3935413360595703, "perplexity": 29.771195617927688, "lr": 0.0026291804804649314, "grad_norm": 0.176772, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:11.307823+00:00", "epoch": 0, "step": 19277, "train_loss": 3.533113718032837, "perplexity": 34.23038561992023, "lr": 0.0026291804804649314, "grad_norm": 0.175951, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:11.614284+00:00", "epoch": 0, "step": 19278, "train_loss": 3.5849955081939697, "perplexity": 36.053195774533336, "lr": 0.0026291804804649314, "grad_norm": 0.197505, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:11.921045+00:00", "epoch": 0, "step": 19279, "train_loss": 3.609348773956299, "perplexity": 36.94198739536112, "lr": 0.0026291804804649314, "grad_norm": 0.206768, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:12.227184+00:00", "epoch": 0, "step": 19280, "train_loss": 3.462256908416748, "perplexity": 31.88886560979632, "lr": 0.0026291804804649314, "grad_norm": 0.188046, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:12.533371+00:00", "epoch": 0, "step": 19281, "train_loss": 3.443483591079712, "perplexity": 31.29579023072876, "lr": 0.0026291804804649314, "grad_norm": 0.177643, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:12.838428+00:00", "epoch": 0, "step": 19282, "train_loss": 3.4731740951538086, "perplexity": 32.23890958153632, "lr": 0.0026291804804649314, "grad_norm": 0.178162, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:13.145072+00:00", "epoch": 0, "step": 19283, "train_loss": 3.5674641132354736, "perplexity": 35.42664119532657, "lr": 0.0026291804804649314, "grad_norm": 0.183815, "tokens_per_sec": 106856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:13.451151+00:00", "epoch": 0, "step": 19284, "train_loss": 3.5026934146881104, "perplexity": 33.20476582905259, "lr": 0.0026291804804649314, "grad_norm": 0.20273, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:13.756617+00:00", "epoch": 0, "step": 19285, "train_loss": 3.462036371231079, "perplexity": 31.881833704547358, "lr": 0.0026291804804649314, "grad_norm": 0.177487, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:14.061393+00:00", "epoch": 0, "step": 19286, "train_loss": 3.5541162490844727, "perplexity": 34.956913108758826, "lr": 0.0026291804804649314, "grad_norm": 0.185798, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:14.367760+00:00", "epoch": 0, "step": 19287, "train_loss": 3.4398598670959473, "perplexity": 31.18258815549121, "lr": 0.0026291804804649314, "grad_norm": 0.189941, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:14.674285+00:00", "epoch": 0, "step": 19288, "train_loss": 3.5166985988616943, "perplexity": 33.673076423276534, "lr": 0.0026291804804649314, "grad_norm": 0.212784, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:14.980929+00:00", "epoch": 0, "step": 19289, "train_loss": 3.526407241821289, "perplexity": 34.00158842207737, "lr": 0.0026291804804649314, "grad_norm": 0.184236, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:15.287185+00:00", "epoch": 0, "step": 19290, "train_loss": 3.428645610809326, "perplexity": 30.83485206530647, "lr": 0.0026291804804649314, "grad_norm": 0.198602, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:15.592799+00:00", "epoch": 0, "step": 19291, "train_loss": 3.580441474914551, "perplexity": 35.88938161183571, "lr": 0.0026291804804649314, "grad_norm": 0.172199, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:15.897830+00:00", "epoch": 0, "step": 19292, "train_loss": 3.419999122619629, "perplexity": 30.569388200057283, "lr": 0.0026291804804649314, "grad_norm": 0.184508, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:16.204734+00:00", "epoch": 0, "step": 19293, "train_loss": 3.5265402793884277, "perplexity": 34.0061122115902, "lr": 0.0026291804804649314, "grad_norm": 0.184225, "tokens_per_sec": 106769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:16.511007+00:00", "epoch": 0, "step": 19294, "train_loss": 3.491469144821167, "perplexity": 32.834150411218744, "lr": 0.0026291804804649314, "grad_norm": 0.181086, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:16.815583+00:00", "epoch": 0, "step": 19295, "train_loss": 3.6081643104553223, "perplexity": 36.898256863359954, "lr": 0.0026291804804649314, "grad_norm": 0.182481, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:17.121170+00:00", "epoch": 0, "step": 19296, "train_loss": 3.434652090072632, "perplexity": 31.020618306860023, "lr": 0.0026291804804649314, "grad_norm": 0.184009, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:17.427259+00:00", "epoch": 0, "step": 19297, "train_loss": 3.5610711574554443, "perplexity": 35.200882645835094, "lr": 0.0026291804804649314, "grad_norm": 0.172866, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:17.734426+00:00", "epoch": 0, "step": 19298, "train_loss": 3.4438154697418213, "perplexity": 31.306178359423967, "lr": 0.0026291804804649314, "grad_norm": 0.169046, "tokens_per_sec": 106680} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:18.040867+00:00", "epoch": 0, "step": 19299, "train_loss": 3.4424490928649902, "perplexity": 31.263431532000617, "lr": 0.0026291804804649314, "grad_norm": 0.177107, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:18.346559+00:00", "epoch": 0, "step": 19300, "train_loss": 3.6153199672698975, "perplexity": 37.163235042338044, "lr": 0.0026291804804649314, "grad_norm": 0.159667, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:18.651459+00:00", "epoch": 0, "step": 19301, "train_loss": 3.557964563369751, "perplexity": 35.091697476808015, "lr": 0.0026291804804649314, "grad_norm": 0.162902, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:18.957007+00:00", "epoch": 0, "step": 19302, "train_loss": 3.5139524936676025, "perplexity": 33.58073346298154, "lr": 0.0026291804804649314, "grad_norm": 0.158099, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:19.263378+00:00", "epoch": 0, "step": 19303, "train_loss": 3.5015461444854736, "perplexity": 33.16669283481336, "lr": 0.0026291804804649314, "grad_norm": 0.139676, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:19.569422+00:00", "epoch": 0, "step": 19304, "train_loss": 3.463808059692383, "perplexity": 31.93836844763492, "lr": 0.0026291804804649314, "grad_norm": 0.160282, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:19.874385+00:00", "epoch": 0, "step": 19305, "train_loss": 3.4655709266662598, "perplexity": 31.99472119917911, "lr": 0.0026291804804649314, "grad_norm": 0.145547, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:20.178717+00:00", "epoch": 0, "step": 19306, "train_loss": 3.5615363121032715, "perplexity": 35.21726030878314, "lr": 0.0026291804804649314, "grad_norm": 0.161238, "tokens_per_sec": 107672} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:20.484138+00:00", "epoch": 0, "step": 19307, "train_loss": 3.466771125793457, "perplexity": 32.03314428870424, "lr": 0.0026291804804649314, "grad_norm": 0.157172, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:20.789351+00:00", "epoch": 0, "step": 19308, "train_loss": 3.517788887023926, "perplexity": 33.709809801231486, "lr": 0.0026291804804649314, "grad_norm": 0.16, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:21.094869+00:00", "epoch": 0, "step": 19309, "train_loss": 3.542750835418701, "perplexity": 34.56186253983424, "lr": 0.0026291804804649314, "grad_norm": 0.154051, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:21.400381+00:00", "epoch": 0, "step": 19310, "train_loss": 3.409362554550171, "perplexity": 30.245957964792392, "lr": 0.0026291804804649314, "grad_norm": 0.165915, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:21.705810+00:00", "epoch": 0, "step": 19311, "train_loss": 3.6074934005737305, "perplexity": 36.873509760682936, "lr": 0.0026291804804649314, "grad_norm": 0.158725, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:22.011294+00:00", "epoch": 0, "step": 19312, "train_loss": 3.3615365028381348, "perplexity": 28.83345955349878, "lr": 0.0026291804804649314, "grad_norm": 0.186857, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:22.316964+00:00", "epoch": 0, "step": 19313, "train_loss": 3.5003957748413086, "perplexity": 33.128560815342844, "lr": 0.0026291804804649314, "grad_norm": 0.187951, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:22.622349+00:00", "epoch": 0, "step": 19314, "train_loss": 3.6990163326263428, "perplexity": 40.40753722843753, "lr": 0.0026291804804649314, "grad_norm": 0.191587, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:22.926907+00:00", "epoch": 0, "step": 19315, "train_loss": 3.574173927307129, "perplexity": 35.665146639699714, "lr": 0.0026291804804649314, "grad_norm": 0.185737, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:23.231828+00:00", "epoch": 0, "step": 19316, "train_loss": 3.4730937480926514, "perplexity": 32.236319383955305, "lr": 0.0026291804804649314, "grad_norm": 0.174056, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:23.538318+00:00", "epoch": 0, "step": 19317, "train_loss": 3.573934555053711, "perplexity": 35.656610414888526, "lr": 0.0026291804804649314, "grad_norm": 0.188279, "tokens_per_sec": 106911} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:23.843720+00:00", "epoch": 0, "step": 19318, "train_loss": 3.5060691833496094, "perplexity": 33.31704684769495, "lr": 0.0026291804804649314, "grad_norm": 0.162152, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:24.149793+00:00", "epoch": 0, "step": 19319, "train_loss": 3.6665279865264893, "perplexity": 39.115859029180854, "lr": 0.0026291804804649314, "grad_norm": 0.198864, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:24.454802+00:00", "epoch": 0, "step": 19320, "train_loss": 3.5220491886138916, "perplexity": 33.853730111807664, "lr": 0.0026291804804649314, "grad_norm": 0.16366, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:24.759382+00:00", "epoch": 0, "step": 19321, "train_loss": 3.5621697902679443, "perplexity": 35.239576741948625, "lr": 0.0026291804804649314, "grad_norm": 0.196864, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:25.064595+00:00", "epoch": 0, "step": 19322, "train_loss": 3.4678525924682617, "perplexity": 32.06780580600135, "lr": 0.0026291804804649314, "grad_norm": 0.180184, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:25.370796+00:00", "epoch": 0, "step": 19323, "train_loss": 3.488527536392212, "perplexity": 32.737707116360184, "lr": 0.0026291804804649314, "grad_norm": 0.175487, "tokens_per_sec": 107014} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:25.676530+00:00", "epoch": 0, "step": 19324, "train_loss": 3.460773468017578, "perplexity": 31.841595448169304, "lr": 0.0026291804804649314, "grad_norm": 0.173255, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:25.981642+00:00", "epoch": 0, "step": 19325, "train_loss": 3.463491439819336, "perplexity": 31.928257726183844, "lr": 0.0026291804804649314, "grad_norm": 0.168338, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:26.286635+00:00", "epoch": 0, "step": 19326, "train_loss": 3.4769093990325928, "perplexity": 32.35955689245066, "lr": 0.0026291804804649314, "grad_norm": 0.198525, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:26.591840+00:00", "epoch": 0, "step": 19327, "train_loss": 3.5227251052856445, "perplexity": 33.87662014739813, "lr": 0.0026291804804649314, "grad_norm": 0.174567, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:26.897554+00:00", "epoch": 0, "step": 19328, "train_loss": 3.538456916809082, "perplexity": 34.4137748808022, "lr": 0.0026291804804649314, "grad_norm": 0.164757, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:27.204512+00:00", "epoch": 0, "step": 19329, "train_loss": 3.5899789333343506, "perplexity": 36.23331260335417, "lr": 0.0026291804804649314, "grad_norm": 0.205596, "tokens_per_sec": 106751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:27.511139+00:00", "epoch": 0, "step": 19330, "train_loss": 3.5062472820281982, "perplexity": 33.3229810981384, "lr": 0.0026291804804649314, "grad_norm": 0.174412, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:27.815750+00:00", "epoch": 0, "step": 19331, "train_loss": 3.518305778503418, "perplexity": 33.72723861870862, "lr": 0.0026291804804649314, "grad_norm": 0.197653, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:28.120264+00:00", "epoch": 0, "step": 19332, "train_loss": 3.5502798557281494, "perplexity": 34.82306155732008, "lr": 0.0026291804804649314, "grad_norm": 0.187518, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:28.425766+00:00", "epoch": 0, "step": 19333, "train_loss": 3.534278154373169, "perplexity": 34.270267940584986, "lr": 0.0026291804804649314, "grad_norm": 0.149045, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:28.731526+00:00", "epoch": 0, "step": 19334, "train_loss": 3.4611294269561768, "perplexity": 31.852931766200488, "lr": 0.0026291804804649314, "grad_norm": 0.174267, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:29.037280+00:00", "epoch": 0, "step": 19335, "train_loss": 3.457899808883667, "perplexity": 31.75022490334924, "lr": 0.0026291804804649314, "grad_norm": 0.165666, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:29.343187+00:00", "epoch": 0, "step": 19336, "train_loss": 3.4628987312316895, "perplexity": 31.90933918078695, "lr": 0.0026291804804649314, "grad_norm": 0.151125, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:29.647737+00:00", "epoch": 0, "step": 19337, "train_loss": 3.5412306785583496, "perplexity": 34.50936300126464, "lr": 0.0026291804804649314, "grad_norm": 0.189168, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:29.952207+00:00", "epoch": 0, "step": 19338, "train_loss": 3.6478946208953857, "perplexity": 38.3937475030735, "lr": 0.0026291804804649314, "grad_norm": 0.172578, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:30.259123+00:00", "epoch": 0, "step": 19339, "train_loss": 3.512587070465088, "perplexity": 33.53491283974371, "lr": 0.0026291804804649314, "grad_norm": 0.168652, "tokens_per_sec": 106762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:30.564738+00:00", "epoch": 0, "step": 19340, "train_loss": 3.463961362838745, "perplexity": 31.94326507533225, "lr": 0.0026291804804649314, "grad_norm": 0.16214, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:30.870560+00:00", "epoch": 0, "step": 19341, "train_loss": 3.474931001663208, "perplexity": 32.295600117039164, "lr": 0.0026291804804649314, "grad_norm": 0.155489, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:31.176241+00:00", "epoch": 0, "step": 19342, "train_loss": 3.500208854675293, "perplexity": 33.122368997961175, "lr": 0.0026291804804649314, "grad_norm": 0.16262, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:31.480842+00:00", "epoch": 0, "step": 19343, "train_loss": 3.4701409339904785, "perplexity": 32.14127192327812, "lr": 0.0026291804804649314, "grad_norm": 0.174805, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:31.786545+00:00", "epoch": 0, "step": 19344, "train_loss": 3.4490549564361572, "perplexity": 31.470637127923638, "lr": 0.0026291804804649314, "grad_norm": 0.182588, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:32.093519+00:00", "epoch": 0, "step": 19345, "train_loss": 3.5447990894317627, "perplexity": 34.63272656246478, "lr": 0.0026291804804649314, "grad_norm": 0.154904, "tokens_per_sec": 106745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:32.399266+00:00", "epoch": 0, "step": 19346, "train_loss": 3.5279288291931152, "perplexity": 34.05336419033061, "lr": 0.0026291804804649314, "grad_norm": 0.191443, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:32.705120+00:00", "epoch": 0, "step": 19347, "train_loss": 3.5549402236938477, "perplexity": 34.9857285875637, "lr": 0.0026291804804649314, "grad_norm": 0.187783, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:33.009292+00:00", "epoch": 0, "step": 19348, "train_loss": 3.49416184425354, "perplexity": 32.92268205044832, "lr": 0.0026291804804649314, "grad_norm": 0.168716, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:33.314772+00:00", "epoch": 0, "step": 19349, "train_loss": 3.6480865478515625, "perplexity": 38.40111700534843, "lr": 0.0026291804804649314, "grad_norm": 0.171298, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:33.621293+00:00", "epoch": 0, "step": 19350, "train_loss": 3.477447032928467, "perplexity": 32.37695916469804, "lr": 0.0026291804804649314, "grad_norm": 0.187636, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:33.928418+00:00", "epoch": 0, "step": 19351, "train_loss": 3.6400249004364014, "perplexity": 38.09278524056598, "lr": 0.0026291804804649314, "grad_norm": 0.20646, "tokens_per_sec": 106633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:34.234401+00:00", "epoch": 0, "step": 19352, "train_loss": 3.474475145339966, "perplexity": 32.28088131859647, "lr": 0.0026291804804649314, "grad_norm": 0.161862, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:34.538880+00:00", "epoch": 0, "step": 19353, "train_loss": 3.486546277999878, "perplexity": 32.67290947102574, "lr": 0.0026291804804649314, "grad_norm": 0.1687, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:34.843615+00:00", "epoch": 0, "step": 19354, "train_loss": 3.5165903568267822, "perplexity": 33.66943177821818, "lr": 0.0026291804804649314, "grad_norm": 0.187092, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:35.148895+00:00", "epoch": 0, "step": 19355, "train_loss": 3.5222387313842773, "perplexity": 33.86014744976174, "lr": 0.0026291804804649314, "grad_norm": 0.195993, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:35.454785+00:00", "epoch": 0, "step": 19356, "train_loss": 3.5386905670166016, "perplexity": 34.421816605883414, "lr": 0.0026291804804649314, "grad_norm": 0.174303, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:35.760603+00:00", "epoch": 0, "step": 19357, "train_loss": 3.634249210357666, "perplexity": 37.873407258243425, "lr": 0.0026291804804649314, "grad_norm": 0.15687, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:36.065805+00:00", "epoch": 0, "step": 19358, "train_loss": 3.484959125518799, "perplexity": 32.62109371232694, "lr": 0.0026291804804649314, "grad_norm": 0.168371, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:36.370655+00:00", "epoch": 0, "step": 19359, "train_loss": 3.517301082611084, "perplexity": 33.69337001727995, "lr": 0.0026291804804649314, "grad_norm": 0.196013, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:36.674933+00:00", "epoch": 0, "step": 19360, "train_loss": 3.558316946029663, "perplexity": 35.1040653614918, "lr": 0.0026291804804649314, "grad_norm": 0.178553, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:36.979388+00:00", "epoch": 0, "step": 19361, "train_loss": 3.5651562213897705, "perplexity": 35.34497461405854, "lr": 0.0026291804804649314, "grad_norm": 0.152938, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:37.285074+00:00", "epoch": 0, "step": 19362, "train_loss": 3.5171244144439697, "perplexity": 33.687417997136826, "lr": 0.0026291804804649314, "grad_norm": 0.183937, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:37.589379+00:00", "epoch": 0, "step": 19363, "train_loss": 3.5005953311920166, "perplexity": 33.13517248972223, "lr": 0.0026291804804649314, "grad_norm": 0.169701, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:37.894070+00:00", "epoch": 0, "step": 19364, "train_loss": 3.4940781593322754, "perplexity": 32.9199270336714, "lr": 0.0026291804804649314, "grad_norm": 0.16461, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:38.198616+00:00", "epoch": 0, "step": 19365, "train_loss": 3.505772829055786, "perplexity": 33.307174660708874, "lr": 0.0026291804804649314, "grad_norm": 0.171924, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:38.503398+00:00", "epoch": 0, "step": 19366, "train_loss": 3.537358522415161, "perplexity": 34.37599573539852, "lr": 0.0026291804804649314, "grad_norm": 0.161398, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:38.808744+00:00", "epoch": 0, "step": 19367, "train_loss": 3.5361547470092773, "perplexity": 34.334639653890385, "lr": 0.0026291804804649314, "grad_norm": 0.170944, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:39.114150+00:00", "epoch": 0, "step": 19368, "train_loss": 3.526707887649536, "perplexity": 34.01181239461058, "lr": 0.0026291804804649314, "grad_norm": 0.15088, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:39.419030+00:00", "epoch": 0, "step": 19369, "train_loss": 3.5041143894195557, "perplexity": 33.25198250116335, "lr": 0.0026291804804649314, "grad_norm": 0.171133, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:39.723655+00:00", "epoch": 0, "step": 19370, "train_loss": 3.5052244663238525, "perplexity": 33.28891525426549, "lr": 0.0026291804804649314, "grad_norm": 0.188128, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:40.028981+00:00", "epoch": 0, "step": 19371, "train_loss": 3.4573700428009033, "perplexity": 31.733409165672345, "lr": 0.0026291804804649314, "grad_norm": 0.210449, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:40.333228+00:00", "epoch": 0, "step": 19372, "train_loss": 3.517193555831909, "perplexity": 33.68974727249698, "lr": 0.0026291804804649314, "grad_norm": 0.201458, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:40.638771+00:00", "epoch": 0, "step": 19373, "train_loss": 3.501974105834961, "perplexity": 33.18088993512539, "lr": 0.0026291804804649314, "grad_norm": 0.166916, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:40.944476+00:00", "epoch": 0, "step": 19374, "train_loss": 3.4455223083496094, "perplexity": 31.359658581380273, "lr": 0.0026291804804649314, "grad_norm": 0.193367, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:41.249209+00:00", "epoch": 0, "step": 19375, "train_loss": 3.4813597202301025, "perplexity": 32.50388823664033, "lr": 0.0026291804804649314, "grad_norm": 0.179239, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:41.553154+00:00", "epoch": 0, "step": 19376, "train_loss": 3.494987964630127, "perplexity": 32.94989138648808, "lr": 0.0026291804804649314, "grad_norm": 0.210832, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:41.858550+00:00", "epoch": 0, "step": 19377, "train_loss": 3.5719964504241943, "perplexity": 35.58757109749163, "lr": 0.0026291804804649314, "grad_norm": 0.218124, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:42.163860+00:00", "epoch": 0, "step": 19378, "train_loss": 3.4946229457855225, "perplexity": 32.937866250038574, "lr": 0.0026291804804649314, "grad_norm": 0.170185, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:42.470399+00:00", "epoch": 0, "step": 19379, "train_loss": 3.3725321292877197, "perplexity": 29.152250947623035, "lr": 0.0026291804804649314, "grad_norm": 0.19723, "tokens_per_sec": 106897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:42.775243+00:00", "epoch": 0, "step": 19380, "train_loss": 3.5591068267822266, "perplexity": 35.131804340858295, "lr": 0.0026291804804649314, "grad_norm": 0.199723, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:43.079481+00:00", "epoch": 0, "step": 19381, "train_loss": 3.4703712463378906, "perplexity": 32.14867530757626, "lr": 0.0026291804804649314, "grad_norm": 0.156887, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:43.384869+00:00", "epoch": 0, "step": 19382, "train_loss": 3.4398059844970703, "perplexity": 31.18090800186761, "lr": 0.0026291804804649314, "grad_norm": 0.190846, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:43.690289+00:00", "epoch": 0, "step": 19383, "train_loss": 3.546428680419922, "perplexity": 34.68920975131455, "lr": 0.0026291804804649314, "grad_norm": 0.16325, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:43.995854+00:00", "epoch": 0, "step": 19384, "train_loss": 3.5680017471313477, "perplexity": 35.4456928794068, "lr": 0.0026291804804649314, "grad_norm": 0.158434, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:44.301025+00:00", "epoch": 0, "step": 19385, "train_loss": 3.4266183376312256, "perplexity": 30.772404717068206, "lr": 0.0026291804804649314, "grad_norm": 0.186567, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:44.605131+00:00", "epoch": 0, "step": 19386, "train_loss": 3.5624165534973145, "perplexity": 35.24827364670112, "lr": 0.0026291804804649314, "grad_norm": 0.187212, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:44.909700+00:00", "epoch": 0, "step": 19387, "train_loss": 3.6717474460601807, "perplexity": 39.3205564126542, "lr": 0.0026291804804649314, "grad_norm": 0.162918, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:45.215540+00:00", "epoch": 0, "step": 19388, "train_loss": 3.5673599243164062, "perplexity": 35.422950324151515, "lr": 0.0026291804804649314, "grad_norm": 0.185248, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:45.521347+00:00", "epoch": 0, "step": 19389, "train_loss": 3.586944818496704, "perplexity": 36.123543182691314, "lr": 0.0026291804804649314, "grad_norm": 0.18269, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:45.827224+00:00", "epoch": 0, "step": 19390, "train_loss": 3.4592785835266113, "perplexity": 31.794031501127094, "lr": 0.0026291804804649314, "grad_norm": 0.163452, "tokens_per_sec": 107189} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:46.132734+00:00", "epoch": 0, "step": 19391, "train_loss": 3.4886767864227295, "perplexity": 32.742593584790086, "lr": 0.0026291804804649314, "grad_norm": 0.16748, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:46.438637+00:00", "epoch": 0, "step": 19392, "train_loss": 3.575296640396118, "perplexity": 35.70521085275266, "lr": 0.0026291804804649314, "grad_norm": 0.162523, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:46.745245+00:00", "epoch": 0, "step": 19393, "train_loss": 3.4567999839782715, "perplexity": 31.715324410981996, "lr": 0.0026291804804649314, "grad_norm": 0.154447, "tokens_per_sec": 106868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:47.050514+00:00", "epoch": 0, "step": 19394, "train_loss": 3.4517927169799805, "perplexity": 31.556914245696248, "lr": 0.0026291804804649314, "grad_norm": 0.147905, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:47.356235+00:00", "epoch": 0, "step": 19395, "train_loss": 3.466782569885254, "perplexity": 32.03351088104568, "lr": 0.0026291804804649314, "grad_norm": 0.170624, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:47.662878+00:00", "epoch": 0, "step": 19396, "train_loss": 3.5085482597351074, "perplexity": 33.39974481670675, "lr": 0.0026291804804649314, "grad_norm": 0.178006, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:47.968261+00:00", "epoch": 0, "step": 19397, "train_loss": 3.5380592346191406, "perplexity": 34.40009185637135, "lr": 0.0026291804804649314, "grad_norm": 0.18173, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:48.273387+00:00", "epoch": 0, "step": 19398, "train_loss": 3.4936623573303223, "perplexity": 32.90624170749683, "lr": 0.0026291804804649314, "grad_norm": 0.212063, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:48.579143+00:00", "epoch": 0, "step": 19399, "train_loss": 3.5582237243652344, "perplexity": 35.1007930546179, "lr": 0.0026291804804649314, "grad_norm": 0.19322, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:48.884978+00:00", "epoch": 0, "step": 19400, "train_loss": 3.356247901916504, "perplexity": 28.681373408671607, "lr": 0.0026291804804649314, "grad_norm": 0.192125, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:49.190482+00:00", "epoch": 0, "step": 19401, "train_loss": 3.5809714794158936, "perplexity": 35.908408187280315, "lr": 0.0026291804804649314, "grad_norm": 0.17405, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:49.495481+00:00", "epoch": 0, "step": 19402, "train_loss": 3.526723861694336, "perplexity": 34.012355705164914, "lr": 0.0026291804804649314, "grad_norm": 0.158868, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:49.800132+00:00", "epoch": 0, "step": 19403, "train_loss": 3.437741756439209, "perplexity": 31.116609882522962, "lr": 0.0026291804804649314, "grad_norm": 0.175134, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:50.104806+00:00", "epoch": 0, "step": 19404, "train_loss": 3.5493850708007812, "perplexity": 34.79191634292598, "lr": 0.0026291804804649314, "grad_norm": 0.186174, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:50.410980+00:00", "epoch": 0, "step": 19405, "train_loss": 3.437241792678833, "perplexity": 31.101056593600333, "lr": 0.0026291804804649314, "grad_norm": 0.185403, "tokens_per_sec": 107027} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:50.717183+00:00", "epoch": 0, "step": 19406, "train_loss": 3.508885622024536, "perplexity": 33.41101453196603, "lr": 0.0026291804804649314, "grad_norm": 0.18961, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:51.021724+00:00", "epoch": 0, "step": 19407, "train_loss": 3.4430346488952637, "perplexity": 31.281743383645576, "lr": 0.0026291804804649314, "grad_norm": 0.18687, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:51.326538+00:00", "epoch": 0, "step": 19408, "train_loss": 3.5186147689819336, "perplexity": 33.737661624530276, "lr": 0.0026291804804649314, "grad_norm": 0.162275, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:51.631629+00:00", "epoch": 0, "step": 19409, "train_loss": 3.5247814655303955, "perplexity": 33.94635435705071, "lr": 0.0026291804804649314, "grad_norm": 0.167795, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:51.937255+00:00", "epoch": 0, "step": 19410, "train_loss": 3.2990572452545166, "perplexity": 27.0870903965413, "lr": 0.0026291804804649314, "grad_norm": 0.177484, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:52.242458+00:00", "epoch": 0, "step": 19411, "train_loss": 3.482787609100342, "perplexity": 32.55033332821632, "lr": 0.0026291804804649314, "grad_norm": 0.153132, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:52.548246+00:00", "epoch": 0, "step": 19412, "train_loss": 3.4784815311431885, "perplexity": 32.41047040182092, "lr": 0.0026291804804649314, "grad_norm": 0.155025, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:52.853184+00:00", "epoch": 0, "step": 19413, "train_loss": 3.521620512008667, "perplexity": 33.83922091980692, "lr": 0.0026291804804649314, "grad_norm": 0.166436, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:53.158375+00:00", "epoch": 0, "step": 19414, "train_loss": 3.3423683643341064, "perplexity": 28.286039083805793, "lr": 0.0026291804804649314, "grad_norm": 0.170062, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:53.464669+00:00", "epoch": 0, "step": 19415, "train_loss": 3.485236644744873, "perplexity": 32.63014794931199, "lr": 0.0026291804804649314, "grad_norm": 0.156749, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:53.770364+00:00", "epoch": 0, "step": 19416, "train_loss": 3.555529832839966, "perplexity": 35.00636257551878, "lr": 0.0026291804804649314, "grad_norm": 0.17344, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:54.075078+00:00", "epoch": 0, "step": 19417, "train_loss": 3.4361555576324463, "perplexity": 31.067291877477633, "lr": 0.0026291804804649314, "grad_norm": 0.209427, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:54.379682+00:00", "epoch": 0, "step": 19418, "train_loss": 3.5772440433502197, "perplexity": 35.77481103364722, "lr": 0.0026291804804649314, "grad_norm": 0.153229, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:54.684152+00:00", "epoch": 0, "step": 19419, "train_loss": 3.641545057296753, "perplexity": 38.15073628555882, "lr": 0.0026291804804649314, "grad_norm": 0.189168, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:54.988831+00:00", "epoch": 0, "step": 19420, "train_loss": 3.4021174907684326, "perplexity": 30.02761597609555, "lr": 0.0026291804804649314, "grad_norm": 0.233406, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:55.293789+00:00", "epoch": 0, "step": 19421, "train_loss": 3.4928975105285645, "perplexity": 32.88108309623713, "lr": 0.0026291804804649314, "grad_norm": 0.184593, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:55.599629+00:00", "epoch": 0, "step": 19422, "train_loss": 3.61350417137146, "perplexity": 37.095815421235415, "lr": 0.0026291804804649314, "grad_norm": 0.158032, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:55.905147+00:00", "epoch": 0, "step": 19423, "train_loss": 3.5738868713378906, "perplexity": 35.65491021574662, "lr": 0.0026291804804649314, "grad_norm": 0.191841, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:56.209606+00:00", "epoch": 0, "step": 19424, "train_loss": 3.5014259815216064, "perplexity": 33.16270766614029, "lr": 0.0026291804804649314, "grad_norm": 0.174532, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:56.515350+00:00", "epoch": 0, "step": 19425, "train_loss": 3.5389022827148438, "perplexity": 34.42910501632912, "lr": 0.0026291804804649314, "grad_norm": 0.209784, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:56.820712+00:00", "epoch": 0, "step": 19426, "train_loss": 3.577004909515381, "perplexity": 35.766257088703696, "lr": 0.0026291804804649314, "grad_norm": 0.209815, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:57.127113+00:00", "epoch": 0, "step": 19427, "train_loss": 3.5104308128356934, "perplexity": 33.46268083147796, "lr": 0.0026291804804649314, "grad_norm": 0.170035, "tokens_per_sec": 106950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:57.434040+00:00", "epoch": 0, "step": 19428, "train_loss": 3.4022719860076904, "perplexity": 30.032255458189837, "lr": 0.0026291804804649314, "grad_norm": 0.199776, "tokens_per_sec": 106756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:57.739901+00:00", "epoch": 0, "step": 19429, "train_loss": 3.444244623184204, "perplexity": 31.319616396918544, "lr": 0.0026291804804649314, "grad_norm": 0.176309, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:58.044854+00:00", "epoch": 0, "step": 19430, "train_loss": 3.560946464538574, "perplexity": 35.19649361874756, "lr": 0.0026291804804649314, "grad_norm": 0.166392, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:58.350069+00:00", "epoch": 0, "step": 19431, "train_loss": 3.5034449100494385, "perplexity": 33.229728435013605, "lr": 0.0026291804804649314, "grad_norm": 0.166357, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:58.656163+00:00", "epoch": 0, "step": 19432, "train_loss": 3.503105401992798, "perplexity": 33.21844858939752, "lr": 0.0026291804804649314, "grad_norm": 0.152242, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:58.962655+00:00", "epoch": 0, "step": 19433, "train_loss": 3.5491786003112793, "perplexity": 34.78473358046771, "lr": 0.0026291804804649314, "grad_norm": 0.161468, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:59.267246+00:00", "epoch": 0, "step": 19434, "train_loss": 3.5109798908233643, "perplexity": 33.48105949812989, "lr": 0.0026291804804649314, "grad_norm": 0.198105, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:59.571339+00:00", "epoch": 0, "step": 19435, "train_loss": 3.470151424407959, "perplexity": 32.14160910040751, "lr": 0.0026291804804649314, "grad_norm": 0.216392, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:39:59.877110+00:00", "epoch": 0, "step": 19436, "train_loss": 3.6111578941345215, "perplexity": 37.008880380661864, "lr": 0.0026291804804649314, "grad_norm": 0.174408, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:00.182281+00:00", "epoch": 0, "step": 19437, "train_loss": 3.492849349975586, "perplexity": 32.8794995632249, "lr": 0.0026291804804649314, "grad_norm": 0.195254, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:00.488849+00:00", "epoch": 0, "step": 19438, "train_loss": 3.4748198986053467, "perplexity": 32.29201217642997, "lr": 0.0026291804804649314, "grad_norm": 0.162034, "tokens_per_sec": 106884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:00.794277+00:00", "epoch": 0, "step": 19439, "train_loss": 3.445517063140869, "perplexity": 31.359494093856377, "lr": 0.0026291804804649314, "grad_norm": 0.170796, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:01.099540+00:00", "epoch": 0, "step": 19440, "train_loss": 3.423738956451416, "perplexity": 30.68392667620098, "lr": 0.0026291804804649314, "grad_norm": 0.170481, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:01.405036+00:00", "epoch": 0, "step": 19441, "train_loss": 3.537688732147217, "perplexity": 34.38734889809869, "lr": 0.0026291804804649314, "grad_norm": 0.156843, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:01.709878+00:00", "epoch": 0, "step": 19442, "train_loss": 3.509845018386841, "perplexity": 33.44308431912762, "lr": 0.0026291804804649314, "grad_norm": 0.175344, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:02.015244+00:00", "epoch": 0, "step": 19443, "train_loss": 3.4128198623657227, "perplexity": 30.350708524777712, "lr": 0.0026291804804649314, "grad_norm": 0.162688, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:02.321762+00:00", "epoch": 0, "step": 19444, "train_loss": 3.442856550216675, "perplexity": 31.27617264257053, "lr": 0.0026291804804649314, "grad_norm": 0.147922, "tokens_per_sec": 106905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:02.627398+00:00", "epoch": 0, "step": 19445, "train_loss": 3.538546085357666, "perplexity": 34.416843643976165, "lr": 0.0026291804804649314, "grad_norm": 0.177557, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:02.931849+00:00", "epoch": 0, "step": 19446, "train_loss": 3.5804216861724854, "perplexity": 35.888671413147094, "lr": 0.0026291804804649314, "grad_norm": 0.189788, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:03.237853+00:00", "epoch": 0, "step": 19447, "train_loss": 3.3918302059173584, "perplexity": 29.72029678741688, "lr": 0.0026291804804649314, "grad_norm": 0.16844, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:03.543041+00:00", "epoch": 0, "step": 19448, "train_loss": 3.6459033489227295, "perplexity": 38.317371178000194, "lr": 0.0026291804804649314, "grad_norm": 0.171352, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:03.848857+00:00", "epoch": 0, "step": 19449, "train_loss": 3.4949100017547607, "perplexity": 32.94732261834817, "lr": 0.0026291804804649314, "grad_norm": 0.19316, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:04.156061+00:00", "epoch": 0, "step": 19450, "train_loss": 3.5658652782440186, "perplexity": 35.37004509772063, "lr": 0.0026291804804649314, "grad_norm": 0.165098, "tokens_per_sec": 106729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:04.461745+00:00", "epoch": 0, "step": 19451, "train_loss": 3.3992631435394287, "perplexity": 29.942028939309818, "lr": 0.0026291804804649314, "grad_norm": 0.170097, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:04.766643+00:00", "epoch": 0, "step": 19452, "train_loss": 3.5135724544525146, "perplexity": 33.56797389211642, "lr": 0.0026291804804649314, "grad_norm": 0.181421, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:05.071692+00:00", "epoch": 0, "step": 19453, "train_loss": 3.5117998123168945, "perplexity": 33.50852259569201, "lr": 0.0026291804804649314, "grad_norm": 0.170614, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:05.377538+00:00", "epoch": 0, "step": 19454, "train_loss": 3.5083603858947754, "perplexity": 33.3934704677934, "lr": 0.0026291804804649314, "grad_norm": 0.187967, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:05.684046+00:00", "epoch": 0, "step": 19455, "train_loss": 3.5970113277435303, "perplexity": 36.489017603725344, "lr": 0.0026291804804649314, "grad_norm": 0.177907, "tokens_per_sec": 106907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:05.989039+00:00", "epoch": 0, "step": 19456, "train_loss": 3.510576009750366, "perplexity": 33.467539862240976, "lr": 0.0026291804804649314, "grad_norm": 0.188524, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:06.295179+00:00", "epoch": 0, "step": 19457, "train_loss": 3.5545198917388916, "perplexity": 34.97102605805888, "lr": 0.0026291804804649314, "grad_norm": 0.200984, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:06.600049+00:00", "epoch": 0, "step": 19458, "train_loss": 3.5129027366638184, "perplexity": 33.5455003491761, "lr": 0.0026291804804649314, "grad_norm": 0.196375, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:06.905465+00:00", "epoch": 0, "step": 19459, "train_loss": 3.5354034900665283, "perplexity": 34.30885520405979, "lr": 0.0026291804804649314, "grad_norm": 0.172962, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:07.211582+00:00", "epoch": 0, "step": 19460, "train_loss": 3.4873762130737305, "perplexity": 32.700037120097164, "lr": 0.0026291804804649314, "grad_norm": 0.179286, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:07.518772+00:00", "epoch": 0, "step": 19461, "train_loss": 3.5128605365753174, "perplexity": 33.54408475596185, "lr": 0.0026291804804649314, "grad_norm": 0.173971, "tokens_per_sec": 106612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:07.824599+00:00", "epoch": 0, "step": 19462, "train_loss": 3.5985267162323, "perplexity": 36.54435455887129, "lr": 0.0026291804804649314, "grad_norm": 0.200008, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:08.129827+00:00", "epoch": 0, "step": 19463, "train_loss": 3.347139596939087, "perplexity": 28.421320829449215, "lr": 0.0026291804804649314, "grad_norm": 0.158128, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:08.435230+00:00", "epoch": 0, "step": 19464, "train_loss": 3.510890007019043, "perplexity": 33.478050228373824, "lr": 0.0026291804804649314, "grad_norm": 0.179355, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:08.740600+00:00", "epoch": 0, "step": 19465, "train_loss": 3.5169177055358887, "perplexity": 33.68045522740462, "lr": 0.0026291804804649314, "grad_norm": 0.207726, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:09.046544+00:00", "epoch": 0, "step": 19466, "train_loss": 3.6052348613739014, "perplexity": 36.79032346856507, "lr": 0.0026291804804649314, "grad_norm": 0.187771, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:09.351341+00:00", "epoch": 0, "step": 19467, "train_loss": 3.492663860321045, "perplexity": 32.873401321807265, "lr": 0.0026291804804649314, "grad_norm": 0.163465, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:09.656613+00:00", "epoch": 0, "step": 19468, "train_loss": 3.3711447715759277, "perplexity": 29.11183439004886, "lr": 0.0026291804804649314, "grad_norm": 0.177168, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:09.962350+00:00", "epoch": 0, "step": 19469, "train_loss": 3.5021324157714844, "perplexity": 33.186143215517085, "lr": 0.0026291804804649314, "grad_norm": 0.228212, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:10.268709+00:00", "epoch": 0, "step": 19470, "train_loss": 3.501817464828491, "perplexity": 33.17569285417902, "lr": 0.0026291804804649314, "grad_norm": 0.205493, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:10.574780+00:00", "epoch": 0, "step": 19471, "train_loss": 3.5143203735351562, "perplexity": 33.59308941137324, "lr": 0.0026291804804649314, "grad_norm": 0.18247, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:10.879502+00:00", "epoch": 0, "step": 19472, "train_loss": 3.4264495372772217, "perplexity": 30.76721076264142, "lr": 0.0026291804804649314, "grad_norm": 0.17685, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:11.184077+00:00", "epoch": 0, "step": 19473, "train_loss": 3.4791715145111084, "perplexity": 32.43284080405988, "lr": 0.0026291804804649314, "grad_norm": 0.213427, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:11.488958+00:00", "epoch": 0, "step": 19474, "train_loss": 3.519052743911743, "perplexity": 33.752441110798195, "lr": 0.0026291804804649314, "grad_norm": 0.180061, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:11.794455+00:00", "epoch": 0, "step": 19475, "train_loss": 3.5015885829925537, "perplexity": 33.16810040960953, "lr": 0.0026291804804649314, "grad_norm": 0.21199, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:12.100271+00:00", "epoch": 0, "step": 19476, "train_loss": 3.57771897315979, "perplexity": 35.79180559313085, "lr": 0.0026291804804649314, "grad_norm": 0.18425, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:12.406528+00:00", "epoch": 0, "step": 19477, "train_loss": 3.4724676609039307, "perplexity": 32.21614295413868, "lr": 0.0026291804804649314, "grad_norm": 0.18337, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:12.711358+00:00", "epoch": 0, "step": 19478, "train_loss": 3.620382785797119, "perplexity": 37.351862848506904, "lr": 0.0026291804804649314, "grad_norm": 0.174532, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:13.016493+00:00", "epoch": 0, "step": 19479, "train_loss": 3.5304696559906006, "perplexity": 34.13999790463387, "lr": 0.0026291804804649314, "grad_norm": 0.16841, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:13.320590+00:00", "epoch": 0, "step": 19480, "train_loss": 3.539997100830078, "perplexity": 34.46681926554682, "lr": 0.0026291804804649314, "grad_norm": 0.150923, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:13.626752+00:00", "epoch": 0, "step": 19481, "train_loss": 3.466290235519409, "perplexity": 32.017743564494666, "lr": 0.0026291804804649314, "grad_norm": 0.164654, "tokens_per_sec": 106943} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:13.933006+00:00", "epoch": 0, "step": 19482, "train_loss": 3.54929256439209, "perplexity": 34.78869801655381, "lr": 0.0026291804804649314, "grad_norm": 0.157999, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:14.237791+00:00", "epoch": 0, "step": 19483, "train_loss": 3.4795703887939453, "perplexity": 32.445780010562764, "lr": 0.0026291804804649314, "grad_norm": 0.175981, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:14.542414+00:00", "epoch": 0, "step": 19484, "train_loss": 3.4659225940704346, "perplexity": 32.00597467835545, "lr": 0.0026291804804649314, "grad_norm": 0.152692, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:14.848115+00:00", "epoch": 0, "step": 19485, "train_loss": 3.4986557960510254, "perplexity": 33.070967941891006, "lr": 0.0026291804804649314, "grad_norm": 0.162497, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:15.153550+00:00", "epoch": 0, "step": 19486, "train_loss": 3.595078945159912, "perplexity": 36.41857494461837, "lr": 0.0026291804804649314, "grad_norm": 0.171264, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:15.459946+00:00", "epoch": 0, "step": 19487, "train_loss": 3.5262553691864014, "perplexity": 33.99642490336186, "lr": 0.0026291804804649314, "grad_norm": 0.164978, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:15.764406+00:00", "epoch": 0, "step": 19488, "train_loss": 3.4972593784332275, "perplexity": 33.02481928851472, "lr": 0.0026291804804649314, "grad_norm": 0.170298, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:16.069242+00:00", "epoch": 0, "step": 19489, "train_loss": 3.4345710277557373, "perplexity": 31.018103805585593, "lr": 0.0026291804804649314, "grad_norm": 0.174608, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:16.374718+00:00", "epoch": 0, "step": 19490, "train_loss": 3.418705940246582, "perplexity": 30.52988195597697, "lr": 0.0026291804804649314, "grad_norm": 0.17087, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:16.680116+00:00", "epoch": 0, "step": 19491, "train_loss": 3.4784433841705322, "perplexity": 32.40923406407415, "lr": 0.0026291804804649314, "grad_norm": 0.189709, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:16.986717+00:00", "epoch": 0, "step": 19492, "train_loss": 3.4454822540283203, "perplexity": 31.358402516695516, "lr": 0.0026291804804649314, "grad_norm": 0.181072, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:17.293856+00:00", "epoch": 0, "step": 19493, "train_loss": 3.524956464767456, "perplexity": 33.95229546299352, "lr": 0.0026291804804649314, "grad_norm": 0.182274, "tokens_per_sec": 106696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:17.599799+00:00", "epoch": 0, "step": 19494, "train_loss": 3.5166003704071045, "perplexity": 33.669768931465754, "lr": 0.0026291804804649314, "grad_norm": 0.170492, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:17.903899+00:00", "epoch": 0, "step": 19495, "train_loss": 3.5061750411987305, "perplexity": 33.32057390529338, "lr": 0.0026291804804649314, "grad_norm": 0.197455, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:18.209957+00:00", "epoch": 0, "step": 19496, "train_loss": 3.480832099914551, "perplexity": 32.48674304834482, "lr": 0.0026291804804649314, "grad_norm": 0.184089, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:18.515934+00:00", "epoch": 0, "step": 19497, "train_loss": 3.471437692642212, "perplexity": 32.18297843151563, "lr": 0.0026291804804649314, "grad_norm": 0.164326, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:18.821266+00:00", "epoch": 0, "step": 19498, "train_loss": 3.418933153152466, "perplexity": 30.536819527295496, "lr": 0.0026291804804649314, "grad_norm": 0.180509, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:19.126289+00:00", "epoch": 0, "step": 19499, "train_loss": 3.607448101043701, "perplexity": 36.87183944585277, "lr": 0.0026291804804649314, "grad_norm": 0.163882, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:19.430300+00:00", "epoch": 0, "step": 19500, "train_loss": 3.5180931091308594, "perplexity": 33.72006663069157, "lr": 0.0026291804804649314, "grad_norm": 0.168993, "tokens_per_sec": 107845} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:40:22.504572+00:00", "step": 19500, "epoch": 0, "val_loss": 3.4561077117919923, "val_ppl": 31.693376371898484, "eval_train_loss": 3.5180931091308594, "eval_train_ppl": 33.72006663069157} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:40:23.463019+00:00", "step": 19500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4561_epoch_0000_step_0019500.pt", "val_loss": 3.4561077117919923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:25.016098+00:00", "epoch": 0, "step": 19501, "train_loss": 3.415118932723999, "perplexity": 30.420567213348647, "lr": 0.0026291804804649314, "grad_norm": 0.159513, "tokens_per_sec": 5866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:25.319810+00:00", "epoch": 0, "step": 19502, "train_loss": 3.4456374645233154, "perplexity": 31.36327004760856, "lr": 0.0026291804804649314, "grad_norm": 0.173977, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:25.624882+00:00", "epoch": 0, "step": 19503, "train_loss": 3.47416615486145, "perplexity": 32.270908374482666, "lr": 0.0026291804804649314, "grad_norm": 0.181057, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:25.929141+00:00", "epoch": 0, "step": 19504, "train_loss": 3.4471070766448975, "perplexity": 31.409395774605024, "lr": 0.0026291804804649314, "grad_norm": 0.179841, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:26.233435+00:00", "epoch": 0, "step": 19505, "train_loss": 3.529604911804199, "perplexity": 34.110488300890125, "lr": 0.0026291804804649314, "grad_norm": 0.160916, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:26.537730+00:00", "epoch": 0, "step": 19506, "train_loss": 3.441849708557129, "perplexity": 31.244698336482998, "lr": 0.0026291804804649314, "grad_norm": 0.174865, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:26.841831+00:00", "epoch": 0, "step": 19507, "train_loss": 3.470276117324829, "perplexity": 32.14561718128359, "lr": 0.0026291804804649314, "grad_norm": 0.159153, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:27.146067+00:00", "epoch": 0, "step": 19508, "train_loss": 3.478332281112671, "perplexity": 32.40563349908728, "lr": 0.0026291804804649314, "grad_norm": 0.169091, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:27.449720+00:00", "epoch": 0, "step": 19509, "train_loss": 3.4473161697387695, "perplexity": 31.41596394900021, "lr": 0.0026291804804649314, "grad_norm": 0.164914, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:27.754075+00:00", "epoch": 0, "step": 19510, "train_loss": 3.523117780685425, "perplexity": 33.889925274884874, "lr": 0.0026291804804649314, "grad_norm": 0.163236, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:28.058482+00:00", "epoch": 0, "step": 19511, "train_loss": 3.5581440925598145, "perplexity": 35.097998026383344, "lr": 0.0026291804804649314, "grad_norm": 0.155072, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:28.362704+00:00", "epoch": 0, "step": 19512, "train_loss": 3.5725860595703125, "perplexity": 35.6085600419276, "lr": 0.0026291804804649314, "grad_norm": 0.184459, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:28.667498+00:00", "epoch": 0, "step": 19513, "train_loss": 3.4948019981384277, "perplexity": 32.94376438051163, "lr": 0.0026291804804649314, "grad_norm": 0.149023, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:28.971478+00:00", "epoch": 0, "step": 19514, "train_loss": 3.4568896293640137, "perplexity": 31.71816767091342, "lr": 0.0026291804804649314, "grad_norm": 0.159193, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:29.275923+00:00", "epoch": 0, "step": 19515, "train_loss": 3.4870004653930664, "perplexity": 32.687752465101504, "lr": 0.0026291804804649314, "grad_norm": 0.178836, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:29.581858+00:00", "epoch": 0, "step": 19516, "train_loss": 3.6212735176086426, "perplexity": 37.38514816290921, "lr": 0.0026291804804649314, "grad_norm": 0.156291, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:29.886542+00:00", "epoch": 0, "step": 19517, "train_loss": 3.487859010696411, "perplexity": 32.71582843197749, "lr": 0.0026291804804649314, "grad_norm": 0.163916, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:30.191586+00:00", "epoch": 0, "step": 19518, "train_loss": 3.5830564498901367, "perplexity": 35.98335426115361, "lr": 0.0026291804804649314, "grad_norm": 0.16469, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:30.496479+00:00", "epoch": 0, "step": 19519, "train_loss": 3.4765498638153076, "perplexity": 32.347924583369775, "lr": 0.0026291804804649314, "grad_norm": 0.173419, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:30.799848+00:00", "epoch": 0, "step": 19520, "train_loss": 3.53210186958313, "perplexity": 34.19576717456161, "lr": 0.0026291804804649314, "grad_norm": 0.161336, "tokens_per_sec": 108085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:31.104735+00:00", "epoch": 0, "step": 19521, "train_loss": 3.54866886138916, "perplexity": 34.767006966222645, "lr": 0.0026291804804649314, "grad_norm": 0.166142, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:31.409693+00:00", "epoch": 0, "step": 19522, "train_loss": 3.5048747062683105, "perplexity": 33.277274157331256, "lr": 0.0026291804804649314, "grad_norm": 0.194559, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:31.715137+00:00", "epoch": 0, "step": 19523, "train_loss": 3.567976713180542, "perplexity": 35.44480554478178, "lr": 0.0026291804804649314, "grad_norm": 0.193218, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:32.019253+00:00", "epoch": 0, "step": 19524, "train_loss": 3.489467144012451, "perplexity": 32.768482171409815, "lr": 0.0026291804804649314, "grad_norm": 0.176919, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:32.322517+00:00", "epoch": 0, "step": 19525, "train_loss": 3.430406332015991, "perplexity": 30.889191467472802, "lr": 0.0026291804804649314, "grad_norm": 0.165909, "tokens_per_sec": 108053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:32.626079+00:00", "epoch": 0, "step": 19526, "train_loss": 3.452868938446045, "perplexity": 31.590894756207806, "lr": 0.0026291804804649314, "grad_norm": 0.204595, "tokens_per_sec": 107944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:32.931235+00:00", "epoch": 0, "step": 19527, "train_loss": 3.584968328475952, "perplexity": 36.052215872155344, "lr": 0.0026291804804649314, "grad_norm": 0.183714, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:33.236969+00:00", "epoch": 0, "step": 19528, "train_loss": 3.4965202808380127, "perplexity": 33.000419741929996, "lr": 0.0026291804804649314, "grad_norm": 0.174123, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:33.541090+00:00", "epoch": 0, "step": 19529, "train_loss": 3.615882635116577, "perplexity": 37.184151483727646, "lr": 0.0026291804804649314, "grad_norm": 0.160055, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:33.844477+00:00", "epoch": 0, "step": 19530, "train_loss": 3.4340837001800537, "perplexity": 31.00299151087377, "lr": 0.0026291804804649314, "grad_norm": 0.19435, "tokens_per_sec": 108066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:34.148355+00:00", "epoch": 0, "step": 19531, "train_loss": 3.554396152496338, "perplexity": 34.96669903749977, "lr": 0.0026291804804649314, "grad_norm": 0.193283, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:34.453229+00:00", "epoch": 0, "step": 19532, "train_loss": 3.6165878772735596, "perplexity": 37.21038456417542, "lr": 0.0026291804804649314, "grad_norm": 0.202713, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:34.757457+00:00", "epoch": 0, "step": 19533, "train_loss": 3.5012011528015137, "perplexity": 33.15525257511159, "lr": 0.0026291804804649314, "grad_norm": 0.163042, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:35.061211+00:00", "epoch": 0, "step": 19534, "train_loss": 3.442960023880005, "perplexity": 31.279409070168423, "lr": 0.0026291804804649314, "grad_norm": 0.203623, "tokens_per_sec": 107877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:35.364660+00:00", "epoch": 0, "step": 19535, "train_loss": 3.4155473709106445, "perplexity": 30.43360333838974, "lr": 0.0026291804804649314, "grad_norm": 0.235007, "tokens_per_sec": 107986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:35.668425+00:00", "epoch": 0, "step": 19536, "train_loss": 3.5177297592163086, "perplexity": 33.707816673007976, "lr": 0.0026291804804649314, "grad_norm": 0.184976, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:35.973598+00:00", "epoch": 0, "step": 19537, "train_loss": 3.4801442623138428, "perplexity": 32.46440512826302, "lr": 0.0026291804804649314, "grad_norm": 0.192784, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:36.278383+00:00", "epoch": 0, "step": 19538, "train_loss": 3.4637813568115234, "perplexity": 31.93751561257405, "lr": 0.0026291804804649314, "grad_norm": 0.164521, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:36.583097+00:00", "epoch": 0, "step": 19539, "train_loss": 3.491990327835083, "perplexity": 32.85126747286286, "lr": 0.0026291804804649314, "grad_norm": 0.17451, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:36.887816+00:00", "epoch": 0, "step": 19540, "train_loss": 3.547292709350586, "perplexity": 34.719195184396945, "lr": 0.0026291804804649314, "grad_norm": 0.162914, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:37.192491+00:00", "epoch": 0, "step": 19541, "train_loss": 3.5438034534454346, "perplexity": 34.59826213345444, "lr": 0.0026291804804649314, "grad_norm": 0.166026, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:37.497436+00:00", "epoch": 0, "step": 19542, "train_loss": 3.50296688079834, "perplexity": 33.21384744890597, "lr": 0.0026291804804649314, "grad_norm": 0.160025, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:37.803661+00:00", "epoch": 0, "step": 19543, "train_loss": 3.5759103298187256, "perplexity": 35.72712948791433, "lr": 0.0026291804804649314, "grad_norm": 0.163482, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:38.108218+00:00", "epoch": 0, "step": 19544, "train_loss": 3.3757824897766113, "perplexity": 29.24716043372863, "lr": 0.0026291804804649314, "grad_norm": 0.155358, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:38.412792+00:00", "epoch": 0, "step": 19545, "train_loss": 3.3076772689819336, "perplexity": 27.321591006836805, "lr": 0.0026291804804649314, "grad_norm": 0.159505, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:38.716736+00:00", "epoch": 0, "step": 19546, "train_loss": 3.4201526641845703, "perplexity": 30.574082232116112, "lr": 0.0026291804804649314, "grad_norm": 0.182672, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:39.021967+00:00", "epoch": 0, "step": 19547, "train_loss": 3.6009292602539062, "perplexity": 36.63225953494366, "lr": 0.0026291804804649314, "grad_norm": 0.215149, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:39.327437+00:00", "epoch": 0, "step": 19548, "train_loss": 3.3553097248077393, "perplexity": 28.65447781907764, "lr": 0.0026291804804649314, "grad_norm": 0.188395, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:39.632275+00:00", "epoch": 0, "step": 19549, "train_loss": 3.587587833404541, "perplexity": 36.14677862905028, "lr": 0.0026291804804649314, "grad_norm": 0.170908, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:39.935999+00:00", "epoch": 0, "step": 19550, "train_loss": 3.5446572303771973, "perplexity": 34.627813945075474, "lr": 0.0026291804804649314, "grad_norm": 0.175696, "tokens_per_sec": 107945} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:40.240171+00:00", "epoch": 0, "step": 19551, "train_loss": 3.560016393661499, "perplexity": 35.16377360338295, "lr": 0.0026291804804649314, "grad_norm": 0.190214, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:40.544267+00:00", "epoch": 0, "step": 19552, "train_loss": 3.526618480682373, "perplexity": 34.00877163755111, "lr": 0.0026291804804649314, "grad_norm": 0.194747, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:40.850079+00:00", "epoch": 0, "step": 19553, "train_loss": 3.527841091156006, "perplexity": 34.05037654606652, "lr": 0.0026291804804649314, "grad_norm": 0.171246, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:41.155448+00:00", "epoch": 0, "step": 19554, "train_loss": 3.53973388671875, "perplexity": 34.457748306198155, "lr": 0.0026291804804649314, "grad_norm": 0.161506, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:41.460226+00:00", "epoch": 0, "step": 19555, "train_loss": 3.39875864982605, "perplexity": 29.926927183634827, "lr": 0.0026291804804649314, "grad_norm": 0.173633, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:41.765611+00:00", "epoch": 0, "step": 19556, "train_loss": 3.4534285068511963, "perplexity": 31.60857696954642, "lr": 0.0026291804804649314, "grad_norm": 0.162494, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:42.069981+00:00", "epoch": 0, "step": 19557, "train_loss": 3.496100425720215, "perplexity": 32.98656725503412, "lr": 0.0026291804804649314, "grad_norm": 0.182035, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:42.374545+00:00", "epoch": 0, "step": 19558, "train_loss": 3.5801241397857666, "perplexity": 35.87799445716759, "lr": 0.0026291804804649314, "grad_norm": 0.179387, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:42.680700+00:00", "epoch": 0, "step": 19559, "train_loss": 3.4784791469573975, "perplexity": 32.410393129330025, "lr": 0.0026291804804649314, "grad_norm": 0.188525, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:42.986051+00:00", "epoch": 0, "step": 19560, "train_loss": 3.530259609222412, "perplexity": 34.132827661477606, "lr": 0.0026291804804649314, "grad_norm": 0.180009, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:43.290547+00:00", "epoch": 0, "step": 19561, "train_loss": 3.5400800704956055, "perplexity": 34.46967908465055, "lr": 0.0026291804804649314, "grad_norm": 0.178537, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:43.595922+00:00", "epoch": 0, "step": 19562, "train_loss": 3.5273263454437256, "perplexity": 34.032853771011474, "lr": 0.0026291804804649314, "grad_norm": 0.187538, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:43.900566+00:00", "epoch": 0, "step": 19563, "train_loss": 3.436030864715576, "perplexity": 31.063418247746288, "lr": 0.0026291804804649314, "grad_norm": 0.159975, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:44.206371+00:00", "epoch": 0, "step": 19564, "train_loss": 3.529444932937622, "perplexity": 34.105031780108995, "lr": 0.0026291804804649314, "grad_norm": 0.147276, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:44.512122+00:00", "epoch": 0, "step": 19565, "train_loss": 3.479372501373291, "perplexity": 32.439360034082604, "lr": 0.0026291804804649314, "grad_norm": 0.175645, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:44.817807+00:00", "epoch": 0, "step": 19566, "train_loss": 3.5365517139434814, "perplexity": 34.34827207616221, "lr": 0.0026291804804649314, "grad_norm": 0.194875, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:45.122218+00:00", "epoch": 0, "step": 19567, "train_loss": 3.4491827487945557, "perplexity": 31.474659091845172, "lr": 0.0026291804804649314, "grad_norm": 0.173493, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:45.426334+00:00", "epoch": 0, "step": 19568, "train_loss": 3.4947330951690674, "perplexity": 32.941494535524335, "lr": 0.0026291804804649314, "grad_norm": 0.173944, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:45.731512+00:00", "epoch": 0, "step": 19569, "train_loss": 3.552446126937866, "perplexity": 34.8985795196676, "lr": 0.0026291804804649314, "grad_norm": 0.160181, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:46.037068+00:00", "epoch": 0, "step": 19570, "train_loss": 3.445767402648926, "perplexity": 31.36734559691044, "lr": 0.0026291804804649314, "grad_norm": 0.160308, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:46.341287+00:00", "epoch": 0, "step": 19571, "train_loss": 3.6024422645568848, "perplexity": 36.687726251357034, "lr": 0.0026291804804649314, "grad_norm": 0.158406, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:46.645391+00:00", "epoch": 0, "step": 19572, "train_loss": 3.5649092197418213, "perplexity": 35.336245425188544, "lr": 0.0026291804804649314, "grad_norm": 0.160339, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:46.949593+00:00", "epoch": 0, "step": 19573, "train_loss": 3.484029531478882, "perplexity": 32.59078342834876, "lr": 0.0026291804804649314, "grad_norm": 0.164136, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:47.254132+00:00", "epoch": 0, "step": 19574, "train_loss": 3.6331467628479004, "perplexity": 37.83167682175969, "lr": 0.0026291804804649314, "grad_norm": 0.162029, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:47.559860+00:00", "epoch": 0, "step": 19575, "train_loss": 3.5850348472595215, "perplexity": 36.0546141014629, "lr": 0.0026291804804649314, "grad_norm": 0.177717, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:47.864708+00:00", "epoch": 0, "step": 19576, "train_loss": 3.4037251472473145, "perplexity": 30.075928892345676, "lr": 0.0026291804804649314, "grad_norm": 0.186067, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:48.170590+00:00", "epoch": 0, "step": 19577, "train_loss": 3.5435259342193604, "perplexity": 34.58866178272622, "lr": 0.0026291804804649314, "grad_norm": 0.172205, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:48.475680+00:00", "epoch": 0, "step": 19578, "train_loss": 3.5783884525299072, "perplexity": 35.81577549137561, "lr": 0.0026291804804649314, "grad_norm": 0.185527, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:48.780979+00:00", "epoch": 0, "step": 19579, "train_loss": 3.449923276901245, "perplexity": 31.497975593748126, "lr": 0.0026291804804649314, "grad_norm": 0.182783, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:49.086629+00:00", "epoch": 0, "step": 19580, "train_loss": 3.5472848415374756, "perplexity": 34.71892202133249, "lr": 0.0026291804804649314, "grad_norm": 0.172559, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:49.391394+00:00", "epoch": 0, "step": 19581, "train_loss": 3.4093050956726074, "perplexity": 30.244220115924804, "lr": 0.0026291804804649314, "grad_norm": 0.153242, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:49.696004+00:00", "epoch": 0, "step": 19582, "train_loss": 3.562411308288574, "perplexity": 35.248088762632996, "lr": 0.0026291804804649314, "grad_norm": 0.202353, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:50.001209+00:00", "epoch": 0, "step": 19583, "train_loss": 3.6275601387023926, "perplexity": 37.62091473500687, "lr": 0.0026291804804649314, "grad_norm": 0.163081, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:50.305705+00:00", "epoch": 0, "step": 19584, "train_loss": 3.475498676300049, "perplexity": 32.313938714793096, "lr": 0.0026291804804649314, "grad_norm": 0.194897, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:50.612456+00:00", "epoch": 0, "step": 19585, "train_loss": 3.4612627029418945, "perplexity": 31.857177279985866, "lr": 0.0026291804804649314, "grad_norm": 0.197646, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:50.917452+00:00", "epoch": 0, "step": 19586, "train_loss": 3.6129860877990723, "perplexity": 37.07660166625652, "lr": 0.0026291804804649314, "grad_norm": 0.182305, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:51.222936+00:00", "epoch": 0, "step": 19587, "train_loss": 3.530988931655884, "perplexity": 34.157730578433096, "lr": 0.0026291804804649314, "grad_norm": 0.173395, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:51.527435+00:00", "epoch": 0, "step": 19588, "train_loss": 3.5978915691375732, "perplexity": 36.521150887896724, "lr": 0.0026291804804649314, "grad_norm": 0.165205, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:51.831969+00:00", "epoch": 0, "step": 19589, "train_loss": 3.4280083179473877, "perplexity": 30.815207494523566, "lr": 0.0026291804804649314, "grad_norm": 0.213164, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:52.136431+00:00", "epoch": 0, "step": 19590, "train_loss": 3.468562364578247, "perplexity": 32.09057471960494, "lr": 0.0026291804804649314, "grad_norm": 0.190035, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:52.442101+00:00", "epoch": 0, "step": 19591, "train_loss": 3.5861711502075195, "perplexity": 36.0956063511517, "lr": 0.0026291804804649314, "grad_norm": 0.195406, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:52.747439+00:00", "epoch": 0, "step": 19592, "train_loss": 3.4542510509490967, "perplexity": 31.63458711375456, "lr": 0.0026291804804649314, "grad_norm": 0.16682, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:53.052225+00:00", "epoch": 0, "step": 19593, "train_loss": 3.4315133094787598, "perplexity": 30.923404039049608, "lr": 0.0026291804804649314, "grad_norm": 0.164175, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:53.355892+00:00", "epoch": 0, "step": 19594, "train_loss": 3.600820302963257, "perplexity": 36.6282684006299, "lr": 0.0026291804804649314, "grad_norm": 0.151478, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:53.660079+00:00", "epoch": 0, "step": 19595, "train_loss": 3.5160937309265137, "perplexity": 33.65271481773116, "lr": 0.0026291804804649314, "grad_norm": 0.164128, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:53.966748+00:00", "epoch": 0, "step": 19596, "train_loss": 3.4135067462921143, "perplexity": 30.37156310013433, "lr": 0.0026291804804649314, "grad_norm": 0.161313, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:54.271646+00:00", "epoch": 0, "step": 19597, "train_loss": 3.475050449371338, "perplexity": 32.29945798285785, "lr": 0.0026291804804649314, "grad_norm": 0.172099, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:54.576887+00:00", "epoch": 0, "step": 19598, "train_loss": 3.6148645877838135, "perplexity": 37.14631552015729, "lr": 0.0026291804804649314, "grad_norm": 0.168949, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:54.882312+00:00", "epoch": 0, "step": 19599, "train_loss": 3.43725848197937, "perplexity": 31.1015756528122, "lr": 0.0026291804804649314, "grad_norm": 0.164484, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:55.186301+00:00", "epoch": 0, "step": 19600, "train_loss": 3.364727258682251, "perplexity": 28.925607014921887, "lr": 0.0026291804804649314, "grad_norm": 0.198105, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:55.491163+00:00", "epoch": 0, "step": 19601, "train_loss": 3.4099154472351074, "perplexity": 30.26268535750044, "lr": 0.0026291804804649314, "grad_norm": 0.210602, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:55.796957+00:00", "epoch": 0, "step": 19602, "train_loss": 3.5190770626068115, "perplexity": 33.75326193610204, "lr": 0.0026291804804649314, "grad_norm": 0.189598, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:56.101684+00:00", "epoch": 0, "step": 19603, "train_loss": 3.534606695175171, "perplexity": 34.28152897165181, "lr": 0.0026291804804649314, "grad_norm": 0.182083, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:56.405783+00:00", "epoch": 0, "step": 19604, "train_loss": 3.4011764526367188, "perplexity": 29.999372135807217, "lr": 0.0026291804804649314, "grad_norm": 0.175665, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:56.710264+00:00", "epoch": 0, "step": 19605, "train_loss": 3.477476119995117, "perplexity": 32.37790092916372, "lr": 0.0026291804804649314, "grad_norm": 0.167447, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:57.014692+00:00", "epoch": 0, "step": 19606, "train_loss": 3.4275619983673096, "perplexity": 30.80145713281068, "lr": 0.0026291804804649314, "grad_norm": 0.184226, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:57.320747+00:00", "epoch": 0, "step": 19607, "train_loss": 3.4460480213165283, "perplexity": 31.376149094793227, "lr": 0.0026291804804649314, "grad_norm": 0.200489, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:57.625157+00:00", "epoch": 0, "step": 19608, "train_loss": 3.6002824306488037, "perplexity": 36.608572366581406, "lr": 0.0026291804804649314, "grad_norm": 0.194026, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:57.929654+00:00", "epoch": 0, "step": 19609, "train_loss": 3.5402324199676514, "perplexity": 34.47493092210786, "lr": 0.0026291804804649314, "grad_norm": 0.180187, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:58.233536+00:00", "epoch": 0, "step": 19610, "train_loss": 3.5102574825286865, "perplexity": 33.4568812373734, "lr": 0.0026291804804649314, "grad_norm": 0.172605, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:58.537929+00:00", "epoch": 0, "step": 19611, "train_loss": 3.5642642974853516, "perplexity": 35.313463641083196, "lr": 0.0026291804804649314, "grad_norm": 0.186898, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:58.843001+00:00", "epoch": 0, "step": 19612, "train_loss": 3.570241689682007, "perplexity": 35.525178183148235, "lr": 0.0026291804804649314, "grad_norm": 0.160864, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:59.149214+00:00", "epoch": 0, "step": 19613, "train_loss": 3.5475380420684814, "perplexity": 34.72771398384206, "lr": 0.0026291804804649314, "grad_norm": 0.16528, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:59.453670+00:00", "epoch": 0, "step": 19614, "train_loss": 3.429149866104126, "perplexity": 30.85040462362516, "lr": 0.0026291804804649314, "grad_norm": 0.17388, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:40:59.756895+00:00", "epoch": 0, "step": 19615, "train_loss": 3.5509214401245117, "perplexity": 34.84541065889961, "lr": 0.0026291804804649314, "grad_norm": 0.174016, "tokens_per_sec": 108065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:00.061453+00:00", "epoch": 0, "step": 19616, "train_loss": 3.3902392387390137, "perplexity": 29.673050364417335, "lr": 0.0026291804804649314, "grad_norm": 0.171863, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:00.366720+00:00", "epoch": 0, "step": 19617, "train_loss": 3.4753530025482178, "perplexity": 32.30923176495192, "lr": 0.0026291804804649314, "grad_norm": 0.160605, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:00.672447+00:00", "epoch": 0, "step": 19618, "train_loss": 3.5391042232513428, "perplexity": 34.43605835032367, "lr": 0.0026291804804649314, "grad_norm": 0.183593, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:00.978955+00:00", "epoch": 0, "step": 19619, "train_loss": 3.470339775085449, "perplexity": 32.1476635644205, "lr": 0.0026291804804649314, "grad_norm": 0.187677, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:01.283609+00:00", "epoch": 0, "step": 19620, "train_loss": 3.4781100749969482, "perplexity": 32.39843356910378, "lr": 0.0026291804804649314, "grad_norm": 0.189774, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:01.588071+00:00", "epoch": 0, "step": 19621, "train_loss": 3.4027278423309326, "perplexity": 30.045948972642122, "lr": 0.0026291804804649314, "grad_norm": 0.158808, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:01.893153+00:00", "epoch": 0, "step": 19622, "train_loss": 3.463050365447998, "perplexity": 31.914178095289788, "lr": 0.0026291804804649314, "grad_norm": 0.157472, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:02.197919+00:00", "epoch": 0, "step": 19623, "train_loss": 3.376274585723877, "perplexity": 29.26155638465114, "lr": 0.0026291804804649314, "grad_norm": 0.178841, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:02.503094+00:00", "epoch": 0, "step": 19624, "train_loss": 3.478900909423828, "perplexity": 32.4240654997179, "lr": 0.0026291804804649314, "grad_norm": 0.166811, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:02.808208+00:00", "epoch": 0, "step": 19625, "train_loss": 3.2927439212799072, "perplexity": 26.916619504542656, "lr": 0.0026291804804649314, "grad_norm": 0.175669, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:03.113675+00:00", "epoch": 0, "step": 19626, "train_loss": 3.479851722717285, "perplexity": 32.454909393291274, "lr": 0.0026291804804649314, "grad_norm": 0.160926, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:03.417550+00:00", "epoch": 0, "step": 19627, "train_loss": 3.4995617866516113, "perplexity": 33.10094350473736, "lr": 0.0026291804804649314, "grad_norm": 0.185718, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:03.721658+00:00", "epoch": 0, "step": 19628, "train_loss": 3.40167498588562, "perplexity": 30.014331548835642, "lr": 0.0026291804804649314, "grad_norm": 0.16861, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:04.026851+00:00", "epoch": 0, "step": 19629, "train_loss": 3.5879127979278564, "perplexity": 36.15852695852861, "lr": 0.0026291804804649314, "grad_norm": 0.162769, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:04.333717+00:00", "epoch": 0, "step": 19630, "train_loss": 3.5599796772003174, "perplexity": 35.16248253775627, "lr": 0.0026291804804649314, "grad_norm": 0.201395, "tokens_per_sec": 106853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:04.638088+00:00", "epoch": 0, "step": 19631, "train_loss": 3.543036460876465, "perplexity": 34.57173569758882, "lr": 0.0026291804804649314, "grad_norm": 0.164807, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:04.942089+00:00", "epoch": 0, "step": 19632, "train_loss": 3.566718578338623, "perplexity": 35.40023924104928, "lr": 0.0026291804804649314, "grad_norm": 0.159374, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:05.246637+00:00", "epoch": 0, "step": 19633, "train_loss": 3.5130863189697266, "perplexity": 33.551659274800635, "lr": 0.0026291804804649314, "grad_norm": 0.15292, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:05.552929+00:00", "epoch": 0, "step": 19634, "train_loss": 3.4949333667755127, "perplexity": 32.948092442218304, "lr": 0.0026291804804649314, "grad_norm": 0.169519, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:05.857097+00:00", "epoch": 0, "step": 19635, "train_loss": 3.633394956588745, "perplexity": 37.84106757246708, "lr": 0.0026291804804649314, "grad_norm": 0.181404, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:06.162302+00:00", "epoch": 0, "step": 19636, "train_loss": 3.45804762840271, "perplexity": 31.75491855322194, "lr": 0.0026291804804649314, "grad_norm": 0.184369, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:06.466877+00:00", "epoch": 0, "step": 19637, "train_loss": 3.5176613330841064, "perplexity": 33.70551025639855, "lr": 0.0026291804804649314, "grad_norm": 0.198969, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:06.771451+00:00", "epoch": 0, "step": 19638, "train_loss": 3.4787135124206543, "perplexity": 32.41798989630525, "lr": 0.0026291804804649314, "grad_norm": 0.174791, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:07.076529+00:00", "epoch": 0, "step": 19639, "train_loss": 3.5595004558563232, "perplexity": 35.14563596255796, "lr": 0.0026291804804649314, "grad_norm": 0.166517, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:07.382700+00:00", "epoch": 0, "step": 19640, "train_loss": 3.3790946006774902, "perplexity": 29.344190871703994, "lr": 0.0026291804804649314, "grad_norm": 0.178124, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:07.688416+00:00", "epoch": 0, "step": 19641, "train_loss": 3.5434250831604004, "perplexity": 34.585173655451044, "lr": 0.0026291804804649314, "grad_norm": 0.173449, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:07.992201+00:00", "epoch": 0, "step": 19642, "train_loss": 3.4046335220336914, "perplexity": 30.10326152008384, "lr": 0.0026291804804649314, "grad_norm": 0.170422, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:08.296014+00:00", "epoch": 0, "step": 19643, "train_loss": 3.5947344303131104, "perplexity": 36.40603036587148, "lr": 0.0026291804804649314, "grad_norm": 0.206821, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:08.600805+00:00", "epoch": 0, "step": 19644, "train_loss": 3.5269925594329834, "perplexity": 34.021495976159045, "lr": 0.0026291804804649314, "grad_norm": 0.193494, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:08.907067+00:00", "epoch": 0, "step": 19645, "train_loss": 3.4409074783325195, "perplexity": 31.215272502483668, "lr": 0.0026291804804649314, "grad_norm": 0.186521, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:09.212387+00:00", "epoch": 0, "step": 19646, "train_loss": 3.4804086685180664, "perplexity": 32.47299005329907, "lr": 0.0026291804804649314, "grad_norm": 0.1852, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:09.517852+00:00", "epoch": 0, "step": 19647, "train_loss": 3.5148420333862305, "perplexity": 33.61061814901453, "lr": 0.0026291804804649314, "grad_norm": 0.157156, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:09.822704+00:00", "epoch": 0, "step": 19648, "train_loss": 3.50297212600708, "perplexity": 33.2140216629258, "lr": 0.0026291804804649314, "grad_norm": 0.159179, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:10.128042+00:00", "epoch": 0, "step": 19649, "train_loss": 3.5373315811157227, "perplexity": 34.375069613879425, "lr": 0.0026291804804649314, "grad_norm": 0.159882, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:10.433479+00:00", "epoch": 0, "step": 19650, "train_loss": 3.55590558052063, "perplexity": 35.019518606584306, "lr": 0.0026291804804649314, "grad_norm": 0.187998, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:10.738065+00:00", "epoch": 0, "step": 19651, "train_loss": 3.3973782062530518, "perplexity": 29.885643250970325, "lr": 0.0026291804804649314, "grad_norm": 0.173655, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:11.042816+00:00", "epoch": 0, "step": 19652, "train_loss": 3.5351033210754395, "perplexity": 34.298558295091475, "lr": 0.0026291804804649314, "grad_norm": 0.165259, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:11.348214+00:00", "epoch": 0, "step": 19653, "train_loss": 3.5086703300476074, "perplexity": 33.40382218285156, "lr": 0.0026291804804649314, "grad_norm": 0.174276, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:11.653161+00:00", "epoch": 0, "step": 19654, "train_loss": 3.418509006500244, "perplexity": 30.523870183927947, "lr": 0.0026291804804649314, "grad_norm": 0.153652, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:11.958983+00:00", "epoch": 0, "step": 19655, "train_loss": 3.4721879959106445, "perplexity": 32.20713448647095, "lr": 0.0026291804804649314, "grad_norm": 0.163494, "tokens_per_sec": 107147} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:12.263939+00:00", "epoch": 0, "step": 19656, "train_loss": 3.488637685775757, "perplexity": 32.74131335322647, "lr": 0.0026291804804649314, "grad_norm": 0.160897, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:12.568754+00:00", "epoch": 0, "step": 19657, "train_loss": 3.5287575721740723, "perplexity": 34.08159737429193, "lr": 0.0026291804804649314, "grad_norm": 0.163805, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:12.872591+00:00", "epoch": 0, "step": 19658, "train_loss": 3.4801318645477295, "perplexity": 32.46400264465619, "lr": 0.0026291804804649314, "grad_norm": 0.145054, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:13.176641+00:00", "epoch": 0, "step": 19659, "train_loss": 3.5440642833709717, "perplexity": 34.60728757259157, "lr": 0.0026291804804649314, "grad_norm": 0.177164, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:13.482388+00:00", "epoch": 0, "step": 19660, "train_loss": 3.496582508087158, "perplexity": 33.00247333116512, "lr": 0.0026291804804649314, "grad_norm": 0.181667, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:13.787559+00:00", "epoch": 0, "step": 19661, "train_loss": 3.535977602005005, "perplexity": 34.328557982686945, "lr": 0.0026291804804649314, "grad_norm": 0.190478, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:14.093784+00:00", "epoch": 0, "step": 19662, "train_loss": 3.5002942085266113, "perplexity": 33.12519624037624, "lr": 0.0026291804804649314, "grad_norm": 0.194449, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:14.399292+00:00", "epoch": 0, "step": 19663, "train_loss": 3.4623522758483887, "perplexity": 31.89190691402584, "lr": 0.0026291804804649314, "grad_norm": 0.190738, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:14.703975+00:00", "epoch": 0, "step": 19664, "train_loss": 3.485595703125, "perplexity": 32.641866181016724, "lr": 0.0026291804804649314, "grad_norm": 0.17696, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:15.008949+00:00", "epoch": 0, "step": 19665, "train_loss": 3.403804063796997, "perplexity": 30.078302474538585, "lr": 0.0026291804804649314, "grad_norm": 0.159964, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:15.315007+00:00", "epoch": 0, "step": 19666, "train_loss": 3.438045024871826, "perplexity": 31.12604799909947, "lr": 0.0026291804804649314, "grad_norm": 0.204319, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:15.620941+00:00", "epoch": 0, "step": 19667, "train_loss": 3.527853488922119, "perplexity": 34.05079869728787, "lr": 0.0026291804804649314, "grad_norm": 0.182151, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:15.927250+00:00", "epoch": 0, "step": 19668, "train_loss": 3.487949848175049, "perplexity": 32.718800390324354, "lr": 0.0026291804804649314, "grad_norm": 0.190115, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:16.231707+00:00", "epoch": 0, "step": 19669, "train_loss": 3.490358591079712, "perplexity": 32.797706562803775, "lr": 0.0026291804804649314, "grad_norm": 0.180381, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:16.536124+00:00", "epoch": 0, "step": 19670, "train_loss": 3.476080894470215, "perplexity": 32.33275795498266, "lr": 0.0026291804804649314, "grad_norm": 0.188925, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:16.840864+00:00", "epoch": 0, "step": 19671, "train_loss": 3.5233311653137207, "perplexity": 33.89715763560249, "lr": 0.0026291804804649314, "grad_norm": 0.177067, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:17.145501+00:00", "epoch": 0, "step": 19672, "train_loss": 3.532383918762207, "perplexity": 34.20541342291537, "lr": 0.0026291804804649314, "grad_norm": 0.160396, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:17.450439+00:00", "epoch": 0, "step": 19673, "train_loss": 3.571183204650879, "perplexity": 35.55864142077713, "lr": 0.0026291804804649314, "grad_norm": 0.168417, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:17.754790+00:00", "epoch": 0, "step": 19674, "train_loss": 3.555665969848633, "perplexity": 35.011128561410295, "lr": 0.0026291804804649314, "grad_norm": 0.196596, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:18.059921+00:00", "epoch": 0, "step": 19675, "train_loss": 3.5444881916046143, "perplexity": 34.62196099661052, "lr": 0.0026291804804649314, "grad_norm": 0.160987, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:18.364062+00:00", "epoch": 0, "step": 19676, "train_loss": 3.4632904529571533, "perplexity": 31.921841210687756, "lr": 0.0026291804804649314, "grad_norm": 0.158654, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:18.670508+00:00", "epoch": 0, "step": 19677, "train_loss": 3.5136640071868896, "perplexity": 33.57104727259971, "lr": 0.0026291804804649314, "grad_norm": 0.149635, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:18.975026+00:00", "epoch": 0, "step": 19678, "train_loss": 3.6153666973114014, "perplexity": 37.16497172243124, "lr": 0.0026291804804649314, "grad_norm": 0.156056, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:19.280295+00:00", "epoch": 0, "step": 19679, "train_loss": 3.5578291416168213, "perplexity": 35.08694561938212, "lr": 0.0026291804804649314, "grad_norm": 0.163021, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:19.585664+00:00", "epoch": 0, "step": 19680, "train_loss": 3.4917681217193604, "perplexity": 32.84396853128591, "lr": 0.0026291804804649314, "grad_norm": 0.169435, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:19.889701+00:00", "epoch": 0, "step": 19681, "train_loss": 3.4837803840637207, "perplexity": 32.58266453034273, "lr": 0.0026291804804649314, "grad_norm": 0.17072, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:20.195809+00:00", "epoch": 0, "step": 19682, "train_loss": 3.5483057498931885, "perplexity": 34.754384958049506, "lr": 0.0026291804804649314, "grad_norm": 0.173562, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:20.502449+00:00", "epoch": 0, "step": 19683, "train_loss": 3.4949707984924316, "perplexity": 32.94932576897025, "lr": 0.0026291804804649314, "grad_norm": 0.158713, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:20.808508+00:00", "epoch": 0, "step": 19684, "train_loss": 3.505638360977173, "perplexity": 33.30269621003929, "lr": 0.0026291804804649314, "grad_norm": 0.177685, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:21.114143+00:00", "epoch": 0, "step": 19685, "train_loss": 3.5031046867370605, "perplexity": 33.21842482972008, "lr": 0.0026291804804649314, "grad_norm": 0.201878, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:21.418023+00:00", "epoch": 0, "step": 19686, "train_loss": 3.531069278717041, "perplexity": 34.160475161959006, "lr": 0.0026291804804649314, "grad_norm": 0.196677, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:21.723273+00:00", "epoch": 0, "step": 19687, "train_loss": 3.5048165321350098, "perplexity": 33.27533833705642, "lr": 0.0026291804804649314, "grad_norm": 0.174251, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:22.028952+00:00", "epoch": 0, "step": 19688, "train_loss": 3.516263008117676, "perplexity": 33.658411936953044, "lr": 0.0026291804804649314, "grad_norm": 0.16572, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:22.334332+00:00", "epoch": 0, "step": 19689, "train_loss": 3.5214078426361084, "perplexity": 33.83202511911636, "lr": 0.0026291804804649314, "grad_norm": 0.221674, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:22.639310+00:00", "epoch": 0, "step": 19690, "train_loss": 3.4384472370147705, "perplexity": 31.138569791606265, "lr": 0.0026291804804649314, "grad_norm": 0.238834, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:22.944744+00:00", "epoch": 0, "step": 19691, "train_loss": 3.516993522644043, "perplexity": 33.6830088789253, "lr": 0.0026291804804649314, "grad_norm": 0.182191, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:23.248802+00:00", "epoch": 0, "step": 19692, "train_loss": 3.4880483150482178, "perplexity": 32.72202226691409, "lr": 0.0026291804804649314, "grad_norm": 0.191647, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:23.554492+00:00", "epoch": 0, "step": 19693, "train_loss": 3.5073790550231934, "perplexity": 33.36071649818339, "lr": 0.0026291804804649314, "grad_norm": 0.215537, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:23.859458+00:00", "epoch": 0, "step": 19694, "train_loss": 3.64262318611145, "perplexity": 38.191889874098734, "lr": 0.0026291804804649314, "grad_norm": 0.19324, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:24.164361+00:00", "epoch": 0, "step": 19695, "train_loss": 3.5033369064331055, "perplexity": 33.22613969797464, "lr": 0.0026291804804649314, "grad_norm": 0.172437, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:24.469158+00:00", "epoch": 0, "step": 19696, "train_loss": 3.5034117698669434, "perplexity": 33.22862721399644, "lr": 0.0026291804804649314, "grad_norm": 0.170566, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:24.773709+00:00", "epoch": 0, "step": 19697, "train_loss": 3.496464967727661, "perplexity": 32.99859443655328, "lr": 0.0026291804804649314, "grad_norm": 0.198098, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:25.078655+00:00", "epoch": 0, "step": 19698, "train_loss": 3.4512197971343994, "perplexity": 31.538839841346455, "lr": 0.0026291804804649314, "grad_norm": 0.180171, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:25.384190+00:00", "epoch": 0, "step": 19699, "train_loss": 3.538745403289795, "perplexity": 34.423704221777065, "lr": 0.0026291804804649314, "grad_norm": 0.184189, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:25.689709+00:00", "epoch": 0, "step": 19700, "train_loss": 3.4108264446258545, "perplexity": 30.29026714645946, "lr": 0.0026291804804649314, "grad_norm": 0.206159, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:25.994775+00:00", "epoch": 0, "step": 19701, "train_loss": 3.456895351409912, "perplexity": 31.718349164243904, "lr": 0.0026291804804649314, "grad_norm": 0.177421, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:26.298957+00:00", "epoch": 0, "step": 19702, "train_loss": 3.460449457168579, "perplexity": 31.831280097027744, "lr": 0.0026291804804649314, "grad_norm": 0.180295, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:26.604939+00:00", "epoch": 0, "step": 19703, "train_loss": 3.467522621154785, "perplexity": 32.0572260955957, "lr": 0.0026291804804649314, "grad_norm": 0.157011, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:26.909822+00:00", "epoch": 0, "step": 19704, "train_loss": 3.3990702629089355, "perplexity": 29.93625425881967, "lr": 0.0026291804804649314, "grad_norm": 0.170388, "tokens_per_sec": 107477} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:27.214560+00:00", "epoch": 0, "step": 19705, "train_loss": 3.4094009399414062, "perplexity": 30.24711899000522, "lr": 0.0026291804804649314, "grad_norm": 0.164783, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:27.519683+00:00", "epoch": 0, "step": 19706, "train_loss": 3.472625255584717, "perplexity": 32.22122044698612, "lr": 0.0026291804804649314, "grad_norm": 0.177373, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:27.823811+00:00", "epoch": 0, "step": 19707, "train_loss": 3.4695301055908203, "perplexity": 32.12164511650197, "lr": 0.0026291804804649314, "grad_norm": 0.173551, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:28.129413+00:00", "epoch": 0, "step": 19708, "train_loss": 3.5788280963897705, "perplexity": 35.83152513902089, "lr": 0.0026291804804649314, "grad_norm": 0.168394, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:28.434428+00:00", "epoch": 0, "step": 19709, "train_loss": 3.3806138038635254, "perplexity": 29.388804540000805, "lr": 0.0026291804804649314, "grad_norm": 0.176248, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:28.739759+00:00", "epoch": 0, "step": 19710, "train_loss": 3.490861177444458, "perplexity": 32.81419438584765, "lr": 0.0026291804804649314, "grad_norm": 0.173804, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:29.044752+00:00", "epoch": 0, "step": 19711, "train_loss": 3.4791417121887207, "perplexity": 32.43187424448522, "lr": 0.0026291804804649314, "grad_norm": 0.159064, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:29.349801+00:00", "epoch": 0, "step": 19712, "train_loss": 3.492610216140747, "perplexity": 32.87163790243877, "lr": 0.0026291804804649314, "grad_norm": 0.154519, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:29.654550+00:00", "epoch": 0, "step": 19713, "train_loss": 3.50136137008667, "perplexity": 33.16056504523125, "lr": 0.0026291804804649314, "grad_norm": 0.170456, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:29.960736+00:00", "epoch": 0, "step": 19714, "train_loss": 3.5070509910583496, "perplexity": 33.34977384430233, "lr": 0.0026291804804649314, "grad_norm": 0.170228, "tokens_per_sec": 107022} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:30.264849+00:00", "epoch": 0, "step": 19715, "train_loss": 3.4876582622528076, "perplexity": 32.70926143951739, "lr": 0.0026291804804649314, "grad_norm": 0.15358, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:30.570299+00:00", "epoch": 0, "step": 19716, "train_loss": 3.576423168182373, "perplexity": 35.74545642951644, "lr": 0.0026291804804649314, "grad_norm": 0.173874, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:30.874718+00:00", "epoch": 0, "step": 19717, "train_loss": 3.5027613639831543, "perplexity": 33.20702214613948, "lr": 0.0026291804804649314, "grad_norm": 0.156696, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:31.180324+00:00", "epoch": 0, "step": 19718, "train_loss": 3.616788864135742, "perplexity": 37.21786411423005, "lr": 0.0026291804804649314, "grad_norm": 0.162508, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:31.486102+00:00", "epoch": 0, "step": 19719, "train_loss": 3.5259618759155273, "perplexity": 33.98644864547285, "lr": 0.0026291804804649314, "grad_norm": 0.170789, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:31.791309+00:00", "epoch": 0, "step": 19720, "train_loss": 3.436077117919922, "perplexity": 31.06485506360659, "lr": 0.0026291804804649314, "grad_norm": 0.179808, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:32.097443+00:00", "epoch": 0, "step": 19721, "train_loss": 3.502143144607544, "perplexity": 33.186499266117096, "lr": 0.0026291804804649314, "grad_norm": 0.165266, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:32.402001+00:00", "epoch": 0, "step": 19722, "train_loss": 3.520031213760376, "perplexity": 33.7854830193623, "lr": 0.0026291804804649314, "grad_norm": 0.203405, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:32.708084+00:00", "epoch": 0, "step": 19723, "train_loss": 3.540083169937134, "perplexity": 34.46978592157094, "lr": 0.0026291804804649314, "grad_norm": 0.190556, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:33.013758+00:00", "epoch": 0, "step": 19724, "train_loss": 3.5668516159057617, "perplexity": 35.40494911604224, "lr": 0.0026291804804649314, "grad_norm": 0.179857, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:33.318982+00:00", "epoch": 0, "step": 19725, "train_loss": 3.5213918685913086, "perplexity": 33.83148468914788, "lr": 0.0026291804804649314, "grad_norm": 0.166794, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:33.624632+00:00", "epoch": 0, "step": 19726, "train_loss": 3.5223405361175537, "perplexity": 33.86359474851425, "lr": 0.0026291804804649314, "grad_norm": 0.173855, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:33.929566+00:00", "epoch": 0, "step": 19727, "train_loss": 3.489161968231201, "perplexity": 32.75848355000893, "lr": 0.0026291804804649314, "grad_norm": 0.150429, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:34.234787+00:00", "epoch": 0, "step": 19728, "train_loss": 3.531127452850342, "perplexity": 34.162462475799266, "lr": 0.0026291804804649314, "grad_norm": 0.172381, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:34.538683+00:00", "epoch": 0, "step": 19729, "train_loss": 3.4512722492218018, "perplexity": 31.540494162716307, "lr": 0.0026291804804649314, "grad_norm": 0.179246, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:34.842876+00:00", "epoch": 0, "step": 19730, "train_loss": 3.5798206329345703, "perplexity": 35.86710689235149, "lr": 0.0026291804804649314, "grad_norm": 0.172645, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:35.147785+00:00", "epoch": 0, "step": 19731, "train_loss": 3.4815616607666016, "perplexity": 32.510452752067714, "lr": 0.0026291804804649314, "grad_norm": 0.177031, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:35.453556+00:00", "epoch": 0, "step": 19732, "train_loss": 3.6218111515045166, "perplexity": 37.405253089824676, "lr": 0.0026291804804649314, "grad_norm": 0.193993, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:35.760089+00:00", "epoch": 0, "step": 19733, "train_loss": 3.499824285507202, "perplexity": 33.10963360504712, "lr": 0.0026291804804649314, "grad_norm": 0.18894, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:36.065530+00:00", "epoch": 0, "step": 19734, "train_loss": 3.5522234439849854, "perplexity": 34.89080906613469, "lr": 0.0026291804804649314, "grad_norm": 0.199521, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:36.370533+00:00", "epoch": 0, "step": 19735, "train_loss": 3.498103380203247, "perplexity": 33.05270406019184, "lr": 0.0026291804804649314, "grad_norm": 0.160183, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:36.675757+00:00", "epoch": 0, "step": 19736, "train_loss": 3.496309280395508, "perplexity": 32.99345737331886, "lr": 0.0026291804804649314, "grad_norm": 0.175667, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:36.980826+00:00", "epoch": 0, "step": 19737, "train_loss": 3.5103542804718018, "perplexity": 33.46011995140813, "lr": 0.0026291804804649314, "grad_norm": 0.198814, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:37.285661+00:00", "epoch": 0, "step": 19738, "train_loss": 3.545743942260742, "perplexity": 34.66546485613171, "lr": 0.0026291804804649314, "grad_norm": 0.213832, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:37.590194+00:00", "epoch": 0, "step": 19739, "train_loss": 3.5335066318511963, "perplexity": 34.24383785404934, "lr": 0.0026291804804649314, "grad_norm": 0.156279, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:37.895165+00:00", "epoch": 0, "step": 19740, "train_loss": 3.5124247074127197, "perplexity": 33.529468450929905, "lr": 0.0026291804804649314, "grad_norm": 0.236885, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:38.200206+00:00", "epoch": 0, "step": 19741, "train_loss": 3.500486135482788, "perplexity": 33.13155446860159, "lr": 0.0026291804804649314, "grad_norm": 0.225165, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:38.505507+00:00", "epoch": 0, "step": 19742, "train_loss": 3.5407800674438477, "perplexity": 34.49381620178527, "lr": 0.0026291804804649314, "grad_norm": 0.189412, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:38.810800+00:00", "epoch": 0, "step": 19743, "train_loss": 3.5399203300476074, "perplexity": 34.46417332242939, "lr": 0.0026291804804649314, "grad_norm": 0.193265, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:39.115592+00:00", "epoch": 0, "step": 19744, "train_loss": 3.504513740539551, "perplexity": 33.265264369505225, "lr": 0.0026291804804649314, "grad_norm": 0.199404, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:39.420336+00:00", "epoch": 0, "step": 19745, "train_loss": 3.440885066986084, "perplexity": 31.2145729340367, "lr": 0.0026291804804649314, "grad_norm": 0.190186, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:39.725187+00:00", "epoch": 0, "step": 19746, "train_loss": 3.4537930488586426, "perplexity": 31.62010172414838, "lr": 0.0026291804804649314, "grad_norm": 0.156693, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:40.029876+00:00", "epoch": 0, "step": 19747, "train_loss": 3.4368233680725098, "perplexity": 31.088045868432523, "lr": 0.0026291804804649314, "grad_norm": 0.164238, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:40.335751+00:00", "epoch": 0, "step": 19748, "train_loss": 3.5570456981658936, "perplexity": 35.05946774670871, "lr": 0.0026291804804649314, "grad_norm": 0.154936, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:40.641975+00:00", "epoch": 0, "step": 19749, "train_loss": 3.4112279415130615, "perplexity": 30.302431036149958, "lr": 0.0026291804804649314, "grad_norm": 0.179179, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:40.946387+00:00", "epoch": 0, "step": 19750, "train_loss": 3.4233577251434326, "perplexity": 30.672231232176504, "lr": 0.0026291804804649314, "grad_norm": 0.166887, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:41.251124+00:00", "epoch": 0, "step": 19751, "train_loss": 3.478739023208618, "perplexity": 32.41881691532062, "lr": 0.0026291804804649314, "grad_norm": 0.156114, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:41.555388+00:00", "epoch": 0, "step": 19752, "train_loss": 3.452646255493164, "perplexity": 31.583860785681075, "lr": 0.0026291804804649314, "grad_norm": 0.163202, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:41.859999+00:00", "epoch": 0, "step": 19753, "train_loss": 3.478363513946533, "perplexity": 32.406645634660414, "lr": 0.0026291804804649314, "grad_norm": 0.149142, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:42.164921+00:00", "epoch": 0, "step": 19754, "train_loss": 3.506514310836792, "perplexity": 33.33188048222291, "lr": 0.0026291804804649314, "grad_norm": 0.144263, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:42.469100+00:00", "epoch": 0, "step": 19755, "train_loss": 3.490065574645996, "perplexity": 32.78809770363826, "lr": 0.0026291804804649314, "grad_norm": 0.162151, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:42.773381+00:00", "epoch": 0, "step": 19756, "train_loss": 3.5889456272125244, "perplexity": 36.1958918365104, "lr": 0.0026291804804649314, "grad_norm": 0.150934, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:43.078202+00:00", "epoch": 0, "step": 19757, "train_loss": 3.499770402908325, "perplexity": 33.107849620003925, "lr": 0.0026291804804649314, "grad_norm": 0.140311, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:43.382797+00:00", "epoch": 0, "step": 19758, "train_loss": 3.5689873695373535, "perplexity": 35.48064617104923, "lr": 0.0026291804804649314, "grad_norm": 0.161655, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:43.688512+00:00", "epoch": 0, "step": 19759, "train_loss": 3.581599712371826, "perplexity": 35.93097412028786, "lr": 0.0026291804804649314, "grad_norm": 0.149789, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:43.993918+00:00", "epoch": 0, "step": 19760, "train_loss": 3.4691734313964844, "perplexity": 32.11019019756773, "lr": 0.0026291804804649314, "grad_norm": 0.156325, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:44.298065+00:00", "epoch": 0, "step": 19761, "train_loss": 3.3681461811065674, "perplexity": 29.024670670362188, "lr": 0.0026291804804649314, "grad_norm": 0.181315, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:44.601801+00:00", "epoch": 0, "step": 19762, "train_loss": 3.5024616718292236, "perplexity": 33.19707175324869, "lr": 0.0026291804804649314, "grad_norm": 0.171334, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:44.906656+00:00", "epoch": 0, "step": 19763, "train_loss": 3.4339258670806885, "perplexity": 30.998098598775915, "lr": 0.0026291804804649314, "grad_norm": 0.156436, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:45.212099+00:00", "epoch": 0, "step": 19764, "train_loss": 3.5061862468719482, "perplexity": 33.32094728684798, "lr": 0.0026291804804649314, "grad_norm": 0.156856, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:45.517720+00:00", "epoch": 0, "step": 19765, "train_loss": 3.517662525177002, "perplexity": 33.70555043652181, "lr": 0.0026291804804649314, "grad_norm": 0.177254, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:45.822331+00:00", "epoch": 0, "step": 19766, "train_loss": 3.4679510593414307, "perplexity": 32.070963578033876, "lr": 0.0026291804804649314, "grad_norm": 0.181062, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:46.126915+00:00", "epoch": 0, "step": 19767, "train_loss": 3.4460489749908447, "perplexity": 31.376179017435035, "lr": 0.0026291804804649314, "grad_norm": 0.194802, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:46.431688+00:00", "epoch": 0, "step": 19768, "train_loss": 3.437439203262329, "perplexity": 31.107196877388382, "lr": 0.0026291804804649314, "grad_norm": 0.203516, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:46.736835+00:00", "epoch": 0, "step": 19769, "train_loss": 3.604184627532959, "perplexity": 36.75170530843811, "lr": 0.0026291804804649314, "grad_norm": 0.225606, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:47.041766+00:00", "epoch": 0, "step": 19770, "train_loss": 3.48283052444458, "perplexity": 32.551730266951004, "lr": 0.0026291804804649314, "grad_norm": 0.249012, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:47.347014+00:00", "epoch": 0, "step": 19771, "train_loss": 3.6435744762420654, "perplexity": 38.22823872841753, "lr": 0.0026291804804649314, "grad_norm": 0.213744, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:47.651455+00:00", "epoch": 0, "step": 19772, "train_loss": 3.4982728958129883, "perplexity": 33.05830748439472, "lr": 0.0026291804804649314, "grad_norm": 0.194302, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:47.956771+00:00", "epoch": 0, "step": 19773, "train_loss": 3.472527027130127, "perplexity": 32.21805556174006, "lr": 0.0026291804804649314, "grad_norm": 0.181282, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:48.262674+00:00", "epoch": 0, "step": 19774, "train_loss": 3.527268409729004, "perplexity": 34.03088211041957, "lr": 0.0026291804804649314, "grad_norm": 0.21116, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:48.568514+00:00", "epoch": 0, "step": 19775, "train_loss": 3.509876012802124, "perplexity": 33.44412088403514, "lr": 0.0026291804804649314, "grad_norm": 0.165987, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:48.873326+00:00", "epoch": 0, "step": 19776, "train_loss": 3.4435036182403564, "perplexity": 31.296417002823418, "lr": 0.0026291804804649314, "grad_norm": 0.179149, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:49.177207+00:00", "epoch": 0, "step": 19777, "train_loss": 3.4529056549072266, "perplexity": 31.592054683362896, "lr": 0.0026291804804649314, "grad_norm": 0.161621, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:49.481554+00:00", "epoch": 0, "step": 19778, "train_loss": 3.5317342281341553, "perplexity": 34.18319770383954, "lr": 0.0026291804804649314, "grad_norm": 0.176672, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:49.786083+00:00", "epoch": 0, "step": 19779, "train_loss": 3.538893461227417, "perplexity": 34.42880130175171, "lr": 0.0026291804804649314, "grad_norm": 0.186836, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:50.091715+00:00", "epoch": 0, "step": 19780, "train_loss": 3.543208599090576, "perplexity": 34.57768732666723, "lr": 0.0026291804804649314, "grad_norm": 0.166087, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:50.396599+00:00", "epoch": 0, "step": 19781, "train_loss": 3.660755157470703, "perplexity": 38.89070038810224, "lr": 0.0026291804804649314, "grad_norm": 0.169852, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:50.702558+00:00", "epoch": 0, "step": 19782, "train_loss": 3.525707721710205, "perplexity": 33.9778119441991, "lr": 0.0026291804804649314, "grad_norm": 0.16125, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:51.007287+00:00", "epoch": 0, "step": 19783, "train_loss": 3.4723217487335205, "perplexity": 32.211442569727986, "lr": 0.0026291804804649314, "grad_norm": 0.168712, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:51.312177+00:00", "epoch": 0, "step": 19784, "train_loss": 3.528388738632202, "perplexity": 34.06902925593406, "lr": 0.0026291804804649314, "grad_norm": 0.168326, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:51.618058+00:00", "epoch": 0, "step": 19785, "train_loss": 3.501556158065796, "perplexity": 33.16702495381893, "lr": 0.0026291804804649314, "grad_norm": 0.157969, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:51.923254+00:00", "epoch": 0, "step": 19786, "train_loss": 3.505612850189209, "perplexity": 33.30184664285426, "lr": 0.0026291804804649314, "grad_norm": 0.172431, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:52.228703+00:00", "epoch": 0, "step": 19787, "train_loss": 3.563469409942627, "perplexity": 35.285404562128015, "lr": 0.0026291804804649314, "grad_norm": 0.189432, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:52.533737+00:00", "epoch": 0, "step": 19788, "train_loss": 3.401923894882202, "perplexity": 30.02180331584098, "lr": 0.0026291804804649314, "grad_norm": 0.177642, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:52.838297+00:00", "epoch": 0, "step": 19789, "train_loss": 3.4428255558013916, "perplexity": 31.27520327090981, "lr": 0.0026291804804649314, "grad_norm": 0.177521, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:53.142994+00:00", "epoch": 0, "step": 19790, "train_loss": 3.426645278930664, "perplexity": 30.77323377680605, "lr": 0.0026291804804649314, "grad_norm": 0.180877, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:53.447796+00:00", "epoch": 0, "step": 19791, "train_loss": 3.623966932296753, "perplexity": 37.48597759687214, "lr": 0.0026291804804649314, "grad_norm": 0.168511, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:53.752196+00:00", "epoch": 0, "step": 19792, "train_loss": 3.508765459060669, "perplexity": 33.407000006637524, "lr": 0.0026291804804649314, "grad_norm": 0.17031, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:54.056997+00:00", "epoch": 0, "step": 19793, "train_loss": 3.518254280090332, "perplexity": 33.725501764164996, "lr": 0.0026291804804649314, "grad_norm": 0.167381, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:54.361472+00:00", "epoch": 0, "step": 19794, "train_loss": 3.5760958194732666, "perplexity": 35.73375711547996, "lr": 0.0026291804804649314, "grad_norm": 0.15668, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:54.666381+00:00", "epoch": 0, "step": 19795, "train_loss": 3.331289052963257, "perplexity": 27.974378927352685, "lr": 0.0026291804804649314, "grad_norm": 0.176518, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:54.971433+00:00", "epoch": 0, "step": 19796, "train_loss": 3.5471901893615723, "perplexity": 34.715635955337255, "lr": 0.0026291804804649314, "grad_norm": 0.161798, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:55.276541+00:00", "epoch": 0, "step": 19797, "train_loss": 3.59578013420105, "perplexity": 36.44412020524515, "lr": 0.0026291804804649314, "grad_norm": 0.175057, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:55.581403+00:00", "epoch": 0, "step": 19798, "train_loss": 3.484297513961792, "perplexity": 32.599518357763635, "lr": 0.0026291804804649314, "grad_norm": 0.18248, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:55.886404+00:00", "epoch": 0, "step": 19799, "train_loss": 3.486802339553833, "perplexity": 32.68127681822937, "lr": 0.0026291804804649314, "grad_norm": 0.168139, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:56.190767+00:00", "epoch": 0, "step": 19800, "train_loss": 3.377635955810547, "perplexity": 29.301419320149996, "lr": 0.0026291804804649314, "grad_norm": 0.15988, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:56.496218+00:00", "epoch": 0, "step": 19801, "train_loss": 3.486204147338867, "perplexity": 32.66173297893061, "lr": 0.0026291804804649314, "grad_norm": 0.175225, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:56.802037+00:00", "epoch": 0, "step": 19802, "train_loss": 3.539689064025879, "perplexity": 34.45620385174236, "lr": 0.0026291804804649314, "grad_norm": 0.179438, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:57.107574+00:00", "epoch": 0, "step": 19803, "train_loss": 3.592298984527588, "perplexity": 36.317473334346154, "lr": 0.0026291804804649314, "grad_norm": 0.171353, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:57.411763+00:00", "epoch": 0, "step": 19804, "train_loss": 3.4404406547546387, "perplexity": 31.200703878042795, "lr": 0.0026291804804649314, "grad_norm": 0.169195, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:57.718216+00:00", "epoch": 0, "step": 19805, "train_loss": 3.5608205795288086, "perplexity": 35.192063186672996, "lr": 0.0026291804804649314, "grad_norm": 0.16896, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:58.022730+00:00", "epoch": 0, "step": 19806, "train_loss": 3.4655747413635254, "perplexity": 31.994843249587376, "lr": 0.0026291804804649314, "grad_norm": 0.168257, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:58.328235+00:00", "epoch": 0, "step": 19807, "train_loss": 3.509552001953125, "perplexity": 33.43328638137646, "lr": 0.0026291804804649314, "grad_norm": 0.168654, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:58.632929+00:00", "epoch": 0, "step": 19808, "train_loss": 3.4101407527923584, "perplexity": 30.26950447685272, "lr": 0.0026291804804649314, "grad_norm": 0.187537, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:58.937608+00:00", "epoch": 0, "step": 19809, "train_loss": 3.4545891284942627, "perplexity": 31.645283865372196, "lr": 0.0026291804804649314, "grad_norm": 0.163347, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:59.242601+00:00", "epoch": 0, "step": 19810, "train_loss": 3.545161247253418, "perplexity": 34.64527134673339, "lr": 0.0026291804804649314, "grad_norm": 0.175046, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:59.547121+00:00", "epoch": 0, "step": 19811, "train_loss": 3.4289259910583496, "perplexity": 30.843498760931833, "lr": 0.0026291804804649314, "grad_norm": 0.192064, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:41:59.852431+00:00", "epoch": 0, "step": 19812, "train_loss": 3.4949073791503906, "perplexity": 32.94723621066919, "lr": 0.0026291804804649314, "grad_norm": 0.181813, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:00.157502+00:00", "epoch": 0, "step": 19813, "train_loss": 3.535064458847046, "perplexity": 34.2972254025852, "lr": 0.0026291804804649314, "grad_norm": 0.192006, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:00.461384+00:00", "epoch": 0, "step": 19814, "train_loss": 3.4748542308807373, "perplexity": 32.29312085371653, "lr": 0.0026291804804649314, "grad_norm": 0.184818, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:00.765938+00:00", "epoch": 0, "step": 19815, "train_loss": 3.494549036026001, "perplexity": 32.935431910226725, "lr": 0.0026291804804649314, "grad_norm": 0.173173, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:01.071062+00:00", "epoch": 0, "step": 19816, "train_loss": 3.3962368965148926, "perplexity": 29.851553932232925, "lr": 0.0026291804804649314, "grad_norm": 0.167304, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:01.376037+00:00", "epoch": 0, "step": 19817, "train_loss": 3.5215442180633545, "perplexity": 33.83663929061924, "lr": 0.0026291804804649314, "grad_norm": 0.181375, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:01.680708+00:00", "epoch": 0, "step": 19818, "train_loss": 3.6220672130584717, "perplexity": 37.41483236344644, "lr": 0.0026291804804649314, "grad_norm": 0.18939, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:01.985939+00:00", "epoch": 0, "step": 19819, "train_loss": 3.538299322128296, "perplexity": 34.408351880264455, "lr": 0.0026291804804649314, "grad_norm": 0.248167, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:02.290805+00:00", "epoch": 0, "step": 19820, "train_loss": 3.3979763984680176, "perplexity": 29.903525958217177, "lr": 0.0026291804804649314, "grad_norm": 0.239836, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:02.595400+00:00", "epoch": 0, "step": 19821, "train_loss": 3.511814594268799, "perplexity": 33.50901792072232, "lr": 0.0026291804804649314, "grad_norm": 0.187959, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:02.901121+00:00", "epoch": 0, "step": 19822, "train_loss": 3.563567876815796, "perplexity": 35.28887917664816, "lr": 0.0026291804804649314, "grad_norm": 0.204993, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:03.206028+00:00", "epoch": 0, "step": 19823, "train_loss": 3.472360372543335, "perplexity": 32.21268672238645, "lr": 0.0026291804804649314, "grad_norm": 0.159058, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:03.510591+00:00", "epoch": 0, "step": 19824, "train_loss": 3.491501569747925, "perplexity": 32.835215073401685, "lr": 0.0026291804804649314, "grad_norm": 0.187111, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:03.815647+00:00", "epoch": 0, "step": 19825, "train_loss": 3.533168315887451, "perplexity": 34.232254576557736, "lr": 0.0026291804804649314, "grad_norm": 0.166576, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:04.119616+00:00", "epoch": 0, "step": 19826, "train_loss": 3.421194076538086, "perplexity": 30.605939044209705, "lr": 0.0026291804804649314, "grad_norm": 0.155597, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:04.425348+00:00", "epoch": 0, "step": 19827, "train_loss": 3.589395523071289, "perplexity": 36.21217988203932, "lr": 0.0026291804804649314, "grad_norm": 0.166143, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:04.731888+00:00", "epoch": 0, "step": 19828, "train_loss": 3.46958589553833, "perplexity": 32.123437231387484, "lr": 0.0026291804804649314, "grad_norm": 0.152513, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:05.036541+00:00", "epoch": 0, "step": 19829, "train_loss": 3.4676079750061035, "perplexity": 32.059962420081796, "lr": 0.0026291804804649314, "grad_norm": 0.178913, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:05.340967+00:00", "epoch": 0, "step": 19830, "train_loss": 3.455493927001953, "perplexity": 31.673929428266057, "lr": 0.0026291804804649314, "grad_norm": 0.15402, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:05.645324+00:00", "epoch": 0, "step": 19831, "train_loss": 3.4833271503448486, "perplexity": 32.56790031419992, "lr": 0.0026291804804649314, "grad_norm": 0.193814, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:05.950485+00:00", "epoch": 0, "step": 19832, "train_loss": 3.556584119796753, "perplexity": 35.04328878897896, "lr": 0.0026291804804649314, "grad_norm": 0.191873, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:06.256209+00:00", "epoch": 0, "step": 19833, "train_loss": 3.6105170249938965, "perplexity": 36.985170129692406, "lr": 0.0026291804804649314, "grad_norm": 0.157828, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:06.561724+00:00", "epoch": 0, "step": 19834, "train_loss": 3.671006441116333, "perplexity": 39.2914304785209, "lr": 0.0026291804804649314, "grad_norm": 0.165027, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:06.866392+00:00", "epoch": 0, "step": 19835, "train_loss": 3.469388484954834, "perplexity": 32.117096350798796, "lr": 0.0026291804804649314, "grad_norm": 0.159227, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:07.171561+00:00", "epoch": 0, "step": 19836, "train_loss": 3.581833839416504, "perplexity": 35.93938751793456, "lr": 0.0026291804804649314, "grad_norm": 0.159786, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:07.476823+00:00", "epoch": 0, "step": 19837, "train_loss": 3.5313944816589355, "perplexity": 34.17158605552791, "lr": 0.0026291804804649314, "grad_norm": 0.160037, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:07.780884+00:00", "epoch": 0, "step": 19838, "train_loss": 3.479907989501953, "perplexity": 32.45673557806582, "lr": 0.0026291804804649314, "grad_norm": 0.15953, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:08.086243+00:00", "epoch": 0, "step": 19839, "train_loss": 3.481851100921631, "perplexity": 32.519863944475134, "lr": 0.0026291804804649314, "grad_norm": 0.173027, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:08.391509+00:00", "epoch": 0, "step": 19840, "train_loss": 3.5248312950134277, "perplexity": 33.948045928483864, "lr": 0.0026291804804649314, "grad_norm": 0.170252, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:08.696144+00:00", "epoch": 0, "step": 19841, "train_loss": 3.49210524559021, "perplexity": 32.855042883700634, "lr": 0.0026291804804649314, "grad_norm": 0.159662, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:09.000093+00:00", "epoch": 0, "step": 19842, "train_loss": 3.586106538772583, "perplexity": 36.09327423757186, "lr": 0.0026291804804649314, "grad_norm": 0.155221, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:09.304527+00:00", "epoch": 0, "step": 19843, "train_loss": 3.5641589164733887, "perplexity": 35.30974246862269, "lr": 0.0026291804804649314, "grad_norm": 0.147453, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:09.610418+00:00", "epoch": 0, "step": 19844, "train_loss": 3.5182435512542725, "perplexity": 33.72513993072657, "lr": 0.0026291804804649314, "grad_norm": 0.165298, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:09.915898+00:00", "epoch": 0, "step": 19845, "train_loss": 3.4703776836395264, "perplexity": 32.14888225896251, "lr": 0.0026291804804649314, "grad_norm": 0.173654, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:10.220724+00:00", "epoch": 0, "step": 19846, "train_loss": 3.4840164184570312, "perplexity": 32.59035606749554, "lr": 0.0026291804804649314, "grad_norm": 0.152194, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:10.525266+00:00", "epoch": 0, "step": 19847, "train_loss": 3.451263904571533, "perplexity": 31.54023096942135, "lr": 0.0026291804804649314, "grad_norm": 0.186115, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:10.829678+00:00", "epoch": 0, "step": 19848, "train_loss": 3.473925828933716, "perplexity": 32.26315377034228, "lr": 0.0026291804804649314, "grad_norm": 0.18069, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:11.135017+00:00", "epoch": 0, "step": 19849, "train_loss": 3.5327982902526855, "perplexity": 34.219590108062086, "lr": 0.0026291804804649314, "grad_norm": 0.157139, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:11.441071+00:00", "epoch": 0, "step": 19850, "train_loss": 3.5113003253936768, "perplexity": 33.49178970611676, "lr": 0.0026291804804649314, "grad_norm": 0.2061, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:11.746412+00:00", "epoch": 0, "step": 19851, "train_loss": 3.479771375656128, "perplexity": 32.45230184145736, "lr": 0.0026291804804649314, "grad_norm": 0.163125, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:12.051477+00:00", "epoch": 0, "step": 19852, "train_loss": 3.5364742279052734, "perplexity": 34.34561066775206, "lr": 0.0026291804804649314, "grad_norm": 0.171355, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:12.356320+00:00", "epoch": 0, "step": 19853, "train_loss": 3.5592706203460693, "perplexity": 35.13755917558509, "lr": 0.0026291804804649314, "grad_norm": 0.186832, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:12.660462+00:00", "epoch": 0, "step": 19854, "train_loss": 3.461768388748169, "perplexity": 31.873291076279973, "lr": 0.0026291804804649314, "grad_norm": 0.196426, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:12.965847+00:00", "epoch": 0, "step": 19855, "train_loss": 3.5189294815063477, "perplexity": 33.74828096011927, "lr": 0.0026291804804649314, "grad_norm": 0.183176, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:13.271447+00:00", "epoch": 0, "step": 19856, "train_loss": 3.5067825317382812, "perplexity": 33.340821988349994, "lr": 0.0026291804804649314, "grad_norm": 0.181871, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:13.575689+00:00", "epoch": 0, "step": 19857, "train_loss": 3.4100632667541504, "perplexity": 30.267159103740298, "lr": 0.0026291804804649314, "grad_norm": 0.184181, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:13.879930+00:00", "epoch": 0, "step": 19858, "train_loss": 3.500540256500244, "perplexity": 33.13334763056277, "lr": 0.0026291804804649314, "grad_norm": 0.188894, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:14.184728+00:00", "epoch": 0, "step": 19859, "train_loss": 3.658642053604126, "perplexity": 38.8086070651443, "lr": 0.0026291804804649314, "grad_norm": 0.200097, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:14.490284+00:00", "epoch": 0, "step": 19860, "train_loss": 3.5130202770233154, "perplexity": 33.549443531083625, "lr": 0.0026291804804649314, "grad_norm": 0.178491, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:14.796680+00:00", "epoch": 0, "step": 19861, "train_loss": 3.5604825019836426, "perplexity": 35.18016755127842, "lr": 0.0026291804804649314, "grad_norm": 0.193002, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:15.102102+00:00", "epoch": 0, "step": 19862, "train_loss": 3.376244306564331, "perplexity": 29.260670382730574, "lr": 0.0026291804804649314, "grad_norm": 0.198691, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:15.406888+00:00", "epoch": 0, "step": 19863, "train_loss": 3.5191118717193604, "perplexity": 33.754436877644885, "lr": 0.0026291804804649314, "grad_norm": 0.177066, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:15.712290+00:00", "epoch": 0, "step": 19864, "train_loss": 3.5038185119628906, "perplexity": 33.24214544450525, "lr": 0.0026291804804649314, "grad_norm": 0.201957, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:16.016572+00:00", "epoch": 0, "step": 19865, "train_loss": 3.5407915115356445, "perplexity": 34.4942109544431, "lr": 0.0026291804804649314, "grad_norm": 0.20364, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:16.322179+00:00", "epoch": 0, "step": 19866, "train_loss": 3.5219480991363525, "perplexity": 33.85030802888912, "lr": 0.0026291804804649314, "grad_norm": 0.18342, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:16.627240+00:00", "epoch": 0, "step": 19867, "train_loss": 3.603407859802246, "perplexity": 36.72316885423906, "lr": 0.0026291804804649314, "grad_norm": 0.159417, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:16.932363+00:00", "epoch": 0, "step": 19868, "train_loss": 3.566376209259033, "perplexity": 35.388121368233904, "lr": 0.0026291804804649314, "grad_norm": 0.200102, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:17.236882+00:00", "epoch": 0, "step": 19869, "train_loss": 3.460101366043091, "perplexity": 31.820201839146524, "lr": 0.0026291804804649314, "grad_norm": 0.214861, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:17.541451+00:00", "epoch": 0, "step": 19870, "train_loss": 3.5029401779174805, "perplexity": 33.21296055533602, "lr": 0.0026291804804649314, "grad_norm": 0.165991, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:17.847440+00:00", "epoch": 0, "step": 19871, "train_loss": 3.4925670623779297, "perplexity": 32.87021939818043, "lr": 0.0026291804804649314, "grad_norm": 0.179627, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:18.151852+00:00", "epoch": 0, "step": 19872, "train_loss": 3.5417661666870117, "perplexity": 34.52784730410018, "lr": 0.0026291804804649314, "grad_norm": 0.183086, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:18.455994+00:00", "epoch": 0, "step": 19873, "train_loss": 3.504258632659912, "perplexity": 33.256779220806465, "lr": 0.0026291804804649314, "grad_norm": 0.181224, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:18.760439+00:00", "epoch": 0, "step": 19874, "train_loss": 3.4762682914733887, "perplexity": 32.338817584688265, "lr": 0.0026291804804649314, "grad_norm": 0.190136, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:19.064521+00:00", "epoch": 0, "step": 19875, "train_loss": 3.4910006523132324, "perplexity": 32.81877146048958, "lr": 0.0026291804804649314, "grad_norm": 0.173735, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:19.369835+00:00", "epoch": 0, "step": 19876, "train_loss": 3.4312398433685303, "perplexity": 30.914948692210093, "lr": 0.0026291804804649314, "grad_norm": 0.168801, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:19.675843+00:00", "epoch": 0, "step": 19877, "train_loss": 3.4091696739196777, "perplexity": 30.240124667933348, "lr": 0.0026291804804649314, "grad_norm": 0.156623, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:19.980556+00:00", "epoch": 0, "step": 19878, "train_loss": 3.537684679031372, "perplexity": 34.38720952247247, "lr": 0.0026291804804649314, "grad_norm": 0.164891, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:20.285338+00:00", "epoch": 0, "step": 19879, "train_loss": 3.436526298522949, "perplexity": 31.0788119282788, "lr": 0.0026291804804649314, "grad_norm": 0.15776, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:20.590381+00:00", "epoch": 0, "step": 19880, "train_loss": 3.4635636806488037, "perplexity": 31.930564333320046, "lr": 0.0026291804804649314, "grad_norm": 0.184218, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:20.895426+00:00", "epoch": 0, "step": 19881, "train_loss": 3.4634928703308105, "perplexity": 31.928303399955553, "lr": 0.0026291804804649314, "grad_norm": 0.163899, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:21.201610+00:00", "epoch": 0, "step": 19882, "train_loss": 3.5537285804748535, "perplexity": 34.94336403730191, "lr": 0.0026291804804649314, "grad_norm": 0.163258, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:21.507572+00:00", "epoch": 0, "step": 19883, "train_loss": 3.477499008178711, "perplexity": 32.37864200898552, "lr": 0.0026291804804649314, "grad_norm": 0.157028, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:21.812641+00:00", "epoch": 0, "step": 19884, "train_loss": 3.490664005279541, "perplexity": 32.80772497791513, "lr": 0.0026291804804649314, "grad_norm": 0.16543, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:22.117792+00:00", "epoch": 0, "step": 19885, "train_loss": 3.5867388248443604, "perplexity": 36.116102728464995, "lr": 0.0026291804804649314, "grad_norm": 0.168089, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:22.422443+00:00", "epoch": 0, "step": 19886, "train_loss": 3.5791327953338623, "perplexity": 35.84244463039021, "lr": 0.0026291804804649314, "grad_norm": 0.166795, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:22.729168+00:00", "epoch": 0, "step": 19887, "train_loss": 3.5516128540039062, "perplexity": 34.86951159036151, "lr": 0.0026291804804649314, "grad_norm": 0.155731, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:23.034580+00:00", "epoch": 0, "step": 19888, "train_loss": 3.4019105434417725, "perplexity": 30.021402484198266, "lr": 0.0026291804804649314, "grad_norm": 0.177799, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:23.339177+00:00", "epoch": 0, "step": 19889, "train_loss": 3.4514565467834473, "perplexity": 31.546307534562263, "lr": 0.0026291804804649314, "grad_norm": 0.180215, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:23.644359+00:00", "epoch": 0, "step": 19890, "train_loss": 3.4825408458709717, "perplexity": 32.542302093794596, "lr": 0.0026291804804649314, "grad_norm": 0.160771, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:23.949223+00:00", "epoch": 0, "step": 19891, "train_loss": 3.4147143363952637, "perplexity": 30.408261653096197, "lr": 0.0026291804804649314, "grad_norm": 0.171715, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:24.254475+00:00", "epoch": 0, "step": 19892, "train_loss": 3.4798569679260254, "perplexity": 32.45507962651214, "lr": 0.0026291804804649314, "grad_norm": 0.161225, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:24.559293+00:00", "epoch": 0, "step": 19893, "train_loss": 3.6007943153381348, "perplexity": 36.6273165312903, "lr": 0.0026291804804649314, "grad_norm": 0.165739, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:24.865797+00:00", "epoch": 0, "step": 19894, "train_loss": 3.617384910583496, "perplexity": 37.2400543024628, "lr": 0.0026291804804649314, "grad_norm": 0.166811, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:25.170722+00:00", "epoch": 0, "step": 19895, "train_loss": 3.4298784732818604, "perplexity": 30.87289064061187, "lr": 0.0026291804804649314, "grad_norm": 0.159805, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:25.476208+00:00", "epoch": 0, "step": 19896, "train_loss": 3.402503490447998, "perplexity": 30.039208863521914, "lr": 0.0026291804804649314, "grad_norm": 0.159236, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:25.781288+00:00", "epoch": 0, "step": 19897, "train_loss": 3.4084441661834717, "perplexity": 30.21819318023691, "lr": 0.0026291804804649314, "grad_norm": 0.187955, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:26.086696+00:00", "epoch": 0, "step": 19898, "train_loss": 3.4927737712860107, "perplexity": 32.87701466763778, "lr": 0.0026291804804649314, "grad_norm": 0.164928, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:26.392043+00:00", "epoch": 0, "step": 19899, "train_loss": 3.4635818004608154, "perplexity": 31.93114291438507, "lr": 0.0026291804804649314, "grad_norm": 0.185506, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:26.697685+00:00", "epoch": 0, "step": 19900, "train_loss": 3.4849562644958496, "perplexity": 32.62100038276271, "lr": 0.0026291804804649314, "grad_norm": 0.207641, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:27.002297+00:00", "epoch": 0, "step": 19901, "train_loss": 3.552396297454834, "perplexity": 34.89684058481705, "lr": 0.0026291804804649314, "grad_norm": 0.190695, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:27.306846+00:00", "epoch": 0, "step": 19902, "train_loss": 3.474278450012207, "perplexity": 32.274532444482574, "lr": 0.0026291804804649314, "grad_norm": 0.14434, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:27.612382+00:00", "epoch": 0, "step": 19903, "train_loss": 3.5265214443206787, "perplexity": 34.005471710194776, "lr": 0.0026291804804649314, "grad_norm": 0.166479, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:27.918681+00:00", "epoch": 0, "step": 19904, "train_loss": 3.5299019813537598, "perplexity": 34.12062299356472, "lr": 0.0026291804804649314, "grad_norm": 0.156883, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:28.223710+00:00", "epoch": 0, "step": 19905, "train_loss": 3.467001438140869, "perplexity": 32.04052276700508, "lr": 0.0026291804804649314, "grad_norm": 0.188857, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:28.527915+00:00", "epoch": 0, "step": 19906, "train_loss": 3.5283873081207275, "perplexity": 34.06898051983164, "lr": 0.0026291804804649314, "grad_norm": 0.176544, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:28.832433+00:00", "epoch": 0, "step": 19907, "train_loss": 3.453925848007202, "perplexity": 31.62430112556701, "lr": 0.0026291804804649314, "grad_norm": 0.173821, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:29.138325+00:00", "epoch": 0, "step": 19908, "train_loss": 3.495389223098755, "perplexity": 32.96311546240608, "lr": 0.0026291804804649314, "grad_norm": 0.186273, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:29.444034+00:00", "epoch": 0, "step": 19909, "train_loss": 3.4683351516723633, "perplexity": 32.08328415515795, "lr": 0.0026291804804649314, "grad_norm": 0.169229, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:29.750665+00:00", "epoch": 0, "step": 19910, "train_loss": 3.5702221393585205, "perplexity": 35.52448366121193, "lr": 0.0026291804804649314, "grad_norm": 0.171486, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:30.055987+00:00", "epoch": 0, "step": 19911, "train_loss": 3.618516683578491, "perplexity": 37.28222544983591, "lr": 0.0026291804804649314, "grad_norm": 0.155441, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:30.361188+00:00", "epoch": 0, "step": 19912, "train_loss": 3.4797112941741943, "perplexity": 32.45035211764231, "lr": 0.0026291804804649314, "grad_norm": 0.184999, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:30.665737+00:00", "epoch": 0, "step": 19913, "train_loss": 3.4253764152526855, "perplexity": 30.734211500422827, "lr": 0.0026291804804649314, "grad_norm": 0.1882, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:30.972385+00:00", "epoch": 0, "step": 19914, "train_loss": 3.4774770736694336, "perplexity": 32.37793180715098, "lr": 0.0026291804804649314, "grad_norm": 0.187526, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:31.278420+00:00", "epoch": 0, "step": 19915, "train_loss": 3.530836582183838, "perplexity": 34.152527062599695, "lr": 0.0026291804804649314, "grad_norm": 0.229505, "tokens_per_sec": 107074} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:31.582989+00:00", "epoch": 0, "step": 19916, "train_loss": 3.582259178161621, "perplexity": 35.95467718333476, "lr": 0.0026291804804649314, "grad_norm": 0.168284, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:31.889025+00:00", "epoch": 0, "step": 19917, "train_loss": 3.463059186935425, "perplexity": 31.914459627052352, "lr": 0.0026291804804649314, "grad_norm": 0.21575, "tokens_per_sec": 107072} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:32.194422+00:00", "epoch": 0, "step": 19918, "train_loss": 3.56742525100708, "perplexity": 35.42526446385682, "lr": 0.0026291804804649314, "grad_norm": 0.242492, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:32.498980+00:00", "epoch": 0, "step": 19919, "train_loss": 3.4290080070495605, "perplexity": 30.846028524794253, "lr": 0.0026291804804649314, "grad_norm": 0.171929, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:32.805408+00:00", "epoch": 0, "step": 19920, "train_loss": 3.431445837020874, "perplexity": 30.921317631361216, "lr": 0.0026291804804649314, "grad_norm": 0.184414, "tokens_per_sec": 106993} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:33.110086+00:00", "epoch": 0, "step": 19921, "train_loss": 3.59378981590271, "perplexity": 36.37165694233395, "lr": 0.0026291804804649314, "grad_norm": 0.191523, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:33.414157+00:00", "epoch": 0, "step": 19922, "train_loss": 3.4349746704101562, "perplexity": 31.030626562530124, "lr": 0.0026291804804649314, "grad_norm": 0.203932, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:33.718345+00:00", "epoch": 0, "step": 19923, "train_loss": 3.3974575996398926, "perplexity": 29.888016067597665, "lr": 0.0026291804804649314, "grad_norm": 0.166544, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:34.024330+00:00", "epoch": 0, "step": 19924, "train_loss": 3.4274742603302, "perplexity": 30.79875479197251, "lr": 0.0026291804804649314, "grad_norm": 0.182463, "tokens_per_sec": 107090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:34.330069+00:00", "epoch": 0, "step": 19925, "train_loss": 3.4780924320220947, "perplexity": 32.397861969397404, "lr": 0.0026291804804649314, "grad_norm": 0.195019, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:34.636186+00:00", "epoch": 0, "step": 19926, "train_loss": 3.4807050228118896, "perplexity": 32.48261498945978, "lr": 0.0026291804804649314, "grad_norm": 0.209078, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:34.941614+00:00", "epoch": 0, "step": 19927, "train_loss": 3.3702847957611084, "perplexity": 29.08680967841405, "lr": 0.0026291804804649314, "grad_norm": 0.17941, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:35.245710+00:00", "epoch": 0, "step": 19928, "train_loss": 3.4784538745880127, "perplexity": 32.409574052253014, "lr": 0.0026291804804649314, "grad_norm": 0.16676, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:35.551284+00:00", "epoch": 0, "step": 19929, "train_loss": 3.5107955932617188, "perplexity": 33.47488958907014, "lr": 0.0026291804804649314, "grad_norm": 0.169623, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:35.858177+00:00", "epoch": 0, "step": 19930, "train_loss": 3.555086612701416, "perplexity": 34.990850488536545, "lr": 0.0026291804804649314, "grad_norm": 0.165615, "tokens_per_sec": 106831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:36.164320+00:00", "epoch": 0, "step": 19931, "train_loss": 3.489123582839966, "perplexity": 32.75722612693508, "lr": 0.0026291804804649314, "grad_norm": 0.177467, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:36.469934+00:00", "epoch": 0, "step": 19932, "train_loss": 3.486895799636841, "perplexity": 32.68433135581005, "lr": 0.0026291804804649314, "grad_norm": 0.172432, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:36.774652+00:00", "epoch": 0, "step": 19933, "train_loss": 3.5049784183502197, "perplexity": 33.280725591689, "lr": 0.0026291804804649314, "grad_norm": 0.171673, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:37.079818+00:00", "epoch": 0, "step": 19934, "train_loss": 3.4509778022766113, "perplexity": 31.53120852769082, "lr": 0.0026291804804649314, "grad_norm": 0.171967, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:37.385319+00:00", "epoch": 0, "step": 19935, "train_loss": 3.568967580795288, "perplexity": 35.479944060640804, "lr": 0.0026291804804649314, "grad_norm": 0.182336, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:37.690845+00:00", "epoch": 0, "step": 19936, "train_loss": 3.5744736194610596, "perplexity": 35.6758368061159, "lr": 0.0026291804804649314, "grad_norm": 0.177581, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:37.996688+00:00", "epoch": 0, "step": 19937, "train_loss": 3.544175148010254, "perplexity": 34.61112450973123, "lr": 0.0026291804804649314, "grad_norm": 0.177149, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:38.301993+00:00", "epoch": 0, "step": 19938, "train_loss": 3.4148731231689453, "perplexity": 30.413090466223043, "lr": 0.0026291804804649314, "grad_norm": 0.163666, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:38.606867+00:00", "epoch": 0, "step": 19939, "train_loss": 3.4514150619506836, "perplexity": 31.544998868414968, "lr": 0.0026291804804649314, "grad_norm": 0.171527, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:38.913642+00:00", "epoch": 0, "step": 19940, "train_loss": 3.4703495502471924, "perplexity": 32.14797781456742, "lr": 0.0026291804804649314, "grad_norm": 0.1535, "tokens_per_sec": 106875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:39.219591+00:00", "epoch": 0, "step": 19941, "train_loss": 3.441786766052246, "perplexity": 31.242731778796074, "lr": 0.0026291804804649314, "grad_norm": 0.165199, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:39.525658+00:00", "epoch": 0, "step": 19942, "train_loss": 3.4754738807678223, "perplexity": 32.31313748341785, "lr": 0.0026291804804649314, "grad_norm": 0.167304, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:39.829944+00:00", "epoch": 0, "step": 19943, "train_loss": 3.6084389686584473, "perplexity": 36.908392664165184, "lr": 0.0026291804804649314, "grad_norm": 0.155474, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:40.135038+00:00", "epoch": 0, "step": 19944, "train_loss": 3.488093852996826, "perplexity": 32.72351239461086, "lr": 0.0026291804804649314, "grad_norm": 0.159449, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:40.440551+00:00", "epoch": 0, "step": 19945, "train_loss": 3.503045082092285, "perplexity": 33.216444916314714, "lr": 0.0026291804804649314, "grad_norm": 0.204397, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:40.745739+00:00", "epoch": 0, "step": 19946, "train_loss": 3.5434412956237793, "perplexity": 34.585734370857665, "lr": 0.0026291804804649314, "grad_norm": 0.191641, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:41.050582+00:00", "epoch": 0, "step": 19947, "train_loss": 3.4891059398651123, "perplexity": 32.75664819711647, "lr": 0.0026291804804649314, "grad_norm": 0.154414, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:41.355725+00:00", "epoch": 0, "step": 19948, "train_loss": 3.532473564147949, "perplexity": 34.208479917842645, "lr": 0.0026291804804649314, "grad_norm": 0.20887, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:41.661391+00:00", "epoch": 0, "step": 19949, "train_loss": 3.493716239929199, "perplexity": 32.908014829089076, "lr": 0.0026291804804649314, "grad_norm": 0.156337, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:41.966480+00:00", "epoch": 0, "step": 19950, "train_loss": 3.4367575645446777, "perplexity": 31.08600023264674, "lr": 0.0026291804804649314, "grad_norm": 0.153998, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:42.271120+00:00", "epoch": 0, "step": 19951, "train_loss": 3.5044336318969727, "perplexity": 33.2625996410669, "lr": 0.0026291804804649314, "grad_norm": 0.186958, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:42.576152+00:00", "epoch": 0, "step": 19952, "train_loss": 3.5996317863464355, "perplexity": 36.58476095477299, "lr": 0.0026291804804649314, "grad_norm": 0.146243, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:42.881582+00:00", "epoch": 0, "step": 19953, "train_loss": 3.510040760040283, "perplexity": 33.44963116447272, "lr": 0.0026291804804649314, "grad_norm": 0.169885, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:43.188481+00:00", "epoch": 0, "step": 19954, "train_loss": 3.4786720275878906, "perplexity": 32.41664506931102, "lr": 0.0026291804804649314, "grad_norm": 0.200246, "tokens_per_sec": 106773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:43.493309+00:00", "epoch": 0, "step": 19955, "train_loss": 3.5210118293762207, "perplexity": 33.81862984108926, "lr": 0.0026291804804649314, "grad_norm": 0.190828, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:43.797479+00:00", "epoch": 0, "step": 19956, "train_loss": 3.504347562789917, "perplexity": 33.25973688201677, "lr": 0.0026291804804649314, "grad_norm": 0.19963, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:44.102488+00:00", "epoch": 0, "step": 19957, "train_loss": 3.574847936630249, "perplexity": 35.68919338399989, "lr": 0.0026291804804649314, "grad_norm": 0.192598, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:44.408253+00:00", "epoch": 0, "step": 19958, "train_loss": 3.4818074703216553, "perplexity": 32.51844511425254, "lr": 0.0026291804804649314, "grad_norm": 0.181267, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:44.714317+00:00", "epoch": 0, "step": 19959, "train_loss": 3.4506547451019287, "perplexity": 31.521023789764445, "lr": 0.0026291804804649314, "grad_norm": 0.160047, "tokens_per_sec": 107064} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:45.020131+00:00", "epoch": 0, "step": 19960, "train_loss": 3.5486080646514893, "perplexity": 34.764893309872875, "lr": 0.0026291804804649314, "grad_norm": 0.158054, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:45.324977+00:00", "epoch": 0, "step": 19961, "train_loss": 3.4711713790893555, "perplexity": 32.17440880934194, "lr": 0.0026291804804649314, "grad_norm": 0.160289, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:45.629984+00:00", "epoch": 0, "step": 19962, "train_loss": 3.462677001953125, "perplexity": 31.902264730366305, "lr": 0.0026291804804649314, "grad_norm": 0.170411, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:45.935643+00:00", "epoch": 0, "step": 19963, "train_loss": 3.4488630294799805, "perplexity": 31.464597643919163, "lr": 0.0026291804804649314, "grad_norm": 0.157028, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:46.240560+00:00", "epoch": 0, "step": 19964, "train_loss": 3.4909679889678955, "perplexity": 32.81769950713071, "lr": 0.0026291804804649314, "grad_norm": 0.17588, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:46.545866+00:00", "epoch": 0, "step": 19965, "train_loss": 3.527845859527588, "perplexity": 34.05053891130151, "lr": 0.0026291804804649314, "grad_norm": 0.183024, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:46.850346+00:00", "epoch": 0, "step": 19966, "train_loss": 3.406038522720337, "perplexity": 30.14558634943445, "lr": 0.0026291804804649314, "grad_norm": 0.171159, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:47.155707+00:00", "epoch": 0, "step": 19967, "train_loss": 3.3558130264282227, "perplexity": 28.668903294076046, "lr": 0.0026291804804649314, "grad_norm": 0.192309, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:47.460680+00:00", "epoch": 0, "step": 19968, "train_loss": 3.4560773372650146, "perplexity": 31.69241371520306, "lr": 0.0026291804804649314, "grad_norm": 0.149143, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:47.766534+00:00", "epoch": 0, "step": 19969, "train_loss": 3.5021910667419434, "perplexity": 33.18808967210269, "lr": 0.0026291804804649314, "grad_norm": 0.176755, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:48.072745+00:00", "epoch": 0, "step": 19970, "train_loss": 3.458437204360962, "perplexity": 31.76729191607238, "lr": 0.0026291804804649314, "grad_norm": 0.180461, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:48.377306+00:00", "epoch": 0, "step": 19971, "train_loss": 3.5030875205993652, "perplexity": 33.21785460255975, "lr": 0.0026291804804649314, "grad_norm": 0.187882, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:48.682032+00:00", "epoch": 0, "step": 19972, "train_loss": 3.427917957305908, "perplexity": 30.812423138407034, "lr": 0.0026291804804649314, "grad_norm": 0.20039, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:48.987018+00:00", "epoch": 0, "step": 19973, "train_loss": 3.496755361557007, "perplexity": 33.008178416251674, "lr": 0.0026291804804649314, "grad_norm": 0.157839, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:49.293709+00:00", "epoch": 0, "step": 19974, "train_loss": 3.4647552967071533, "perplexity": 31.96863598542282, "lr": 0.0026291804804649314, "grad_norm": 0.192644, "tokens_per_sec": 106843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:49.599446+00:00", "epoch": 0, "step": 19975, "train_loss": 3.4883291721343994, "perplexity": 32.73121376943083, "lr": 0.0026291804804649314, "grad_norm": 0.187601, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:49.903983+00:00", "epoch": 0, "step": 19976, "train_loss": 3.4196860790252686, "perplexity": 30.559820146584798, "lr": 0.0026291804804649314, "grad_norm": 0.202142, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:50.208986+00:00", "epoch": 0, "step": 19977, "train_loss": 3.5111236572265625, "perplexity": 33.48587329565209, "lr": 0.0026291804804649314, "grad_norm": 0.170119, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:50.513536+00:00", "epoch": 0, "step": 19978, "train_loss": 3.5049984455108643, "perplexity": 33.28139211680107, "lr": 0.0026291804804649314, "grad_norm": 0.187198, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:50.818918+00:00", "epoch": 0, "step": 19979, "train_loss": 3.4706337451934814, "perplexity": 32.15711540576285, "lr": 0.0026291804804649314, "grad_norm": 0.174601, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:51.125640+00:00", "epoch": 0, "step": 19980, "train_loss": 3.567146062850952, "perplexity": 35.41537553009164, "lr": 0.0026291804804649314, "grad_norm": 0.179797, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:51.431510+00:00", "epoch": 0, "step": 19981, "train_loss": 3.611149311065674, "perplexity": 37.008562732256784, "lr": 0.0026291804804649314, "grad_norm": 0.174952, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:51.736473+00:00", "epoch": 0, "step": 19982, "train_loss": 3.504385232925415, "perplexity": 33.26098980441047, "lr": 0.0026291804804649314, "grad_norm": 0.194448, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:52.041059+00:00", "epoch": 0, "step": 19983, "train_loss": 3.547377109527588, "perplexity": 34.72212561427853, "lr": 0.0026291804804649314, "grad_norm": 0.175305, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:52.346027+00:00", "epoch": 0, "step": 19984, "train_loss": 3.530393123626709, "perplexity": 34.13738518987087, "lr": 0.0026291804804649314, "grad_norm": 0.155037, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:52.652135+00:00", "epoch": 0, "step": 19985, "train_loss": 3.5167076587677, "perplexity": 33.673381499565835, "lr": 0.0026291804804649314, "grad_norm": 0.176728, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:52.958001+00:00", "epoch": 0, "step": 19986, "train_loss": 3.4466114044189453, "perplexity": 31.393830867350406, "lr": 0.0026291804804649314, "grad_norm": 0.183252, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:53.263613+00:00", "epoch": 0, "step": 19987, "train_loss": 3.4981467723846436, "perplexity": 33.05413832023967, "lr": 0.0026291804804649314, "grad_norm": 0.154867, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:53.568510+00:00", "epoch": 0, "step": 19988, "train_loss": 3.6200497150421143, "perplexity": 37.33942410695263, "lr": 0.0026291804804649314, "grad_norm": 0.182279, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:53.873239+00:00", "epoch": 0, "step": 19989, "train_loss": 3.5382957458496094, "perplexity": 34.40822882662903, "lr": 0.0026291804804649314, "grad_norm": 0.152393, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:54.178598+00:00", "epoch": 0, "step": 19990, "train_loss": 3.4341635704040527, "perplexity": 31.005467825640974, "lr": 0.0026291804804649314, "grad_norm": 0.171301, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:54.483345+00:00", "epoch": 0, "step": 19991, "train_loss": 3.5923731327056885, "perplexity": 36.32016630866544, "lr": 0.0026291804804649314, "grad_norm": 0.174005, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:54.787352+00:00", "epoch": 0, "step": 19992, "train_loss": 3.4791622161865234, "perplexity": 32.43253923438092, "lr": 0.0026291804804649314, "grad_norm": 0.176813, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:55.091789+00:00", "epoch": 0, "step": 19993, "train_loss": 3.600097894668579, "perplexity": 36.60181739108229, "lr": 0.0026291804804649314, "grad_norm": 0.186218, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:55.395916+00:00", "epoch": 0, "step": 19994, "train_loss": 3.4388678073883057, "perplexity": 31.15166850580736, "lr": 0.0026291804804649314, "grad_norm": 0.16834, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:55.700503+00:00", "epoch": 0, "step": 19995, "train_loss": 3.5192043781280518, "perplexity": 33.757559523808, "lr": 0.0026291804804649314, "grad_norm": 0.162872, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:56.005866+00:00", "epoch": 0, "step": 19996, "train_loss": 3.577479362487793, "perplexity": 35.78323052192101, "lr": 0.0026291804804649314, "grad_norm": 0.180343, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:56.311231+00:00", "epoch": 0, "step": 19997, "train_loss": 3.660411834716797, "perplexity": 38.87735061751478, "lr": 0.0026291804804649314, "grad_norm": 0.162587, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:56.615956+00:00", "epoch": 0, "step": 19998, "train_loss": 3.602046489715576, "perplexity": 36.67320904528389, "lr": 0.0026291804804649314, "grad_norm": 0.167381, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:56.921943+00:00", "epoch": 0, "step": 19999, "train_loss": 3.538231134414673, "perplexity": 34.40600573341031, "lr": 0.0026291804804649314, "grad_norm": 0.168739, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:42:57.228156+00:00", "epoch": 0, "step": 20000, "train_loss": 3.4970126152038574, "perplexity": 33.01667098285017, "lr": 0.0026291804804649314, "grad_norm": 0.190043, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:43:00.302660+00:00", "step": 20000, "epoch": 0, "val_loss": 3.458089566230774, "val_ppl": 31.756250313461788, "eval_train_loss": 3.4970126152038574, "eval_train_ppl": 33.01667098285017} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:43:01.254387+00:00", "step": 20000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4581_epoch_0000_step_0020000.pt", "val_loss": 3.458089566230774} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T22:43:03.358215+00:00", "step": 20000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0020000.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:04.696634+00:00", "epoch": 0, "step": 20001, "train_loss": 3.526594877243042, "perplexity": 34.00796892304649, "lr": 0.0026291804804649314, "grad_norm": 0.166008, "tokens_per_sec": 4387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:05.000100+00:00", "epoch": 0, "step": 20002, "train_loss": 3.4372005462646484, "perplexity": 31.09977381299373, "lr": 0.0026291804804649314, "grad_norm": 0.161378, "tokens_per_sec": 107974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:05.303631+00:00", "epoch": 0, "step": 20003, "train_loss": 3.4264745712280273, "perplexity": 30.767980997123043, "lr": 0.0026291804804649314, "grad_norm": 0.18728, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:05.606975+00:00", "epoch": 0, "step": 20004, "train_loss": 3.515554904937744, "perplexity": 33.63458674477173, "lr": 0.0026291804804649314, "grad_norm": 0.173294, "tokens_per_sec": 108023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:05.910075+00:00", "epoch": 0, "step": 20005, "train_loss": 3.458298683166504, "perplexity": 31.762891777614723, "lr": 0.0026291804804649314, "grad_norm": 0.188433, "tokens_per_sec": 108109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:06.217198+00:00", "epoch": 0, "step": 20006, "train_loss": 3.5981547832489014, "perplexity": 36.5307650354065, "lr": 0.0026291804804649314, "grad_norm": 0.162172, "tokens_per_sec": 106693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:06.523285+00:00", "epoch": 0, "step": 20007, "train_loss": 3.5519373416900635, "perplexity": 34.88082815343829, "lr": 0.0026291804804649314, "grad_norm": 0.174765, "tokens_per_sec": 107056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:06.827694+00:00", "epoch": 0, "step": 20008, "train_loss": 3.5578949451446533, "perplexity": 35.089254540151465, "lr": 0.0026291804804649314, "grad_norm": 0.142996, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:07.131693+00:00", "epoch": 0, "step": 20009, "train_loss": 3.499197244644165, "perplexity": 33.088879019483215, "lr": 0.0026291804804649314, "grad_norm": 0.156292, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:08.127534+00:00", "epoch": 0, "step": 20010, "train_loss": 3.5362861156463623, "perplexity": 34.33915044498827, "lr": 0.0026291804804649314, "grad_norm": 0.18618, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:08.431327+00:00", "epoch": 0, "step": 20011, "train_loss": 3.5553643703460693, "perplexity": 35.000570814637605, "lr": 0.0026291804804649314, "grad_norm": 0.173434, "tokens_per_sec": 32954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:08.737049+00:00", "epoch": 0, "step": 20012, "train_loss": 3.4479668140411377, "perplexity": 31.436411218175284, "lr": 0.0026291804804649314, "grad_norm": 0.161572, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:09.041030+00:00", "epoch": 0, "step": 20013, "train_loss": 3.6739978790283203, "perplexity": 39.409144332330925, "lr": 0.0026291804804649314, "grad_norm": 0.194883, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:09.344816+00:00", "epoch": 0, "step": 20014, "train_loss": 3.60552978515625, "perplexity": 36.80117541008541, "lr": 0.0026291804804649314, "grad_norm": 0.159048, "tokens_per_sec": 107872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:09.648348+00:00", "epoch": 0, "step": 20015, "train_loss": 3.4589569568634033, "perplexity": 31.783807337135464, "lr": 0.0026291804804649314, "grad_norm": 0.21323, "tokens_per_sec": 107950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:09.951898+00:00", "epoch": 0, "step": 20016, "train_loss": 3.5057685375213623, "perplexity": 33.30703172212897, "lr": 0.0026291804804649314, "grad_norm": 0.179378, "tokens_per_sec": 107950} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:10.256939+00:00", "epoch": 0, "step": 20017, "train_loss": 3.4932291507720947, "perplexity": 32.8919895950597, "lr": 0.0026291804804649314, "grad_norm": 0.19018, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:10.562657+00:00", "epoch": 0, "step": 20018, "train_loss": 3.625363826751709, "perplexity": 37.53837814161018, "lr": 0.0026291804804649314, "grad_norm": 0.212743, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:10.866882+00:00", "epoch": 0, "step": 20019, "train_loss": 3.521047830581665, "perplexity": 33.81984737444622, "lr": 0.0026291804804649314, "grad_norm": 0.180456, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:11.171125+00:00", "epoch": 0, "step": 20020, "train_loss": 3.523625373840332, "perplexity": 33.90713193559688, "lr": 0.0026291804804649314, "grad_norm": 0.169301, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:11.474556+00:00", "epoch": 0, "step": 20021, "train_loss": 3.5648579597473145, "perplexity": 35.33443413586589, "lr": 0.0026291804804649314, "grad_norm": 0.1679, "tokens_per_sec": 107922} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:11.779413+00:00", "epoch": 0, "step": 20022, "train_loss": 3.4305853843688965, "perplexity": 30.894722745063753, "lr": 0.0026291804804649314, "grad_norm": 0.189071, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:12.084513+00:00", "epoch": 0, "step": 20023, "train_loss": 3.554569721221924, "perplexity": 34.97276868962531, "lr": 0.0026291804804649314, "grad_norm": 0.160812, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:12.388370+00:00", "epoch": 0, "step": 20024, "train_loss": 3.5277090072631836, "perplexity": 34.045879336791245, "lr": 0.0026291804804649314, "grad_norm": 0.188487, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:12.692303+00:00", "epoch": 0, "step": 20025, "train_loss": 3.510965585708618, "perplexity": 33.48058055115765, "lr": 0.0026291804804649314, "grad_norm": 0.20311, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:12.996387+00:00", "epoch": 0, "step": 20026, "train_loss": 3.4447543621063232, "perplexity": 31.335585294060333, "lr": 0.0026291804804649314, "grad_norm": 0.167916, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:13.300478+00:00", "epoch": 0, "step": 20027, "train_loss": 3.5275704860687256, "perplexity": 34.04116358754229, "lr": 0.0026291804804649314, "grad_norm": 0.183056, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:13.605154+00:00", "epoch": 0, "step": 20028, "train_loss": 3.378260850906372, "perplexity": 29.31973535558821, "lr": 0.0026291804804649314, "grad_norm": 0.223427, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:13.908760+00:00", "epoch": 0, "step": 20029, "train_loss": 3.551846504211426, "perplexity": 34.87765981086034, "lr": 0.0026291804804649314, "grad_norm": 0.186754, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:14.213021+00:00", "epoch": 0, "step": 20030, "train_loss": 3.4812915325164795, "perplexity": 32.50167194638035, "lr": 0.0026291804804649314, "grad_norm": 0.191292, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:14.516657+00:00", "epoch": 0, "step": 20031, "train_loss": 3.432126522064209, "perplexity": 30.942372474838667, "lr": 0.0026291804804649314, "grad_norm": 0.156566, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:14.820651+00:00", "epoch": 0, "step": 20032, "train_loss": 3.548271417617798, "perplexity": 34.75319178141644, "lr": 0.0026291804804649314, "grad_norm": 0.175261, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:15.124380+00:00", "epoch": 0, "step": 20033, "train_loss": 3.4373202323913574, "perplexity": 31.103496247220352, "lr": 0.0026291804804649314, "grad_norm": 0.160596, "tokens_per_sec": 107887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:15.428462+00:00", "epoch": 0, "step": 20034, "train_loss": 3.5404975414276123, "perplexity": 34.48407217784397, "lr": 0.0026291804804649314, "grad_norm": 0.156943, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:15.732433+00:00", "epoch": 0, "step": 20035, "train_loss": 3.5966265201568604, "perplexity": 36.47497905416471, "lr": 0.0026291804804649314, "grad_norm": 0.191381, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:16.036395+00:00", "epoch": 0, "step": 20036, "train_loss": 3.4873790740966797, "perplexity": 32.70013067578763, "lr": 0.0026291804804649314, "grad_norm": 0.180437, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:16.340235+00:00", "epoch": 0, "step": 20037, "train_loss": 3.572812795639038, "perplexity": 35.61663470221826, "lr": 0.0026291804804649314, "grad_norm": 0.182698, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:16.644270+00:00", "epoch": 0, "step": 20038, "train_loss": 3.5429294109344482, "perplexity": 34.568034993370595, "lr": 0.0026291804804649314, "grad_norm": 0.197608, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:16.948446+00:00", "epoch": 0, "step": 20039, "train_loss": 3.522919178009033, "perplexity": 33.88319531333888, "lr": 0.0026291804804649314, "grad_norm": 0.168954, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:17.252668+00:00", "epoch": 0, "step": 20040, "train_loss": 3.465477228164673, "perplexity": 31.991723482187044, "lr": 0.0026291804804649314, "grad_norm": 0.222878, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:17.555833+00:00", "epoch": 0, "step": 20041, "train_loss": 3.4764161109924316, "perplexity": 32.343598246478905, "lr": 0.0026291804804649314, "grad_norm": 0.187725, "tokens_per_sec": 108018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:17.859930+00:00", "epoch": 0, "step": 20042, "train_loss": 3.471219539642334, "perplexity": 32.17595838397584, "lr": 0.0026291804804649314, "grad_norm": 0.177461, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:18.163809+00:00", "epoch": 0, "step": 20043, "train_loss": 3.565598487854004, "perplexity": 35.36060996825107, "lr": 0.0026291804804649314, "grad_norm": 0.191392, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:18.467872+00:00", "epoch": 0, "step": 20044, "train_loss": 3.6028788089752197, "perplexity": 36.70374556979109, "lr": 0.0026291804804649314, "grad_norm": 0.142652, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:18.771506+00:00", "epoch": 0, "step": 20045, "train_loss": 3.457059144973755, "perplexity": 31.72354485119115, "lr": 0.0026291804804649314, "grad_norm": 0.168577, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:19.075029+00:00", "epoch": 0, "step": 20046, "train_loss": 3.567333936691284, "perplexity": 35.422029777759235, "lr": 0.0026291804804649314, "grad_norm": 0.158836, "tokens_per_sec": 107959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:19.377732+00:00", "epoch": 0, "step": 20047, "train_loss": 3.4273273944854736, "perplexity": 30.794231838975247, "lr": 0.0026291804804649314, "grad_norm": 0.153494, "tokens_per_sec": 108250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:19.682507+00:00", "epoch": 0, "step": 20048, "train_loss": 3.5454747676849365, "perplexity": 34.65613505006358, "lr": 0.0026291804804649314, "grad_norm": 0.181718, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:19.987090+00:00", "epoch": 0, "step": 20049, "train_loss": 3.439967155456543, "perplexity": 31.18593386372811, "lr": 0.0026291804804649314, "grad_norm": 0.167835, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:20.291666+00:00", "epoch": 0, "step": 20050, "train_loss": 3.4884626865386963, "perplexity": 32.735584149687035, "lr": 0.0026291804804649314, "grad_norm": 0.16986, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:20.595425+00:00", "epoch": 0, "step": 20051, "train_loss": 3.5448315143585205, "perplexity": 34.6338495442932, "lr": 0.0026291804804649314, "grad_norm": 0.165171, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:20.899359+00:00", "epoch": 0, "step": 20052, "train_loss": 3.4557530879974365, "perplexity": 31.682139139120025, "lr": 0.0026291804804649314, "grad_norm": 0.166011, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:21.203302+00:00", "epoch": 0, "step": 20053, "train_loss": 3.545231342315674, "perplexity": 34.64769989429891, "lr": 0.0026291804804649314, "grad_norm": 0.193098, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:21.507885+00:00", "epoch": 0, "step": 20054, "train_loss": 3.4568891525268555, "perplexity": 31.71815254651609, "lr": 0.0026291804804649314, "grad_norm": 0.182956, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:21.811771+00:00", "epoch": 0, "step": 20055, "train_loss": 3.5542681217193604, "perplexity": 34.9622225104264, "lr": 0.0026291804804649314, "grad_norm": 0.200757, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:22.114999+00:00", "epoch": 0, "step": 20056, "train_loss": 3.490128993988037, "perplexity": 32.79017716915987, "lr": 0.0026291804804649314, "grad_norm": 0.178206, "tokens_per_sec": 108065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:22.419203+00:00", "epoch": 0, "step": 20057, "train_loss": 3.473879337310791, "perplexity": 32.26165383883021, "lr": 0.0026291804804649314, "grad_norm": 0.159651, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:22.723096+00:00", "epoch": 0, "step": 20058, "train_loss": 3.3664162158966064, "perplexity": 28.974502407058846, "lr": 0.0026291804804649314, "grad_norm": 0.171779, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:23.027772+00:00", "epoch": 0, "step": 20059, "train_loss": 3.4858033657073975, "perplexity": 32.648645379110675, "lr": 0.0026291804804649314, "grad_norm": 0.175676, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:23.332544+00:00", "epoch": 0, "step": 20060, "train_loss": 3.535645008087158, "perplexity": 34.31714241157344, "lr": 0.0026291804804649314, "grad_norm": 0.192769, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:23.636372+00:00", "epoch": 0, "step": 20061, "train_loss": 3.539907455444336, "perplexity": 34.46372961272708, "lr": 0.0026291804804649314, "grad_norm": 0.181011, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:23.939779+00:00", "epoch": 0, "step": 20062, "train_loss": 3.4388444423675537, "perplexity": 31.15094065492942, "lr": 0.0026291804804649314, "grad_norm": 0.227271, "tokens_per_sec": 108001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:24.244147+00:00", "epoch": 0, "step": 20063, "train_loss": 3.556771993637085, "perplexity": 35.049873124714466, "lr": 0.0026291804804649314, "grad_norm": 0.203956, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:24.549578+00:00", "epoch": 0, "step": 20064, "train_loss": 3.405151128768921, "perplexity": 30.118847204278595, "lr": 0.0026291804804649314, "grad_norm": 0.166964, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:24.853947+00:00", "epoch": 0, "step": 20065, "train_loss": 3.496349573135376, "perplexity": 32.99478679689703, "lr": 0.0026291804804649314, "grad_norm": 0.158344, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:25.158436+00:00", "epoch": 0, "step": 20066, "train_loss": 3.549614429473877, "perplexity": 34.799897085884176, "lr": 0.0026291804804649314, "grad_norm": 0.166945, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:25.463260+00:00", "epoch": 0, "step": 20067, "train_loss": 3.541980266571045, "perplexity": 34.53524050361724, "lr": 0.0026291804804649314, "grad_norm": 0.161034, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:25.768531+00:00", "epoch": 0, "step": 20068, "train_loss": 3.3882687091827393, "perplexity": 29.614636313857744, "lr": 0.0026291804804649314, "grad_norm": 0.158732, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:26.074676+00:00", "epoch": 0, "step": 20069, "train_loss": 3.5345311164855957, "perplexity": 34.27893811652345, "lr": 0.0026291804804649314, "grad_norm": 0.171961, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:26.379880+00:00", "epoch": 0, "step": 20070, "train_loss": 3.5494208335876465, "perplexity": 34.79316062106408, "lr": 0.0026291804804649314, "grad_norm": 0.158944, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:26.684435+00:00", "epoch": 0, "step": 20071, "train_loss": 3.519122838973999, "perplexity": 33.754807073179315, "lr": 0.0026291804804649314, "grad_norm": 0.176871, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:26.989661+00:00", "epoch": 0, "step": 20072, "train_loss": 3.5500779151916504, "perplexity": 34.81603007958074, "lr": 0.0026291804804649314, "grad_norm": 0.200646, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:27.294428+00:00", "epoch": 0, "step": 20073, "train_loss": 3.5811052322387695, "perplexity": 35.913211359452106, "lr": 0.0026291804804649314, "grad_norm": 0.169696, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:27.599717+00:00", "epoch": 0, "step": 20074, "train_loss": 3.427232027053833, "perplexity": 30.791295212206922, "lr": 0.0026291804804649314, "grad_norm": 0.165162, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:27.905662+00:00", "epoch": 0, "step": 20075, "train_loss": 3.528965711593628, "perplexity": 34.08869183647995, "lr": 0.0026291804804649314, "grad_norm": 0.191073, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:28.210732+00:00", "epoch": 0, "step": 20076, "train_loss": 3.520188331604004, "perplexity": 33.79079173863657, "lr": 0.0026291804804649314, "grad_norm": 0.166671, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:28.514791+00:00", "epoch": 0, "step": 20077, "train_loss": 3.4758265018463135, "perplexity": 32.32453378597685, "lr": 0.0026291804804649314, "grad_norm": 0.188249, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:28.819492+00:00", "epoch": 0, "step": 20078, "train_loss": 3.4020750522613525, "perplexity": 30.02634167594224, "lr": 0.0026291804804649314, "grad_norm": 0.17935, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:29.124736+00:00", "epoch": 0, "step": 20079, "train_loss": 3.471456527709961, "perplexity": 32.1835846058034, "lr": 0.0026291804804649314, "grad_norm": 0.178605, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:29.429645+00:00", "epoch": 0, "step": 20080, "train_loss": 3.4855434894561768, "perplexity": 32.640161873920626, "lr": 0.0026291804804649314, "grad_norm": 0.166834, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:29.734638+00:00", "epoch": 0, "step": 20081, "train_loss": 3.520488739013672, "perplexity": 33.80094426772454, "lr": 0.0026291804804649314, "grad_norm": 0.169721, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:30.039362+00:00", "epoch": 0, "step": 20082, "train_loss": 3.4318456649780273, "perplexity": 30.933683310529638, "lr": 0.0026291804804649314, "grad_norm": 0.149949, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:30.343479+00:00", "epoch": 0, "step": 20083, "train_loss": 3.4835705757141113, "perplexity": 32.575829132358606, "lr": 0.0026291804804649314, "grad_norm": 0.164188, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:30.647308+00:00", "epoch": 0, "step": 20084, "train_loss": 3.45576810836792, "perplexity": 31.68261502016155, "lr": 0.0026291804804649314, "grad_norm": 0.174931, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:30.952522+00:00", "epoch": 0, "step": 20085, "train_loss": 3.331566095352173, "perplexity": 27.98213008976986, "lr": 0.0026291804804649314, "grad_norm": 0.150374, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:31.256922+00:00", "epoch": 0, "step": 20086, "train_loss": 3.523313283920288, "perplexity": 33.89655151260974, "lr": 0.0026291804804649314, "grad_norm": 0.153656, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:31.561523+00:00", "epoch": 0, "step": 20087, "train_loss": 3.534960985183716, "perplexity": 34.29367672663115, "lr": 0.0026291804804649314, "grad_norm": 0.164471, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:31.865535+00:00", "epoch": 0, "step": 20088, "train_loss": 3.534221649169922, "perplexity": 34.268331546838354, "lr": 0.0026291804804649314, "grad_norm": 0.156907, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:32.169770+00:00", "epoch": 0, "step": 20089, "train_loss": 3.5426833629608154, "perplexity": 34.559530644689794, "lr": 0.0026291804804649314, "grad_norm": 0.171582, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:32.475554+00:00", "epoch": 0, "step": 20090, "train_loss": 3.5815417766571045, "perplexity": 35.92889249392239, "lr": 0.0026291804804649314, "grad_norm": 0.177169, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:32.780376+00:00", "epoch": 0, "step": 20091, "train_loss": 3.5131640434265137, "perplexity": 33.55426716063902, "lr": 0.0026291804804649314, "grad_norm": 0.178764, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:33.084771+00:00", "epoch": 0, "step": 20092, "train_loss": 3.598212957382202, "perplexity": 36.5328902428167, "lr": 0.0026291804804649314, "grad_norm": 0.156472, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:33.389425+00:00", "epoch": 0, "step": 20093, "train_loss": 3.4095945358276367, "perplexity": 30.25297527467003, "lr": 0.0026291804804649314, "grad_norm": 0.176842, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:33.693221+00:00", "epoch": 0, "step": 20094, "train_loss": 3.505728006362915, "perplexity": 33.30568177690644, "lr": 0.0026291804804649314, "grad_norm": 0.196575, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:33.997613+00:00", "epoch": 0, "step": 20095, "train_loss": 3.5775535106658936, "perplexity": 35.785883881640444, "lr": 0.0026291804804649314, "grad_norm": 0.239676, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:34.303930+00:00", "epoch": 0, "step": 20096, "train_loss": 3.5914037227630615, "perplexity": 36.28497423885514, "lr": 0.0026291804804649314, "grad_norm": 0.192808, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:34.610257+00:00", "epoch": 0, "step": 20097, "train_loss": 3.5731523036956787, "perplexity": 35.628728889570944, "lr": 0.0026291804804649314, "grad_norm": 0.154156, "tokens_per_sec": 106970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:34.915245+00:00", "epoch": 0, "step": 20098, "train_loss": 3.3795928955078125, "perplexity": 29.358816573968642, "lr": 0.0026291804804649314, "grad_norm": 0.19563, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:35.220512+00:00", "epoch": 0, "step": 20099, "train_loss": 3.4603421688079834, "perplexity": 31.827865154365547, "lr": 0.0026291804804649314, "grad_norm": 0.169514, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:35.525650+00:00", "epoch": 0, "step": 20100, "train_loss": 3.581265449523926, "perplexity": 35.91896573764048, "lr": 0.0026291804804649314, "grad_norm": 0.165433, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:35.830723+00:00", "epoch": 0, "step": 20101, "train_loss": 3.4795968532562256, "perplexity": 32.4466386820461, "lr": 0.0026291804804649314, "grad_norm": 0.152039, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:36.136330+00:00", "epoch": 0, "step": 20102, "train_loss": 3.5530290603637695, "perplexity": 34.918928998805704, "lr": 0.0026291804804649314, "grad_norm": 0.158221, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:36.441575+00:00", "epoch": 0, "step": 20103, "train_loss": 3.493335723876953, "perplexity": 32.89549518331321, "lr": 0.0026291804804649314, "grad_norm": 0.160081, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:36.746575+00:00", "epoch": 0, "step": 20104, "train_loss": 3.398040294647217, "perplexity": 29.90543674031568, "lr": 0.0026291804804649314, "grad_norm": 0.161237, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:37.051479+00:00", "epoch": 0, "step": 20105, "train_loss": 3.501948118209839, "perplexity": 33.18002765380091, "lr": 0.0026291804804649314, "grad_norm": 0.156076, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:37.357874+00:00", "epoch": 0, "step": 20106, "train_loss": 3.414144992828369, "perplexity": 30.39095383245926, "lr": 0.0026291804804649314, "grad_norm": 0.174844, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:37.663810+00:00", "epoch": 0, "step": 20107, "train_loss": 3.5110514163970947, "perplexity": 33.48345433576465, "lr": 0.0026291804804649314, "grad_norm": 0.144318, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:37.967998+00:00", "epoch": 0, "step": 20108, "train_loss": 3.5019218921661377, "perplexity": 33.17915748435626, "lr": 0.0026291804804649314, "grad_norm": 0.163505, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:38.272159+00:00", "epoch": 0, "step": 20109, "train_loss": 3.5068066120147705, "perplexity": 33.34162485422843, "lr": 0.0026291804804649314, "grad_norm": 0.172203, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:38.576614+00:00", "epoch": 0, "step": 20110, "train_loss": 3.484525680541992, "perplexity": 32.60695732701332, "lr": 0.0026291804804649314, "grad_norm": 0.15401, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:38.881139+00:00", "epoch": 0, "step": 20111, "train_loss": 3.5399887561798096, "perplexity": 34.466531653194195, "lr": 0.0026291804804649314, "grad_norm": 0.186583, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:39.186982+00:00", "epoch": 0, "step": 20112, "train_loss": 3.4916412830352783, "perplexity": 32.839802909723936, "lr": 0.0026291804804649314, "grad_norm": 0.179268, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:39.490758+00:00", "epoch": 0, "step": 20113, "train_loss": 3.436041831970215, "perplexity": 31.063758930032332, "lr": 0.0026291804804649314, "grad_norm": 0.172261, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:39.795289+00:00", "epoch": 0, "step": 20114, "train_loss": 3.4828782081604004, "perplexity": 32.55328249141414, "lr": 0.0026291804804649314, "grad_norm": 0.190371, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:40.099922+00:00", "epoch": 0, "step": 20115, "train_loss": 3.5292296409606934, "perplexity": 34.09769003073207, "lr": 0.0026291804804649314, "grad_norm": 0.182166, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:40.405384+00:00", "epoch": 0, "step": 20116, "train_loss": 3.544625997543335, "perplexity": 34.62673243720488, "lr": 0.0026291804804649314, "grad_norm": 0.18245, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:40.711228+00:00", "epoch": 0, "step": 20117, "train_loss": 3.549133062362671, "perplexity": 34.783149591123646, "lr": 0.0026291804804649314, "grad_norm": 0.176819, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:41.014972+00:00", "epoch": 0, "step": 20118, "train_loss": 3.4684650897979736, "perplexity": 32.08745326782197, "lr": 0.0026291804804649314, "grad_norm": 0.186781, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:41.320182+00:00", "epoch": 0, "step": 20119, "train_loss": 3.4550180435180664, "perplexity": 31.65885991433214, "lr": 0.0026291804804649314, "grad_norm": 0.150864, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:41.625570+00:00", "epoch": 0, "step": 20120, "train_loss": 3.521798849105835, "perplexity": 33.845256246381496, "lr": 0.0026291804804649314, "grad_norm": 0.172127, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:41.931020+00:00", "epoch": 0, "step": 20121, "train_loss": 3.50231671333313, "perplexity": 33.19225990442024, "lr": 0.0026291804804649314, "grad_norm": 0.157839, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:42.237684+00:00", "epoch": 0, "step": 20122, "train_loss": 3.4672181606292725, "perplexity": 32.04746742133309, "lr": 0.0026291804804649314, "grad_norm": 0.179709, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:42.542302+00:00", "epoch": 0, "step": 20123, "train_loss": 3.5465738773345947, "perplexity": 34.694246883222036, "lr": 0.0026291804804649314, "grad_norm": 0.18239, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:42.847151+00:00", "epoch": 0, "step": 20124, "train_loss": 3.574418544769287, "perplexity": 35.67387202450545, "lr": 0.0026291804804649314, "grad_norm": 0.183117, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:43.152590+00:00", "epoch": 0, "step": 20125, "train_loss": 3.501457452774048, "perplexity": 33.163751354507895, "lr": 0.0026291804804649314, "grad_norm": 0.204021, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:43.457636+00:00", "epoch": 0, "step": 20126, "train_loss": 3.5197231769561768, "perplexity": 33.77507744987615, "lr": 0.0026291804804649314, "grad_norm": 0.242934, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:43.763405+00:00", "epoch": 0, "step": 20127, "train_loss": 3.4258689880371094, "perplexity": 30.749354065660253, "lr": 0.0026291804804649314, "grad_norm": 0.25785, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:44.068385+00:00", "epoch": 0, "step": 20128, "train_loss": 3.6292243003845215, "perplexity": 37.683574142987474, "lr": 0.0026291804804649314, "grad_norm": 0.192753, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:44.373279+00:00", "epoch": 0, "step": 20129, "train_loss": 3.420201063156128, "perplexity": 30.575562022062332, "lr": 0.0026291804804649314, "grad_norm": 0.197936, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:44.678730+00:00", "epoch": 0, "step": 20130, "train_loss": 3.586406946182251, "perplexity": 36.10411855336686, "lr": 0.0026291804804649314, "grad_norm": 0.204723, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:44.984527+00:00", "epoch": 0, "step": 20131, "train_loss": 3.4688289165496826, "perplexity": 32.099129665679165, "lr": 0.0026291804804649314, "grad_norm": 0.210448, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:45.290216+00:00", "epoch": 0, "step": 20132, "train_loss": 3.468595266342163, "perplexity": 32.09163057348792, "lr": 0.0026291804804649314, "grad_norm": 0.177375, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:45.595456+00:00", "epoch": 0, "step": 20133, "train_loss": 3.6611461639404297, "perplexity": 38.90590987687666, "lr": 0.0026291804804649314, "grad_norm": 0.177895, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:45.900946+00:00", "epoch": 0, "step": 20134, "train_loss": 3.4068214893341064, "perplexity": 30.16919857968628, "lr": 0.0026291804804649314, "grad_norm": 0.173825, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:46.205662+00:00", "epoch": 0, "step": 20135, "train_loss": 3.5143067836761475, "perplexity": 33.59263288912651, "lr": 0.0026291804804649314, "grad_norm": 0.207172, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:46.511452+00:00", "epoch": 0, "step": 20136, "train_loss": 3.487428665161133, "perplexity": 32.70175235028555, "lr": 0.0026291804804649314, "grad_norm": 0.196533, "tokens_per_sec": 107159} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:46.816482+00:00", "epoch": 0, "step": 20137, "train_loss": 3.4986679553985596, "perplexity": 33.07137006572828, "lr": 0.0026291804804649314, "grad_norm": 0.172501, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:47.121901+00:00", "epoch": 0, "step": 20138, "train_loss": 3.5817811489105225, "perplexity": 35.93749390330977, "lr": 0.0026291804804649314, "grad_norm": 0.207408, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:47.426758+00:00", "epoch": 0, "step": 20139, "train_loss": 3.4707796573638916, "perplexity": 32.161807862600796, "lr": 0.0026291804804649314, "grad_norm": 0.170015, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:47.732005+00:00", "epoch": 0, "step": 20140, "train_loss": 3.5506465435028076, "perplexity": 34.83583308970861, "lr": 0.0026291804804649314, "grad_norm": 0.186065, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:48.038878+00:00", "epoch": 0, "step": 20141, "train_loss": 3.53324294090271, "perplexity": 34.23480925439811, "lr": 0.0026291804804649314, "grad_norm": 0.175663, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:48.345739+00:00", "epoch": 0, "step": 20142, "train_loss": 3.4784178733825684, "perplexity": 32.408407289521755, "lr": 0.0026291804804649314, "grad_norm": 0.193511, "tokens_per_sec": 106785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:48.651806+00:00", "epoch": 0, "step": 20143, "train_loss": 3.4723949432373047, "perplexity": 32.213800356570516, "lr": 0.0026291804804649314, "grad_norm": 0.173339, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:48.957421+00:00", "epoch": 0, "step": 20144, "train_loss": 3.5782179832458496, "perplexity": 35.80967052213928, "lr": 0.0026291804804649314, "grad_norm": 0.167361, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:49.263717+00:00", "epoch": 0, "step": 20145, "train_loss": 3.4978392124176025, "perplexity": 33.04397375373433, "lr": 0.0026291804804649314, "grad_norm": 0.174174, "tokens_per_sec": 106981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:49.569553+00:00", "epoch": 0, "step": 20146, "train_loss": 3.5401318073272705, "perplexity": 34.47146248276819, "lr": 0.0026291804804649314, "grad_norm": 0.176896, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:49.873758+00:00", "epoch": 0, "step": 20147, "train_loss": 3.4394848346710205, "perplexity": 31.17089586647056, "lr": 0.0026291804804649314, "grad_norm": 0.166486, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:50.177355+00:00", "epoch": 0, "step": 20148, "train_loss": 3.5150959491729736, "perplexity": 33.61915349914897, "lr": 0.0026291804804649314, "grad_norm": 0.152422, "tokens_per_sec": 107933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:50.482037+00:00", "epoch": 0, "step": 20149, "train_loss": 3.6033577919006348, "perplexity": 36.72133024826197, "lr": 0.0026291804804649314, "grad_norm": 0.160979, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:50.787985+00:00", "epoch": 0, "step": 20150, "train_loss": 3.476705551147461, "perplexity": 32.352961137502525, "lr": 0.0026291804804649314, "grad_norm": 0.157092, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:51.093415+00:00", "epoch": 0, "step": 20151, "train_loss": 3.489001750946045, "perplexity": 32.75323549513449, "lr": 0.0026291804804649314, "grad_norm": 0.16505, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:51.397383+00:00", "epoch": 0, "step": 20152, "train_loss": 3.602952718734741, "perplexity": 36.70645843505206, "lr": 0.0026291804804649314, "grad_norm": 0.178505, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:51.701957+00:00", "epoch": 0, "step": 20153, "train_loss": 3.4851691722869873, "perplexity": 32.62794638730191, "lr": 0.0026291804804649314, "grad_norm": 0.182508, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:52.006378+00:00", "epoch": 0, "step": 20154, "train_loss": 3.4680392742156982, "perplexity": 32.07379283884316, "lr": 0.0026291804804649314, "grad_norm": 0.18253, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:52.311567+00:00", "epoch": 0, "step": 20155, "train_loss": 3.506340980529785, "perplexity": 33.326103557818314, "lr": 0.0026291804804649314, "grad_norm": 0.173786, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:52.616397+00:00", "epoch": 0, "step": 20156, "train_loss": 3.509120464324951, "perplexity": 33.41886177287391, "lr": 0.0026291804804649314, "grad_norm": 0.180943, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:52.921174+00:00", "epoch": 0, "step": 20157, "train_loss": 3.485558032989502, "perplexity": 32.64063658065452, "lr": 0.0026291804804649314, "grad_norm": 0.20654, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:53.225517+00:00", "epoch": 0, "step": 20158, "train_loss": 3.5470268726348877, "perplexity": 34.709966774256905, "lr": 0.0026291804804649314, "grad_norm": 0.204713, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:53.530206+00:00", "epoch": 0, "step": 20159, "train_loss": 3.4801623821258545, "perplexity": 32.46499338251053, "lr": 0.0026291804804649314, "grad_norm": 0.170331, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:53.835360+00:00", "epoch": 0, "step": 20160, "train_loss": 3.5656886100769043, "perplexity": 35.36379688862854, "lr": 0.0026291804804649314, "grad_norm": 0.225076, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:54.140407+00:00", "epoch": 0, "step": 20161, "train_loss": 3.474179267883301, "perplexity": 32.27133154638384, "lr": 0.0026291804804649314, "grad_norm": 0.238266, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:54.446048+00:00", "epoch": 0, "step": 20162, "train_loss": 3.5984957218170166, "perplexity": 36.54322190552289, "lr": 0.0026291804804649314, "grad_norm": 0.160546, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:54.750807+00:00", "epoch": 0, "step": 20163, "train_loss": 3.580906629562378, "perplexity": 35.90607960777423, "lr": 0.0026291804804649314, "grad_norm": 0.179453, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:55.055549+00:00", "epoch": 0, "step": 20164, "train_loss": 3.3992702960968018, "perplexity": 29.942243102155576, "lr": 0.0026291804804649314, "grad_norm": 0.156247, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:55.360161+00:00", "epoch": 0, "step": 20165, "train_loss": 3.464478015899658, "perplexity": 31.95977292505732, "lr": 0.0026291804804649314, "grad_norm": 0.180064, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:55.665562+00:00", "epoch": 0, "step": 20166, "train_loss": 3.49552845954895, "perplexity": 32.96770544912975, "lr": 0.0026291804804649314, "grad_norm": 0.177133, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:55.971087+00:00", "epoch": 0, "step": 20167, "train_loss": 3.5171291828155518, "perplexity": 33.68757863164646, "lr": 0.0026291804804649314, "grad_norm": 0.175763, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:56.275077+00:00", "epoch": 0, "step": 20168, "train_loss": 3.4966020584106445, "perplexity": 33.00311854650166, "lr": 0.0026291804804649314, "grad_norm": 0.165285, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:56.579573+00:00", "epoch": 0, "step": 20169, "train_loss": 3.5079562664031982, "perplexity": 33.379978241905306, "lr": 0.0026291804804649314, "grad_norm": 0.184771, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:56.883931+00:00", "epoch": 0, "step": 20170, "train_loss": 3.4686524868011475, "perplexity": 32.0934669238568, "lr": 0.0026291804804649314, "grad_norm": 0.142643, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:57.189439+00:00", "epoch": 0, "step": 20171, "train_loss": 3.41282320022583, "perplexity": 30.350809831366, "lr": 0.0026291804804649314, "grad_norm": 0.188465, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:57.494169+00:00", "epoch": 0, "step": 20172, "train_loss": 3.5469889640808105, "perplexity": 34.708650994544236, "lr": 0.0026291804804649314, "grad_norm": 0.156921, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:57.799300+00:00", "epoch": 0, "step": 20173, "train_loss": 3.599247455596924, "perplexity": 36.57070300779822, "lr": 0.0026291804804649314, "grad_norm": 0.164978, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:58.103962+00:00", "epoch": 0, "step": 20174, "train_loss": 3.516733169555664, "perplexity": 33.67424054501871, "lr": 0.0026291804804649314, "grad_norm": 0.168243, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:58.408366+00:00", "epoch": 0, "step": 20175, "train_loss": 3.466991424560547, "perplexity": 32.04020192826316, "lr": 0.0026291804804649314, "grad_norm": 0.182004, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:58.712220+00:00", "epoch": 0, "step": 20176, "train_loss": 3.5033814907073975, "perplexity": 33.227621094323844, "lr": 0.0026291804804649314, "grad_norm": 0.17203, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:59.016840+00:00", "epoch": 0, "step": 20177, "train_loss": 3.528921127319336, "perplexity": 34.08717205077238, "lr": 0.0026291804804649314, "grad_norm": 0.163664, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:59.321004+00:00", "epoch": 0, "step": 20178, "train_loss": 3.4859931468963623, "perplexity": 32.6548420658375, "lr": 0.0026291804804649314, "grad_norm": 0.183696, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:59.624697+00:00", "epoch": 0, "step": 20179, "train_loss": 3.492499589920044, "perplexity": 32.86800163850608, "lr": 0.0026291804804649314, "grad_norm": 0.198166, "tokens_per_sec": 107899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:43:59.929234+00:00", "epoch": 0, "step": 20180, "train_loss": 3.418440341949463, "perplexity": 30.521774348049302, "lr": 0.0026291804804649314, "grad_norm": 0.192747, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:00.234005+00:00", "epoch": 0, "step": 20181, "train_loss": 3.5350801944732666, "perplexity": 34.29776509515072, "lr": 0.0026291804804649314, "grad_norm": 0.168005, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:00.538953+00:00", "epoch": 0, "step": 20182, "train_loss": 3.6276917457580566, "perplexity": 37.62586623864579, "lr": 0.0026291804804649314, "grad_norm": 0.175388, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:00.844377+00:00", "epoch": 0, "step": 20183, "train_loss": 3.4954628944396973, "perplexity": 32.96554398877913, "lr": 0.0026291804804649314, "grad_norm": 0.161449, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:01.147903+00:00", "epoch": 0, "step": 20184, "train_loss": 3.4737114906311035, "perplexity": 32.256239281772565, "lr": 0.0026291804804649314, "grad_norm": 0.179722, "tokens_per_sec": 107959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:01.452387+00:00", "epoch": 0, "step": 20185, "train_loss": 3.4631638526916504, "perplexity": 31.917800152920034, "lr": 0.0026291804804649314, "grad_norm": 0.154453, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:01.758009+00:00", "epoch": 0, "step": 20186, "train_loss": 3.458425283432007, "perplexity": 31.766913222699547, "lr": 0.0026291804804649314, "grad_norm": 0.167274, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:02.062402+00:00", "epoch": 0, "step": 20187, "train_loss": 3.4534196853637695, "perplexity": 31.60829813611197, "lr": 0.0026291804804649314, "grad_norm": 0.1749, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:02.367317+00:00", "epoch": 0, "step": 20188, "train_loss": 3.4872570037841797, "perplexity": 32.6961392042423, "lr": 0.0026291804804649314, "grad_norm": 0.176444, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:02.671336+00:00", "epoch": 0, "step": 20189, "train_loss": 3.382673501968384, "perplexity": 29.44939898672751, "lr": 0.0026291804804649314, "grad_norm": 0.175919, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:02.975585+00:00", "epoch": 0, "step": 20190, "train_loss": 3.542844533920288, "perplexity": 34.56510108628738, "lr": 0.0026291804804649314, "grad_norm": 0.159338, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:03.280380+00:00", "epoch": 0, "step": 20191, "train_loss": 3.409026861190796, "perplexity": 30.235806301573973, "lr": 0.0026291804804649314, "grad_norm": 0.170768, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:03.585015+00:00", "epoch": 0, "step": 20192, "train_loss": 3.506192922592163, "perplexity": 33.32116972891184, "lr": 0.0026291804804649314, "grad_norm": 0.162334, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:03.890681+00:00", "epoch": 0, "step": 20193, "train_loss": 3.510610342025757, "perplexity": 33.46868889876058, "lr": 0.0026291804804649314, "grad_norm": 0.175537, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:04.195412+00:00", "epoch": 0, "step": 20194, "train_loss": 3.48968505859375, "perplexity": 32.77562367957312, "lr": 0.0026291804804649314, "grad_norm": 0.189509, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:04.499907+00:00", "epoch": 0, "step": 20195, "train_loss": 3.481736660003662, "perplexity": 32.516142554336824, "lr": 0.0026291804804649314, "grad_norm": 0.182538, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:04.805287+00:00", "epoch": 0, "step": 20196, "train_loss": 3.4639008045196533, "perplexity": 31.94133070346473, "lr": 0.0026291804804649314, "grad_norm": 0.180492, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:05.110095+00:00", "epoch": 0, "step": 20197, "train_loss": 3.501932382583618, "perplexity": 33.17950554939559, "lr": 0.0026291804804649314, "grad_norm": 0.220696, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:05.414686+00:00", "epoch": 0, "step": 20198, "train_loss": 3.4980435371398926, "perplexity": 33.05072614431159, "lr": 0.0026291804804649314, "grad_norm": 0.226039, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:05.719470+00:00", "epoch": 0, "step": 20199, "train_loss": 3.547092914581299, "perplexity": 34.71225916371864, "lr": 0.0026291804804649314, "grad_norm": 0.182317, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:06.024424+00:00", "epoch": 0, "step": 20200, "train_loss": 3.4510159492492676, "perplexity": 31.53241137078261, "lr": 0.0026291804804649314, "grad_norm": 0.168533, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:06.329284+00:00", "epoch": 0, "step": 20201, "train_loss": 3.4874939918518066, "perplexity": 32.70388871732597, "lr": 0.0026291804804649314, "grad_norm": 0.196968, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:06.634365+00:00", "epoch": 0, "step": 20202, "train_loss": 3.557377815246582, "perplexity": 35.07111352856173, "lr": 0.0026291804804649314, "grad_norm": 0.180567, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:06.941548+00:00", "epoch": 0, "step": 20203, "train_loss": 3.485690116882324, "perplexity": 32.64494816773763, "lr": 0.0026291804804649314, "grad_norm": 0.181998, "tokens_per_sec": 106673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:07.246813+00:00", "epoch": 0, "step": 20204, "train_loss": 3.491790533065796, "perplexity": 32.84470461709129, "lr": 0.0026291804804649314, "grad_norm": 0.190746, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:07.550646+00:00", "epoch": 0, "step": 20205, "train_loss": 3.501821756362915, "perplexity": 33.17583522911244, "lr": 0.0026291804804649314, "grad_norm": 0.183012, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:07.854763+00:00", "epoch": 0, "step": 20206, "train_loss": 3.639983654022217, "perplexity": 38.09121408217105, "lr": 0.0026291804804649314, "grad_norm": 0.184937, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:08.158877+00:00", "epoch": 0, "step": 20207, "train_loss": 3.454742431640625, "perplexity": 31.650135558836723, "lr": 0.0026291804804649314, "grad_norm": 0.177575, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:08.463179+00:00", "epoch": 0, "step": 20208, "train_loss": 3.510744094848633, "perplexity": 33.473165729766464, "lr": 0.0026291804804649314, "grad_norm": 0.223326, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:08.768313+00:00", "epoch": 0, "step": 20209, "train_loss": 3.474565267562866, "perplexity": 32.28379067447492, "lr": 0.0026291804804649314, "grad_norm": 0.190495, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:09.073218+00:00", "epoch": 0, "step": 20210, "train_loss": 3.5003268718719482, "perplexity": 33.126278237771125, "lr": 0.0026291804804649314, "grad_norm": 0.16826, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:09.378046+00:00", "epoch": 0, "step": 20211, "train_loss": 3.525193452835083, "perplexity": 33.96034270539935, "lr": 0.0026291804804649314, "grad_norm": 0.180777, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:09.682680+00:00", "epoch": 0, "step": 20212, "train_loss": 3.4647727012634277, "perplexity": 31.96919239018882, "lr": 0.0026291804804649314, "grad_norm": 0.176145, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:09.987186+00:00", "epoch": 0, "step": 20213, "train_loss": 3.499556541442871, "perplexity": 33.100769883834516, "lr": 0.0026291804804649314, "grad_norm": 0.168993, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:10.292944+00:00", "epoch": 0, "step": 20214, "train_loss": 3.583197593688965, "perplexity": 35.98843344690796, "lr": 0.0026291804804649314, "grad_norm": 0.185056, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:10.598431+00:00", "epoch": 0, "step": 20215, "train_loss": 3.477823495864868, "perplexity": 32.389150184406496, "lr": 0.0026291804804649314, "grad_norm": 0.174559, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:10.902665+00:00", "epoch": 0, "step": 20216, "train_loss": 3.4936866760253906, "perplexity": 32.9070419540852, "lr": 0.0026291804804649314, "grad_norm": 0.187121, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:11.207008+00:00", "epoch": 0, "step": 20217, "train_loss": 3.523782968521118, "perplexity": 33.912475940312945, "lr": 0.0026291804804649314, "grad_norm": 0.186242, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:11.512153+00:00", "epoch": 0, "step": 20218, "train_loss": 3.443849802017212, "perplexity": 31.30725319021142, "lr": 0.0026291804804649314, "grad_norm": 0.161401, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:11.817689+00:00", "epoch": 0, "step": 20219, "train_loss": 3.565892219543457, "perplexity": 35.37099802553326, "lr": 0.0026291804804649314, "grad_norm": 0.160044, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:12.123611+00:00", "epoch": 0, "step": 20220, "train_loss": 3.5692620277404785, "perplexity": 35.49039255997385, "lr": 0.0026291804804649314, "grad_norm": 0.172677, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:12.427347+00:00", "epoch": 0, "step": 20221, "train_loss": 3.4867937564849854, "perplexity": 32.680996313784206, "lr": 0.0026291804804649314, "grad_norm": 0.172081, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:12.732001+00:00", "epoch": 0, "step": 20222, "train_loss": 3.4844253063201904, "perplexity": 32.60368459329809, "lr": 0.0026291804804649314, "grad_norm": 0.16046, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:13.036146+00:00", "epoch": 0, "step": 20223, "train_loss": 3.413111448287964, "perplexity": 30.35955965448328, "lr": 0.0026291804804649314, "grad_norm": 0.15937, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:13.341611+00:00", "epoch": 0, "step": 20224, "train_loss": 3.4696943759918213, "perplexity": 32.12692218544733, "lr": 0.0026291804804649314, "grad_norm": 0.186892, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:13.647276+00:00", "epoch": 0, "step": 20225, "train_loss": 3.4634108543395996, "perplexity": 31.925684875886414, "lr": 0.0026291804804649314, "grad_norm": 0.179881, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:13.952841+00:00", "epoch": 0, "step": 20226, "train_loss": 3.5392391681671143, "perplexity": 34.440705634873915, "lr": 0.0026291804804649314, "grad_norm": 0.160842, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:14.257588+00:00", "epoch": 0, "step": 20227, "train_loss": 3.5430545806884766, "perplexity": 34.572362136616036, "lr": 0.0026291804804649314, "grad_norm": 0.182166, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:14.562205+00:00", "epoch": 0, "step": 20228, "train_loss": 3.516775131225586, "perplexity": 33.67565360203224, "lr": 0.0026291804804649314, "grad_norm": 0.161884, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:14.866192+00:00", "epoch": 0, "step": 20229, "train_loss": 3.5137908458709717, "perplexity": 33.57530565011682, "lr": 0.0026291804804649314, "grad_norm": 0.186738, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:15.171876+00:00", "epoch": 0, "step": 20230, "train_loss": 3.4383928775787354, "perplexity": 31.13687716251905, "lr": 0.0026291804804649314, "grad_norm": 0.170519, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:15.477014+00:00", "epoch": 0, "step": 20231, "train_loss": 3.513108730316162, "perplexity": 33.55241122108617, "lr": 0.0026291804804649314, "grad_norm": 0.170756, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:15.782535+00:00", "epoch": 0, "step": 20232, "train_loss": 3.508197784423828, "perplexity": 33.38804108180034, "lr": 0.0026291804804649314, "grad_norm": 0.179999, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:16.087085+00:00", "epoch": 0, "step": 20233, "train_loss": 3.465707778930664, "perplexity": 31.999100048845115, "lr": 0.0026291804804649314, "grad_norm": 0.18285, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:16.392847+00:00", "epoch": 0, "step": 20234, "train_loss": 3.603450059890747, "perplexity": 36.72471860791439, "lr": 0.0026291804804649314, "grad_norm": 0.168185, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:16.698475+00:00", "epoch": 0, "step": 20235, "train_loss": 3.510524272918701, "perplexity": 33.46580840255539, "lr": 0.0026291804804649314, "grad_norm": 0.194636, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:17.002954+00:00", "epoch": 0, "step": 20236, "train_loss": 3.5591607093811035, "perplexity": 35.13369738478002, "lr": 0.0026291804804649314, "grad_norm": 0.152321, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:17.307234+00:00", "epoch": 0, "step": 20237, "train_loss": 3.5449390411376953, "perplexity": 34.63757381081072, "lr": 0.0026291804804649314, "grad_norm": 0.175608, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:17.611875+00:00", "epoch": 0, "step": 20238, "train_loss": 3.451143980026245, "perplexity": 31.536448748359167, "lr": 0.0026291804804649314, "grad_norm": 0.174822, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:17.916861+00:00", "epoch": 0, "step": 20239, "train_loss": 3.548158645629883, "perplexity": 34.749272815871784, "lr": 0.0026291804804649314, "grad_norm": 0.175475, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:18.223143+00:00", "epoch": 0, "step": 20240, "train_loss": 3.5257997512817383, "perplexity": 33.980939051564924, "lr": 0.0026291804804649314, "grad_norm": 0.184898, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:18.529014+00:00", "epoch": 0, "step": 20241, "train_loss": 3.614407539367676, "perplexity": 37.1293417346998, "lr": 0.0026291804804649314, "grad_norm": 0.179628, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:18.833824+00:00", "epoch": 0, "step": 20242, "train_loss": 3.5324630737304688, "perplexity": 34.20812105848923, "lr": 0.0026291804804649314, "grad_norm": 0.169656, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:19.139282+00:00", "epoch": 0, "step": 20243, "train_loss": 3.5230164527893066, "perplexity": 33.88649145403103, "lr": 0.0026291804804649314, "grad_norm": 0.166053, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:19.443367+00:00", "epoch": 0, "step": 20244, "train_loss": 3.57841157913208, "perplexity": 35.81660379814483, "lr": 0.0026291804804649314, "grad_norm": 0.172774, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:19.748429+00:00", "epoch": 0, "step": 20245, "train_loss": 3.4339232444763184, "perplexity": 30.998017303133665, "lr": 0.0026291804804649314, "grad_norm": 0.153312, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:20.054471+00:00", "epoch": 0, "step": 20246, "train_loss": 3.466477870941162, "perplexity": 32.02375179097239, "lr": 0.0026291804804649314, "grad_norm": 0.162308, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:20.359193+00:00", "epoch": 0, "step": 20247, "train_loss": 3.4746344089508057, "perplexity": 32.286022897738725, "lr": 0.0026291804804649314, "grad_norm": 0.159567, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:20.664605+00:00", "epoch": 0, "step": 20248, "train_loss": 3.559171438217163, "perplexity": 35.134074330481525, "lr": 0.0026291804804649314, "grad_norm": 0.163699, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:20.970192+00:00", "epoch": 0, "step": 20249, "train_loss": 3.4916324615478516, "perplexity": 32.83951321509324, "lr": 0.0026291804804649314, "grad_norm": 0.189584, "tokens_per_sec": 107229} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:21.275312+00:00", "epoch": 0, "step": 20250, "train_loss": 3.3213486671447754, "perplexity": 27.697680331588955, "lr": 0.0026291804804649314, "grad_norm": 0.192556, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:21.579909+00:00", "epoch": 0, "step": 20251, "train_loss": 3.63832688331604, "perplexity": 38.02815792373858, "lr": 0.0026291804804649314, "grad_norm": 0.185184, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:21.884368+00:00", "epoch": 0, "step": 20252, "train_loss": 3.5855369567871094, "perplexity": 36.07272201241362, "lr": 0.0026291804804649314, "grad_norm": 0.15421, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:22.188865+00:00", "epoch": 0, "step": 20253, "train_loss": 3.4479639530181885, "perplexity": 31.436321278010006, "lr": 0.0026291804804649314, "grad_norm": 0.152475, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:22.493918+00:00", "epoch": 0, "step": 20254, "train_loss": 3.437540054321289, "perplexity": 31.110334229334715, "lr": 0.0026291804804649314, "grad_norm": 0.173125, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:22.798630+00:00", "epoch": 0, "step": 20255, "train_loss": 3.4372880458831787, "perplexity": 31.102495150394997, "lr": 0.0026291804804649314, "grad_norm": 0.144168, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:23.103900+00:00", "epoch": 0, "step": 20256, "train_loss": 3.5092787742614746, "perplexity": 33.424152729554315, "lr": 0.0026291804804649314, "grad_norm": 0.192571, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:23.408791+00:00", "epoch": 0, "step": 20257, "train_loss": 3.335724353790283, "perplexity": 28.09872927509431, "lr": 0.0026291804804649314, "grad_norm": 0.17679, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:23.713861+00:00", "epoch": 0, "step": 20258, "train_loss": 3.544651746749878, "perplexity": 34.62762405956955, "lr": 0.0026291804804649314, "grad_norm": 0.176173, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:24.018202+00:00", "epoch": 0, "step": 20259, "train_loss": 3.4219655990600586, "perplexity": 30.629561326878225, "lr": 0.0026291804804649314, "grad_norm": 0.230307, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:24.322387+00:00", "epoch": 0, "step": 20260, "train_loss": 3.610806465148926, "perplexity": 36.99587667244548, "lr": 0.0026291804804649314, "grad_norm": 0.196709, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:24.627070+00:00", "epoch": 0, "step": 20261, "train_loss": 3.4588871002197266, "perplexity": 31.781587104581526, "lr": 0.0026291804804649314, "grad_norm": 0.14771, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:24.932420+00:00", "epoch": 0, "step": 20262, "train_loss": 3.478100299835205, "perplexity": 32.39811687072331, "lr": 0.0026291804804649314, "grad_norm": 0.169988, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:25.238297+00:00", "epoch": 0, "step": 20263, "train_loss": 3.640423059463501, "perplexity": 38.107955246713914, "lr": 0.0026291804804649314, "grad_norm": 0.169054, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:25.543422+00:00", "epoch": 0, "step": 20264, "train_loss": 3.487745761871338, "perplexity": 32.712123612633455, "lr": 0.0026291804804649314, "grad_norm": 0.150114, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:25.848112+00:00", "epoch": 0, "step": 20265, "train_loss": 3.6350655555725098, "perplexity": 37.904337656252004, "lr": 0.0026291804804649314, "grad_norm": 0.156646, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:26.153159+00:00", "epoch": 0, "step": 20266, "train_loss": 3.5049750804901123, "perplexity": 33.2806145054681, "lr": 0.0026291804804649314, "grad_norm": 0.160484, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:26.458920+00:00", "epoch": 0, "step": 20267, "train_loss": 3.4400124549865723, "perplexity": 31.187346603873653, "lr": 0.0026291804804649314, "grad_norm": 0.16335, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:26.764590+00:00", "epoch": 0, "step": 20268, "train_loss": 3.4929397106170654, "perplexity": 32.8824707101323, "lr": 0.0026291804804649314, "grad_norm": 0.184297, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:27.069957+00:00", "epoch": 0, "step": 20269, "train_loss": 3.5461795330047607, "perplexity": 34.68056810093816, "lr": 0.0026291804804649314, "grad_norm": 0.186055, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:27.374551+00:00", "epoch": 0, "step": 20270, "train_loss": 3.5115249156951904, "perplexity": 33.49931248200472, "lr": 0.0026291804804649314, "grad_norm": 0.163629, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:27.679022+00:00", "epoch": 0, "step": 20271, "train_loss": 3.5333704948425293, "perplexity": 34.239176317709614, "lr": 0.0026291804804649314, "grad_norm": 0.179657, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:27.983996+00:00", "epoch": 0, "step": 20272, "train_loss": 3.5119564533233643, "perplexity": 33.51377181550751, "lr": 0.0026291804804649314, "grad_norm": 0.16485, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:28.289878+00:00", "epoch": 0, "step": 20273, "train_loss": 3.5485646724700928, "perplexity": 34.763384818044756, "lr": 0.0026291804804649314, "grad_norm": 0.16284, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:28.594411+00:00", "epoch": 0, "step": 20274, "train_loss": 3.467235565185547, "perplexity": 32.04802519813719, "lr": 0.0026291804804649314, "grad_norm": 0.17082, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:28.899237+00:00", "epoch": 0, "step": 20275, "train_loss": 3.470477819442749, "perplexity": 32.1521016742969, "lr": 0.0026291804804649314, "grad_norm": 0.183595, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:29.203756+00:00", "epoch": 0, "step": 20276, "train_loss": 3.459024429321289, "perplexity": 31.785951941087504, "lr": 0.0026291804804649314, "grad_norm": 0.188392, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:29.508122+00:00", "epoch": 0, "step": 20277, "train_loss": 3.5690927505493164, "perplexity": 35.48438535446384, "lr": 0.0026291804804649314, "grad_norm": 0.136777, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:29.814264+00:00", "epoch": 0, "step": 20278, "train_loss": 3.4223170280456543, "perplexity": 30.640327334177282, "lr": 0.0026291804804649314, "grad_norm": 0.182189, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:30.120441+00:00", "epoch": 0, "step": 20279, "train_loss": 3.3955376148223877, "perplexity": 29.830686584002923, "lr": 0.0026291804804649314, "grad_norm": 0.168721, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:30.425119+00:00", "epoch": 0, "step": 20280, "train_loss": 3.4904446601867676, "perplexity": 32.800529553605514, "lr": 0.0026291804804649314, "grad_norm": 0.161324, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:30.730251+00:00", "epoch": 0, "step": 20281, "train_loss": 3.4311654567718506, "perplexity": 30.91264911992009, "lr": 0.0026291804804649314, "grad_norm": 0.172278, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:31.035108+00:00", "epoch": 0, "step": 20282, "train_loss": 3.4614291191101074, "perplexity": 31.862479270515117, "lr": 0.0026291804804649314, "grad_norm": 0.187902, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:31.340102+00:00", "epoch": 0, "step": 20283, "train_loss": 3.4997804164886475, "perplexity": 33.10818114977529, "lr": 0.0026291804804649314, "grad_norm": 0.153278, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:31.645911+00:00", "epoch": 0, "step": 20284, "train_loss": 3.494438886642456, "perplexity": 32.93180429249862, "lr": 0.0026291804804649314, "grad_norm": 0.179964, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:31.950714+00:00", "epoch": 0, "step": 20285, "train_loss": 3.442145824432373, "perplexity": 31.253951757652484, "lr": 0.0026291804804649314, "grad_norm": 0.193913, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:32.254996+00:00", "epoch": 0, "step": 20286, "train_loss": 3.603731155395508, "perplexity": 36.73504321226063, "lr": 0.0026291804804649314, "grad_norm": 0.211369, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:32.559987+00:00", "epoch": 0, "step": 20287, "train_loss": 3.43738055229187, "perplexity": 31.105372463605605, "lr": 0.0026291804804649314, "grad_norm": 0.196637, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:32.865503+00:00", "epoch": 0, "step": 20288, "train_loss": 3.505089282989502, "perplexity": 33.284415451860504, "lr": 0.0026291804804649314, "grad_norm": 0.178213, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:33.171820+00:00", "epoch": 0, "step": 20289, "train_loss": 3.4458298683166504, "perplexity": 31.369305040296233, "lr": 0.0026291804804649314, "grad_norm": 0.190998, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:33.476851+00:00", "epoch": 0, "step": 20290, "train_loss": 3.534497022628784, "perplexity": 34.27776943523819, "lr": 0.0026291804804649314, "grad_norm": 0.188829, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:33.781006+00:00", "epoch": 0, "step": 20291, "train_loss": 3.5190703868865967, "perplexity": 33.753036609521125, "lr": 0.0026291804804649314, "grad_norm": 0.169909, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:34.085330+00:00", "epoch": 0, "step": 20292, "train_loss": 3.4897520542144775, "perplexity": 32.77781957638318, "lr": 0.0026291804804649314, "grad_norm": 0.209635, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:34.388822+00:00", "epoch": 0, "step": 20293, "train_loss": 3.4850471019744873, "perplexity": 32.62396372677743, "lr": 0.0026291804804649314, "grad_norm": 0.192782, "tokens_per_sec": 107970} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:34.694735+00:00", "epoch": 0, "step": 20294, "train_loss": 3.4902849197387695, "perplexity": 32.7952904007835, "lr": 0.0026291804804649314, "grad_norm": 0.188852, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:35.000724+00:00", "epoch": 0, "step": 20295, "train_loss": 3.4957339763641357, "perplexity": 32.97448156323648, "lr": 0.0026291804804649314, "grad_norm": 0.184224, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:35.306780+00:00", "epoch": 0, "step": 20296, "train_loss": 3.471785545349121, "perplexity": 32.19417531499964, "lr": 0.0026291804804649314, "grad_norm": 0.187184, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:35.611548+00:00", "epoch": 0, "step": 20297, "train_loss": 3.5193541049957275, "perplexity": 33.76261431586553, "lr": 0.0026291804804649314, "grad_norm": 0.164046, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:35.915993+00:00", "epoch": 0, "step": 20298, "train_loss": 3.4529786109924316, "perplexity": 31.594359600074007, "lr": 0.0026291804804649314, "grad_norm": 0.185852, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:36.221274+00:00", "epoch": 0, "step": 20299, "train_loss": 3.447695016860962, "perplexity": 31.427868051308263, "lr": 0.0026291804804649314, "grad_norm": 0.181761, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:36.526845+00:00", "epoch": 0, "step": 20300, "train_loss": 3.5179622173309326, "perplexity": 33.715653239321306, "lr": 0.0026291804804649314, "grad_norm": 0.205899, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:36.831418+00:00", "epoch": 0, "step": 20301, "train_loss": 3.5346665382385254, "perplexity": 34.283580544747544, "lr": 0.0026291804804649314, "grad_norm": 0.182596, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:37.134642+00:00", "epoch": 0, "step": 20302, "train_loss": 3.4907848834991455, "perplexity": 32.81169095699515, "lr": 0.0026291804804649314, "grad_norm": 0.173684, "tokens_per_sec": 108067} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:37.438984+00:00", "epoch": 0, "step": 20303, "train_loss": 3.5213871002197266, "perplexity": 33.83132336844233, "lr": 0.0026291804804649314, "grad_norm": 0.193308, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:37.743801+00:00", "epoch": 0, "step": 20304, "train_loss": 3.5314700603485107, "perplexity": 34.17416879682156, "lr": 0.0026291804804649314, "grad_norm": 0.162028, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:38.049450+00:00", "epoch": 0, "step": 20305, "train_loss": 3.549891471862793, "perplexity": 34.80953946811951, "lr": 0.0026291804804649314, "grad_norm": 0.173029, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:38.354151+00:00", "epoch": 0, "step": 20306, "train_loss": 3.6229076385498047, "perplexity": 37.4462899593498, "lr": 0.0026291804804649314, "grad_norm": 0.199046, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:38.658121+00:00", "epoch": 0, "step": 20307, "train_loss": 3.5776147842407227, "perplexity": 35.78807667785384, "lr": 0.0026291804804649314, "grad_norm": 0.174491, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:38.963897+00:00", "epoch": 0, "step": 20308, "train_loss": 3.5231947898864746, "perplexity": 33.8925352114472, "lr": 0.0026291804804649314, "grad_norm": 0.175065, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:39.268857+00:00", "epoch": 0, "step": 20309, "train_loss": 3.6243131160736084, "perplexity": 37.498956880654866, "lr": 0.0026291804804649314, "grad_norm": 0.153741, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:39.575365+00:00", "epoch": 0, "step": 20310, "train_loss": 3.397585391998291, "perplexity": 29.891835771718096, "lr": 0.0026291804804649314, "grad_norm": 0.183459, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:39.881212+00:00", "epoch": 0, "step": 20311, "train_loss": 3.5106863975524902, "perplexity": 33.47123447432517, "lr": 0.0026291804804649314, "grad_norm": 0.196583, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:40.187041+00:00", "epoch": 0, "step": 20312, "train_loss": 3.4299044609069824, "perplexity": 30.87369296414547, "lr": 0.0026291804804649314, "grad_norm": 0.159585, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:40.490993+00:00", "epoch": 0, "step": 20313, "train_loss": 3.446587562561035, "perplexity": 31.393082389018204, "lr": 0.0026291804804649314, "grad_norm": 0.191094, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:40.795242+00:00", "epoch": 0, "step": 20314, "train_loss": 3.506925344467163, "perplexity": 33.34558382213846, "lr": 0.0026291804804649314, "grad_norm": 0.177864, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:41.100091+00:00", "epoch": 0, "step": 20315, "train_loss": 3.418318033218384, "perplexity": 30.51804149684331, "lr": 0.0026291804804649314, "grad_norm": 0.186421, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:41.404940+00:00", "epoch": 0, "step": 20316, "train_loss": 3.523858070373535, "perplexity": 33.91502292571657, "lr": 0.0026291804804649314, "grad_norm": 0.187018, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:41.710827+00:00", "epoch": 0, "step": 20317, "train_loss": 3.367353916168213, "perplexity": 29.00168454819444, "lr": 0.0026291804804649314, "grad_norm": 0.173817, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:42.015560+00:00", "epoch": 0, "step": 20318, "train_loss": 3.4773108959198, "perplexity": 32.372551762339725, "lr": 0.0026291804804649314, "grad_norm": 0.212846, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:42.319292+00:00", "epoch": 0, "step": 20319, "train_loss": 3.50753116607666, "perplexity": 33.365791417880246, "lr": 0.0026291804804649314, "grad_norm": 0.204819, "tokens_per_sec": 107884} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:42.624725+00:00", "epoch": 0, "step": 20320, "train_loss": 3.4796969890594482, "perplexity": 32.44988791495196, "lr": 0.0026291804804649314, "grad_norm": 0.168136, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:42.930123+00:00", "epoch": 0, "step": 20321, "train_loss": 3.513207197189331, "perplexity": 33.5557151847692, "lr": 0.0026291804804649314, "grad_norm": 0.179552, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:43.234372+00:00", "epoch": 0, "step": 20322, "train_loss": 3.517366647720337, "perplexity": 33.695579199188074, "lr": 0.0026291804804649314, "grad_norm": 0.180153, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:43.540013+00:00", "epoch": 0, "step": 20323, "train_loss": 3.2512166500091553, "perplexity": 25.82173683014374, "lr": 0.0026291804804649314, "grad_norm": 0.154332, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:43.844646+00:00", "epoch": 0, "step": 20324, "train_loss": 3.480363130569458, "perplexity": 32.47151133361605, "lr": 0.0026291804804649314, "grad_norm": 0.15973, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:44.149347+00:00", "epoch": 0, "step": 20325, "train_loss": 3.591512441635132, "perplexity": 36.28891931477571, "lr": 0.0026291804804649314, "grad_norm": 0.192204, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:44.455604+00:00", "epoch": 0, "step": 20326, "train_loss": 3.4968764781951904, "perplexity": 33.01217649796623, "lr": 0.0026291804804649314, "grad_norm": 0.200858, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:44.762067+00:00", "epoch": 0, "step": 20327, "train_loss": 3.5330777168273926, "perplexity": 34.22915330695759, "lr": 0.0026291804804649314, "grad_norm": 0.173662, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:45.068281+00:00", "epoch": 0, "step": 20328, "train_loss": 3.5266313552856445, "perplexity": 34.00920948981228, "lr": 0.0026291804804649314, "grad_norm": 0.179166, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:45.374018+00:00", "epoch": 0, "step": 20329, "train_loss": 3.440248727798462, "perplexity": 31.194716196534035, "lr": 0.0026291804804649314, "grad_norm": 0.205662, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:45.679116+00:00", "epoch": 0, "step": 20330, "train_loss": 3.4188737869262695, "perplexity": 30.535006725370263, "lr": 0.0026291804804649314, "grad_norm": 0.217884, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:45.984577+00:00", "epoch": 0, "step": 20331, "train_loss": 3.511866807937622, "perplexity": 33.510767595164694, "lr": 0.0026291804804649314, "grad_norm": 0.183155, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:46.291126+00:00", "epoch": 0, "step": 20332, "train_loss": 3.4078292846679688, "perplexity": 30.199618283084174, "lr": 0.0026291804804649314, "grad_norm": 0.179251, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:46.597371+00:00", "epoch": 0, "step": 20333, "train_loss": 3.4990878105163574, "perplexity": 33.085258164993725, "lr": 0.0026291804804649314, "grad_norm": 0.161495, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:46.903424+00:00", "epoch": 0, "step": 20334, "train_loss": 3.5426137447357178, "perplexity": 34.55712475525395, "lr": 0.0026291804804649314, "grad_norm": 0.169707, "tokens_per_sec": 107075} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:47.208774+00:00", "epoch": 0, "step": 20335, "train_loss": 3.5424234867095947, "perplexity": 34.55055061032128, "lr": 0.0026291804804649314, "grad_norm": 0.175568, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:47.513670+00:00", "epoch": 0, "step": 20336, "train_loss": 3.638334035873413, "perplexity": 38.02842992329267, "lr": 0.0026291804804649314, "grad_norm": 0.166091, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:47.819649+00:00", "epoch": 0, "step": 20337, "train_loss": 3.4745712280273438, "perplexity": 32.283983101435915, "lr": 0.0026291804804649314, "grad_norm": 0.176765, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:48.126031+00:00", "epoch": 0, "step": 20338, "train_loss": 3.613570213317871, "perplexity": 37.09826538198874, "lr": 0.0026291804804649314, "grad_norm": 0.186487, "tokens_per_sec": 106952} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:48.431268+00:00", "epoch": 0, "step": 20339, "train_loss": 3.513235330581665, "perplexity": 33.55665923414914, "lr": 0.0026291804804649314, "grad_norm": 0.179219, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:48.736687+00:00", "epoch": 0, "step": 20340, "train_loss": 3.414824962615967, "perplexity": 30.41162579023849, "lr": 0.0026291804804649314, "grad_norm": 0.14957, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:49.041187+00:00", "epoch": 0, "step": 20341, "train_loss": 3.4981307983398438, "perplexity": 33.05361031617052, "lr": 0.0026291804804649314, "grad_norm": 0.171554, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:49.345853+00:00", "epoch": 0, "step": 20342, "train_loss": 3.506216287612915, "perplexity": 33.32194828782953, "lr": 0.0026291804804649314, "grad_norm": 0.167993, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:49.651202+00:00", "epoch": 0, "step": 20343, "train_loss": 3.4932010173797607, "perplexity": 32.891064244828456, "lr": 0.0026291804804649314, "grad_norm": 0.177265, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:49.956874+00:00", "epoch": 0, "step": 20344, "train_loss": 3.4084315299987793, "perplexity": 30.21781133997932, "lr": 0.0026291804804649314, "grad_norm": 0.186225, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:50.261459+00:00", "epoch": 0, "step": 20345, "train_loss": 3.508023738861084, "perplexity": 33.38223054706489, "lr": 0.0026291804804649314, "grad_norm": 0.179213, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:50.566860+00:00", "epoch": 0, "step": 20346, "train_loss": 3.5457346439361572, "perplexity": 34.66514252688616, "lr": 0.0026291804804649314, "grad_norm": 0.190268, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:50.871175+00:00", "epoch": 0, "step": 20347, "train_loss": 3.5223586559295654, "perplexity": 33.864208356044344, "lr": 0.0026291804804649314, "grad_norm": 0.179404, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:51.177484+00:00", "epoch": 0, "step": 20348, "train_loss": 3.4387149810791016, "perplexity": 31.146908075052863, "lr": 0.0026291804804649314, "grad_norm": 0.186764, "tokens_per_sec": 106977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:51.483493+00:00", "epoch": 0, "step": 20349, "train_loss": 3.4383394718170166, "perplexity": 31.13521431827977, "lr": 0.0026291804804649314, "grad_norm": 0.176478, "tokens_per_sec": 107083} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:51.789581+00:00", "epoch": 0, "step": 20350, "train_loss": 3.3197555541992188, "perplexity": 27.65358992831208, "lr": 0.0026291804804649314, "grad_norm": 0.180812, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:52.094302+00:00", "epoch": 0, "step": 20351, "train_loss": 3.558765172958374, "perplexity": 35.11980347575386, "lr": 0.0026291804804649314, "grad_norm": 0.174098, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:52.400915+00:00", "epoch": 0, "step": 20352, "train_loss": 3.404050827026367, "perplexity": 30.085725609431876, "lr": 0.0026291804804649314, "grad_norm": 0.170662, "tokens_per_sec": 106871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:52.707354+00:00", "epoch": 0, "step": 20353, "train_loss": 3.476722478866577, "perplexity": 32.353508803976595, "lr": 0.0026291804804649314, "grad_norm": 0.160333, "tokens_per_sec": 106932} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:53.011165+00:00", "epoch": 0, "step": 20354, "train_loss": 3.4769840240478516, "perplexity": 32.36197181498323, "lr": 0.0026291804804649314, "grad_norm": 0.177999, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:53.315990+00:00", "epoch": 0, "step": 20355, "train_loss": 3.4403886795043945, "perplexity": 31.199082255793417, "lr": 0.0026291804804649314, "grad_norm": 0.189537, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:53.621684+00:00", "epoch": 0, "step": 20356, "train_loss": 3.4791579246520996, "perplexity": 32.432400049321004, "lr": 0.0026291804804649314, "grad_norm": 0.182885, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:53.928326+00:00", "epoch": 0, "step": 20357, "train_loss": 3.5653412342071533, "perplexity": 35.35151449235421, "lr": 0.0026291804804649314, "grad_norm": 0.162594, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:54.234743+00:00", "epoch": 0, "step": 20358, "train_loss": 3.4749112129211426, "perplexity": 32.29496103406194, "lr": 0.0026291804804649314, "grad_norm": 0.17012, "tokens_per_sec": 106938} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:54.540301+00:00", "epoch": 0, "step": 20359, "train_loss": 3.466230630874634, "perplexity": 32.0158352151368, "lr": 0.0026291804804649314, "grad_norm": 0.164216, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:54.846008+00:00", "epoch": 0, "step": 20360, "train_loss": 3.573209047317505, "perplexity": 35.63075065004966, "lr": 0.0026291804804649314, "grad_norm": 0.166962, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:55.151798+00:00", "epoch": 0, "step": 20361, "train_loss": 3.4436569213867188, "perplexity": 31.301215209800535, "lr": 0.0026291804804649314, "grad_norm": 0.171802, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:55.458355+00:00", "epoch": 0, "step": 20362, "train_loss": 3.510354995727539, "perplexity": 33.460143883959454, "lr": 0.0026291804804649314, "grad_norm": 0.165763, "tokens_per_sec": 106890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:55.763744+00:00", "epoch": 0, "step": 20363, "train_loss": 3.5210914611816406, "perplexity": 33.82132298686894, "lr": 0.0026291804804649314, "grad_norm": 0.196663, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:56.068021+00:00", "epoch": 0, "step": 20364, "train_loss": 3.3501229286193848, "perplexity": 28.506237660830376, "lr": 0.0026291804804649314, "grad_norm": 0.1739, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:56.373041+00:00", "epoch": 0, "step": 20365, "train_loss": 3.543846845626831, "perplexity": 34.59976346009362, "lr": 0.0026291804804649314, "grad_norm": 0.183647, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:56.678687+00:00", "epoch": 0, "step": 20366, "train_loss": 3.4264097213745117, "perplexity": 30.765985762758437, "lr": 0.0026291804804649314, "grad_norm": 0.171316, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:56.982964+00:00", "epoch": 0, "step": 20367, "train_loss": 3.499643325805664, "perplexity": 33.103642637710095, "lr": 0.0026291804804649314, "grad_norm": 0.161971, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:57.288078+00:00", "epoch": 0, "step": 20368, "train_loss": 3.4833662509918213, "perplexity": 32.56917376506896, "lr": 0.0026291804804649314, "grad_norm": 0.180589, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:57.592696+00:00", "epoch": 0, "step": 20369, "train_loss": 3.5174779891967773, "perplexity": 33.69933112359415, "lr": 0.0026291804804649314, "grad_norm": 0.172962, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:57.896911+00:00", "epoch": 0, "step": 20370, "train_loss": 3.6015982627868652, "perplexity": 36.65677480883636, "lr": 0.0026291804804649314, "grad_norm": 0.158745, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:58.201577+00:00", "epoch": 0, "step": 20371, "train_loss": 3.486650228500366, "perplexity": 32.676306012850716, "lr": 0.0026291804804649314, "grad_norm": 0.180292, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:58.505842+00:00", "epoch": 0, "step": 20372, "train_loss": 3.535590410232544, "perplexity": 34.31526882036878, "lr": 0.0026291804804649314, "grad_norm": 0.188607, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:58.810475+00:00", "epoch": 0, "step": 20373, "train_loss": 3.2953546047210693, "perplexity": 26.98698208462258, "lr": 0.0026291804804649314, "grad_norm": 0.166746, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:59.114451+00:00", "epoch": 0, "step": 20374, "train_loss": 3.472647190093994, "perplexity": 32.221927211396185, "lr": 0.0026291804804649314, "grad_norm": 0.174501, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:59.418994+00:00", "epoch": 0, "step": 20375, "train_loss": 3.541585922241211, "perplexity": 34.52162441223585, "lr": 0.0026291804804649314, "grad_norm": 0.182674, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:44:59.724589+00:00", "epoch": 0, "step": 20376, "train_loss": 3.4559497833251953, "perplexity": 31.688371480778244, "lr": 0.0026291804804649314, "grad_norm": 0.155277, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:00.030125+00:00", "epoch": 0, "step": 20377, "train_loss": 3.4222874641418457, "perplexity": 30.639421499877376, "lr": 0.0026291804804649314, "grad_norm": 0.15949, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:00.335648+00:00", "epoch": 0, "step": 20378, "train_loss": 3.466409683227539, "perplexity": 32.02156823900269, "lr": 0.0026291804804649314, "grad_norm": 0.165915, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:00.640831+00:00", "epoch": 0, "step": 20379, "train_loss": 3.5339579582214355, "perplexity": 34.259296489273524, "lr": 0.0026291804804649314, "grad_norm": 0.176865, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:00.945245+00:00", "epoch": 0, "step": 20380, "train_loss": 3.5868921279907227, "perplexity": 36.121639865066996, "lr": 0.0026291804804649314, "grad_norm": 0.186813, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:01.249969+00:00", "epoch": 0, "step": 20381, "train_loss": 3.522886037826538, "perplexity": 33.88207243666895, "lr": 0.0026291804804649314, "grad_norm": 0.197826, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:01.555729+00:00", "epoch": 0, "step": 20382, "train_loss": 3.456670045852661, "perplexity": 31.711203648902757, "lr": 0.0026291804804649314, "grad_norm": 0.210582, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:01.862159+00:00", "epoch": 0, "step": 20383, "train_loss": 3.5307765007019043, "perplexity": 34.15047518980239, "lr": 0.0026291804804649314, "grad_norm": 0.223576, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:02.167441+00:00", "epoch": 0, "step": 20384, "train_loss": 3.4973714351654053, "perplexity": 33.02852014919425, "lr": 0.0026291804804649314, "grad_norm": 0.195729, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:02.472342+00:00", "epoch": 0, "step": 20385, "train_loss": 3.4708218574523926, "perplexity": 32.16316512237699, "lr": 0.0026291804804649314, "grad_norm": 0.227084, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:02.776507+00:00", "epoch": 0, "step": 20386, "train_loss": 3.471299886703491, "perplexity": 32.178543731533054, "lr": 0.0026291804804649314, "grad_norm": 0.205793, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:03.081487+00:00", "epoch": 0, "step": 20387, "train_loss": 3.459198236465454, "perplexity": 31.791477046756462, "lr": 0.0026291804804649314, "grad_norm": 0.183148, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:03.387314+00:00", "epoch": 0, "step": 20388, "train_loss": 3.448556661605835, "perplexity": 31.45495937853098, "lr": 0.0026291804804649314, "grad_norm": 0.197353, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:03.691839+00:00", "epoch": 0, "step": 20389, "train_loss": 3.609440326690674, "perplexity": 36.94536969014722, "lr": 0.0026291804804649314, "grad_norm": 0.198289, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:03.996442+00:00", "epoch": 0, "step": 20390, "train_loss": 3.450155258178711, "perplexity": 31.505283381966205, "lr": 0.0026291804804649314, "grad_norm": 0.224102, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:04.300210+00:00", "epoch": 0, "step": 20391, "train_loss": 3.505324602127075, "perplexity": 33.292248833435416, "lr": 0.0026291804804649314, "grad_norm": 0.189227, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:04.605414+00:00", "epoch": 0, "step": 20392, "train_loss": 3.6213860511779785, "perplexity": 37.38935548380011, "lr": 0.0026291804804649314, "grad_norm": 0.169889, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:04.911155+00:00", "epoch": 0, "step": 20393, "train_loss": 3.5372824668884277, "perplexity": 34.373381350356354, "lr": 0.0026291804804649314, "grad_norm": 0.19199, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:05.217718+00:00", "epoch": 0, "step": 20394, "train_loss": 3.475555658340454, "perplexity": 32.31578008141651, "lr": 0.0026291804804649314, "grad_norm": 0.194544, "tokens_per_sec": 106889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:05.523929+00:00", "epoch": 0, "step": 20395, "train_loss": 3.4785335063934326, "perplexity": 32.4121549879086, "lr": 0.0026291804804649314, "grad_norm": 0.185908, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:05.828852+00:00", "epoch": 0, "step": 20396, "train_loss": 3.5053865909576416, "perplexity": 33.294312644973495, "lr": 0.0026291804804649314, "grad_norm": 0.192148, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:06.134949+00:00", "epoch": 0, "step": 20397, "train_loss": 3.474442958831787, "perplexity": 32.27984232646674, "lr": 0.0026291804804649314, "grad_norm": 0.170132, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:06.440711+00:00", "epoch": 0, "step": 20398, "train_loss": 3.436960458755493, "perplexity": 31.092308042018715, "lr": 0.0026291804804649314, "grad_norm": 0.192111, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:06.747182+00:00", "epoch": 0, "step": 20399, "train_loss": 3.530623435974121, "perplexity": 34.1452483566459, "lr": 0.0026291804804649314, "grad_norm": 0.165807, "tokens_per_sec": 106920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:07.052264+00:00", "epoch": 0, "step": 20400, "train_loss": 3.458890438079834, "perplexity": 31.781693187250315, "lr": 0.0026291804804649314, "grad_norm": 0.174365, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:07.356857+00:00", "epoch": 0, "step": 20401, "train_loss": 3.5258634090423584, "perplexity": 33.98310227090083, "lr": 0.0026291804804649314, "grad_norm": 0.177004, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:07.660689+00:00", "epoch": 0, "step": 20402, "train_loss": 3.4915032386779785, "perplexity": 32.835269873124666, "lr": 0.0026291804804649314, "grad_norm": 0.169436, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:07.965186+00:00", "epoch": 0, "step": 20403, "train_loss": 3.4583675861358643, "perplexity": 31.765080410574463, "lr": 0.0026291804804649314, "grad_norm": 0.196471, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:08.269126+00:00", "epoch": 0, "step": 20404, "train_loss": 3.4362740516662598, "perplexity": 31.070973384325846, "lr": 0.0026291804804649314, "grad_norm": 0.202778, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:08.574564+00:00", "epoch": 0, "step": 20405, "train_loss": 3.550243616104126, "perplexity": 34.82179960552836, "lr": 0.0026291804804649314, "grad_norm": 0.175319, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:08.879412+00:00", "epoch": 0, "step": 20406, "train_loss": 3.517313241958618, "perplexity": 33.69377970916638, "lr": 0.0026291804804649314, "grad_norm": 0.180959, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:09.184475+00:00", "epoch": 0, "step": 20407, "train_loss": 3.4728689193725586, "perplexity": 32.22907254820664, "lr": 0.0026291804804649314, "grad_norm": 0.149317, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:09.490075+00:00", "epoch": 0, "step": 20408, "train_loss": 3.553150177001953, "perplexity": 34.92315851822243, "lr": 0.0026291804804649314, "grad_norm": 0.161681, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:09.795355+00:00", "epoch": 0, "step": 20409, "train_loss": 3.5686585903167725, "perplexity": 35.46898278929925, "lr": 0.0026291804804649314, "grad_norm": 0.156437, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:10.100145+00:00", "epoch": 0, "step": 20410, "train_loss": 3.521388292312622, "perplexity": 33.8313636985466, "lr": 0.0026291804804649314, "grad_norm": 0.17047, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:10.405302+00:00", "epoch": 0, "step": 20411, "train_loss": 3.4799389839172363, "perplexity": 32.45774157119706, "lr": 0.0026291804804649314, "grad_norm": 0.160462, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:10.710648+00:00", "epoch": 0, "step": 20412, "train_loss": 3.4905216693878174, "perplexity": 32.803055593443354, "lr": 0.0026291804804649314, "grad_norm": 0.157984, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:11.015918+00:00", "epoch": 0, "step": 20413, "train_loss": 3.5700669288635254, "perplexity": 35.51897031639399, "lr": 0.0026291804804649314, "grad_norm": 0.149239, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:11.321430+00:00", "epoch": 0, "step": 20414, "train_loss": 3.587926149368286, "perplexity": 36.15900973017017, "lr": 0.0026291804804649314, "grad_norm": 0.173024, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:11.627695+00:00", "epoch": 0, "step": 20415, "train_loss": 3.5439956188201904, "perplexity": 34.60491136031834, "lr": 0.0026291804804649314, "grad_norm": 0.154028, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:11.932012+00:00", "epoch": 0, "step": 20416, "train_loss": 3.5032131671905518, "perplexity": 33.222028574974296, "lr": 0.0026291804804649314, "grad_norm": 0.143607, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:12.236573+00:00", "epoch": 0, "step": 20417, "train_loss": 3.612515687942505, "perplexity": 37.059164939589095, "lr": 0.0026291804804649314, "grad_norm": 0.171679, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:12.541375+00:00", "epoch": 0, "step": 20418, "train_loss": 3.3470511436462402, "perplexity": 28.418806981215514, "lr": 0.0026291804804649314, "grad_norm": 0.162833, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:12.847130+00:00", "epoch": 0, "step": 20419, "train_loss": 3.5457358360290527, "perplexity": 34.66518385098091, "lr": 0.0026291804804649314, "grad_norm": 0.177485, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:13.153212+00:00", "epoch": 0, "step": 20420, "train_loss": 3.4497604370117188, "perplexity": 31.492846884472698, "lr": 0.0026291804804649314, "grad_norm": 0.187259, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:13.459687+00:00", "epoch": 0, "step": 20421, "train_loss": 3.501854419708252, "perplexity": 33.17691888057311, "lr": 0.0026291804804649314, "grad_norm": 0.173092, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:13.764918+00:00", "epoch": 0, "step": 20422, "train_loss": 3.499072790145874, "perplexity": 33.08476121589074, "lr": 0.0026291804804649314, "grad_norm": 0.163887, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:14.070592+00:00", "epoch": 0, "step": 20423, "train_loss": 3.414944648742676, "perplexity": 30.41526585776469, "lr": 0.0026291804804649314, "grad_norm": 0.178196, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:14.375793+00:00", "epoch": 0, "step": 20424, "train_loss": 3.458418846130371, "perplexity": 31.766708730155287, "lr": 0.0026291804804649314, "grad_norm": 0.160177, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:14.680883+00:00", "epoch": 0, "step": 20425, "train_loss": 3.4870541095733643, "perplexity": 32.689506019821856, "lr": 0.0026291804804649314, "grad_norm": 0.176181, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:14.986490+00:00", "epoch": 0, "step": 20426, "train_loss": 3.5065579414367676, "perplexity": 33.333334803892896, "lr": 0.0026291804804649314, "grad_norm": 0.174516, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:15.291261+00:00", "epoch": 0, "step": 20427, "train_loss": 3.5894582271575928, "perplexity": 36.21445060488294, "lr": 0.0026291804804649314, "grad_norm": 0.142697, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:15.596982+00:00", "epoch": 0, "step": 20428, "train_loss": 3.461982488632202, "perplexity": 31.880115874771374, "lr": 0.0026291804804649314, "grad_norm": 0.172928, "tokens_per_sec": 107183} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:15.902030+00:00", "epoch": 0, "step": 20429, "train_loss": 3.5438835620880127, "perplexity": 34.60103386428783, "lr": 0.0026291804804649314, "grad_norm": 0.148406, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:16.208299+00:00", "epoch": 0, "step": 20430, "train_loss": 3.5196666717529297, "perplexity": 33.773169036178324, "lr": 0.0026291804804649314, "grad_norm": 0.168869, "tokens_per_sec": 107054} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:16.513134+00:00", "epoch": 0, "step": 20431, "train_loss": 3.483246326446533, "perplexity": 32.56526815590866, "lr": 0.0026291804804649314, "grad_norm": 0.16427, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:16.817785+00:00", "epoch": 0, "step": 20432, "train_loss": 3.5335235595703125, "perplexity": 34.24441752902427, "lr": 0.0026291804804649314, "grad_norm": 0.171476, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:17.122146+00:00", "epoch": 0, "step": 20433, "train_loss": 3.55129075050354, "perplexity": 34.858281807295505, "lr": 0.0026291804804649314, "grad_norm": 0.155062, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:17.425877+00:00", "epoch": 0, "step": 20434, "train_loss": 3.5134761333465576, "perplexity": 33.56474074345912, "lr": 0.0026291804804649314, "grad_norm": 0.143355, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:17.730662+00:00", "epoch": 0, "step": 20435, "train_loss": 3.5557503700256348, "perplexity": 35.01408363156038, "lr": 0.0026291804804649314, "grad_norm": 0.170868, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:18.036500+00:00", "epoch": 0, "step": 20436, "train_loss": 3.579629421234131, "perplexity": 35.860249337496015, "lr": 0.0026291804804649314, "grad_norm": 0.165473, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:18.340924+00:00", "epoch": 0, "step": 20437, "train_loss": 3.52817440032959, "perplexity": 34.061727740556854, "lr": 0.0026291804804649314, "grad_norm": 0.169649, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:18.645108+00:00", "epoch": 0, "step": 20438, "train_loss": 3.4098429679870605, "perplexity": 30.260492020308526, "lr": 0.0026291804804649314, "grad_norm": 0.186723, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:18.948879+00:00", "epoch": 0, "step": 20439, "train_loss": 3.5089566707611084, "perplexity": 33.41338842666624, "lr": 0.0026291804804649314, "grad_norm": 0.160997, "tokens_per_sec": 107864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:19.252558+00:00", "epoch": 0, "step": 20440, "train_loss": 3.541823625564575, "perplexity": 34.52983129244928, "lr": 0.0026291804804649314, "grad_norm": 0.179412, "tokens_per_sec": 107966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:19.558581+00:00", "epoch": 0, "step": 20441, "train_loss": 3.5555717945098877, "perplexity": 35.00783153177005, "lr": 0.0026291804804649314, "grad_norm": 0.205998, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:19.864113+00:00", "epoch": 0, "step": 20442, "train_loss": 3.479828119277954, "perplexity": 32.45414335484702, "lr": 0.0026291804804649314, "grad_norm": 0.19447, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:20.169594+00:00", "epoch": 0, "step": 20443, "train_loss": 3.4004361629486084, "perplexity": 29.977172128197502, "lr": 0.0026291804804649314, "grad_norm": 0.154019, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:20.475270+00:00", "epoch": 0, "step": 20444, "train_loss": 3.5074148178100586, "perplexity": 33.36190959171124, "lr": 0.0026291804804649314, "grad_norm": 0.202974, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:20.779812+00:00", "epoch": 0, "step": 20445, "train_loss": 3.45110821723938, "perplexity": 31.535320937231045, "lr": 0.0026291804804649314, "grad_norm": 0.212931, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:21.084082+00:00", "epoch": 0, "step": 20446, "train_loss": 3.3451788425445557, "perplexity": 28.365648197749508, "lr": 0.0026291804804649314, "grad_norm": 0.170377, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:21.389585+00:00", "epoch": 0, "step": 20447, "train_loss": 3.469670295715332, "perplexity": 32.12614856959283, "lr": 0.0026291804804649314, "grad_norm": 0.153025, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:21.694648+00:00", "epoch": 0, "step": 20448, "train_loss": 3.43916392326355, "perplexity": 31.160894375288066, "lr": 0.0026291804804649314, "grad_norm": 0.181843, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:21.999610+00:00", "epoch": 0, "step": 20449, "train_loss": 3.4419400691986084, "perplexity": 31.24752175502855, "lr": 0.0026291804804649314, "grad_norm": 0.171622, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:22.304450+00:00", "epoch": 0, "step": 20450, "train_loss": 3.4585301876068115, "perplexity": 31.770245879319006, "lr": 0.0026291804804649314, "grad_norm": 0.166231, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:22.609732+00:00", "epoch": 0, "step": 20451, "train_loss": 3.5167412757873535, "perplexity": 33.674513517320925, "lr": 0.0026291804804649314, "grad_norm": 0.187273, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:22.915755+00:00", "epoch": 0, "step": 20452, "train_loss": 3.5840017795562744, "perplexity": 36.01738647672091, "lr": 0.0026291804804649314, "grad_norm": 0.178543, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:23.220248+00:00", "epoch": 0, "step": 20453, "train_loss": 3.336829423904419, "perplexity": 28.129797504185486, "lr": 0.0026291804804649314, "grad_norm": 0.184942, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:23.525410+00:00", "epoch": 0, "step": 20454, "train_loss": 3.5207018852233887, "perplexity": 33.80814957874514, "lr": 0.0026291804804649314, "grad_norm": 0.185739, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:23.829601+00:00", "epoch": 0, "step": 20455, "train_loss": 3.563702344894409, "perplexity": 35.29362472348259, "lr": 0.0026291804804649314, "grad_norm": 0.210311, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:24.133424+00:00", "epoch": 0, "step": 20456, "train_loss": 3.458458423614502, "perplexity": 31.767966001445597, "lr": 0.0026291804804649314, "grad_norm": 0.195312, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:24.439187+00:00", "epoch": 0, "step": 20457, "train_loss": 3.481640577316284, "perplexity": 32.51301846606484, "lr": 0.0026291804804649314, "grad_norm": 0.173516, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:24.745578+00:00", "epoch": 0, "step": 20458, "train_loss": 3.518659830093384, "perplexity": 33.739181915313544, "lr": 0.0026291804804649314, "grad_norm": 0.217106, "tokens_per_sec": 106947} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:25.050363+00:00", "epoch": 0, "step": 20459, "train_loss": 3.445279598236084, "perplexity": 31.352048198681654, "lr": 0.0026291804804649314, "grad_norm": 0.190742, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:25.354305+00:00", "epoch": 0, "step": 20460, "train_loss": 3.4712536334991455, "perplexity": 32.177055405194494, "lr": 0.0026291804804649314, "grad_norm": 0.167641, "tokens_per_sec": 107878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:25.659056+00:00", "epoch": 0, "step": 20461, "train_loss": 3.459437608718872, "perplexity": 31.799087955140433, "lr": 0.0026291804804649314, "grad_norm": 0.197114, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:25.963181+00:00", "epoch": 0, "step": 20462, "train_loss": 3.534061908721924, "perplexity": 34.26285794539433, "lr": 0.0026291804804649314, "grad_norm": 0.16343, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:26.268353+00:00", "epoch": 0, "step": 20463, "train_loss": 3.42268705368042, "perplexity": 30.651667138628323, "lr": 0.0026291804804649314, "grad_norm": 0.18245, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:26.573735+00:00", "epoch": 0, "step": 20464, "train_loss": 3.4708569049835205, "perplexity": 32.1642923816615, "lr": 0.0026291804804649314, "grad_norm": 0.179066, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:26.878704+00:00", "epoch": 0, "step": 20465, "train_loss": 3.3843789100646973, "perplexity": 29.499665080109455, "lr": 0.0026291804804649314, "grad_norm": 0.164849, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:27.183227+00:00", "epoch": 0, "step": 20466, "train_loss": 3.4903345108032227, "perplexity": 32.79691679447049, "lr": 0.0026291804804649314, "grad_norm": 0.182163, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:27.488061+00:00", "epoch": 0, "step": 20467, "train_loss": 3.5458672046661377, "perplexity": 34.669738068071815, "lr": 0.0026291804804649314, "grad_norm": 0.172128, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:27.793999+00:00", "epoch": 0, "step": 20468, "train_loss": 3.496943235397339, "perplexity": 33.014380372067485, "lr": 0.0026291804804649314, "grad_norm": 0.176737, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:28.099692+00:00", "epoch": 0, "step": 20469, "train_loss": 3.496778964996338, "perplexity": 33.00895753198322, "lr": 0.0026291804804649314, "grad_norm": 0.171257, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:28.406006+00:00", "epoch": 0, "step": 20470, "train_loss": 3.576493501663208, "perplexity": 35.74797062030602, "lr": 0.0026291804804649314, "grad_norm": 0.20388, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:28.711629+00:00", "epoch": 0, "step": 20471, "train_loss": 3.4934282302856445, "perplexity": 32.89853836818963, "lr": 0.0026291804804649314, "grad_norm": 0.179135, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:29.016901+00:00", "epoch": 0, "step": 20472, "train_loss": 3.4948134422302246, "perplexity": 32.94414139413262, "lr": 0.0026291804804649314, "grad_norm": 0.162516, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:29.322089+00:00", "epoch": 0, "step": 20473, "train_loss": 3.505493640899658, "perplexity": 33.29787698998968, "lr": 0.0026291804804649314, "grad_norm": 0.156088, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:29.627905+00:00", "epoch": 0, "step": 20474, "train_loss": 3.4711501598358154, "perplexity": 32.17372609964723, "lr": 0.0026291804804649314, "grad_norm": 0.162881, "tokens_per_sec": 107149} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:29.932744+00:00", "epoch": 0, "step": 20475, "train_loss": 3.488325834274292, "perplexity": 32.73110451740045, "lr": 0.0026291804804649314, "grad_norm": 0.181431, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:30.237738+00:00", "epoch": 0, "step": 20476, "train_loss": 3.5025689601898193, "perplexity": 33.200633603722814, "lr": 0.0026291804804649314, "grad_norm": 0.166483, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:30.542361+00:00", "epoch": 0, "step": 20477, "train_loss": 3.5156195163726807, "perplexity": 33.636759993892426, "lr": 0.0026291804804649314, "grad_norm": 0.173774, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:30.846470+00:00", "epoch": 0, "step": 20478, "train_loss": 3.4602715969085693, "perplexity": 31.825619080723083, "lr": 0.0026291804804649314, "grad_norm": 0.170676, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:31.151321+00:00", "epoch": 0, "step": 20479, "train_loss": 3.391888380050659, "perplexity": 29.722025790215064, "lr": 0.0026291804804649314, "grad_norm": 0.178273, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:31.456795+00:00", "epoch": 0, "step": 20480, "train_loss": 3.498386859893799, "perplexity": 33.06207515870602, "lr": 0.0026291804804649314, "grad_norm": 0.17805, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:31.760959+00:00", "epoch": 0, "step": 20481, "train_loss": 3.4347784519195557, "perplexity": 31.024538377150442, "lr": 0.0026291804804649314, "grad_norm": 0.156505, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:32.065955+00:00", "epoch": 0, "step": 20482, "train_loss": 3.3390610218048096, "perplexity": 28.192641996972704, "lr": 0.0026291804804649314, "grad_norm": 0.177092, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:32.371261+00:00", "epoch": 0, "step": 20483, "train_loss": 3.5723538398742676, "perplexity": 35.600291992977134, "lr": 0.0026291804804649314, "grad_norm": 0.204846, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:32.677108+00:00", "epoch": 0, "step": 20484, "train_loss": 3.5221638679504395, "perplexity": 33.8576126577362, "lr": 0.0026291804804649314, "grad_norm": 0.157012, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:32.983358+00:00", "epoch": 0, "step": 20485, "train_loss": 3.5027947425842285, "perplexity": 33.20813056858325, "lr": 0.0026291804804649314, "grad_norm": 0.168384, "tokens_per_sec": 106998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:33.288547+00:00", "epoch": 0, "step": 20486, "train_loss": 3.5963964462280273, "perplexity": 36.46658807773909, "lr": 0.0026291804804649314, "grad_norm": 0.177921, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:33.593474+00:00", "epoch": 0, "step": 20487, "train_loss": 3.334132194519043, "perplexity": 28.05402721860894, "lr": 0.0026291804804649314, "grad_norm": 0.149739, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:33.897549+00:00", "epoch": 0, "step": 20488, "train_loss": 3.463090181350708, "perplexity": 31.915448812397106, "lr": 0.0026291804804649314, "grad_norm": 0.229685, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:34.203328+00:00", "epoch": 0, "step": 20489, "train_loss": 3.5452957153320312, "perplexity": 34.64993034304054, "lr": 0.0026291804804649314, "grad_norm": 0.198024, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:34.508925+00:00", "epoch": 0, "step": 20490, "train_loss": 3.5517961978912354, "perplexity": 34.87590528827056, "lr": 0.0026291804804649314, "grad_norm": 0.174971, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:34.815805+00:00", "epoch": 0, "step": 20491, "train_loss": 3.4095022678375244, "perplexity": 30.25018402222012, "lr": 0.0026291804804649314, "grad_norm": 0.186065, "tokens_per_sec": 106713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:35.121518+00:00", "epoch": 0, "step": 20492, "train_loss": 3.4904944896698, "perplexity": 32.80216402775852, "lr": 0.0026291804804649314, "grad_norm": 0.162228, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:35.426559+00:00", "epoch": 0, "step": 20493, "train_loss": 3.4797260761260986, "perplexity": 32.450831800731905, "lr": 0.0026291804804649314, "grad_norm": 0.167433, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:35.730500+00:00", "epoch": 0, "step": 20494, "train_loss": 3.547842025756836, "perplexity": 34.73827224711559, "lr": 0.0026291804804649314, "grad_norm": 0.154544, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:36.035896+00:00", "epoch": 0, "step": 20495, "train_loss": 3.6605076789855957, "perplexity": 38.881076967329335, "lr": 0.0026291804804649314, "grad_norm": 0.157472, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:36.341732+00:00", "epoch": 0, "step": 20496, "train_loss": 3.4918837547302246, "perplexity": 32.847766597842615, "lr": 0.0026291804804649314, "grad_norm": 0.150744, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:36.646980+00:00", "epoch": 0, "step": 20497, "train_loss": 3.423954725265503, "perplexity": 30.690548024986033, "lr": 0.0026291804804649314, "grad_norm": 0.152728, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:36.952505+00:00", "epoch": 0, "step": 20498, "train_loss": 3.543701410293579, "perplexity": 34.59473179786397, "lr": 0.0026291804804649314, "grad_norm": 0.162377, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:37.257034+00:00", "epoch": 0, "step": 20499, "train_loss": 3.367737054824829, "perplexity": 29.012798343578204, "lr": 0.0026291804804649314, "grad_norm": 0.18914, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:37.561768+00:00", "epoch": 0, "step": 20500, "train_loss": 3.5254976749420166, "perplexity": 33.97067576410212, "lr": 0.0026291804804649314, "grad_norm": 0.16789, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:45:40.635498+00:00", "step": 20500, "epoch": 0, "val_loss": 3.458132040500641, "val_ppl": 31.75759916565312, "eval_train_loss": 3.5254976749420166, "eval_train_ppl": 33.97067576410212} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:45:41.670318+00:00", "step": 20500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4581_epoch_0000_step_0020500.pt", "val_loss": 3.458132040500641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:42.966312+00:00", "epoch": 0, "step": 20501, "train_loss": 3.5105578899383545, "perplexity": 33.4669334422043, "lr": 0.0026291804804649314, "grad_norm": 0.165841, "tokens_per_sec": 6063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:43.269772+00:00", "epoch": 0, "step": 20502, "train_loss": 3.5167183876037598, "perplexity": 33.673742777693555, "lr": 0.0026291804804649314, "grad_norm": 0.184896, "tokens_per_sec": 107980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:43.574686+00:00", "epoch": 0, "step": 20503, "train_loss": 3.5763418674468994, "perplexity": 35.742550415751005, "lr": 0.0026291804804649314, "grad_norm": 0.158113, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:43.879521+00:00", "epoch": 0, "step": 20504, "train_loss": 3.4939987659454346, "perplexity": 32.91731351291915, "lr": 0.0026291804804649314, "grad_norm": 0.186948, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:44.182896+00:00", "epoch": 0, "step": 20505, "train_loss": 3.6159305572509766, "perplexity": 37.18593347033054, "lr": 0.0026291804804649314, "grad_norm": 0.185456, "tokens_per_sec": 108012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:44.486528+00:00", "epoch": 0, "step": 20506, "train_loss": 3.5021896362304688, "perplexity": 33.18804219619356, "lr": 0.0026291804804649314, "grad_norm": 0.165125, "tokens_per_sec": 107921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:44.790328+00:00", "epoch": 0, "step": 20507, "train_loss": 3.587925672531128, "perplexity": 36.158992488214835, "lr": 0.0026291804804649314, "grad_norm": 0.218975, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:45.094284+00:00", "epoch": 0, "step": 20508, "train_loss": 3.503133773803711, "perplexity": 33.219391070309605, "lr": 0.0026291804804649314, "grad_norm": 0.186379, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:45.398803+00:00", "epoch": 0, "step": 20509, "train_loss": 3.4554286003112793, "perplexity": 31.67186034285986, "lr": 0.0026291804804649314, "grad_norm": 0.185634, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:45.702776+00:00", "epoch": 0, "step": 20510, "train_loss": 3.5741238594055176, "perplexity": 35.663361005348655, "lr": 0.0026291804804649314, "grad_norm": 0.181204, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:46.007134+00:00", "epoch": 0, "step": 20511, "train_loss": 3.509148359298706, "perplexity": 33.41979400414821, "lr": 0.0026291804804649314, "grad_norm": 0.159226, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:46.312033+00:00", "epoch": 0, "step": 20512, "train_loss": 3.5282373428344727, "perplexity": 34.063871738495074, "lr": 0.0026291804804649314, "grad_norm": 0.176142, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:46.617730+00:00", "epoch": 0, "step": 20513, "train_loss": 3.3949570655822754, "perplexity": 29.813373427630705, "lr": 0.0026291804804649314, "grad_norm": 0.186723, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:46.922767+00:00", "epoch": 0, "step": 20514, "train_loss": 3.5415115356445312, "perplexity": 34.51905656159199, "lr": 0.0026291804804649314, "grad_norm": 0.159814, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:47.227171+00:00", "epoch": 0, "step": 20515, "train_loss": 3.4489054679870605, "perplexity": 31.465932982803743, "lr": 0.0026291804804649314, "grad_norm": 0.187447, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:47.531314+00:00", "epoch": 0, "step": 20516, "train_loss": 3.4929046630859375, "perplexity": 32.88131828091154, "lr": 0.0026291804804649314, "grad_norm": 0.183982, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:47.836896+00:00", "epoch": 0, "step": 20517, "train_loss": 3.430396318435669, "perplexity": 30.888882157621605, "lr": 0.0026291804804649314, "grad_norm": 0.164645, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:48.143079+00:00", "epoch": 0, "step": 20518, "train_loss": 3.5306649208068848, "perplexity": 34.1466648959459, "lr": 0.0026291804804649314, "grad_norm": 0.16784, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:48.449183+00:00", "epoch": 0, "step": 20519, "train_loss": 3.5234572887420654, "perplexity": 33.90143313094903, "lr": 0.0026291804804649314, "grad_norm": 0.172999, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:48.754295+00:00", "epoch": 0, "step": 20520, "train_loss": 3.6040329933166504, "perplexity": 36.7461329148991, "lr": 0.0026291804804649314, "grad_norm": 0.159996, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:49.059106+00:00", "epoch": 0, "step": 20521, "train_loss": 3.533531904220581, "perplexity": 34.24470328790447, "lr": 0.0026291804804649314, "grad_norm": 0.191984, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:49.363373+00:00", "epoch": 0, "step": 20522, "train_loss": 3.470585346221924, "perplexity": 32.155559072111735, "lr": 0.0026291804804649314, "grad_norm": 0.164984, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:49.668435+00:00", "epoch": 0, "step": 20523, "train_loss": 3.491600275039673, "perplexity": 32.838456242842774, "lr": 0.0026291804804649314, "grad_norm": 0.152595, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:49.974302+00:00", "epoch": 0, "step": 20524, "train_loss": 3.5242819786071777, "perplexity": 33.92940283084371, "lr": 0.0026291804804649314, "grad_norm": 0.163957, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:50.280225+00:00", "epoch": 0, "step": 20525, "train_loss": 3.4940783977508545, "perplexity": 32.919934882394564, "lr": 0.0026291804804649314, "grad_norm": 0.188345, "tokens_per_sec": 107107} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:50.584781+00:00", "epoch": 0, "step": 20526, "train_loss": 3.332240104675293, "perplexity": 28.000996663740263, "lr": 0.0026291804804649314, "grad_norm": 0.174419, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:50.889312+00:00", "epoch": 0, "step": 20527, "train_loss": 3.43575382232666, "perplexity": 31.05481355613423, "lr": 0.0026291804804649314, "grad_norm": 0.162929, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:51.193955+00:00", "epoch": 0, "step": 20528, "train_loss": 3.496328830718994, "perplexity": 32.99410241238878, "lr": 0.0026291804804649314, "grad_norm": 0.160543, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:51.499302+00:00", "epoch": 0, "step": 20529, "train_loss": 3.4802041053771973, "perplexity": 32.466347955847674, "lr": 0.0026291804804649314, "grad_norm": 0.179999, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:51.803794+00:00", "epoch": 0, "step": 20530, "train_loss": 3.5658724308013916, "perplexity": 35.37029808490223, "lr": 0.0026291804804649314, "grad_norm": 0.182702, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:52.107634+00:00", "epoch": 0, "step": 20531, "train_loss": 3.4424562454223633, "perplexity": 31.263655146288034, "lr": 0.0026291804804649314, "grad_norm": 0.204625, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:52.412397+00:00", "epoch": 0, "step": 20532, "train_loss": 3.4711990356445312, "perplexity": 32.1752986549594, "lr": 0.0026291804804649314, "grad_norm": 0.180375, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:52.717050+00:00", "epoch": 0, "step": 20533, "train_loss": 3.5426836013793945, "perplexity": 34.559538884324965, "lr": 0.0026291804804649314, "grad_norm": 0.16421, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:53.023269+00:00", "epoch": 0, "step": 20534, "train_loss": 3.561457872390747, "perplexity": 35.21449798534796, "lr": 0.0026291804804649314, "grad_norm": 0.173669, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:53.328013+00:00", "epoch": 0, "step": 20535, "train_loss": 3.584242343902588, "perplexity": 36.02605201802295, "lr": 0.0026291804804649314, "grad_norm": 0.170147, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:53.631992+00:00", "epoch": 0, "step": 20536, "train_loss": 3.444748640060425, "perplexity": 31.335405990916016, "lr": 0.0026291804804649314, "grad_norm": 0.171307, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:53.936885+00:00", "epoch": 0, "step": 20537, "train_loss": 3.5829057693481445, "perplexity": 35.97793267830462, "lr": 0.0026291804804649314, "grad_norm": 0.159972, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:54.240819+00:00", "epoch": 0, "step": 20538, "train_loss": 3.354686975479126, "perplexity": 28.63663881744371, "lr": 0.0026291804804649314, "grad_norm": 0.160435, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:54.545125+00:00", "epoch": 0, "step": 20539, "train_loss": 3.5292649269104004, "perplexity": 34.09889322133535, "lr": 0.0026291804804649314, "grad_norm": 0.168824, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:54.850515+00:00", "epoch": 0, "step": 20540, "train_loss": 3.451566696166992, "perplexity": 31.549782532271227, "lr": 0.0026291804804649314, "grad_norm": 0.165148, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:55.155068+00:00", "epoch": 0, "step": 20541, "train_loss": 3.3820817470550537, "perplexity": 29.43197731537069, "lr": 0.0026291804804649314, "grad_norm": 0.166324, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:55.459606+00:00", "epoch": 0, "step": 20542, "train_loss": 3.477126121520996, "perplexity": 32.36657069614107, "lr": 0.0026291804804649314, "grad_norm": 0.170997, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:55.764352+00:00", "epoch": 0, "step": 20543, "train_loss": 3.467252016067505, "perplexity": 32.04855242075334, "lr": 0.0026291804804649314, "grad_norm": 0.158073, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:56.070310+00:00", "epoch": 0, "step": 20544, "train_loss": 3.413607120513916, "perplexity": 30.37461177514704, "lr": 0.0026291804804649314, "grad_norm": 0.189466, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:56.376139+00:00", "epoch": 0, "step": 20545, "train_loss": 3.498657464981079, "perplexity": 33.07102313506937, "lr": 0.0026291804804649314, "grad_norm": 0.182335, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:56.680340+00:00", "epoch": 0, "step": 20546, "train_loss": 3.4701344966888428, "perplexity": 32.14106502088174, "lr": 0.0026291804804649314, "grad_norm": 0.188292, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:56.984515+00:00", "epoch": 0, "step": 20547, "train_loss": 3.5143308639526367, "perplexity": 33.59344181875407, "lr": 0.0026291804804649314, "grad_norm": 0.209808, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:57.288790+00:00", "epoch": 0, "step": 20548, "train_loss": 3.519740104675293, "perplexity": 33.77564918973947, "lr": 0.0026291804804649314, "grad_norm": 0.226503, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:57.594191+00:00", "epoch": 0, "step": 20549, "train_loss": 3.4774200916290283, "perplexity": 32.37608689909632, "lr": 0.0026291804804649314, "grad_norm": 0.211255, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:57.898960+00:00", "epoch": 0, "step": 20550, "train_loss": 3.448611259460449, "perplexity": 31.45667679871333, "lr": 0.0026291804804649314, "grad_norm": 0.171873, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:58.202561+00:00", "epoch": 0, "step": 20551, "train_loss": 3.5477957725524902, "perplexity": 34.73666552786897, "lr": 0.0026291804804649314, "grad_norm": 0.187254, "tokens_per_sec": 107872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:58.506973+00:00", "epoch": 0, "step": 20552, "train_loss": 3.4556305408477783, "perplexity": 31.67825682116181, "lr": 0.0026291804804649314, "grad_norm": 0.210822, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:58.811180+00:00", "epoch": 0, "step": 20553, "train_loss": 3.477471113204956, "perplexity": 32.377738820213736, "lr": 0.0026291804804649314, "grad_norm": 0.159852, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:59.117194+00:00", "epoch": 0, "step": 20554, "train_loss": 3.4271531105041504, "perplexity": 30.78886536530734, "lr": 0.0026291804804649314, "grad_norm": 0.192617, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:59.422106+00:00", "epoch": 0, "step": 20555, "train_loss": 3.4804491996765137, "perplexity": 32.47430624787744, "lr": 0.0026291804804649314, "grad_norm": 0.20534, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:45:59.726383+00:00", "epoch": 0, "step": 20556, "train_loss": 3.584559679031372, "perplexity": 36.03748616401182, "lr": 0.0026291804804649314, "grad_norm": 0.22077, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:00.031132+00:00", "epoch": 0, "step": 20557, "train_loss": 3.5536913871765137, "perplexity": 34.94206440250728, "lr": 0.0026291804804649314, "grad_norm": 0.185032, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:00.336071+00:00", "epoch": 0, "step": 20558, "train_loss": 3.4843831062316895, "perplexity": 32.602308743953486, "lr": 0.0026291804804649314, "grad_norm": 0.18668, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:00.641316+00:00", "epoch": 0, "step": 20559, "train_loss": 3.536372184753418, "perplexity": 34.34210611219812, "lr": 0.0026291804804649314, "grad_norm": 0.155134, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:00.947021+00:00", "epoch": 0, "step": 20560, "train_loss": 3.6108994483947754, "perplexity": 36.99931682907752, "lr": 0.0026291804804649314, "grad_norm": 0.171302, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:01.251717+00:00", "epoch": 0, "step": 20561, "train_loss": 3.4928669929504395, "perplexity": 32.88007966052619, "lr": 0.0026291804804649314, "grad_norm": 0.165735, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:01.556031+00:00", "epoch": 0, "step": 20562, "train_loss": 3.3928825855255127, "perplexity": 29.75159028513555, "lr": 0.0026291804804649314, "grad_norm": 0.167691, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:01.861503+00:00", "epoch": 0, "step": 20563, "train_loss": 3.422667980194092, "perplexity": 30.65108251004969, "lr": 0.0026291804804649314, "grad_norm": 0.156272, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:02.167325+00:00", "epoch": 0, "step": 20564, "train_loss": 3.476499557495117, "perplexity": 32.346297319249366, "lr": 0.0026291804804649314, "grad_norm": 0.154268, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:02.472114+00:00", "epoch": 0, "step": 20565, "train_loss": 3.5059213638305664, "perplexity": 33.312122301834954, "lr": 0.0026291804804649314, "grad_norm": 0.167116, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:02.776760+00:00", "epoch": 0, "step": 20566, "train_loss": 3.513461112976074, "perplexity": 33.56423659240425, "lr": 0.0026291804804649314, "grad_norm": 0.159104, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:03.082372+00:00", "epoch": 0, "step": 20567, "train_loss": 3.4961588382720947, "perplexity": 32.988494140881855, "lr": 0.0026291804804649314, "grad_norm": 0.167626, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:03.387079+00:00", "epoch": 0, "step": 20568, "train_loss": 3.458324909210205, "perplexity": 31.763724803525996, "lr": 0.0026291804804649314, "grad_norm": 0.154277, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:03.691382+00:00", "epoch": 0, "step": 20569, "train_loss": 3.511929750442505, "perplexity": 33.51287691319986, "lr": 0.0026291804804649314, "grad_norm": 0.161793, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:03.996282+00:00", "epoch": 0, "step": 20570, "train_loss": 3.4244396686553955, "perplexity": 30.70543481271584, "lr": 0.0026291804804649314, "grad_norm": 0.191178, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:04.300587+00:00", "epoch": 0, "step": 20571, "train_loss": 3.5236711502075195, "perplexity": 33.90868411644497, "lr": 0.0026291804804649314, "grad_norm": 0.183832, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:04.604937+00:00", "epoch": 0, "step": 20572, "train_loss": 3.5614330768585205, "perplexity": 35.213624833953496, "lr": 0.0026291804804649314, "grad_norm": 0.155794, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:04.911439+00:00", "epoch": 0, "step": 20573, "train_loss": 3.5256597995758057, "perplexity": 33.97618369394343, "lr": 0.0026291804804649314, "grad_norm": 0.153872, "tokens_per_sec": 106909} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:05.216270+00:00", "epoch": 0, "step": 20574, "train_loss": 3.5451574325561523, "perplexity": 34.64513918576359, "lr": 0.0026291804804649314, "grad_norm": 0.175725, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:05.520399+00:00", "epoch": 0, "step": 20575, "train_loss": 3.4580414295196533, "perplexity": 31.754721708805466, "lr": 0.0026291804804649314, "grad_norm": 0.183082, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:05.824967+00:00", "epoch": 0, "step": 20576, "train_loss": 3.5627691745758057, "perplexity": 35.26070512264079, "lr": 0.0026291804804649314, "grad_norm": 0.165253, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:06.129584+00:00", "epoch": 0, "step": 20577, "train_loss": 3.408822774887085, "perplexity": 30.22963621726243, "lr": 0.0026291804804649314, "grad_norm": 0.169533, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:06.434722+00:00", "epoch": 0, "step": 20578, "train_loss": 3.452401876449585, "perplexity": 31.576143295024636, "lr": 0.0026291804804649314, "grad_norm": 0.158047, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:06.738932+00:00", "epoch": 0, "step": 20579, "train_loss": 3.589271068572998, "perplexity": 36.20767339379226, "lr": 0.0026291804804649314, "grad_norm": 0.193653, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:07.042607+00:00", "epoch": 0, "step": 20580, "train_loss": 3.4695658683776855, "perplexity": 32.1227938965917, "lr": 0.0026291804804649314, "grad_norm": 0.158418, "tokens_per_sec": 107990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:07.346670+00:00", "epoch": 0, "step": 20581, "train_loss": 3.5736441612243652, "perplexity": 35.646257458538734, "lr": 0.0026291804804649314, "grad_norm": 0.154936, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:07.652092+00:00", "epoch": 0, "step": 20582, "train_loss": 3.516669988632202, "perplexity": 33.67211304261369, "lr": 0.0026291804804649314, "grad_norm": 0.187038, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:07.958614+00:00", "epoch": 0, "step": 20583, "train_loss": 3.490342378616333, "perplexity": 32.79717483549753, "lr": 0.0026291804804649314, "grad_norm": 0.153563, "tokens_per_sec": 106898} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:08.263525+00:00", "epoch": 0, "step": 20584, "train_loss": 3.560089588165283, "perplexity": 35.16634749253915, "lr": 0.0026291804804649314, "grad_norm": 0.166674, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:08.568024+00:00", "epoch": 0, "step": 20585, "train_loss": 3.54898738861084, "perplexity": 34.77808296826815, "lr": 0.0026291804804649314, "grad_norm": 0.171265, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:08.872388+00:00", "epoch": 0, "step": 20586, "train_loss": 3.542933464050293, "perplexity": 34.56817510190489, "lr": 0.0026291804804649314, "grad_norm": 0.157577, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:09.177339+00:00", "epoch": 0, "step": 20587, "train_loss": 3.5353174209594727, "perplexity": 34.30590239860276, "lr": 0.0026291804804649314, "grad_norm": 0.147445, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:09.482615+00:00", "epoch": 0, "step": 20588, "train_loss": 3.5596373081207275, "perplexity": 35.15044605155165, "lr": 0.0026291804804649314, "grad_norm": 0.147889, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:09.788130+00:00", "epoch": 0, "step": 20589, "train_loss": 3.5536606311798096, "perplexity": 34.940989741015905, "lr": 0.0026291804804649314, "grad_norm": 0.172908, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:10.092282+00:00", "epoch": 0, "step": 20590, "train_loss": 3.4654433727264404, "perplexity": 31.990640406702845, "lr": 0.0026291804804649314, "grad_norm": 0.187307, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:10.396510+00:00", "epoch": 0, "step": 20591, "train_loss": 3.4181249141693115, "perplexity": 30.512148450737936, "lr": 0.0026291804804649314, "grad_norm": 0.1928, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:10.701108+00:00", "epoch": 0, "step": 20592, "train_loss": 3.487748384475708, "perplexity": 32.71220940370429, "lr": 0.0026291804804649314, "grad_norm": 0.224957, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:11.006955+00:00", "epoch": 0, "step": 20593, "train_loss": 3.634005069732666, "perplexity": 37.86416194954816, "lr": 0.0026291804804649314, "grad_norm": 0.181884, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:11.312195+00:00", "epoch": 0, "step": 20594, "train_loss": 3.6004300117492676, "perplexity": 36.613975498667955, "lr": 0.0026291804804649314, "grad_norm": 0.168848, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:11.616781+00:00", "epoch": 0, "step": 20595, "train_loss": 3.520181655883789, "perplexity": 33.79056616151803, "lr": 0.0026291804804649314, "grad_norm": 0.176914, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:11.921446+00:00", "epoch": 0, "step": 20596, "train_loss": 3.452315092086792, "perplexity": 31.573403098454147, "lr": 0.0026291804804649314, "grad_norm": 0.186574, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:12.226083+00:00", "epoch": 0, "step": 20597, "train_loss": 3.480839729309082, "perplexity": 32.486990903470065, "lr": 0.0026291804804649314, "grad_norm": 0.170366, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:12.530083+00:00", "epoch": 0, "step": 20598, "train_loss": 3.5420076847076416, "perplexity": 34.536187408539895, "lr": 0.0026291804804649314, "grad_norm": 0.160458, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:12.836241+00:00", "epoch": 0, "step": 20599, "train_loss": 3.4992034435272217, "perplexity": 33.089084134210474, "lr": 0.0026291804804649314, "grad_norm": 0.148958, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:13.141661+00:00", "epoch": 0, "step": 20600, "train_loss": 3.437668561935425, "perplexity": 31.114332401053744, "lr": 0.0026291804804649314, "grad_norm": 0.162709, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:13.446283+00:00", "epoch": 0, "step": 20601, "train_loss": 3.5209836959838867, "perplexity": 33.817678421691134, "lr": 0.0026291804804649314, "grad_norm": 0.168784, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:13.751183+00:00", "epoch": 0, "step": 20602, "train_loss": 3.412121534347534, "perplexity": 30.32952117336457, "lr": 0.0026291804804649314, "grad_norm": 0.155959, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:14.056875+00:00", "epoch": 0, "step": 20603, "train_loss": 3.370858669281006, "perplexity": 29.10350661877452, "lr": 0.0026291804804649314, "grad_norm": 0.161551, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:14.362809+00:00", "epoch": 0, "step": 20604, "train_loss": 3.6475517749786377, "perplexity": 38.38058661971989, "lr": 0.0026291804804649314, "grad_norm": 0.173705, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:14.668113+00:00", "epoch": 0, "step": 20605, "train_loss": 3.5556840896606445, "perplexity": 35.01176296222574, "lr": 0.0026291804804649314, "grad_norm": 0.173198, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:14.972001+00:00", "epoch": 0, "step": 20606, "train_loss": 3.545612096786499, "perplexity": 34.66089467276351, "lr": 0.0026291804804649314, "grad_norm": 0.170751, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:15.275961+00:00", "epoch": 0, "step": 20607, "train_loss": 3.5381393432617188, "perplexity": 34.40284771141695, "lr": 0.0026291804804649314, "grad_norm": 0.177866, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:15.580726+00:00", "epoch": 0, "step": 20608, "train_loss": 3.537959337234497, "perplexity": 34.39665554880532, "lr": 0.0026291804804649314, "grad_norm": 0.145324, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:15.886016+00:00", "epoch": 0, "step": 20609, "train_loss": 3.5156469345092773, "perplexity": 33.63768226381601, "lr": 0.0026291804804649314, "grad_norm": 0.173867, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:16.190662+00:00", "epoch": 0, "step": 20610, "train_loss": 3.523629903793335, "perplexity": 33.90728553365891, "lr": 0.0026291804804649314, "grad_norm": 0.165688, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:16.495402+00:00", "epoch": 0, "step": 20611, "train_loss": 3.5177597999572754, "perplexity": 33.7088292960071, "lr": 0.0026291804804649314, "grad_norm": 0.148243, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:16.800414+00:00", "epoch": 0, "step": 20612, "train_loss": 3.467564344406128, "perplexity": 32.05856365520091, "lr": 0.0026291804804649314, "grad_norm": 0.161711, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:17.104755+00:00", "epoch": 0, "step": 20613, "train_loss": 3.6241955757141113, "perplexity": 37.494549498810045, "lr": 0.0026291804804649314, "grad_norm": 0.145896, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:17.409795+00:00", "epoch": 0, "step": 20614, "train_loss": 3.547358751296997, "perplexity": 34.72148818334097, "lr": 0.0026291804804649314, "grad_norm": 0.150488, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:17.715561+00:00", "epoch": 0, "step": 20615, "train_loss": 3.411646842956543, "perplexity": 30.31512742733463, "lr": 0.0026291804804649314, "grad_norm": 0.17229, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:18.020539+00:00", "epoch": 0, "step": 20616, "train_loss": 3.4650628566741943, "perplexity": 31.978469770214588, "lr": 0.0026291804804649314, "grad_norm": 0.189133, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:18.325272+00:00", "epoch": 0, "step": 20617, "train_loss": 3.5797736644744873, "perplexity": 35.86542230913457, "lr": 0.0026291804804649314, "grad_norm": 0.155246, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:18.630069+00:00", "epoch": 0, "step": 20618, "train_loss": 3.4824140071868896, "perplexity": 32.53817473278007, "lr": 0.0026291804804649314, "grad_norm": 0.185253, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:18.935285+00:00", "epoch": 0, "step": 20619, "train_loss": 3.484927177429199, "perplexity": 32.62005154734986, "lr": 0.0026291804804649314, "grad_norm": 0.185926, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:19.241228+00:00", "epoch": 0, "step": 20620, "train_loss": 3.4725043773651123, "perplexity": 32.21732583861642, "lr": 0.0026291804804649314, "grad_norm": 0.20437, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:19.546698+00:00", "epoch": 0, "step": 20621, "train_loss": 3.478736400604248, "perplexity": 32.41873189370119, "lr": 0.0026291804804649314, "grad_norm": 0.157374, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:19.850526+00:00", "epoch": 0, "step": 20622, "train_loss": 3.5395846366882324, "perplexity": 34.45260586997581, "lr": 0.0026291804804649314, "grad_norm": 0.193892, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:20.154363+00:00", "epoch": 0, "step": 20623, "train_loss": 3.3473763465881348, "perplexity": 28.428050363757073, "lr": 0.0026291804804649314, "grad_norm": 0.168836, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:20.459076+00:00", "epoch": 0, "step": 20624, "train_loss": 3.5637881755828857, "perplexity": 35.296654129597584, "lr": 0.0026291804804649314, "grad_norm": 0.189046, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:20.763933+00:00", "epoch": 0, "step": 20625, "train_loss": 3.5063552856445312, "perplexity": 33.32658029496363, "lr": 0.0026291804804649314, "grad_norm": 0.212357, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:21.068889+00:00", "epoch": 0, "step": 20626, "train_loss": 3.463561534881592, "perplexity": 31.930495817835553, "lr": 0.0026291804804649314, "grad_norm": 0.181363, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:21.375443+00:00", "epoch": 0, "step": 20627, "train_loss": 3.4737942218780518, "perplexity": 32.258907991061506, "lr": 0.0026291804804649314, "grad_norm": 0.208087, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:21.679943+00:00", "epoch": 0, "step": 20628, "train_loss": 3.4389193058013916, "perplexity": 31.153272808609557, "lr": 0.0026291804804649314, "grad_norm": 0.20569, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:21.984036+00:00", "epoch": 0, "step": 20629, "train_loss": 3.451140880584717, "perplexity": 31.53635100313174, "lr": 0.0026291804804649314, "grad_norm": 0.20046, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:22.289524+00:00", "epoch": 0, "step": 20630, "train_loss": 3.5559935569763184, "perplexity": 35.02259963523824, "lr": 0.0026291804804649314, "grad_norm": 0.170279, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:22.595124+00:00", "epoch": 0, "step": 20631, "train_loss": 3.519557237625122, "perplexity": 33.76947330110501, "lr": 0.0026291804804649314, "grad_norm": 0.200648, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:22.901553+00:00", "epoch": 0, "step": 20632, "train_loss": 3.523996114730835, "perplexity": 33.919705026420516, "lr": 0.0026291804804649314, "grad_norm": 0.186711, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:23.206146+00:00", "epoch": 0, "step": 20633, "train_loss": 3.5036561489105225, "perplexity": 33.23674858644064, "lr": 0.0026291804804649314, "grad_norm": 0.209093, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:23.510546+00:00", "epoch": 0, "step": 20634, "train_loss": 3.557523250579834, "perplexity": 35.07621447856547, "lr": 0.0026291804804649314, "grad_norm": 0.192287, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:23.815148+00:00", "epoch": 0, "step": 20635, "train_loss": 3.5125339031219482, "perplexity": 33.53312992492245, "lr": 0.0026291804804649314, "grad_norm": 0.168449, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:24.120649+00:00", "epoch": 0, "step": 20636, "train_loss": 3.440016508102417, "perplexity": 31.187473010058497, "lr": 0.0026291804804649314, "grad_norm": 0.172028, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:24.425346+00:00", "epoch": 0, "step": 20637, "train_loss": 3.4471817016601562, "perplexity": 31.41173978870393, "lr": 0.0026291804804649314, "grad_norm": 0.160946, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:24.730680+00:00", "epoch": 0, "step": 20638, "train_loss": 3.447234630584717, "perplexity": 31.413402422309836, "lr": 0.0026291804804649314, "grad_norm": 0.166708, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:25.034914+00:00", "epoch": 0, "step": 20639, "train_loss": 3.6117820739746094, "perplexity": 37.03198778853855, "lr": 0.0026291804804649314, "grad_norm": 0.188452, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:25.339517+00:00", "epoch": 0, "step": 20640, "train_loss": 3.5261058807373047, "perplexity": 33.991343210364896, "lr": 0.0026291804804649314, "grad_norm": 0.192457, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:25.644511+00:00", "epoch": 0, "step": 20641, "train_loss": 3.5535614490509033, "perplexity": 34.937524391120526, "lr": 0.0026291804804649314, "grad_norm": 0.187611, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:25.949843+00:00", "epoch": 0, "step": 20642, "train_loss": 3.516798734664917, "perplexity": 33.67644847265977, "lr": 0.0026291804804649314, "grad_norm": 0.16816, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:26.254314+00:00", "epoch": 0, "step": 20643, "train_loss": 3.4717440605163574, "perplexity": 32.192839772723296, "lr": 0.0026291804804649314, "grad_norm": 0.158763, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:26.558364+00:00", "epoch": 0, "step": 20644, "train_loss": 3.4454400539398193, "perplexity": 31.357079217255936, "lr": 0.0026291804804649314, "grad_norm": 0.160033, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:26.862516+00:00", "epoch": 0, "step": 20645, "train_loss": 3.5296528339385986, "perplexity": 34.11212298746343, "lr": 0.0026291804804649314, "grad_norm": 0.176878, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:27.166859+00:00", "epoch": 0, "step": 20646, "train_loss": 3.5332202911376953, "perplexity": 34.23403385279454, "lr": 0.0026291804804649314, "grad_norm": 0.17511, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:27.472508+00:00", "epoch": 0, "step": 20647, "train_loss": 3.503249406814575, "perplexity": 33.22323255061483, "lr": 0.0026291804804649314, "grad_norm": 0.172122, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:27.777304+00:00", "epoch": 0, "step": 20648, "train_loss": 3.482753038406372, "perplexity": 32.54920806005497, "lr": 0.0026291804804649314, "grad_norm": 0.179691, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:28.081291+00:00", "epoch": 0, "step": 20649, "train_loss": 3.5517189502716064, "perplexity": 34.87321131165761, "lr": 0.0026291804804649314, "grad_norm": 0.194397, "tokens_per_sec": 107795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:28.385818+00:00", "epoch": 0, "step": 20650, "train_loss": 3.5934128761291504, "perplexity": 36.357949601785336, "lr": 0.0026291804804649314, "grad_norm": 0.164494, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:28.690472+00:00", "epoch": 0, "step": 20651, "train_loss": 3.597766399383545, "perplexity": 36.516579830508284, "lr": 0.0026291804804649314, "grad_norm": 0.177539, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:28.996240+00:00", "epoch": 0, "step": 20652, "train_loss": 3.400646924972534, "perplexity": 29.983490843516037, "lr": 0.0026291804804649314, "grad_norm": 0.159254, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:29.301569+00:00", "epoch": 0, "step": 20653, "train_loss": 3.470736265182495, "perplexity": 32.16041232187798, "lr": 0.0026291804804649314, "grad_norm": 0.176677, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:29.606217+00:00", "epoch": 0, "step": 20654, "train_loss": 3.4853062629699707, "perplexity": 32.63241968137295, "lr": 0.0026291804804649314, "grad_norm": 0.158049, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:29.910563+00:00", "epoch": 0, "step": 20655, "train_loss": 3.3914456367492676, "perplexity": 29.708869475042775, "lr": 0.0026291804804649314, "grad_norm": 0.188262, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:30.215732+00:00", "epoch": 0, "step": 20656, "train_loss": 3.4871461391448975, "perplexity": 32.69251455948968, "lr": 0.0026291804804649314, "grad_norm": 0.212647, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:30.520981+00:00", "epoch": 0, "step": 20657, "train_loss": 3.5175893306732178, "perplexity": 33.70308346576829, "lr": 0.0026291804804649314, "grad_norm": 0.204897, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:30.827156+00:00", "epoch": 0, "step": 20658, "train_loss": 3.3979313373565674, "perplexity": 29.902178502460377, "lr": 0.0026291804804649314, "grad_norm": 0.18861, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:31.132254+00:00", "epoch": 0, "step": 20659, "train_loss": 3.489832639694214, "perplexity": 32.78046109913121, "lr": 0.0026291804804649314, "grad_norm": 0.178798, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:31.436935+00:00", "epoch": 0, "step": 20660, "train_loss": 3.5683753490448, "perplexity": 35.45893793212464, "lr": 0.0026291804804649314, "grad_norm": 0.203223, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:31.741404+00:00", "epoch": 0, "step": 20661, "train_loss": 3.6185646057128906, "perplexity": 37.28401213646521, "lr": 0.0026291804804649314, "grad_norm": 0.201394, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:32.045399+00:00", "epoch": 0, "step": 20662, "train_loss": 3.4673705101013184, "perplexity": 32.05235020801068, "lr": 0.0026291804804649314, "grad_norm": 0.166658, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:32.350661+00:00", "epoch": 0, "step": 20663, "train_loss": 3.504725933074951, "perplexity": 33.27232375924097, "lr": 0.0026291804804649314, "grad_norm": 0.167647, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:32.655693+00:00", "epoch": 0, "step": 20664, "train_loss": 3.4600980281829834, "perplexity": 31.820095627941456, "lr": 0.0026291804804649314, "grad_norm": 0.154832, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:32.960600+00:00", "epoch": 0, "step": 20665, "train_loss": 3.5408966541290283, "perplexity": 34.4978379559124, "lr": 0.0026291804804649314, "grad_norm": 0.16399, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:33.264627+00:00", "epoch": 0, "step": 20666, "train_loss": 3.5388216972351074, "perplexity": 34.42633064217313, "lr": 0.0026291804804649314, "grad_norm": 0.163703, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:33.569968+00:00", "epoch": 0, "step": 20667, "train_loss": 3.499069929122925, "perplexity": 33.08466655976504, "lr": 0.0026291804804649314, "grad_norm": 0.177648, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:33.875774+00:00", "epoch": 0, "step": 20668, "train_loss": 3.4598875045776367, "perplexity": 31.81339745177422, "lr": 0.0026291804804649314, "grad_norm": 0.163013, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:34.181483+00:00", "epoch": 0, "step": 20669, "train_loss": 3.4901607036590576, "perplexity": 32.7912169513761, "lr": 0.0026291804804649314, "grad_norm": 0.171802, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:34.486695+00:00", "epoch": 0, "step": 20670, "train_loss": 3.5317225456237793, "perplexity": 34.182798360610356, "lr": 0.0026291804804649314, "grad_norm": 0.198419, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:34.791604+00:00", "epoch": 0, "step": 20671, "train_loss": 3.548957109451294, "perplexity": 34.77702993308784, "lr": 0.0026291804804649314, "grad_norm": 0.20338, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:35.096331+00:00", "epoch": 0, "step": 20672, "train_loss": 3.5529112815856934, "perplexity": 34.9148165322019, "lr": 0.0026291804804649314, "grad_norm": 0.183998, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:35.401458+00:00", "epoch": 0, "step": 20673, "train_loss": 3.4622316360473633, "perplexity": 31.888059712788873, "lr": 0.0026291804804649314, "grad_norm": 0.176965, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:35.706223+00:00", "epoch": 0, "step": 20674, "train_loss": 3.4419424533843994, "perplexity": 31.247596255014734, "lr": 0.0026291804804649314, "grad_norm": 0.146726, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:36.011030+00:00", "epoch": 0, "step": 20675, "train_loss": 3.4993300437927246, "perplexity": 33.09327348622806, "lr": 0.0026291804804649314, "grad_norm": 0.154264, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:36.315297+00:00", "epoch": 0, "step": 20676, "train_loss": 3.4549179077148438, "perplexity": 31.655689887684737, "lr": 0.0026291804804649314, "grad_norm": 0.17096, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:36.620320+00:00", "epoch": 0, "step": 20677, "train_loss": 3.509003162384033, "perplexity": 33.41494190543321, "lr": 0.0026291804804649314, "grad_norm": 0.141614, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:36.925473+00:00", "epoch": 0, "step": 20678, "train_loss": 3.4535751342773438, "perplexity": 31.61321199363417, "lr": 0.0026291804804649314, "grad_norm": 0.173856, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:37.230995+00:00", "epoch": 0, "step": 20679, "train_loss": 3.4011776447296143, "perplexity": 29.999407897866927, "lr": 0.0026291804804649314, "grad_norm": 0.200194, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:37.535548+00:00", "epoch": 0, "step": 20680, "train_loss": 3.5100202560424805, "perplexity": 33.44894532034012, "lr": 0.0026291804804649314, "grad_norm": 0.193511, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:37.839507+00:00", "epoch": 0, "step": 20681, "train_loss": 3.4648971557617188, "perplexity": 31.973171347582948, "lr": 0.0026291804804649314, "grad_norm": 0.17511, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:38.143439+00:00", "epoch": 0, "step": 20682, "train_loss": 3.526162624359131, "perplexity": 33.99327205701374, "lr": 0.0026291804804649314, "grad_norm": 0.192334, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:38.447172+00:00", "epoch": 0, "step": 20683, "train_loss": 3.400494337081909, "perplexity": 29.978916074930673, "lr": 0.0026291804804649314, "grad_norm": 0.233543, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:38.752655+00:00", "epoch": 0, "step": 20684, "train_loss": 3.4561359882354736, "perplexity": 31.69427256053465, "lr": 0.0026291804804649314, "grad_norm": 0.259091, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:39.057060+00:00", "epoch": 0, "step": 20685, "train_loss": 3.4452877044677734, "perplexity": 31.352302346678382, "lr": 0.0026291804804649314, "grad_norm": 0.19785, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:39.361508+00:00", "epoch": 0, "step": 20686, "train_loss": 3.516871452331543, "perplexity": 33.67889743445327, "lr": 0.0026291804804649314, "grad_norm": 0.214556, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:39.666713+00:00", "epoch": 0, "step": 20687, "train_loss": 3.4859542846679688, "perplexity": 32.65357305056553, "lr": 0.0026291804804649314, "grad_norm": 0.203609, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:39.971558+00:00", "epoch": 0, "step": 20688, "train_loss": 3.5324935913085938, "perplexity": 34.209165023425705, "lr": 0.0026291804804649314, "grad_norm": 0.195679, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:40.276887+00:00", "epoch": 0, "step": 20689, "train_loss": 3.622772693634033, "perplexity": 37.441237113840856, "lr": 0.0026291804804649314, "grad_norm": 0.183072, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:40.582882+00:00", "epoch": 0, "step": 20690, "train_loss": 3.4626245498657227, "perplexity": 31.900591433872666, "lr": 0.0026291804804649314, "grad_norm": 0.14954, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:40.887527+00:00", "epoch": 0, "step": 20691, "train_loss": 3.4898695945739746, "perplexity": 32.78167251951344, "lr": 0.0026291804804649314, "grad_norm": 0.176608, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:41.192025+00:00", "epoch": 0, "step": 20692, "train_loss": 3.4577410221099854, "perplexity": 31.745183787814995, "lr": 0.0026291804804649314, "grad_norm": 0.171864, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:41.497074+00:00", "epoch": 0, "step": 20693, "train_loss": 3.5476348400115967, "perplexity": 34.731075717826634, "lr": 0.0026291804804649314, "grad_norm": 0.207795, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:41.802044+00:00", "epoch": 0, "step": 20694, "train_loss": 3.4789223670959473, "perplexity": 32.42476125214875, "lr": 0.0026291804804649314, "grad_norm": 0.18394, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:42.108339+00:00", "epoch": 0, "step": 20695, "train_loss": 3.6441822052001953, "perplexity": 38.251478197044854, "lr": 0.0026291804804649314, "grad_norm": 0.184575, "tokens_per_sec": 106987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:42.414140+00:00", "epoch": 0, "step": 20696, "train_loss": 3.4542341232299805, "perplexity": 31.634051616881923, "lr": 0.0026291804804649314, "grad_norm": 0.165247, "tokens_per_sec": 107150} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:42.719563+00:00", "epoch": 0, "step": 20697, "train_loss": 3.506840229034424, "perplexity": 33.34274571912639, "lr": 0.0026291804804649314, "grad_norm": 0.160784, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:43.023470+00:00", "epoch": 0, "step": 20698, "train_loss": 3.3913557529449463, "perplexity": 29.706199248839116, "lr": 0.0026291804804649314, "grad_norm": 0.170392, "tokens_per_sec": 107821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:43.327575+00:00", "epoch": 0, "step": 20699, "train_loss": 3.5335116386413574, "perplexity": 34.24400930618899, "lr": 0.0026291804804649314, "grad_norm": 0.186017, "tokens_per_sec": 107753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:43.633647+00:00", "epoch": 0, "step": 20700, "train_loss": 3.4900779724121094, "perplexity": 32.78850420532474, "lr": 0.0026291804804649314, "grad_norm": 0.159705, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:43.938674+00:00", "epoch": 0, "step": 20701, "train_loss": 3.405379056930542, "perplexity": 30.12571292016729, "lr": 0.0026291804804649314, "grad_norm": 0.155308, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:44.243317+00:00", "epoch": 0, "step": 20702, "train_loss": 3.454284906387329, "perplexity": 31.635658134694435, "lr": 0.0026291804804649314, "grad_norm": 0.16468, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:44.548479+00:00", "epoch": 0, "step": 20703, "train_loss": 3.418632745742798, "perplexity": 30.52764741819555, "lr": 0.0026291804804649314, "grad_norm": 0.142426, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:44.853044+00:00", "epoch": 0, "step": 20704, "train_loss": 3.5475974082946777, "perplexity": 34.729775698363206, "lr": 0.0026291804804649314, "grad_norm": 0.163325, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:45.157462+00:00", "epoch": 0, "step": 20705, "train_loss": 3.4952094554901123, "perplexity": 32.95719029455858, "lr": 0.0026291804804649314, "grad_norm": 0.152741, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:45.461995+00:00", "epoch": 0, "step": 20706, "train_loss": 3.426513671875, "perplexity": 30.769184068606357, "lr": 0.0026291804804649314, "grad_norm": 0.15072, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:45.766882+00:00", "epoch": 0, "step": 20707, "train_loss": 3.5600104331970215, "perplexity": 35.16356401158412, "lr": 0.0026291804804649314, "grad_norm": 0.152144, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:46.071096+00:00", "epoch": 0, "step": 20708, "train_loss": 3.5615105628967285, "perplexity": 35.216353503948376, "lr": 0.0026291804804649314, "grad_norm": 0.156, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:46.375351+00:00", "epoch": 0, "step": 20709, "train_loss": 3.509251832962036, "perplexity": 33.42325225157723, "lr": 0.0026291804804649314, "grad_norm": 0.159443, "tokens_per_sec": 107700} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:46.679674+00:00", "epoch": 0, "step": 20710, "train_loss": 3.381080150604248, "perplexity": 29.402513109436665, "lr": 0.0026291804804649314, "grad_norm": 0.165996, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:46.986274+00:00", "epoch": 0, "step": 20711, "train_loss": 3.5145645141601562, "perplexity": 33.601291850451375, "lr": 0.0026291804804649314, "grad_norm": 0.207559, "tokens_per_sec": 106822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:47.293021+00:00", "epoch": 0, "step": 20712, "train_loss": 3.4660277366638184, "perplexity": 32.00934004645557, "lr": 0.0026291804804649314, "grad_norm": 0.184675, "tokens_per_sec": 106824} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:47.598757+00:00", "epoch": 0, "step": 20713, "train_loss": 3.5114071369171143, "perplexity": 33.49536720625364, "lr": 0.0026291804804649314, "grad_norm": 0.152278, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:47.902871+00:00", "epoch": 0, "step": 20714, "train_loss": 3.479032278060913, "perplexity": 32.428325284806306, "lr": 0.0026291804804649314, "grad_norm": 0.178983, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:48.206482+00:00", "epoch": 0, "step": 20715, "train_loss": 3.4508309364318848, "perplexity": 31.52657801015591, "lr": 0.0026291804804649314, "grad_norm": 0.175254, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:48.510871+00:00", "epoch": 0, "step": 20716, "train_loss": 3.567836284637451, "perplexity": 35.439828431851524, "lr": 0.0026291804804649314, "grad_norm": 0.155031, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:48.815983+00:00", "epoch": 0, "step": 20717, "train_loss": 3.6218016147613525, "perplexity": 37.40489636723396, "lr": 0.0026291804804649314, "grad_norm": 0.176847, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:49.120683+00:00", "epoch": 0, "step": 20718, "train_loss": 3.5045266151428223, "perplexity": 33.26569264934366, "lr": 0.0026291804804649314, "grad_norm": 0.168399, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:49.425415+00:00", "epoch": 0, "step": 20719, "train_loss": 3.3563859462738037, "perplexity": 28.68533298372249, "lr": 0.0026291804804649314, "grad_norm": 0.155137, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:49.729521+00:00", "epoch": 0, "step": 20720, "train_loss": 3.5058586597442627, "perplexity": 33.310033561130155, "lr": 0.0026291804804649314, "grad_norm": 0.171138, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:50.034783+00:00", "epoch": 0, "step": 20721, "train_loss": 3.537346601486206, "perplexity": 34.37558594403814, "lr": 0.0026291804804649314, "grad_norm": 0.145725, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:50.340022+00:00", "epoch": 0, "step": 20722, "train_loss": 3.4459965229034424, "perplexity": 31.374533314511513, "lr": 0.0026291804804649314, "grad_norm": 0.1556, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:50.646611+00:00", "epoch": 0, "step": 20723, "train_loss": 3.5681347846984863, "perplexity": 35.450408801843494, "lr": 0.0026291804804649314, "grad_norm": 0.171402, "tokens_per_sec": 106880} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:50.951153+00:00", "epoch": 0, "step": 20724, "train_loss": 3.5594139099121094, "perplexity": 35.14259438192872, "lr": 0.0026291804804649314, "grad_norm": 0.165875, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:51.255713+00:00", "epoch": 0, "step": 20725, "train_loss": 3.5296690464019775, "perplexity": 34.11267603349125, "lr": 0.0026291804804649314, "grad_norm": 0.236022, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:51.559718+00:00", "epoch": 0, "step": 20726, "train_loss": 3.545276165008545, "perplexity": 34.64925293231538, "lr": 0.0026291804804649314, "grad_norm": 0.241086, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:51.864822+00:00", "epoch": 0, "step": 20727, "train_loss": 3.524156093597412, "perplexity": 33.925131896465956, "lr": 0.0026291804804649314, "grad_norm": 0.175362, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:52.169504+00:00", "epoch": 0, "step": 20728, "train_loss": 3.5494906902313232, "perplexity": 34.79559123938441, "lr": 0.0026291804804649314, "grad_norm": 0.199388, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:52.475050+00:00", "epoch": 0, "step": 20729, "train_loss": 3.462789535522461, "perplexity": 31.90585500809591, "lr": 0.0026291804804649314, "grad_norm": 0.184549, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:52.780857+00:00", "epoch": 0, "step": 20730, "train_loss": 3.4606471061706543, "perplexity": 31.83757213956118, "lr": 0.0026291804804649314, "grad_norm": 0.170605, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:53.085403+00:00", "epoch": 0, "step": 20731, "train_loss": 3.598897933959961, "perplexity": 36.55792298939464, "lr": 0.0026291804804649314, "grad_norm": 0.183137, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:53.390509+00:00", "epoch": 0, "step": 20732, "train_loss": 3.5116846561431885, "perplexity": 33.50466410461214, "lr": 0.0026291804804649314, "grad_norm": 0.189342, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:53.696238+00:00", "epoch": 0, "step": 20733, "train_loss": 3.5148372650146484, "perplexity": 33.6104578814802, "lr": 0.0026291804804649314, "grad_norm": 0.165815, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:54.000833+00:00", "epoch": 0, "step": 20734, "train_loss": 3.4263482093811035, "perplexity": 30.764093343848824, "lr": 0.0026291804804649314, "grad_norm": 0.175867, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:54.306805+00:00", "epoch": 0, "step": 20735, "train_loss": 3.5083816051483154, "perplexity": 33.39417905982771, "lr": 0.0026291804804649314, "grad_norm": 0.162636, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:54.610885+00:00", "epoch": 0, "step": 20736, "train_loss": 3.5556182861328125, "perplexity": 35.009459140507836, "lr": 0.0026291804804649314, "grad_norm": 0.204674, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:54.915836+00:00", "epoch": 0, "step": 20737, "train_loss": 3.524059772491455, "perplexity": 33.92186434761144, "lr": 0.0026291804804649314, "grad_norm": 0.180071, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:55.220785+00:00", "epoch": 0, "step": 20738, "train_loss": 3.4527506828308105, "perplexity": 31.587159176393605, "lr": 0.0026291804804649314, "grad_norm": 0.190403, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:55.525867+00:00", "epoch": 0, "step": 20739, "train_loss": 3.4140799045562744, "perplexity": 30.38897580216098, "lr": 0.0026291804804649314, "grad_norm": 0.207283, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:55.830357+00:00", "epoch": 0, "step": 20740, "train_loss": 3.5203402042388916, "perplexity": 33.79592402492992, "lr": 0.0026291804804649314, "grad_norm": 0.215788, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:56.134686+00:00", "epoch": 0, "step": 20741, "train_loss": 3.5105323791503906, "perplexity": 33.46607968525151, "lr": 0.0026291804804649314, "grad_norm": 0.167438, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:56.439971+00:00", "epoch": 0, "step": 20742, "train_loss": 3.521808385848999, "perplexity": 33.84557902143675, "lr": 0.0026291804804649314, "grad_norm": 0.191431, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:56.745026+00:00", "epoch": 0, "step": 20743, "train_loss": 3.4653003215789795, "perplexity": 31.986064436191377, "lr": 0.0026291804804649314, "grad_norm": 0.205258, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:57.049858+00:00", "epoch": 0, "step": 20744, "train_loss": 3.5793001651763916, "perplexity": 35.84844407675306, "lr": 0.0026291804804649314, "grad_norm": 0.198629, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:57.354388+00:00", "epoch": 0, "step": 20745, "train_loss": 3.458091974258423, "perplexity": 31.756326783482635, "lr": 0.0026291804804649314, "grad_norm": 0.153902, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:57.659846+00:00", "epoch": 0, "step": 20746, "train_loss": 3.509500026702881, "perplexity": 33.431548723108314, "lr": 0.0026291804804649314, "grad_norm": 0.170888, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:57.965480+00:00", "epoch": 0, "step": 20747, "train_loss": 3.443727493286133, "perplexity": 31.30342427395987, "lr": 0.0026291804804649314, "grad_norm": 0.185534, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:58.270005+00:00", "epoch": 0, "step": 20748, "train_loss": 3.463106632232666, "perplexity": 31.915973853996842, "lr": 0.0026291804804649314, "grad_norm": 0.205988, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:58.575768+00:00", "epoch": 0, "step": 20749, "train_loss": 3.4649391174316406, "perplexity": 31.97451302339467, "lr": 0.0026291804804649314, "grad_norm": 0.208675, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:58.882188+00:00", "epoch": 0, "step": 20750, "train_loss": 3.4897289276123047, "perplexity": 32.777061545555114, "lr": 0.0026291804804649314, "grad_norm": 0.187055, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:59.188184+00:00", "epoch": 0, "step": 20751, "train_loss": 3.423084020614624, "perplexity": 30.66383725236714, "lr": 0.0026291804804649314, "grad_norm": 0.219872, "tokens_per_sec": 106985} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:59.493341+00:00", "epoch": 0, "step": 20752, "train_loss": 3.5155389308929443, "perplexity": 33.63404946866749, "lr": 0.0026291804804649314, "grad_norm": 0.193816, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:46:59.798364+00:00", "epoch": 0, "step": 20753, "train_loss": 3.522303342819214, "perplexity": 33.86233527315408, "lr": 0.0026291804804649314, "grad_norm": 0.174006, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:00.103108+00:00", "epoch": 0, "step": 20754, "train_loss": 3.516741991043091, "perplexity": 33.67453760321853, "lr": 0.0026291804804649314, "grad_norm": 0.199638, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:00.409648+00:00", "epoch": 0, "step": 20755, "train_loss": 3.4694509506225586, "perplexity": 32.119102629328836, "lr": 0.0026291804804649314, "grad_norm": 0.173367, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:00.714675+00:00", "epoch": 0, "step": 20756, "train_loss": 3.489935874938965, "perplexity": 32.783845372740984, "lr": 0.0026291804804649314, "grad_norm": 0.168215, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:01.018451+00:00", "epoch": 0, "step": 20757, "train_loss": 3.5691561698913574, "perplexity": 35.48663582219659, "lr": 0.0026291804804649314, "grad_norm": 0.174348, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:01.322816+00:00", "epoch": 0, "step": 20758, "train_loss": 3.508509397506714, "perplexity": 33.39844685341644, "lr": 0.0026291804804649314, "grad_norm": 0.17193, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:01.628023+00:00", "epoch": 0, "step": 20759, "train_loss": 3.607405185699463, "perplexity": 36.870257112123696, "lr": 0.0026291804804649314, "grad_norm": 0.161847, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:01.934616+00:00", "epoch": 0, "step": 20760, "train_loss": 3.4357988834381104, "perplexity": 31.056212952077882, "lr": 0.0026291804804649314, "grad_norm": 0.167248, "tokens_per_sec": 106935} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:02.240308+00:00", "epoch": 0, "step": 20761, "train_loss": 3.397793769836426, "perplexity": 29.8980652168507, "lr": 0.0026291804804649314, "grad_norm": 0.168785, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:02.545808+00:00", "epoch": 0, "step": 20762, "train_loss": 3.5873379707336426, "perplexity": 36.13774802664985, "lr": 0.0026291804804649314, "grad_norm": 0.174948, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:02.851042+00:00", "epoch": 0, "step": 20763, "train_loss": 3.4724349975585938, "perplexity": 32.21509068432137, "lr": 0.0026291804804649314, "grad_norm": 0.16498, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:03.157575+00:00", "epoch": 0, "step": 20764, "train_loss": 3.4857306480407715, "perplexity": 32.64627133211883, "lr": 0.0026291804804649314, "grad_norm": 0.162014, "tokens_per_sec": 106899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:03.462838+00:00", "epoch": 0, "step": 20765, "train_loss": 3.4076733589172363, "perplexity": 30.194909752031712, "lr": 0.0026291804804649314, "grad_norm": 0.178951, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:03.768069+00:00", "epoch": 0, "step": 20766, "train_loss": 3.402940034866333, "perplexity": 30.052325175200508, "lr": 0.0026291804804649314, "grad_norm": 0.194143, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:04.073460+00:00", "epoch": 0, "step": 20767, "train_loss": 3.416874885559082, "perplexity": 30.474031220992217, "lr": 0.0026291804804649314, "grad_norm": 0.16469, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:04.378832+00:00", "epoch": 0, "step": 20768, "train_loss": 3.432403802871704, "perplexity": 30.950953390470936, "lr": 0.0026291804804649314, "grad_norm": 0.16647, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:04.684519+00:00", "epoch": 0, "step": 20769, "train_loss": 3.5445523262023926, "perplexity": 34.624181533359184, "lr": 0.0026291804804649314, "grad_norm": 0.188419, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:04.991349+00:00", "epoch": 0, "step": 20770, "train_loss": 3.5279788970947266, "perplexity": 34.05506921350154, "lr": 0.0026291804804649314, "grad_norm": 0.172532, "tokens_per_sec": 106852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:05.297284+00:00", "epoch": 0, "step": 20771, "train_loss": 3.5338025093078613, "perplexity": 34.253971332760074, "lr": 0.0026291804804649314, "grad_norm": 0.187528, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:05.602707+00:00", "epoch": 0, "step": 20772, "train_loss": 3.4542129039764404, "perplexity": 31.633380373041838, "lr": 0.0026291804804649314, "grad_norm": 0.159798, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:05.907090+00:00", "epoch": 0, "step": 20773, "train_loss": 3.5172853469848633, "perplexity": 33.69283983517463, "lr": 0.0026291804804649314, "grad_norm": 0.180071, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:06.212222+00:00", "epoch": 0, "step": 20774, "train_loss": 3.488980531692505, "perplexity": 32.75254050329989, "lr": 0.0026291804804649314, "grad_norm": 0.172144, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:06.518542+00:00", "epoch": 0, "step": 20775, "train_loss": 3.5217370986938477, "perplexity": 33.84316635239104, "lr": 0.0026291804804649314, "grad_norm": 0.200165, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:06.824342+00:00", "epoch": 0, "step": 20776, "train_loss": 3.546306848526001, "perplexity": 34.684983756627624, "lr": 0.0026291804804649314, "grad_norm": 0.173504, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:07.130145+00:00", "epoch": 0, "step": 20777, "train_loss": 3.4876294136047363, "perplexity": 32.708317835156365, "lr": 0.0026291804804649314, "grad_norm": 0.175876, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:07.435283+00:00", "epoch": 0, "step": 20778, "train_loss": 3.5251598358154297, "perplexity": 33.95920107908033, "lr": 0.0026291804804649314, "grad_norm": 0.187172, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:07.741659+00:00", "epoch": 0, "step": 20779, "train_loss": 3.4575159549713135, "perplexity": 31.738039794102527, "lr": 0.0026291804804649314, "grad_norm": 0.218833, "tokens_per_sec": 106954} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:08.048165+00:00", "epoch": 0, "step": 20780, "train_loss": 3.3773093223571777, "perplexity": 29.291850059274278, "lr": 0.0026291804804649314, "grad_norm": 0.189447, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:08.352818+00:00", "epoch": 0, "step": 20781, "train_loss": 3.5244357585906982, "perplexity": 33.93462089505853, "lr": 0.0026291804804649314, "grad_norm": 0.169249, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:08.657430+00:00", "epoch": 0, "step": 20782, "train_loss": 3.4733145236968994, "perplexity": 32.243437162533056, "lr": 0.0026291804804649314, "grad_norm": 0.206706, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:08.963755+00:00", "epoch": 0, "step": 20783, "train_loss": 3.5235679149627686, "perplexity": 33.90518372582597, "lr": 0.0026291804804649314, "grad_norm": 0.18505, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:09.268455+00:00", "epoch": 0, "step": 20784, "train_loss": 3.439218521118164, "perplexity": 31.16259573971382, "lr": 0.0026291804804649314, "grad_norm": 0.186231, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:09.573236+00:00", "epoch": 0, "step": 20785, "train_loss": 3.571885108947754, "perplexity": 35.58360894536274, "lr": 0.0026291804804649314, "grad_norm": 0.178134, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:09.877651+00:00", "epoch": 0, "step": 20786, "train_loss": 3.522096872329712, "perplexity": 33.85534442194162, "lr": 0.0026291804804649314, "grad_norm": 0.185371, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:10.182529+00:00", "epoch": 0, "step": 20787, "train_loss": 3.3976006507873535, "perplexity": 29.89229188841471, "lr": 0.0026291804804649314, "grad_norm": 0.16248, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:10.488366+00:00", "epoch": 0, "step": 20788, "train_loss": 3.440732717514038, "perplexity": 31.209817772562428, "lr": 0.0026291804804649314, "grad_norm": 0.178065, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:10.793200+00:00", "epoch": 0, "step": 20789, "train_loss": 3.551968574523926, "perplexity": 34.88191759756211, "lr": 0.0026291804804649314, "grad_norm": 0.160914, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:11.099018+00:00", "epoch": 0, "step": 20790, "train_loss": 3.5157833099365234, "perplexity": 33.64226992992166, "lr": 0.0026291804804649314, "grad_norm": 0.189594, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:11.404196+00:00", "epoch": 0, "step": 20791, "train_loss": 3.457048177719116, "perplexity": 31.72319693290458, "lr": 0.0026291804804649314, "grad_norm": 0.18641, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:11.708559+00:00", "epoch": 0, "step": 20792, "train_loss": 3.3934319019317627, "perplexity": 29.767937811362312, "lr": 0.0026291804804649314, "grad_norm": 0.173927, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:12.014113+00:00", "epoch": 0, "step": 20793, "train_loss": 3.4351911544799805, "perplexity": 31.037344926040674, "lr": 0.0026291804804649314, "grad_norm": 0.179435, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:12.318831+00:00", "epoch": 0, "step": 20794, "train_loss": 3.4462661743164062, "perplexity": 31.38299464250411, "lr": 0.0026291804804649314, "grad_norm": 0.227333, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:12.624206+00:00", "epoch": 0, "step": 20795, "train_loss": 3.4640309810638428, "perplexity": 31.945488986162083, "lr": 0.0026291804804649314, "grad_norm": 0.191363, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:12.929186+00:00", "epoch": 0, "step": 20796, "train_loss": 3.5090157985687256, "perplexity": 33.41536414547836, "lr": 0.0026291804804649314, "grad_norm": 0.181373, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:13.233262+00:00", "epoch": 0, "step": 20797, "train_loss": 3.4739813804626465, "perplexity": 32.264946087644866, "lr": 0.0026291804804649314, "grad_norm": 0.176731, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:13.539077+00:00", "epoch": 0, "step": 20798, "train_loss": 3.4823155403137207, "perplexity": 32.53497095819093, "lr": 0.0026291804804649314, "grad_norm": 0.203823, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:13.844354+00:00", "epoch": 0, "step": 20799, "train_loss": 3.4764487743377686, "perplexity": 32.344654713851654, "lr": 0.0026291804804649314, "grad_norm": 0.169858, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:14.150074+00:00", "epoch": 0, "step": 20800, "train_loss": 3.4509849548339844, "perplexity": 31.531434057275412, "lr": 0.0026291804804649314, "grad_norm": 0.187446, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:14.454590+00:00", "epoch": 0, "step": 20801, "train_loss": 3.420259475708008, "perplexity": 30.577348070828528, "lr": 0.0026291804804649314, "grad_norm": 0.206788, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:14.758771+00:00", "epoch": 0, "step": 20802, "train_loss": 3.496730089187622, "perplexity": 33.00734423191497, "lr": 0.0026291804804649314, "grad_norm": 0.173038, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:15.063312+00:00", "epoch": 0, "step": 20803, "train_loss": 3.4083375930786133, "perplexity": 30.214972905166885, "lr": 0.0026291804804649314, "grad_norm": 0.180817, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:15.367351+00:00", "epoch": 0, "step": 20804, "train_loss": 3.5932111740112305, "perplexity": 36.35061686588647, "lr": 0.0026291804804649314, "grad_norm": 0.18567, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:15.671847+00:00", "epoch": 0, "step": 20805, "train_loss": 3.571140766143799, "perplexity": 35.55713239714201, "lr": 0.0026291804804649314, "grad_norm": 0.169537, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:15.976143+00:00", "epoch": 0, "step": 20806, "train_loss": 3.4815001487731934, "perplexity": 32.508453030816376, "lr": 0.0026291804804649314, "grad_norm": 0.169178, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:16.280580+00:00", "epoch": 0, "step": 20807, "train_loss": 3.466636896133423, "perplexity": 32.028844779203844, "lr": 0.0026291804804649314, "grad_norm": 0.166272, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:16.584915+00:00", "epoch": 0, "step": 20808, "train_loss": 3.405468225479126, "perplexity": 30.128399306032225, "lr": 0.0026291804804649314, "grad_norm": 0.160388, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:16.890861+00:00", "epoch": 0, "step": 20809, "train_loss": 3.495593547821045, "perplexity": 32.969851329947446, "lr": 0.0026291804804649314, "grad_norm": 0.16717, "tokens_per_sec": 107104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:17.196374+00:00", "epoch": 0, "step": 20810, "train_loss": 3.524127244949341, "perplexity": 33.92415321639197, "lr": 0.0026291804804649314, "grad_norm": 0.153304, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:17.501218+00:00", "epoch": 0, "step": 20811, "train_loss": 3.5127205848693848, "perplexity": 33.53939053256625, "lr": 0.0026291804804649314, "grad_norm": 0.143881, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:17.805778+00:00", "epoch": 0, "step": 20812, "train_loss": 3.5599517822265625, "perplexity": 35.16150169490908, "lr": 0.0026291804804649314, "grad_norm": 0.142683, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:18.110489+00:00", "epoch": 0, "step": 20813, "train_loss": 3.4781014919281006, "perplexity": 32.39815549231128, "lr": 0.0026291804804649314, "grad_norm": 0.153506, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:18.415605+00:00", "epoch": 0, "step": 20814, "train_loss": 3.4378716945648193, "perplexity": 31.120653379182947, "lr": 0.0026291804804649314, "grad_norm": 0.155058, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:18.721347+00:00", "epoch": 0, "step": 20815, "train_loss": 3.558438301086426, "perplexity": 35.10832567583646, "lr": 0.0026291804804649314, "grad_norm": 0.15937, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:19.026392+00:00", "epoch": 0, "step": 20816, "train_loss": 3.559644937515259, "perplexity": 35.150714229195536, "lr": 0.0026291804804649314, "grad_norm": 0.183874, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:19.332073+00:00", "epoch": 0, "step": 20817, "train_loss": 3.408444404602051, "perplexity": 30.218200384816452, "lr": 0.0026291804804649314, "grad_norm": 0.19279, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:19.637311+00:00", "epoch": 0, "step": 20818, "train_loss": 3.5197017192840576, "perplexity": 33.774352723113935, "lr": 0.0026291804804649314, "grad_norm": 0.184929, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:19.943502+00:00", "epoch": 0, "step": 20819, "train_loss": 3.5266380310058594, "perplexity": 34.00943652653738, "lr": 0.0026291804804649314, "grad_norm": 0.194379, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:20.248232+00:00", "epoch": 0, "step": 20820, "train_loss": 3.4284558296203613, "perplexity": 30.829000745672722, "lr": 0.0026291804804649314, "grad_norm": 0.195803, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:20.553391+00:00", "epoch": 0, "step": 20821, "train_loss": 3.6258223056793213, "perplexity": 37.555592642906355, "lr": 0.0026291804804649314, "grad_norm": 0.185964, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:20.859884+00:00", "epoch": 0, "step": 20822, "train_loss": 3.5045132637023926, "perplexity": 33.26524850739488, "lr": 0.0026291804804649314, "grad_norm": 0.181839, "tokens_per_sec": 106913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:21.164863+00:00", "epoch": 0, "step": 20823, "train_loss": 3.3985507488250732, "perplexity": 29.920705992235565, "lr": 0.0026291804804649314, "grad_norm": 0.171805, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:21.469852+00:00", "epoch": 0, "step": 20824, "train_loss": 3.5132811069488525, "perplexity": 33.55819537126294, "lr": 0.0026291804804649314, "grad_norm": 0.181202, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:21.775783+00:00", "epoch": 0, "step": 20825, "train_loss": 3.6346936225891113, "perplexity": 37.89024240427082, "lr": 0.0026291804804649314, "grad_norm": 0.170105, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:22.080594+00:00", "epoch": 0, "step": 20826, "train_loss": 3.529741048812866, "perplexity": 34.11513231683562, "lr": 0.0026291804804649314, "grad_norm": 0.151716, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:22.386506+00:00", "epoch": 0, "step": 20827, "train_loss": 3.4720356464385986, "perplexity": 32.20222812028646, "lr": 0.0026291804804649314, "grad_norm": 0.203192, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:22.690961+00:00", "epoch": 0, "step": 20828, "train_loss": 3.5331854820251465, "perplexity": 34.23284221719717, "lr": 0.0026291804804649314, "grad_norm": 0.188819, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:22.995544+00:00", "epoch": 0, "step": 20829, "train_loss": 3.579444169998169, "perplexity": 35.85360679727273, "lr": 0.0026291804804649314, "grad_norm": 0.159351, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:23.301340+00:00", "epoch": 0, "step": 20830, "train_loss": 3.462843418121338, "perplexity": 31.90757422480065, "lr": 0.0026291804804649314, "grad_norm": 0.207771, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:23.606971+00:00", "epoch": 0, "step": 20831, "train_loss": 3.5261898040771484, "perplexity": 33.99419599711891, "lr": 0.0026291804804649314, "grad_norm": 0.180349, "tokens_per_sec": 107158} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:23.912588+00:00", "epoch": 0, "step": 20832, "train_loss": 3.4511091709136963, "perplexity": 31.535351011671025, "lr": 0.0026291804804649314, "grad_norm": 0.20016, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:24.218018+00:00", "epoch": 0, "step": 20833, "train_loss": 3.5723206996917725, "perplexity": 35.599112212352786, "lr": 0.0026291804804649314, "grad_norm": 0.171713, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:24.521972+00:00", "epoch": 0, "step": 20834, "train_loss": 3.545919179916382, "perplexity": 34.671540083213486, "lr": 0.0026291804804649314, "grad_norm": 0.148923, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:24.826775+00:00", "epoch": 0, "step": 20835, "train_loss": 3.4994020462036133, "perplexity": 33.09565636748883, "lr": 0.0026291804804649314, "grad_norm": 0.164657, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:25.131151+00:00", "epoch": 0, "step": 20836, "train_loss": 3.413367986679077, "perplexity": 30.367349046168233, "lr": 0.0026291804804649314, "grad_norm": 0.164726, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:25.436604+00:00", "epoch": 0, "step": 20837, "train_loss": 3.4200241565704346, "perplexity": 30.57015348219662, "lr": 0.0026291804804649314, "grad_norm": 0.156596, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:25.740920+00:00", "epoch": 0, "step": 20838, "train_loss": 3.5033719539642334, "perplexity": 33.22730421254653, "lr": 0.0026291804804649314, "grad_norm": 0.180383, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:26.045795+00:00", "epoch": 0, "step": 20839, "train_loss": 3.586428165435791, "perplexity": 36.1048846639404, "lr": 0.0026291804804649314, "grad_norm": 0.235779, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:26.351045+00:00", "epoch": 0, "step": 20840, "train_loss": 3.5217983722686768, "perplexity": 33.845240107709536, "lr": 0.0026291804804649314, "grad_norm": 0.183659, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:26.656733+00:00", "epoch": 0, "step": 20841, "train_loss": 3.545670986175537, "perplexity": 34.66293589177678, "lr": 0.0026291804804649314, "grad_norm": 0.173513, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:26.962402+00:00", "epoch": 0, "step": 20842, "train_loss": 3.474106550216675, "perplexity": 32.26898493577593, "lr": 0.0026291804804649314, "grad_norm": 0.165053, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:27.267565+00:00", "epoch": 0, "step": 20843, "train_loss": 3.4367125034332275, "perplexity": 31.08459949448536, "lr": 0.0026291804804649314, "grad_norm": 0.17499, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:27.572591+00:00", "epoch": 0, "step": 20844, "train_loss": 3.4895312786102295, "perplexity": 32.77058383222755, "lr": 0.0026291804804649314, "grad_norm": 0.18382, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:27.877723+00:00", "epoch": 0, "step": 20845, "train_loss": 3.486396074295044, "perplexity": 32.66800224752629, "lr": 0.0026291804804649314, "grad_norm": 0.186751, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:28.183249+00:00", "epoch": 0, "step": 20846, "train_loss": 3.4389150142669678, "perplexity": 31.153139113553763, "lr": 0.0026291804804649314, "grad_norm": 0.182977, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:28.489707+00:00", "epoch": 0, "step": 20847, "train_loss": 3.43351149559021, "perplexity": 30.985256531334603, "lr": 0.0026291804804649314, "grad_norm": 0.167763, "tokens_per_sec": 106925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:28.796502+00:00", "epoch": 0, "step": 20848, "train_loss": 3.5500690937042236, "perplexity": 34.81572295176381, "lr": 0.0026291804804649314, "grad_norm": 0.167176, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:29.101162+00:00", "epoch": 0, "step": 20849, "train_loss": 3.5057406425476074, "perplexity": 33.3061026363117, "lr": 0.0026291804804649314, "grad_norm": 0.16578, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:29.405258+00:00", "epoch": 0, "step": 20850, "train_loss": 3.5413992404937744, "perplexity": 34.5151804565694, "lr": 0.0026291804804649314, "grad_norm": 0.162254, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:29.710134+00:00", "epoch": 0, "step": 20851, "train_loss": 3.4077539443969727, "perplexity": 30.197343121365474, "lr": 0.0026291804804649314, "grad_norm": 0.15929, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:30.015449+00:00", "epoch": 0, "step": 20852, "train_loss": 3.485886812210083, "perplexity": 32.65136990805961, "lr": 0.0026291804804649314, "grad_norm": 0.180284, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:30.320795+00:00", "epoch": 0, "step": 20853, "train_loss": 3.518115758895874, "perplexity": 33.7208303909265, "lr": 0.0026291804804649314, "grad_norm": 0.155643, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:30.625331+00:00", "epoch": 0, "step": 20854, "train_loss": 3.500821828842163, "perplexity": 33.14267837842934, "lr": 0.0026291804804649314, "grad_norm": 0.165471, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:30.930415+00:00", "epoch": 0, "step": 20855, "train_loss": 3.4633142948150635, "perplexity": 31.92260229576294, "lr": 0.0026291804804649314, "grad_norm": 0.141375, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:31.235439+00:00", "epoch": 0, "step": 20856, "train_loss": 3.5077741146087646, "perplexity": 33.37389857269853, "lr": 0.0026291804804649314, "grad_norm": 0.170393, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:31.542373+00:00", "epoch": 0, "step": 20857, "train_loss": 3.4866459369659424, "perplexity": 32.67616578165952, "lr": 0.0026291804804649314, "grad_norm": 0.171726, "tokens_per_sec": 106759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:31.846778+00:00", "epoch": 0, "step": 20858, "train_loss": 3.3675475120544434, "perplexity": 29.00729969853419, "lr": 0.0026291804804649314, "grad_norm": 0.162276, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:32.151652+00:00", "epoch": 0, "step": 20859, "train_loss": 3.4339756965637207, "perplexity": 30.9996432564885, "lr": 0.0026291804804649314, "grad_norm": 0.156282, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:32.457035+00:00", "epoch": 0, "step": 20860, "train_loss": 3.6292428970336914, "perplexity": 37.68427493771148, "lr": 0.0026291804804649314, "grad_norm": 0.177388, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:32.761767+00:00", "epoch": 0, "step": 20861, "train_loss": 3.450249671936035, "perplexity": 31.508258054568657, "lr": 0.0026291804804649314, "grad_norm": 0.184034, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:33.067346+00:00", "epoch": 0, "step": 20862, "train_loss": 3.453341007232666, "perplexity": 31.605811352116298, "lr": 0.0026291804804649314, "grad_norm": 0.172218, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:33.372579+00:00", "epoch": 0, "step": 20863, "train_loss": 3.5026490688323975, "perplexity": 33.20329336794717, "lr": 0.0026291804804649314, "grad_norm": 0.185622, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:33.678324+00:00", "epoch": 0, "step": 20864, "train_loss": 3.5237855911254883, "perplexity": 33.91256487943717, "lr": 0.0026291804804649314, "grad_norm": 0.204625, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:33.984178+00:00", "epoch": 0, "step": 20865, "train_loss": 3.3853871822357178, "perplexity": 29.52942377137319, "lr": 0.0026291804804649314, "grad_norm": 0.205421, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:34.288995+00:00", "epoch": 0, "step": 20866, "train_loss": 3.5917656421661377, "perplexity": 36.29810885176457, "lr": 0.0026291804804649314, "grad_norm": 0.168491, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:34.593020+00:00", "epoch": 0, "step": 20867, "train_loss": 3.5138285160064697, "perplexity": 33.57657046025268, "lr": 0.0026291804804649314, "grad_norm": 0.203279, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:34.897689+00:00", "epoch": 0, "step": 20868, "train_loss": 3.452204942703247, "perplexity": 31.569925499097664, "lr": 0.0026291804804649314, "grad_norm": 0.189037, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:35.203518+00:00", "epoch": 0, "step": 20869, "train_loss": 3.516117572784424, "perplexity": 33.65351717054089, "lr": 0.0026291804804649314, "grad_norm": 0.159035, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:35.508789+00:00", "epoch": 0, "step": 20870, "train_loss": 3.4119303226470947, "perplexity": 30.32372236846484, "lr": 0.0026291804804649314, "grad_norm": 0.205811, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:35.812919+00:00", "epoch": 0, "step": 20871, "train_loss": 3.5573887825012207, "perplexity": 35.07149816450346, "lr": 0.0026291804804649314, "grad_norm": 0.178418, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:36.117505+00:00", "epoch": 0, "step": 20872, "train_loss": 3.4618828296661377, "perplexity": 31.876938893695232, "lr": 0.0026291804804649314, "grad_norm": 0.146611, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:36.422558+00:00", "epoch": 0, "step": 20873, "train_loss": 3.5950350761413574, "perplexity": 36.41697733252148, "lr": 0.0026291804804649314, "grad_norm": 0.165977, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:36.729169+00:00", "epoch": 0, "step": 20874, "train_loss": 3.471909761428833, "perplexity": 32.19817459762929, "lr": 0.0026291804804649314, "grad_norm": 0.165881, "tokens_per_sec": 106872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:37.035602+00:00", "epoch": 0, "step": 20875, "train_loss": 3.5191490650177, "perplexity": 33.75569233983321, "lr": 0.0026291804804649314, "grad_norm": 0.169418, "tokens_per_sec": 106934} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:37.341316+00:00", "epoch": 0, "step": 20876, "train_loss": 3.494079351425171, "perplexity": 32.91996627730593, "lr": 0.0026291804804649314, "grad_norm": 0.157573, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:37.645269+00:00", "epoch": 0, "step": 20877, "train_loss": 3.506354570388794, "perplexity": 33.32655645794439, "lr": 0.0026291804804649314, "grad_norm": 0.171744, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:37.950070+00:00", "epoch": 0, "step": 20878, "train_loss": 3.5100104808807373, "perplexity": 33.44861835308756, "lr": 0.0026291804804649314, "grad_norm": 0.192445, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:38.255545+00:00", "epoch": 0, "step": 20879, "train_loss": 3.4941697120666504, "perplexity": 32.92294108097679, "lr": 0.0026291804804649314, "grad_norm": 0.169277, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:38.561629+00:00", "epoch": 0, "step": 20880, "train_loss": 3.466615915298462, "perplexity": 32.02817279434697, "lr": 0.0026291804804649314, "grad_norm": 0.192417, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:38.867032+00:00", "epoch": 0, "step": 20881, "train_loss": 3.3653042316436768, "perplexity": 28.942301123623874, "lr": 0.0026291804804649314, "grad_norm": 0.15525, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:39.172128+00:00", "epoch": 0, "step": 20882, "train_loss": 3.4703519344329834, "perplexity": 32.14805446141071, "lr": 0.0026291804804649314, "grad_norm": 0.178515, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:39.476531+00:00", "epoch": 0, "step": 20883, "train_loss": 3.504293203353882, "perplexity": 33.25792895061669, "lr": 0.0026291804804649314, "grad_norm": 0.17179, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:39.783025+00:00", "epoch": 0, "step": 20884, "train_loss": 3.518300771713257, "perplexity": 33.72706975392488, "lr": 0.0026291804804649314, "grad_norm": 0.143126, "tokens_per_sec": 106912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:40.087976+00:00", "epoch": 0, "step": 20885, "train_loss": 3.5926549434661865, "perplexity": 36.33040316471465, "lr": 0.0026291804804649314, "grad_norm": 0.165019, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:40.392577+00:00", "epoch": 0, "step": 20886, "train_loss": 3.597484827041626, "perplexity": 36.50629921904239, "lr": 0.0026291804804649314, "grad_norm": 0.157757, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:40.697046+00:00", "epoch": 0, "step": 20887, "train_loss": 3.4504575729370117, "perplexity": 31.514809333942402, "lr": 0.0026291804804649314, "grad_norm": 0.148108, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:41.001816+00:00", "epoch": 0, "step": 20888, "train_loss": 3.5644772052764893, "perplexity": 35.32098295305609, "lr": 0.0026291804804649314, "grad_norm": 0.174013, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:41.307308+00:00", "epoch": 0, "step": 20889, "train_loss": 3.477515935897827, "perplexity": 32.37919011018185, "lr": 0.0026291804804649314, "grad_norm": 0.162754, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:41.613551+00:00", "epoch": 0, "step": 20890, "train_loss": 3.447585105895996, "perplexity": 31.424413973827896, "lr": 0.0026291804804649314, "grad_norm": 0.170493, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:41.919302+00:00", "epoch": 0, "step": 20891, "train_loss": 3.599438428878784, "perplexity": 36.577687701895314, "lr": 0.0026291804804649314, "grad_norm": 0.180227, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:42.224714+00:00", "epoch": 0, "step": 20892, "train_loss": 3.4261722564697266, "perplexity": 30.758680788250526, "lr": 0.0026291804804649314, "grad_norm": 0.161287, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:42.529978+00:00", "epoch": 0, "step": 20893, "train_loss": 3.510498523712158, "perplexity": 33.46494669563688, "lr": 0.0026291804804649314, "grad_norm": 0.213164, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:42.834210+00:00", "epoch": 0, "step": 20894, "train_loss": 3.518903970718384, "perplexity": 33.747420025861146, "lr": 0.0026291804804649314, "grad_norm": 0.186893, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:43.139692+00:00", "epoch": 0, "step": 20895, "train_loss": 3.540851593017578, "perplexity": 34.496283480014945, "lr": 0.0026291804804649314, "grad_norm": 0.208463, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:43.445992+00:00", "epoch": 0, "step": 20896, "train_loss": 3.433281183242798, "perplexity": 30.978121065892147, "lr": 0.0026291804804649314, "grad_norm": 0.170442, "tokens_per_sec": 106980} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:43.751741+00:00", "epoch": 0, "step": 20897, "train_loss": 3.3856570720672607, "perplexity": 29.537394538146476, "lr": 0.0026291804804649314, "grad_norm": 0.174431, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:44.057096+00:00", "epoch": 0, "step": 20898, "train_loss": 3.481813669204712, "perplexity": 32.51864669291577, "lr": 0.0026291804804649314, "grad_norm": 0.17069, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:44.361282+00:00", "epoch": 0, "step": 20899, "train_loss": 3.454845905303955, "perplexity": 31.653410683749552, "lr": 0.0026291804804649314, "grad_norm": 0.208223, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:44.667326+00:00", "epoch": 0, "step": 20900, "train_loss": 3.4630205631256104, "perplexity": 31.913226992838055, "lr": 0.0026291804804649314, "grad_norm": 0.17132, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:44.972086+00:00", "epoch": 0, "step": 20901, "train_loss": 3.4312362670898438, "perplexity": 30.914838131935685, "lr": 0.0026291804804649314, "grad_norm": 0.168915, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:45.276823+00:00", "epoch": 0, "step": 20902, "train_loss": 3.4844789505004883, "perplexity": 32.605433638145406, "lr": 0.0026291804804649314, "grad_norm": 0.195108, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:45.581799+00:00", "epoch": 0, "step": 20903, "train_loss": 3.478996753692627, "perplexity": 32.427173309497704, "lr": 0.0026291804804649314, "grad_norm": 0.169982, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:45.886091+00:00", "epoch": 0, "step": 20904, "train_loss": 3.5134358406066895, "perplexity": 33.563388355337494, "lr": 0.0026291804804649314, "grad_norm": 0.195159, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:46.190365+00:00", "epoch": 0, "step": 20905, "train_loss": 3.5216872692108154, "perplexity": 33.84148000692274, "lr": 0.0026291804804649314, "grad_norm": 0.177586, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:46.494936+00:00", "epoch": 0, "step": 20906, "train_loss": 3.501084327697754, "perplexity": 33.15137943553369, "lr": 0.0026291804804649314, "grad_norm": 0.181953, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:46.800409+00:00", "epoch": 0, "step": 20907, "train_loss": 3.5991673469543457, "perplexity": 36.567773495763326, "lr": 0.0026291804804649314, "grad_norm": 0.206504, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:47.105456+00:00", "epoch": 0, "step": 20908, "train_loss": 3.5416293144226074, "perplexity": 34.523122413324984, "lr": 0.0026291804804649314, "grad_norm": 0.163497, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:47.409674+00:00", "epoch": 0, "step": 20909, "train_loss": 3.5810530185699463, "perplexity": 35.91133624788146, "lr": 0.0026291804804649314, "grad_norm": 0.180741, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:47.715531+00:00", "epoch": 0, "step": 20910, "train_loss": 3.509814500808716, "perplexity": 33.442063732762165, "lr": 0.0026291804804649314, "grad_norm": 0.169888, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:48.021637+00:00", "epoch": 0, "step": 20911, "train_loss": 3.4159905910491943, "perplexity": 30.44709511397041, "lr": 0.0026291804804649314, "grad_norm": 0.183615, "tokens_per_sec": 106948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:48.328137+00:00", "epoch": 0, "step": 20912, "train_loss": 3.422900438308716, "perplexity": 30.658208431108157, "lr": 0.0026291804804649314, "grad_norm": 0.178421, "tokens_per_sec": 106910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:48.632589+00:00", "epoch": 0, "step": 20913, "train_loss": 3.4448635578155518, "perplexity": 31.339007192345512, "lr": 0.0026291804804649314, "grad_norm": 0.151914, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:48.937027+00:00", "epoch": 0, "step": 20914, "train_loss": 3.390828847885132, "perplexity": 29.69055102507322, "lr": 0.0026291804804649314, "grad_norm": 0.172733, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:49.241923+00:00", "epoch": 0, "step": 20915, "train_loss": 3.478419542312622, "perplexity": 32.408461376931804, "lr": 0.0026291804804649314, "grad_norm": 0.184516, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:49.546596+00:00", "epoch": 0, "step": 20916, "train_loss": 3.480024814605713, "perplexity": 32.46052756106252, "lr": 0.0026291804804649314, "grad_norm": 0.168643, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:49.852582+00:00", "epoch": 0, "step": 20917, "train_loss": 3.576293706893921, "perplexity": 35.74082907620878, "lr": 0.0026291804804649314, "grad_norm": 0.154165, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:50.158637+00:00", "epoch": 0, "step": 20918, "train_loss": 3.5258615016937256, "perplexity": 33.983037453338994, "lr": 0.0026291804804649314, "grad_norm": 0.190806, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:50.463564+00:00", "epoch": 0, "step": 20919, "train_loss": 3.4659039974212646, "perplexity": 32.00537948000738, "lr": 0.0026291804804649314, "grad_norm": 0.186078, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:50.767898+00:00", "epoch": 0, "step": 20920, "train_loss": 3.423433780670166, "perplexity": 30.674564113592094, "lr": 0.0026291804804649314, "grad_norm": 0.178233, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:51.072216+00:00", "epoch": 0, "step": 20921, "train_loss": 3.6017510890960693, "perplexity": 36.66237735653516, "lr": 0.0026291804804649314, "grad_norm": 0.190274, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:51.377264+00:00", "epoch": 0, "step": 20922, "train_loss": 3.512275218963623, "perplexity": 33.524456557311574, "lr": 0.0026291804804649314, "grad_norm": 0.156572, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:51.682737+00:00", "epoch": 0, "step": 20923, "train_loss": 3.4785377979278564, "perplexity": 32.41229408608596, "lr": 0.0026291804804649314, "grad_norm": 0.157114, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:51.988507+00:00", "epoch": 0, "step": 20924, "train_loss": 3.516613483428955, "perplexity": 33.670210446776245, "lr": 0.0026291804804649314, "grad_norm": 0.157813, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:52.293512+00:00", "epoch": 0, "step": 20925, "train_loss": 3.513533353805542, "perplexity": 33.56666138827971, "lr": 0.0026291804804649314, "grad_norm": 0.160864, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:52.598117+00:00", "epoch": 0, "step": 20926, "train_loss": 3.5344858169555664, "perplexity": 34.27738533190734, "lr": 0.0026291804804649314, "grad_norm": 0.154064, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:52.903380+00:00", "epoch": 0, "step": 20927, "train_loss": 3.5344133377075195, "perplexity": 34.274901022824984, "lr": 0.0026291804804649314, "grad_norm": 0.163368, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:53.209480+00:00", "epoch": 0, "step": 20928, "train_loss": 3.4997646808624268, "perplexity": 33.1076601759108, "lr": 0.0026291804804649314, "grad_norm": 0.181279, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:53.514549+00:00", "epoch": 0, "step": 20929, "train_loss": 3.5088984966278076, "perplexity": 33.41144468829207, "lr": 0.0026291804804649314, "grad_norm": 0.173095, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:53.820593+00:00", "epoch": 0, "step": 20930, "train_loss": 3.518929958343506, "perplexity": 33.74829705255749, "lr": 0.0026291804804649314, "grad_norm": 0.168176, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:54.125420+00:00", "epoch": 0, "step": 20931, "train_loss": 3.613734483718872, "perplexity": 37.10436002949087, "lr": 0.0026291804804649314, "grad_norm": 0.178204, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:54.431202+00:00", "epoch": 0, "step": 20932, "train_loss": 3.5288338661193848, "perplexity": 34.08419769301119, "lr": 0.0026291804804649314, "grad_norm": 0.178636, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:54.735729+00:00", "epoch": 0, "step": 20933, "train_loss": 3.514465808868408, "perplexity": 33.597975388815016, "lr": 0.0026291804804649314, "grad_norm": 0.174957, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:55.040765+00:00", "epoch": 0, "step": 20934, "train_loss": 3.472008466720581, "perplexity": 32.201352884700995, "lr": 0.0026291804804649314, "grad_norm": 0.183464, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:55.345435+00:00", "epoch": 0, "step": 20935, "train_loss": 3.449427843093872, "perplexity": 31.482374296799264, "lr": 0.0026291804804649314, "grad_norm": 0.167826, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:55.649999+00:00", "epoch": 0, "step": 20936, "train_loss": 3.5010604858398438, "perplexity": 33.15058905447778, "lr": 0.0026291804804649314, "grad_norm": 0.155514, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:55.954442+00:00", "epoch": 0, "step": 20937, "train_loss": 3.5153584480285645, "perplexity": 33.62797964684459, "lr": 0.0026291804804649314, "grad_norm": 0.176575, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:56.260399+00:00", "epoch": 0, "step": 20938, "train_loss": 3.4944515228271484, "perplexity": 32.93222042748909, "lr": 0.0026291804804649314, "grad_norm": 0.193031, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:56.564535+00:00", "epoch": 0, "step": 20939, "train_loss": 3.463205575942993, "perplexity": 31.919131895100236, "lr": 0.0026291804804649314, "grad_norm": 0.169973, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:56.869244+00:00", "epoch": 0, "step": 20940, "train_loss": 3.535752534866333, "perplexity": 34.320832621762094, "lr": 0.0026291804804649314, "grad_norm": 0.180392, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:57.174058+00:00", "epoch": 0, "step": 20941, "train_loss": 3.4781508445739746, "perplexity": 32.39975446646274, "lr": 0.0026291804804649314, "grad_norm": 0.177731, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:57.479951+00:00", "epoch": 0, "step": 20942, "train_loss": 3.4344778060913086, "perplexity": 31.01521238109521, "lr": 0.0026291804804649314, "grad_norm": 0.160165, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:57.784695+00:00", "epoch": 0, "step": 20943, "train_loss": 3.466862201690674, "perplexity": 32.03606186891961, "lr": 0.0026291804804649314, "grad_norm": 0.181943, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:58.089759+00:00", "epoch": 0, "step": 20944, "train_loss": 3.501193046569824, "perplexity": 33.15498381204183, "lr": 0.0026291804804649314, "grad_norm": 0.179314, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:58.396229+00:00", "epoch": 0, "step": 20945, "train_loss": 3.5522353649139404, "perplexity": 34.891224999469905, "lr": 0.0026291804804649314, "grad_norm": 0.169113, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:58.701989+00:00", "epoch": 0, "step": 20946, "train_loss": 3.507689952850342, "perplexity": 33.371089884902936, "lr": 0.0026291804804649314, "grad_norm": 0.205494, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:59.006972+00:00", "epoch": 0, "step": 20947, "train_loss": 3.4073128700256348, "perplexity": 30.184026784195552, "lr": 0.0026291804804649314, "grad_norm": 0.196059, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:59.313081+00:00", "epoch": 0, "step": 20948, "train_loss": 3.5180318355560303, "perplexity": 33.71800054496451, "lr": 0.0026291804804649314, "grad_norm": 0.174446, "tokens_per_sec": 107046} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:59.618008+00:00", "epoch": 0, "step": 20949, "train_loss": 3.465745210647583, "perplexity": 32.00029785251759, "lr": 0.0026291804804649314, "grad_norm": 0.195117, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:47:59.924116+00:00", "epoch": 0, "step": 20950, "train_loss": 3.474653959274292, "perplexity": 32.286654106100606, "lr": 0.0026291804804649314, "grad_norm": 0.198078, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:00.228220+00:00", "epoch": 0, "step": 20951, "train_loss": 3.4932126998901367, "perplexity": 32.891448497272286, "lr": 0.0026291804804649314, "grad_norm": 0.163618, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:00.532699+00:00", "epoch": 0, "step": 20952, "train_loss": 3.367253065109253, "perplexity": 28.998759845078315, "lr": 0.0026291804804649314, "grad_norm": 0.17902, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:00.838301+00:00", "epoch": 0, "step": 20953, "train_loss": 3.438674211502075, "perplexity": 31.14563825467016, "lr": 0.0026291804804649314, "grad_norm": 0.198189, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:01.143158+00:00", "epoch": 0, "step": 20954, "train_loss": 3.4778263568878174, "perplexity": 32.38924285064104, "lr": 0.0026291804804649314, "grad_norm": 0.166696, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:01.448778+00:00", "epoch": 0, "step": 20955, "train_loss": 3.4829769134521484, "perplexity": 32.55649583124402, "lr": 0.0026291804804649314, "grad_norm": 0.173935, "tokens_per_sec": 107218} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:01.754034+00:00", "epoch": 0, "step": 20956, "train_loss": 3.488473415374756, "perplexity": 32.73593536628676, "lr": 0.0026291804804649314, "grad_norm": 0.157666, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:02.058830+00:00", "epoch": 0, "step": 20957, "train_loss": 3.4121010303497314, "perplexity": 30.328899303304507, "lr": 0.0026291804804649314, "grad_norm": 0.170869, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:02.362866+00:00", "epoch": 0, "step": 20958, "train_loss": 3.4905123710632324, "perplexity": 32.802750581403124, "lr": 0.0026291804804649314, "grad_norm": 0.184978, "tokens_per_sec": 107776} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:02.668793+00:00", "epoch": 0, "step": 20959, "train_loss": 3.4547278881073, "perplexity": 31.64967525738269, "lr": 0.0026291804804649314, "grad_norm": 0.172848, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:02.975312+00:00", "epoch": 0, "step": 20960, "train_loss": 3.4989798069000244, "perplexity": 33.0816850304238, "lr": 0.0026291804804649314, "grad_norm": 0.163646, "tokens_per_sec": 106992} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:03.281652+00:00", "epoch": 0, "step": 20961, "train_loss": 3.514291524887085, "perplexity": 33.59212031013788, "lr": 0.0026291804804649314, "grad_norm": 0.165166, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:03.586147+00:00", "epoch": 0, "step": 20962, "train_loss": 3.5261616706848145, "perplexity": 33.99323963851871, "lr": 0.0026291804804649314, "grad_norm": 0.160696, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:03.890676+00:00", "epoch": 0, "step": 20963, "train_loss": 3.479630708694458, "perplexity": 32.447737195813076, "lr": 0.0026291804804649314, "grad_norm": 0.168465, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:04.195447+00:00", "epoch": 0, "step": 20964, "train_loss": 3.397472620010376, "perplexity": 29.88846500004357, "lr": 0.0026291804804649314, "grad_norm": 0.161987, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:04.500849+00:00", "epoch": 0, "step": 20965, "train_loss": 3.6085093021392822, "perplexity": 36.91098865118461, "lr": 0.0026291804804649314, "grad_norm": 0.167116, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:04.806932+00:00", "epoch": 0, "step": 20966, "train_loss": 3.619503974914551, "perplexity": 37.31905204430901, "lr": 0.0026291804804649314, "grad_norm": 0.197202, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:05.113157+00:00", "epoch": 0, "step": 20967, "train_loss": 3.524260997772217, "perplexity": 33.92869097111033, "lr": 0.0026291804804649314, "grad_norm": 0.195087, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:05.418681+00:00", "epoch": 0, "step": 20968, "train_loss": 3.5312745571136475, "perplexity": 34.16748828932471, "lr": 0.0026291804804649314, "grad_norm": 0.183554, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:05.724096+00:00", "epoch": 0, "step": 20969, "train_loss": 3.534290313720703, "perplexity": 34.270684647216406, "lr": 0.0026291804804649314, "grad_norm": 0.171064, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:06.031366+00:00", "epoch": 0, "step": 20970, "train_loss": 3.507964849472046, "perplexity": 33.380264745786235, "lr": 0.0026291804804649314, "grad_norm": 0.188987, "tokens_per_sec": 106710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:06.336456+00:00", "epoch": 0, "step": 20971, "train_loss": 3.5674126148223877, "perplexity": 35.424816826500496, "lr": 0.0026291804804649314, "grad_norm": 0.186336, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:06.641616+00:00", "epoch": 0, "step": 20972, "train_loss": 3.5297763347625732, "perplexity": 34.1163361229174, "lr": 0.0026291804804649314, "grad_norm": 0.170494, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:06.946304+00:00", "epoch": 0, "step": 20973, "train_loss": 3.5000903606414795, "perplexity": 33.11844442737284, "lr": 0.0026291804804649314, "grad_norm": 0.19758, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:07.251308+00:00", "epoch": 0, "step": 20974, "train_loss": 3.5899033546447754, "perplexity": 36.23057424055089, "lr": 0.0026291804804649314, "grad_norm": 0.19078, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:07.556503+00:00", "epoch": 0, "step": 20975, "train_loss": 3.4693048000335693, "perplexity": 32.11440874657698, "lr": 0.0026291804804649314, "grad_norm": 0.188126, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:07.863163+00:00", "epoch": 0, "step": 20976, "train_loss": 3.377251148223877, "perplexity": 29.29014608084852, "lr": 0.0026291804804649314, "grad_norm": 0.195174, "tokens_per_sec": 106854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:08.169764+00:00", "epoch": 0, "step": 20977, "train_loss": 3.4776899814605713, "perplexity": 32.384826054987144, "lr": 0.0026291804804649314, "grad_norm": 0.155377, "tokens_per_sec": 106876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:08.475155+00:00", "epoch": 0, "step": 20978, "train_loss": 3.4740517139434814, "perplexity": 32.267215473418126, "lr": 0.0026291804804649314, "grad_norm": 0.16434, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:08.780753+00:00", "epoch": 0, "step": 20979, "train_loss": 3.5255510807037354, "perplexity": 33.972490042363425, "lr": 0.0026291804804649314, "grad_norm": 0.162528, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:09.086472+00:00", "epoch": 0, "step": 20980, "train_loss": 3.483842611312866, "perplexity": 32.58469212301139, "lr": 0.0026291804804649314, "grad_norm": 0.148545, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:09.392902+00:00", "epoch": 0, "step": 20981, "train_loss": 3.5273513793945312, "perplexity": 34.033705758462816, "lr": 0.0026291804804649314, "grad_norm": 0.167211, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:09.698367+00:00", "epoch": 0, "step": 20982, "train_loss": 3.4486916065216064, "perplexity": 31.45920435178725, "lr": 0.0026291804804649314, "grad_norm": 0.159927, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:10.003486+00:00", "epoch": 0, "step": 20983, "train_loss": 3.5567502975463867, "perplexity": 35.04911268773747, "lr": 0.0026291804804649314, "grad_norm": 0.182514, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:10.308826+00:00", "epoch": 0, "step": 20984, "train_loss": 3.5222244262695312, "perplexity": 33.85966307993164, "lr": 0.0026291804804649314, "grad_norm": 0.179541, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:10.613169+00:00", "epoch": 0, "step": 20985, "train_loss": 3.4237170219421387, "perplexity": 30.683253646707946, "lr": 0.0026291804804649314, "grad_norm": 0.164464, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:10.920605+00:00", "epoch": 0, "step": 20986, "train_loss": 3.498495578765869, "perplexity": 33.0656698256261, "lr": 0.0026291804804649314, "grad_norm": 0.15808, "tokens_per_sec": 106584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:11.227196+00:00", "epoch": 0, "step": 20987, "train_loss": 3.4786057472229004, "perplexity": 32.414496553447066, "lr": 0.0026291804804649314, "grad_norm": 0.165519, "tokens_per_sec": 106879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:11.532381+00:00", "epoch": 0, "step": 20988, "train_loss": 3.515347719192505, "perplexity": 33.62761885969935, "lr": 0.0026291804804649314, "grad_norm": 0.18549, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:11.837931+00:00", "epoch": 0, "step": 20989, "train_loss": 3.518932819366455, "perplexity": 33.74839360734798, "lr": 0.0026291804804649314, "grad_norm": 0.186237, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:12.143434+00:00", "epoch": 0, "step": 20990, "train_loss": 3.533423662185669, "perplexity": 34.24099677213965, "lr": 0.0026291804804649314, "grad_norm": 0.153906, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:12.448785+00:00", "epoch": 0, "step": 20991, "train_loss": 3.390814781188965, "perplexity": 29.69013338005037, "lr": 0.0026291804804649314, "grad_norm": 0.173338, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:12.753743+00:00", "epoch": 0, "step": 20992, "train_loss": 3.3947701454162598, "perplexity": 29.807801227714616, "lr": 0.0026291804804649314, "grad_norm": 0.160727, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:13.057656+00:00", "epoch": 0, "step": 20993, "train_loss": 3.5311119556427, "perplexity": 34.161933057126994, "lr": 0.0026291804804649314, "grad_norm": 0.182518, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:13.362258+00:00", "epoch": 0, "step": 20994, "train_loss": 3.5578415393829346, "perplexity": 35.08738062182407, "lr": 0.0026291804804649314, "grad_norm": 0.191967, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:13.667227+00:00", "epoch": 0, "step": 20995, "train_loss": 3.479555368423462, "perplexity": 32.44529266658644, "lr": 0.0026291804804649314, "grad_norm": 0.180137, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:13.974193+00:00", "epoch": 0, "step": 20996, "train_loss": 3.373155355453491, "perplexity": 29.17042505590639, "lr": 0.0026291804804649314, "grad_norm": 0.19171, "tokens_per_sec": 106748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:14.279894+00:00", "epoch": 0, "step": 20997, "train_loss": 3.4441769123077393, "perplexity": 31.317495790036652, "lr": 0.0026291804804649314, "grad_norm": 0.170377, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:14.584318+00:00", "epoch": 0, "step": 20998, "train_loss": 3.426952600479126, "perplexity": 30.782692508023036, "lr": 0.0026291804804649314, "grad_norm": 0.188051, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:14.889294+00:00", "epoch": 0, "step": 20999, "train_loss": 3.490877151489258, "perplexity": 32.81471856544546, "lr": 0.0026291804804649314, "grad_norm": 0.172157, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:15.194816+00:00", "epoch": 0, "step": 21000, "train_loss": 3.4253571033477783, "perplexity": 30.733617969984053, "lr": 0.0026291804804649314, "grad_norm": 0.204448, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:48:18.268382+00:00", "step": 21000, "epoch": 0, "val_loss": 3.461587977409363, "val_ppl": 31.867541291843498, "eval_train_loss": 3.4253571033477783, "eval_train_ppl": 30.733617969984053} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:48:19.335055+00:00", "step": 21000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4616_epoch_0000_step_0021000.pt", "val_loss": 3.461587977409363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:20.616156+00:00", "epoch": 0, "step": 21001, "train_loss": 3.3091394901275635, "perplexity": 27.361570437200182, "lr": 0.0026291804804649314, "grad_norm": 0.216383, "tokens_per_sec": 6044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:20.919713+00:00", "epoch": 0, "step": 21002, "train_loss": 3.4846670627593994, "perplexity": 32.61156769684759, "lr": 0.0026291804804649314, "grad_norm": 0.168396, "tokens_per_sec": 107940} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:21.223174+00:00", "epoch": 0, "step": 21003, "train_loss": 3.550529718399048, "perplexity": 34.83176362760564, "lr": 0.0026291804804649314, "grad_norm": 0.200629, "tokens_per_sec": 107981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:21.527327+00:00", "epoch": 0, "step": 21004, "train_loss": 3.4464492797851562, "perplexity": 31.38874156657946, "lr": 0.0026291804804649314, "grad_norm": 0.179782, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:21.830575+00:00", "epoch": 0, "step": 21005, "train_loss": 3.463346004486084, "perplexity": 31.923614567029166, "lr": 0.0026291804804649314, "grad_norm": 0.184414, "tokens_per_sec": 108057} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:22.134876+00:00", "epoch": 0, "step": 21006, "train_loss": 3.460110902786255, "perplexity": 31.820505301685913, "lr": 0.0026291804804649314, "grad_norm": 0.183619, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:22.440801+00:00", "epoch": 0, "step": 21007, "train_loss": 3.510495901107788, "perplexity": 33.46485893043652, "lr": 0.0026291804804649314, "grad_norm": 0.190632, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:22.746236+00:00", "epoch": 0, "step": 21008, "train_loss": 3.5347039699554443, "perplexity": 34.2848638620477, "lr": 0.0026291804804649314, "grad_norm": 0.188444, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:23.050459+00:00", "epoch": 0, "step": 21009, "train_loss": 3.4578583240509033, "perplexity": 31.748907777899465, "lr": 0.0026291804804649314, "grad_norm": 0.184099, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:23.354955+00:00", "epoch": 0, "step": 21010, "train_loss": 3.611483573913574, "perplexity": 37.02093538757667, "lr": 0.0026291804804649314, "grad_norm": 0.177069, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:23.659884+00:00", "epoch": 0, "step": 21011, "train_loss": 3.6062259674072266, "perplexity": 36.82680465549873, "lr": 0.0026291804804649314, "grad_norm": 0.180403, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:23.965548+00:00", "epoch": 0, "step": 21012, "train_loss": 3.468498945236206, "perplexity": 32.08853962100349, "lr": 0.0026291804804649314, "grad_norm": 0.174939, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:24.271081+00:00", "epoch": 0, "step": 21013, "train_loss": 3.539461612701416, "perplexity": 34.44836763375318, "lr": 0.0026291804804649314, "grad_norm": 0.149126, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:24.574862+00:00", "epoch": 0, "step": 21014, "train_loss": 3.517719268798828, "perplexity": 33.70746306579347, "lr": 0.0026291804804649314, "grad_norm": 0.177193, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:24.878631+00:00", "epoch": 0, "step": 21015, "train_loss": 3.453920364379883, "perplexity": 31.624127710160874, "lr": 0.0026291804804649314, "grad_norm": 0.171115, "tokens_per_sec": 107875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:25.182154+00:00", "epoch": 0, "step": 21016, "train_loss": 3.447814702987671, "perplexity": 31.431629756213496, "lr": 0.0026291804804649314, "grad_norm": 0.172675, "tokens_per_sec": 107957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:25.487499+00:00", "epoch": 0, "step": 21017, "train_loss": 3.385934352874756, "perplexity": 29.545585826346336, "lr": 0.0026291804804649314, "grad_norm": 0.217913, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:25.792783+00:00", "epoch": 0, "step": 21018, "train_loss": 3.5481045246124268, "perplexity": 34.747392200761986, "lr": 0.0026291804804649314, "grad_norm": 0.191796, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:26.096575+00:00", "epoch": 0, "step": 21019, "train_loss": 3.584204912185669, "perplexity": 36.02470352628044, "lr": 0.0026291804804649314, "grad_norm": 0.162248, "tokens_per_sec": 107864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:26.400326+00:00", "epoch": 0, "step": 21020, "train_loss": 3.3979756832122803, "perplexity": 29.90350456955632, "lr": 0.0026291804804649314, "grad_norm": 0.178731, "tokens_per_sec": 107937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:26.703936+00:00", "epoch": 0, "step": 21021, "train_loss": 3.5132548809051514, "perplexity": 33.55731528410526, "lr": 0.0026291804804649314, "grad_norm": 0.156015, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:27.009058+00:00", "epoch": 0, "step": 21022, "train_loss": 3.5213911533355713, "perplexity": 33.83146049099301, "lr": 0.0026291804804649314, "grad_norm": 0.173488, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:27.313476+00:00", "epoch": 0, "step": 21023, "train_loss": 3.454623222351074, "perplexity": 31.64636279354137, "lr": 0.0026291804804649314, "grad_norm": 0.186692, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:27.617706+00:00", "epoch": 0, "step": 21024, "train_loss": 3.5380570888519287, "perplexity": 34.400018041861344, "lr": 0.0026291804804649314, "grad_norm": 0.187909, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:27.921634+00:00", "epoch": 0, "step": 21025, "train_loss": 3.540044069290161, "perplexity": 34.468438156989805, "lr": 0.0026291804804649314, "grad_norm": 0.198202, "tokens_per_sec": 107815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:28.225701+00:00", "epoch": 0, "step": 21026, "train_loss": 3.5157723426818848, "perplexity": 33.64190096860396, "lr": 0.0026291804804649314, "grad_norm": 0.182666, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:28.531078+00:00", "epoch": 0, "step": 21027, "train_loss": 3.555166721343994, "perplexity": 34.99365367034988, "lr": 0.0026291804804649314, "grad_norm": 0.215359, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:28.836371+00:00", "epoch": 0, "step": 21028, "train_loss": 3.500215530395508, "perplexity": 33.12259011436752, "lr": 0.0026291804804649314, "grad_norm": 0.172282, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:29.140938+00:00", "epoch": 0, "step": 21029, "train_loss": 3.502713680267334, "perplexity": 33.20543874968364, "lr": 0.0026291804804649314, "grad_norm": 0.161563, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:29.444968+00:00", "epoch": 0, "step": 21030, "train_loss": 3.523881196975708, "perplexity": 33.91580727402908, "lr": 0.0026291804804649314, "grad_norm": 0.173583, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:29.749404+00:00", "epoch": 0, "step": 21031, "train_loss": 3.5997743606567383, "perplexity": 36.589977373688534, "lr": 0.0026291804804649314, "grad_norm": 0.158719, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:30.054562+00:00", "epoch": 0, "step": 21032, "train_loss": 3.5631022453308105, "perplexity": 35.2724513883782, "lr": 0.0026291804804649314, "grad_norm": 0.172011, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:30.359212+00:00", "epoch": 0, "step": 21033, "train_loss": 3.4792959690093994, "perplexity": 32.436877468175595, "lr": 0.0026291804804649314, "grad_norm": 0.182994, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:30.663529+00:00", "epoch": 0, "step": 21034, "train_loss": 3.5386693477630615, "perplexity": 34.42108620837882, "lr": 0.0026291804804649314, "grad_norm": 0.190709, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:30.967616+00:00", "epoch": 0, "step": 21035, "train_loss": 3.522243022918701, "perplexity": 33.860292762061924, "lr": 0.0026291804804649314, "grad_norm": 0.160844, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:31.271530+00:00", "epoch": 0, "step": 21036, "train_loss": 3.5333251953125, "perplexity": 34.23762533424351, "lr": 0.0026291804804649314, "grad_norm": 0.162227, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:31.576291+00:00", "epoch": 0, "step": 21037, "train_loss": 3.5248770713806152, "perplexity": 33.94959998226878, "lr": 0.0026291804804649314, "grad_norm": 0.159631, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:31.881885+00:00", "epoch": 0, "step": 21038, "train_loss": 3.495346784591675, "perplexity": 32.96171658668045, "lr": 0.0026291804804649314, "grad_norm": 0.183505, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:32.186382+00:00", "epoch": 0, "step": 21039, "train_loss": 3.4682374000549316, "perplexity": 32.08014811551823, "lr": 0.0026291804804649314, "grad_norm": 0.197771, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:32.491172+00:00", "epoch": 0, "step": 21040, "train_loss": 3.410965919494629, "perplexity": 30.294492172130457, "lr": 0.0026291804804649314, "grad_norm": 0.193959, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:32.795342+00:00", "epoch": 0, "step": 21041, "train_loss": 3.5484797954559326, "perplexity": 34.76043433095534, "lr": 0.0026291804804649314, "grad_norm": 0.196751, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:33.099212+00:00", "epoch": 0, "step": 21042, "train_loss": 3.5486340522766113, "perplexity": 34.76579677862707, "lr": 0.0026291804804649314, "grad_norm": 0.179116, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:33.404608+00:00", "epoch": 0, "step": 21043, "train_loss": 3.4855117797851562, "perplexity": 32.63912688153527, "lr": 0.0026291804804649314, "grad_norm": 0.161582, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:33.709927+00:00", "epoch": 0, "step": 21044, "train_loss": 3.3372490406036377, "perplexity": 28.14160371383201, "lr": 0.0026291804804649314, "grad_norm": 0.181905, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:34.014261+00:00", "epoch": 0, "step": 21045, "train_loss": 3.55964994430542, "perplexity": 35.15089022188628, "lr": 0.0026291804804649314, "grad_norm": 0.15085, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:34.319892+00:00", "epoch": 0, "step": 21046, "train_loss": 3.5627942085266113, "perplexity": 35.26158784844721, "lr": 0.0026291804804649314, "grad_norm": 0.168414, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:34.624893+00:00", "epoch": 0, "step": 21047, "train_loss": 3.535452365875244, "perplexity": 34.310532118083934, "lr": 0.0026291804804649314, "grad_norm": 0.175415, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:34.930019+00:00", "epoch": 0, "step": 21048, "train_loss": 3.4158360958099365, "perplexity": 30.442391546074848, "lr": 0.0026291804804649314, "grad_norm": 0.184228, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:35.235318+00:00", "epoch": 0, "step": 21049, "train_loss": 3.4442551136016846, "perplexity": 31.319944954493227, "lr": 0.0026291804804649314, "grad_norm": 0.190553, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:35.539697+00:00", "epoch": 0, "step": 21050, "train_loss": 3.5045464038848877, "perplexity": 33.2663509420685, "lr": 0.0026291804804649314, "grad_norm": 0.184942, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:35.843877+00:00", "epoch": 0, "step": 21051, "train_loss": 3.5518863201141357, "perplexity": 34.87904852401637, "lr": 0.0026291804804649314, "grad_norm": 0.20821, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:36.148330+00:00", "epoch": 0, "step": 21052, "train_loss": 3.548820734024048, "perplexity": 34.77228752415374, "lr": 0.0026291804804649314, "grad_norm": 0.215997, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:36.452649+00:00", "epoch": 0, "step": 21053, "train_loss": 3.492309093475342, "perplexity": 32.86174099738335, "lr": 0.0026291804804649314, "grad_norm": 0.179801, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:36.758064+00:00", "epoch": 0, "step": 21054, "train_loss": 3.4903321266174316, "perplexity": 32.796838600620696, "lr": 0.0026291804804649314, "grad_norm": 0.187515, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:37.062668+00:00", "epoch": 0, "step": 21055, "train_loss": 3.6154699325561523, "perplexity": 37.16880865543312, "lr": 0.0026291804804649314, "grad_norm": 0.202157, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:37.367187+00:00", "epoch": 0, "step": 21056, "train_loss": 3.563136339187622, "perplexity": 35.27365398278565, "lr": 0.0026291804804649314, "grad_norm": 0.170863, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:37.671158+00:00", "epoch": 0, "step": 21057, "train_loss": 3.430652141571045, "perplexity": 30.89678525915843, "lr": 0.0026291804804649314, "grad_norm": 0.187724, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:37.974608+00:00", "epoch": 0, "step": 21058, "train_loss": 3.3863401412963867, "perplexity": 29.5575775158648, "lr": 0.0026291804804649314, "grad_norm": 0.178942, "tokens_per_sec": 107986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:38.279598+00:00", "epoch": 0, "step": 21059, "train_loss": 3.5046794414520264, "perplexity": 33.270776910868946, "lr": 0.0026291804804649314, "grad_norm": 0.169897, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:38.586030+00:00", "epoch": 0, "step": 21060, "train_loss": 3.4588186740875244, "perplexity": 31.779412487901865, "lr": 0.0026291804804649314, "grad_norm": 0.208854, "tokens_per_sec": 107003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:38.890239+00:00", "epoch": 0, "step": 21061, "train_loss": 3.547121047973633, "perplexity": 34.71323575106179, "lr": 0.0026291804804649314, "grad_norm": 0.165753, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:39.195109+00:00", "epoch": 0, "step": 21062, "train_loss": 3.5715150833129883, "perplexity": 35.57044453361059, "lr": 0.0026291804804649314, "grad_norm": 0.203667, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:39.499339+00:00", "epoch": 0, "step": 21063, "train_loss": 3.461076259613037, "perplexity": 31.851238275466873, "lr": 0.0026291804804649314, "grad_norm": 0.200219, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:39.804296+00:00", "epoch": 0, "step": 21064, "train_loss": 3.571289300918579, "perplexity": 35.56241426005491, "lr": 0.0026291804804649314, "grad_norm": 0.172893, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:40.109031+00:00", "epoch": 0, "step": 21065, "train_loss": 3.470492124557495, "perplexity": 32.15256161709044, "lr": 0.0026291804804649314, "grad_norm": 0.200309, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:40.414388+00:00", "epoch": 0, "step": 21066, "train_loss": 3.3763461112976074, "perplexity": 29.263649409111295, "lr": 0.0026291804804649314, "grad_norm": 0.187778, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:40.719980+00:00", "epoch": 0, "step": 21067, "train_loss": 3.4839956760406494, "perplexity": 32.58968007177087, "lr": 0.0026291804804649314, "grad_norm": 0.182383, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:41.025718+00:00", "epoch": 0, "step": 21068, "train_loss": 3.515561580657959, "perplexity": 33.63481128061185, "lr": 0.0026291804804649314, "grad_norm": 0.193749, "tokens_per_sec": 107176} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:41.331270+00:00", "epoch": 0, "step": 21069, "train_loss": 3.454085111618042, "perplexity": 31.629338127049426, "lr": 0.0026291804804649314, "grad_norm": 0.17661, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:41.636652+00:00", "epoch": 0, "step": 21070, "train_loss": 3.5474987030029297, "perplexity": 34.72634785489648, "lr": 0.0026291804804649314, "grad_norm": 0.183798, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:41.942103+00:00", "epoch": 0, "step": 21071, "train_loss": 3.474295139312744, "perplexity": 32.275071088349016, "lr": 0.0026291804804649314, "grad_norm": 0.185935, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:42.247264+00:00", "epoch": 0, "step": 21072, "train_loss": 3.5022943019866943, "perplexity": 33.191516029520194, "lr": 0.0026291804804649314, "grad_norm": 0.168153, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:42.552021+00:00", "epoch": 0, "step": 21073, "train_loss": 3.487898588180542, "perplexity": 32.71712326778109, "lr": 0.0026291804804649314, "grad_norm": 0.188047, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:42.857586+00:00", "epoch": 0, "step": 21074, "train_loss": 3.4719719886779785, "perplexity": 32.20017826380268, "lr": 0.0026291804804649314, "grad_norm": 0.181367, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:43.162912+00:00", "epoch": 0, "step": 21075, "train_loss": 3.4589226245880127, "perplexity": 31.78271614544066, "lr": 0.0026291804804649314, "grad_norm": 0.172186, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:43.467633+00:00", "epoch": 0, "step": 21076, "train_loss": 3.4658701419830322, "perplexity": 32.00429594220122, "lr": 0.0026291804804649314, "grad_norm": 0.177235, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:43.772599+00:00", "epoch": 0, "step": 21077, "train_loss": 3.594054698944092, "perplexity": 36.38129245353035, "lr": 0.0026291804804649314, "grad_norm": 0.19589, "tokens_per_sec": 107448} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:44.076797+00:00", "epoch": 0, "step": 21078, "train_loss": 3.517423152923584, "perplexity": 33.697483228532526, "lr": 0.0026291804804649314, "grad_norm": 0.159475, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:44.380886+00:00", "epoch": 0, "step": 21079, "train_loss": 3.5828232765197754, "perplexity": 35.97496487929187, "lr": 0.0026291804804649314, "grad_norm": 0.201, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:44.686145+00:00", "epoch": 0, "step": 21080, "train_loss": 3.5230624675750732, "perplexity": 33.88805076955118, "lr": 0.0026291804804649314, "grad_norm": 0.174617, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:44.991416+00:00", "epoch": 0, "step": 21081, "train_loss": 3.511247396469116, "perplexity": 33.49001706861832, "lr": 0.0026291804804649314, "grad_norm": 0.180127, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:45.296629+00:00", "epoch": 0, "step": 21082, "train_loss": 3.4524714946746826, "perplexity": 31.578341646598044, "lr": 0.0026291804804649314, "grad_norm": 0.169793, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:45.600599+00:00", "epoch": 0, "step": 21083, "train_loss": 3.445343017578125, "perplexity": 31.354036588000557, "lr": 0.0026291804804649314, "grad_norm": 0.194049, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:45.905608+00:00", "epoch": 0, "step": 21084, "train_loss": 3.530688762664795, "perplexity": 34.14747902558359, "lr": 0.0026291804804649314, "grad_norm": 0.15253, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:46.210104+00:00", "epoch": 0, "step": 21085, "train_loss": 3.4767467975616455, "perplexity": 32.35429560865858, "lr": 0.0026291804804649314, "grad_norm": 0.190238, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:46.515121+00:00", "epoch": 0, "step": 21086, "train_loss": 3.449079990386963, "perplexity": 31.47142497216671, "lr": 0.0026291804804649314, "grad_norm": 0.168621, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:46.820625+00:00", "epoch": 0, "step": 21087, "train_loss": 3.57212495803833, "perplexity": 35.592144665209105, "lr": 0.0026291804804649314, "grad_norm": 0.184722, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:47.125093+00:00", "epoch": 0, "step": 21088, "train_loss": 3.648322105407715, "perplexity": 38.41016374409574, "lr": 0.0026291804804649314, "grad_norm": 0.168097, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:47.429311+00:00", "epoch": 0, "step": 21089, "train_loss": 3.564492702484131, "perplexity": 35.32153033390444, "lr": 0.0026291804804649314, "grad_norm": 0.16483, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:47.734422+00:00", "epoch": 0, "step": 21090, "train_loss": 3.3800830841064453, "perplexity": 29.37321145893856, "lr": 0.0026291804804649314, "grad_norm": 0.180026, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:48.040269+00:00", "epoch": 0, "step": 21091, "train_loss": 3.4111340045928955, "perplexity": 30.299584652797215, "lr": 0.0026291804804649314, "grad_norm": 0.190002, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:48.345748+00:00", "epoch": 0, "step": 21092, "train_loss": 3.531277656555176, "perplexity": 34.16759418962095, "lr": 0.0026291804804649314, "grad_norm": 0.160447, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:48.651587+00:00", "epoch": 0, "step": 21093, "train_loss": 3.4549195766448975, "perplexity": 31.655742718861045, "lr": 0.0026291804804649314, "grad_norm": 0.187664, "tokens_per_sec": 107142} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:48.957004+00:00", "epoch": 0, "step": 21094, "train_loss": 3.5403037071228027, "perplexity": 34.47738862945778, "lr": 0.0026291804804649314, "grad_norm": 0.174406, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:49.261233+00:00", "epoch": 0, "step": 21095, "train_loss": 3.4071714878082275, "perplexity": 30.17975960121753, "lr": 0.0026291804804649314, "grad_norm": 0.152961, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:49.566013+00:00", "epoch": 0, "step": 21096, "train_loss": 3.6100692749023438, "perplexity": 36.968613723223605, "lr": 0.0026291804804649314, "grad_norm": 0.161643, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:49.870823+00:00", "epoch": 0, "step": 21097, "train_loss": 3.6534063816070557, "perplexity": 38.60594891662757, "lr": 0.0026291804804649314, "grad_norm": 0.172426, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:50.175870+00:00", "epoch": 0, "step": 21098, "train_loss": 3.5860841274261475, "perplexity": 36.09246534776312, "lr": 0.0026291804804649314, "grad_norm": 0.172354, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:50.480424+00:00", "epoch": 0, "step": 21099, "train_loss": 3.5091934204101562, "perplexity": 33.421299971140485, "lr": 0.0026291804804649314, "grad_norm": 0.152003, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:50.784754+00:00", "epoch": 0, "step": 21100, "train_loss": 3.5012459754943848, "perplexity": 33.156738716121, "lr": 0.0026291804804649314, "grad_norm": 0.143631, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:51.088649+00:00", "epoch": 0, "step": 21101, "train_loss": 3.5755929946899414, "perplexity": 35.715793813376195, "lr": 0.0026291804804649314, "grad_norm": 0.157217, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:51.394025+00:00", "epoch": 0, "step": 21102, "train_loss": 3.4738123416900635, "perplexity": 32.25949252170579, "lr": 0.0026291804804649314, "grad_norm": 0.175369, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:51.699531+00:00", "epoch": 0, "step": 21103, "train_loss": 3.5028634071350098, "perplexity": 33.21041086823801, "lr": 0.0026291804804649314, "grad_norm": 0.190008, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:52.003877+00:00", "epoch": 0, "step": 21104, "train_loss": 3.584432363510132, "perplexity": 36.03289832473443, "lr": 0.0026291804804649314, "grad_norm": 0.198684, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:52.308474+00:00", "epoch": 0, "step": 21105, "train_loss": 3.549684762954712, "perplexity": 34.80234476985498, "lr": 0.0026291804804649314, "grad_norm": 0.191146, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:52.613639+00:00", "epoch": 0, "step": 21106, "train_loss": 3.555393934249878, "perplexity": 35.00160558344224, "lr": 0.0026291804804649314, "grad_norm": 0.173983, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:52.918163+00:00", "epoch": 0, "step": 21107, "train_loss": 3.498577833175659, "perplexity": 33.06838973464264, "lr": 0.0026291804804649314, "grad_norm": 0.166189, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:53.223909+00:00", "epoch": 0, "step": 21108, "train_loss": 3.3940980434417725, "perplexity": 29.787774076552154, "lr": 0.0026291804804649314, "grad_norm": 0.16564, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:53.529524+00:00", "epoch": 0, "step": 21109, "train_loss": 3.522216320037842, "perplexity": 33.85938860677027, "lr": 0.0026291804804649314, "grad_norm": 0.180651, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:53.834382+00:00", "epoch": 0, "step": 21110, "train_loss": 3.5176424980163574, "perplexity": 33.704875416807994, "lr": 0.0026291804804649314, "grad_norm": 0.164093, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:54.139192+00:00", "epoch": 0, "step": 21111, "train_loss": 3.4191555976867676, "perplexity": 30.543613031455223, "lr": 0.0026291804804649314, "grad_norm": 0.190169, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:54.443529+00:00", "epoch": 0, "step": 21112, "train_loss": 3.5084221363067627, "perplexity": 33.395532592020324, "lr": 0.0026291804804649314, "grad_norm": 0.193196, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:54.747708+00:00", "epoch": 0, "step": 21113, "train_loss": 3.4934303760528564, "perplexity": 32.89860896087031, "lr": 0.0026291804804649314, "grad_norm": 0.183969, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:55.053599+00:00", "epoch": 0, "step": 21114, "train_loss": 3.449795722961426, "perplexity": 31.493958159090067, "lr": 0.0026291804804649314, "grad_norm": 0.194895, "tokens_per_sec": 107124} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:55.359977+00:00", "epoch": 0, "step": 21115, "train_loss": 3.428712844848633, "perplexity": 30.836925286656985, "lr": 0.0026291804804649314, "grad_norm": 0.166445, "tokens_per_sec": 106953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:55.666713+00:00", "epoch": 0, "step": 21116, "train_loss": 3.5403923988342285, "perplexity": 34.480446623668186, "lr": 0.0026291804804649314, "grad_norm": 0.196199, "tokens_per_sec": 106829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:55.972239+00:00", "epoch": 0, "step": 21117, "train_loss": 3.543168544769287, "perplexity": 34.576302368606584, "lr": 0.0026291804804649314, "grad_norm": 0.216304, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:56.277293+00:00", "epoch": 0, "step": 21118, "train_loss": 3.555202007293701, "perplexity": 34.99488847643888, "lr": 0.0026291804804649314, "grad_norm": 0.190854, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:56.583045+00:00", "epoch": 0, "step": 21119, "train_loss": 3.5283656120300293, "perplexity": 34.06824136415871, "lr": 0.0026291804804649314, "grad_norm": 0.192359, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:56.888758+00:00", "epoch": 0, "step": 21120, "train_loss": 3.4197003841400146, "perplexity": 30.560257311445454, "lr": 0.0026291804804649314, "grad_norm": 0.183733, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:57.193003+00:00", "epoch": 0, "step": 21121, "train_loss": 3.532536029815674, "perplexity": 34.210616840124004, "lr": 0.0026291804804649314, "grad_norm": 0.170853, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:57.497301+00:00", "epoch": 0, "step": 21122, "train_loss": 3.488007068634033, "perplexity": 32.720672628664765, "lr": 0.0026291804804649314, "grad_norm": 0.182735, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:57.801288+00:00", "epoch": 0, "step": 21123, "train_loss": 3.542029619216919, "perplexity": 34.536944951171144, "lr": 0.0026291804804649314, "grad_norm": 0.152767, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:58.106324+00:00", "epoch": 0, "step": 21124, "train_loss": 3.6508076190948486, "perplexity": 38.505751474903384, "lr": 0.0026291804804649314, "grad_norm": 0.171743, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:58.411552+00:00", "epoch": 0, "step": 21125, "train_loss": 3.4264252185821533, "perplexity": 30.76646255332255, "lr": 0.0026291804804649314, "grad_norm": 0.155492, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:58.716457+00:00", "epoch": 0, "step": 21126, "train_loss": 3.485673189163208, "perplexity": 32.644395567901626, "lr": 0.0026291804804649314, "grad_norm": 0.166799, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:59.020710+00:00", "epoch": 0, "step": 21127, "train_loss": 3.3985278606414795, "perplexity": 29.920021169460764, "lr": 0.0026291804804649314, "grad_norm": 0.151842, "tokens_per_sec": 107699} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:59.324728+00:00", "epoch": 0, "step": 21128, "train_loss": 3.4206771850585938, "perplexity": 30.590123182989984, "lr": 0.0026291804804649314, "grad_norm": 0.162229, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:59.629918+00:00", "epoch": 0, "step": 21129, "train_loss": 3.4238369464874268, "perplexity": 30.6869335426, "lr": 0.0026291804804649314, "grad_norm": 0.155784, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:48:59.935936+00:00", "epoch": 0, "step": 21130, "train_loss": 3.481881856918335, "perplexity": 32.52086414068437, "lr": 0.0026291804804649314, "grad_norm": 0.159042, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:00.240449+00:00", "epoch": 0, "step": 21131, "train_loss": 3.4720089435577393, "perplexity": 32.20136823950626, "lr": 0.0026291804804649314, "grad_norm": 0.173772, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:00.545249+00:00", "epoch": 0, "step": 21132, "train_loss": 3.5622718334198, "perplexity": 35.24317288290708, "lr": 0.0026291804804649314, "grad_norm": 0.168693, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:00.848891+00:00", "epoch": 0, "step": 21133, "train_loss": 3.599508047103882, "perplexity": 36.580234264233844, "lr": 0.0026291804804649314, "grad_norm": 0.151614, "tokens_per_sec": 107917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:01.154102+00:00", "epoch": 0, "step": 21134, "train_loss": 3.516451358795166, "perplexity": 33.66475211871464, "lr": 0.0026291804804649314, "grad_norm": 0.189581, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:01.459316+00:00", "epoch": 0, "step": 21135, "train_loss": 3.5302858352661133, "perplexity": 34.13372284224597, "lr": 0.0026291804804649314, "grad_norm": 0.207029, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:01.763810+00:00", "epoch": 0, "step": 21136, "train_loss": 3.5368449687957764, "perplexity": 34.35834635070924, "lr": 0.0026291804804649314, "grad_norm": 0.171842, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:02.068138+00:00", "epoch": 0, "step": 21137, "train_loss": 3.4944865703582764, "perplexity": 32.93337464073568, "lr": 0.0026291804804649314, "grad_norm": 0.170653, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:02.373065+00:00", "epoch": 0, "step": 21138, "train_loss": 3.5913233757019043, "perplexity": 36.2820589649293, "lr": 0.0026291804804649314, "grad_norm": 0.17004, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:02.677665+00:00", "epoch": 0, "step": 21139, "train_loss": 3.4631757736206055, "perplexity": 31.918180645015962, "lr": 0.0026291804804649314, "grad_norm": 0.165151, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:02.983594+00:00", "epoch": 0, "step": 21140, "train_loss": 3.588890552520752, "perplexity": 36.19389841381815, "lr": 0.0026291804804649314, "grad_norm": 0.154352, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:03.288934+00:00", "epoch": 0, "step": 21141, "train_loss": 3.4405927658081055, "perplexity": 31.20545021095429, "lr": 0.0026291804804649314, "grad_norm": 0.200676, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:03.594560+00:00", "epoch": 0, "step": 21142, "train_loss": 3.483649492263794, "perplexity": 32.57840000583768, "lr": 0.0026291804804649314, "grad_norm": 0.18961, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:03.899032+00:00", "epoch": 0, "step": 21143, "train_loss": 3.308051824569702, "perplexity": 27.33182637815324, "lr": 0.0026291804804649314, "grad_norm": 0.158519, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:04.203853+00:00", "epoch": 0, "step": 21144, "train_loss": 3.375762462615967, "perplexity": 29.24657470201352, "lr": 0.0026291804804649314, "grad_norm": 0.169935, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:04.509096+00:00", "epoch": 0, "step": 21145, "train_loss": 3.4916930198669434, "perplexity": 32.841501981030895, "lr": 0.0026291804804649314, "grad_norm": 0.161493, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:04.813824+00:00", "epoch": 0, "step": 21146, "train_loss": 3.4893243312835693, "perplexity": 32.763802749198085, "lr": 0.0026291804804649314, "grad_norm": 0.155963, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:05.118685+00:00", "epoch": 0, "step": 21147, "train_loss": 3.5669620037078857, "perplexity": 35.40885760628033, "lr": 0.0026291804804649314, "grad_norm": 0.159472, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:05.424168+00:00", "epoch": 0, "step": 21148, "train_loss": 3.4706547260284424, "perplexity": 32.157790095971755, "lr": 0.0026291804804649314, "grad_norm": 0.165521, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:05.728887+00:00", "epoch": 0, "step": 21149, "train_loss": 3.5329787731170654, "perplexity": 34.22576671507179, "lr": 0.0026291804804649314, "grad_norm": 0.155669, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:06.033571+00:00", "epoch": 0, "step": 21150, "train_loss": 3.4823098182678223, "perplexity": 32.53478479212643, "lr": 0.0026291804804649314, "grad_norm": 0.19744, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:06.340020+00:00", "epoch": 0, "step": 21151, "train_loss": 3.440838575363159, "perplexity": 31.213121751616264, "lr": 0.0026291804804649314, "grad_norm": 0.17207, "tokens_per_sec": 106864} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:06.646026+00:00", "epoch": 0, "step": 21152, "train_loss": 3.5066678524017334, "perplexity": 33.336998704234446, "lr": 0.0026291804804649314, "grad_norm": 0.172402, "tokens_per_sec": 107082} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:06.950742+00:00", "epoch": 0, "step": 21153, "train_loss": 3.505721092224121, "perplexity": 33.305451497596096, "lr": 0.0026291804804649314, "grad_norm": 0.161067, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:07.255618+00:00", "epoch": 0, "step": 21154, "train_loss": 3.517308473587036, "perplexity": 33.69361904508778, "lr": 0.0026291804804649314, "grad_norm": 0.167958, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:07.559738+00:00", "epoch": 0, "step": 21155, "train_loss": 3.4515085220336914, "perplexity": 31.547947204401407, "lr": 0.0026291804804649314, "grad_norm": 0.17224, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:07.865503+00:00", "epoch": 0, "step": 21156, "train_loss": 3.561845541000366, "perplexity": 35.22815218730212, "lr": 0.0026291804804649314, "grad_norm": 0.153831, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:08.170091+00:00", "epoch": 0, "step": 21157, "train_loss": 3.6136908531188965, "perplexity": 37.102741179317036, "lr": 0.0026291804804649314, "grad_norm": 0.176626, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:08.475192+00:00", "epoch": 0, "step": 21158, "train_loss": 3.457927942276001, "perplexity": 31.75111815744821, "lr": 0.0026291804804649314, "grad_norm": 0.176527, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:08.780572+00:00", "epoch": 0, "step": 21159, "train_loss": 3.5244317054748535, "perplexity": 33.93448335438763, "lr": 0.0026291804804649314, "grad_norm": 0.166326, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:09.085441+00:00", "epoch": 0, "step": 21160, "train_loss": 3.5552031993865967, "perplexity": 34.99493019362168, "lr": 0.0026291804804649314, "grad_norm": 0.218514, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:09.389695+00:00", "epoch": 0, "step": 21161, "train_loss": 3.4904327392578125, "perplexity": 32.80013854315363, "lr": 0.0026291804804649314, "grad_norm": 0.189267, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:09.694707+00:00", "epoch": 0, "step": 21162, "train_loss": 3.4318783283233643, "perplexity": 30.934693724611815, "lr": 0.0026291804804649314, "grad_norm": 0.197123, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:10.001124+00:00", "epoch": 0, "step": 21163, "train_loss": 3.428678274154663, "perplexity": 30.835859251176835, "lr": 0.0026291804804649314, "grad_norm": 0.159628, "tokens_per_sec": 106939} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:10.306618+00:00", "epoch": 0, "step": 21164, "train_loss": 3.470766544342041, "perplexity": 32.16138612687666, "lr": 0.0026291804804649314, "grad_norm": 0.19132, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:10.611279+00:00", "epoch": 0, "step": 21165, "train_loss": 3.593295097351074, "perplexity": 36.35366765907383, "lr": 0.0026291804804649314, "grad_norm": 0.196021, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:10.916033+00:00", "epoch": 0, "step": 21166, "train_loss": 3.4946837425231934, "perplexity": 32.93986882572683, "lr": 0.0026291804804649314, "grad_norm": 0.218603, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:11.221996+00:00", "epoch": 0, "step": 21167, "train_loss": 3.4607114791870117, "perplexity": 31.83962168608034, "lr": 0.0026291804804649314, "grad_norm": 0.260751, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:11.526863+00:00", "epoch": 0, "step": 21168, "train_loss": 3.618661403656006, "perplexity": 37.28762132682938, "lr": 0.0026291804804649314, "grad_norm": 0.233615, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:11.831541+00:00", "epoch": 0, "step": 21169, "train_loss": 3.4626052379608154, "perplexity": 31.89997537863303, "lr": 0.0026291804804649314, "grad_norm": 0.195895, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:12.136680+00:00", "epoch": 0, "step": 21170, "train_loss": 3.517385244369507, "perplexity": 33.69620582987962, "lr": 0.0026291804804649314, "grad_norm": 0.217829, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:12.441229+00:00", "epoch": 0, "step": 21171, "train_loss": 3.474066734313965, "perplexity": 32.26770014258895, "lr": 0.0026291804804649314, "grad_norm": 0.183154, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:12.746779+00:00", "epoch": 0, "step": 21172, "train_loss": 3.5279994010925293, "perplexity": 34.05576748572453, "lr": 0.0026291804804649314, "grad_norm": 0.164438, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:13.053343+00:00", "epoch": 0, "step": 21173, "train_loss": 3.4552762508392334, "perplexity": 31.667035519196915, "lr": 0.0026291804804649314, "grad_norm": 0.169549, "tokens_per_sec": 106887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:13.357958+00:00", "epoch": 0, "step": 21174, "train_loss": 3.47269606590271, "perplexity": 32.223502122634244, "lr": 0.0026291804804649314, "grad_norm": 0.16807, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:13.663594+00:00", "epoch": 0, "step": 21175, "train_loss": 3.470371723175049, "perplexity": 32.14869063726289, "lr": 0.0026291804804649314, "grad_norm": 0.173604, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:13.970118+00:00", "epoch": 0, "step": 21176, "train_loss": 3.4323675632476807, "perplexity": 30.949831759880755, "lr": 0.0026291804804649314, "grad_norm": 0.166646, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:14.275729+00:00", "epoch": 0, "step": 21177, "train_loss": 3.5276589393615723, "perplexity": 34.044174773726645, "lr": 0.0026291804804649314, "grad_norm": 0.159498, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:14.581234+00:00", "epoch": 0, "step": 21178, "train_loss": 3.388021469116211, "perplexity": 29.60731529426727, "lr": 0.0026291804804649314, "grad_norm": 0.159707, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:14.886181+00:00", "epoch": 0, "step": 21179, "train_loss": 3.5433900356292725, "perplexity": 34.58396155174154, "lr": 0.0026291804804649314, "grad_norm": 0.1583, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:15.191647+00:00", "epoch": 0, "step": 21180, "train_loss": 3.5867984294891357, "perplexity": 36.11825548009516, "lr": 0.0026291804804649314, "grad_norm": 0.166675, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:15.497331+00:00", "epoch": 0, "step": 21181, "train_loss": 3.5189931392669678, "perplexity": 33.75042936849067, "lr": 0.0026291804804649314, "grad_norm": 0.178208, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:15.804189+00:00", "epoch": 0, "step": 21182, "train_loss": 3.482067346572876, "perplexity": 32.52689698403692, "lr": 0.0026291804804649314, "grad_norm": 0.167176, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:16.108875+00:00", "epoch": 0, "step": 21183, "train_loss": 3.5147860050201416, "perplexity": 33.60873505375027, "lr": 0.0026291804804649314, "grad_norm": 0.158079, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:16.413787+00:00", "epoch": 0, "step": 21184, "train_loss": 3.4730384349823, "perplexity": 32.23453634217714, "lr": 0.0026291804804649314, "grad_norm": 0.165968, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:16.718971+00:00", "epoch": 0, "step": 21185, "train_loss": 3.5040900707244873, "perplexity": 33.251173866173, "lr": 0.0026291804804649314, "grad_norm": 0.160412, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:17.023519+00:00", "epoch": 0, "step": 21186, "train_loss": 3.406635284423828, "perplexity": 30.1635814497564, "lr": 0.0026291804804649314, "grad_norm": 0.149959, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:17.328945+00:00", "epoch": 0, "step": 21187, "train_loss": 3.4715898036956787, "perplexity": 32.18787419060864, "lr": 0.0026291804804649314, "grad_norm": 0.157447, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:17.635387+00:00", "epoch": 0, "step": 21188, "train_loss": 3.5755391120910645, "perplexity": 35.71386940543109, "lr": 0.0026291804804649314, "grad_norm": 0.153275, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:17.941982+00:00", "epoch": 0, "step": 21189, "train_loss": 3.597033977508545, "perplexity": 36.489844080759404, "lr": 0.0026291804804649314, "grad_norm": 0.161916, "tokens_per_sec": 106878} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:18.246830+00:00", "epoch": 0, "step": 21190, "train_loss": 3.424391746520996, "perplexity": 30.70396337799938, "lr": 0.0026291804804649314, "grad_norm": 0.162854, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:18.551789+00:00", "epoch": 0, "step": 21191, "train_loss": 3.4360437393188477, "perplexity": 31.06381817950696, "lr": 0.0026291804804649314, "grad_norm": 0.210815, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:18.856518+00:00", "epoch": 0, "step": 21192, "train_loss": 3.4973247051239014, "perplexity": 33.02697676113844, "lr": 0.0026291804804649314, "grad_norm": 0.159287, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:19.162166+00:00", "epoch": 0, "step": 21193, "train_loss": 3.510789155960083, "perplexity": 33.47467410180221, "lr": 0.0026291804804649314, "grad_norm": 0.200153, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:19.468508+00:00", "epoch": 0, "step": 21194, "train_loss": 3.5443203449249268, "perplexity": 34.616150303079415, "lr": 0.0026291804804649314, "grad_norm": 0.19472, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:19.773250+00:00", "epoch": 0, "step": 21195, "train_loss": 3.4449427127838135, "perplexity": 31.341487928645183, "lr": 0.0026291804804649314, "grad_norm": 0.203329, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:20.077982+00:00", "epoch": 0, "step": 21196, "train_loss": 3.4473767280578613, "perplexity": 31.417866504576814, "lr": 0.0026291804804649314, "grad_norm": 0.185199, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:20.383134+00:00", "epoch": 0, "step": 21197, "train_loss": 3.468900680541992, "perplexity": 32.10143331003201, "lr": 0.0026291804804649314, "grad_norm": 0.18367, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:20.688750+00:00", "epoch": 0, "step": 21198, "train_loss": 3.4322409629821777, "perplexity": 30.945913750978402, "lr": 0.0026291804804649314, "grad_norm": 0.212029, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:20.995074+00:00", "epoch": 0, "step": 21199, "train_loss": 3.457343578338623, "perplexity": 31.732569369174886, "lr": 0.0026291804804649314, "grad_norm": 0.182139, "tokens_per_sec": 106972} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:21.301263+00:00", "epoch": 0, "step": 21200, "train_loss": 3.468165397644043, "perplexity": 32.07783835066756, "lr": 0.0026291804804649314, "grad_norm": 0.175829, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:21.605916+00:00", "epoch": 0, "step": 21201, "train_loss": 3.5050954818725586, "perplexity": 33.284621778699, "lr": 0.0026291804804649314, "grad_norm": 0.209541, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:21.910368+00:00", "epoch": 0, "step": 21202, "train_loss": 3.4659323692321777, "perplexity": 32.00628754346383, "lr": 0.0026291804804649314, "grad_norm": 0.209223, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:22.217697+00:00", "epoch": 0, "step": 21203, "train_loss": 3.5003626346588135, "perplexity": 33.127462946983506, "lr": 0.0026291804804649314, "grad_norm": 0.190549, "tokens_per_sec": 106622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:22.523635+00:00", "epoch": 0, "step": 21204, "train_loss": 3.54382061958313, "perplexity": 34.598856057083914, "lr": 0.0026291804804649314, "grad_norm": 0.218405, "tokens_per_sec": 107106} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:22.828544+00:00", "epoch": 0, "step": 21205, "train_loss": 3.529158592224121, "perplexity": 34.09526751899457, "lr": 0.0026291804804649314, "grad_norm": 0.194017, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:23.133950+00:00", "epoch": 0, "step": 21206, "train_loss": 3.5266878604888916, "perplexity": 34.01113124140075, "lr": 0.0026291804804649314, "grad_norm": 0.184989, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:23.744512+00:00", "epoch": 0, "step": 21207, "train_loss": 3.6132984161376953, "perplexity": 37.08818354823756, "lr": 0.0026291804804649314, "grad_norm": 0.208249, "tokens_per_sec": 107077} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:24.050344+00:00", "epoch": 0, "step": 21208, "train_loss": 3.48770809173584, "perplexity": 32.71089136571414, "lr": 0.0026291804804649314, "grad_norm": 0.182015, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:24.355449+00:00", "epoch": 0, "step": 21209, "train_loss": 3.5213143825531006, "perplexity": 33.82886332299357, "lr": 0.0026291804804649314, "grad_norm": 0.203625, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:24.660314+00:00", "epoch": 0, "step": 21210, "train_loss": 3.472909450531006, "perplexity": 32.23037885632561, "lr": 0.0026291804804649314, "grad_norm": 0.183776, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:24.966900+00:00", "epoch": 0, "step": 21211, "train_loss": 3.5043749809265137, "perplexity": 33.26064881452745, "lr": 0.0026291804804649314, "grad_norm": 0.18551, "tokens_per_sec": 106819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:25.271631+00:00", "epoch": 0, "step": 21212, "train_loss": 3.5541961193084717, "perplexity": 34.9597052367416, "lr": 0.0026291804804649314, "grad_norm": 0.159709, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:25.576037+00:00", "epoch": 0, "step": 21213, "train_loss": 3.357184648513794, "perplexity": 28.70825317541723, "lr": 0.0026291804804649314, "grad_norm": 0.18712, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:25.880525+00:00", "epoch": 0, "step": 21214, "train_loss": 3.6204917430877686, "perplexity": 37.35593282800653, "lr": 0.0026291804804649314, "grad_norm": 0.175192, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:26.184992+00:00", "epoch": 0, "step": 21215, "train_loss": 3.5681307315826416, "perplexity": 35.450265117521056, "lr": 0.0026291804804649314, "grad_norm": 0.183651, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:26.491037+00:00", "epoch": 0, "step": 21216, "train_loss": 3.455312967300415, "perplexity": 31.668198242022694, "lr": 0.0026291804804649314, "grad_norm": 0.173054, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:26.796219+00:00", "epoch": 0, "step": 21217, "train_loss": 3.4284074306488037, "perplexity": 30.827508689849758, "lr": 0.0026291804804649314, "grad_norm": 0.16922, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:27.100607+00:00", "epoch": 0, "step": 21218, "train_loss": 3.425529718399048, "perplexity": 30.73892351292075, "lr": 0.0026291804804649314, "grad_norm": 0.163038, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:27.405512+00:00", "epoch": 0, "step": 21219, "train_loss": 3.553335666656494, "perplexity": 34.92963700365889, "lr": 0.0026291804804649314, "grad_norm": 0.152501, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:27.710869+00:00", "epoch": 0, "step": 21220, "train_loss": 3.4241631031036377, "perplexity": 30.69694392139304, "lr": 0.0026291804804649314, "grad_norm": 0.15607, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:28.016524+00:00", "epoch": 0, "step": 21221, "train_loss": 3.4194397926330566, "perplexity": 30.552294605491475, "lr": 0.0026291804804649314, "grad_norm": 0.165777, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:28.321359+00:00", "epoch": 0, "step": 21222, "train_loss": 3.495847463607788, "perplexity": 32.97822395861301, "lr": 0.0026291804804649314, "grad_norm": 0.150333, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:28.626043+00:00", "epoch": 0, "step": 21223, "train_loss": 3.4513461589813232, "perplexity": 31.542825399204563, "lr": 0.0026291804804649314, "grad_norm": 0.155338, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:28.931019+00:00", "epoch": 0, "step": 21224, "train_loss": 3.3997979164123535, "perplexity": 29.958045406351236, "lr": 0.0026291804804649314, "grad_norm": 0.142293, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:29.235505+00:00", "epoch": 0, "step": 21225, "train_loss": 3.5251076221466064, "perplexity": 33.95742799089181, "lr": 0.0026291804804649314, "grad_norm": 0.16447, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:29.540898+00:00", "epoch": 0, "step": 21226, "train_loss": 3.5821964740753174, "perplexity": 35.9524227488355, "lr": 0.0026291804804649314, "grad_norm": 0.172261, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:29.845164+00:00", "epoch": 0, "step": 21227, "train_loss": 3.4997153282165527, "perplexity": 33.106026265601656, "lr": 0.0026291804804649314, "grad_norm": 0.177709, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:30.149209+00:00", "epoch": 0, "step": 21228, "train_loss": 3.6109864711761475, "perplexity": 37.00253675263818, "lr": 0.0026291804804649314, "grad_norm": 0.20541, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:30.452864+00:00", "epoch": 0, "step": 21229, "train_loss": 3.4494762420654297, "perplexity": 31.483898048211124, "lr": 0.0026291804804649314, "grad_norm": 0.191145, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:30.757395+00:00", "epoch": 0, "step": 21230, "train_loss": 3.526569128036499, "perplexity": 34.0070932561045, "lr": 0.0026291804804649314, "grad_norm": 0.164426, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:31.062989+00:00", "epoch": 0, "step": 21231, "train_loss": 3.442610263824463, "perplexity": 31.268470695329526, "lr": 0.0026291804804649314, "grad_norm": 0.197349, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:31.369091+00:00", "epoch": 0, "step": 21232, "train_loss": 3.4417715072631836, "perplexity": 31.24225505617924, "lr": 0.0026291804804649314, "grad_norm": 0.158066, "tokens_per_sec": 107049} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:31.672951+00:00", "epoch": 0, "step": 21233, "train_loss": 3.544391632080078, "perplexity": 34.618618077916096, "lr": 0.0026291804804649314, "grad_norm": 0.193707, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:31.976917+00:00", "epoch": 0, "step": 21234, "train_loss": 3.5621345043182373, "perplexity": 35.23833330195412, "lr": 0.0026291804804649314, "grad_norm": 0.186604, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:32.282514+00:00", "epoch": 0, "step": 21235, "train_loss": 3.569932222366333, "perplexity": 35.5141860025651, "lr": 0.0026291804804649314, "grad_norm": 0.189984, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:32.586437+00:00", "epoch": 0, "step": 21236, "train_loss": 3.539027690887451, "perplexity": 34.43342297822193, "lr": 0.0026291804804649314, "grad_norm": 0.168921, "tokens_per_sec": 107817} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:32.892066+00:00", "epoch": 0, "step": 21237, "train_loss": 3.561335802078247, "perplexity": 35.21019960293214, "lr": 0.0026291804804649314, "grad_norm": 0.198859, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:33.196379+00:00", "epoch": 0, "step": 21238, "train_loss": 3.4967041015625, "perplexity": 33.00648646057257, "lr": 0.0026291804804649314, "grad_norm": 0.24003, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:33.500986+00:00", "epoch": 0, "step": 21239, "train_loss": 3.5719149112701416, "perplexity": 35.5846694353507, "lr": 0.0026291804804649314, "grad_norm": 0.222982, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:33.805155+00:00", "epoch": 0, "step": 21240, "train_loss": 3.446932315826416, "perplexity": 31.40390712250608, "lr": 0.0026291804804649314, "grad_norm": 0.188412, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:34.109838+00:00", "epoch": 0, "step": 21241, "train_loss": 3.5868375301361084, "perplexity": 36.1196677548622, "lr": 0.0026291804804649314, "grad_norm": 0.198734, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:34.414512+00:00", "epoch": 0, "step": 21242, "train_loss": 3.5405266284942627, "perplexity": 34.48507523293768, "lr": 0.0026291804804649314, "grad_norm": 0.200875, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:34.720124+00:00", "epoch": 0, "step": 21243, "train_loss": 3.4411420822143555, "perplexity": 31.222596585680897, "lr": 0.0026291804804649314, "grad_norm": 0.177622, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:35.024231+00:00", "epoch": 0, "step": 21244, "train_loss": 3.5484771728515625, "perplexity": 34.760343168207896, "lr": 0.0026291804804649314, "grad_norm": 0.170952, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:35.328429+00:00", "epoch": 0, "step": 21245, "train_loss": 3.6269187927246094, "perplexity": 37.59679444821228, "lr": 0.0026291804804649314, "grad_norm": 0.175705, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:35.633448+00:00", "epoch": 0, "step": 21246, "train_loss": 3.3473122119903564, "perplexity": 28.426227200645915, "lr": 0.0026291804804649314, "grad_norm": 0.218502, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:35.939573+00:00", "epoch": 0, "step": 21247, "train_loss": 3.5508017539978027, "perplexity": 34.84124039623053, "lr": 0.0026291804804649314, "grad_norm": 0.190927, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:36.244286+00:00", "epoch": 0, "step": 21248, "train_loss": 3.432892322540283, "perplexity": 30.966077233810147, "lr": 0.0026291804804649314, "grad_norm": 0.200901, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:36.548576+00:00", "epoch": 0, "step": 21249, "train_loss": 3.593581199645996, "perplexity": 36.364070014817884, "lr": 0.0026291804804649314, "grad_norm": 0.174734, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:36.852685+00:00", "epoch": 0, "step": 21250, "train_loss": 3.5257887840270996, "perplexity": 33.9805663759971, "lr": 0.0026291804804649314, "grad_norm": 0.189112, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:37.156551+00:00", "epoch": 0, "step": 21251, "train_loss": 3.5065295696258545, "perplexity": 33.3323890902366, "lr": 0.0026291804804649314, "grad_norm": 0.150828, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:37.462231+00:00", "epoch": 0, "step": 21252, "train_loss": 3.576425790786743, "perplexity": 35.745550175829614, "lr": 0.0026291804804649314, "grad_norm": 0.179306, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:37.766900+00:00", "epoch": 0, "step": 21253, "train_loss": 3.497577667236328, "perplexity": 33.03533239173144, "lr": 0.0026291804804649314, "grad_norm": 0.182392, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:38.071508+00:00", "epoch": 0, "step": 21254, "train_loss": 3.4201464653015137, "perplexity": 30.573892707543212, "lr": 0.0026291804804649314, "grad_norm": 0.187821, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:38.375885+00:00", "epoch": 0, "step": 21255, "train_loss": 3.6588680744171143, "perplexity": 38.81737960941556, "lr": 0.0026291804804649314, "grad_norm": 0.172189, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:38.680694+00:00", "epoch": 0, "step": 21256, "train_loss": 3.4556474685668945, "perplexity": 31.678793066334062, "lr": 0.0026291804804649314, "grad_norm": 0.204525, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:38.986329+00:00", "epoch": 0, "step": 21257, "train_loss": 3.4897682666778564, "perplexity": 32.77835098989045, "lr": 0.0026291804804649314, "grad_norm": 0.164006, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:39.290898+00:00", "epoch": 0, "step": 21258, "train_loss": 3.445172071456909, "perplexity": 31.34867719515814, "lr": 0.0026291804804649314, "grad_norm": 0.181105, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:39.595577+00:00", "epoch": 0, "step": 21259, "train_loss": 3.5722789764404297, "perplexity": 35.59762693263193, "lr": 0.0026291804804649314, "grad_norm": 0.184022, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:39.900288+00:00", "epoch": 0, "step": 21260, "train_loss": 3.3979835510253906, "perplexity": 29.90373984566717, "lr": 0.0026291804804649314, "grad_norm": 0.175805, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:40.204293+00:00", "epoch": 0, "step": 21261, "train_loss": 3.545919418334961, "perplexity": 34.671548349553795, "lr": 0.0026291804804649314, "grad_norm": 0.182154, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:40.509098+00:00", "epoch": 0, "step": 21262, "train_loss": 3.5193305015563965, "perplexity": 33.761817411451744, "lr": 0.0026291804804649314, "grad_norm": 0.207175, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:40.814545+00:00", "epoch": 0, "step": 21263, "train_loss": 3.5133097171783447, "perplexity": 33.55915549266844, "lr": 0.0026291804804649314, "grad_norm": 0.183949, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:41.120536+00:00", "epoch": 0, "step": 21264, "train_loss": 3.3780837059020996, "perplexity": 29.31454197094803, "lr": 0.0026291804804649314, "grad_norm": 0.177994, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:41.425696+00:00", "epoch": 0, "step": 21265, "train_loss": 3.5509262084960938, "perplexity": 34.84557681516171, "lr": 0.0026291804804649314, "grad_norm": 0.191992, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:41.730205+00:00", "epoch": 0, "step": 21266, "train_loss": 3.478915214538574, "perplexity": 32.42452933301299, "lr": 0.0026291804804649314, "grad_norm": 0.227219, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:42.034472+00:00", "epoch": 0, "step": 21267, "train_loss": 3.5060853958129883, "perplexity": 33.317587003475474, "lr": 0.0026291804804649314, "grad_norm": 0.196413, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:42.338940+00:00", "epoch": 0, "step": 21268, "train_loss": 3.486231803894043, "perplexity": 32.66263630244223, "lr": 0.0026291804804649314, "grad_norm": 0.193978, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:42.644382+00:00", "epoch": 0, "step": 21269, "train_loss": 3.6511549949645996, "perplexity": 38.51912976732584, "lr": 0.0026291804804649314, "grad_norm": 0.187804, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:42.949105+00:00", "epoch": 0, "step": 21270, "train_loss": 3.4651215076446533, "perplexity": 31.98034539350343, "lr": 0.0026291804804649314, "grad_norm": 0.194088, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:43.254127+00:00", "epoch": 0, "step": 21271, "train_loss": 3.456897497177124, "perplexity": 31.718417224510574, "lr": 0.0026291804804649314, "grad_norm": 0.183691, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:43.558645+00:00", "epoch": 0, "step": 21272, "train_loss": 3.594050168991089, "perplexity": 36.38112764835863, "lr": 0.0026291804804649314, "grad_norm": 0.188062, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:43.863156+00:00", "epoch": 0, "step": 21273, "train_loss": 3.3453049659729004, "perplexity": 28.369225996164786, "lr": 0.0026291804804649314, "grad_norm": 0.184046, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:44.168650+00:00", "epoch": 0, "step": 21274, "train_loss": 3.5543787479400635, "perplexity": 34.96609046291464, "lr": 0.0026291804804649314, "grad_norm": 0.20327, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:44.473217+00:00", "epoch": 0, "step": 21275, "train_loss": 3.46077299118042, "perplexity": 31.841580264917038, "lr": 0.0026291804804649314, "grad_norm": 0.166427, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:44.777034+00:00", "epoch": 0, "step": 21276, "train_loss": 3.561283826828003, "perplexity": 35.20836959155469, "lr": 0.0026291804804649314, "grad_norm": 0.190069, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:45.081516+00:00", "epoch": 0, "step": 21277, "train_loss": 3.505288600921631, "perplexity": 33.29105029392002, "lr": 0.0026291804804649314, "grad_norm": 0.192558, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:45.385988+00:00", "epoch": 0, "step": 21278, "train_loss": 3.4645843505859375, "perplexity": 31.96317153817689, "lr": 0.0026291804804649314, "grad_norm": 0.168721, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:45.691035+00:00", "epoch": 0, "step": 21279, "train_loss": 3.390727996826172, "perplexity": 29.68755685254651, "lr": 0.0026291804804649314, "grad_norm": 0.21057, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:45.996077+00:00", "epoch": 0, "step": 21280, "train_loss": 3.498011350631714, "perplexity": 33.049662373963855, "lr": 0.0026291804804649314, "grad_norm": 0.212638, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:46.300135+00:00", "epoch": 0, "step": 21281, "train_loss": 3.5650994777679443, "perplexity": 35.34296906908707, "lr": 0.0026291804804649314, "grad_norm": 0.194935, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:46.604913+00:00", "epoch": 0, "step": 21282, "train_loss": 3.477771282196045, "perplexity": 32.38745907219527, "lr": 0.0026291804804649314, "grad_norm": 0.183885, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:46.910165+00:00", "epoch": 0, "step": 21283, "train_loss": 3.530351161956787, "perplexity": 34.13595275823536, "lr": 0.0026291804804649314, "grad_norm": 0.195944, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:47.216052+00:00", "epoch": 0, "step": 21284, "train_loss": 3.4978718757629395, "perplexity": 33.045053098087756, "lr": 0.0026291804804649314, "grad_norm": 0.187787, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:47.522090+00:00", "epoch": 0, "step": 21285, "train_loss": 3.571913957595825, "perplexity": 35.58463549918159, "lr": 0.0026291804804649314, "grad_norm": 0.160813, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:47.826753+00:00", "epoch": 0, "step": 21286, "train_loss": 3.51774525642395, "perplexity": 33.70833905408982, "lr": 0.0026291804804649314, "grad_norm": 0.183077, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:48.131417+00:00", "epoch": 0, "step": 21287, "train_loss": 3.5556247234344482, "perplexity": 35.009684507681804, "lr": 0.0026291804804649314, "grad_norm": 0.199217, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:48.436275+00:00", "epoch": 0, "step": 21288, "train_loss": 3.391402006149292, "perplexity": 29.707573287519907, "lr": 0.0026291804804649314, "grad_norm": 0.195067, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:48.741632+00:00", "epoch": 0, "step": 21289, "train_loss": 3.5620503425598145, "perplexity": 35.23536770665615, "lr": 0.0026291804804649314, "grad_norm": 0.179756, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:49.048812+00:00", "epoch": 0, "step": 21290, "train_loss": 3.5171828269958496, "perplexity": 33.689385822660576, "lr": 0.0026291804804649314, "grad_norm": 0.185779, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:49.354872+00:00", "epoch": 0, "step": 21291, "train_loss": 3.3957345485687256, "perplexity": 29.83656183136599, "lr": 0.0026291804804649314, "grad_norm": 0.151132, "tokens_per_sec": 107006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:49.660353+00:00", "epoch": 0, "step": 21292, "train_loss": 3.503471612930298, "perplexity": 33.23061577634023, "lr": 0.0026291804804649314, "grad_norm": 0.151331, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:49.964445+00:00", "epoch": 0, "step": 21293, "train_loss": 3.59420108795166, "perplexity": 36.38661866466654, "lr": 0.0026291804804649314, "grad_norm": 0.194824, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:50.269559+00:00", "epoch": 0, "step": 21294, "train_loss": 3.56632137298584, "perplexity": 35.38618086874813, "lr": 0.0026291804804649314, "grad_norm": 0.177483, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:50.575894+00:00", "epoch": 0, "step": 21295, "train_loss": 3.5918869972229004, "perplexity": 36.30251407811749, "lr": 0.0026291804804649314, "grad_norm": 0.181346, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:50.883026+00:00", "epoch": 0, "step": 21296, "train_loss": 3.520233631134033, "perplexity": 33.79232248029233, "lr": 0.0026291804804649314, "grad_norm": 0.226102, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:51.188627+00:00", "epoch": 0, "step": 21297, "train_loss": 3.507011651992798, "perplexity": 33.34846192116801, "lr": 0.0026291804804649314, "grad_norm": 0.17914, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:51.494032+00:00", "epoch": 0, "step": 21298, "train_loss": 3.4676432609558105, "perplexity": 32.0610937062625, "lr": 0.0026291804804649314, "grad_norm": 0.14927, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:51.799432+00:00", "epoch": 0, "step": 21299, "train_loss": 3.4332094192504883, "perplexity": 30.975898032018055, "lr": 0.0026291804804649314, "grad_norm": 0.172852, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:52.104799+00:00", "epoch": 0, "step": 21300, "train_loss": 3.5528764724731445, "perplexity": 34.91360119957606, "lr": 0.0026291804804649314, "grad_norm": 0.152846, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:52.411051+00:00", "epoch": 0, "step": 21301, "train_loss": 3.487088680267334, "perplexity": 32.690636138264864, "lr": 0.0026291804804649314, "grad_norm": 0.190004, "tokens_per_sec": 106937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:52.716457+00:00", "epoch": 0, "step": 21302, "train_loss": 3.561373233795166, "perplexity": 35.211517605823744, "lr": 0.0026291804804649314, "grad_norm": 0.17583, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:53.020663+00:00", "epoch": 0, "step": 21303, "train_loss": 3.490816831588745, "perplexity": 32.812739244583064, "lr": 0.0026291804804649314, "grad_norm": 0.168953, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:53.324700+00:00", "epoch": 0, "step": 21304, "train_loss": 3.399332284927368, "perplexity": 29.94409924431946, "lr": 0.0026291804804649314, "grad_norm": 0.189855, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:53.629441+00:00", "epoch": 0, "step": 21305, "train_loss": 3.492535352706909, "perplexity": 32.869177110862324, "lr": 0.0026291804804649314, "grad_norm": 0.183068, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:53.935355+00:00", "epoch": 0, "step": 21306, "train_loss": 3.4871792793273926, "perplexity": 32.69359801334123, "lr": 0.0026291804804649314, "grad_norm": 0.190853, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:54.239707+00:00", "epoch": 0, "step": 21307, "train_loss": 3.4262771606445312, "perplexity": 30.7619076715305, "lr": 0.0026291804804649314, "grad_norm": 0.181193, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:54.545102+00:00", "epoch": 0, "step": 21308, "train_loss": 3.458664894104004, "perplexity": 31.774525826118197, "lr": 0.0026291804804649314, "grad_norm": 0.15764, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:54.849697+00:00", "epoch": 0, "step": 21309, "train_loss": 3.5711934566497803, "perplexity": 35.559005969798584, "lr": 0.0026291804804649314, "grad_norm": 0.179093, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:55.153998+00:00", "epoch": 0, "step": 21310, "train_loss": 3.4240472316741943, "perplexity": 30.693387228684838, "lr": 0.0026291804804649314, "grad_norm": 0.144089, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:55.457917+00:00", "epoch": 0, "step": 21311, "train_loss": 3.3657002449035645, "perplexity": 28.953764928400023, "lr": 0.0026291804804649314, "grad_norm": 0.155874, "tokens_per_sec": 107763} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:55.763914+00:00", "epoch": 0, "step": 21312, "train_loss": 3.5290472507476807, "perplexity": 34.09147151289975, "lr": 0.0026291804804649314, "grad_norm": 0.144016, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:56.068480+00:00", "epoch": 0, "step": 21313, "train_loss": 3.5014231204986572, "perplexity": 33.16261278700832, "lr": 0.0026291804804649314, "grad_norm": 0.162866, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:56.373828+00:00", "epoch": 0, "step": 21314, "train_loss": 3.553675651550293, "perplexity": 34.94151457156843, "lr": 0.0026291804804649314, "grad_norm": 0.163867, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:56.678030+00:00", "epoch": 0, "step": 21315, "train_loss": 3.6033811569213867, "perplexity": 36.722188252928866, "lr": 0.0026291804804649314, "grad_norm": 0.168438, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:56.983114+00:00", "epoch": 0, "step": 21316, "train_loss": 3.533637523651123, "perplexity": 34.24832038497932, "lr": 0.0026291804804649314, "grad_norm": 0.180096, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:57.289622+00:00", "epoch": 0, "step": 21317, "train_loss": 3.4108824729919434, "perplexity": 30.291964308180226, "lr": 0.0026291804804649314, "grad_norm": 0.171197, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:57.594463+00:00", "epoch": 0, "step": 21318, "train_loss": 3.466672420501709, "perplexity": 32.029982603891696, "lr": 0.0026291804804649314, "grad_norm": 0.175309, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:57.899544+00:00", "epoch": 0, "step": 21319, "train_loss": 3.400460720062256, "perplexity": 29.97790829005925, "lr": 0.0026291804804649314, "grad_norm": 0.146232, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:58.205230+00:00", "epoch": 0, "step": 21320, "train_loss": 3.5903568267822266, "perplexity": 36.247007522228905, "lr": 0.0026291804804649314, "grad_norm": 0.154582, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:58.509517+00:00", "epoch": 0, "step": 21321, "train_loss": 3.516507148742676, "perplexity": 33.666630325860325, "lr": 0.0026291804804649314, "grad_norm": 0.173006, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:58.814225+00:00", "epoch": 0, "step": 21322, "train_loss": 3.4538261890411377, "perplexity": 31.621149637453954, "lr": 0.0026291804804649314, "grad_norm": 0.166879, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:59.120414+00:00", "epoch": 0, "step": 21323, "train_loss": 3.5480306148529053, "perplexity": 34.74482412426456, "lr": 0.0026291804804649314, "grad_norm": 0.184121, "tokens_per_sec": 107012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:59.425313+00:00", "epoch": 0, "step": 21324, "train_loss": 3.410583972930908, "perplexity": 30.28292350439273, "lr": 0.0026291804804649314, "grad_norm": 0.222033, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:49:59.730804+00:00", "epoch": 0, "step": 21325, "train_loss": 3.5211751461029053, "perplexity": 33.82415344005164, "lr": 0.0026291804804649314, "grad_norm": 0.200062, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:00.035006+00:00", "epoch": 0, "step": 21326, "train_loss": 3.461768865585327, "perplexity": 31.873306274653135, "lr": 0.0026291804804649314, "grad_norm": 0.17385, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:00.339693+00:00", "epoch": 0, "step": 21327, "train_loss": 3.4652349948883057, "perplexity": 31.98397496070408, "lr": 0.0026291804804649314, "grad_norm": 0.216516, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:00.645473+00:00", "epoch": 0, "step": 21328, "train_loss": 3.4332616329193115, "perplexity": 30.977515439524424, "lr": 0.0026291804804649314, "grad_norm": 0.16882, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:00.950610+00:00", "epoch": 0, "step": 21329, "train_loss": 3.4290459156036377, "perplexity": 30.847197875298704, "lr": 0.0026291804804649314, "grad_norm": 0.193769, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:01.255622+00:00", "epoch": 0, "step": 21330, "train_loss": 3.3810739517211914, "perplexity": 29.40233084726124, "lr": 0.0026291804804649314, "grad_norm": 0.173301, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:01.560112+00:00", "epoch": 0, "step": 21331, "train_loss": 3.508868455886841, "perplexity": 33.41044099881273, "lr": 0.0026291804804649314, "grad_norm": 0.163994, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:01.864889+00:00", "epoch": 0, "step": 21332, "train_loss": 3.4800894260406494, "perplexity": 32.46262495008396, "lr": 0.0026291804804649314, "grad_norm": 0.159364, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:02.171429+00:00", "epoch": 0, "step": 21333, "train_loss": 3.374521255493164, "perplexity": 29.210296164422974, "lr": 0.0026291804804649314, "grad_norm": 0.151421, "tokens_per_sec": 106902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:02.476341+00:00", "epoch": 0, "step": 21334, "train_loss": 3.465120315551758, "perplexity": 31.980307269983616, "lr": 0.0026291804804649314, "grad_norm": 0.173308, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:02.780519+00:00", "epoch": 0, "step": 21335, "train_loss": 3.409590482711792, "perplexity": 30.252852656105084, "lr": 0.0026291804804649314, "grad_norm": 0.171837, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:03.085396+00:00", "epoch": 0, "step": 21336, "train_loss": 3.5292181968688965, "perplexity": 34.09729981587012, "lr": 0.0026291804804649314, "grad_norm": 0.159254, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:03.390898+00:00", "epoch": 0, "step": 21337, "train_loss": 3.5979387760162354, "perplexity": 36.52287497812943, "lr": 0.0026291804804649314, "grad_norm": 0.169745, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:03.695427+00:00", "epoch": 0, "step": 21338, "train_loss": 3.5482587814331055, "perplexity": 34.75275263644101, "lr": 0.0026291804804649314, "grad_norm": 0.175602, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:04.000896+00:00", "epoch": 0, "step": 21339, "train_loss": 3.5960609912872314, "perplexity": 36.45435723215793, "lr": 0.0026291804804649314, "grad_norm": 0.172938, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:04.305634+00:00", "epoch": 0, "step": 21340, "train_loss": 3.526872396469116, "perplexity": 34.017408097978056, "lr": 0.0026291804804649314, "grad_norm": 0.16467, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:04.610079+00:00", "epoch": 0, "step": 21341, "train_loss": 3.4355547428131104, "perplexity": 31.048631794309582, "lr": 0.0026291804804649314, "grad_norm": 0.162353, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:04.915118+00:00", "epoch": 0, "step": 21342, "train_loss": 3.518634557723999, "perplexity": 33.738329257019835, "lr": 0.0026291804804649314, "grad_norm": 0.154537, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:05.221009+00:00", "epoch": 0, "step": 21343, "train_loss": 3.4985718727111816, "perplexity": 33.06819263226772, "lr": 0.0026291804804649314, "grad_norm": 0.169491, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:05.526600+00:00", "epoch": 0, "step": 21344, "train_loss": 3.53836727142334, "perplexity": 34.41068998295366, "lr": 0.0026291804804649314, "grad_norm": 0.142369, "tokens_per_sec": 107228} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:05.832722+00:00", "epoch": 0, "step": 21345, "train_loss": 3.458037853240967, "perplexity": 31.75460814527409, "lr": 0.0026291804804649314, "grad_norm": 0.155096, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:06.137100+00:00", "epoch": 0, "step": 21346, "train_loss": 3.531505584716797, "perplexity": 34.1753828341436, "lr": 0.0026291804804649314, "grad_norm": 0.159205, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:06.441129+00:00", "epoch": 0, "step": 21347, "train_loss": 3.5564913749694824, "perplexity": 35.04003885592281, "lr": 0.0026291804804649314, "grad_norm": 0.177688, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:06.745175+00:00", "epoch": 0, "step": 21348, "train_loss": 3.531108856201172, "perplexity": 34.16182717437708, "lr": 0.0026291804804649314, "grad_norm": 0.187172, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:07.049482+00:00", "epoch": 0, "step": 21349, "train_loss": 3.5080790519714355, "perplexity": 33.384077073135, "lr": 0.0026291804804649314, "grad_norm": 0.152481, "tokens_per_sec": 107682} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:07.354187+00:00", "epoch": 0, "step": 21350, "train_loss": 3.5496413707733154, "perplexity": 34.800834652961576, "lr": 0.0026291804804649314, "grad_norm": 0.195564, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:07.658250+00:00", "epoch": 0, "step": 21351, "train_loss": 3.4973087310791016, "perplexity": 33.02644919094578, "lr": 0.0026291804804649314, "grad_norm": 0.174368, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:07.963033+00:00", "epoch": 0, "step": 21352, "train_loss": 3.4508543014526367, "perplexity": 31.527314637910983, "lr": 0.0026291804804649314, "grad_norm": 0.155893, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:08.267523+00:00", "epoch": 0, "step": 21353, "train_loss": 3.503495454788208, "perplexity": 33.23140806540462, "lr": 0.0026291804804649314, "grad_norm": 0.181613, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:08.571620+00:00", "epoch": 0, "step": 21354, "train_loss": 3.4439680576324463, "perplexity": 31.310955667613783, "lr": 0.0026291804804649314, "grad_norm": 0.170194, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:08.876512+00:00", "epoch": 0, "step": 21355, "train_loss": 3.527421236038208, "perplexity": 34.03608332196232, "lr": 0.0026291804804649314, "grad_norm": 0.151767, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:09.181774+00:00", "epoch": 0, "step": 21356, "train_loss": 3.4450621604919434, "perplexity": 31.345231821142878, "lr": 0.0026291804804649314, "grad_norm": 0.170132, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:09.486659+00:00", "epoch": 0, "step": 21357, "train_loss": 3.608959436416626, "perplexity": 36.92760729241658, "lr": 0.0026291804804649314, "grad_norm": 0.15756, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:09.790550+00:00", "epoch": 0, "step": 21358, "train_loss": 3.56105899810791, "perplexity": 35.2004546286717, "lr": 0.0026291804804649314, "grad_norm": 0.192732, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:10.095297+00:00", "epoch": 0, "step": 21359, "train_loss": 3.525111675262451, "perplexity": 33.95756562456017, "lr": 0.0026291804804649314, "grad_norm": 0.16087, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:10.401599+00:00", "epoch": 0, "step": 21360, "train_loss": 3.4437711238861084, "perplexity": 31.304790090937722, "lr": 0.0026291804804649314, "grad_norm": 0.166467, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:10.706760+00:00", "epoch": 0, "step": 21361, "train_loss": 3.4936130046844482, "perplexity": 32.904617737476734, "lr": 0.0026291804804649314, "grad_norm": 0.181349, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:11.011666+00:00", "epoch": 0, "step": 21362, "train_loss": 3.500885248184204, "perplexity": 33.14478033193713, "lr": 0.0026291804804649314, "grad_norm": 0.159751, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:11.316715+00:00", "epoch": 0, "step": 21363, "train_loss": 3.510727882385254, "perplexity": 33.47262305169198, "lr": 0.0026291804804649314, "grad_norm": 0.183672, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:11.622227+00:00", "epoch": 0, "step": 21364, "train_loss": 3.505845546722412, "perplexity": 33.30959676879606, "lr": 0.0026291804804649314, "grad_norm": 0.214226, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:11.927724+00:00", "epoch": 0, "step": 21365, "train_loss": 3.3914811611175537, "perplexity": 29.709924882609606, "lr": 0.0026291804804649314, "grad_norm": 0.186752, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:12.232484+00:00", "epoch": 0, "step": 21366, "train_loss": 3.5226194858551025, "perplexity": 33.87304230701753, "lr": 0.0026291804804649314, "grad_norm": 0.155541, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:12.537996+00:00", "epoch": 0, "step": 21367, "train_loss": 3.555788278579712, "perplexity": 35.01541099000215, "lr": 0.0026291804804649314, "grad_norm": 0.179575, "tokens_per_sec": 107255} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:12.843105+00:00", "epoch": 0, "step": 21368, "train_loss": 3.4480574131011963, "perplexity": 31.439259456505173, "lr": 0.0026291804804649314, "grad_norm": 0.17612, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:13.149117+00:00", "epoch": 0, "step": 21369, "train_loss": 3.55609393119812, "perplexity": 35.026115177854074, "lr": 0.0026291804804649314, "grad_norm": 0.180855, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:13.454777+00:00", "epoch": 0, "step": 21370, "train_loss": 3.5178961753845215, "perplexity": 33.7134266654813, "lr": 0.0026291804804649314, "grad_norm": 0.169623, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:13.759934+00:00", "epoch": 0, "step": 21371, "train_loss": 3.499997138977051, "perplexity": 33.11535721475982, "lr": 0.0026291804804649314, "grad_norm": 0.171134, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:14.066017+00:00", "epoch": 0, "step": 21372, "train_loss": 3.384800672531128, "perplexity": 29.512109555734416, "lr": 0.0026291804804649314, "grad_norm": 0.155879, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:14.370743+00:00", "epoch": 0, "step": 21373, "train_loss": 3.5329179763793945, "perplexity": 34.22368596336333, "lr": 0.0026291804804649314, "grad_norm": 0.160581, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:14.674982+00:00", "epoch": 0, "step": 21374, "train_loss": 3.406428337097168, "perplexity": 30.15733982307913, "lr": 0.0026291804804649314, "grad_norm": 0.175305, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:14.980825+00:00", "epoch": 0, "step": 21375, "train_loss": 3.5930936336517334, "perplexity": 36.34634445240754, "lr": 0.0026291804804649314, "grad_norm": 0.172419, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:15.285037+00:00", "epoch": 0, "step": 21376, "train_loss": 3.419809341430664, "perplexity": 30.563587255691218, "lr": 0.0026291804804649314, "grad_norm": 0.180778, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:15.590796+00:00", "epoch": 0, "step": 21377, "train_loss": 3.5399632453918457, "perplexity": 34.46565239602866, "lr": 0.0026291804804649314, "grad_norm": 0.175186, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:15.895158+00:00", "epoch": 0, "step": 21378, "train_loss": 3.4954833984375, "perplexity": 32.96621992115028, "lr": 0.0026291804804649314, "grad_norm": 0.163437, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:16.200013+00:00", "epoch": 0, "step": 21379, "train_loss": 3.5225682258605957, "perplexity": 33.871306019556364, "lr": 0.0026291804804649314, "grad_norm": 0.180352, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:16.503898+00:00", "epoch": 0, "step": 21380, "train_loss": 3.462982654571533, "perplexity": 31.912017231477115, "lr": 0.0026291804804649314, "grad_norm": 0.182818, "tokens_per_sec": 107887} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:16.808363+00:00", "epoch": 0, "step": 21381, "train_loss": 3.4859936237335205, "perplexity": 32.65485763688331, "lr": 0.0026291804804649314, "grad_norm": 0.184867, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:17.112852+00:00", "epoch": 0, "step": 21382, "train_loss": 3.4871411323547363, "perplexity": 32.6923508753392, "lr": 0.0026291804804649314, "grad_norm": 0.200217, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:17.419144+00:00", "epoch": 0, "step": 21383, "train_loss": 3.454775333404541, "perplexity": 31.651176921256024, "lr": 0.0026291804804649314, "grad_norm": 0.176242, "tokens_per_sec": 106982} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:17.724064+00:00", "epoch": 0, "step": 21384, "train_loss": 3.43200945854187, "perplexity": 30.938750463734067, "lr": 0.0026291804804649314, "grad_norm": 0.190507, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:18.027545+00:00", "epoch": 0, "step": 21385, "train_loss": 3.467820882797241, "perplexity": 32.066788962550866, "lr": 0.0026291804804649314, "grad_norm": 0.177288, "tokens_per_sec": 107981} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:18.331001+00:00", "epoch": 0, "step": 21386, "train_loss": 3.490079402923584, "perplexity": 32.78855110968979, "lr": 0.0026291804804649314, "grad_norm": 0.172937, "tokens_per_sec": 107976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:18.635653+00:00", "epoch": 0, "step": 21387, "train_loss": 3.464125394821167, "perplexity": 31.948505222191674, "lr": 0.0026291804804649314, "grad_norm": 0.179101, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:18.940662+00:00", "epoch": 0, "step": 21388, "train_loss": 3.4216954708099365, "perplexity": 30.621288534482762, "lr": 0.0026291804804649314, "grad_norm": 0.159017, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:19.244989+00:00", "epoch": 0, "step": 21389, "train_loss": 3.4829516410827637, "perplexity": 32.55567306185221, "lr": 0.0026291804804649314, "grad_norm": 0.183555, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:19.550160+00:00", "epoch": 0, "step": 21390, "train_loss": 3.472175359725952, "perplexity": 32.206727513742464, "lr": 0.0026291804804649314, "grad_norm": 0.167051, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:19.854747+00:00", "epoch": 0, "step": 21391, "train_loss": 3.4219279289245605, "perplexity": 30.628407528884797, "lr": 0.0026291804804649314, "grad_norm": 0.184249, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:20.159894+00:00", "epoch": 0, "step": 21392, "train_loss": 3.392047882080078, "perplexity": 29.726766891744653, "lr": 0.0026291804804649314, "grad_norm": 0.171362, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:20.465855+00:00", "epoch": 0, "step": 21393, "train_loss": 3.3346543312072754, "perplexity": 28.068679080276837, "lr": 0.0026291804804649314, "grad_norm": 0.180163, "tokens_per_sec": 107099} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:20.770444+00:00", "epoch": 0, "step": 21394, "train_loss": 3.4916882514953613, "perplexity": 32.8413453809195, "lr": 0.0026291804804649314, "grad_norm": 0.16497, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:21.075745+00:00", "epoch": 0, "step": 21395, "train_loss": 3.544412612915039, "perplexity": 34.6193444130481, "lr": 0.0026291804804649314, "grad_norm": 0.154194, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:21.381756+00:00", "epoch": 0, "step": 21396, "train_loss": 3.506337881088257, "perplexity": 33.326000265669045, "lr": 0.0026291804804649314, "grad_norm": 0.166659, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:21.688218+00:00", "epoch": 0, "step": 21397, "train_loss": 3.5193066596984863, "perplexity": 33.76101247659384, "lr": 0.0026291804804649314, "grad_norm": 0.165748, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:21.995044+00:00", "epoch": 0, "step": 21398, "train_loss": 3.4379403591156006, "perplexity": 31.122790338233074, "lr": 0.0026291804804649314, "grad_norm": 0.16326, "tokens_per_sec": 106797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:22.301977+00:00", "epoch": 0, "step": 21399, "train_loss": 3.4331963062286377, "perplexity": 30.97549184705348, "lr": 0.0026291804804649314, "grad_norm": 0.161601, "tokens_per_sec": 106761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:22.607876+00:00", "epoch": 0, "step": 21400, "train_loss": 3.4561755657196045, "perplexity": 31.695526964926877, "lr": 0.0026291804804649314, "grad_norm": 0.169773, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:22.913002+00:00", "epoch": 0, "step": 21401, "train_loss": 3.4884088039398193, "perplexity": 32.73382031885762, "lr": 0.0026291804804649314, "grad_norm": 0.18604, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:23.219221+00:00", "epoch": 0, "step": 21402, "train_loss": 3.5345561504364014, "perplexity": 34.2797962645153, "lr": 0.0026291804804649314, "grad_norm": 0.209171, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:23.524832+00:00", "epoch": 0, "step": 21403, "train_loss": 3.391660213470459, "perplexity": 29.71524499083921, "lr": 0.0026291804804649314, "grad_norm": 0.208176, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:23.831091+00:00", "epoch": 0, "step": 21404, "train_loss": 3.5351247787475586, "perplexity": 34.29929427020566, "lr": 0.0026291804804649314, "grad_norm": 0.189585, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:24.136048+00:00", "epoch": 0, "step": 21405, "train_loss": 3.6054530143737793, "perplexity": 36.79835026349908, "lr": 0.0026291804804649314, "grad_norm": 0.222577, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:24.440727+00:00", "epoch": 0, "step": 21406, "train_loss": 3.5820693969726562, "perplexity": 35.9478543093971, "lr": 0.0026291804804649314, "grad_norm": 0.182496, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:24.746090+00:00", "epoch": 0, "step": 21407, "train_loss": 3.569862127304077, "perplexity": 35.511696720730484, "lr": 0.0026291804804649314, "grad_norm": 0.167032, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:25.051943+00:00", "epoch": 0, "step": 21408, "train_loss": 3.426342487335205, "perplexity": 30.763917310798323, "lr": 0.0026291804804649314, "grad_norm": 0.152947, "tokens_per_sec": 107135} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:25.357837+00:00", "epoch": 0, "step": 21409, "train_loss": 3.462681293487549, "perplexity": 31.90240164032737, "lr": 0.0026291804804649314, "grad_norm": 0.172741, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:25.663675+00:00", "epoch": 0, "step": 21410, "train_loss": 3.528041362762451, "perplexity": 34.05719655258151, "lr": 0.0026291804804649314, "grad_norm": 0.18475, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:25.968128+00:00", "epoch": 0, "step": 21411, "train_loss": 3.4179985523223877, "perplexity": 30.50829312289439, "lr": 0.0026291804804649314, "grad_norm": 0.182123, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:26.272903+00:00", "epoch": 0, "step": 21412, "train_loss": 3.5077412128448486, "perplexity": 33.37280053063059, "lr": 0.0026291804804649314, "grad_norm": 0.17739, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:26.579301+00:00", "epoch": 0, "step": 21413, "train_loss": 3.493859052658081, "perplexity": 32.912714848092186, "lr": 0.0026291804804649314, "grad_norm": 0.1511, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:26.885616+00:00", "epoch": 0, "step": 21414, "train_loss": 3.402920722961426, "perplexity": 30.05174481315845, "lr": 0.0026291804804649314, "grad_norm": 0.161857, "tokens_per_sec": 106975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:27.190145+00:00", "epoch": 0, "step": 21415, "train_loss": 3.4703023433685303, "perplexity": 32.14646024469965, "lr": 0.0026291804804649314, "grad_norm": 0.162094, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:27.495139+00:00", "epoch": 0, "step": 21416, "train_loss": 3.49076509475708, "perplexity": 32.811041661330464, "lr": 0.0026291804804649314, "grad_norm": 0.170113, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:27.799768+00:00", "epoch": 0, "step": 21417, "train_loss": 3.461435079574585, "perplexity": 31.862669186256966, "lr": 0.0026291804804649314, "grad_norm": 0.166046, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:28.105928+00:00", "epoch": 0, "step": 21418, "train_loss": 3.3943099975585938, "perplexity": 29.794088387047037, "lr": 0.0026291804804649314, "grad_norm": 0.156497, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:28.411317+00:00", "epoch": 0, "step": 21419, "train_loss": 3.6051766872406006, "perplexity": 36.78818328563568, "lr": 0.0026291804804649314, "grad_norm": 0.153686, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:28.715810+00:00", "epoch": 0, "step": 21420, "train_loss": 3.5658323764801025, "perplexity": 35.368881379991414, "lr": 0.0026291804804649314, "grad_norm": 0.180194, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:29.020960+00:00", "epoch": 0, "step": 21421, "train_loss": 3.37552547454834, "perplexity": 29.23964443401803, "lr": 0.0026291804804649314, "grad_norm": 0.19502, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:29.326447+00:00", "epoch": 0, "step": 21422, "train_loss": 3.543567180633545, "perplexity": 34.59008847041887, "lr": 0.0026291804804649314, "grad_norm": 0.165208, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:29.633405+00:00", "epoch": 0, "step": 21423, "train_loss": 3.5448970794677734, "perplexity": 34.636120390865756, "lr": 0.0026291804804649314, "grad_norm": 0.153905, "tokens_per_sec": 106749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:29.938290+00:00", "epoch": 0, "step": 21424, "train_loss": 3.45587158203125, "perplexity": 31.685893506017106, "lr": 0.0026291804804649314, "grad_norm": 0.165135, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:30.242448+00:00", "epoch": 0, "step": 21425, "train_loss": 3.4966490268707275, "perplexity": 33.00466868856133, "lr": 0.0026291804804649314, "grad_norm": 0.163629, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:30.546711+00:00", "epoch": 0, "step": 21426, "train_loss": 3.4133832454681396, "perplexity": 30.36781241867696, "lr": 0.0026291804804649314, "grad_norm": 0.160512, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:30.852340+00:00", "epoch": 0, "step": 21427, "train_loss": 3.6277880668640137, "perplexity": 37.629490578241885, "lr": 0.0026291804804649314, "grad_norm": 0.176933, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:31.158024+00:00", "epoch": 0, "step": 21428, "train_loss": 3.467346668243408, "perplexity": 32.051586029541085, "lr": 0.0026291804804649314, "grad_norm": 0.167534, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:31.463127+00:00", "epoch": 0, "step": 21429, "train_loss": 3.520956516265869, "perplexity": 33.816759279218694, "lr": 0.0026291804804649314, "grad_norm": 0.160102, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:31.768494+00:00", "epoch": 0, "step": 21430, "train_loss": 3.5065102577209473, "perplexity": 33.33174538452378, "lr": 0.0026291804804649314, "grad_norm": 0.159886, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:32.074378+00:00", "epoch": 0, "step": 21431, "train_loss": 3.432461738586426, "perplexity": 30.9527466080221, "lr": 0.0026291804804649314, "grad_norm": 0.158466, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:32.380432+00:00", "epoch": 0, "step": 21432, "train_loss": 3.4148850440979004, "perplexity": 30.413453020674783, "lr": 0.0026291804804649314, "grad_norm": 0.14367, "tokens_per_sec": 107066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:32.686579+00:00", "epoch": 0, "step": 21433, "train_loss": 3.428300380706787, "perplexity": 30.824208783462538, "lr": 0.0026291804804649314, "grad_norm": 0.173824, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:32.991945+00:00", "epoch": 0, "step": 21434, "train_loss": 3.502899408340454, "perplexity": 33.211606504584616, "lr": 0.0026291804804649314, "grad_norm": 0.166176, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:33.296075+00:00", "epoch": 0, "step": 21435, "train_loss": 3.440901279449463, "perplexity": 31.215079003259582, "lr": 0.0026291804804649314, "grad_norm": 0.16821, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:33.601786+00:00", "epoch": 0, "step": 21436, "train_loss": 3.5636236667633057, "perplexity": 35.29084799628476, "lr": 0.0026291804804649314, "grad_norm": 0.175377, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:33.907209+00:00", "epoch": 0, "step": 21437, "train_loss": 3.5530176162719727, "perplexity": 34.918529385663405, "lr": 0.0026291804804649314, "grad_norm": 0.24017, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:34.213311+00:00", "epoch": 0, "step": 21438, "train_loss": 3.429487705230713, "perplexity": 30.860828858136053, "lr": 0.0026291804804649314, "grad_norm": 0.196288, "tokens_per_sec": 107051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:34.518490+00:00", "epoch": 0, "step": 21439, "train_loss": 3.5351831912994385, "perplexity": 34.30129783802784, "lr": 0.0026291804804649314, "grad_norm": 0.178021, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:34.823308+00:00", "epoch": 0, "step": 21440, "train_loss": 3.5083563327789307, "perplexity": 33.39333512046342, "lr": 0.0026291804804649314, "grad_norm": 0.210876, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:35.128674+00:00", "epoch": 0, "step": 21441, "train_loss": 3.4017069339752197, "perplexity": 30.015290464706922, "lr": 0.0026291804804649314, "grad_norm": 0.181942, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:35.434795+00:00", "epoch": 0, "step": 21442, "train_loss": 3.4097650051116943, "perplexity": 30.258132917303048, "lr": 0.0026291804804649314, "grad_norm": 0.170696, "tokens_per_sec": 107043} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:35.739687+00:00", "epoch": 0, "step": 21443, "train_loss": 3.517122268676758, "perplexity": 33.68734571185739, "lr": 0.0026291804804649314, "grad_norm": 0.196411, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:36.043994+00:00", "epoch": 0, "step": 21444, "train_loss": 3.5239815711975098, "perplexity": 33.919211717647315, "lr": 0.0026291804804649314, "grad_norm": 0.196251, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:36.348776+00:00", "epoch": 0, "step": 21445, "train_loss": 3.5262086391448975, "perplexity": 33.99483628613351, "lr": 0.0026291804804649314, "grad_norm": 0.16814, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:36.653967+00:00", "epoch": 0, "step": 21446, "train_loss": 3.441307783126831, "perplexity": 31.227770627084855, "lr": 0.0026291804804649314, "grad_norm": 0.168638, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:36.959507+00:00", "epoch": 0, "step": 21447, "train_loss": 3.4745912551879883, "perplexity": 32.28462966442612, "lr": 0.0026291804804649314, "grad_norm": 0.180382, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:37.265189+00:00", "epoch": 0, "step": 21448, "train_loss": 3.5598554611206055, "perplexity": 35.15811506328338, "lr": 0.0026291804804649314, "grad_norm": 0.21823, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:37.570298+00:00", "epoch": 0, "step": 21449, "train_loss": 3.477905035018921, "perplexity": 32.39179127598774, "lr": 0.0026291804804649314, "grad_norm": 0.180562, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:37.874669+00:00", "epoch": 0, "step": 21450, "train_loss": 3.541297197341919, "perplexity": 34.51165859846255, "lr": 0.0026291804804649314, "grad_norm": 0.20104, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:38.179309+00:00", "epoch": 0, "step": 21451, "train_loss": 3.595503568649292, "perplexity": 36.434042410681904, "lr": 0.0026291804804649314, "grad_norm": 0.179519, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:38.485218+00:00", "epoch": 0, "step": 21452, "train_loss": 3.4777002334594727, "perplexity": 32.38515806589017, "lr": 0.0026291804804649314, "grad_norm": 0.186858, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:38.790508+00:00", "epoch": 0, "step": 21453, "train_loss": 3.4304921627044678, "perplexity": 30.891842821825097, "lr": 0.0026291804804649314, "grad_norm": 0.181707, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:39.096493+00:00", "epoch": 0, "step": 21454, "train_loss": 3.5528039932250977, "perplexity": 34.91107077971707, "lr": 0.0026291804804649314, "grad_norm": 0.174144, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:39.400874+00:00", "epoch": 0, "step": 21455, "train_loss": 3.534599781036377, "perplexity": 34.281291945221845, "lr": 0.0026291804804649314, "grad_norm": 0.173429, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:39.704636+00:00", "epoch": 0, "step": 21456, "train_loss": 3.4110770225524902, "perplexity": 30.297858169829993, "lr": 0.0026291804804649314, "grad_norm": 0.170929, "tokens_per_sec": 107875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:40.010125+00:00", "epoch": 0, "step": 21457, "train_loss": 3.578554153442383, "perplexity": 35.82171068977587, "lr": 0.0026291804804649314, "grad_norm": 0.183738, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:40.315127+00:00", "epoch": 0, "step": 21458, "train_loss": 3.5371904373168945, "perplexity": 34.370218128355745, "lr": 0.0026291804804649314, "grad_norm": 0.174082, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:40.620888+00:00", "epoch": 0, "step": 21459, "train_loss": 3.5400707721710205, "perplexity": 34.46935857587618, "lr": 0.0026291804804649314, "grad_norm": 0.170033, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:40.925781+00:00", "epoch": 0, "step": 21460, "train_loss": 3.4265387058258057, "perplexity": 30.769954352488245, "lr": 0.0026291804804649314, "grad_norm": 0.181224, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:41.230362+00:00", "epoch": 0, "step": 21461, "train_loss": 3.4805963039398193, "perplexity": 32.47908370815817, "lr": 0.0026291804804649314, "grad_norm": 0.148538, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:41.534793+00:00", "epoch": 0, "step": 21462, "train_loss": 3.417492151260376, "perplexity": 30.492847602000957, "lr": 0.0026291804804649314, "grad_norm": 0.168188, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:41.839892+00:00", "epoch": 0, "step": 21463, "train_loss": 3.471872329711914, "perplexity": 32.19696938722913, "lr": 0.0026291804804649314, "grad_norm": 0.165152, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:42.145317+00:00", "epoch": 0, "step": 21464, "train_loss": 3.3675601482391357, "perplexity": 29.00766624244646, "lr": 0.0026291804804649314, "grad_norm": 0.171051, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:42.450681+00:00", "epoch": 0, "step": 21465, "train_loss": 3.452505111694336, "perplexity": 31.579403234173405, "lr": 0.0026291804804649314, "grad_norm": 0.1464, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:42.756071+00:00", "epoch": 0, "step": 21466, "train_loss": 3.4894826412200928, "perplexity": 32.76898999531704, "lr": 0.0026291804804649314, "grad_norm": 0.182062, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:43.060804+00:00", "epoch": 0, "step": 21467, "train_loss": 3.572343349456787, "perplexity": 35.59991853301058, "lr": 0.0026291804804649314, "grad_norm": 0.177694, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:43.366452+00:00", "epoch": 0, "step": 21468, "train_loss": 3.505284309387207, "perplexity": 33.29090742453825, "lr": 0.0026291804804649314, "grad_norm": 0.160209, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:43.672127+00:00", "epoch": 0, "step": 21469, "train_loss": 3.5094640254974365, "perplexity": 33.43034516871924, "lr": 0.0026291804804649314, "grad_norm": 0.147345, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:43.976701+00:00", "epoch": 0, "step": 21470, "train_loss": 3.4776225090026855, "perplexity": 32.382641044889844, "lr": 0.0026291804804649314, "grad_norm": 0.205081, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:44.280997+00:00", "epoch": 0, "step": 21471, "train_loss": 3.588228702545166, "perplexity": 36.169951408573176, "lr": 0.0026291804804649314, "grad_norm": 0.198704, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:44.585973+00:00", "epoch": 0, "step": 21472, "train_loss": 3.531283140182495, "perplexity": 34.1677815524876, "lr": 0.0026291804804649314, "grad_norm": 0.176106, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:44.889692+00:00", "epoch": 0, "step": 21473, "train_loss": 3.3923144340515137, "perplexity": 29.73469167620059, "lr": 0.0026291804804649314, "grad_norm": 0.16852, "tokens_per_sec": 107890} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:45.194780+00:00", "epoch": 0, "step": 21474, "train_loss": 3.5864551067352295, "perplexity": 36.10585738955251, "lr": 0.0026291804804649314, "grad_norm": 0.18207, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:45.499588+00:00", "epoch": 0, "step": 21475, "train_loss": 3.4480035305023193, "perplexity": 31.437565473137415, "lr": 0.0026291804804649314, "grad_norm": 0.160446, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:45.804007+00:00", "epoch": 0, "step": 21476, "train_loss": 3.450148820877075, "perplexity": 31.505080573606723, "lr": 0.0026291804804649314, "grad_norm": 0.196889, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:46.108529+00:00", "epoch": 0, "step": 21477, "train_loss": 3.445821523666382, "perplexity": 31.369043275508673, "lr": 0.0026291804804649314, "grad_norm": 0.163667, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:46.413295+00:00", "epoch": 0, "step": 21478, "train_loss": 3.4624557495117188, "perplexity": 31.895207057200825, "lr": 0.0026291804804649314, "grad_norm": 0.17733, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:46.718844+00:00", "epoch": 0, "step": 21479, "train_loss": 3.5730550289154053, "perplexity": 35.62526328135766, "lr": 0.0026291804804649314, "grad_norm": 0.176689, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:47.024771+00:00", "epoch": 0, "step": 21480, "train_loss": 3.556657075881958, "perplexity": 35.04584550340452, "lr": 0.0026291804804649314, "grad_norm": 0.165437, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:47.329730+00:00", "epoch": 0, "step": 21481, "train_loss": 3.4457201957702637, "perplexity": 31.365864877383242, "lr": 0.0026291804804649314, "grad_norm": 0.160579, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:47.634937+00:00", "epoch": 0, "step": 21482, "train_loss": 3.6192753314971924, "perplexity": 37.31052026412188, "lr": 0.0026291804804649314, "grad_norm": 0.163383, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:47.939336+00:00", "epoch": 0, "step": 21483, "train_loss": 3.4706003665924072, "perplexity": 32.156042064149446, "lr": 0.0026291804804649314, "grad_norm": 0.165503, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:48.244326+00:00", "epoch": 0, "step": 21484, "train_loss": 3.460702657699585, "perplexity": 31.839340814496815, "lr": 0.0026291804804649314, "grad_norm": 0.161868, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:48.549636+00:00", "epoch": 0, "step": 21485, "train_loss": 3.499216079711914, "perplexity": 33.089502256630624, "lr": 0.0026291804804649314, "grad_norm": 0.168341, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:48.855702+00:00", "epoch": 0, "step": 21486, "train_loss": 3.5236804485321045, "perplexity": 33.90899941186199, "lr": 0.0026291804804649314, "grad_norm": 0.15726, "tokens_per_sec": 107063} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:49.159668+00:00", "epoch": 0, "step": 21487, "train_loss": 3.4322564601898193, "perplexity": 30.946393329945515, "lr": 0.0026291804804649314, "grad_norm": 0.180502, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:49.464511+00:00", "epoch": 0, "step": 21488, "train_loss": 3.5904626846313477, "perplexity": 36.25084475557934, "lr": 0.0026291804804649314, "grad_norm": 0.207816, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:49.771161+00:00", "epoch": 0, "step": 21489, "train_loss": 3.4391419887542725, "perplexity": 31.160210883857353, "lr": 0.0026291804804649314, "grad_norm": 0.191238, "tokens_per_sec": 106858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:50.078868+00:00", "epoch": 0, "step": 21490, "train_loss": 3.469172954559326, "perplexity": 32.11017488623954, "lr": 0.0026291804804649314, "grad_norm": 0.156124, "tokens_per_sec": 106550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:50.384403+00:00", "epoch": 0, "step": 21491, "train_loss": 3.5611493587493896, "perplexity": 35.203635508043305, "lr": 0.0026291804804649314, "grad_norm": 0.165169, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:50.690093+00:00", "epoch": 0, "step": 21492, "train_loss": 3.4211254119873047, "perplexity": 30.6038375733031, "lr": 0.0026291804804649314, "grad_norm": 0.168059, "tokens_per_sec": 107186} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:50.994663+00:00", "epoch": 0, "step": 21493, "train_loss": 3.4999234676361084, "perplexity": 33.11291765185206, "lr": 0.0026291804804649314, "grad_norm": 0.165177, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:51.298896+00:00", "epoch": 0, "step": 21494, "train_loss": 3.407254934310913, "perplexity": 30.18227810168671, "lr": 0.0026291804804649314, "grad_norm": 0.161044, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:51.604005+00:00", "epoch": 0, "step": 21495, "train_loss": 3.563145637512207, "perplexity": 35.27398197019455, "lr": 0.0026291804804649314, "grad_norm": 0.153059, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:51.909416+00:00", "epoch": 0, "step": 21496, "train_loss": 3.466904640197754, "perplexity": 32.037421460407366, "lr": 0.0026291804804649314, "grad_norm": 0.172257, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:52.214552+00:00", "epoch": 0, "step": 21497, "train_loss": 3.5732593536376953, "perplexity": 35.63254314708708, "lr": 0.0026291804804649314, "grad_norm": 0.187153, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:52.519368+00:00", "epoch": 0, "step": 21498, "train_loss": 3.4211437702178955, "perplexity": 30.60439941076739, "lr": 0.0026291804804649314, "grad_norm": 0.201252, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:52.824239+00:00", "epoch": 0, "step": 21499, "train_loss": 3.5403847694396973, "perplexity": 34.48018355974079, "lr": 0.0026291804804649314, "grad_norm": 0.159035, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:53.129355+00:00", "epoch": 0, "step": 21500, "train_loss": 3.467787504196167, "perplexity": 32.06571863585746, "lr": 0.0026291804804649314, "grad_norm": 0.215065, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:50:56.203736+00:00", "step": 21500, "epoch": 0, "val_loss": 3.455549347400665, "val_ppl": 31.675684858706624, "eval_train_loss": 3.467787504196167, "eval_train_ppl": 32.06571863585746} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:50:57.180853+00:00", "step": 21500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4555_epoch_0000_step_0021500.pt", "val_loss": 3.455549347400665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:58.467169+00:00", "epoch": 0, "step": 21501, "train_loss": 3.576862335205078, "perplexity": 35.76115810276801, "lr": 0.0026291804804649314, "grad_norm": 0.224598, "tokens_per_sec": 6139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:58.772198+00:00", "epoch": 0, "step": 21502, "train_loss": 3.6253044605255127, "perplexity": 37.53614969591025, "lr": 0.0026291804804649314, "grad_norm": 0.158349, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:59.075464+00:00", "epoch": 0, "step": 21503, "train_loss": 3.4955356121063232, "perplexity": 32.96794125337774, "lr": 0.0026291804804649314, "grad_norm": 0.184025, "tokens_per_sec": 108051} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:59.378824+00:00", "epoch": 0, "step": 21504, "train_loss": 3.5194475650787354, "perplexity": 33.76576992006127, "lr": 0.0026291804804649314, "grad_norm": 0.196286, "tokens_per_sec": 108019} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:59.683724+00:00", "epoch": 0, "step": 21505, "train_loss": 3.500065326690674, "perplexity": 33.11761535224184, "lr": 0.0026291804804649314, "grad_norm": 0.19282, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:50:59.987921+00:00", "epoch": 0, "step": 21506, "train_loss": 3.5844814777374268, "perplexity": 36.03466809615297, "lr": 0.0026291804804649314, "grad_norm": 0.187425, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:00.293833+00:00", "epoch": 0, "step": 21507, "train_loss": 3.436225175857544, "perplexity": 31.06945480248536, "lr": 0.0026291804804649314, "grad_norm": 0.173299, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:00.598212+00:00", "epoch": 0, "step": 21508, "train_loss": 3.4869942665100098, "perplexity": 32.68754983817462, "lr": 0.0026291804804649314, "grad_norm": 0.156909, "tokens_per_sec": 107655} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:00.901533+00:00", "epoch": 0, "step": 21509, "train_loss": 3.306586742401123, "perplexity": 27.291812325786616, "lr": 0.0026291804804649314, "grad_norm": 0.161692, "tokens_per_sec": 108031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:01.205551+00:00", "epoch": 0, "step": 21510, "train_loss": 3.5340795516967773, "perplexity": 34.26346244946808, "lr": 0.0026291804804649314, "grad_norm": 0.162618, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:01.509971+00:00", "epoch": 0, "step": 21511, "train_loss": 3.4779865741729736, "perplexity": 32.39443258293014, "lr": 0.0026291804804649314, "grad_norm": 0.186712, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:01.815867+00:00", "epoch": 0, "step": 21512, "train_loss": 3.443755626678467, "perplexity": 31.30430495786462, "lr": 0.0026291804804649314, "grad_norm": 0.203026, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:02.138700+00:00", "epoch": 0, "step": 21513, "train_loss": 3.4997854232788086, "perplexity": 33.1083469159059, "lr": 0.0026291804804649314, "grad_norm": 0.200487, "tokens_per_sec": 101502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:02.442672+00:00", "epoch": 0, "step": 21514, "train_loss": 3.5489253997802734, "perplexity": 34.775927182393616, "lr": 0.0026291804804649314, "grad_norm": 0.148166, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:02.747651+00:00", "epoch": 0, "step": 21515, "train_loss": 3.327188491821289, "perplexity": 27.859903144032238, "lr": 0.0026291804804649314, "grad_norm": 0.235253, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:03.051095+00:00", "epoch": 0, "step": 21516, "train_loss": 3.3939762115478516, "perplexity": 29.784145196681862, "lr": 0.0026291804804649314, "grad_norm": 0.21591, "tokens_per_sec": 107978} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:03.355375+00:00", "epoch": 0, "step": 21517, "train_loss": 3.5132663249969482, "perplexity": 33.55769931929929, "lr": 0.0026291804804649314, "grad_norm": 0.187079, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:03.661422+00:00", "epoch": 0, "step": 21518, "train_loss": 3.468451738357544, "perplexity": 32.08702485696108, "lr": 0.0026291804804649314, "grad_norm": 0.202636, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:03.965606+00:00", "epoch": 0, "step": 21519, "train_loss": 3.5280041694641113, "perplexity": 34.05592987666558, "lr": 0.0026291804804649314, "grad_norm": 0.179229, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:04.270613+00:00", "epoch": 0, "step": 21520, "train_loss": 3.447394371032715, "perplexity": 31.418420814095324, "lr": 0.0026291804804649314, "grad_norm": 0.174058, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:04.575475+00:00", "epoch": 0, "step": 21521, "train_loss": 3.566030502319336, "perplexity": 35.37588956352577, "lr": 0.0026291804804649314, "grad_norm": 0.183293, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:04.881113+00:00", "epoch": 0, "step": 21522, "train_loss": 3.4036030769348145, "perplexity": 30.072257738381072, "lr": 0.0026291804804649314, "grad_norm": 0.151557, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:05.185625+00:00", "epoch": 0, "step": 21523, "train_loss": 3.4556972980499268, "perplexity": 31.680371643545158, "lr": 0.0026291804804649314, "grad_norm": 0.163911, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:05.490140+00:00", "epoch": 0, "step": 21524, "train_loss": 3.4950287342071533, "perplexity": 32.95123476700732, "lr": 0.0026291804804649314, "grad_norm": 0.155209, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:05.793853+00:00", "epoch": 0, "step": 21525, "train_loss": 3.5315487384796143, "perplexity": 34.17685766233058, "lr": 0.0026291804804649314, "grad_norm": 0.153843, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:06.097757+00:00", "epoch": 0, "step": 21526, "train_loss": 3.5286998748779297, "perplexity": 34.07963101500256, "lr": 0.0026291804804649314, "grad_norm": 0.185643, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:06.403101+00:00", "epoch": 0, "step": 21527, "train_loss": 3.40274715423584, "perplexity": 30.04652922275438, "lr": 0.0026291804804649314, "grad_norm": 0.164101, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:06.707537+00:00", "epoch": 0, "step": 21528, "train_loss": 3.4460513591766357, "perplexity": 31.3762538241644, "lr": 0.0026291804804649314, "grad_norm": 0.147399, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:07.011960+00:00", "epoch": 0, "step": 21529, "train_loss": 3.506021738052368, "perplexity": 33.315466148002734, "lr": 0.0026291804804649314, "grad_norm": 0.162667, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:07.316643+00:00", "epoch": 0, "step": 21530, "train_loss": 3.365103244781494, "perplexity": 28.93648468587006, "lr": 0.0026291804804649314, "grad_norm": 0.170205, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:07.620287+00:00", "epoch": 0, "step": 21531, "train_loss": 3.4620373249053955, "perplexity": 31.881864109447818, "lr": 0.0026291804804649314, "grad_norm": 0.166306, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:07.924794+00:00", "epoch": 0, "step": 21532, "train_loss": 3.444241523742676, "perplexity": 31.31951932374927, "lr": 0.0026291804804649314, "grad_norm": 0.163916, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:08.229100+00:00", "epoch": 0, "step": 21533, "train_loss": 3.492459535598755, "perplexity": 32.86668515937384, "lr": 0.0026291804804649314, "grad_norm": 0.147195, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:08.533801+00:00", "epoch": 0, "step": 21534, "train_loss": 3.5119948387145996, "perplexity": 33.51505827944097, "lr": 0.0026291804804649314, "grad_norm": 0.170936, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:08.837519+00:00", "epoch": 0, "step": 21535, "train_loss": 3.6193466186523438, "perplexity": 37.31318011977437, "lr": 0.0026291804804649314, "grad_norm": 0.17955, "tokens_per_sec": 107889} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:09.141801+00:00", "epoch": 0, "step": 21536, "train_loss": 3.6221399307250977, "perplexity": 37.41755318167769, "lr": 0.0026291804804649314, "grad_norm": 0.157663, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:09.445625+00:00", "epoch": 0, "step": 21537, "train_loss": 3.4684362411499023, "perplexity": 32.086527601527315, "lr": 0.0026291804804649314, "grad_norm": 0.151226, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:09.750731+00:00", "epoch": 0, "step": 21538, "train_loss": 3.52980899810791, "perplexity": 34.11745049478527, "lr": 0.0026291804804649314, "grad_norm": 0.180763, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:10.055258+00:00", "epoch": 0, "step": 21539, "train_loss": 3.529604434967041, "perplexity": 34.110472035745694, "lr": 0.0026291804804649314, "grad_norm": 0.151324, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:10.359371+00:00", "epoch": 0, "step": 21540, "train_loss": 3.609856367111206, "perplexity": 36.9607436551635, "lr": 0.0026291804804649314, "grad_norm": 0.186357, "tokens_per_sec": 107808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:10.663472+00:00", "epoch": 0, "step": 21541, "train_loss": 3.4455432891845703, "perplexity": 31.36031654010364, "lr": 0.0026291804804649314, "grad_norm": 0.161627, "tokens_per_sec": 107695} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:10.967744+00:00", "epoch": 0, "step": 21542, "train_loss": 3.564338207244873, "perplexity": 35.31607374714375, "lr": 0.0026291804804649314, "grad_norm": 0.191661, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:11.272943+00:00", "epoch": 0, "step": 21543, "train_loss": 3.519949197769165, "perplexity": 33.78271218311192, "lr": 0.0026291804804649314, "grad_norm": 0.174879, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:11.578693+00:00", "epoch": 0, "step": 21544, "train_loss": 3.5500404834747314, "perplexity": 34.81472688018921, "lr": 0.0026291804804649314, "grad_norm": 0.153833, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:11.882660+00:00", "epoch": 0, "step": 21545, "train_loss": 3.552557945251465, "perplexity": 34.90248203815893, "lr": 0.0026291804804649314, "grad_norm": 0.183631, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:12.186340+00:00", "epoch": 0, "step": 21546, "train_loss": 3.498039722442627, "perplexity": 33.05060006603742, "lr": 0.0026291804804649314, "grad_norm": 0.184686, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:12.490119+00:00", "epoch": 0, "step": 21547, "train_loss": 3.4660942554473877, "perplexity": 32.011469339636534, "lr": 0.0026291804804649314, "grad_norm": 0.201263, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:12.795236+00:00", "epoch": 0, "step": 21548, "train_loss": 3.4669318199157715, "perplexity": 32.03829224032239, "lr": 0.0026291804804649314, "grad_norm": 0.181482, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:13.100951+00:00", "epoch": 0, "step": 21549, "train_loss": 3.442721128463745, "perplexity": 31.27193745522106, "lr": 0.0026291804804649314, "grad_norm": 0.158679, "tokens_per_sec": 107185} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:13.406218+00:00", "epoch": 0, "step": 21550, "train_loss": 3.3786089420318604, "perplexity": 29.329943071771794, "lr": 0.0026291804804649314, "grad_norm": 0.179088, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:13.710406+00:00", "epoch": 0, "step": 21551, "train_loss": 3.4358325004577637, "perplexity": 31.05725698694762, "lr": 0.0026291804804649314, "grad_norm": 0.170138, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:14.013791+00:00", "epoch": 0, "step": 21552, "train_loss": 3.5732083320617676, "perplexity": 35.63072516495995, "lr": 0.0026291804804649314, "grad_norm": 0.155881, "tokens_per_sec": 108009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:14.319906+00:00", "epoch": 0, "step": 21553, "train_loss": 3.5237202644348145, "perplexity": 33.91034955616199, "lr": 0.0026291804804649314, "grad_norm": 0.165252, "tokens_per_sec": 107044} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:14.625592+00:00", "epoch": 0, "step": 21554, "train_loss": 3.6466102600097656, "perplexity": 38.3444677288097, "lr": 0.0026291804804649314, "grad_norm": 0.16041, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:14.931582+00:00", "epoch": 0, "step": 21555, "train_loss": 3.49701189994812, "perplexity": 33.016647367495274, "lr": 0.0026291804804649314, "grad_norm": 0.19776, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:15.235673+00:00", "epoch": 0, "step": 21556, "train_loss": 3.527527093887329, "perplexity": 34.039686499244226, "lr": 0.0026291804804649314, "grad_norm": 0.170107, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:15.540007+00:00", "epoch": 0, "step": 21557, "train_loss": 3.627660036087036, "perplexity": 37.624673153721766, "lr": 0.0026291804804649314, "grad_norm": 0.176501, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:15.845307+00:00", "epoch": 0, "step": 21558, "train_loss": 3.4844508171081543, "perplexity": 32.604516349591925, "lr": 0.0026291804804649314, "grad_norm": 0.174926, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:16.150197+00:00", "epoch": 0, "step": 21559, "train_loss": 3.4272208213806152, "perplexity": 30.790950176948, "lr": 0.0026291804804649314, "grad_norm": 0.192243, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:16.455045+00:00", "epoch": 0, "step": 21560, "train_loss": 3.5383338928222656, "perplexity": 34.40954142142883, "lr": 0.0026291804804649314, "grad_norm": 0.168799, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:16.759699+00:00", "epoch": 0, "step": 21561, "train_loss": 3.5708653926849365, "perplexity": 35.54734225464116, "lr": 0.0026291804804649314, "grad_norm": 0.158285, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:17.064644+00:00", "epoch": 0, "step": 21562, "train_loss": 3.405766248703003, "perplexity": 30.13737960682629, "lr": 0.0026291804804649314, "grad_norm": 0.184034, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:17.369198+00:00", "epoch": 0, "step": 21563, "train_loss": 3.510272264480591, "perplexity": 33.457375799038005, "lr": 0.0026291804804649314, "grad_norm": 0.214056, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:17.674662+00:00", "epoch": 0, "step": 21564, "train_loss": 3.555177688598633, "perplexity": 34.99403745676496, "lr": 0.0026291804804649314, "grad_norm": 0.198764, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:17.980650+00:00", "epoch": 0, "step": 21565, "train_loss": 3.516378164291382, "perplexity": 33.66228813406446, "lr": 0.0026291804804649314, "grad_norm": 0.166842, "tokens_per_sec": 107087} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:18.285282+00:00", "epoch": 0, "step": 21566, "train_loss": 3.4681761264801025, "perplexity": 32.07818251038259, "lr": 0.0026291804804649314, "grad_norm": 0.185845, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:18.589109+00:00", "epoch": 0, "step": 21567, "train_loss": 3.483069896697998, "perplexity": 32.559523180644106, "lr": 0.0026291804804649314, "grad_norm": 0.173215, "tokens_per_sec": 107853} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:18.892275+00:00", "epoch": 0, "step": 21568, "train_loss": 3.476335048675537, "perplexity": 32.340976505731966, "lr": 0.0026291804804649314, "grad_norm": 0.170255, "tokens_per_sec": 108084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:19.196993+00:00", "epoch": 0, "step": 21569, "train_loss": 3.499077796936035, "perplexity": 33.084926864762366, "lr": 0.0026291804804649314, "grad_norm": 0.174271, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:19.502844+00:00", "epoch": 0, "step": 21570, "train_loss": 3.521723985671997, "perplexity": 33.842722569120845, "lr": 0.0026291804804649314, "grad_norm": 0.190991, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:19.808304+00:00", "epoch": 0, "step": 21571, "train_loss": 3.5396409034729004, "perplexity": 34.454544461870206, "lr": 0.0026291804804649314, "grad_norm": 0.148312, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:20.112764+00:00", "epoch": 0, "step": 21572, "train_loss": 3.55692982673645, "perplexity": 35.055405591413795, "lr": 0.0026291804804649314, "grad_norm": 0.175918, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:20.415997+00:00", "epoch": 0, "step": 21573, "train_loss": 3.5561325550079346, "perplexity": 35.02746804599153, "lr": 0.0026291804804649314, "grad_norm": 0.1709, "tokens_per_sec": 108056} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:20.720659+00:00", "epoch": 0, "step": 21574, "train_loss": 3.413477897644043, "perplexity": 30.37068693423724, "lr": 0.0026291804804649314, "grad_norm": 0.167407, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:21.025115+00:00", "epoch": 0, "step": 21575, "train_loss": 3.4625675678253174, "perplexity": 31.898773724871504, "lr": 0.0026291804804649314, "grad_norm": 0.185842, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:21.329318+00:00", "epoch": 0, "step": 21576, "train_loss": 3.5125999450683594, "perplexity": 33.53534459122158, "lr": 0.0026291804804649314, "grad_norm": 0.233779, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:21.634595+00:00", "epoch": 0, "step": 21577, "train_loss": 3.5540778636932373, "perplexity": 34.95557129972586, "lr": 0.0026291804804649314, "grad_norm": 0.192068, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:21.939265+00:00", "epoch": 0, "step": 21578, "train_loss": 3.577977180480957, "perplexity": 35.80104849261359, "lr": 0.0026291804804649314, "grad_norm": 0.199372, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:22.243154+00:00", "epoch": 0, "step": 21579, "train_loss": 3.5055365562438965, "perplexity": 33.29930601050635, "lr": 0.0026291804804649314, "grad_norm": 0.251267, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:22.547712+00:00", "epoch": 0, "step": 21580, "train_loss": 3.496638059616089, "perplexity": 33.004306719940466, "lr": 0.0026291804804649314, "grad_norm": 0.197131, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:22.853068+00:00", "epoch": 0, "step": 21581, "train_loss": 3.5081851482391357, "perplexity": 33.387619187012284, "lr": 0.0026291804804649314, "grad_norm": 0.215596, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:23.159115+00:00", "epoch": 0, "step": 21582, "train_loss": 3.4408164024353027, "perplexity": 31.212429672992226, "lr": 0.0026291804804649314, "grad_norm": 0.198647, "tokens_per_sec": 107069} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:23.463490+00:00", "epoch": 0, "step": 21583, "train_loss": 3.5354197025299072, "perplexity": 34.30941143962732, "lr": 0.0026291804804649314, "grad_norm": 0.149098, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:23.768164+00:00", "epoch": 0, "step": 21584, "train_loss": 3.4922077655792236, "perplexity": 32.85841135500098, "lr": 0.0026291804804649314, "grad_norm": 0.165724, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:24.073814+00:00", "epoch": 0, "step": 21585, "train_loss": 3.4997267723083496, "perplexity": 33.10640513617318, "lr": 0.0026291804804649314, "grad_norm": 0.159804, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:24.379744+00:00", "epoch": 0, "step": 21586, "train_loss": 3.418260097503662, "perplexity": 30.516273463513915, "lr": 0.0026291804804649314, "grad_norm": 0.154942, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:24.685691+00:00", "epoch": 0, "step": 21587, "train_loss": 3.482370376586914, "perplexity": 32.53675510366423, "lr": 0.0026291804804649314, "grad_norm": 0.17711, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:24.990381+00:00", "epoch": 0, "step": 21588, "train_loss": 3.444509744644165, "perplexity": 31.327921000158735, "lr": 0.0026291804804649314, "grad_norm": 0.184923, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:25.295769+00:00", "epoch": 0, "step": 21589, "train_loss": 3.4786453247070312, "perplexity": 32.41577946305702, "lr": 0.0026291804804649314, "grad_norm": 0.187854, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:25.600933+00:00", "epoch": 0, "step": 21590, "train_loss": 3.468721866607666, "perplexity": 32.09569363962615, "lr": 0.0026291804804649314, "grad_norm": 0.202254, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:25.906629+00:00", "epoch": 0, "step": 21591, "train_loss": 3.4712491035461426, "perplexity": 32.17690964497588, "lr": 0.0026291804804649314, "grad_norm": 0.185113, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:26.211869+00:00", "epoch": 0, "step": 21592, "train_loss": 3.459888219833374, "perplexity": 31.81342020649741, "lr": 0.0026291804804649314, "grad_norm": 0.240503, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:26.517324+00:00", "epoch": 0, "step": 21593, "train_loss": 3.5919229984283447, "perplexity": 36.30382103591085, "lr": 0.0026291804804649314, "grad_norm": 0.191132, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:26.821812+00:00", "epoch": 0, "step": 21594, "train_loss": 3.473813056945801, "perplexity": 32.25951559550114, "lr": 0.0026291804804649314, "grad_norm": 0.240946, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:27.126929+00:00", "epoch": 0, "step": 21595, "train_loss": 3.5616800785064697, "perplexity": 35.22232373159455, "lr": 0.0026291804804649314, "grad_norm": 0.266912, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:27.431619+00:00", "epoch": 0, "step": 21596, "train_loss": 3.5038058757781982, "perplexity": 33.24172539326977, "lr": 0.0026291804804649314, "grad_norm": 0.225775, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:27.736089+00:00", "epoch": 0, "step": 21597, "train_loss": 3.5699729919433594, "perplexity": 35.51563393042236, "lr": 0.0026291804804649314, "grad_norm": 0.184249, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:28.041685+00:00", "epoch": 0, "step": 21598, "train_loss": 3.5074710845947266, "perplexity": 33.36378681190642, "lr": 0.0026291804804649314, "grad_norm": 0.227327, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:28.345746+00:00", "epoch": 0, "step": 21599, "train_loss": 3.374295711517334, "perplexity": 29.203708701000146, "lr": 0.0026291804804649314, "grad_norm": 0.211176, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:28.649952+00:00", "epoch": 0, "step": 21600, "train_loss": 3.481217384338379, "perplexity": 32.49926209596442, "lr": 0.0026291804804649314, "grad_norm": 0.188501, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:28.955155+00:00", "epoch": 0, "step": 21601, "train_loss": 3.570847749710083, "perplexity": 35.54671509930811, "lr": 0.0026291804804649314, "grad_norm": 0.171548, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:29.259681+00:00", "epoch": 0, "step": 21602, "train_loss": 3.42169451713562, "perplexity": 30.621259331760275, "lr": 0.0026291804804649314, "grad_norm": 0.183347, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:29.564948+00:00", "epoch": 0, "step": 21603, "train_loss": 3.5474963188171387, "perplexity": 34.72626506093005, "lr": 0.0026291804804649314, "grad_norm": 0.163189, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:29.869616+00:00", "epoch": 0, "step": 21604, "train_loss": 3.453658103942871, "perplexity": 31.615835040074664, "lr": 0.0026291804804649314, "grad_norm": 0.167579, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:30.174680+00:00", "epoch": 0, "step": 21605, "train_loss": 3.562047243118286, "perplexity": 35.23525849686346, "lr": 0.0026291804804649314, "grad_norm": 0.165804, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:30.479690+00:00", "epoch": 0, "step": 21606, "train_loss": 3.4557833671569824, "perplexity": 31.68309846218945, "lr": 0.0026291804804649314, "grad_norm": 0.180917, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:30.784556+00:00", "epoch": 0, "step": 21607, "train_loss": 3.459019899368286, "perplexity": 31.785807952545188, "lr": 0.0026291804804649314, "grad_norm": 0.158657, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:31.090431+00:00", "epoch": 0, "step": 21608, "train_loss": 3.4245548248291016, "perplexity": 30.708970936700197, "lr": 0.0026291804804649314, "grad_norm": 0.159923, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:31.396120+00:00", "epoch": 0, "step": 21609, "train_loss": 3.502413749694824, "perplexity": 33.19548091683291, "lr": 0.0026291804804649314, "grad_norm": 0.170213, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:31.702014+00:00", "epoch": 0, "step": 21610, "train_loss": 3.6036343574523926, "perplexity": 36.73148750773265, "lr": 0.0026291804804649314, "grad_norm": 0.162812, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:32.006280+00:00", "epoch": 0, "step": 21611, "train_loss": 3.4919111728668213, "perplexity": 32.84866723474093, "lr": 0.0026291804804649314, "grad_norm": 0.183451, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:32.311390+00:00", "epoch": 0, "step": 21612, "train_loss": 3.5225677490234375, "perplexity": 33.87128986846291, "lr": 0.0026291804804649314, "grad_norm": 0.15851, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:32.615912+00:00", "epoch": 0, "step": 21613, "train_loss": 3.4700489044189453, "perplexity": 32.13831411189963, "lr": 0.0026291804804649314, "grad_norm": 0.159299, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:32.921656+00:00", "epoch": 0, "step": 21614, "train_loss": 3.5350639820098877, "perplexity": 34.2972090483976, "lr": 0.0026291804804649314, "grad_norm": 0.166596, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:33.226155+00:00", "epoch": 0, "step": 21615, "train_loss": 3.5496466159820557, "perplexity": 34.80101719108239, "lr": 0.0026291804804649314, "grad_norm": 0.157726, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:33.530205+00:00", "epoch": 0, "step": 21616, "train_loss": 3.7003233432769775, "perplexity": 40.46038483863252, "lr": 0.0026291804804649314, "grad_norm": 0.167198, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:33.836762+00:00", "epoch": 0, "step": 21617, "train_loss": 3.5391104221343994, "perplexity": 34.43627181608394, "lr": 0.0026291804804649314, "grad_norm": 0.167162, "tokens_per_sec": 106893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:34.142937+00:00", "epoch": 0, "step": 21618, "train_loss": 3.4407341480255127, "perplexity": 31.209862418596803, "lr": 0.0026291804804649314, "grad_norm": 0.184623, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:34.448472+00:00", "epoch": 0, "step": 21619, "train_loss": 3.5970542430877686, "perplexity": 36.4905835760786, "lr": 0.0026291804804649314, "grad_norm": 0.181286, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:34.753123+00:00", "epoch": 0, "step": 21620, "train_loss": 3.4712204933166504, "perplexity": 32.17598906937559, "lr": 0.0026291804804649314, "grad_norm": 0.177861, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:35.057184+00:00", "epoch": 0, "step": 21621, "train_loss": 3.609768867492676, "perplexity": 36.95750974567807, "lr": 0.0026291804804649314, "grad_norm": 0.188231, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:35.361857+00:00", "epoch": 0, "step": 21622, "train_loss": 3.5223276615142822, "perplexity": 33.863158770973044, "lr": 0.0026291804804649314, "grad_norm": 0.186418, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:35.667272+00:00", "epoch": 0, "step": 21623, "train_loss": 3.5844802856445312, "perplexity": 36.034625139506744, "lr": 0.0026291804804649314, "grad_norm": 0.166854, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:35.972369+00:00", "epoch": 0, "step": 21624, "train_loss": 3.4443747997283936, "perplexity": 31.323693741728007, "lr": 0.0026291804804649314, "grad_norm": 0.173572, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:36.277606+00:00", "epoch": 0, "step": 21625, "train_loss": 3.4748358726501465, "perplexity": 32.29252801459916, "lr": 0.0026291804804649314, "grad_norm": 0.159233, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:36.581896+00:00", "epoch": 0, "step": 21626, "train_loss": 3.423541784286499, "perplexity": 30.67787725635828, "lr": 0.0026291804804649314, "grad_norm": 0.17214, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:36.887410+00:00", "epoch": 0, "step": 21627, "train_loss": 3.4795713424682617, "perplexity": 32.44581095328459, "lr": 0.0026291804804649314, "grad_norm": 0.173472, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:37.192898+00:00", "epoch": 0, "step": 21628, "train_loss": 3.490398406982422, "perplexity": 32.79901245909494, "lr": 0.0026291804804649314, "grad_norm": 0.176586, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:37.497425+00:00", "epoch": 0, "step": 21629, "train_loss": 3.483825206756592, "perplexity": 32.58412500583888, "lr": 0.0026291804804649314, "grad_norm": 0.167215, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:37.801640+00:00", "epoch": 0, "step": 21630, "train_loss": 3.5469141006469727, "perplexity": 34.70605268300738, "lr": 0.0026291804804649314, "grad_norm": 0.225947, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:38.107374+00:00", "epoch": 0, "step": 21631, "train_loss": 3.462677001953125, "perplexity": 31.902264730366305, "lr": 0.0026291804804649314, "grad_norm": 0.22125, "tokens_per_sec": 107121} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:38.411296+00:00", "epoch": 0, "step": 21632, "train_loss": 3.422112464904785, "perplexity": 30.634060093625607, "lr": 0.0026291804804649314, "grad_norm": 0.178776, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:38.715433+00:00", "epoch": 0, "step": 21633, "train_loss": 3.3942973613739014, "perplexity": 29.793711905822082, "lr": 0.0026291804804649314, "grad_norm": 0.211819, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:39.019549+00:00", "epoch": 0, "step": 21634, "train_loss": 3.437795877456665, "perplexity": 31.118293990682, "lr": 0.0026291804804649314, "grad_norm": 0.177172, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:39.324709+00:00", "epoch": 0, "step": 21635, "train_loss": 3.560234785079956, "perplexity": 35.171453908404324, "lr": 0.0026291804804649314, "grad_norm": 0.190423, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:39.629535+00:00", "epoch": 0, "step": 21636, "train_loss": 3.4600014686584473, "perplexity": 31.81702324297333, "lr": 0.0026291804804649314, "grad_norm": 0.182714, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:39.934917+00:00", "epoch": 0, "step": 21637, "train_loss": 3.4674196243286133, "perplexity": 32.05392447308322, "lr": 0.0026291804804649314, "grad_norm": 0.188193, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:40.239179+00:00", "epoch": 0, "step": 21638, "train_loss": 3.442349910736084, "perplexity": 31.260330912069943, "lr": 0.0026291804804649314, "grad_norm": 0.1797, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:40.544197+00:00", "epoch": 0, "step": 21639, "train_loss": 3.456160068511963, "perplexity": 31.695035776570226, "lr": 0.0026291804804649314, "grad_norm": 0.165062, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:40.850081+00:00", "epoch": 0, "step": 21640, "train_loss": 3.4257419109344482, "perplexity": 30.745446775105727, "lr": 0.0026291804804649314, "grad_norm": 0.191212, "tokens_per_sec": 107196} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:41.156073+00:00", "epoch": 0, "step": 21641, "train_loss": 3.506653308868408, "perplexity": 33.33651387000844, "lr": 0.0026291804804649314, "grad_norm": 0.156082, "tokens_per_sec": 107018} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:41.460807+00:00", "epoch": 0, "step": 21642, "train_loss": 3.394610643386841, "perplexity": 29.803047202074726, "lr": 0.0026291804804649314, "grad_norm": 0.182043, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:41.765453+00:00", "epoch": 0, "step": 21643, "train_loss": 3.4646034240722656, "perplexity": 31.963781193106332, "lr": 0.0026291804804649314, "grad_norm": 0.195977, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:42.070586+00:00", "epoch": 0, "step": 21644, "train_loss": 3.53879976272583, "perplexity": 34.42557552578585, "lr": 0.0026291804804649314, "grad_norm": 0.156945, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:42.375682+00:00", "epoch": 0, "step": 21645, "train_loss": 3.4458277225494385, "perplexity": 31.369237729142235, "lr": 0.0026291804804649314, "grad_norm": 0.170882, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:42.681400+00:00", "epoch": 0, "step": 21646, "train_loss": 3.6165497303009033, "perplexity": 37.2089651277267, "lr": 0.0026291804804649314, "grad_norm": 0.172892, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:42.985340+00:00", "epoch": 0, "step": 21647, "train_loss": 3.5266008377075195, "perplexity": 34.00817162694132, "lr": 0.0026291804804649314, "grad_norm": 0.198713, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:43.289910+00:00", "epoch": 0, "step": 21648, "train_loss": 3.536545753479004, "perplexity": 34.34806734511678, "lr": 0.0026291804804649314, "grad_norm": 0.15356, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:43.593800+00:00", "epoch": 0, "step": 21649, "train_loss": 3.503840923309326, "perplexity": 33.242890454091366, "lr": 0.0026291804804649314, "grad_norm": 0.21738, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:43.898382+00:00", "epoch": 0, "step": 21650, "train_loss": 3.4658915996551514, "perplexity": 32.0049826872579, "lr": 0.0026291804804649314, "grad_norm": 0.19722, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:44.203620+00:00", "epoch": 0, "step": 21651, "train_loss": 3.5291500091552734, "perplexity": 34.094974878221954, "lr": 0.0026291804804649314, "grad_norm": 0.185869, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:44.508523+00:00", "epoch": 0, "step": 21652, "train_loss": 3.556825876235962, "perplexity": 35.0517617538504, "lr": 0.0026291804804649314, "grad_norm": 0.191907, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:44.812411+00:00", "epoch": 0, "step": 21653, "train_loss": 3.462372064590454, "perplexity": 31.89253802099012, "lr": 0.0026291804804649314, "grad_norm": 0.171253, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:45.116485+00:00", "epoch": 0, "step": 21654, "train_loss": 3.50895357131958, "perplexity": 33.413284863983044, "lr": 0.0026291804804649314, "grad_norm": 0.194845, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:45.421775+00:00", "epoch": 0, "step": 21655, "train_loss": 3.691335916519165, "perplexity": 40.09837927923952, "lr": 0.0026291804804649314, "grad_norm": 0.182653, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:45.727646+00:00", "epoch": 0, "step": 21656, "train_loss": 3.4483015537261963, "perplexity": 31.446935993997624, "lr": 0.0026291804804649314, "grad_norm": 0.229997, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:46.032152+00:00", "epoch": 0, "step": 21657, "train_loss": 3.46419620513916, "perplexity": 31.95076758610427, "lr": 0.0026291804804649314, "grad_norm": 0.191142, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:46.336373+00:00", "epoch": 0, "step": 21658, "train_loss": 3.4604387283325195, "perplexity": 31.830938586274026, "lr": 0.0026291804804649314, "grad_norm": 0.191267, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:46.640770+00:00", "epoch": 0, "step": 21659, "train_loss": 3.4697232246398926, "perplexity": 32.127849017087925, "lr": 0.0026291804804649314, "grad_norm": 0.169853, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:46.944867+00:00", "epoch": 0, "step": 21660, "train_loss": 3.5531833171844482, "perplexity": 34.9243158972468, "lr": 0.0026291804804649314, "grad_norm": 0.173531, "tokens_per_sec": 107825} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:47.249542+00:00", "epoch": 0, "step": 21661, "train_loss": 3.4143683910369873, "perplexity": 30.39774387551712, "lr": 0.0026291804804649314, "grad_norm": 0.163907, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:47.553985+00:00", "epoch": 0, "step": 21662, "train_loss": 3.447587728500366, "perplexity": 31.42449638774138, "lr": 0.0026291804804649314, "grad_norm": 0.166343, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:47.858177+00:00", "epoch": 0, "step": 21663, "train_loss": 3.520141124725342, "perplexity": 33.78919661848169, "lr": 0.0026291804804649314, "grad_norm": 0.177362, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:48.162476+00:00", "epoch": 0, "step": 21664, "train_loss": 3.4282755851745605, "perplexity": 30.823444490275858, "lr": 0.0026291804804649314, "grad_norm": 0.178538, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:48.466667+00:00", "epoch": 0, "step": 21665, "train_loss": 3.485187292098999, "perplexity": 32.628537604913134, "lr": 0.0026291804804649314, "grad_norm": 0.160515, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:48.771503+00:00", "epoch": 0, "step": 21666, "train_loss": 3.5052649974823, "perplexity": 33.290264519907666, "lr": 0.0026291804804649314, "grad_norm": 0.162981, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:49.077284+00:00", "epoch": 0, "step": 21667, "train_loss": 3.534818410873413, "perplexity": 34.288787677858764, "lr": 0.0026291804804649314, "grad_norm": 0.170128, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:49.382643+00:00", "epoch": 0, "step": 21668, "train_loss": 3.466831684112549, "perplexity": 32.03508422081651, "lr": 0.0026291804804649314, "grad_norm": 0.157264, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:49.687134+00:00", "epoch": 0, "step": 21669, "train_loss": 3.5099363327026367, "perplexity": 33.44613829092388, "lr": 0.0026291804804649314, "grad_norm": 0.157949, "tokens_per_sec": 107613} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:49.991687+00:00", "epoch": 0, "step": 21670, "train_loss": 3.519223213195801, "perplexity": 33.75819535571662, "lr": 0.0026291804804649314, "grad_norm": 0.172102, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:50.296952+00:00", "epoch": 0, "step": 21671, "train_loss": 3.523433208465576, "perplexity": 33.900616784894815, "lr": 0.0026291804804649314, "grad_norm": 0.187866, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:50.602577+00:00", "epoch": 0, "step": 21672, "train_loss": 3.4984500408172607, "perplexity": 33.06416411713658, "lr": 0.0026291804804649314, "grad_norm": 0.19093, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:50.908837+00:00", "epoch": 0, "step": 21673, "train_loss": 3.5213282108306885, "perplexity": 33.8293311211405, "lr": 0.0026291804804649314, "grad_norm": 0.170135, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:51.212755+00:00", "epoch": 0, "step": 21674, "train_loss": 3.5482654571533203, "perplexity": 34.75298463686869, "lr": 0.0026291804804649314, "grad_norm": 0.208218, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:51.517315+00:00", "epoch": 0, "step": 21675, "train_loss": 3.5535430908203125, "perplexity": 34.936883005878855, "lr": 0.0026291804804649314, "grad_norm": 0.171783, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:51.821841+00:00", "epoch": 0, "step": 21676, "train_loss": 3.450592517852783, "perplexity": 31.519062384190825, "lr": 0.0026291804804649314, "grad_norm": 0.164072, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:52.126960+00:00", "epoch": 0, "step": 21677, "train_loss": 3.55130934715271, "perplexity": 34.85893006056059, "lr": 0.0026291804804649314, "grad_norm": 0.198613, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:52.431649+00:00", "epoch": 0, "step": 21678, "train_loss": 3.4732112884521484, "perplexity": 32.24010867521752, "lr": 0.0026291804804649314, "grad_norm": 0.170548, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:52.735667+00:00", "epoch": 0, "step": 21679, "train_loss": 3.5902438163757324, "perplexity": 36.24291146462755, "lr": 0.0026291804804649314, "grad_norm": 0.205721, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:53.040407+00:00", "epoch": 0, "step": 21680, "train_loss": 3.5868947505950928, "perplexity": 36.12173459796179, "lr": 0.0026291804804649314, "grad_norm": 0.16831, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:53.344832+00:00", "epoch": 0, "step": 21681, "train_loss": 3.555325508117676, "perplexity": 34.99921064089056, "lr": 0.0026291804804649314, "grad_norm": 0.180021, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:53.649389+00:00", "epoch": 0, "step": 21682, "train_loss": 3.4994213581085205, "perplexity": 33.09629551382899, "lr": 0.0026291804804649314, "grad_norm": 0.189757, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:53.954649+00:00", "epoch": 0, "step": 21683, "train_loss": 3.4716744422912598, "perplexity": 32.190598642370134, "lr": 0.0026291804804649314, "grad_norm": 0.171597, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:54.259803+00:00", "epoch": 0, "step": 21684, "train_loss": 3.402527093887329, "perplexity": 30.0399179005337, "lr": 0.0026291804804649314, "grad_norm": 0.19515, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:54.564448+00:00", "epoch": 0, "step": 21685, "train_loss": 3.575127124786377, "perplexity": 35.699158775139324, "lr": 0.0026291804804649314, "grad_norm": 0.152945, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:54.869303+00:00", "epoch": 0, "step": 21686, "train_loss": 3.5552022457122803, "perplexity": 34.994896819871464, "lr": 0.0026291804804649314, "grad_norm": 0.1971, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:55.174231+00:00", "epoch": 0, "step": 21687, "train_loss": 3.4746501445770264, "perplexity": 32.28653094252439, "lr": 0.0026291804804649314, "grad_norm": 0.18609, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:55.480696+00:00", "epoch": 0, "step": 21688, "train_loss": 3.7024593353271484, "perplexity": 40.54690026423173, "lr": 0.0026291804804649314, "grad_norm": 0.187005, "tokens_per_sec": 106924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:55.785671+00:00", "epoch": 0, "step": 21689, "train_loss": 3.4621615409851074, "perplexity": 31.88582459559434, "lr": 0.0026291804804649314, "grad_norm": 0.223779, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:56.089716+00:00", "epoch": 0, "step": 21690, "train_loss": 3.5164380073547363, "perplexity": 33.66430264878269, "lr": 0.0026291804804649314, "grad_norm": 0.179581, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:56.394866+00:00", "epoch": 0, "step": 21691, "train_loss": 3.486870527267456, "perplexity": 32.683505355752466, "lr": 0.0026291804804649314, "grad_norm": 0.193743, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:56.699347+00:00", "epoch": 0, "step": 21692, "train_loss": 3.549997568130493, "perplexity": 34.813232826259735, "lr": 0.0026291804804649314, "grad_norm": 0.184226, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:57.003147+00:00", "epoch": 0, "step": 21693, "train_loss": 3.5633223056793213, "perplexity": 35.280214310447896, "lr": 0.0026291804804649314, "grad_norm": 0.180454, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:57.308798+00:00", "epoch": 0, "step": 21694, "train_loss": 3.492535352706909, "perplexity": 32.869177110862324, "lr": 0.0026291804804649314, "grad_norm": 0.17925, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:57.614367+00:00", "epoch": 0, "step": 21695, "train_loss": 3.5288662910461426, "perplexity": 34.085302888542834, "lr": 0.0026291804804649314, "grad_norm": 0.22816, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:57.918038+00:00", "epoch": 0, "step": 21696, "train_loss": 3.6092309951782227, "perplexity": 36.937636669442604, "lr": 0.0026291804804649314, "grad_norm": 0.197928, "tokens_per_sec": 107902} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:58.222332+00:00", "epoch": 0, "step": 21697, "train_loss": 3.4367294311523438, "perplexity": 31.08512569030808, "lr": 0.0026291804804649314, "grad_norm": 0.190478, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:58.527978+00:00", "epoch": 0, "step": 21698, "train_loss": 3.422936201095581, "perplexity": 30.659304873687756, "lr": 0.0026291804804649314, "grad_norm": 0.211058, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:58.834239+00:00", "epoch": 0, "step": 21699, "train_loss": 3.409208059310913, "perplexity": 30.241285469228494, "lr": 0.0026291804804649314, "grad_norm": 0.189956, "tokens_per_sec": 106994} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:59.138856+00:00", "epoch": 0, "step": 21700, "train_loss": 3.4848103523254395, "perplexity": 32.61624092903546, "lr": 0.0026291804804649314, "grad_norm": 0.180883, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:59.443738+00:00", "epoch": 0, "step": 21701, "train_loss": 3.485434055328369, "perplexity": 32.63659012171381, "lr": 0.0026291804804649314, "grad_norm": 0.209395, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:51:59.748203+00:00", "epoch": 0, "step": 21702, "train_loss": 3.4737119674682617, "perplexity": 32.256254662749704, "lr": 0.0026291804804649314, "grad_norm": 0.175096, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:00.052356+00:00", "epoch": 0, "step": 21703, "train_loss": 3.5533127784729004, "perplexity": 34.9288375368635, "lr": 0.0026291804804649314, "grad_norm": 0.162059, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:00.357623+00:00", "epoch": 0, "step": 21704, "train_loss": 3.499126672744751, "perplexity": 33.086543956837204, "lr": 0.0026291804804649314, "grad_norm": 0.170605, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:00.966458+00:00", "epoch": 0, "step": 21705, "train_loss": 3.460566282272339, "perplexity": 31.83499900685467, "lr": 0.0026291804804649314, "grad_norm": 0.165825, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:01.270231+00:00", "epoch": 0, "step": 21706, "train_loss": 3.4588799476623535, "perplexity": 31.781359785769308, "lr": 0.0026291804804649314, "grad_norm": 0.15833, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:01.574139+00:00", "epoch": 0, "step": 21707, "train_loss": 3.5391767024993896, "perplexity": 34.43855434039125, "lr": 0.0026291804804649314, "grad_norm": 0.162133, "tokens_per_sec": 107822} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:01.879595+00:00", "epoch": 0, "step": 21708, "train_loss": 3.447420835494995, "perplexity": 31.419252296710187, "lr": 0.0026291804804649314, "grad_norm": 0.183017, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:02.184947+00:00", "epoch": 0, "step": 21709, "train_loss": 3.5037546157836914, "perplexity": 33.24002146628073, "lr": 0.0026291804804649314, "grad_norm": 0.171547, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:02.490146+00:00", "epoch": 0, "step": 21710, "train_loss": 3.375185966491699, "perplexity": 29.229719024133022, "lr": 0.0026291804804649314, "grad_norm": 0.163105, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:02.795571+00:00", "epoch": 0, "step": 21711, "train_loss": 3.5895345211029053, "perplexity": 36.21721365359752, "lr": 0.0026291804804649314, "grad_norm": 0.163669, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:03.100680+00:00", "epoch": 0, "step": 21712, "train_loss": 3.422750473022461, "perplexity": 30.653611108832717, "lr": 0.0026291804804649314, "grad_norm": 0.190842, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:03.404724+00:00", "epoch": 0, "step": 21713, "train_loss": 3.51497745513916, "perplexity": 33.615170066048655, "lr": 0.0026291804804649314, "grad_norm": 0.18342, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:03.709994+00:00", "epoch": 0, "step": 21714, "train_loss": 3.475771903991699, "perplexity": 32.32276898395837, "lr": 0.0026291804804649314, "grad_norm": 0.173993, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:04.015301+00:00", "epoch": 0, "step": 21715, "train_loss": 3.616980791091919, "perplexity": 37.22500791112657, "lr": 0.0026291804804649314, "grad_norm": 0.189061, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:04.319891+00:00", "epoch": 0, "step": 21716, "train_loss": 3.4911866188049316, "perplexity": 32.824875219809726, "lr": 0.0026291804804649314, "grad_norm": 0.18666, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:04.625023+00:00", "epoch": 0, "step": 21717, "train_loss": 3.553953170776367, "perplexity": 34.95121285931861, "lr": 0.0026291804804649314, "grad_norm": 0.188177, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:04.930204+00:00", "epoch": 0, "step": 21718, "train_loss": 3.472266912460327, "perplexity": 32.209676262692575, "lr": 0.0026291804804649314, "grad_norm": 0.180597, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:05.235440+00:00", "epoch": 0, "step": 21719, "train_loss": 3.4902071952819824, "perplexity": 32.792741503709024, "lr": 0.0026291804804649314, "grad_norm": 0.163369, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:05.541130+00:00", "epoch": 0, "step": 21720, "train_loss": 3.6413931846618652, "perplexity": 38.14494267267299, "lr": 0.0026291804804649314, "grad_norm": 0.177124, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:05.845059+00:00", "epoch": 0, "step": 21721, "train_loss": 3.52150821685791, "perplexity": 33.83542115274393, "lr": 0.0026291804804649314, "grad_norm": 0.176323, "tokens_per_sec": 107735} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:06.149105+00:00", "epoch": 0, "step": 21722, "train_loss": 3.6129283905029297, "perplexity": 37.07446250830263, "lr": 0.0026291804804649314, "grad_norm": 0.177499, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:06.452983+00:00", "epoch": 0, "step": 21723, "train_loss": 3.5041637420654297, "perplexity": 33.253623614976654, "lr": 0.0026291804804649314, "grad_norm": 0.175493, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:06.758171+00:00", "epoch": 0, "step": 21724, "train_loss": 3.4664433002471924, "perplexity": 32.022644726785565, "lr": 0.0026291804804649314, "grad_norm": 0.17402, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:07.063923+00:00", "epoch": 0, "step": 21725, "train_loss": 3.5682032108306885, "perplexity": 35.45283461919649, "lr": 0.0026291804804649314, "grad_norm": 0.167687, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:07.368843+00:00", "epoch": 0, "step": 21726, "train_loss": 3.5867233276367188, "perplexity": 36.11554303405867, "lr": 0.0026291804804649314, "grad_norm": 0.173023, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:07.673337+00:00", "epoch": 0, "step": 21727, "train_loss": 3.497006416320801, "perplexity": 33.016466317002184, "lr": 0.0026291804804649314, "grad_norm": 0.188975, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:07.977498+00:00", "epoch": 0, "step": 21728, "train_loss": 3.5732319355010986, "perplexity": 35.631566182545114, "lr": 0.0026291804804649314, "grad_norm": 0.208918, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:08.282708+00:00", "epoch": 0, "step": 21729, "train_loss": 3.48520565032959, "perplexity": 32.62913661262867, "lr": 0.0026291804804649314, "grad_norm": 0.195309, "tokens_per_sec": 107363} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:08.588257+00:00", "epoch": 0, "step": 21730, "train_loss": 3.433573007583618, "perplexity": 30.987162554851164, "lr": 0.0026291804804649314, "grad_norm": 0.173175, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:08.894218+00:00", "epoch": 0, "step": 21731, "train_loss": 3.474324941635132, "perplexity": 32.27603297475583, "lr": 0.0026291804804649314, "grad_norm": 0.189397, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:09.198673+00:00", "epoch": 0, "step": 21732, "train_loss": 3.5089547634124756, "perplexity": 33.41332469574628, "lr": 0.0026291804804649314, "grad_norm": 0.180222, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:09.502918+00:00", "epoch": 0, "step": 21733, "train_loss": 3.6016862392425537, "perplexity": 36.65999988382427, "lr": 0.0026291804804649314, "grad_norm": 0.157446, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:09.807564+00:00", "epoch": 0, "step": 21734, "train_loss": 3.5084874629974365, "perplexity": 33.397714282908396, "lr": 0.0026291804804649314, "grad_norm": 0.158679, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:10.112739+00:00", "epoch": 0, "step": 21735, "train_loss": 3.449246406555176, "perplexity": 31.476662761933138, "lr": 0.0026291804804649314, "grad_norm": 0.177128, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:10.417248+00:00", "epoch": 0, "step": 21736, "train_loss": 3.4397592544555664, "perplexity": 31.179450950786848, "lr": 0.0026291804804649314, "grad_norm": 0.16552, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:10.721319+00:00", "epoch": 0, "step": 21737, "train_loss": 3.522583484649658, "perplexity": 33.87182285861335, "lr": 0.0026291804804649314, "grad_norm": 0.165084, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:11.025920+00:00", "epoch": 0, "step": 21738, "train_loss": 3.5273544788360596, "perplexity": 34.033811244107284, "lr": 0.0026291804804649314, "grad_norm": 0.173076, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:11.330554+00:00", "epoch": 0, "step": 21739, "train_loss": 3.5383245944976807, "perplexity": 34.40922147183137, "lr": 0.0026291804804649314, "grad_norm": 0.160793, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:11.635616+00:00", "epoch": 0, "step": 21740, "train_loss": 3.5018868446350098, "perplexity": 33.17799465717876, "lr": 0.0026291804804649314, "grad_norm": 0.163479, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:11.939729+00:00", "epoch": 0, "step": 21741, "train_loss": 3.5621719360351562, "perplexity": 35.239652357958086, "lr": 0.0026291804804649314, "grad_norm": 0.158823, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:12.245272+00:00", "epoch": 0, "step": 21742, "train_loss": 3.463326930999756, "perplexity": 31.92300567821001, "lr": 0.0026291804804649314, "grad_norm": 0.172265, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:12.549612+00:00", "epoch": 0, "step": 21743, "train_loss": 3.4192874431610107, "perplexity": 30.547640334085436, "lr": 0.0026291804804649314, "grad_norm": 0.163, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:12.853279+00:00", "epoch": 0, "step": 21744, "train_loss": 3.5204861164093018, "perplexity": 33.80085562133663, "lr": 0.0026291804804649314, "grad_norm": 0.152432, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:13.157447+00:00", "epoch": 0, "step": 21745, "train_loss": 3.582681894302368, "perplexity": 35.96987901851969, "lr": 0.0026291804804649314, "grad_norm": 0.156764, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:13.462845+00:00", "epoch": 0, "step": 21746, "train_loss": 3.5083608627319336, "perplexity": 33.39348639104475, "lr": 0.0026291804804649314, "grad_norm": 0.17513, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:13.767340+00:00", "epoch": 0, "step": 21747, "train_loss": 3.500969648361206, "perplexity": 33.147577875318746, "lr": 0.0026291804804649314, "grad_norm": 0.187354, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:14.071319+00:00", "epoch": 0, "step": 21748, "train_loss": 3.493889331817627, "perplexity": 32.913711432523954, "lr": 0.0026291804804649314, "grad_norm": 0.195691, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:14.375453+00:00", "epoch": 0, "step": 21749, "train_loss": 3.531054735183716, "perplexity": 34.159978351562785, "lr": 0.0026291804804649314, "grad_norm": 0.140266, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:14.681223+00:00", "epoch": 0, "step": 21750, "train_loss": 3.452084541320801, "perplexity": 31.566124665241265, "lr": 0.0026291804804649314, "grad_norm": 0.197749, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:14.986345+00:00", "epoch": 0, "step": 21751, "train_loss": 3.618330240249634, "perplexity": 37.275275075561424, "lr": 0.0026291804804649314, "grad_norm": 0.177228, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:15.291529+00:00", "epoch": 0, "step": 21752, "train_loss": 3.4228806495666504, "perplexity": 30.657601749732077, "lr": 0.0026291804804649314, "grad_norm": 0.157022, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:15.595441+00:00", "epoch": 0, "step": 21753, "train_loss": 3.6239283084869385, "perplexity": 37.48452977356314, "lr": 0.0026291804804649314, "grad_norm": 0.190086, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:15.899382+00:00", "epoch": 0, "step": 21754, "train_loss": 3.4046146869659424, "perplexity": 30.102694528453323, "lr": 0.0026291804804649314, "grad_norm": 0.183123, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:16.204538+00:00", "epoch": 0, "step": 21755, "train_loss": 3.3718583583831787, "perplexity": 29.132615624730096, "lr": 0.0026291804804649314, "grad_norm": 0.172739, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:16.508613+00:00", "epoch": 0, "step": 21756, "train_loss": 3.4748013019561768, "perplexity": 32.29141165879236, "lr": 0.0026291804804649314, "grad_norm": 0.174937, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:16.814080+00:00", "epoch": 0, "step": 21757, "train_loss": 3.535618543624878, "perplexity": 34.31623423886973, "lr": 0.0026291804804649314, "grad_norm": 0.165158, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:17.118721+00:00", "epoch": 0, "step": 21758, "train_loss": 3.4608936309814453, "perplexity": 31.845421858544416, "lr": 0.0026291804804649314, "grad_norm": 0.175452, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:17.423366+00:00", "epoch": 0, "step": 21759, "train_loss": 3.5012784004211426, "perplexity": 33.15781383837569, "lr": 0.0026291804804649314, "grad_norm": 0.165834, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:17.729185+00:00", "epoch": 0, "step": 21760, "train_loss": 3.531797170639038, "perplexity": 34.18534934764214, "lr": 0.0026291804804649314, "grad_norm": 0.202904, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:18.033798+00:00", "epoch": 0, "step": 21761, "train_loss": 3.4881536960601807, "perplexity": 32.725470728432036, "lr": 0.0026291804804649314, "grad_norm": 0.184311, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:18.338639+00:00", "epoch": 0, "step": 21762, "train_loss": 3.5251710414886475, "perplexity": 33.95958161692245, "lr": 0.0026291804804649314, "grad_norm": 0.160511, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:18.643977+00:00", "epoch": 0, "step": 21763, "train_loss": 3.519860029220581, "perplexity": 33.77969996199901, "lr": 0.0026291804804649314, "grad_norm": 0.176476, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:18.947987+00:00", "epoch": 0, "step": 21764, "train_loss": 3.4988019466400146, "perplexity": 33.075801636549215, "lr": 0.0026291804804649314, "grad_norm": 0.181154, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:19.251487+00:00", "epoch": 0, "step": 21765, "train_loss": 3.4936625957489014, "perplexity": 32.906249552957156, "lr": 0.0026291804804649314, "grad_norm": 0.187241, "tokens_per_sec": 107963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:19.556807+00:00", "epoch": 0, "step": 21766, "train_loss": 3.5172665119171143, "perplexity": 33.69220523423007, "lr": 0.0026291804804649314, "grad_norm": 0.199479, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:19.862765+00:00", "epoch": 0, "step": 21767, "train_loss": 3.5664172172546387, "perplexity": 35.38957259391518, "lr": 0.0026291804804649314, "grad_norm": 0.235007, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:20.168526+00:00", "epoch": 0, "step": 21768, "train_loss": 3.411297559738159, "perplexity": 30.304540711049896, "lr": 0.0026291804804649314, "grad_norm": 0.231867, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:20.472654+00:00", "epoch": 0, "step": 21769, "train_loss": 3.5367438793182373, "perplexity": 34.354873258977015, "lr": 0.0026291804804649314, "grad_norm": 0.186682, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:20.776320+00:00", "epoch": 0, "step": 21770, "train_loss": 3.448655128479004, "perplexity": 31.458056802520964, "lr": 0.0026291804804649314, "grad_norm": 0.185375, "tokens_per_sec": 107999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:21.080386+00:00", "epoch": 0, "step": 21771, "train_loss": 3.4420247077941895, "perplexity": 31.25016661331225, "lr": 0.0026291804804649314, "grad_norm": 0.201983, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:21.385143+00:00", "epoch": 0, "step": 21772, "train_loss": 3.4280920028686523, "perplexity": 30.817786370641514, "lr": 0.0026291804804649314, "grad_norm": 0.167103, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:21.690964+00:00", "epoch": 0, "step": 21773, "train_loss": 3.5139126777648926, "perplexity": 33.57939644238257, "lr": 0.0026291804804649314, "grad_norm": 0.169578, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:21.995639+00:00", "epoch": 0, "step": 21774, "train_loss": 3.5722732543945312, "perplexity": 35.59742324195952, "lr": 0.0026291804804649314, "grad_norm": 0.186922, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:22.300582+00:00", "epoch": 0, "step": 21775, "train_loss": 3.414098024368286, "perplexity": 30.389526449678545, "lr": 0.0026291804804649314, "grad_norm": 0.163854, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:22.604861+00:00", "epoch": 0, "step": 21776, "train_loss": 3.5751023292541504, "perplexity": 35.69827360647161, "lr": 0.0026291804804649314, "grad_norm": 0.168485, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:22.909922+00:00", "epoch": 0, "step": 21777, "train_loss": 3.441462516784668, "perplexity": 31.232602988114884, "lr": 0.0026291804804649314, "grad_norm": 0.191037, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:23.215349+00:00", "epoch": 0, "step": 21778, "train_loss": 3.4747939109802246, "perplexity": 32.29117299462731, "lr": 0.0026291804804649314, "grad_norm": 0.180965, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:23.520820+00:00", "epoch": 0, "step": 21779, "train_loss": 3.5041637420654297, "perplexity": 33.253623614976654, "lr": 0.0026291804804649314, "grad_norm": 0.175446, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:23.825121+00:00", "epoch": 0, "step": 21780, "train_loss": 3.422764301300049, "perplexity": 30.654034998407028, "lr": 0.0026291804804649314, "grad_norm": 0.194175, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:24.130268+00:00", "epoch": 0, "step": 21781, "train_loss": 3.5879478454589844, "perplexity": 36.159794247835286, "lr": 0.0026291804804649314, "grad_norm": 0.185248, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:24.435257+00:00", "epoch": 0, "step": 21782, "train_loss": 3.560985803604126, "perplexity": 35.197878243151955, "lr": 0.0026291804804649314, "grad_norm": 0.1979, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:24.739386+00:00", "epoch": 0, "step": 21783, "train_loss": 3.4644317626953125, "perplexity": 31.95829471733556, "lr": 0.0026291804804649314, "grad_norm": 0.171377, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:25.043549+00:00", "epoch": 0, "step": 21784, "train_loss": 3.3997409343719482, "perplexity": 29.956338384432684, "lr": 0.0026291804804649314, "grad_norm": 0.176227, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:25.347508+00:00", "epoch": 0, "step": 21785, "train_loss": 3.547654867172241, "perplexity": 34.73177128962453, "lr": 0.0026291804804649314, "grad_norm": 0.16243, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:25.651849+00:00", "epoch": 0, "step": 21786, "train_loss": 3.4972035884857178, "perplexity": 33.02297688697431, "lr": 0.0026291804804649314, "grad_norm": 0.175049, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:25.955667+00:00", "epoch": 0, "step": 21787, "train_loss": 3.512037515640259, "perplexity": 33.51648862961289, "lr": 0.0026291804804649314, "grad_norm": 0.150182, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:26.260973+00:00", "epoch": 0, "step": 21788, "train_loss": 3.4260787963867188, "perplexity": 30.755806213721925, "lr": 0.0026291804804649314, "grad_norm": 0.174492, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:26.565560+00:00", "epoch": 0, "step": 21789, "train_loss": 3.5155327320098877, "perplexity": 33.633840975774326, "lr": 0.0026291804804649314, "grad_norm": 0.153953, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:26.871140+00:00", "epoch": 0, "step": 21790, "train_loss": 3.5683720111846924, "perplexity": 35.458819575347796, "lr": 0.0026291804804649314, "grad_norm": 0.172839, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:27.175455+00:00", "epoch": 0, "step": 21791, "train_loss": 3.442274570465088, "perplexity": 31.25797583898464, "lr": 0.0026291804804649314, "grad_norm": 0.193409, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:27.480122+00:00", "epoch": 0, "step": 21792, "train_loss": 3.4743223190307617, "perplexity": 32.2759483276017, "lr": 0.0026291804804649314, "grad_norm": 0.152912, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:27.784952+00:00", "epoch": 0, "step": 21793, "train_loss": 3.451613664627075, "perplexity": 31.55126441177325, "lr": 0.0026291804804649314, "grad_norm": 0.193571, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:28.090268+00:00", "epoch": 0, "step": 21794, "train_loss": 3.5764822959899902, "perplexity": 35.74757004247343, "lr": 0.0026291804804649314, "grad_norm": 0.164699, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:28.394508+00:00", "epoch": 0, "step": 21795, "train_loss": 3.418961524963379, "perplexity": 30.537685924455584, "lr": 0.0026291804804649314, "grad_norm": 0.205765, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:28.698364+00:00", "epoch": 0, "step": 21796, "train_loss": 3.3923704624176025, "perplexity": 29.73635770906348, "lr": 0.0026291804804649314, "grad_norm": 0.211001, "tokens_per_sec": 107840} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:29.001974+00:00", "epoch": 0, "step": 21797, "train_loss": 3.431086778640747, "perplexity": 30.910217066135854, "lr": 0.0026291804804649314, "grad_norm": 0.171708, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:29.306568+00:00", "epoch": 0, "step": 21798, "train_loss": 3.568178176879883, "perplexity": 35.45194710578774, "lr": 0.0026291804804649314, "grad_norm": 0.184766, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:29.611288+00:00", "epoch": 0, "step": 21799, "train_loss": 3.557600498199463, "perplexity": 35.07892413729552, "lr": 0.0026291804804649314, "grad_norm": 0.181534, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:29.915908+00:00", "epoch": 0, "step": 21800, "train_loss": 3.4768006801605225, "perplexity": 32.35603898915932, "lr": 0.0026291804804649314, "grad_norm": 0.188293, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:30.219846+00:00", "epoch": 0, "step": 21801, "train_loss": 3.444976568222046, "perplexity": 31.34254902641573, "lr": 0.0026291804804649314, "grad_norm": 0.191563, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:30.523451+00:00", "epoch": 0, "step": 21802, "train_loss": 3.486895799636841, "perplexity": 32.68433135581005, "lr": 0.0026291804804649314, "grad_norm": 0.21071, "tokens_per_sec": 107930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:30.828273+00:00", "epoch": 0, "step": 21803, "train_loss": 3.4949331283569336, "perplexity": 32.94808458678186, "lr": 0.0026291804804649314, "grad_norm": 0.208354, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:31.133506+00:00", "epoch": 0, "step": 21804, "train_loss": 3.523047685623169, "perplexity": 33.887549841716925, "lr": 0.0026291804804649314, "grad_norm": 0.175623, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:31.438531+00:00", "epoch": 0, "step": 21805, "train_loss": 3.45930552482605, "perplexity": 31.794888085188813, "lr": 0.0026291804804649314, "grad_norm": 0.184244, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:31.743929+00:00", "epoch": 0, "step": 21806, "train_loss": 3.5054993629455566, "perplexity": 33.29806752251526, "lr": 0.0026291804804649314, "grad_norm": 0.154949, "tokens_per_sec": 107297} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:32.048774+00:00", "epoch": 0, "step": 21807, "train_loss": 3.4909114837646484, "perplexity": 32.81584518873978, "lr": 0.0026291804804649314, "grad_norm": 0.189789, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:32.352736+00:00", "epoch": 0, "step": 21808, "train_loss": 3.395782470703125, "perplexity": 29.83799169735293, "lr": 0.0026291804804649314, "grad_norm": 0.16817, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:32.656646+00:00", "epoch": 0, "step": 21809, "train_loss": 3.5111467838287354, "perplexity": 33.48664771907706, "lr": 0.0026291804804649314, "grad_norm": 0.163229, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:32.961442+00:00", "epoch": 0, "step": 21810, "train_loss": 3.4389290809631348, "perplexity": 31.1535773383785, "lr": 0.0026291804804649314, "grad_norm": 0.163139, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:33.264928+00:00", "epoch": 0, "step": 21811, "train_loss": 3.376756191253662, "perplexity": 29.27565230608051, "lr": 0.0026291804804649314, "grad_norm": 0.149483, "tokens_per_sec": 107912} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:33.568799+00:00", "epoch": 0, "step": 21812, "train_loss": 3.448789358139038, "perplexity": 31.46227969020295, "lr": 0.0026291804804649314, "grad_norm": 0.185513, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:33.873008+00:00", "epoch": 0, "step": 21813, "train_loss": 3.430208683013916, "perplexity": 30.88308685290971, "lr": 0.0026291804804649314, "grad_norm": 0.16464, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:34.177308+00:00", "epoch": 0, "step": 21814, "train_loss": 3.5125651359558105, "perplexity": 33.53417727595406, "lr": 0.0026291804804649314, "grad_norm": 0.16351, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:34.482091+00:00", "epoch": 0, "step": 21815, "train_loss": 3.569974184036255, "perplexity": 35.515676268382485, "lr": 0.0026291804804649314, "grad_norm": 0.153797, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:34.786069+00:00", "epoch": 0, "step": 21816, "train_loss": 3.5294713973999023, "perplexity": 34.10593436337925, "lr": 0.0026291804804649314, "grad_norm": 0.192924, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:35.091039+00:00", "epoch": 0, "step": 21817, "train_loss": 3.4871044158935547, "perplexity": 32.69115054994334, "lr": 0.0026291804804649314, "grad_norm": 0.185911, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:35.395037+00:00", "epoch": 0, "step": 21818, "train_loss": 3.4887301921844482, "perplexity": 32.74434227463577, "lr": 0.0026291804804649314, "grad_norm": 0.21163, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:35.700709+00:00", "epoch": 0, "step": 21819, "train_loss": 3.5191566944122314, "perplexity": 33.75594987631017, "lr": 0.0026291804804649314, "grad_norm": 0.199654, "tokens_per_sec": 107199} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:36.006763+00:00", "epoch": 0, "step": 21820, "train_loss": 3.552812337875366, "perplexity": 34.91136210160872, "lr": 0.0026291804804649314, "grad_norm": 0.168445, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:36.311676+00:00", "epoch": 0, "step": 21821, "train_loss": 3.529968500137329, "perplexity": 34.122892731390145, "lr": 0.0026291804804649314, "grad_norm": 0.189977, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:36.616008+00:00", "epoch": 0, "step": 21822, "train_loss": 3.5442562103271484, "perplexity": 34.613930281393955, "lr": 0.0026291804804649314, "grad_norm": 0.174023, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:36.921211+00:00", "epoch": 0, "step": 21823, "train_loss": 3.4794023036956787, "perplexity": 32.4403268167545, "lr": 0.0026291804804649314, "grad_norm": 0.143759, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:37.225204+00:00", "epoch": 0, "step": 21824, "train_loss": 3.4644875526428223, "perplexity": 31.960077718656652, "lr": 0.0026291804804649314, "grad_norm": 0.14861, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:37.531036+00:00", "epoch": 0, "step": 21825, "train_loss": 3.5148067474365234, "perplexity": 33.609432187356916, "lr": 0.0026291804804649314, "grad_norm": 0.156384, "tokens_per_sec": 107145} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:37.837399+00:00", "epoch": 0, "step": 21826, "train_loss": 3.3990955352783203, "perplexity": 29.937010828455406, "lr": 0.0026291804804649314, "grad_norm": 0.153575, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:38.143055+00:00", "epoch": 0, "step": 21827, "train_loss": 3.4398908615112305, "perplexity": 31.183554656556097, "lr": 0.0026291804804649314, "grad_norm": 0.168678, "tokens_per_sec": 107202} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:38.447950+00:00", "epoch": 0, "step": 21828, "train_loss": 3.438697099685669, "perplexity": 31.14635112991486, "lr": 0.0026291804804649314, "grad_norm": 0.158341, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:38.753082+00:00", "epoch": 0, "step": 21829, "train_loss": 3.53997802734375, "perplexity": 34.46616186941022, "lr": 0.0026291804804649314, "grad_norm": 0.178717, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:39.057264+00:00", "epoch": 0, "step": 21830, "train_loss": 3.447021722793579, "perplexity": 31.406714976117943, "lr": 0.0026291804804649314, "grad_norm": 0.165724, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:39.362776+00:00", "epoch": 0, "step": 21831, "train_loss": 3.4972188472747803, "perplexity": 33.023480781457245, "lr": 0.0026291804804649314, "grad_norm": 0.158705, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:39.668120+00:00", "epoch": 0, "step": 21832, "train_loss": 3.5655922889709473, "perplexity": 35.36039077264445, "lr": 0.0026291804804649314, "grad_norm": 0.186555, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:39.972198+00:00", "epoch": 0, "step": 21833, "train_loss": 3.38242769241333, "perplexity": 29.44216093269385, "lr": 0.0026291804804649314, "grad_norm": 0.177888, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:40.277132+00:00", "epoch": 0, "step": 21834, "train_loss": 3.47536301612854, "perplexity": 32.30955529765921, "lr": 0.0026291804804649314, "grad_norm": 0.232543, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:40.582820+00:00", "epoch": 0, "step": 21835, "train_loss": 3.501425266265869, "perplexity": 33.162683946331846, "lr": 0.0026291804804649314, "grad_norm": 0.178136, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:40.889154+00:00", "epoch": 0, "step": 21836, "train_loss": 3.426255226135254, "perplexity": 30.761232931581365, "lr": 0.0026291804804649314, "grad_norm": 0.212134, "tokens_per_sec": 106968} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:41.195244+00:00", "epoch": 0, "step": 21837, "train_loss": 3.574652671813965, "perplexity": 35.68222522055115, "lr": 0.0026291804804649314, "grad_norm": 0.19302, "tokens_per_sec": 107060} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:41.500022+00:00", "epoch": 0, "step": 21838, "train_loss": 3.4652884006500244, "perplexity": 31.985683134862416, "lr": 0.0026291804804649314, "grad_norm": 0.166715, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:41.805016+00:00", "epoch": 0, "step": 21839, "train_loss": 3.4688291549682617, "perplexity": 32.09913731870896, "lr": 0.0026291804804649314, "grad_norm": 0.17259, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:42.109724+00:00", "epoch": 0, "step": 21840, "train_loss": 3.4674532413482666, "perplexity": 32.055002048604536, "lr": 0.0026291804804649314, "grad_norm": 0.182026, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:42.416299+00:00", "epoch": 0, "step": 21841, "train_loss": 3.4825382232666016, "perplexity": 32.54221674832282, "lr": 0.0026291804804649314, "grad_norm": 0.190012, "tokens_per_sec": 106791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:42.722096+00:00", "epoch": 0, "step": 21842, "train_loss": 3.5671916007995605, "perplexity": 35.41698831036354, "lr": 0.0026291804804649314, "grad_norm": 0.15887, "tokens_per_sec": 107156} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:43.027024+00:00", "epoch": 0, "step": 21843, "train_loss": 3.564692258834839, "perplexity": 35.32857967294609, "lr": 0.0026291804804649314, "grad_norm": 0.175671, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:43.331051+00:00", "epoch": 0, "step": 21844, "train_loss": 3.46889591217041, "perplexity": 32.10128023883462, "lr": 0.0026291804804649314, "grad_norm": 0.16928, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:43.635850+00:00", "epoch": 0, "step": 21845, "train_loss": 3.406475305557251, "perplexity": 30.15875630015543, "lr": 0.0026291804804649314, "grad_norm": 0.179018, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:43.941360+00:00", "epoch": 0, "step": 21846, "train_loss": 3.5688042640686035, "perplexity": 35.47415006545491, "lr": 0.0026291804804649314, "grad_norm": 0.189676, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:44.246467+00:00", "epoch": 0, "step": 21847, "train_loss": 3.4692342281341553, "perplexity": 32.11214245172248, "lr": 0.0026291804804649314, "grad_norm": 0.174434, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:44.550675+00:00", "epoch": 0, "step": 21848, "train_loss": 3.5002682209014893, "perplexity": 33.124335406379814, "lr": 0.0026291804804649314, "grad_norm": 0.172234, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:44.855915+00:00", "epoch": 0, "step": 21849, "train_loss": 3.4218297004699707, "perplexity": 30.625399095505987, "lr": 0.0026291804804649314, "grad_norm": 0.181196, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:45.161851+00:00", "epoch": 0, "step": 21850, "train_loss": 3.5249056816101074, "perplexity": 33.95057130201022, "lr": 0.0026291804804649314, "grad_norm": 0.198693, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:45.466675+00:00", "epoch": 0, "step": 21851, "train_loss": 3.500093460083008, "perplexity": 33.11854707621393, "lr": 0.0026291804804649314, "grad_norm": 0.174998, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:45.770878+00:00", "epoch": 0, "step": 21852, "train_loss": 3.498447895050049, "perplexity": 33.06409316921345, "lr": 0.0026291804804649314, "grad_norm": 0.161816, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:46.075219+00:00", "epoch": 0, "step": 21853, "train_loss": 3.444251298904419, "perplexity": 31.31982547861273, "lr": 0.0026291804804649314, "grad_norm": 0.161732, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:46.380913+00:00", "epoch": 0, "step": 21854, "train_loss": 3.3537142276763916, "perplexity": 28.60879613410635, "lr": 0.0026291804804649314, "grad_norm": 0.185343, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:46.684843+00:00", "epoch": 0, "step": 21855, "train_loss": 3.4556987285614014, "perplexity": 31.68041696271273, "lr": 0.0026291804804649314, "grad_norm": 0.168166, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:46.989199+00:00", "epoch": 0, "step": 21856, "train_loss": 3.57540225982666, "perplexity": 35.70898221595142, "lr": 0.0026291804804649314, "grad_norm": 0.182095, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:47.294677+00:00", "epoch": 0, "step": 21857, "train_loss": 3.4839141368865967, "perplexity": 32.587022845162444, "lr": 0.0026291804804649314, "grad_norm": 0.183487, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:47.598629+00:00", "epoch": 0, "step": 21858, "train_loss": 3.5208394527435303, "perplexity": 33.81280080196455, "lr": 0.0026291804804649314, "grad_norm": 0.16511, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:47.904351+00:00", "epoch": 0, "step": 21859, "train_loss": 3.484423875808716, "perplexity": 32.603637953386524, "lr": 0.0026291804804649314, "grad_norm": 0.193733, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:48.209527+00:00", "epoch": 0, "step": 21860, "train_loss": 3.5559754371643066, "perplexity": 35.02196503806609, "lr": 0.0026291804804649314, "grad_norm": 0.176167, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:48.514424+00:00", "epoch": 0, "step": 21861, "train_loss": 3.4278881549835205, "perplexity": 30.811504870322445, "lr": 0.0026291804804649314, "grad_norm": 0.179173, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:48.819047+00:00", "epoch": 0, "step": 21862, "train_loss": 3.538095712661743, "perplexity": 34.40134672727509, "lr": 0.0026291804804649314, "grad_norm": 0.175978, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:49.123088+00:00", "epoch": 0, "step": 21863, "train_loss": 3.583120822906494, "perplexity": 35.9856706927631, "lr": 0.0026291804804649314, "grad_norm": 0.175894, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:49.428184+00:00", "epoch": 0, "step": 21864, "train_loss": 3.5438013076782227, "perplexity": 34.598187893717615, "lr": 0.0026291804804649314, "grad_norm": 0.193909, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:49.733513+00:00", "epoch": 0, "step": 21865, "train_loss": 3.519990921020508, "perplexity": 33.78412173710876, "lr": 0.0026291804804649314, "grad_norm": 0.212067, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:50.038717+00:00", "epoch": 0, "step": 21866, "train_loss": 3.5501818656921387, "perplexity": 34.81964941144495, "lr": 0.0026291804804649314, "grad_norm": 0.233714, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:50.343321+00:00", "epoch": 0, "step": 21867, "train_loss": 3.4926705360412598, "perplexity": 32.87362077616951, "lr": 0.0026291804804649314, "grad_norm": 0.165692, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:50.648939+00:00", "epoch": 0, "step": 21868, "train_loss": 3.4715256690979004, "perplexity": 32.185809900441, "lr": 0.0026291804804649314, "grad_norm": 0.185036, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:50.953515+00:00", "epoch": 0, "step": 21869, "train_loss": 3.448789358139038, "perplexity": 31.46227969020295, "lr": 0.0026291804804649314, "grad_norm": 0.238962, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:51.258591+00:00", "epoch": 0, "step": 21870, "train_loss": 3.4274253845214844, "perplexity": 30.797249514710735, "lr": 0.0026291804804649314, "grad_norm": 0.1685, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:51.563968+00:00", "epoch": 0, "step": 21871, "train_loss": 3.425030469894409, "perplexity": 30.72358098150933, "lr": 0.0026291804804649314, "grad_norm": 0.208993, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:51.867787+00:00", "epoch": 0, "step": 21872, "train_loss": 3.5389347076416016, "perplexity": 34.43022139563677, "lr": 0.0026291804804649314, "grad_norm": 0.17025, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:52.171832+00:00", "epoch": 0, "step": 21873, "train_loss": 3.6025118827819824, "perplexity": 36.69028047465074, "lr": 0.0026291804804649314, "grad_norm": 0.180382, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:52.476262+00:00", "epoch": 0, "step": 21874, "train_loss": 3.536968946456909, "perplexity": 34.36260628219291, "lr": 0.0026291804804649314, "grad_norm": 0.171098, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:52.781457+00:00", "epoch": 0, "step": 21875, "train_loss": 3.428312301635742, "perplexity": 30.824576238855744, "lr": 0.0026291804804649314, "grad_norm": 0.169432, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:53.086661+00:00", "epoch": 0, "step": 21876, "train_loss": 3.547529697418213, "perplexity": 34.727424194423335, "lr": 0.0026291804804649314, "grad_norm": 0.161277, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:53.392321+00:00", "epoch": 0, "step": 21877, "train_loss": 3.5116569995880127, "perplexity": 33.50373749383417, "lr": 0.0026291804804649314, "grad_norm": 0.145598, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:53.696798+00:00", "epoch": 0, "step": 21878, "train_loss": 3.592670202255249, "perplexity": 36.33095752690254, "lr": 0.0026291804804649314, "grad_norm": 0.172652, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:54.001253+00:00", "epoch": 0, "step": 21879, "train_loss": 3.3929708003997803, "perplexity": 29.754214933696634, "lr": 0.0026291804804649314, "grad_norm": 0.165792, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:54.306445+00:00", "epoch": 0, "step": 21880, "train_loss": 3.511138916015625, "perplexity": 33.48638425342757, "lr": 0.0026291804804649314, "grad_norm": 0.163289, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:54.612506+00:00", "epoch": 0, "step": 21881, "train_loss": 3.3621585369110107, "perplexity": 28.851400527149917, "lr": 0.0026291804804649314, "grad_norm": 0.152738, "tokens_per_sec": 107007} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:54.916609+00:00", "epoch": 0, "step": 21882, "train_loss": 3.42362904548645, "perplexity": 30.68055436154163, "lr": 0.0026291804804649314, "grad_norm": 0.15677, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:55.221386+00:00", "epoch": 0, "step": 21883, "train_loss": 3.4836854934692383, "perplexity": 32.57957288862181, "lr": 0.0026291804804649314, "grad_norm": 0.168537, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:55.525441+00:00", "epoch": 0, "step": 21884, "train_loss": 3.4935593605041504, "perplexity": 32.90285264357413, "lr": 0.0026291804804649314, "grad_norm": 0.161579, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:55.829494+00:00", "epoch": 0, "step": 21885, "train_loss": 3.5487780570983887, "perplexity": 34.77080358148935, "lr": 0.0026291804804649314, "grad_norm": 0.167824, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:56.135190+00:00", "epoch": 0, "step": 21886, "train_loss": 3.549771547317505, "perplexity": 34.8053652002307, "lr": 0.0026291804804649314, "grad_norm": 0.157958, "tokens_per_sec": 107192} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:56.440396+00:00", "epoch": 0, "step": 21887, "train_loss": 3.472661018371582, "perplexity": 32.22237278823086, "lr": 0.0026291804804649314, "grad_norm": 0.169065, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:56.745860+00:00", "epoch": 0, "step": 21888, "train_loss": 3.4867115020751953, "perplexity": 32.67830826827438, "lr": 0.0026291804804649314, "grad_norm": 0.17917, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:57.049961+00:00", "epoch": 0, "step": 21889, "train_loss": 3.4893174171447754, "perplexity": 32.7635762165016, "lr": 0.0026291804804649314, "grad_norm": 0.188436, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:57.353786+00:00", "epoch": 0, "step": 21890, "train_loss": 3.4307010173797607, "perplexity": 30.898295401429106, "lr": 0.0026291804804649314, "grad_norm": 0.200031, "tokens_per_sec": 107919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:57.659505+00:00", "epoch": 0, "step": 21891, "train_loss": 3.466418504714966, "perplexity": 32.02185071811024, "lr": 0.0026291804804649314, "grad_norm": 0.187215, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:57.964231+00:00", "epoch": 0, "step": 21892, "train_loss": 3.5566742420196533, "perplexity": 35.0464471103777, "lr": 0.0026291804804649314, "grad_norm": 0.159919, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:58.267752+00:00", "epoch": 0, "step": 21893, "train_loss": 3.387096643447876, "perplexity": 29.579946346815934, "lr": 0.0026291804804649314, "grad_norm": 0.179739, "tokens_per_sec": 107959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:58.572280+00:00", "epoch": 0, "step": 21894, "train_loss": 3.4910075664520264, "perplexity": 32.81899837481496, "lr": 0.0026291804804649314, "grad_norm": 0.171254, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:58.876476+00:00", "epoch": 0, "step": 21895, "train_loss": 3.4123566150665283, "perplexity": 30.336651897123748, "lr": 0.0026291804804649314, "grad_norm": 0.161526, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:59.180766+00:00", "epoch": 0, "step": 21896, "train_loss": 3.526700258255005, "perplexity": 34.01155290606497, "lr": 0.0026291804804649314, "grad_norm": 0.165582, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:59.485811+00:00", "epoch": 0, "step": 21897, "train_loss": 3.5302915573120117, "perplexity": 34.133918157533564, "lr": 0.0026291804804649314, "grad_norm": 0.191759, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:52:59.789204+00:00", "epoch": 0, "step": 21898, "train_loss": 3.5131287574768066, "perplexity": 33.55308318734447, "lr": 0.0026291804804649314, "grad_norm": 0.172301, "tokens_per_sec": 108005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:00.093258+00:00", "epoch": 0, "step": 21899, "train_loss": 3.4846179485321045, "perplexity": 32.60996604423157, "lr": 0.0026291804804649314, "grad_norm": 0.199494, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:00.398116+00:00", "epoch": 0, "step": 21900, "train_loss": 3.5434980392456055, "perplexity": 34.58769694637068, "lr": 0.0026291804804649314, "grad_norm": 0.191743, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:00.703000+00:00", "epoch": 0, "step": 21901, "train_loss": 3.443216323852539, "perplexity": 31.287427009308892, "lr": 0.0026291804804649314, "grad_norm": 0.186141, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:01.007156+00:00", "epoch": 0, "step": 21902, "train_loss": 3.581895112991333, "perplexity": 35.94158972015267, "lr": 0.0026291804804649314, "grad_norm": 0.152374, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:01.311515+00:00", "epoch": 0, "step": 21903, "train_loss": 3.49428391456604, "perplexity": 32.926701177837636, "lr": 0.0026291804804649314, "grad_norm": 0.186088, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:01.616150+00:00", "epoch": 0, "step": 21904, "train_loss": 3.424711227416992, "perplexity": 30.713774274843622, "lr": 0.0026291804804649314, "grad_norm": 0.1926, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:01.920028+00:00", "epoch": 0, "step": 21905, "train_loss": 3.5378832817077637, "perplexity": 34.39403959252992, "lr": 0.0026291804804649314, "grad_norm": 0.230145, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:02.224471+00:00", "epoch": 0, "step": 21906, "train_loss": 3.563143253326416, "perplexity": 35.2738978705682, "lr": 0.0026291804804649314, "grad_norm": 0.193881, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:02.530068+00:00", "epoch": 0, "step": 21907, "train_loss": 3.4962852001190186, "perplexity": 32.99266289130868, "lr": 0.0026291804804649314, "grad_norm": 0.162356, "tokens_per_sec": 107226} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:02.835575+00:00", "epoch": 0, "step": 21908, "train_loss": 3.553473711013794, "perplexity": 34.93445917577895, "lr": 0.0026291804804649314, "grad_norm": 0.18144, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:03.139795+00:00", "epoch": 0, "step": 21909, "train_loss": 3.5717837810516357, "perplexity": 35.58000351580039, "lr": 0.0026291804804649314, "grad_norm": 0.162561, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:03.444262+00:00", "epoch": 0, "step": 21910, "train_loss": 3.5127768516540527, "perplexity": 33.54127773932427, "lr": 0.0026291804804649314, "grad_norm": 0.170012, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:03.748960+00:00", "epoch": 0, "step": 21911, "train_loss": 3.4975109100341797, "perplexity": 33.03312711897866, "lr": 0.0026291804804649314, "grad_norm": 0.159384, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:04.054288+00:00", "epoch": 0, "step": 21912, "train_loss": 3.4546890258789062, "perplexity": 31.648445304373766, "lr": 0.0026291804804649314, "grad_norm": 0.170313, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:04.358761+00:00", "epoch": 0, "step": 21913, "train_loss": 3.4993515014648438, "perplexity": 33.09398359845852, "lr": 0.0026291804804649314, "grad_norm": 0.176464, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:04.663128+00:00", "epoch": 0, "step": 21914, "train_loss": 3.467672824859619, "perplexity": 32.062041571364055, "lr": 0.0026291804804649314, "grad_norm": 0.164245, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:04.968146+00:00", "epoch": 0, "step": 21915, "train_loss": 3.5792903900146484, "perplexity": 35.84809365412669, "lr": 0.0026291804804649314, "grad_norm": 0.167346, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:05.272296+00:00", "epoch": 0, "step": 21916, "train_loss": 3.483093500137329, "perplexity": 32.56029170644404, "lr": 0.0026291804804649314, "grad_norm": 0.164897, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:05.577589+00:00", "epoch": 0, "step": 21917, "train_loss": 3.517334461212158, "perplexity": 33.69449467360623, "lr": 0.0026291804804649314, "grad_norm": 0.175747, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:05.882480+00:00", "epoch": 0, "step": 21918, "train_loss": 3.468196392059326, "perplexity": 32.0788325999188, "lr": 0.0026291804804649314, "grad_norm": 0.200287, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:06.187552+00:00", "epoch": 0, "step": 21919, "train_loss": 3.4932477474212646, "perplexity": 32.89260128153834, "lr": 0.0026291804804649314, "grad_norm": 0.180953, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:06.492857+00:00", "epoch": 0, "step": 21920, "train_loss": 3.507357597351074, "perplexity": 33.36000066254723, "lr": 0.0026291804804649314, "grad_norm": 0.16096, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:06.797766+00:00", "epoch": 0, "step": 21921, "train_loss": 3.449500322341919, "perplexity": 31.48465619830928, "lr": 0.0026291804804649314, "grad_norm": 0.202305, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:07.101742+00:00", "epoch": 0, "step": 21922, "train_loss": 3.585559606552124, "perplexity": 36.07353906034358, "lr": 0.0026291804804649314, "grad_norm": 0.200049, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:07.406988+00:00", "epoch": 0, "step": 21923, "train_loss": 3.505221128463745, "perplexity": 33.288804140708685, "lr": 0.0026291804804649314, "grad_norm": 0.181032, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:07.712140+00:00", "epoch": 0, "step": 21924, "train_loss": 3.4535205364227295, "perplexity": 31.611486027199312, "lr": 0.0026291804804649314, "grad_norm": 0.166914, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:08.016400+00:00", "epoch": 0, "step": 21925, "train_loss": 3.5713047981262207, "perplexity": 35.562965382443345, "lr": 0.0026291804804649314, "grad_norm": 0.16407, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:08.320662+00:00", "epoch": 0, "step": 21926, "train_loss": 3.563912868499756, "perplexity": 35.30105564677007, "lr": 0.0026291804804649314, "grad_norm": 0.183184, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:08.625106+00:00", "epoch": 0, "step": 21927, "train_loss": 3.432314157485962, "perplexity": 30.94817890467694, "lr": 0.0026291804804649314, "grad_norm": 0.171576, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:08.929412+00:00", "epoch": 0, "step": 21928, "train_loss": 3.478727102279663, "perplexity": 32.41843045521085, "lr": 0.0026291804804649314, "grad_norm": 0.166916, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:09.234303+00:00", "epoch": 0, "step": 21929, "train_loss": 3.55108642578125, "perplexity": 34.851160126138254, "lr": 0.0026291804804649314, "grad_norm": 0.174857, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:09.538442+00:00", "epoch": 0, "step": 21930, "train_loss": 3.4551992416381836, "perplexity": 31.664596959988916, "lr": 0.0026291804804649314, "grad_norm": 0.156548, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:09.842803+00:00", "epoch": 0, "step": 21931, "train_loss": 3.570514440536499, "perplexity": 35.53486902738662, "lr": 0.0026291804804649314, "grad_norm": 0.178175, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:10.146699+00:00", "epoch": 0, "step": 21932, "train_loss": 3.4854745864868164, "perplexity": 32.63791294732686, "lr": 0.0026291804804649314, "grad_norm": 0.16557, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:10.452241+00:00", "epoch": 0, "step": 21933, "train_loss": 3.5847864151000977, "perplexity": 36.04565808835139, "lr": 0.0026291804804649314, "grad_norm": 0.158296, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:10.757732+00:00", "epoch": 0, "step": 21934, "train_loss": 3.424650192260742, "perplexity": 30.71189971203943, "lr": 0.0026291804804649314, "grad_norm": 0.180252, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:11.062173+00:00", "epoch": 0, "step": 21935, "train_loss": 3.4016637802124023, "perplexity": 30.013995219928855, "lr": 0.0026291804804649314, "grad_norm": 0.175046, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:11.366740+00:00", "epoch": 0, "step": 21936, "train_loss": 3.5815329551696777, "perplexity": 35.92857554904696, "lr": 0.0026291804804649314, "grad_norm": 0.163759, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:11.671498+00:00", "epoch": 0, "step": 21937, "train_loss": 3.441695213317871, "perplexity": 31.239871552205162, "lr": 0.0026291804804649314, "grad_norm": 0.18925, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:11.975447+00:00", "epoch": 0, "step": 21938, "train_loss": 3.528815269470215, "perplexity": 34.08356384703818, "lr": 0.0026291804804649314, "grad_norm": 0.162599, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:12.280759+00:00", "epoch": 0, "step": 21939, "train_loss": 3.4436216354370117, "perplexity": 31.300110736181185, "lr": 0.0026291804804649314, "grad_norm": 0.1952, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:12.585719+00:00", "epoch": 0, "step": 21940, "train_loss": 3.5494697093963623, "perplexity": 34.79486120648562, "lr": 0.0026291804804649314, "grad_norm": 0.198535, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:12.889974+00:00", "epoch": 0, "step": 21941, "train_loss": 3.464735269546509, "perplexity": 31.967995750825416, "lr": 0.0026291804804649314, "grad_norm": 0.179809, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:13.195166+00:00", "epoch": 0, "step": 21942, "train_loss": 3.426426649093628, "perplexity": 30.766506565131746, "lr": 0.0026291804804649314, "grad_norm": 0.183517, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:13.501965+00:00", "epoch": 0, "step": 21943, "train_loss": 3.4747602939605713, "perplexity": 32.290087479876114, "lr": 0.0026291804804649314, "grad_norm": 0.165174, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:13.806970+00:00", "epoch": 0, "step": 21944, "train_loss": 3.520596742630005, "perplexity": 33.80459508908833, "lr": 0.0026291804804649314, "grad_norm": 0.167228, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:14.112851+00:00", "epoch": 0, "step": 21945, "train_loss": 3.5128705501556396, "perplexity": 33.54442065403066, "lr": 0.0026291804804649314, "grad_norm": 0.178124, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:14.416959+00:00", "epoch": 0, "step": 21946, "train_loss": 3.530461549758911, "perplexity": 34.13972115902266, "lr": 0.0026291804804649314, "grad_norm": 0.156009, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:14.722180+00:00", "epoch": 0, "step": 21947, "train_loss": 3.4430744647979736, "perplexity": 31.282988919292634, "lr": 0.0026291804804649314, "grad_norm": 0.191218, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:15.026718+00:00", "epoch": 0, "step": 21948, "train_loss": 3.464754343032837, "perplexity": 31.968605497770287, "lr": 0.0026291804804649314, "grad_norm": 0.166486, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:15.332278+00:00", "epoch": 0, "step": 21949, "train_loss": 3.527865409851074, "perplexity": 34.051204616859465, "lr": 0.0026291804804649314, "grad_norm": 0.163829, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:15.637862+00:00", "epoch": 0, "step": 21950, "train_loss": 3.4408349990844727, "perplexity": 31.213010124993808, "lr": 0.0026291804804649314, "grad_norm": 0.188963, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:15.942124+00:00", "epoch": 0, "step": 21951, "train_loss": 3.530586004257202, "perplexity": 34.14397026529601, "lr": 0.0026291804804649314, "grad_norm": 0.1523, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:16.246076+00:00", "epoch": 0, "step": 21952, "train_loss": 3.4533164501190186, "perplexity": 31.6050352141449, "lr": 0.0026291804804649314, "grad_norm": 0.174848, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:16.550546+00:00", "epoch": 0, "step": 21953, "train_loss": 3.4418230056762695, "perplexity": 31.243864024165152, "lr": 0.0026291804804649314, "grad_norm": 0.169449, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:16.854609+00:00", "epoch": 0, "step": 21954, "train_loss": 3.5032763481140137, "perplexity": 33.22412763972867, "lr": 0.0026291804804649314, "grad_norm": 0.166628, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:17.159886+00:00", "epoch": 0, "step": 21955, "train_loss": 3.4838180541992188, "perplexity": 32.58389194684881, "lr": 0.0026291804804649314, "grad_norm": 0.154395, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:17.464904+00:00", "epoch": 0, "step": 21956, "train_loss": 3.419513702392578, "perplexity": 30.55455280168894, "lr": 0.0026291804804649314, "grad_norm": 0.151034, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:17.768989+00:00", "epoch": 0, "step": 21957, "train_loss": 3.626333475112915, "perplexity": 37.574794821291114, "lr": 0.0026291804804649314, "grad_norm": 0.15368, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:18.073108+00:00", "epoch": 0, "step": 21958, "train_loss": 3.5098376274108887, "perplexity": 33.44283714300909, "lr": 0.0026291804804649314, "grad_norm": 0.16248, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:18.377274+00:00", "epoch": 0, "step": 21959, "train_loss": 3.4656240940093994, "perplexity": 31.99642231872137, "lr": 0.0026291804804649314, "grad_norm": 0.155914, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:18.681994+00:00", "epoch": 0, "step": 21960, "train_loss": 3.5687735080718994, "perplexity": 35.4730590393903, "lr": 0.0026291804804649314, "grad_norm": 0.193576, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:18.986670+00:00", "epoch": 0, "step": 21961, "train_loss": 3.431699752807617, "perplexity": 30.929170038937652, "lr": 0.0026291804804649314, "grad_norm": 0.199186, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:19.291129+00:00", "epoch": 0, "step": 21962, "train_loss": 3.459873676300049, "perplexity": 31.81295753032493, "lr": 0.0026291804804649314, "grad_norm": 0.168208, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:19.595410+00:00", "epoch": 0, "step": 21963, "train_loss": 3.447449207305908, "perplexity": 31.420143730441115, "lr": 0.0026291804804649314, "grad_norm": 0.162402, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:19.899575+00:00", "epoch": 0, "step": 21964, "train_loss": 3.479038715362549, "perplexity": 32.4285340363896, "lr": 0.0026291804804649314, "grad_norm": 0.174864, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:20.203926+00:00", "epoch": 0, "step": 21965, "train_loss": 3.5498263835906982, "perplexity": 34.80727384907653, "lr": 0.0026291804804649314, "grad_norm": 0.180348, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:20.509156+00:00", "epoch": 0, "step": 21966, "train_loss": 3.498532772064209, "perplexity": 33.06689966981958, "lr": 0.0026291804804649314, "grad_norm": 0.176229, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:20.814144+00:00", "epoch": 0, "step": 21967, "train_loss": 3.5474355220794678, "perplexity": 34.72415388147991, "lr": 0.0026291804804649314, "grad_norm": 0.154945, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:21.118324+00:00", "epoch": 0, "step": 21968, "train_loss": 3.5087389945983887, "perplexity": 33.40611592004444, "lr": 0.0026291804804649314, "grad_norm": 0.170871, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:21.423224+00:00", "epoch": 0, "step": 21969, "train_loss": 3.4523046016693115, "perplexity": 31.573071882011664, "lr": 0.0026291804804649314, "grad_norm": 0.178233, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:21.727665+00:00", "epoch": 0, "step": 21970, "train_loss": 3.614548444747925, "perplexity": 37.13457382732167, "lr": 0.0026291804804649314, "grad_norm": 0.183303, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:22.032546+00:00", "epoch": 0, "step": 21971, "train_loss": 3.5178797245025635, "perplexity": 33.71287205444076, "lr": 0.0026291804804649314, "grad_norm": 0.150978, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:22.337621+00:00", "epoch": 0, "step": 21972, "train_loss": 3.4925334453582764, "perplexity": 32.869114417942086, "lr": 0.0026291804804649314, "grad_norm": 0.158024, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:22.642364+00:00", "epoch": 0, "step": 21973, "train_loss": 3.477965831756592, "perplexity": 32.39376065108982, "lr": 0.0026291804804649314, "grad_norm": 0.150413, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:22.947932+00:00", "epoch": 0, "step": 21974, "train_loss": 3.491353988647461, "perplexity": 32.83036957378806, "lr": 0.0026291804804649314, "grad_norm": 0.162601, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:23.252819+00:00", "epoch": 0, "step": 21975, "train_loss": 3.4567744731903076, "perplexity": 31.714515338385826, "lr": 0.0026291804804649314, "grad_norm": 0.160612, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:23.558043+00:00", "epoch": 0, "step": 21976, "train_loss": 3.4378511905670166, "perplexity": 31.12001528791617, "lr": 0.0026291804804649314, "grad_norm": 0.170132, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:23.862581+00:00", "epoch": 0, "step": 21977, "train_loss": 3.4124176502227783, "perplexity": 30.338503555919957, "lr": 0.0026291804804649314, "grad_norm": 0.199442, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:24.167749+00:00", "epoch": 0, "step": 21978, "train_loss": 3.582943916320801, "perplexity": 35.97930515369645, "lr": 0.0026291804804649314, "grad_norm": 0.160602, "tokens_per_sec": 107379} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:24.472791+00:00", "epoch": 0, "step": 21979, "train_loss": 3.532604694366455, "perplexity": 34.212965977411585, "lr": 0.0026291804804649314, "grad_norm": 0.176254, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:24.777378+00:00", "epoch": 0, "step": 21980, "train_loss": 3.564976930618286, "perplexity": 35.338638154343236, "lr": 0.0026291804804649314, "grad_norm": 0.190897, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:25.082204+00:00", "epoch": 0, "step": 21981, "train_loss": 3.4981911182403564, "perplexity": 33.05560416679021, "lr": 0.0026291804804649314, "grad_norm": 0.164622, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:25.388082+00:00", "epoch": 0, "step": 21982, "train_loss": 3.5650646686553955, "perplexity": 35.34173883311077, "lr": 0.0026291804804649314, "grad_norm": 0.173525, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:25.692737+00:00", "epoch": 0, "step": 21983, "train_loss": 3.4200897216796875, "perplexity": 30.57215788335823, "lr": 0.0026291804804649314, "grad_norm": 0.150724, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:25.997013+00:00", "epoch": 0, "step": 21984, "train_loss": 3.531399965286255, "perplexity": 34.17177344028452, "lr": 0.0026291804804649314, "grad_norm": 0.181677, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:26.300870+00:00", "epoch": 0, "step": 21985, "train_loss": 3.5832765102386475, "perplexity": 35.991273641971794, "lr": 0.0026291804804649314, "grad_norm": 0.187051, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:26.605311+00:00", "epoch": 0, "step": 21986, "train_loss": 3.5161619186401367, "perplexity": 33.65500959764879, "lr": 0.0026291804804649314, "grad_norm": 0.181579, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:26.908962+00:00", "epoch": 0, "step": 21987, "train_loss": 3.410073757171631, "perplexity": 30.26747662054068, "lr": 0.0026291804804649314, "grad_norm": 0.203944, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:27.214060+00:00", "epoch": 0, "step": 21988, "train_loss": 3.516002893447876, "perplexity": 33.649658028805405, "lr": 0.0026291804804649314, "grad_norm": 0.17909, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:27.518071+00:00", "epoch": 0, "step": 21989, "train_loss": 3.5110511779785156, "perplexity": 33.48344635268799, "lr": 0.0026291804804649314, "grad_norm": 0.172875, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:27.822064+00:00", "epoch": 0, "step": 21990, "train_loss": 3.464287042617798, "perplexity": 31.953670045096363, "lr": 0.0026291804804649314, "grad_norm": 0.194357, "tokens_per_sec": 107846} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:28.126954+00:00", "epoch": 0, "step": 21991, "train_loss": 3.5129337310791016, "perplexity": 33.54654008845778, "lr": 0.0026291804804649314, "grad_norm": 0.207249, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:28.431666+00:00", "epoch": 0, "step": 21992, "train_loss": 3.554976463317871, "perplexity": 34.986996480187734, "lr": 0.0026291804804649314, "grad_norm": 0.175715, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:28.737417+00:00", "epoch": 0, "step": 21993, "train_loss": 3.5786378383636475, "perplexity": 35.8247085522507, "lr": 0.0026291804804649314, "grad_norm": 0.226372, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:29.042995+00:00", "epoch": 0, "step": 21994, "train_loss": 3.437535524368286, "perplexity": 31.11019330130195, "lr": 0.0026291804804649314, "grad_norm": 0.168617, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:29.347132+00:00", "epoch": 0, "step": 21995, "train_loss": 3.517319440841675, "perplexity": 33.6939885736139, "lr": 0.0026291804804649314, "grad_norm": 0.19056, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:29.651173+00:00", "epoch": 0, "step": 21996, "train_loss": 3.402141571044922, "perplexity": 30.02833905809654, "lr": 0.0026291804804649314, "grad_norm": 0.173631, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:29.955938+00:00", "epoch": 0, "step": 21997, "train_loss": 3.474055290222168, "perplexity": 32.26733087017944, "lr": 0.0026291804804649314, "grad_norm": 0.165697, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:30.261126+00:00", "epoch": 0, "step": 21998, "train_loss": 3.55560302734375, "perplexity": 35.00892494263123, "lr": 0.0026291804804649314, "grad_norm": 0.185846, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:30.566125+00:00", "epoch": 0, "step": 21999, "train_loss": 3.411114454269409, "perplexity": 30.298992291906192, "lr": 0.0026291804804649314, "grad_norm": 0.163925, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:30.871043+00:00", "epoch": 0, "step": 22000, "train_loss": 3.5466692447662354, "perplexity": 34.69755574221618, "lr": 0.0026291804804649314, "grad_norm": 0.181608, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:53:33.944450+00:00", "step": 22000, "epoch": 0, "val_loss": 3.4509812951087953, "val_ppl": 31.531318661103104, "eval_train_loss": 3.5466692447662354, "eval_train_ppl": 34.69755574221618} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:53:34.914197+00:00", "step": 22000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4510_epoch_0000_step_0022000.pt", "val_loss": 3.4509812951087953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:36.205314+00:00", "epoch": 0, "step": 22001, "train_loss": 3.5285568237304688, "perplexity": 34.07475623336069, "lr": 0.0026291804804649314, "grad_norm": 0.16101, "tokens_per_sec": 6143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:36.508370+00:00", "epoch": 0, "step": 22002, "train_loss": 3.508920669555664, "perplexity": 33.41218552605797, "lr": 0.0026291804804649314, "grad_norm": 0.165065, "tokens_per_sec": 108118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:36.813322+00:00", "epoch": 0, "step": 22003, "train_loss": 3.5564281940460205, "perplexity": 35.0378250638452, "lr": 0.0026291804804649314, "grad_norm": 0.18385, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:37.117881+00:00", "epoch": 0, "step": 22004, "train_loss": 3.598483085632324, "perplexity": 36.54276014153911, "lr": 0.0026291804804649314, "grad_norm": 0.166271, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:37.420961+00:00", "epoch": 0, "step": 22005, "train_loss": 3.544766426086426, "perplexity": 34.63159536023164, "lr": 0.0026291804804649314, "grad_norm": 0.174919, "tokens_per_sec": 108117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:37.725790+00:00", "epoch": 0, "step": 22006, "train_loss": 3.5086874961853027, "perplexity": 33.40439560238439, "lr": 0.0026291804804649314, "grad_norm": 0.18834, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:38.031338+00:00", "epoch": 0, "step": 22007, "train_loss": 3.459688186645508, "perplexity": 31.807057103073724, "lr": 0.0026291804804649314, "grad_norm": 0.173743, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:38.335796+00:00", "epoch": 0, "step": 22008, "train_loss": 3.588197946548462, "perplexity": 36.168838982773835, "lr": 0.0026291804804649314, "grad_norm": 0.177384, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:38.640495+00:00", "epoch": 0, "step": 22009, "train_loss": 3.405863046646118, "perplexity": 30.140296984378907, "lr": 0.0026291804804649314, "grad_norm": 0.186318, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:38.944542+00:00", "epoch": 0, "step": 22010, "train_loss": 3.6079704761505127, "perplexity": 36.891105408513184, "lr": 0.0026291804804649314, "grad_norm": 0.189402, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:39.248015+00:00", "epoch": 0, "step": 22011, "train_loss": 3.4307615756988525, "perplexity": 30.900166606919374, "lr": 0.0026291804804649314, "grad_norm": 0.172772, "tokens_per_sec": 107919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:39.553080+00:00", "epoch": 0, "step": 22012, "train_loss": 3.4792895317077637, "perplexity": 32.43666866288328, "lr": 0.0026291804804649314, "grad_norm": 0.1712, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:39.857306+00:00", "epoch": 0, "step": 22013, "train_loss": 3.4126839637756348, "perplexity": 30.34658418653325, "lr": 0.0026291804804649314, "grad_norm": 0.170554, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:40.160802+00:00", "epoch": 0, "step": 22014, "train_loss": 3.4463131427764893, "perplexity": 31.384468688051815, "lr": 0.0026291804804649314, "grad_norm": 0.164827, "tokens_per_sec": 107969} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:40.464825+00:00", "epoch": 0, "step": 22015, "train_loss": 3.5575051307678223, "perplexity": 35.07557890991124, "lr": 0.0026291804804649314, "grad_norm": 0.185714, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:40.768274+00:00", "epoch": 0, "step": 22016, "train_loss": 3.5131468772888184, "perplexity": 33.55369116841247, "lr": 0.0026291804804649314, "grad_norm": 0.209679, "tokens_per_sec": 107986} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:41.071319+00:00", "epoch": 0, "step": 22017, "train_loss": 3.5323355197906494, "perplexity": 34.203757956145765, "lr": 0.0026291804804649314, "grad_norm": 0.176921, "tokens_per_sec": 108129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:41.375023+00:00", "epoch": 0, "step": 22018, "train_loss": 3.5190110206604004, "perplexity": 33.75103287859252, "lr": 0.0026291804804649314, "grad_norm": 0.181519, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:41.680846+00:00", "epoch": 0, "step": 22019, "train_loss": 3.470313310623169, "perplexity": 32.14681280504819, "lr": 0.0026291804804649314, "grad_norm": 0.201387, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:41.985216+00:00", "epoch": 0, "step": 22020, "train_loss": 3.5367867946624756, "perplexity": 34.35634764182578, "lr": 0.0026291804804649314, "grad_norm": 0.188514, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:42.289214+00:00", "epoch": 0, "step": 22021, "train_loss": 3.4241456985473633, "perplexity": 30.69640965935442, "lr": 0.0026291804804649314, "grad_norm": 0.17251, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:42.593026+00:00", "epoch": 0, "step": 22022, "train_loss": 3.6059393882751465, "perplexity": 36.816252373887345, "lr": 0.0026291804804649314, "grad_norm": 0.205008, "tokens_per_sec": 107856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:42.897119+00:00", "epoch": 0, "step": 22023, "train_loss": 3.467785358428955, "perplexity": 32.065649830363604, "lr": 0.0026291804804649314, "grad_norm": 0.178946, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:43.201459+00:00", "epoch": 0, "step": 22024, "train_loss": 3.4899446964263916, "perplexity": 32.78413457629634, "lr": 0.0026291804804649314, "grad_norm": 0.188315, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:43.505406+00:00", "epoch": 0, "step": 22025, "train_loss": 3.496721029281616, "perplexity": 33.00704518983338, "lr": 0.0026291804804649314, "grad_norm": 0.183187, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:43.809465+00:00", "epoch": 0, "step": 22026, "train_loss": 3.57944655418396, "perplexity": 35.85369227903451, "lr": 0.0026291804804649314, "grad_norm": 0.164347, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:44.114259+00:00", "epoch": 0, "step": 22027, "train_loss": 3.4822723865509033, "perplexity": 32.53356698206458, "lr": 0.0026291804804649314, "grad_norm": 0.184873, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:44.418508+00:00", "epoch": 0, "step": 22028, "train_loss": 3.4403231143951416, "perplexity": 31.19703675161431, "lr": 0.0026291804804649314, "grad_norm": 0.185168, "tokens_per_sec": 107701} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:44.723211+00:00", "epoch": 0, "step": 22029, "train_loss": 3.484942674636841, "perplexity": 32.62055707097906, "lr": 0.0026291804804649314, "grad_norm": 0.15782, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:45.026688+00:00", "epoch": 0, "step": 22030, "train_loss": 3.437861442565918, "perplexity": 31.12033433191413, "lr": 0.0026291804804649314, "grad_norm": 0.194232, "tokens_per_sec": 108037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:45.330715+00:00", "epoch": 0, "step": 22031, "train_loss": 3.5456557273864746, "perplexity": 34.662406981385004, "lr": 0.0026291804804649314, "grad_norm": 0.161261, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:45.633928+00:00", "epoch": 0, "step": 22032, "train_loss": 3.485001802444458, "perplexity": 32.62248591002537, "lr": 0.0026291804804649314, "grad_norm": 0.170769, "tokens_per_sec": 108066} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:45.939272+00:00", "epoch": 0, "step": 22033, "train_loss": 3.5354504585266113, "perplexity": 34.31046667599981, "lr": 0.0026291804804649314, "grad_norm": 0.198558, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:46.243236+00:00", "epoch": 0, "step": 22034, "train_loss": 3.520190954208374, "perplexity": 33.790880358630865, "lr": 0.0026291804804649314, "grad_norm": 0.159067, "tokens_per_sec": 107802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:46.547149+00:00", "epoch": 0, "step": 22035, "train_loss": 3.5399532318115234, "perplexity": 34.46530727317799, "lr": 0.0026291804804649314, "grad_norm": 0.164866, "tokens_per_sec": 107821} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:46.851205+00:00", "epoch": 0, "step": 22036, "train_loss": 3.532008171081543, "perplexity": 34.192563232521366, "lr": 0.0026291804804649314, "grad_norm": 0.21127, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:47.156264+00:00", "epoch": 0, "step": 22037, "train_loss": 3.546090602874756, "perplexity": 34.67748409064105, "lr": 0.0026291804804649314, "grad_norm": 0.211531, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:47.461638+00:00", "epoch": 0, "step": 22038, "train_loss": 3.4621005058288574, "perplexity": 31.883878498698756, "lr": 0.0026291804804649314, "grad_norm": 0.188791, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:47.766723+00:00", "epoch": 0, "step": 22039, "train_loss": 3.4911818504333496, "perplexity": 32.82471869898072, "lr": 0.0026291804804649314, "grad_norm": 0.162721, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:48.070996+00:00", "epoch": 0, "step": 22040, "train_loss": 3.6119959354400635, "perplexity": 37.039908350637006, "lr": 0.0026291804804649314, "grad_norm": 0.194318, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:48.375298+00:00", "epoch": 0, "step": 22041, "train_loss": 3.477280855178833, "perplexity": 32.37157928150489, "lr": 0.0026291804804649314, "grad_norm": 0.214626, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:48.678990+00:00", "epoch": 0, "step": 22042, "train_loss": 3.522162914276123, "perplexity": 33.857580368615984, "lr": 0.0026291804804649314, "grad_norm": 0.226424, "tokens_per_sec": 107901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:48.983871+00:00", "epoch": 0, "step": 22043, "train_loss": 3.5457260608673096, "perplexity": 34.6648449948581, "lr": 0.0026291804804649314, "grad_norm": 0.208764, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:49.289635+00:00", "epoch": 0, "step": 22044, "train_loss": 3.547356128692627, "perplexity": 34.721397122733734, "lr": 0.0026291804804649314, "grad_norm": 0.191184, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:49.594737+00:00", "epoch": 0, "step": 22045, "train_loss": 3.539135694503784, "perplexity": 34.437142113262695, "lr": 0.0026291804804649314, "grad_norm": 0.233172, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:49.898416+00:00", "epoch": 0, "step": 22046, "train_loss": 3.48945951461792, "perplexity": 32.768232168684825, "lr": 0.0026291804804649314, "grad_norm": 0.184218, "tokens_per_sec": 107904} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:50.202387+00:00", "epoch": 0, "step": 22047, "train_loss": 3.462667226791382, "perplexity": 31.90195288209278, "lr": 0.0026291804804649314, "grad_norm": 0.220372, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:50.507113+00:00", "epoch": 0, "step": 22048, "train_loss": 3.331099271774292, "perplexity": 27.96907042020263, "lr": 0.0026291804804649314, "grad_norm": 0.217235, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:50.811668+00:00", "epoch": 0, "step": 22049, "train_loss": 3.53293776512146, "perplexity": 34.22436321375833, "lr": 0.0026291804804649314, "grad_norm": 0.200346, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:51.116157+00:00", "epoch": 0, "step": 22050, "train_loss": 3.5117013454437256, "perplexity": 33.50522327868688, "lr": 0.0026291804804649314, "grad_norm": 0.216465, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:51.420695+00:00", "epoch": 0, "step": 22051, "train_loss": 3.5003445148468018, "perplexity": 33.126862689020776, "lr": 0.0026291804804649314, "grad_norm": 0.158543, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:51.725911+00:00", "epoch": 0, "step": 22052, "train_loss": 3.6142776012420654, "perplexity": 37.12451753106043, "lr": 0.0026291804804649314, "grad_norm": 0.202013, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:52.030842+00:00", "epoch": 0, "step": 22053, "train_loss": 3.5267159938812256, "perplexity": 34.0120881033595, "lr": 0.0026291804804649314, "grad_norm": 0.154635, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:52.336298+00:00", "epoch": 0, "step": 22054, "train_loss": 3.5679049491882324, "perplexity": 35.44226197529869, "lr": 0.0026291804804649314, "grad_norm": 0.170132, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:52.641537+00:00", "epoch": 0, "step": 22055, "train_loss": 3.4775021076202393, "perplexity": 32.37874236484872, "lr": 0.0026291804804649314, "grad_norm": 0.171608, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:52.946166+00:00", "epoch": 0, "step": 22056, "train_loss": 3.484131097793579, "perplexity": 32.594093722219, "lr": 0.0026291804804649314, "grad_norm": 0.143999, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:53.251169+00:00", "epoch": 0, "step": 22057, "train_loss": 3.5972166061401367, "perplexity": 36.49650877961476, "lr": 0.0026291804804649314, "grad_norm": 0.180212, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:53.555550+00:00", "epoch": 0, "step": 22058, "train_loss": 3.518065929412842, "perplexity": 33.719150141244036, "lr": 0.0026291804804649314, "grad_norm": 0.148567, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:53.859584+00:00", "epoch": 0, "step": 22059, "train_loss": 3.4477195739746094, "perplexity": 31.42863983851209, "lr": 0.0026291804804649314, "grad_norm": 0.159472, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:54.163426+00:00", "epoch": 0, "step": 22060, "train_loss": 3.461655378341675, "perplexity": 31.86968926622398, "lr": 0.0026291804804649314, "grad_norm": 0.173881, "tokens_per_sec": 107933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:54.467988+00:00", "epoch": 0, "step": 22061, "train_loss": 3.506560802459717, "perplexity": 33.33343017146517, "lr": 0.0026291804804649314, "grad_norm": 0.162971, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:54.772313+00:00", "epoch": 0, "step": 22062, "train_loss": 3.4947292804718018, "perplexity": 32.941368873934884, "lr": 0.0026291804804649314, "grad_norm": 0.149, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:55.077120+00:00", "epoch": 0, "step": 22063, "train_loss": 3.5660507678985596, "perplexity": 35.3766064836827, "lr": 0.0026291804804649314, "grad_norm": 0.193789, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:55.381739+00:00", "epoch": 0, "step": 22064, "train_loss": 3.4902517795562744, "perplexity": 32.79420357688351, "lr": 0.0026291804804649314, "grad_norm": 0.172556, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:55.685556+00:00", "epoch": 0, "step": 22065, "train_loss": 3.463670015335083, "perplexity": 31.93395984038804, "lr": 0.0026291804804649314, "grad_norm": 0.181229, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:55.989905+00:00", "epoch": 0, "step": 22066, "train_loss": 3.4880058765411377, "perplexity": 32.72063362260664, "lr": 0.0026291804804649314, "grad_norm": 0.168966, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:56.294808+00:00", "epoch": 0, "step": 22067, "train_loss": 3.5023083686828613, "perplexity": 33.19198292777535, "lr": 0.0026291804804649314, "grad_norm": 0.189577, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:56.599633+00:00", "epoch": 0, "step": 22068, "train_loss": 3.4923174381256104, "perplexity": 32.86201521826332, "lr": 0.0026291804804649314, "grad_norm": 0.167015, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:56.904168+00:00", "epoch": 0, "step": 22069, "train_loss": 3.5153305530548096, "perplexity": 33.62704160831825, "lr": 0.0026291804804649314, "grad_norm": 0.15906, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:57.207873+00:00", "epoch": 0, "step": 22070, "train_loss": 3.4756455421447754, "perplexity": 32.318684877214935, "lr": 0.0026291804804649314, "grad_norm": 0.163044, "tokens_per_sec": 107953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:57.512454+00:00", "epoch": 0, "step": 22071, "train_loss": 3.4729018211364746, "perplexity": 32.23013295898745, "lr": 0.0026291804804649314, "grad_norm": 0.142043, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:57.817858+00:00", "epoch": 0, "step": 22072, "train_loss": 3.5219333171844482, "perplexity": 33.84980765896212, "lr": 0.0026291804804649314, "grad_norm": 0.175985, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:58.123426+00:00", "epoch": 0, "step": 22073, "train_loss": 3.4373714923858643, "perplexity": 31.1050906531314, "lr": 0.0026291804804649314, "grad_norm": 0.159303, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:58.427155+00:00", "epoch": 0, "step": 22074, "train_loss": 3.5490262508392334, "perplexity": 34.77943454833409, "lr": 0.0026291804804649314, "grad_norm": 0.166213, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:58.730527+00:00", "epoch": 0, "step": 22075, "train_loss": 3.47072696685791, "perplexity": 32.160113285315695, "lr": 0.0026291804804649314, "grad_norm": 0.193531, "tokens_per_sec": 108012} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:59.035731+00:00", "epoch": 0, "step": 22076, "train_loss": 3.485081911087036, "perplexity": 32.625099357767624, "lr": 0.0026291804804649314, "grad_norm": 0.18754, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:59.340866+00:00", "epoch": 0, "step": 22077, "train_loss": 3.4886677265167236, "perplexity": 32.742296941313604, "lr": 0.0026291804804649314, "grad_norm": 0.177641, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:59.646520+00:00", "epoch": 0, "step": 22078, "train_loss": 3.504650354385376, "perplexity": 33.26980917563779, "lr": 0.0026291804804649314, "grad_norm": 0.172766, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:53:59.950908+00:00", "epoch": 0, "step": 22079, "train_loss": 3.500650405883789, "perplexity": 33.13699744938801, "lr": 0.0026291804804649314, "grad_norm": 0.179478, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:00.255550+00:00", "epoch": 0, "step": 22080, "train_loss": 3.4752278327941895, "perplexity": 32.305187879450415, "lr": 0.0026291804804649314, "grad_norm": 0.16146, "tokens_per_sec": 107620} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:00.559283+00:00", "epoch": 0, "step": 22081, "train_loss": 3.5507731437683105, "perplexity": 34.840243594606434, "lr": 0.0026291804804649314, "grad_norm": 0.171306, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:00.864596+00:00", "epoch": 0, "step": 22082, "train_loss": 3.614039182662964, "perplexity": 37.11566741139924, "lr": 0.0026291804804649314, "grad_norm": 0.175706, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:01.169160+00:00", "epoch": 0, "step": 22083, "train_loss": 3.4621853828430176, "perplexity": 31.886584821956266, "lr": 0.0026291804804649314, "grad_norm": 0.18274, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:01.473256+00:00", "epoch": 0, "step": 22084, "train_loss": 3.5961172580718994, "perplexity": 36.45640845933396, "lr": 0.0026291804804649314, "grad_norm": 0.19183, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:01.778168+00:00", "epoch": 0, "step": 22085, "train_loss": 3.487395763397217, "perplexity": 32.70067642265014, "lr": 0.0026291804804649314, "grad_norm": 0.206511, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:02.082196+00:00", "epoch": 0, "step": 22086, "train_loss": 3.586500406265259, "perplexity": 36.10749300496959, "lr": 0.0026291804804649314, "grad_norm": 0.190987, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:02.386054+00:00", "epoch": 0, "step": 22087, "train_loss": 3.5093793869018555, "perplexity": 33.42751579099334, "lr": 0.0026291804804649314, "grad_norm": 0.173395, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:02.691722+00:00", "epoch": 0, "step": 22088, "train_loss": 3.535402774810791, "perplexity": 34.30883066446304, "lr": 0.0026291804804649314, "grad_norm": 0.180674, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:02.996701+00:00", "epoch": 0, "step": 22089, "train_loss": 3.4884510040283203, "perplexity": 32.73520171811943, "lr": 0.0026291804804649314, "grad_norm": 0.18253, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:03.300936+00:00", "epoch": 0, "step": 22090, "train_loss": 3.6171436309814453, "perplexity": 37.231070120873824, "lr": 0.0026291804804649314, "grad_norm": 0.181011, "tokens_per_sec": 107764} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:03.605823+00:00", "epoch": 0, "step": 22091, "train_loss": 3.542208433151245, "perplexity": 34.54312119035983, "lr": 0.0026291804804649314, "grad_norm": 0.173487, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:03.910323+00:00", "epoch": 0, "step": 22092, "train_loss": 3.5198888778686523, "perplexity": 33.78067447473179, "lr": 0.0026291804804649314, "grad_norm": 0.164943, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:04.215544+00:00", "epoch": 0, "step": 22093, "train_loss": 3.4869658946990967, "perplexity": 32.68662244634736, "lr": 0.0026291804804649314, "grad_norm": 0.164321, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:04.520440+00:00", "epoch": 0, "step": 22094, "train_loss": 3.579159736633301, "perplexity": 35.84341028543155, "lr": 0.0026291804804649314, "grad_norm": 0.16179, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:04.824508+00:00", "epoch": 0, "step": 22095, "train_loss": 3.6529016494750977, "perplexity": 38.58646817041747, "lr": 0.0026291804804649314, "grad_norm": 0.18578, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:05.128709+00:00", "epoch": 0, "step": 22096, "train_loss": 3.367299795150757, "perplexity": 29.000114989992177, "lr": 0.0026291804804649314, "grad_norm": 0.164008, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:05.433001+00:00", "epoch": 0, "step": 22097, "train_loss": 3.514505386352539, "perplexity": 33.59930513846669, "lr": 0.0026291804804649314, "grad_norm": 0.144285, "tokens_per_sec": 107686} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:05.738498+00:00", "epoch": 0, "step": 22098, "train_loss": 3.570802927017212, "perplexity": 35.54512183552209, "lr": 0.0026291804804649314, "grad_norm": 0.164495, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:06.042087+00:00", "epoch": 0, "step": 22099, "train_loss": 3.6401047706604004, "perplexity": 38.09582784136087, "lr": 0.0026291804804649314, "grad_norm": 0.149855, "tokens_per_sec": 107937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:06.347219+00:00", "epoch": 0, "step": 22100, "train_loss": 3.5750112533569336, "perplexity": 35.695022502224674, "lr": 0.0026291804804649314, "grad_norm": 0.177734, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:06.651213+00:00", "epoch": 0, "step": 22101, "train_loss": 3.6299002170562744, "perplexity": 37.70905370906365, "lr": 0.0026291804804649314, "grad_norm": 0.179617, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:06.955562+00:00", "epoch": 0, "step": 22102, "train_loss": 3.480003833770752, "perplexity": 32.45984651923545, "lr": 0.0026291804804649314, "grad_norm": 0.159818, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:07.261417+00:00", "epoch": 0, "step": 22103, "train_loss": 3.6547770500183105, "perplexity": 38.65890115298269, "lr": 0.0026291804804649314, "grad_norm": 0.171427, "tokens_per_sec": 107136} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:07.565749+00:00", "epoch": 0, "step": 22104, "train_loss": 3.53692364692688, "perplexity": 34.36104970753406, "lr": 0.0026291804804649314, "grad_norm": 0.193805, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:07.869971+00:00", "epoch": 0, "step": 22105, "train_loss": 3.5600695610046387, "perplexity": 35.16564321750097, "lr": 0.0026291804804649314, "grad_norm": 0.176537, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:08.173576+00:00", "epoch": 0, "step": 22106, "train_loss": 3.4547550678253174, "perplexity": 31.650535498322032, "lr": 0.0026291804804649314, "grad_norm": 0.172819, "tokens_per_sec": 107924} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:08.477565+00:00", "epoch": 0, "step": 22107, "train_loss": 3.4840807914733887, "perplexity": 32.59245407454656, "lr": 0.0026291804804649314, "grad_norm": 0.180203, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:08.781367+00:00", "epoch": 0, "step": 22108, "train_loss": 3.3389663696289062, "perplexity": 28.189973628348664, "lr": 0.0026291804804649314, "grad_norm": 0.1977, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:09.086438+00:00", "epoch": 0, "step": 22109, "train_loss": 3.42710542678833, "perplexity": 30.787397272803165, "lr": 0.0026291804804649314, "grad_norm": 0.191917, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:09.392293+00:00", "epoch": 0, "step": 22110, "train_loss": 3.375542163848877, "perplexity": 29.240132427303706, "lr": 0.0026291804804649314, "grad_norm": 0.156421, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:09.697610+00:00", "epoch": 0, "step": 22111, "train_loss": 3.367884397506714, "perplexity": 29.017073482042576, "lr": 0.0026291804804649314, "grad_norm": 0.199246, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:10.002157+00:00", "epoch": 0, "step": 22112, "train_loss": 3.6160213947296143, "perplexity": 37.18931150019131, "lr": 0.0026291804804649314, "grad_norm": 0.178988, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:10.307406+00:00", "epoch": 0, "step": 22113, "train_loss": 3.4988157749176025, "perplexity": 33.0762590210781, "lr": 0.0026291804804649314, "grad_norm": 0.182919, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:10.612780+00:00", "epoch": 0, "step": 22114, "train_loss": 3.5316877365112305, "perplexity": 34.18160850844395, "lr": 0.0026291804804649314, "grad_norm": 0.186812, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:10.918366+00:00", "epoch": 0, "step": 22115, "train_loss": 3.431300401687622, "perplexity": 30.916820906226196, "lr": 0.0026291804804649314, "grad_norm": 0.173105, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:11.223757+00:00", "epoch": 0, "step": 22116, "train_loss": 3.53952693939209, "perplexity": 34.45061810511693, "lr": 0.0026291804804649314, "grad_norm": 0.180561, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:11.529107+00:00", "epoch": 0, "step": 22117, "train_loss": 3.447981834411621, "perplexity": 31.436883408264674, "lr": 0.0026291804804649314, "grad_norm": 0.168466, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:11.834609+00:00", "epoch": 0, "step": 22118, "train_loss": 3.5516650676727295, "perplexity": 34.871332303024346, "lr": 0.0026291804804649314, "grad_norm": 0.17427, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:12.139894+00:00", "epoch": 0, "step": 22119, "train_loss": 3.484224557876587, "perplexity": 32.597140111279515, "lr": 0.0026291804804649314, "grad_norm": 0.164599, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:12.444218+00:00", "epoch": 0, "step": 22120, "train_loss": 3.473823308944702, "perplexity": 32.25984632171489, "lr": 0.0026291804804649314, "grad_norm": 0.18154, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:12.748774+00:00", "epoch": 0, "step": 22121, "train_loss": 3.428436279296875, "perplexity": 30.828398034627003, "lr": 0.0026291804804649314, "grad_norm": 0.186853, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:13.053565+00:00", "epoch": 0, "step": 22122, "train_loss": 3.5324246883392334, "perplexity": 34.20680799158043, "lr": 0.0026291804804649314, "grad_norm": 0.162803, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:13.357253+00:00", "epoch": 0, "step": 22123, "train_loss": 3.5428123474121094, "perplexity": 34.5639885742826, "lr": 0.0026291804804649314, "grad_norm": 0.182475, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:13.661429+00:00", "epoch": 0, "step": 22124, "train_loss": 3.438368558883667, "perplexity": 31.13611996350504, "lr": 0.0026291804804649314, "grad_norm": 0.1574, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:13.966560+00:00", "epoch": 0, "step": 22125, "train_loss": 3.47727632522583, "perplexity": 32.37143264010426, "lr": 0.0026291804804649314, "grad_norm": 0.172324, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:14.272092+00:00", "epoch": 0, "step": 22126, "train_loss": 3.439800500869751, "perplexity": 31.180737017857457, "lr": 0.0026291804804649314, "grad_norm": 0.178892, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:14.578734+00:00", "epoch": 0, "step": 22127, "train_loss": 3.472871780395508, "perplexity": 32.229164756454736, "lr": 0.0026291804804649314, "grad_norm": 0.183965, "tokens_per_sec": 106861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:14.883269+00:00", "epoch": 0, "step": 22128, "train_loss": 3.364189386367798, "perplexity": 28.91005291516462, "lr": 0.0026291804804649314, "grad_norm": 0.193494, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:15.187437+00:00", "epoch": 0, "step": 22129, "train_loss": 3.412158250808716, "perplexity": 30.33063478649523, "lr": 0.0026291804804649314, "grad_norm": 0.192682, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:15.492832+00:00", "epoch": 0, "step": 22130, "train_loss": 3.627678394317627, "perplexity": 37.62536388248769, "lr": 0.0026291804804649314, "grad_norm": 0.164226, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:15.798427+00:00", "epoch": 0, "step": 22131, "train_loss": 3.471985340118408, "perplexity": 32.200608185434625, "lr": 0.0026291804804649314, "grad_norm": 0.18182, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:16.103466+00:00", "epoch": 0, "step": 22132, "train_loss": 3.5659103393554688, "perplexity": 35.37163894717482, "lr": 0.0026291804804649314, "grad_norm": 0.179886, "tokens_per_sec": 107422} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:16.408703+00:00", "epoch": 0, "step": 22133, "train_loss": 3.6105690002441406, "perplexity": 36.98709249312245, "lr": 0.0026291804804649314, "grad_norm": 0.184273, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:16.713439+00:00", "epoch": 0, "step": 22134, "train_loss": 3.459042549133301, "perplexity": 31.786527901779426, "lr": 0.0026291804804649314, "grad_norm": 0.228991, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:17.019130+00:00", "epoch": 0, "step": 22135, "train_loss": 3.5336995124816895, "perplexity": 34.25044346411177, "lr": 0.0026291804804649314, "grad_norm": 0.207777, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:17.323638+00:00", "epoch": 0, "step": 22136, "train_loss": 3.4213743209838867, "perplexity": 30.61145609192461, "lr": 0.0026291804804649314, "grad_norm": 0.15134, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:17.627925+00:00", "epoch": 0, "step": 22137, "train_loss": 3.5155692100524902, "perplexity": 33.635067894836, "lr": 0.0026291804804649314, "grad_norm": 0.181803, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:17.932187+00:00", "epoch": 0, "step": 22138, "train_loss": 3.522907257080078, "perplexity": 33.88279139658231, "lr": 0.0026291804804649314, "grad_norm": 0.180643, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:18.237736+00:00", "epoch": 0, "step": 22139, "train_loss": 3.5780627727508545, "perplexity": 35.80411291676242, "lr": 0.0026291804804649314, "grad_norm": 0.174728, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:18.541729+00:00", "epoch": 0, "step": 22140, "train_loss": 3.5134222507476807, "perplexity": 33.56293223672119, "lr": 0.0026291804804649314, "grad_norm": 0.188405, "tokens_per_sec": 107849} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:18.845850+00:00", "epoch": 0, "step": 22141, "train_loss": 3.4774982929229736, "perplexity": 32.37861884998434, "lr": 0.0026291804804649314, "grad_norm": 0.187658, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:19.150947+00:00", "epoch": 0, "step": 22142, "train_loss": 3.398559808731079, "perplexity": 29.92097707224746, "lr": 0.0026291804804649314, "grad_norm": 0.154789, "tokens_per_sec": 107401} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:19.454608+00:00", "epoch": 0, "step": 22143, "train_loss": 3.4225168228149414, "perplexity": 30.64644972289815, "lr": 0.0026291804804649314, "grad_norm": 0.183264, "tokens_per_sec": 107910} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:19.758770+00:00", "epoch": 0, "step": 22144, "train_loss": 3.39323091506958, "perplexity": 29.761955448156304, "lr": 0.0026291804804649314, "grad_norm": 0.185276, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:20.064037+00:00", "epoch": 0, "step": 22145, "train_loss": 3.537476062774658, "perplexity": 34.38003653976947, "lr": 0.0026291804804649314, "grad_norm": 0.153785, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:20.369524+00:00", "epoch": 0, "step": 22146, "train_loss": 3.5636022090911865, "perplexity": 35.290090744964104, "lr": 0.0026291804804649314, "grad_norm": 0.200385, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:20.674641+00:00", "epoch": 0, "step": 22147, "train_loss": 3.526395082473755, "perplexity": 34.00117498746059, "lr": 0.0026291804804649314, "grad_norm": 0.172767, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:20.979354+00:00", "epoch": 0, "step": 22148, "train_loss": 3.6386210918426514, "perplexity": 38.03934777804563, "lr": 0.0026291804804649314, "grad_norm": 0.18593, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:21.283978+00:00", "epoch": 0, "step": 22149, "train_loss": 3.4739065170288086, "perplexity": 32.262530713400885, "lr": 0.0026291804804649314, "grad_norm": 0.216631, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:21.589702+00:00", "epoch": 0, "step": 22150, "train_loss": 3.5854523181915283, "perplexity": 36.06966899708699, "lr": 0.0026291804804649314, "grad_norm": 0.202451, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:21.894102+00:00", "epoch": 0, "step": 22151, "train_loss": 3.480299472808838, "perplexity": 32.46944433571152, "lr": 0.0026291804804649314, "grad_norm": 0.173579, "tokens_per_sec": 107585} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:22.200014+00:00", "epoch": 0, "step": 22152, "train_loss": 3.5643088817596436, "perplexity": 35.31503810133019, "lr": 0.0026291804804649314, "grad_norm": 0.178435, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:22.504855+00:00", "epoch": 0, "step": 22153, "train_loss": 3.4973807334899902, "perplexity": 33.02882726052297, "lr": 0.0026291804804649314, "grad_norm": 0.16726, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:22.808833+00:00", "epoch": 0, "step": 22154, "train_loss": 3.568347930908203, "perplexity": 35.45796572744893, "lr": 0.0026291804804649314, "grad_norm": 0.155186, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:23.113150+00:00", "epoch": 0, "step": 22155, "train_loss": 3.6375277042388916, "perplexity": 37.99777875639779, "lr": 0.0026291804804649314, "grad_norm": 0.185768, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:23.418184+00:00", "epoch": 0, "step": 22156, "train_loss": 3.4477453231811523, "perplexity": 31.42944911146968, "lr": 0.0026291804804649314, "grad_norm": 0.177155, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:23.723552+00:00", "epoch": 0, "step": 22157, "train_loss": 3.422895908355713, "perplexity": 30.658069551179373, "lr": 0.0026291804804649314, "grad_norm": 0.178143, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:24.028516+00:00", "epoch": 0, "step": 22158, "train_loss": 3.510118007659912, "perplexity": 33.452215168660445, "lr": 0.0026291804804649314, "grad_norm": 0.17236, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:24.332976+00:00", "epoch": 0, "step": 22159, "train_loss": 3.52486252784729, "perplexity": 33.94910623872046, "lr": 0.0026291804804649314, "grad_norm": 0.172859, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:24.637426+00:00", "epoch": 0, "step": 22160, "train_loss": 3.442789316177368, "perplexity": 31.27406988983879, "lr": 0.0026291804804649314, "grad_norm": 0.180702, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:24.940900+00:00", "epoch": 0, "step": 22161, "train_loss": 3.5442988872528076, "perplexity": 34.615407529045314, "lr": 0.0026291804804649314, "grad_norm": 0.170652, "tokens_per_sec": 107920} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:25.246428+00:00", "epoch": 0, "step": 22162, "train_loss": 3.4067764282226562, "perplexity": 30.167839152695596, "lr": 0.0026291804804649314, "grad_norm": 0.184503, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:25.551074+00:00", "epoch": 0, "step": 22163, "train_loss": 3.5890681743621826, "perplexity": 36.200327811686805, "lr": 0.0026291804804649314, "grad_norm": 0.196732, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:25.855244+00:00", "epoch": 0, "step": 22164, "train_loss": 3.5948410034179688, "perplexity": 36.40991047631718, "lr": 0.0026291804804649314, "grad_norm": 0.18165, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:26.159672+00:00", "epoch": 0, "step": 22165, "train_loss": 3.524536371231079, "perplexity": 33.938035318631314, "lr": 0.0026291804804649314, "grad_norm": 0.219197, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:26.463495+00:00", "epoch": 0, "step": 22166, "train_loss": 3.6073856353759766, "perplexity": 36.86953629371627, "lr": 0.0026291804804649314, "grad_norm": 0.199189, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:26.768833+00:00", "epoch": 0, "step": 22167, "train_loss": 3.4896328449249268, "perplexity": 32.77391238868962, "lr": 0.0026291804804649314, "grad_norm": 0.161537, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:27.074151+00:00", "epoch": 0, "step": 22168, "train_loss": 3.491990566253662, "perplexity": 32.851275305216305, "lr": 0.0026291804804649314, "grad_norm": 0.198101, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:27.379659+00:00", "epoch": 0, "step": 22169, "train_loss": 3.567033052444458, "perplexity": 35.41137345024943, "lr": 0.0026291804804649314, "grad_norm": 0.161281, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:27.683841+00:00", "epoch": 0, "step": 22170, "train_loss": 3.5239062309265137, "perplexity": 33.91665633130749, "lr": 0.0026291804804649314, "grad_norm": 0.171723, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:27.988185+00:00", "epoch": 0, "step": 22171, "train_loss": 3.4763083457946777, "perplexity": 32.34011292001962, "lr": 0.0026291804804649314, "grad_norm": 0.158424, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:28.292125+00:00", "epoch": 0, "step": 22172, "train_loss": 3.4073970317840576, "perplexity": 30.186567231868768, "lr": 0.0026291804804649314, "grad_norm": 0.147338, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:28.598727+00:00", "epoch": 0, "step": 22173, "train_loss": 3.4546525478363037, "perplexity": 31.647290852093857, "lr": 0.0026291804804649314, "grad_norm": 0.151275, "tokens_per_sec": 106881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:28.903255+00:00", "epoch": 0, "step": 22174, "train_loss": 3.45931339263916, "perplexity": 31.795138242410225, "lr": 0.0026291804804649314, "grad_norm": 0.158847, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:29.207902+00:00", "epoch": 0, "step": 22175, "train_loss": 3.4305014610290527, "perplexity": 30.892130065542123, "lr": 0.0026291804804649314, "grad_norm": 0.169039, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:29.512627+00:00", "epoch": 0, "step": 22176, "train_loss": 3.4605112075805664, "perplexity": 31.83324575237721, "lr": 0.0026291804804649314, "grad_norm": 0.172772, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:29.817209+00:00", "epoch": 0, "step": 22177, "train_loss": 3.3856263160705566, "perplexity": 29.536486100107442, "lr": 0.0026291804804649314, "grad_norm": 0.158308, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:30.121451+00:00", "epoch": 0, "step": 22178, "train_loss": 3.5504705905914307, "perplexity": 34.82970416267332, "lr": 0.0026291804804649314, "grad_norm": 0.160109, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:30.426526+00:00", "epoch": 0, "step": 22179, "train_loss": 3.495543956756592, "perplexity": 32.968216360465405, "lr": 0.0026291804804649314, "grad_norm": 0.15565, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:30.730814+00:00", "epoch": 0, "step": 22180, "train_loss": 3.4586594104766846, "perplexity": 31.77435158693805, "lr": 0.0026291804804649314, "grad_norm": 0.178527, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:31.035502+00:00", "epoch": 0, "step": 22181, "train_loss": 3.4669322967529297, "perplexity": 32.03830751737426, "lr": 0.0026291804804649314, "grad_norm": 0.217828, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:31.339988+00:00", "epoch": 0, "step": 22182, "train_loss": 3.529263734817505, "perplexity": 34.09885257231123, "lr": 0.0026291804804649314, "grad_norm": 0.241763, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:31.644596+00:00", "epoch": 0, "step": 22183, "train_loss": 3.408663034439087, "perplexity": 30.22480770729471, "lr": 0.0026291804804649314, "grad_norm": 0.185244, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:31.949233+00:00", "epoch": 0, "step": 22184, "train_loss": 3.530600070953369, "perplexity": 34.14445056152975, "lr": 0.0026291804804649314, "grad_norm": 0.164249, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:32.254341+00:00", "epoch": 0, "step": 22185, "train_loss": 3.4940667152404785, "perplexity": 32.919550297160185, "lr": 0.0026291804804649314, "grad_norm": 0.170469, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:32.558468+00:00", "epoch": 0, "step": 22186, "train_loss": 3.4899017810821533, "perplexity": 32.78272766406472, "lr": 0.0026291804804649314, "grad_norm": 0.168629, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:32.863885+00:00", "epoch": 0, "step": 22187, "train_loss": 3.478659152984619, "perplexity": 32.41622772055297, "lr": 0.0026291804804649314, "grad_norm": 0.168102, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:33.168622+00:00", "epoch": 0, "step": 22188, "train_loss": 3.5097365379333496, "perplexity": 33.43945659494668, "lr": 0.0026291804804649314, "grad_norm": 0.165812, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:33.474149+00:00", "epoch": 0, "step": 22189, "train_loss": 3.5130202770233154, "perplexity": 33.549443531083625, "lr": 0.0026291804804649314, "grad_norm": 0.156541, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:33.779694+00:00", "epoch": 0, "step": 22190, "train_loss": 3.633287191390991, "perplexity": 37.83698984205957, "lr": 0.0026291804804649314, "grad_norm": 0.18087, "tokens_per_sec": 107302} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:34.083406+00:00", "epoch": 0, "step": 22191, "train_loss": 3.4961206912994385, "perplexity": 32.98723575369987, "lr": 0.0026291804804649314, "grad_norm": 0.173436, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:34.387295+00:00", "epoch": 0, "step": 22192, "train_loss": 3.3735806941986084, "perplexity": 29.182835006923607, "lr": 0.0026291804804649314, "grad_norm": 0.214268, "tokens_per_sec": 107829} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:34.691473+00:00", "epoch": 0, "step": 22193, "train_loss": 3.5058770179748535, "perplexity": 33.31064508002044, "lr": 0.0026291804804649314, "grad_norm": 0.185816, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:34.997238+00:00", "epoch": 0, "step": 22194, "train_loss": 3.4411706924438477, "perplexity": 31.22348988411323, "lr": 0.0026291804804649314, "grad_norm": 0.159705, "tokens_per_sec": 107168} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:35.302023+00:00", "epoch": 0, "step": 22195, "train_loss": 3.5029549598693848, "perplexity": 33.213451511350186, "lr": 0.0026291804804649314, "grad_norm": 0.211087, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:35.606389+00:00", "epoch": 0, "step": 22196, "train_loss": 3.5877304077148438, "perplexity": 36.15193259848607, "lr": 0.0026291804804649314, "grad_norm": 0.199104, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:35.911169+00:00", "epoch": 0, "step": 22197, "train_loss": 3.520080804824829, "perplexity": 33.78715851897285, "lr": 0.0026291804804649314, "grad_norm": 0.194013, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:36.215295+00:00", "epoch": 0, "step": 22198, "train_loss": 3.5259287357330322, "perplexity": 33.985322347025345, "lr": 0.0026291804804649314, "grad_norm": 0.245571, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:36.520366+00:00", "epoch": 0, "step": 22199, "train_loss": 3.5112619400024414, "perplexity": 33.49050413533944, "lr": 0.0026291804804649314, "grad_norm": 0.167363, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:36.825777+00:00", "epoch": 0, "step": 22200, "train_loss": 3.622400999069214, "perplexity": 37.42732299556679, "lr": 0.0026291804804649314, "grad_norm": 0.202008, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:37.131107+00:00", "epoch": 0, "step": 22201, "train_loss": 3.4870901107788086, "perplexity": 32.690682902628424, "lr": 0.0026291804804649314, "grad_norm": 0.184405, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:37.435302+00:00", "epoch": 0, "step": 22202, "train_loss": 3.4752509593963623, "perplexity": 32.305934997317735, "lr": 0.0026291804804649314, "grad_norm": 0.190283, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:37.739448+00:00", "epoch": 0, "step": 22203, "train_loss": 3.472421169281006, "perplexity": 32.21464520818496, "lr": 0.0026291804804649314, "grad_norm": 0.167392, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:38.044377+00:00", "epoch": 0, "step": 22204, "train_loss": 3.4458019733428955, "perplexity": 31.368430006560004, "lr": 0.0026291804804649314, "grad_norm": 0.200342, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:38.349344+00:00", "epoch": 0, "step": 22205, "train_loss": 3.3273329734802246, "perplexity": 27.863928679857313, "lr": 0.0026291804804649314, "grad_norm": 0.188427, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:38.653870+00:00", "epoch": 0, "step": 22206, "train_loss": 3.6426122188568115, "perplexity": 38.191471016214216, "lr": 0.0026291804804649314, "grad_norm": 0.182488, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:38.957785+00:00", "epoch": 0, "step": 22207, "train_loss": 3.482757329940796, "perplexity": 32.549347746401565, "lr": 0.0026291804804649314, "grad_norm": 0.16405, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:39.261551+00:00", "epoch": 0, "step": 22208, "train_loss": 3.459925889968872, "perplexity": 31.814618644919772, "lr": 0.0026291804804649314, "grad_norm": 0.232808, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:39.565719+00:00", "epoch": 0, "step": 22209, "train_loss": 3.5117533206939697, "perplexity": 33.50696476630801, "lr": 0.0026291804804649314, "grad_norm": 0.163088, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:39.870353+00:00", "epoch": 0, "step": 22210, "train_loss": 3.4543867111206055, "perplexity": 31.638878958377628, "lr": 0.0026291804804649314, "grad_norm": 0.171529, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:40.176600+00:00", "epoch": 0, "step": 22211, "train_loss": 3.543476104736328, "perplexity": 34.58693829053153, "lr": 0.0026291804804649314, "grad_norm": 0.176353, "tokens_per_sec": 106949} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:40.481081+00:00", "epoch": 0, "step": 22212, "train_loss": 3.4853620529174805, "perplexity": 32.63424029313954, "lr": 0.0026291804804649314, "grad_norm": 0.164799, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:40.785177+00:00", "epoch": 0, "step": 22213, "train_loss": 3.4264323711395264, "perplexity": 30.76668261299812, "lr": 0.0026291804804649314, "grad_norm": 0.163502, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:41.089940+00:00", "epoch": 0, "step": 22214, "train_loss": 3.4590845108032227, "perplexity": 31.787861745556164, "lr": 0.0026291804804649314, "grad_norm": 0.165339, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:41.394105+00:00", "epoch": 0, "step": 22215, "train_loss": 3.4188334941864014, "perplexity": 30.533776411073937, "lr": 0.0026291804804649314, "grad_norm": 0.166072, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:41.700628+00:00", "epoch": 0, "step": 22216, "train_loss": 3.545436143875122, "perplexity": 34.654796523944164, "lr": 0.0026291804804649314, "grad_norm": 0.172463, "tokens_per_sec": 106903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:42.006627+00:00", "epoch": 0, "step": 22217, "train_loss": 3.4804022312164307, "perplexity": 32.4727810155399, "lr": 0.0026291804804649314, "grad_norm": 0.170318, "tokens_per_sec": 107096} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:42.310941+00:00", "epoch": 0, "step": 22218, "train_loss": 3.485438346862793, "perplexity": 32.63673018306433, "lr": 0.0026291804804649314, "grad_norm": 0.185098, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:42.615706+00:00", "epoch": 0, "step": 22219, "train_loss": 3.405728578567505, "perplexity": 30.136244349035735, "lr": 0.0026291804804649314, "grad_norm": 0.174309, "tokens_per_sec": 107518} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:42.920998+00:00", "epoch": 0, "step": 22220, "train_loss": 3.639075517654419, "perplexity": 38.05663776775004, "lr": 0.0026291804804649314, "grad_norm": 0.167522, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:43.225386+00:00", "epoch": 0, "step": 22221, "train_loss": 3.618884801864624, "perplexity": 37.295952245158894, "lr": 0.0026291804804649314, "grad_norm": 0.173, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:43.531046+00:00", "epoch": 0, "step": 22222, "train_loss": 3.5059762001037598, "perplexity": 33.31394906456012, "lr": 0.0026291804804649314, "grad_norm": 0.16403, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:43.835204+00:00", "epoch": 0, "step": 22223, "train_loss": 3.4409596920013428, "perplexity": 31.216902408935667, "lr": 0.0026291804804649314, "grad_norm": 0.168729, "tokens_per_sec": 107728} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:44.139078+00:00", "epoch": 0, "step": 22224, "train_loss": 3.428567409515381, "perplexity": 30.832440834258275, "lr": 0.0026291804804649314, "grad_norm": 0.1729, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:44.443473+00:00", "epoch": 0, "step": 22225, "train_loss": 3.410423994064331, "perplexity": 30.278079264112222, "lr": 0.0026291804804649314, "grad_norm": 0.172523, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:44.748660+00:00", "epoch": 0, "step": 22226, "train_loss": 3.4750282764434814, "perplexity": 32.29874181724597, "lr": 0.0026291804804649314, "grad_norm": 0.167526, "tokens_per_sec": 107368} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:45.054610+00:00", "epoch": 0, "step": 22227, "train_loss": 3.4800188541412354, "perplexity": 32.46033408181768, "lr": 0.0026291804804649314, "grad_norm": 0.160559, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:45.358048+00:00", "epoch": 0, "step": 22228, "train_loss": 3.500606060028076, "perplexity": 33.135527993462745, "lr": 0.0026291804804649314, "grad_norm": 0.190434, "tokens_per_sec": 107989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:45.662213+00:00", "epoch": 0, "step": 22229, "train_loss": 3.4766101837158203, "perplexity": 32.34987586581243, "lr": 0.0026291804804649314, "grad_norm": 0.196021, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:45.966932+00:00", "epoch": 0, "step": 22230, "train_loss": 3.4765090942382812, "perplexity": 32.34660579905015, "lr": 0.0026291804804649314, "grad_norm": 0.166207, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:46.271049+00:00", "epoch": 0, "step": 22231, "train_loss": 3.4740488529205322, "perplexity": 32.26712315630621, "lr": 0.0026291804804649314, "grad_norm": 0.149019, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:46.577363+00:00", "epoch": 0, "step": 22232, "train_loss": 3.5657222270965576, "perplexity": 35.364985734066174, "lr": 0.0026291804804649314, "grad_norm": 0.162486, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:46.881779+00:00", "epoch": 0, "step": 22233, "train_loss": 3.4251348972320557, "perplexity": 30.72678953080141, "lr": 0.0026291804804649314, "grad_norm": 0.178383, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:47.185258+00:00", "epoch": 0, "step": 22234, "train_loss": 3.4671146869659424, "perplexity": 32.04415152403553, "lr": 0.0026291804804649314, "grad_norm": 0.178495, "tokens_per_sec": 107975} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:47.489655+00:00", "epoch": 0, "step": 22235, "train_loss": 3.5942606925964355, "perplexity": 36.38878754078353, "lr": 0.0026291804804649314, "grad_norm": 0.215803, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:47.793510+00:00", "epoch": 0, "step": 22236, "train_loss": 3.4997360706329346, "perplexity": 33.10671297170515, "lr": 0.0026291804804649314, "grad_norm": 0.191982, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:48.097880+00:00", "epoch": 0, "step": 22237, "train_loss": 3.5077667236328125, "perplexity": 33.3736519079283, "lr": 0.0026291804804649314, "grad_norm": 0.185283, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:48.403417+00:00", "epoch": 0, "step": 22238, "train_loss": 3.485678195953369, "perplexity": 32.64455901194933, "lr": 0.0026291804804649314, "grad_norm": 0.184294, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:48.707663+00:00", "epoch": 0, "step": 22239, "train_loss": 3.5280447006225586, "perplexity": 34.05731023092897, "lr": 0.0026291804804649314, "grad_norm": 0.218985, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:49.011918+00:00", "epoch": 0, "step": 22240, "train_loss": 3.5390145778656006, "perplexity": 34.43297145495444, "lr": 0.0026291804804649314, "grad_norm": 0.20791, "tokens_per_sec": 107766} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:49.316768+00:00", "epoch": 0, "step": 22241, "train_loss": 3.4341225624084473, "perplexity": 31.00419637962254, "lr": 0.0026291804804649314, "grad_norm": 0.224417, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:49.620938+00:00", "epoch": 0, "step": 22242, "train_loss": 3.393066644668579, "perplexity": 29.757066841337956, "lr": 0.0026291804804649314, "grad_norm": 0.172995, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:49.925381+00:00", "epoch": 0, "step": 22243, "train_loss": 3.463343620300293, "perplexity": 31.92353845529165, "lr": 0.0026291804804649314, "grad_norm": 0.173242, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:50.230350+00:00", "epoch": 0, "step": 22244, "train_loss": 3.580211877822876, "perplexity": 35.88114246007446, "lr": 0.0026291804804649314, "grad_norm": 0.165724, "tokens_per_sec": 107447} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:50.533763+00:00", "epoch": 0, "step": 22245, "train_loss": 3.5950145721435547, "perplexity": 36.41623064655333, "lr": 0.0026291804804649314, "grad_norm": 0.177759, "tokens_per_sec": 107999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:50.837428+00:00", "epoch": 0, "step": 22246, "train_loss": 3.3879594802856445, "perplexity": 29.605480028299553, "lr": 0.0026291804804649314, "grad_norm": 0.207285, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:51.142028+00:00", "epoch": 0, "step": 22247, "train_loss": 3.577545166015625, "perplexity": 35.78558526220084, "lr": 0.0026291804804649314, "grad_norm": 0.162371, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:51.447683+00:00", "epoch": 0, "step": 22248, "train_loss": 3.5157928466796875, "perplexity": 33.64259076913931, "lr": 0.0026291804804649314, "grad_norm": 0.223939, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:51.752308+00:00", "epoch": 0, "step": 22249, "train_loss": 3.4979183673858643, "perplexity": 33.04658945194944, "lr": 0.0026291804804649314, "grad_norm": 0.169205, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:52.056577+00:00", "epoch": 0, "step": 22250, "train_loss": 3.4578843116760254, "perplexity": 31.749732867333844, "lr": 0.0026291804804649314, "grad_norm": 0.21584, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:52.361428+00:00", "epoch": 0, "step": 22251, "train_loss": 3.5362048149108887, "perplexity": 34.3363587602861, "lr": 0.0026291804804649314, "grad_norm": 0.175518, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:52.666998+00:00", "epoch": 0, "step": 22252, "train_loss": 3.4146244525909424, "perplexity": 30.405528565688005, "lr": 0.0026291804804649314, "grad_norm": 0.185307, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:52.974142+00:00", "epoch": 0, "step": 22253, "train_loss": 3.4268720149993896, "perplexity": 30.78021196992872, "lr": 0.0026291804804649314, "grad_norm": 0.175871, "tokens_per_sec": 106685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:53.278710+00:00", "epoch": 0, "step": 22254, "train_loss": 3.4810776710510254, "perplexity": 32.49472183439525, "lr": 0.0026291804804649314, "grad_norm": 0.190525, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:53.583281+00:00", "epoch": 0, "step": 22255, "train_loss": 3.572889804840088, "perplexity": 35.61937761641422, "lr": 0.0026291804804649314, "grad_norm": 0.217852, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:53.886800+00:00", "epoch": 0, "step": 22256, "train_loss": 3.491180181503296, "perplexity": 32.82466391686689, "lr": 0.0026291804804649314, "grad_norm": 0.18768, "tokens_per_sec": 107959} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:54.191821+00:00", "epoch": 0, "step": 22257, "train_loss": 3.4155004024505615, "perplexity": 30.43217395247445, "lr": 0.0026291804804649314, "grad_norm": 0.194726, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:54.498108+00:00", "epoch": 0, "step": 22258, "train_loss": 3.4512429237365723, "perplexity": 31.539569235982636, "lr": 0.0026291804804649314, "grad_norm": 0.200762, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:54.803271+00:00", "epoch": 0, "step": 22259, "train_loss": 3.3899850845336914, "perplexity": 29.665509792157142, "lr": 0.0026291804804649314, "grad_norm": 0.193774, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:55.107722+00:00", "epoch": 0, "step": 22260, "train_loss": 3.543182849884033, "perplexity": 34.57679699011726, "lr": 0.0026291804804649314, "grad_norm": 0.168562, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:55.412953+00:00", "epoch": 0, "step": 22261, "train_loss": 3.5014936923980713, "perplexity": 33.1649532181656, "lr": 0.0026291804804649314, "grad_norm": 0.191365, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:55.719625+00:00", "epoch": 0, "step": 22262, "train_loss": 3.504182815551758, "perplexity": 33.254257883560875, "lr": 0.0026291804804649314, "grad_norm": 0.168937, "tokens_per_sec": 106850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:56.025413+00:00", "epoch": 0, "step": 22263, "train_loss": 3.4823927879333496, "perplexity": 32.5374843043259, "lr": 0.0026291804804649314, "grad_norm": 0.17918, "tokens_per_sec": 107160} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:56.330540+00:00", "epoch": 0, "step": 22264, "train_loss": 3.4705278873443604, "perplexity": 32.15371150286016, "lr": 0.0026291804804649314, "grad_norm": 0.179914, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:56.636209+00:00", "epoch": 0, "step": 22265, "train_loss": 3.424452781677246, "perplexity": 30.7058374563934, "lr": 0.0026291804804649314, "grad_norm": 0.172057, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:56.940543+00:00", "epoch": 0, "step": 22266, "train_loss": 3.5012779235839844, "perplexity": 33.157798027501734, "lr": 0.0026291804804649314, "grad_norm": 0.197753, "tokens_per_sec": 107671} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:57.245605+00:00", "epoch": 0, "step": 22267, "train_loss": 3.5757362842559814, "perplexity": 35.72091188064655, "lr": 0.0026291804804649314, "grad_norm": 0.153073, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:57.551045+00:00", "epoch": 0, "step": 22268, "train_loss": 3.539560556411743, "perplexity": 34.451776251689445, "lr": 0.0026291804804649314, "grad_norm": 0.202701, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:57.857046+00:00", "epoch": 0, "step": 22269, "train_loss": 3.5324881076812744, "perplexity": 34.20897743362815, "lr": 0.0026291804804649314, "grad_norm": 0.191616, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:58.161603+00:00", "epoch": 0, "step": 22270, "train_loss": 3.5271782875061035, "perplexity": 34.02781530987203, "lr": 0.0026291804804649314, "grad_norm": 0.20257, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:58.466457+00:00", "epoch": 0, "step": 22271, "train_loss": 3.4811360836029053, "perplexity": 32.49661998945772, "lr": 0.0026291804804649314, "grad_norm": 0.182891, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:58.771378+00:00", "epoch": 0, "step": 22272, "train_loss": 3.516763925552368, "perplexity": 33.67527624577685, "lr": 0.0026291804804649314, "grad_norm": 0.182302, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:59.076400+00:00", "epoch": 0, "step": 22273, "train_loss": 3.547488212585449, "perplexity": 34.7259835629207, "lr": 0.0026291804804649314, "grad_norm": 0.196443, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:59.381257+00:00", "epoch": 0, "step": 22274, "train_loss": 3.562490940093994, "perplexity": 35.250895743339754, "lr": 0.0026291804804649314, "grad_norm": 0.21592, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:59.686566+00:00", "epoch": 0, "step": 22275, "train_loss": 3.6232919692993164, "perplexity": 37.46068448598873, "lr": 0.0026291804804649314, "grad_norm": 0.200478, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:54:59.991406+00:00", "epoch": 0, "step": 22276, "train_loss": 3.4793341159820557, "perplexity": 32.43811486045466, "lr": 0.0026291804804649314, "grad_norm": 0.195749, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:00.296337+00:00", "epoch": 0, "step": 22277, "train_loss": 3.487776279449463, "perplexity": 32.71312192265436, "lr": 0.0026291804804649314, "grad_norm": 0.200865, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:00.600780+00:00", "epoch": 0, "step": 22278, "train_loss": 3.5510621070861816, "perplexity": 34.850312601707756, "lr": 0.0026291804804649314, "grad_norm": 0.18191, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:00.905068+00:00", "epoch": 0, "step": 22279, "train_loss": 3.5735368728637695, "perplexity": 35.64243323516563, "lr": 0.0026291804804649314, "grad_norm": 0.168897, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:01.210672+00:00", "epoch": 0, "step": 22280, "train_loss": 3.397252082824707, "perplexity": 29.88187420887248, "lr": 0.0026291804804649314, "grad_norm": 0.170073, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:01.515367+00:00", "epoch": 0, "step": 22281, "train_loss": 3.64564847946167, "perplexity": 38.30760649467148, "lr": 0.0026291804804649314, "grad_norm": 0.179375, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:01.819981+00:00", "epoch": 0, "step": 22282, "train_loss": 3.385010242462158, "perplexity": 29.518295054623216, "lr": 0.0026291804804649314, "grad_norm": 0.155207, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:02.124504+00:00", "epoch": 0, "step": 22283, "train_loss": 3.4893226623535156, "perplexity": 32.76374806874863, "lr": 0.0026291804804649314, "grad_norm": 0.160859, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:02.429051+00:00", "epoch": 0, "step": 22284, "train_loss": 3.3604860305786133, "perplexity": 28.803186707268207, "lr": 0.0026291804804649314, "grad_norm": 0.170158, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:02.733452+00:00", "epoch": 0, "step": 22285, "train_loss": 3.5452520847320557, "perplexity": 34.64841857877039, "lr": 0.0026291804804649314, "grad_norm": 0.160128, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:03.038986+00:00", "epoch": 0, "step": 22286, "train_loss": 3.4294068813323975, "perplexity": 30.858334666439, "lr": 0.0026291804804649314, "grad_norm": 0.165346, "tokens_per_sec": 107248} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:03.343694+00:00", "epoch": 0, "step": 22287, "train_loss": 3.4970052242279053, "perplexity": 33.01642695833071, "lr": 0.0026291804804649314, "grad_norm": 0.153989, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:03.647842+00:00", "epoch": 0, "step": 22288, "train_loss": 3.510529041290283, "perplexity": 33.4659679803456, "lr": 0.0026291804804649314, "grad_norm": 0.159553, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:03.952275+00:00", "epoch": 0, "step": 22289, "train_loss": 3.464630126953125, "perplexity": 31.964634729543235, "lr": 0.0026291804804649314, "grad_norm": 0.165931, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:04.257086+00:00", "epoch": 0, "step": 22290, "train_loss": 3.4518916606903076, "perplexity": 31.56003675835214, "lr": 0.0026291804804649314, "grad_norm": 0.182574, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:04.562937+00:00", "epoch": 0, "step": 22291, "train_loss": 3.4906868934631348, "perplexity": 32.80847589574125, "lr": 0.0026291804804649314, "grad_norm": 0.191288, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:04.866797+00:00", "epoch": 0, "step": 22292, "train_loss": 3.6018662452697754, "perplexity": 36.666599498728715, "lr": 0.0026291804804649314, "grad_norm": 0.196566, "tokens_per_sec": 107839} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:05.171479+00:00", "epoch": 0, "step": 22293, "train_loss": 3.394784927368164, "perplexity": 29.808241848455346, "lr": 0.0026291804804649314, "grad_norm": 0.177143, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:05.477397+00:00", "epoch": 0, "step": 22294, "train_loss": 3.4372503757476807, "perplexity": 31.101323537255904, "lr": 0.0026291804804649314, "grad_norm": 0.151802, "tokens_per_sec": 107114} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:05.782399+00:00", "epoch": 0, "step": 22295, "train_loss": 3.535876989364624, "perplexity": 34.32510426957435, "lr": 0.0026291804804649314, "grad_norm": 0.182219, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:06.087635+00:00", "epoch": 0, "step": 22296, "train_loss": 3.466604471206665, "perplexity": 32.02780626309474, "lr": 0.0026291804804649314, "grad_norm": 0.184911, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:06.392948+00:00", "epoch": 0, "step": 22297, "train_loss": 3.4060192108154297, "perplexity": 30.145004186358854, "lr": 0.0026291804804649314, "grad_norm": 0.174747, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:06.697467+00:00", "epoch": 0, "step": 22298, "train_loss": 3.557234287261963, "perplexity": 35.06608020353877, "lr": 0.0026291804804649314, "grad_norm": 0.142277, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:07.002677+00:00", "epoch": 0, "step": 22299, "train_loss": 3.4505956172943115, "perplexity": 31.51916007583311, "lr": 0.0026291804804649314, "grad_norm": 0.177583, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:07.307552+00:00", "epoch": 0, "step": 22300, "train_loss": 3.544369697570801, "perplexity": 34.617858743844536, "lr": 0.0026291804804649314, "grad_norm": 0.18103, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:07.613021+00:00", "epoch": 0, "step": 22301, "train_loss": 3.4578983783721924, "perplexity": 31.75017948432068, "lr": 0.0026291804804649314, "grad_norm": 0.181099, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:07.918606+00:00", "epoch": 0, "step": 22302, "train_loss": 3.509188413619995, "perplexity": 33.421132638123524, "lr": 0.0026291804804649314, "grad_norm": 0.173235, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:08.223500+00:00", "epoch": 0, "step": 22303, "train_loss": 3.4953396320343018, "perplexity": 32.9614808269546, "lr": 0.0026291804804649314, "grad_norm": 0.165375, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:08.528125+00:00", "epoch": 0, "step": 22304, "train_loss": 3.5057592391967773, "perplexity": 33.306722023976896, "lr": 0.0026291804804649314, "grad_norm": 0.183941, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:08.831928+00:00", "epoch": 0, "step": 22305, "train_loss": 3.6141364574432373, "perplexity": 37.11927800539841, "lr": 0.0026291804804649314, "grad_norm": 0.178246, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:09.136312+00:00", "epoch": 0, "step": 22306, "train_loss": 3.4444520473480225, "perplexity": 31.326113515967236, "lr": 0.0026291804804649314, "grad_norm": 0.179398, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:09.441141+00:00", "epoch": 0, "step": 22307, "train_loss": 3.434515953063965, "perplexity": 31.016395540120662, "lr": 0.0026291804804649314, "grad_norm": 0.187659, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:09.747219+00:00", "epoch": 0, "step": 22308, "train_loss": 3.560593366622925, "perplexity": 35.18406800407104, "lr": 0.0026291804804649314, "grad_norm": 0.179517, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:10.051525+00:00", "epoch": 0, "step": 22309, "train_loss": 3.526088237762451, "perplexity": 33.99074350724169, "lr": 0.0026291804804649314, "grad_norm": 0.164617, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:10.355764+00:00", "epoch": 0, "step": 22310, "train_loss": 3.5021486282348633, "perplexity": 33.18668124901007, "lr": 0.0026291804804649314, "grad_norm": 0.177401, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:10.659845+00:00", "epoch": 0, "step": 22311, "train_loss": 3.5093159675598145, "perplexity": 33.42539590715734, "lr": 0.0026291804804649314, "grad_norm": 0.165028, "tokens_per_sec": 107691} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:10.965473+00:00", "epoch": 0, "step": 22312, "train_loss": 3.487637758255005, "perplexity": 32.70859077576837, "lr": 0.0026291804804649314, "grad_norm": 0.187314, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:11.270226+00:00", "epoch": 0, "step": 22313, "train_loss": 3.5169637203216553, "perplexity": 33.68200506199381, "lr": 0.0026291804804649314, "grad_norm": 0.174109, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:11.574876+00:00", "epoch": 0, "step": 22314, "train_loss": 3.43733286857605, "perplexity": 31.10388927922672, "lr": 0.0026291804804649314, "grad_norm": 0.174852, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:11.880095+00:00", "epoch": 0, "step": 22315, "train_loss": 3.4857420921325684, "perplexity": 32.646644941182586, "lr": 0.0026291804804649314, "grad_norm": 0.170988, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:12.184596+00:00", "epoch": 0, "step": 22316, "train_loss": 3.3944029808044434, "perplexity": 29.7968588668945, "lr": 0.0026291804804649314, "grad_norm": 0.173558, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:12.490772+00:00", "epoch": 0, "step": 22317, "train_loss": 3.484288454055786, "perplexity": 32.59922301052939, "lr": 0.0026291804804649314, "grad_norm": 0.17141, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:12.795318+00:00", "epoch": 0, "step": 22318, "train_loss": 3.447753429412842, "perplexity": 31.429703886898682, "lr": 0.0026291804804649314, "grad_norm": 0.186695, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:13.100516+00:00", "epoch": 0, "step": 22319, "train_loss": 3.5774917602539062, "perplexity": 35.783674156793836, "lr": 0.0026291804804649314, "grad_norm": 0.162376, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:13.404993+00:00", "epoch": 0, "step": 22320, "train_loss": 3.5931644439697266, "perplexity": 36.348918239740385, "lr": 0.0026291804804649314, "grad_norm": 0.175929, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:13.708636+00:00", "epoch": 0, "step": 22321, "train_loss": 3.454503059387207, "perplexity": 31.64256030125643, "lr": 0.0026291804804649314, "grad_norm": 0.173834, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:14.012695+00:00", "epoch": 0, "step": 22322, "train_loss": 3.4551010131835938, "perplexity": 31.66148674832247, "lr": 0.0026291804804649314, "grad_norm": 0.182691, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:14.317707+00:00", "epoch": 0, "step": 22323, "train_loss": 3.468111753463745, "perplexity": 32.076117607477855, "lr": 0.0026291804804649314, "grad_norm": 0.175085, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:14.622661+00:00", "epoch": 0, "step": 22324, "train_loss": 3.4780397415161133, "perplexity": 32.39615495462965, "lr": 0.0026291804804649314, "grad_norm": 0.159372, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:14.927022+00:00", "epoch": 0, "step": 22325, "train_loss": 3.4186317920684814, "perplexity": 30.527618304776148, "lr": 0.0026291804804649314, "grad_norm": 0.186364, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:15.231047+00:00", "epoch": 0, "step": 22326, "train_loss": 3.568819761276245, "perplexity": 35.4746998199842, "lr": 0.0026291804804649314, "grad_norm": 0.189554, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:15.534778+00:00", "epoch": 0, "step": 22327, "train_loss": 3.4302167892456055, "perplexity": 30.883337199381707, "lr": 0.0026291804804649314, "grad_norm": 0.15064, "tokens_per_sec": 107879} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:15.840351+00:00", "epoch": 0, "step": 22328, "train_loss": 3.5592093467712402, "perplexity": 35.13540623768341, "lr": 0.0026291804804649314, "grad_norm": 0.176247, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:16.144760+00:00", "epoch": 0, "step": 22329, "train_loss": 3.536048412322998, "perplexity": 34.3309888848594, "lr": 0.0026291804804649314, "grad_norm": 0.165987, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:16.448664+00:00", "epoch": 0, "step": 22330, "train_loss": 3.428924560546875, "perplexity": 30.843454638984497, "lr": 0.0026291804804649314, "grad_norm": 0.152385, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:16.753929+00:00", "epoch": 0, "step": 22331, "train_loss": 3.521247386932373, "perplexity": 33.826597013214005, "lr": 0.0026291804804649314, "grad_norm": 0.186592, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:17.058686+00:00", "epoch": 0, "step": 22332, "train_loss": 3.504617691040039, "perplexity": 33.26872249011888, "lr": 0.0026291804804649314, "grad_norm": 0.155103, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:17.363606+00:00", "epoch": 0, "step": 22333, "train_loss": 3.520961046218872, "perplexity": 33.81691246789591, "lr": 0.0026291804804649314, "grad_norm": 0.196008, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:17.667882+00:00", "epoch": 0, "step": 22334, "train_loss": 3.605672597885132, "perplexity": 36.806431461678386, "lr": 0.0026291804804649314, "grad_norm": 0.180106, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:17.972594+00:00", "epoch": 0, "step": 22335, "train_loss": 3.556870222091675, "perplexity": 35.053316188685734, "lr": 0.0026291804804649314, "grad_norm": 0.166639, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:18.277926+00:00", "epoch": 0, "step": 22336, "train_loss": 3.4532034397125244, "perplexity": 31.601463718079998, "lr": 0.0026291804804649314, "grad_norm": 0.180799, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:18.584169+00:00", "epoch": 0, "step": 22337, "train_loss": 3.539414167404175, "perplexity": 34.44673325948332, "lr": 0.0026291804804649314, "grad_norm": 0.193035, "tokens_per_sec": 107005} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:18.889082+00:00", "epoch": 0, "step": 22338, "train_loss": 3.4912092685699463, "perplexity": 32.825618703939924, "lr": 0.0026291804804649314, "grad_norm": 0.221618, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:19.194604+00:00", "epoch": 0, "step": 22339, "train_loss": 3.4924285411834717, "perplexity": 32.865666491471615, "lr": 0.0026291804804649314, "grad_norm": 0.176931, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:19.499555+00:00", "epoch": 0, "step": 22340, "train_loss": 3.4824318885803223, "perplexity": 32.538756565886025, "lr": 0.0026291804804649314, "grad_norm": 0.180613, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:19.803581+00:00", "epoch": 0, "step": 22341, "train_loss": 3.55092191696167, "perplexity": 34.84542727449017, "lr": 0.0026291804804649314, "grad_norm": 0.173225, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:20.107600+00:00", "epoch": 0, "step": 22342, "train_loss": 3.577895164489746, "perplexity": 35.798112354541864, "lr": 0.0026291804804649314, "grad_norm": 0.164469, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:20.411463+00:00", "epoch": 0, "step": 22343, "train_loss": 3.5180954933166504, "perplexity": 33.72014702569114, "lr": 0.0026291804804649314, "grad_norm": 0.157766, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:20.716415+00:00", "epoch": 0, "step": 22344, "train_loss": 3.599970817565918, "perplexity": 36.59716643369746, "lr": 0.0026291804804649314, "grad_norm": 0.165772, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:21.022448+00:00", "epoch": 0, "step": 22345, "train_loss": 3.5307846069335938, "perplexity": 34.15075202258862, "lr": 0.0026291804804649314, "grad_norm": 0.166626, "tokens_per_sec": 107073} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:21.326380+00:00", "epoch": 0, "step": 22346, "train_loss": 3.580721616744995, "perplexity": 35.899437137314855, "lr": 0.0026291804804649314, "grad_norm": 0.155799, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:21.631059+00:00", "epoch": 0, "step": 22347, "train_loss": 3.5534791946411133, "perplexity": 34.93465074385891, "lr": 0.0026291804804649314, "grad_norm": 0.167323, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:21.936360+00:00", "epoch": 0, "step": 22348, "train_loss": 3.398178815841675, "perplexity": 29.90957956406156, "lr": 0.0026291804804649314, "grad_norm": 0.164905, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:22.241715+00:00", "epoch": 0, "step": 22349, "train_loss": 3.445413589477539, "perplexity": 31.35624937999642, "lr": 0.0026291804804649314, "grad_norm": 0.163608, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:22.547756+00:00", "epoch": 0, "step": 22350, "train_loss": 3.411660671234131, "perplexity": 30.31554663623027, "lr": 0.0026291804804649314, "grad_norm": 0.162052, "tokens_per_sec": 107128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:22.851535+00:00", "epoch": 0, "step": 22351, "train_loss": 3.465296506881714, "perplexity": 31.985942419271563, "lr": 0.0026291804804649314, "grad_norm": 0.184673, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:23.156885+00:00", "epoch": 0, "step": 22352, "train_loss": 3.4814682006835938, "perplexity": 32.5074144644364, "lr": 0.0026291804804649314, "grad_norm": 0.160455, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:23.461721+00:00", "epoch": 0, "step": 22353, "train_loss": 3.4059886932373047, "perplexity": 30.144084247875735, "lr": 0.0026291804804649314, "grad_norm": 0.178731, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:23.766794+00:00", "epoch": 0, "step": 22354, "train_loss": 3.4646754264831543, "perplexity": 31.96608274527101, "lr": 0.0026291804804649314, "grad_norm": 0.185493, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:24.071879+00:00", "epoch": 0, "step": 22355, "train_loss": 3.548166275024414, "perplexity": 34.74953793279511, "lr": 0.0026291804804649314, "grad_norm": 0.177715, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:24.377130+00:00", "epoch": 0, "step": 22356, "train_loss": 3.5314536094665527, "perplexity": 34.17360660622895, "lr": 0.0026291804804649314, "grad_norm": 0.17466, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:24.680248+00:00", "epoch": 0, "step": 22357, "train_loss": 3.4695322513580322, "perplexity": 32.12171404214881, "lr": 0.0026291804804649314, "grad_norm": 0.159936, "tokens_per_sec": 108102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:24.984831+00:00", "epoch": 0, "step": 22358, "train_loss": 3.487027406692505, "perplexity": 32.68863312749168, "lr": 0.0026291804804649314, "grad_norm": 0.189252, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:25.289577+00:00", "epoch": 0, "step": 22359, "train_loss": 3.560809850692749, "perplexity": 35.1916856188219, "lr": 0.0026291804804649314, "grad_norm": 0.202138, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:25.595839+00:00", "epoch": 0, "step": 22360, "train_loss": 3.4960269927978516, "perplexity": 32.98414504393796, "lr": 0.0026291804804649314, "grad_norm": 0.160522, "tokens_per_sec": 107052} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:25.900381+00:00", "epoch": 0, "step": 22361, "train_loss": 3.6423511505126953, "perplexity": 38.181501733505335, "lr": 0.0026291804804649314, "grad_norm": 0.193497, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:26.203956+00:00", "epoch": 0, "step": 22362, "train_loss": 3.4578652381896973, "perplexity": 31.749127295013285, "lr": 0.0026291804804649314, "grad_norm": 0.158795, "tokens_per_sec": 107941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:26.508831+00:00", "epoch": 0, "step": 22363, "train_loss": 3.4286880493164062, "perplexity": 30.83616067816175, "lr": 0.0026291804804649314, "grad_norm": 0.178251, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:26.813040+00:00", "epoch": 0, "step": 22364, "train_loss": 3.52829647064209, "perplexity": 34.065885920096406, "lr": 0.0026291804804649314, "grad_norm": 0.159847, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:27.118302+00:00", "epoch": 0, "step": 22365, "train_loss": 3.478604793548584, "perplexity": 32.414465640588965, "lr": 0.0026291804804649314, "grad_norm": 0.171603, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:27.422514+00:00", "epoch": 0, "step": 22366, "train_loss": 3.4285671710968018, "perplexity": 30.832433483232418, "lr": 0.0026291804804649314, "grad_norm": 0.172402, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:27.727625+00:00", "epoch": 0, "step": 22367, "train_loss": 3.4536221027374268, "perplexity": 31.614696852390278, "lr": 0.0026291804804649314, "grad_norm": 0.180888, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:28.032578+00:00", "epoch": 0, "step": 22368, "train_loss": 3.4694595336914062, "perplexity": 32.119378310981126, "lr": 0.0026291804804649314, "grad_norm": 0.169787, "tokens_per_sec": 107452} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:28.336860+00:00", "epoch": 0, "step": 22369, "train_loss": 3.495553731918335, "perplexity": 32.968538631687835, "lr": 0.0026291804804649314, "grad_norm": 0.17852, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:28.641917+00:00", "epoch": 0, "step": 22370, "train_loss": 3.558901786804199, "perplexity": 35.1246016549127, "lr": 0.0026291804804649314, "grad_norm": 0.208406, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:28.947372+00:00", "epoch": 0, "step": 22371, "train_loss": 3.5625784397125244, "perplexity": 35.253980318217735, "lr": 0.0026291804804649314, "grad_norm": 0.278044, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:29.251899+00:00", "epoch": 0, "step": 22372, "train_loss": 3.4223365783691406, "perplexity": 30.64092636834403, "lr": 0.0026291804804649314, "grad_norm": 0.211375, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:29.556797+00:00", "epoch": 0, "step": 22373, "train_loss": 3.53658127784729, "perplexity": 34.34928756018462, "lr": 0.0026291804804649314, "grad_norm": 0.163939, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:29.860620+00:00", "epoch": 0, "step": 22374, "train_loss": 3.42004656791687, "perplexity": 30.570838608174164, "lr": 0.0026291804804649314, "grad_norm": 0.214927, "tokens_per_sec": 107852} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:30.165904+00:00", "epoch": 0, "step": 22375, "train_loss": 3.43918776512146, "perplexity": 31.161637317760643, "lr": 0.0026291804804649314, "grad_norm": 0.231585, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:30.470253+00:00", "epoch": 0, "step": 22376, "train_loss": 3.5443098545074463, "perplexity": 34.6157871671159, "lr": 0.0026291804804649314, "grad_norm": 0.191898, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:30.776411+00:00", "epoch": 0, "step": 22377, "train_loss": 3.4964711666107178, "perplexity": 32.99879899161523, "lr": 0.0026291804804649314, "grad_norm": 0.18011, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:31.080597+00:00", "epoch": 0, "step": 22378, "train_loss": 3.4432380199432373, "perplexity": 31.288105831526867, "lr": 0.0026291804804649314, "grad_norm": 0.191958, "tokens_per_sec": 107722} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:31.386372+00:00", "epoch": 0, "step": 22379, "train_loss": 3.6843533515930176, "perplexity": 39.819364993809685, "lr": 0.0026291804804649314, "grad_norm": 0.164922, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:31.691101+00:00", "epoch": 0, "step": 22380, "train_loss": 3.361621379852295, "perplexity": 28.8359069553162, "lr": 0.0026291804804649314, "grad_norm": 0.171455, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:31.995513+00:00", "epoch": 0, "step": 22381, "train_loss": 3.5155441761016846, "perplexity": 33.63422588674041, "lr": 0.0026291804804649314, "grad_norm": 0.185914, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:32.300669+00:00", "epoch": 0, "step": 22382, "train_loss": 3.5113515853881836, "perplexity": 33.49350653907517, "lr": 0.0026291804804649314, "grad_norm": 0.157064, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:32.607661+00:00", "epoch": 0, "step": 22383, "train_loss": 3.5321311950683594, "perplexity": 34.19676999673085, "lr": 0.0026291804804649314, "grad_norm": 0.160077, "tokens_per_sec": 106732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:32.912489+00:00", "epoch": 0, "step": 22384, "train_loss": 3.527280569076538, "perplexity": 34.03129590625778, "lr": 0.0026291804804649314, "grad_norm": 0.16598, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:33.217338+00:00", "epoch": 0, "step": 22385, "train_loss": 3.4495413303375244, "perplexity": 31.48594734742584, "lr": 0.0026291804804649314, "grad_norm": 0.168872, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:33.522418+00:00", "epoch": 0, "step": 22386, "train_loss": 3.621633529663086, "perplexity": 37.39860968991561, "lr": 0.0026291804804649314, "grad_norm": 0.21146, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:33.827406+00:00", "epoch": 0, "step": 22387, "train_loss": 3.455399990081787, "perplexity": 31.67095421662931, "lr": 0.0026291804804649314, "grad_norm": 0.206103, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:34.132930+00:00", "epoch": 0, "step": 22388, "train_loss": 3.5456461906433105, "perplexity": 34.66207641648843, "lr": 0.0026291804804649314, "grad_norm": 0.17261, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:34.437305+00:00", "epoch": 0, "step": 22389, "train_loss": 3.5526037216186523, "perplexity": 34.90407978356168, "lr": 0.0026291804804649314, "grad_norm": 0.183844, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:34.741741+00:00", "epoch": 0, "step": 22390, "train_loss": 3.5329160690307617, "perplexity": 34.22362068692495, "lr": 0.0026291804804649314, "grad_norm": 0.169895, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:35.046842+00:00", "epoch": 0, "step": 22391, "train_loss": 3.42787766456604, "perplexity": 30.811181646468533, "lr": 0.0026291804804649314, "grad_norm": 0.175138, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:35.352343+00:00", "epoch": 0, "step": 22392, "train_loss": 3.5486197471618652, "perplexity": 34.765299453472075, "lr": 0.0026291804804649314, "grad_norm": 0.158623, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:35.657938+00:00", "epoch": 0, "step": 22393, "train_loss": 3.51472806930542, "perplexity": 33.6067879640676, "lr": 0.0026291804804649314, "grad_norm": 0.157194, "tokens_per_sec": 107227} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:35.964414+00:00", "epoch": 0, "step": 22394, "train_loss": 3.3882946968078613, "perplexity": 29.615405937924702, "lr": 0.0026291804804649314, "grad_norm": 0.163793, "tokens_per_sec": 106923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:36.269164+00:00", "epoch": 0, "step": 22395, "train_loss": 3.476931095123291, "perplexity": 32.36025897594816, "lr": 0.0026291804804649314, "grad_norm": 0.15925, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:36.573204+00:00", "epoch": 0, "step": 22396, "train_loss": 3.530566453933716, "perplexity": 34.14330274615735, "lr": 0.0026291804804649314, "grad_norm": 0.188708, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:36.878093+00:00", "epoch": 0, "step": 22397, "train_loss": 3.4148271083831787, "perplexity": 30.41169104657798, "lr": 0.0026291804804649314, "grad_norm": 0.1945, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:37.183455+00:00", "epoch": 0, "step": 22398, "train_loss": 3.5163462162017822, "perplexity": 33.66121270544606, "lr": 0.0026291804804649314, "grad_norm": 0.187172, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:37.488465+00:00", "epoch": 0, "step": 22399, "train_loss": 3.408815383911133, "perplexity": 30.229412791573775, "lr": 0.0026291804804649314, "grad_norm": 0.162171, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:37.791889+00:00", "epoch": 0, "step": 22400, "train_loss": 3.5313074588775635, "perplexity": 34.16861247845181, "lr": 0.0026291804804649314, "grad_norm": 0.185302, "tokens_per_sec": 108062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:38.095654+00:00", "epoch": 0, "step": 22401, "train_loss": 3.5441110134124756, "perplexity": 34.60890481036267, "lr": 0.0026291804804649314, "grad_norm": 0.193889, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:38.399863+00:00", "epoch": 0, "step": 22402, "train_loss": 3.4460511207580566, "perplexity": 31.376246343483437, "lr": 0.0026291804804649314, "grad_norm": 0.225699, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:38.705134+00:00", "epoch": 0, "step": 22403, "train_loss": 3.6382670402526855, "perplexity": 38.02588227036641, "lr": 0.0026291804804649314, "grad_norm": 0.199613, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:39.009906+00:00", "epoch": 0, "step": 22404, "train_loss": 3.5811350345611572, "perplexity": 35.91428167250385, "lr": 0.0026291804804649314, "grad_norm": 0.179298, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:39.314008+00:00", "epoch": 0, "step": 22405, "train_loss": 3.462000608444214, "perplexity": 31.88069354171153, "lr": 0.0026291804804649314, "grad_norm": 0.180168, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:39.619133+00:00", "epoch": 0, "step": 22406, "train_loss": 3.4624369144439697, "perplexity": 31.894606314472565, "lr": 0.0026291804804649314, "grad_norm": 0.192763, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:39.924160+00:00", "epoch": 0, "step": 22407, "train_loss": 3.440122365951538, "perplexity": 31.190774623618637, "lr": 0.0026291804804649314, "grad_norm": 0.173735, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:40.229207+00:00", "epoch": 0, "step": 22408, "train_loss": 3.532583475112915, "perplexity": 34.21224001151441, "lr": 0.0026291804804649314, "grad_norm": 0.191882, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:40.534300+00:00", "epoch": 0, "step": 22409, "train_loss": 3.4909234046936035, "perplexity": 32.81623638643059, "lr": 0.0026291804804649314, "grad_norm": 0.173135, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:40.839375+00:00", "epoch": 0, "step": 22410, "train_loss": 3.4600255489349365, "perplexity": 31.817789414914866, "lr": 0.0026291804804649314, "grad_norm": 0.179166, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:41.143770+00:00", "epoch": 0, "step": 22411, "train_loss": 3.4702742099761963, "perplexity": 32.14555586844308, "lr": 0.0026291804804649314, "grad_norm": 0.15925, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:41.447400+00:00", "epoch": 0, "step": 22412, "train_loss": 3.5538296699523926, "perplexity": 34.94689662226647, "lr": 0.0026291804804649314, "grad_norm": 0.178445, "tokens_per_sec": 107921} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:41.751958+00:00", "epoch": 0, "step": 22413, "train_loss": 3.48329758644104, "perplexity": 32.566937494160214, "lr": 0.0026291804804649314, "grad_norm": 0.158825, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:42.056195+00:00", "epoch": 0, "step": 22414, "train_loss": 3.4065678119659424, "perplexity": 30.161546307436147, "lr": 0.0026291804804649314, "grad_norm": 0.164989, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:42.361926+00:00", "epoch": 0, "step": 22415, "train_loss": 3.4016623497009277, "perplexity": 30.013952284595003, "lr": 0.0026291804804649314, "grad_norm": 0.163313, "tokens_per_sec": 107178} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:42.666430+00:00", "epoch": 0, "step": 22416, "train_loss": 3.4871296882629395, "perplexity": 32.691976743215534, "lr": 0.0026291804804649314, "grad_norm": 0.171747, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:42.971095+00:00", "epoch": 0, "step": 22417, "train_loss": 3.558504819869995, "perplexity": 35.11066111662805, "lr": 0.0026291804804649314, "grad_norm": 0.155063, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:43.275057+00:00", "epoch": 0, "step": 22418, "train_loss": 3.5104918479919434, "perplexity": 33.46472329376142, "lr": 0.0026291804804649314, "grad_norm": 0.180803, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:43.581827+00:00", "epoch": 0, "step": 22419, "train_loss": 3.492729663848877, "perplexity": 32.87556457876027, "lr": 0.0026291804804649314, "grad_norm": 0.17416, "tokens_per_sec": 106816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:43.887678+00:00", "epoch": 0, "step": 22420, "train_loss": 3.5046379566192627, "perplexity": 33.26939670688185, "lr": 0.0026291804804649314, "grad_norm": 0.198827, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:44.193463+00:00", "epoch": 0, "step": 22421, "train_loss": 3.6065452098846436, "perplexity": 36.83856321266789, "lr": 0.0026291804804649314, "grad_norm": 0.202529, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:44.498475+00:00", "epoch": 0, "step": 22422, "train_loss": 3.4404776096343994, "perplexity": 31.201856917608147, "lr": 0.0026291804804649314, "grad_norm": 0.163327, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:44.802941+00:00", "epoch": 0, "step": 22423, "train_loss": 3.5173499584198, "perplexity": 33.69501684823268, "lr": 0.0026291804804649314, "grad_norm": 0.200597, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:45.108335+00:00", "epoch": 0, "step": 22424, "train_loss": 3.478203773498535, "perplexity": 32.401469396006966, "lr": 0.0026291804804649314, "grad_norm": 0.186481, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:45.414022+00:00", "epoch": 0, "step": 22425, "train_loss": 3.558880567550659, "perplexity": 35.123856344992184, "lr": 0.0026291804804649314, "grad_norm": 0.167367, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:45.719188+00:00", "epoch": 0, "step": 22426, "train_loss": 3.424323797225952, "perplexity": 30.70187713621292, "lr": 0.0026291804804649314, "grad_norm": 0.169646, "tokens_per_sec": 107378} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:46.022956+00:00", "epoch": 0, "step": 22427, "train_loss": 3.572831392288208, "perplexity": 35.61729705843722, "lr": 0.0026291804804649314, "grad_norm": 0.161634, "tokens_per_sec": 107874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:46.327143+00:00", "epoch": 0, "step": 22428, "train_loss": 3.437669515609741, "perplexity": 31.114362074007577, "lr": 0.0026291804804649314, "grad_norm": 0.164513, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:46.631093+00:00", "epoch": 0, "step": 22429, "train_loss": 3.3640949726104736, "perplexity": 28.907323537291894, "lr": 0.0026291804804649314, "grad_norm": 0.173133, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:46.935907+00:00", "epoch": 0, "step": 22430, "train_loss": 3.4269628524780273, "perplexity": 30.7830080937705, "lr": 0.0026291804804649314, "grad_norm": 0.190646, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:47.240841+00:00", "epoch": 0, "step": 22431, "train_loss": 3.5190799236297607, "perplexity": 33.7533585050972, "lr": 0.0026291804804649314, "grad_norm": 0.168109, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:47.545045+00:00", "epoch": 0, "step": 22432, "train_loss": 3.4220449924468994, "perplexity": 30.63199320802578, "lr": 0.0026291804804649314, "grad_norm": 0.217889, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:47.850495+00:00", "epoch": 0, "step": 22433, "train_loss": 3.484395742416382, "perplexity": 32.60272071535104, "lr": 0.0026291804804649314, "grad_norm": 0.17687, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:48.154890+00:00", "epoch": 0, "step": 22434, "train_loss": 3.4949727058410645, "perplexity": 32.94938861488164, "lr": 0.0026291804804649314, "grad_norm": 0.171476, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:48.459357+00:00", "epoch": 0, "step": 22435, "train_loss": 3.5710787773132324, "perplexity": 35.554928320401196, "lr": 0.0026291804804649314, "grad_norm": 0.213982, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:48.763983+00:00", "epoch": 0, "step": 22436, "train_loss": 3.5080275535583496, "perplexity": 33.38235789041137, "lr": 0.0026291804804649314, "grad_norm": 0.196636, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:49.068452+00:00", "epoch": 0, "step": 22437, "train_loss": 3.5073230266571045, "perplexity": 33.35884740410808, "lr": 0.0026291804804649314, "grad_norm": 0.188502, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:49.372207+00:00", "epoch": 0, "step": 22438, "train_loss": 3.4039435386657715, "perplexity": 30.082497934402983, "lr": 0.0026291804804649314, "grad_norm": 0.212024, "tokens_per_sec": 107869} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:49.676535+00:00", "epoch": 0, "step": 22439, "train_loss": 3.4449000358581543, "perplexity": 31.34015039883584, "lr": 0.0026291804804649314, "grad_norm": 0.23627, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:49.981962+00:00", "epoch": 0, "step": 22440, "train_loss": 3.4969797134399414, "perplexity": 33.01558469400671, "lr": 0.0026291804804649314, "grad_norm": 0.20938, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:50.287582+00:00", "epoch": 0, "step": 22441, "train_loss": 3.5158228874206543, "perplexity": 33.64360143267453, "lr": 0.0026291804804649314, "grad_norm": 0.183727, "tokens_per_sec": 107143} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:50.591951+00:00", "epoch": 0, "step": 22442, "train_loss": 3.440913677215576, "perplexity": 31.215466002907235, "lr": 0.0026291804804649314, "grad_norm": 0.182906, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:50.895748+00:00", "epoch": 0, "step": 22443, "train_loss": 3.47621488571167, "perplexity": 32.33709055161924, "lr": 0.0026291804804649314, "grad_norm": 0.193273, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:51.199663+00:00", "epoch": 0, "step": 22444, "train_loss": 3.5160231590270996, "perplexity": 33.65033996552593, "lr": 0.0026291804804649314, "grad_norm": 0.178708, "tokens_per_sec": 107811} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:51.504830+00:00", "epoch": 0, "step": 22445, "train_loss": 3.5513203144073486, "perplexity": 34.85931236941943, "lr": 0.0026291804804649314, "grad_norm": 0.176778, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:51.809793+00:00", "epoch": 0, "step": 22446, "train_loss": 3.504531145095825, "perplexity": 33.26584334170929, "lr": 0.0026291804804649314, "grad_norm": 0.160838, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:52.114855+00:00", "epoch": 0, "step": 22447, "train_loss": 3.4992499351501465, "perplexity": 33.09062253519407, "lr": 0.0026291804804649314, "grad_norm": 0.178297, "tokens_per_sec": 107415} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:52.418267+00:00", "epoch": 0, "step": 22448, "train_loss": 3.4364266395568848, "perplexity": 31.07571480034636, "lr": 0.0026291804804649314, "grad_norm": 0.152487, "tokens_per_sec": 108003} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:52.722161+00:00", "epoch": 0, "step": 22449, "train_loss": 3.424485445022583, "perplexity": 30.706840428146215, "lr": 0.0026291804804649314, "grad_norm": 0.157266, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:53.027728+00:00", "epoch": 0, "step": 22450, "train_loss": 3.5661683082580566, "perplexity": 35.38076490711308, "lr": 0.0026291804804649314, "grad_norm": 0.157695, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:53.333766+00:00", "epoch": 0, "step": 22451, "train_loss": 3.5581815242767334, "perplexity": 35.099311829298685, "lr": 0.0026291804804649314, "grad_norm": 0.16691, "tokens_per_sec": 106946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:53.639473+00:00", "epoch": 0, "step": 22452, "train_loss": 3.449354410171509, "perplexity": 31.480062538932323, "lr": 0.0026291804804649314, "grad_norm": 0.189598, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:53.944394+00:00", "epoch": 0, "step": 22453, "train_loss": 3.442044734954834, "perplexity": 31.250792471686246, "lr": 0.0026291804804649314, "grad_norm": 0.156361, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:54.248697+00:00", "epoch": 0, "step": 22454, "train_loss": 3.4607386589050293, "perplexity": 31.84048708978021, "lr": 0.0026291804804649314, "grad_norm": 0.178294, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:54.552640+00:00", "epoch": 0, "step": 22455, "train_loss": 3.383420705795288, "perplexity": 29.47141191339777, "lr": 0.0026291804804649314, "grad_norm": 0.164393, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:54.858352+00:00", "epoch": 0, "step": 22456, "train_loss": 3.5263748168945312, "perplexity": 34.00048594093717, "lr": 0.0026291804804649314, "grad_norm": 0.192964, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:55.164625+00:00", "epoch": 0, "step": 22457, "train_loss": 3.4052364826202393, "perplexity": 30.121418073599987, "lr": 0.0026291804804649314, "grad_norm": 0.173636, "tokens_per_sec": 106990} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:55.469765+00:00", "epoch": 0, "step": 22458, "train_loss": 3.4250357151031494, "perplexity": 30.72374213352746, "lr": 0.0026291804804649314, "grad_norm": 0.167365, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:55.774200+00:00", "epoch": 0, "step": 22459, "train_loss": 3.5018973350524902, "perplexity": 33.178342710019486, "lr": 0.0026291804804649314, "grad_norm": 0.177614, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:56.078248+00:00", "epoch": 0, "step": 22460, "train_loss": 3.558516502380371, "perplexity": 35.111071299686834, "lr": 0.0026291804804649314, "grad_norm": 0.181106, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:56.383196+00:00", "epoch": 0, "step": 22461, "train_loss": 3.3630478382110596, "perplexity": 28.877069527192788, "lr": 0.0026291804804649314, "grad_norm": 0.178951, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:56.688586+00:00", "epoch": 0, "step": 22462, "train_loss": 3.574488639831543, "perplexity": 35.676372674426496, "lr": 0.0026291804804649314, "grad_norm": 0.156792, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:56.993290+00:00", "epoch": 0, "step": 22463, "train_loss": 3.489870071411133, "perplexity": 32.78168815103673, "lr": 0.0026291804804649314, "grad_norm": 0.168895, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:57.296895+00:00", "epoch": 0, "step": 22464, "train_loss": 3.5208566188812256, "perplexity": 33.81338124214093, "lr": 0.0026291804804649314, "grad_norm": 0.157336, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:57.602085+00:00", "epoch": 0, "step": 22465, "train_loss": 3.531482458114624, "perplexity": 34.174592482799795, "lr": 0.0026291804804649314, "grad_norm": 0.165651, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:57.907608+00:00", "epoch": 0, "step": 22466, "train_loss": 3.3809361457824707, "perplexity": 29.398279310627572, "lr": 0.0026291804804649314, "grad_norm": 0.147018, "tokens_per_sec": 107253} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:58.213222+00:00", "epoch": 0, "step": 22467, "train_loss": 3.4800875186920166, "perplexity": 32.4625630325997, "lr": 0.0026291804804649314, "grad_norm": 0.162997, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:58.518379+00:00", "epoch": 0, "step": 22468, "train_loss": 3.4932026863098145, "perplexity": 32.89111913775988, "lr": 0.0026291804804649314, "grad_norm": 0.146229, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:58.822960+00:00", "epoch": 0, "step": 22469, "train_loss": 3.4693732261657715, "perplexity": 32.11660628653919, "lr": 0.0026291804804649314, "grad_norm": 0.172687, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:59.127125+00:00", "epoch": 0, "step": 22470, "train_loss": 3.568554401397705, "perplexity": 35.46528750682909, "lr": 0.0026291804804649314, "grad_norm": 0.1557, "tokens_per_sec": 107788} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:59.431729+00:00", "epoch": 0, "step": 22471, "train_loss": 3.4324326515197754, "perplexity": 30.951846296512272, "lr": 0.0026291804804649314, "grad_norm": 0.169718, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:55:59.736033+00:00", "epoch": 0, "step": 22472, "train_loss": 3.517176866531372, "perplexity": 33.689185018871555, "lr": 0.0026291804804649314, "grad_norm": 0.182848, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:00.042113+00:00", "epoch": 0, "step": 22473, "train_loss": 3.4507017135620117, "perplexity": 31.522504318480895, "lr": 0.0026291804804649314, "grad_norm": 0.199546, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:00.346528+00:00", "epoch": 0, "step": 22474, "train_loss": 3.4999985694885254, "perplexity": 33.115404586692186, "lr": 0.0026291804804649314, "grad_norm": 0.189053, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:00.650837+00:00", "epoch": 0, "step": 22475, "train_loss": 3.5537822246551514, "perplexity": 34.94523859570167, "lr": 0.0026291804804649314, "grad_norm": 0.192201, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:00.954783+00:00", "epoch": 0, "step": 22476, "train_loss": 3.472438097000122, "perplexity": 32.21519053326601, "lr": 0.0026291804804649314, "grad_norm": 0.235894, "tokens_per_sec": 107809} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:01.260661+00:00", "epoch": 0, "step": 22477, "train_loss": 3.3454322814941406, "perplexity": 28.37283806889125, "lr": 0.0026291804804649314, "grad_norm": 0.192865, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:01.566353+00:00", "epoch": 0, "step": 22478, "train_loss": 3.5118117332458496, "perplexity": 33.50892205079019, "lr": 0.0026291804804649314, "grad_norm": 0.191696, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:01.870453+00:00", "epoch": 0, "step": 22479, "train_loss": 3.4387917518615723, "perplexity": 31.14929933934577, "lr": 0.0026291804804649314, "grad_norm": 0.20937, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:02.174659+00:00", "epoch": 0, "step": 22480, "train_loss": 3.5042619705200195, "perplexity": 33.25689022746839, "lr": 0.0026291804804649314, "grad_norm": 0.169805, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:02.479408+00:00", "epoch": 0, "step": 22481, "train_loss": 3.4419572353363037, "perplexity": 31.248058158893613, "lr": 0.0026291804804649314, "grad_norm": 0.18711, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:02.784589+00:00", "epoch": 0, "step": 22482, "train_loss": 3.5089330673217773, "perplexity": 33.412599765087265, "lr": 0.0026291804804649314, "grad_norm": 0.188234, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:03.089751+00:00", "epoch": 0, "step": 22483, "train_loss": 3.495147228240967, "perplexity": 32.95513952307458, "lr": 0.0026291804804649314, "grad_norm": 0.207876, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:03.393716+00:00", "epoch": 0, "step": 22484, "train_loss": 3.556988000869751, "perplexity": 35.05744496857051, "lr": 0.0026291804804649314, "grad_norm": 0.189141, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:03.697134+00:00", "epoch": 0, "step": 22485, "train_loss": 3.5474133491516113, "perplexity": 34.723383953856825, "lr": 0.0026291804804649314, "grad_norm": 0.197192, "tokens_per_sec": 107996} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:04.000818+00:00", "epoch": 0, "step": 22486, "train_loss": 3.41766619682312, "perplexity": 30.498155208690818, "lr": 0.0026291804804649314, "grad_norm": 0.167638, "tokens_per_sec": 107903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:04.305402+00:00", "epoch": 0, "step": 22487, "train_loss": 3.474921464920044, "perplexity": 32.29529212366415, "lr": 0.0026291804804649314, "grad_norm": 0.195472, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:04.610208+00:00", "epoch": 0, "step": 22488, "train_loss": 3.521973133087158, "perplexity": 33.85115544644213, "lr": 0.0026291804804649314, "grad_norm": 0.201087, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:04.914438+00:00", "epoch": 0, "step": 22489, "train_loss": 3.497398853302002, "perplexity": 33.02942574208607, "lr": 0.0026291804804649314, "grad_norm": 0.179285, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:05.219145+00:00", "epoch": 0, "step": 22490, "train_loss": 3.4569737911224365, "perplexity": 31.72083724001479, "lr": 0.0026291804804649314, "grad_norm": 0.217982, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:05.523308+00:00", "epoch": 0, "step": 22491, "train_loss": 3.497469186782837, "perplexity": 33.03174889826536, "lr": 0.0026291804804649314, "grad_norm": 0.186742, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:05.829219+00:00", "epoch": 0, "step": 22492, "train_loss": 3.3753244876861572, "perplexity": 29.233768240170562, "lr": 0.0026291804804649314, "grad_norm": 0.192756, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:06.135238+00:00", "epoch": 0, "step": 22493, "train_loss": 3.574760913848877, "perplexity": 35.68608774625927, "lr": 0.0026291804804649314, "grad_norm": 0.169046, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:06.440460+00:00", "epoch": 0, "step": 22494, "train_loss": 3.5204362869262695, "perplexity": 33.79917138413765, "lr": 0.0026291804804649314, "grad_norm": 0.18811, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:06.745383+00:00", "epoch": 0, "step": 22495, "train_loss": 3.5573298931121826, "perplexity": 35.069432886215985, "lr": 0.0026291804804649314, "grad_norm": 0.16201, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:07.050300+00:00", "epoch": 0, "step": 22496, "train_loss": 3.4191465377807617, "perplexity": 30.543336310445614, "lr": 0.0026291804804649314, "grad_norm": 0.154523, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:07.356830+00:00", "epoch": 0, "step": 22497, "train_loss": 3.5239014625549316, "perplexity": 33.916494604472874, "lr": 0.0026291804804649314, "grad_norm": 0.181539, "tokens_per_sec": 106901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:07.661025+00:00", "epoch": 0, "step": 22498, "train_loss": 3.5783350467681885, "perplexity": 35.813862773679475, "lr": 0.0026291804804649314, "grad_norm": 0.18415, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:07.966932+00:00", "epoch": 0, "step": 22499, "train_loss": 3.5766286849975586, "perplexity": 35.752803476824205, "lr": 0.0026291804804649314, "grad_norm": 0.159815, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:08.272332+00:00", "epoch": 0, "step": 22500, "train_loss": 3.406219959259033, "perplexity": 30.15105635649323, "lr": 0.0026291804804649314, "grad_norm": 0.169735, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:56:11.345826+00:00", "step": 22500, "epoch": 0, "val_loss": 3.451837730407715, "val_ppl": 31.5583347625461, "eval_train_loss": 3.406219959259033, "eval_train_ppl": 30.15105635649323} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:56:12.284114+00:00", "step": 22500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4518_epoch_0000_step_0022500.pt", "val_loss": 3.451837730407715} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T22:56:14.612997+00:00", "step": 22500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0022500.pt", "category": "last"} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:15.968790+00:00", "epoch": 0, "step": 22501, "train_loss": 3.4270758628845215, "perplexity": 30.78648709060601, "lr": 0.0026291804804649314, "grad_norm": 0.176097, "tokens_per_sec": 4257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:16.272475+00:00", "epoch": 0, "step": 22502, "train_loss": 3.5351345539093018, "perplexity": 34.299629552993544, "lr": 0.0026291804804649314, "grad_norm": 0.171932, "tokens_per_sec": 107896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:16.576835+00:00", "epoch": 0, "step": 22503, "train_loss": 3.4615280628204346, "perplexity": 31.865632018404078, "lr": 0.0026291804804649314, "grad_norm": 0.174324, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:17.063261+00:00", "epoch": 0, "step": 22504, "train_loss": 3.5616085529327393, "perplexity": 35.21980452477646, "lr": 0.0026291804804649314, "grad_norm": 0.196804, "tokens_per_sec": 67365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:17.367449+00:00", "epoch": 0, "step": 22505, "train_loss": 3.552581787109375, "perplexity": 34.90331418809635, "lr": 0.0026291804804649314, "grad_norm": 0.166393, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:17.672196+00:00", "epoch": 0, "step": 22506, "train_loss": 3.5200963020324707, "perplexity": 33.78768212964128, "lr": 0.0026291804804649314, "grad_norm": 0.192748, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:17.976533+00:00", "epoch": 0, "step": 22507, "train_loss": 3.4464333057403564, "perplexity": 31.38824016542018, "lr": 0.0026291804804649314, "grad_norm": 0.185646, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:18.279987+00:00", "epoch": 0, "step": 22508, "train_loss": 3.421969413757324, "perplexity": 30.629678169604926, "lr": 0.0026291804804649314, "grad_norm": 0.168099, "tokens_per_sec": 107983} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:18.584146+00:00", "epoch": 0, "step": 22509, "train_loss": 3.4945058822631836, "perplexity": 32.9340106530763, "lr": 0.0026291804804649314, "grad_norm": 0.189601, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:18.887594+00:00", "epoch": 0, "step": 22510, "train_loss": 3.5433425903320312, "perplexity": 34.58232074433054, "lr": 0.0026291804804649314, "grad_norm": 0.167595, "tokens_per_sec": 108048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:19.191604+00:00", "epoch": 0, "step": 22511, "train_loss": 3.40771746635437, "perplexity": 30.196241601487376, "lr": 0.0026291804804649314, "grad_norm": 0.176224, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:19.495463+00:00", "epoch": 0, "step": 22512, "train_loss": 3.5712785720825195, "perplexity": 35.562032718789176, "lr": 0.0026291804804649314, "grad_norm": 0.162566, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:19.798596+00:00", "epoch": 0, "step": 22513, "train_loss": 3.524167776107788, "perplexity": 33.925528229486424, "lr": 0.0026291804804649314, "grad_norm": 0.207646, "tokens_per_sec": 108098} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:20.101682+00:00", "epoch": 0, "step": 22514, "train_loss": 3.560328722000122, "perplexity": 35.1747579616461, "lr": 0.0026291804804649314, "grad_norm": 0.187259, "tokens_per_sec": 108115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:20.405842+00:00", "epoch": 0, "step": 22515, "train_loss": 3.3620874881744385, "perplexity": 28.849350744412224, "lr": 0.0026291804804649314, "grad_norm": 0.203055, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:20.710675+00:00", "epoch": 0, "step": 22516, "train_loss": 3.4989473819732666, "perplexity": 33.080612376600115, "lr": 0.0026291804804649314, "grad_norm": 0.170048, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:21.015295+00:00", "epoch": 0, "step": 22517, "train_loss": 3.5097219944000244, "perplexity": 33.438970270631756, "lr": 0.0026291804804649314, "grad_norm": 0.175861, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:21.318368+00:00", "epoch": 0, "step": 22518, "train_loss": 3.4892678260803223, "perplexity": 32.76195147616836, "lr": 0.0026291804804649314, "grad_norm": 0.186914, "tokens_per_sec": 108111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:21.621559+00:00", "epoch": 0, "step": 22519, "train_loss": 3.6177585124969482, "perplexity": 37.25396985728436, "lr": 0.0026291804804649314, "grad_norm": 0.156473, "tokens_per_sec": 108078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:21.925193+00:00", "epoch": 0, "step": 22520, "train_loss": 3.6046195030212402, "perplexity": 36.767691199913536, "lr": 0.0026291804804649314, "grad_norm": 0.18048, "tokens_per_sec": 107987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:22.227896+00:00", "epoch": 0, "step": 22521, "train_loss": 3.4989657402038574, "perplexity": 33.081219683684736, "lr": 0.0026291804804649314, "grad_norm": 0.149788, "tokens_per_sec": 108182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:22.533413+00:00", "epoch": 0, "step": 22522, "train_loss": 3.4745233058929443, "perplexity": 32.282436021128774, "lr": 0.0026291804804649314, "grad_norm": 0.190539, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:22.837559+00:00", "epoch": 0, "step": 22523, "train_loss": 3.4732675552368164, "perplexity": 32.24192277350628, "lr": 0.0026291804804649314, "grad_norm": 0.158148, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:23.140997+00:00", "epoch": 0, "step": 22524, "train_loss": 3.541792392730713, "perplexity": 34.5287528448068, "lr": 0.0026291804804649314, "grad_norm": 0.167854, "tokens_per_sec": 107989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:23.444894+00:00", "epoch": 0, "step": 22525, "train_loss": 3.5201547145843506, "perplexity": 33.78965581201993, "lr": 0.0026291804804649314, "grad_norm": 0.165715, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:23.749012+00:00", "epoch": 0, "step": 22526, "train_loss": 3.467663526535034, "perplexity": 32.06174344948069, "lr": 0.0026291804804649314, "grad_norm": 0.174393, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:24.053259+00:00", "epoch": 0, "step": 22527, "train_loss": 3.454465389251709, "perplexity": 31.64136834417311, "lr": 0.0026291804804649314, "grad_norm": 0.17799, "tokens_per_sec": 107702} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:24.357573+00:00", "epoch": 0, "step": 22528, "train_loss": 3.4721453189849854, "perplexity": 32.205760014316105, "lr": 0.0026291804804649314, "grad_norm": 0.205788, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:24.660856+00:00", "epoch": 0, "step": 22529, "train_loss": 3.604342222213745, "perplexity": 36.757497638112724, "lr": 0.0026291804804649314, "grad_norm": 0.161046, "tokens_per_sec": 108038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:24.964886+00:00", "epoch": 0, "step": 22530, "train_loss": 3.508232355117798, "perplexity": 33.38919534950263, "lr": 0.0026291804804649314, "grad_norm": 0.162622, "tokens_per_sec": 107875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:25.268453+00:00", "epoch": 0, "step": 22531, "train_loss": 3.4825754165649414, "perplexity": 32.54342712320776, "lr": 0.0026291804804649314, "grad_norm": 0.152649, "tokens_per_sec": 107848} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:25.572676+00:00", "epoch": 0, "step": 22532, "train_loss": 3.4894182682037354, "perplexity": 32.76688062448207, "lr": 0.0026291804804649314, "grad_norm": 0.17222, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:25.877255+00:00", "epoch": 0, "step": 22533, "train_loss": 3.4592113494873047, "perplexity": 31.791893931822944, "lr": 0.0026291804804649314, "grad_norm": 0.190927, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:26.181022+00:00", "epoch": 0, "step": 22534, "train_loss": 3.374183177947998, "perplexity": 29.200422488330265, "lr": 0.0026291804804649314, "grad_norm": 0.186785, "tokens_per_sec": 107872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:26.484754+00:00", "epoch": 0, "step": 22535, "train_loss": 3.544525384902954, "perplexity": 34.62324872548227, "lr": 0.0026291804804649314, "grad_norm": 0.17963, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:26.790146+00:00", "epoch": 0, "step": 22536, "train_loss": 3.42594313621521, "perplexity": 30.75163415877332, "lr": 0.0026291804804649314, "grad_norm": 0.181077, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:27.095505+00:00", "epoch": 0, "step": 22537, "train_loss": 3.524782419204712, "perplexity": 33.94638673083243, "lr": 0.0026291804804649314, "grad_norm": 0.208598, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:27.400630+00:00", "epoch": 0, "step": 22538, "train_loss": 3.4331679344177246, "perplexity": 30.97461302872275, "lr": 0.0026291804804649314, "grad_norm": 0.157681, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:27.704602+00:00", "epoch": 0, "step": 22539, "train_loss": 3.51473331451416, "perplexity": 33.60696423914786, "lr": 0.0026291804804649314, "grad_norm": 0.173549, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:28.007810+00:00", "epoch": 0, "step": 22540, "train_loss": 3.3785622119903564, "perplexity": 29.328572514338095, "lr": 0.0026291804804649314, "grad_norm": 0.18196, "tokens_per_sec": 108128} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:28.312439+00:00", "epoch": 0, "step": 22541, "train_loss": 3.5038132667541504, "perplexity": 33.2419710829707, "lr": 0.0026291804804649314, "grad_norm": 0.166997, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:28.617387+00:00", "epoch": 0, "step": 22542, "train_loss": 3.4500341415405273, "perplexity": 31.50146779902791, "lr": 0.0026291804804649314, "grad_norm": 0.16363, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:28.922835+00:00", "epoch": 0, "step": 22543, "train_loss": 3.4861743450164795, "perplexity": 32.66075959793921, "lr": 0.0026291804804649314, "grad_norm": 0.168592, "tokens_per_sec": 107279} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:29.227599+00:00", "epoch": 0, "step": 22544, "train_loss": 3.589595079421997, "perplexity": 36.219406973589784, "lr": 0.0026291804804649314, "grad_norm": 0.180318, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:29.531379+00:00", "epoch": 0, "step": 22545, "train_loss": 3.4552204608917236, "perplexity": 31.665268866228704, "lr": 0.0026291804804649314, "grad_norm": 0.173432, "tokens_per_sec": 107868} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:29.835523+00:00", "epoch": 0, "step": 22546, "train_loss": 3.4587411880493164, "perplexity": 31.7769501225323, "lr": 0.0026291804804649314, "grad_norm": 0.160426, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:30.139613+00:00", "epoch": 0, "step": 22547, "train_loss": 3.49314284324646, "perplexity": 32.889150891327056, "lr": 0.0026291804804649314, "grad_norm": 0.171979, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:30.444781+00:00", "epoch": 0, "step": 22548, "train_loss": 3.5339348316192627, "perplexity": 34.258504197314444, "lr": 0.0026291804804649314, "grad_norm": 0.215857, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:30.749495+00:00", "epoch": 0, "step": 22549, "train_loss": 3.6035282611846924, "perplexity": 36.727590640726184, "lr": 0.0026291804804649314, "grad_norm": 0.19115, "tokens_per_sec": 107537} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:31.053504+00:00", "epoch": 0, "step": 22550, "train_loss": 3.50179386138916, "perplexity": 33.17490980296686, "lr": 0.0026291804804649314, "grad_norm": 0.173983, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:31.357853+00:00", "epoch": 0, "step": 22551, "train_loss": 3.375763177871704, "perplexity": 29.246595620801354, "lr": 0.0026291804804649314, "grad_norm": 0.185063, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:31.662550+00:00", "epoch": 0, "step": 22552, "train_loss": 3.45021653175354, "perplexity": 31.50721388244875, "lr": 0.0026291804804649314, "grad_norm": 0.175332, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:31.966819+00:00", "epoch": 0, "step": 22553, "train_loss": 3.4289348125457764, "perplexity": 30.843770847668452, "lr": 0.0026291804804649314, "grad_norm": 0.164763, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:32.271538+00:00", "epoch": 0, "step": 22554, "train_loss": 3.5162200927734375, "perplexity": 33.65696750561261, "lr": 0.0026291804804649314, "grad_norm": 0.22132, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:32.576288+00:00", "epoch": 0, "step": 22555, "train_loss": 3.390329599380493, "perplexity": 29.675731761427468, "lr": 0.0026291804804649314, "grad_norm": 0.183844, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:32.880678+00:00", "epoch": 0, "step": 22556, "train_loss": 3.47830867767334, "perplexity": 32.40486862370988, "lr": 0.0026291804804649314, "grad_norm": 0.186177, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:33.184556+00:00", "epoch": 0, "step": 22557, "train_loss": 3.4991073608398438, "perplexity": 33.08590499881637, "lr": 0.0026291804804649314, "grad_norm": 0.198345, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:33.489092+00:00", "epoch": 0, "step": 22558, "train_loss": 3.4937541484832764, "perplexity": 32.90926234799446, "lr": 0.0026291804804649314, "grad_norm": 0.194099, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:33.793945+00:00", "epoch": 0, "step": 22559, "train_loss": 3.416790723800659, "perplexity": 30.47146658086225, "lr": 0.0026291804804649314, "grad_norm": 0.186583, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:34.099241+00:00", "epoch": 0, "step": 22560, "train_loss": 3.5074942111968994, "perplexity": 33.36455841185321, "lr": 0.0026291804804649314, "grad_norm": 0.195538, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:34.402957+00:00", "epoch": 0, "step": 22561, "train_loss": 3.4982292652130127, "perplexity": 33.05686516206992, "lr": 0.0026291804804649314, "grad_norm": 0.160611, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:34.707363+00:00", "epoch": 0, "step": 22562, "train_loss": 3.5095531940460205, "perplexity": 33.43332623698339, "lr": 0.0026291804804649314, "grad_norm": 0.192742, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:35.011254+00:00", "epoch": 0, "step": 22563, "train_loss": 3.5175771713256836, "perplexity": 33.70267366075494, "lr": 0.0026291804804649314, "grad_norm": 0.207934, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:35.316466+00:00", "epoch": 0, "step": 22564, "train_loss": 3.5510358810424805, "perplexity": 34.84939862787147, "lr": 0.0026291804804649314, "grad_norm": 0.179574, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:35.621014+00:00", "epoch": 0, "step": 22565, "train_loss": 3.4589152336120605, "perplexity": 31.782481241018022, "lr": 0.0026291804804649314, "grad_norm": 0.17542, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:35.924541+00:00", "epoch": 0, "step": 22566, "train_loss": 3.4980318546295166, "perplexity": 33.050340031115866, "lr": 0.0026291804804649314, "grad_norm": 0.187262, "tokens_per_sec": 107958} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:36.227850+00:00", "epoch": 0, "step": 22567, "train_loss": 3.5363924503326416, "perplexity": 34.34280208192233, "lr": 0.0026291804804649314, "grad_norm": 0.187218, "tokens_per_sec": 108035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:36.532524+00:00", "epoch": 0, "step": 22568, "train_loss": 3.5408785343170166, "perplexity": 34.49721286723709, "lr": 0.0026291804804649314, "grad_norm": 0.172502, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:36.838736+00:00", "epoch": 0, "step": 22569, "train_loss": 3.559441328048706, "perplexity": 35.14355793959126, "lr": 0.0026291804804649314, "grad_norm": 0.16877, "tokens_per_sec": 107010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:37.143401+00:00", "epoch": 0, "step": 22570, "train_loss": 3.4174976348876953, "perplexity": 30.493014813871575, "lr": 0.0026291804804649314, "grad_norm": 0.177664, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:37.447550+00:00", "epoch": 0, "step": 22571, "train_loss": 3.428328037261963, "perplexity": 30.825061286682104, "lr": 0.0026291804804649314, "grad_norm": 0.182471, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:37.751587+00:00", "epoch": 0, "step": 22572, "train_loss": 3.599228858947754, "perplexity": 36.570022921588176, "lr": 0.0026291804804649314, "grad_norm": 0.179935, "tokens_per_sec": 107775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:38.055300+00:00", "epoch": 0, "step": 22573, "train_loss": 3.470919132232666, "perplexity": 32.16629393937219, "lr": 0.0026291804804649314, "grad_norm": 0.151497, "tokens_per_sec": 107900} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:38.359448+00:00", "epoch": 0, "step": 22574, "train_loss": 3.4973530769348145, "perplexity": 33.02791380957096, "lr": 0.0026291804804649314, "grad_norm": 0.174229, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:38.664192+00:00", "epoch": 0, "step": 22575, "train_loss": 3.5144236087799072, "perplexity": 33.596557581196166, "lr": 0.0026291804804649314, "grad_norm": 0.180253, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:38.969441+00:00", "epoch": 0, "step": 22576, "train_loss": 3.5744950771331787, "perplexity": 35.676602334737865, "lr": 0.0026291804804649314, "grad_norm": 0.180382, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:39.273597+00:00", "epoch": 0, "step": 22577, "train_loss": 3.5040714740753174, "perplexity": 33.2505555115078, "lr": 0.0026291804804649314, "grad_norm": 0.170389, "tokens_per_sec": 107736} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:39.577553+00:00", "epoch": 0, "step": 22578, "train_loss": 3.4781174659729004, "perplexity": 32.398673026032085, "lr": 0.0026291804804649314, "grad_norm": 0.18143, "tokens_per_sec": 107805} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:39.883208+00:00", "epoch": 0, "step": 22579, "train_loss": 3.4375576972961426, "perplexity": 31.110883113021163, "lr": 0.0026291804804649314, "grad_norm": 0.18533, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:40.186829+00:00", "epoch": 0, "step": 22580, "train_loss": 3.4661741256713867, "perplexity": 32.01402620497057, "lr": 0.0026291804804649314, "grad_norm": 0.173241, "tokens_per_sec": 108013} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:40.491665+00:00", "epoch": 0, "step": 22581, "train_loss": 3.4267845153808594, "perplexity": 30.777518830949106, "lr": 0.0026291804804649314, "grad_norm": 0.161128, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:40.795901+00:00", "epoch": 0, "step": 22582, "train_loss": 3.520352840423584, "perplexity": 33.79635107916591, "lr": 0.0026291804804649314, "grad_norm": 0.161751, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:41.099764+00:00", "epoch": 0, "step": 22583, "train_loss": 3.4099347591400146, "perplexity": 30.263269793245566, "lr": 0.0026291804804649314, "grad_norm": 0.165288, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:41.404656+00:00", "epoch": 0, "step": 22584, "train_loss": 3.453745126724243, "perplexity": 31.61858645769152, "lr": 0.0026291804804649314, "grad_norm": 0.156645, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:41.710385+00:00", "epoch": 0, "step": 22585, "train_loss": 3.437692165374756, "perplexity": 31.115066814978213, "lr": 0.0026291804804649314, "grad_norm": 0.172159, "tokens_per_sec": 107184} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:42.015855+00:00", "epoch": 0, "step": 22586, "train_loss": 3.4752395153045654, "perplexity": 32.30556528734754, "lr": 0.0026291804804649314, "grad_norm": 0.171432, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:42.320093+00:00", "epoch": 0, "step": 22587, "train_loss": 3.632155179977417, "perplexity": 37.79418217166013, "lr": 0.0026291804804649314, "grad_norm": 0.202183, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:42.623504+00:00", "epoch": 0, "step": 22588, "train_loss": 3.568777561187744, "perplexity": 35.473202816099324, "lr": 0.0026291804804649314, "grad_norm": 0.153199, "tokens_per_sec": 107997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:42.927675+00:00", "epoch": 0, "step": 22589, "train_loss": 3.5639326572418213, "perplexity": 35.30175421716679, "lr": 0.0026291804804649314, "grad_norm": 0.205967, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:43.233116+00:00", "epoch": 0, "step": 22590, "train_loss": 3.4960896968841553, "perplexity": 32.98621334946037, "lr": 0.0026291804804649314, "grad_norm": 0.203025, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:43.537473+00:00", "epoch": 0, "step": 22591, "train_loss": 3.5252161026000977, "perplexity": 33.96111190789254, "lr": 0.0026291804804649314, "grad_norm": 0.187386, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:43.842124+00:00", "epoch": 0, "step": 22592, "train_loss": 3.477457284927368, "perplexity": 32.3772910949493, "lr": 0.0026291804804649314, "grad_norm": 0.190088, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:44.146466+00:00", "epoch": 0, "step": 22593, "train_loss": 3.559070587158203, "perplexity": 35.13053120054692, "lr": 0.0026291804804649314, "grad_norm": 0.215019, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:44.451295+00:00", "epoch": 0, "step": 22594, "train_loss": 3.5927443504333496, "perplexity": 36.33365150108721, "lr": 0.0026291804804649314, "grad_norm": 0.179121, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:44.755532+00:00", "epoch": 0, "step": 22595, "train_loss": 3.5206775665283203, "perplexity": 33.80732741866168, "lr": 0.0026291804804649314, "grad_norm": 0.183128, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:45.061103+00:00", "epoch": 0, "step": 22596, "train_loss": 3.5495972633361816, "perplexity": 34.799299711186336, "lr": 0.0026291804804649314, "grad_norm": 0.172068, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:45.365262+00:00", "epoch": 0, "step": 22597, "train_loss": 3.534430742263794, "perplexity": 34.275497567459915, "lr": 0.0026291804804649314, "grad_norm": 0.188368, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:45.668817+00:00", "epoch": 0, "step": 22598, "train_loss": 3.6154181957244873, "perplexity": 37.16688570878053, "lr": 0.0026291804804649314, "grad_norm": 0.171289, "tokens_per_sec": 107948} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:45.972691+00:00", "epoch": 0, "step": 22599, "train_loss": 3.464431047439575, "perplexity": 31.958271858990084, "lr": 0.0026291804804649314, "grad_norm": 0.166294, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:46.276644+00:00", "epoch": 0, "step": 22600, "train_loss": 3.457195520401001, "perplexity": 31.727871458188723, "lr": 0.0026291804804649314, "grad_norm": 0.153449, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:46.581725+00:00", "epoch": 0, "step": 22601, "train_loss": 3.4616408348083496, "perplexity": 31.869225771706503, "lr": 0.0026291804804649314, "grad_norm": 0.169997, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:46.887418+00:00", "epoch": 0, "step": 22602, "train_loss": 3.567678689956665, "perplexity": 35.43424374347284, "lr": 0.0026291804804649314, "grad_norm": 0.153987, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:47.192191+00:00", "epoch": 0, "step": 22603, "train_loss": 3.5477588176727295, "perplexity": 34.7353818622901, "lr": 0.0026291804804649314, "grad_norm": 0.162721, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:47.496757+00:00", "epoch": 0, "step": 22604, "train_loss": 3.5296714305877686, "perplexity": 34.11275736454569, "lr": 0.0026291804804649314, "grad_norm": 0.163903, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:47.800201+00:00", "epoch": 0, "step": 22605, "train_loss": 3.474592924118042, "perplexity": 32.2846835452598, "lr": 0.0026291804804649314, "grad_norm": 0.160945, "tokens_per_sec": 107987} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:48.105044+00:00", "epoch": 0, "step": 22606, "train_loss": 3.5948407649993896, "perplexity": 36.40990179551909, "lr": 0.0026291804804649314, "grad_norm": 0.169218, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:48.410952+00:00", "epoch": 0, "step": 22607, "train_loss": 3.439589738845825, "perplexity": 31.17416599510143, "lr": 0.0026291804804649314, "grad_norm": 0.186411, "tokens_per_sec": 107116} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:48.715918+00:00", "epoch": 0, "step": 22608, "train_loss": 3.5147459506988525, "perplexity": 33.60738890563801, "lr": 0.0026291804804649314, "grad_norm": 0.19267, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:49.019993+00:00", "epoch": 0, "step": 22609, "train_loss": 3.423985004425049, "perplexity": 30.69147732305534, "lr": 0.0026291804804649314, "grad_norm": 0.171772, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:49.323984+00:00", "epoch": 0, "step": 22610, "train_loss": 3.4353299140930176, "perplexity": 31.041651954826225, "lr": 0.0026291804804649314, "grad_norm": 0.170518, "tokens_per_sec": 107851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:49.628694+00:00", "epoch": 0, "step": 22611, "train_loss": 3.5496535301208496, "perplexity": 34.801257810977255, "lr": 0.0026291804804649314, "grad_norm": 0.174984, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:49.934301+00:00", "epoch": 0, "step": 22612, "train_loss": 3.5519866943359375, "perplexity": 34.882549657077966, "lr": 0.0026291804804649314, "grad_norm": 0.194631, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:50.239520+00:00", "epoch": 0, "step": 22613, "train_loss": 3.458434581756592, "perplexity": 31.767208603143022, "lr": 0.0026291804804649314, "grad_norm": 0.165688, "tokens_per_sec": 107362} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:50.543397+00:00", "epoch": 0, "step": 22614, "train_loss": 3.5048930644989014, "perplexity": 33.27788507481134, "lr": 0.0026291804804649314, "grad_norm": 0.180255, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:50.848788+00:00", "epoch": 0, "step": 22615, "train_loss": 3.490367889404297, "perplexity": 32.79801152794287, "lr": 0.0026291804804649314, "grad_norm": 0.174121, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:51.152705+00:00", "epoch": 0, "step": 22616, "train_loss": 3.5785155296325684, "perplexity": 35.820327145554, "lr": 0.0026291804804649314, "grad_norm": 0.195417, "tokens_per_sec": 107819} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:51.457572+00:00", "epoch": 0, "step": 22617, "train_loss": 3.4482810497283936, "perplexity": 31.44629121270142, "lr": 0.0026291804804649314, "grad_norm": 0.188948, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:51.762185+00:00", "epoch": 0, "step": 22618, "train_loss": 3.5535597801208496, "perplexity": 34.937466082884725, "lr": 0.0026291804804649314, "grad_norm": 0.160655, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:52.068273+00:00", "epoch": 0, "step": 22619, "train_loss": 3.4725656509399414, "perplexity": 32.2192999698224, "lr": 0.0026291804804649314, "grad_norm": 0.169692, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:52.373515+00:00", "epoch": 0, "step": 22620, "train_loss": 3.4769322872161865, "perplexity": 32.36029755240598, "lr": 0.0026291804804649314, "grad_norm": 0.165223, "tokens_per_sec": 107410} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:52.678426+00:00", "epoch": 0, "step": 22621, "train_loss": 3.523690700531006, "perplexity": 33.909347048668685, "lr": 0.0026291804804649314, "grad_norm": 0.158489, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:52.983551+00:00", "epoch": 0, "step": 22622, "train_loss": 3.4631311893463135, "perplexity": 31.916757627817514, "lr": 0.0026291804804649314, "grad_norm": 0.203265, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:53.288568+00:00", "epoch": 0, "step": 22623, "train_loss": 3.5825047492980957, "perplexity": 35.963507698487696, "lr": 0.0026291804804649314, "grad_norm": 0.170749, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:53.594135+00:00", "epoch": 0, "step": 22624, "train_loss": 3.5094456672668457, "perplexity": 33.42973145236729, "lr": 0.0026291804804649314, "grad_norm": 0.169924, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:53.899159+00:00", "epoch": 0, "step": 22625, "train_loss": 3.4860472679138184, "perplexity": 32.65660942694023, "lr": 0.0026291804804649314, "grad_norm": 0.187702, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:54.203865+00:00", "epoch": 0, "step": 22626, "train_loss": 3.3889083862304688, "perplexity": 29.633586177234733, "lr": 0.0026291804804649314, "grad_norm": 0.164079, "tokens_per_sec": 107541} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:54.508495+00:00", "epoch": 0, "step": 22627, "train_loss": 3.403639316558838, "perplexity": 30.07334756544239, "lr": 0.0026291804804649314, "grad_norm": 0.184665, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:54.814195+00:00", "epoch": 0, "step": 22628, "train_loss": 3.479114294052124, "perplexity": 32.43098503511739, "lr": 0.0026291804804649314, "grad_norm": 0.166427, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:55.119057+00:00", "epoch": 0, "step": 22629, "train_loss": 3.5469439029693604, "perplexity": 34.70708701939098, "lr": 0.0026291804804649314, "grad_norm": 0.184067, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:55.422979+00:00", "epoch": 0, "step": 22630, "train_loss": 3.48307466506958, "perplexity": 32.55967843691932, "lr": 0.0026291804804649314, "grad_norm": 0.186409, "tokens_per_sec": 107886} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:55.726960+00:00", "epoch": 0, "step": 22631, "train_loss": 3.5135912895202637, "perplexity": 33.568606153133196, "lr": 0.0026291804804649314, "grad_norm": 0.175834, "tokens_per_sec": 107727} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:56.030960+00:00", "epoch": 0, "step": 22632, "train_loss": 3.4189612865448, "perplexity": 30.537678643704762, "lr": 0.0026291804804649314, "grad_norm": 0.168831, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:56.336040+00:00", "epoch": 0, "step": 22633, "train_loss": 3.5668272972106934, "perplexity": 35.40408812434992, "lr": 0.0026291804804649314, "grad_norm": 0.174592, "tokens_per_sec": 107408} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:56.641251+00:00", "epoch": 0, "step": 22634, "train_loss": 3.4444167613983154, "perplexity": 31.32500816380301, "lr": 0.0026291804804649314, "grad_norm": 0.177667, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:56.946821+00:00", "epoch": 0, "step": 22635, "train_loss": 3.4744791984558105, "perplexity": 32.28101215701317, "lr": 0.0026291804804649314, "grad_norm": 0.169506, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:57.250856+00:00", "epoch": 0, "step": 22636, "train_loss": 3.569619655609131, "perplexity": 35.50308718326294, "lr": 0.0026291804804649314, "grad_norm": 0.171789, "tokens_per_sec": 107777} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:57.554765+00:00", "epoch": 0, "step": 22637, "train_loss": 3.4882986545562744, "perplexity": 32.730214907299, "lr": 0.0026291804804649314, "grad_norm": 0.2042, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:57.859593+00:00", "epoch": 0, "step": 22638, "train_loss": 3.5407469272613525, "perplexity": 34.49267308936297, "lr": 0.0026291804804649314, "grad_norm": 0.190189, "tokens_per_sec": 107496} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:58.164471+00:00", "epoch": 0, "step": 22639, "train_loss": 3.5509657859802246, "perplexity": 34.84695594271616, "lr": 0.0026291804804649314, "grad_norm": 0.181897, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:58.469972+00:00", "epoch": 0, "step": 22640, "train_loss": 3.4397811889648438, "perplexity": 31.180134864243616, "lr": 0.0026291804804649314, "grad_norm": 0.185463, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:58.773714+00:00", "epoch": 0, "step": 22641, "train_loss": 3.4539167881011963, "perplexity": 31.624014613669196, "lr": 0.0026291804804649314, "grad_norm": 0.213623, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:59.077483+00:00", "epoch": 0, "step": 22642, "train_loss": 3.536285638809204, "perplexity": 34.33913407080926, "lr": 0.0026291804804649314, "grad_norm": 0.184388, "tokens_per_sec": 107871} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:59.381847+00:00", "epoch": 0, "step": 22643, "train_loss": 3.540297031402588, "perplexity": 34.4771584688258, "lr": 0.0026291804804649314, "grad_norm": 0.171937, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:59.687850+00:00", "epoch": 0, "step": 22644, "train_loss": 3.5180139541625977, "perplexity": 33.71739762552154, "lr": 0.0026291804804649314, "grad_norm": 0.205486, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:56:59.993203+00:00", "epoch": 0, "step": 22645, "train_loss": 3.5826573371887207, "perplexity": 35.96899571295851, "lr": 0.0026291804804649314, "grad_norm": 0.178085, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:00.297197+00:00", "epoch": 0, "step": 22646, "train_loss": 3.5254268646240234, "perplexity": 33.96827037491301, "lr": 0.0026291804804649314, "grad_norm": 0.176204, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:00.601716+00:00", "epoch": 0, "step": 22647, "train_loss": 3.3537821769714355, "perplexity": 28.610740147682147, "lr": 0.0026291804804649314, "grad_norm": 0.153471, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:00.905094+00:00", "epoch": 0, "step": 22648, "train_loss": 3.454244375228882, "perplexity": 31.634375930806776, "lr": 0.0026291804804649314, "grad_norm": 0.190146, "tokens_per_sec": 108011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:01.210513+00:00", "epoch": 0, "step": 22649, "train_loss": 3.524580478668213, "perplexity": 33.939532271403706, "lr": 0.0026291804804649314, "grad_norm": 0.191583, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:01.516404+00:00", "epoch": 0, "step": 22650, "train_loss": 3.4625332355499268, "perplexity": 31.897678586186768, "lr": 0.0026291804804649314, "grad_norm": 0.180496, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:01.821698+00:00", "epoch": 0, "step": 22651, "train_loss": 3.5547831058502197, "perplexity": 34.980232137136795, "lr": 0.0026291804804649314, "grad_norm": 0.157888, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:02.126223+00:00", "epoch": 0, "step": 22652, "train_loss": 3.5509345531463623, "perplexity": 34.845867590526844, "lr": 0.0026291804804649314, "grad_norm": 0.194819, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:02.430251+00:00", "epoch": 0, "step": 22653, "train_loss": 3.565598964691162, "perplexity": 35.36062682950786, "lr": 0.0026291804804649314, "grad_norm": 0.189491, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:02.735313+00:00", "epoch": 0, "step": 22654, "train_loss": 3.5026466846466064, "perplexity": 33.20321420522127, "lr": 0.0026291804804649314, "grad_norm": 0.157372, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:03.039959+00:00", "epoch": 0, "step": 22655, "train_loss": 3.336961507797241, "perplexity": 28.133513242733343, "lr": 0.0026291804804649314, "grad_norm": 0.172526, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:03.343980+00:00", "epoch": 0, "step": 22656, "train_loss": 3.5789854526519775, "perplexity": 35.83716389752129, "lr": 0.0026291804804649314, "grad_norm": 0.161416, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:03.647911+00:00", "epoch": 0, "step": 22657, "train_loss": 3.603550672531128, "perplexity": 36.7284137647074, "lr": 0.0026291804804649314, "grad_norm": 0.166858, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:03.952913+00:00", "epoch": 0, "step": 22658, "train_loss": 3.4917898178100586, "perplexity": 32.844681124736276, "lr": 0.0026291804804649314, "grad_norm": 0.168659, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:04.257581+00:00", "epoch": 0, "step": 22659, "train_loss": 3.5176899433135986, "perplexity": 33.70647459257701, "lr": 0.0026291804804649314, "grad_norm": 0.171912, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:04.562801+00:00", "epoch": 0, "step": 22660, "train_loss": 3.474590301513672, "perplexity": 32.28459887541868, "lr": 0.0026291804804649314, "grad_norm": 0.170981, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:04.868327+00:00", "epoch": 0, "step": 22661, "train_loss": 3.4996743202209473, "perplexity": 33.104668681658126, "lr": 0.0026291804804649314, "grad_norm": 0.19836, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:05.172640+00:00", "epoch": 0, "step": 22662, "train_loss": 3.424415349960327, "perplexity": 30.704688105689193, "lr": 0.0026291804804649314, "grad_norm": 0.180161, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:05.477727+00:00", "epoch": 0, "step": 22663, "train_loss": 3.541365146636963, "perplexity": 34.514003721008926, "lr": 0.0026291804804649314, "grad_norm": 0.167181, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:05.782330+00:00", "epoch": 0, "step": 22664, "train_loss": 3.469874620437622, "perplexity": 32.132713406633975, "lr": 0.0026291804804649314, "grad_norm": 0.155018, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:06.086716+00:00", "epoch": 0, "step": 22665, "train_loss": 3.521535873413086, "perplexity": 33.836356936876165, "lr": 0.0026291804804649314, "grad_norm": 0.162641, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:06.391991+00:00", "epoch": 0, "step": 22666, "train_loss": 3.4331729412078857, "perplexity": 30.974768112498744, "lr": 0.0026291804804649314, "grad_norm": 0.179831, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:06.696993+00:00", "epoch": 0, "step": 22667, "train_loss": 3.5284106731414795, "perplexity": 34.0697765515681, "lr": 0.0026291804804649314, "grad_norm": 0.1642, "tokens_per_sec": 107436} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:07.000783+00:00", "epoch": 0, "step": 22668, "train_loss": 3.5204145908355713, "perplexity": 33.7984380822047, "lr": 0.0026291804804649314, "grad_norm": 0.175746, "tokens_per_sec": 107865} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:07.305767+00:00", "epoch": 0, "step": 22669, "train_loss": 3.5428552627563477, "perplexity": 34.56547193157968, "lr": 0.0026291804804649314, "grad_norm": 0.187993, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:07.609822+00:00", "epoch": 0, "step": 22670, "train_loss": 3.4888386726379395, "perplexity": 32.747894588409814, "lr": 0.0026291804804649314, "grad_norm": 0.200027, "tokens_per_sec": 107828} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:07.915277+00:00", "epoch": 0, "step": 22671, "train_loss": 3.5627543926239014, "perplexity": 35.26018390444588, "lr": 0.0026291804804649314, "grad_norm": 0.186479, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:08.219264+00:00", "epoch": 0, "step": 22672, "train_loss": 3.4844167232513428, "perplexity": 32.60340475482948, "lr": 0.0026291804804649314, "grad_norm": 0.178504, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:08.523362+00:00", "epoch": 0, "step": 22673, "train_loss": 3.460857629776001, "perplexity": 31.844275405606595, "lr": 0.0026291804804649314, "grad_norm": 0.165877, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:08.827328+00:00", "epoch": 0, "step": 22674, "train_loss": 3.539534330368042, "perplexity": 34.45087272974784, "lr": 0.0026291804804649314, "grad_norm": 0.224955, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:09.132268+00:00", "epoch": 0, "step": 22675, "train_loss": 3.4390957355499268, "perplexity": 31.15876965758681, "lr": 0.0026291804804649314, "grad_norm": 0.191991, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:09.437589+00:00", "epoch": 0, "step": 22676, "train_loss": 3.4326350688934326, "perplexity": 30.95811212208404, "lr": 0.0026291804804649314, "grad_norm": 0.18069, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:09.740973+00:00", "epoch": 0, "step": 22677, "train_loss": 3.4937963485717773, "perplexity": 32.91065115108165, "lr": 0.0026291804804649314, "grad_norm": 0.160883, "tokens_per_sec": 108009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:10.044357+00:00", "epoch": 0, "step": 22678, "train_loss": 3.518476963043213, "perplexity": 33.73301269473229, "lr": 0.0026291804804649314, "grad_norm": 0.181768, "tokens_per_sec": 108010} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:10.348533+00:00", "epoch": 0, "step": 22679, "train_loss": 3.384601593017578, "perplexity": 29.50623488410302, "lr": 0.0026291804804649314, "grad_norm": 0.167635, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:10.653970+00:00", "epoch": 0, "step": 22680, "train_loss": 3.420074939727783, "perplexity": 30.57170597053087, "lr": 0.0026291804804649314, "grad_norm": 0.166894, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:10.958277+00:00", "epoch": 0, "step": 22681, "train_loss": 3.543820858001709, "perplexity": 34.598864306094995, "lr": 0.0026291804804649314, "grad_norm": 0.170478, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:11.262801+00:00", "epoch": 0, "step": 22682, "train_loss": 3.465327262878418, "perplexity": 31.986926193939595, "lr": 0.0026291804804649314, "grad_norm": 0.167593, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:11.567186+00:00", "epoch": 0, "step": 22683, "train_loss": 3.451401710510254, "perplexity": 31.544577700053328, "lr": 0.0026291804804649314, "grad_norm": 0.156012, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:11.872370+00:00", "epoch": 0, "step": 22684, "train_loss": 3.5683887004852295, "perplexity": 35.45941136318263, "lr": 0.0026291804804649314, "grad_norm": 0.176163, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:12.177767+00:00", "epoch": 0, "step": 22685, "train_loss": 3.6554505825042725, "perplexity": 38.68494794947795, "lr": 0.0026291804804649314, "grad_norm": 0.169129, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:12.483547+00:00", "epoch": 0, "step": 22686, "train_loss": 3.454866409301758, "perplexity": 31.654059711866473, "lr": 0.0026291804804649314, "grad_norm": 0.148563, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:12.787586+00:00", "epoch": 0, "step": 22687, "train_loss": 3.460071563720703, "perplexity": 31.819253537363743, "lr": 0.0026291804804649314, "grad_norm": 0.172588, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:13.092338+00:00", "epoch": 0, "step": 22688, "train_loss": 3.465949773788452, "perplexity": 32.006844603544195, "lr": 0.0026291804804649314, "grad_norm": 0.161245, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:13.396693+00:00", "epoch": 0, "step": 22689, "train_loss": 3.4196622371673584, "perplexity": 30.559091552380657, "lr": 0.0026291804804649314, "grad_norm": 0.17945, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:13.702278+00:00", "epoch": 0, "step": 22690, "train_loss": 3.5706889629364014, "perplexity": 35.54107119920271, "lr": 0.0026291804804649314, "grad_norm": 0.176112, "tokens_per_sec": 107292} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:14.006617+00:00", "epoch": 0, "step": 22691, "train_loss": 3.544130563735962, "perplexity": 34.60958143226129, "lr": 0.0026291804804649314, "grad_norm": 0.173001, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:14.311123+00:00", "epoch": 0, "step": 22692, "train_loss": 3.5810210704803467, "perplexity": 35.91018896762017, "lr": 0.0026291804804649314, "grad_norm": 0.178912, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:14.615553+00:00", "epoch": 0, "step": 22693, "train_loss": 3.4649336338043213, "perplexity": 31.974337687562272, "lr": 0.0026291804804649314, "grad_norm": 0.190477, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:14.920992+00:00", "epoch": 0, "step": 22694, "train_loss": 3.5977237224578857, "perplexity": 36.515021448399246, "lr": 0.0026291804804649314, "grad_norm": 0.202436, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:15.227005+00:00", "epoch": 0, "step": 22695, "train_loss": 3.563243865966797, "perplexity": 35.277447049112595, "lr": 0.0026291804804649314, "grad_norm": 0.18047, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:15.532395+00:00", "epoch": 0, "step": 22696, "train_loss": 3.465228319168091, "perplexity": 31.98376144534857, "lr": 0.0026291804804649314, "grad_norm": 0.158163, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:15.837672+00:00", "epoch": 0, "step": 22697, "train_loss": 3.5554635524749756, "perplexity": 35.00404241792159, "lr": 0.0026291804804649314, "grad_norm": 0.154281, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:16.142966+00:00", "epoch": 0, "step": 22698, "train_loss": 3.4926538467407227, "perplexity": 32.873072143010795, "lr": 0.0026291804804649314, "grad_norm": 0.174807, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:16.448692+00:00", "epoch": 0, "step": 22699, "train_loss": 3.4460830688476562, "perplexity": 31.37724877062565, "lr": 0.0026291804804649314, "grad_norm": 0.175848, "tokens_per_sec": 107177} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:16.754564+00:00", "epoch": 0, "step": 22700, "train_loss": 3.451510429382324, "perplexity": 31.54800737739276, "lr": 0.0026291804804649314, "grad_norm": 0.168855, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:17.058570+00:00", "epoch": 0, "step": 22701, "train_loss": 3.4042208194732666, "perplexity": 30.090840390269705, "lr": 0.0026291804804649314, "grad_norm": 0.170796, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:17.363361+00:00", "epoch": 0, "step": 22702, "train_loss": 3.4513821601867676, "perplexity": 31.543960999383422, "lr": 0.0026291804804649314, "grad_norm": 0.161232, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:17.668019+00:00", "epoch": 0, "step": 22703, "train_loss": 3.5501224994659424, "perplexity": 34.817582361618996, "lr": 0.0026291804804649314, "grad_norm": 0.178157, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:17.973524+00:00", "epoch": 0, "step": 22704, "train_loss": 3.510427474975586, "perplexity": 33.46256913791694, "lr": 0.0026291804804649314, "grad_norm": 0.16336, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:18.279116+00:00", "epoch": 0, "step": 22705, "train_loss": 3.4407150745391846, "perplexity": 31.209267143389667, "lr": 0.0026291804804649314, "grad_norm": 0.169029, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:18.583180+00:00", "epoch": 0, "step": 22706, "train_loss": 3.5345618724823, "perplexity": 34.27999241564411, "lr": 0.0026291804804649314, "grad_norm": 0.167183, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:18.887290+00:00", "epoch": 0, "step": 22707, "train_loss": 3.51163387298584, "perplexity": 33.502962675185344, "lr": 0.0026291804804649314, "grad_norm": 0.169346, "tokens_per_sec": 107750} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:19.192297+00:00", "epoch": 0, "step": 22708, "train_loss": 3.394105911254883, "perplexity": 29.788008442113533, "lr": 0.0026291804804649314, "grad_norm": 0.175533, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:19.497337+00:00", "epoch": 0, "step": 22709, "train_loss": 3.4453935623168945, "perplexity": 31.355621409641127, "lr": 0.0026291804804649314, "grad_norm": 0.186445, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:19.802135+00:00", "epoch": 0, "step": 22710, "train_loss": 3.4464762210845947, "perplexity": 31.389587231456606, "lr": 0.0026291804804649314, "grad_norm": 0.151217, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:20.107114+00:00", "epoch": 0, "step": 22711, "train_loss": 3.533379316329956, "perplexity": 34.239478359505235, "lr": 0.0026291804804649314, "grad_norm": 0.165137, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:20.411231+00:00", "epoch": 0, "step": 22712, "train_loss": 3.5705490112304688, "perplexity": 35.536097513703716, "lr": 0.0026291804804649314, "grad_norm": 0.164717, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:20.716806+00:00", "epoch": 0, "step": 22713, "train_loss": 3.4915614128112793, "perplexity": 32.83718009205336, "lr": 0.0026291804804649314, "grad_norm": 0.162623, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:21.022893+00:00", "epoch": 0, "step": 22714, "train_loss": 3.5748653411865234, "perplexity": 35.68981454398003, "lr": 0.0026291804804649314, "grad_norm": 0.188205, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:21.327462+00:00", "epoch": 0, "step": 22715, "train_loss": 3.498255968093872, "perplexity": 33.05774788738753, "lr": 0.0026291804804649314, "grad_norm": 0.201805, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:21.631832+00:00", "epoch": 0, "step": 22716, "train_loss": 3.589505910873413, "perplexity": 36.21617748562589, "lr": 0.0026291804804649314, "grad_norm": 0.188814, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:21.937008+00:00", "epoch": 0, "step": 22717, "train_loss": 3.557443618774414, "perplexity": 35.07342140748936, "lr": 0.0026291804804649314, "grad_norm": 0.179191, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:22.240678+00:00", "epoch": 0, "step": 22718, "train_loss": 3.4959330558776855, "perplexity": 32.9810467604621, "lr": 0.0026291804804649314, "grad_norm": 0.185563, "tokens_per_sec": 107905} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:22.546303+00:00", "epoch": 0, "step": 22719, "train_loss": 3.543269157409668, "perplexity": 34.579781356694646, "lr": 0.0026291804804649314, "grad_norm": 0.203729, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:22.850867+00:00", "epoch": 0, "step": 22720, "train_loss": 3.5198535919189453, "perplexity": 33.77948251258108, "lr": 0.0026291804804649314, "grad_norm": 0.17488, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:23.155599+00:00", "epoch": 0, "step": 22721, "train_loss": 3.4605205059051514, "perplexity": 31.833541749604944, "lr": 0.0026291804804649314, "grad_norm": 0.17908, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:23.460432+00:00", "epoch": 0, "step": 22722, "train_loss": 3.4628565311431885, "perplexity": 31.907992632261948, "lr": 0.0026291804804649314, "grad_norm": 0.177884, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:23.765891+00:00", "epoch": 0, "step": 22723, "train_loss": 3.562830686569214, "perplexity": 35.26287414561165, "lr": 0.0026291804804649314, "grad_norm": 0.157524, "tokens_per_sec": 107274} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:24.070797+00:00", "epoch": 0, "step": 22724, "train_loss": 3.607745409011841, "perplexity": 36.88280336726983, "lr": 0.0026291804804649314, "grad_norm": 0.172761, "tokens_per_sec": 107470} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:24.375769+00:00", "epoch": 0, "step": 22725, "train_loss": 3.5635945796966553, "perplexity": 35.28982150396585, "lr": 0.0026291804804649314, "grad_norm": 0.152468, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:24.681313+00:00", "epoch": 0, "step": 22726, "train_loss": 3.422767162322998, "perplexity": 30.6541227004301, "lr": 0.0026291804804649314, "grad_norm": 0.166357, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:24.985498+00:00", "epoch": 0, "step": 22727, "train_loss": 3.6239616870880127, "perplexity": 37.485780975610474, "lr": 0.0026291804804649314, "grad_norm": 0.163557, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:25.289070+00:00", "epoch": 0, "step": 22728, "train_loss": 3.397139072418213, "perplexity": 29.878497436930104, "lr": 0.0026291804804649314, "grad_norm": 0.172115, "tokens_per_sec": 107941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:25.594344+00:00", "epoch": 0, "step": 22729, "train_loss": 3.5557470321655273, "perplexity": 35.01396675964248, "lr": 0.0026291804804649314, "grad_norm": 0.161942, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:25.899917+00:00", "epoch": 0, "step": 22730, "train_loss": 3.5916800498962402, "perplexity": 36.2950021471918, "lr": 0.0026291804804649314, "grad_norm": 0.199612, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:26.205383+00:00", "epoch": 0, "step": 22731, "train_loss": 3.493217706680298, "perplexity": 32.891613178265274, "lr": 0.0026291804804649314, "grad_norm": 0.184804, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:26.509810+00:00", "epoch": 0, "step": 22732, "train_loss": 3.513645648956299, "perplexity": 33.57043097322981, "lr": 0.0026291804804649314, "grad_norm": 0.156414, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:26.813994+00:00", "epoch": 0, "step": 22733, "train_loss": 3.488101005554199, "perplexity": 32.72374645224777, "lr": 0.0026291804804649314, "grad_norm": 0.182164, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:27.117616+00:00", "epoch": 0, "step": 22734, "train_loss": 3.5430216789245605, "perplexity": 34.57122466363154, "lr": 0.0026291804804649314, "grad_norm": 0.199616, "tokens_per_sec": 107925} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:27.422129+00:00", "epoch": 0, "step": 22735, "train_loss": 3.3885931968688965, "perplexity": 29.624247457936235, "lr": 0.0026291804804649314, "grad_norm": 0.183009, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:27.727470+00:00", "epoch": 0, "step": 22736, "train_loss": 3.4376845359802246, "perplexity": 31.11482942676318, "lr": 0.0026291804804649314, "grad_norm": 0.159501, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:28.031937+00:00", "epoch": 0, "step": 22737, "train_loss": 3.515509843826294, "perplexity": 33.633071167056904, "lr": 0.0026291804804649314, "grad_norm": 0.154468, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:28.335630+00:00", "epoch": 0, "step": 22738, "train_loss": 3.3318793773651123, "perplexity": 27.99089776111587, "lr": 0.0026291804804649314, "grad_norm": 0.156638, "tokens_per_sec": 107901} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:28.639508+00:00", "epoch": 0, "step": 22739, "train_loss": 3.4886763095855713, "perplexity": 32.74257797190853, "lr": 0.0026291804804649314, "grad_norm": 0.176874, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:28.943680+00:00", "epoch": 0, "step": 22740, "train_loss": 3.447709321975708, "perplexity": 31.428317633782616, "lr": 0.0026291804804649314, "grad_norm": 0.179978, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:29.248533+00:00", "epoch": 0, "step": 22741, "train_loss": 3.4752910137176514, "perplexity": 32.307229015533, "lr": 0.0026291804804649314, "grad_norm": 0.195585, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:29.553270+00:00", "epoch": 0, "step": 22742, "train_loss": 3.4380688667297363, "perplexity": 31.126790110759796, "lr": 0.0026291804804649314, "grad_norm": 0.175934, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:29.857935+00:00", "epoch": 0, "step": 22743, "train_loss": 3.4867372512817383, "perplexity": 32.67914971961676, "lr": 0.0026291804804649314, "grad_norm": 0.178769, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:30.162157+00:00", "epoch": 0, "step": 22744, "train_loss": 3.4725797176361084, "perplexity": 32.21975319211345, "lr": 0.0026291804804649314, "grad_norm": 0.191527, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:30.466594+00:00", "epoch": 0, "step": 22745, "train_loss": 3.450500965118408, "perplexity": 31.516176859935452, "lr": 0.0026291804804649314, "grad_norm": 0.168123, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:30.772559+00:00", "epoch": 0, "step": 22746, "train_loss": 3.485440731048584, "perplexity": 32.63680799518546, "lr": 0.0026291804804649314, "grad_norm": 0.169126, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:31.077867+00:00", "epoch": 0, "step": 22747, "train_loss": 3.4460644721984863, "perplexity": 31.376665264363993, "lr": 0.0026291804804649314, "grad_norm": 0.186811, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:31.381894+00:00", "epoch": 0, "step": 22748, "train_loss": 3.4685122966766357, "perplexity": 32.088968052088795, "lr": 0.0026291804804649314, "grad_norm": 0.166087, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:31.686162+00:00", "epoch": 0, "step": 22749, "train_loss": 3.4923007488250732, "perplexity": 32.861466778791645, "lr": 0.0026291804804649314, "grad_norm": 0.167793, "tokens_per_sec": 107698} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:31.990480+00:00", "epoch": 0, "step": 22750, "train_loss": 3.4961466789245605, "perplexity": 32.98809302475562, "lr": 0.0026291804804649314, "grad_norm": 0.164446, "tokens_per_sec": 107733} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:32.294689+00:00", "epoch": 0, "step": 22751, "train_loss": 3.5252256393432617, "perplexity": 33.961435787838745, "lr": 0.0026291804804649314, "grad_norm": 0.161628, "tokens_per_sec": 107656} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:32.600064+00:00", "epoch": 0, "step": 22752, "train_loss": 3.4019618034362793, "perplexity": 30.02294142056729, "lr": 0.0026291804804649314, "grad_norm": 0.179233, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:32.904811+00:00", "epoch": 0, "step": 22753, "train_loss": 3.428612470626831, "perplexity": 30.833830209614153, "lr": 0.0026291804804649314, "grad_norm": 0.175744, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:33.209362+00:00", "epoch": 0, "step": 22754, "train_loss": 3.527691125869751, "perplexity": 34.04527055447102, "lr": 0.0026291804804649314, "grad_norm": 0.179131, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:33.513550+00:00", "epoch": 0, "step": 22755, "train_loss": 3.358266592025757, "perplexity": 28.739330692707647, "lr": 0.0026291804804649314, "grad_norm": 0.173943, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:33.817513+00:00", "epoch": 0, "step": 22756, "train_loss": 3.4382452964782715, "perplexity": 31.13228228698913, "lr": 0.0026291804804649314, "grad_norm": 0.204632, "tokens_per_sec": 107796} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:34.122063+00:00", "epoch": 0, "step": 22757, "train_loss": 3.5668485164642334, "perplexity": 35.4048393806427, "lr": 0.0026291804804649314, "grad_norm": 0.188425, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:34.426652+00:00", "epoch": 0, "step": 22758, "train_loss": 3.5620367527008057, "perplexity": 35.234888866230584, "lr": 0.0026291804804649314, "grad_norm": 0.175227, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:34.730566+00:00", "epoch": 0, "step": 22759, "train_loss": 3.470576524734497, "perplexity": 32.15527541350283, "lr": 0.0026291804804649314, "grad_norm": 0.221871, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:35.034523+00:00", "epoch": 0, "step": 22760, "train_loss": 3.4150660037994385, "perplexity": 30.41895712805194, "lr": 0.0026291804804649314, "grad_norm": 0.219278, "tokens_per_sec": 107875} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:35.339445+00:00", "epoch": 0, "step": 22761, "train_loss": 3.499023675918579, "perplexity": 33.08313632331138, "lr": 0.0026291804804649314, "grad_norm": 0.172077, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:35.645514+00:00", "epoch": 0, "step": 22762, "train_loss": 3.5115067958831787, "perplexity": 33.49870548625937, "lr": 0.0026291804804649314, "grad_norm": 0.180901, "tokens_per_sec": 107061} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:35.949869+00:00", "epoch": 0, "step": 22763, "train_loss": 3.4606997966766357, "perplexity": 31.839249721542366, "lr": 0.0026291804804649314, "grad_norm": 0.197069, "tokens_per_sec": 107664} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:36.254456+00:00", "epoch": 0, "step": 22764, "train_loss": 3.394899368286133, "perplexity": 29.811653326218117, "lr": 0.0026291804804649314, "grad_norm": 0.138927, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:36.558476+00:00", "epoch": 0, "step": 22765, "train_loss": 3.5008468627929688, "perplexity": 33.143508080994756, "lr": 0.0026291804804649314, "grad_norm": 0.188136, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:36.863117+00:00", "epoch": 0, "step": 22766, "train_loss": 3.4436464309692383, "perplexity": 31.300886848707663, "lr": 0.0026291804804649314, "grad_norm": 0.167561, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:37.169026+00:00", "epoch": 0, "step": 22767, "train_loss": 3.4574050903320312, "perplexity": 31.73452136280764, "lr": 0.0026291804804649314, "grad_norm": 0.186636, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:37.474671+00:00", "epoch": 0, "step": 22768, "train_loss": 3.5168299674987793, "perplexity": 33.67750030000568, "lr": 0.0026291804804649314, "grad_norm": 0.172997, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:37.780014+00:00", "epoch": 0, "step": 22769, "train_loss": 3.5264909267425537, "perplexity": 34.00443396139005, "lr": 0.0026291804804649314, "grad_norm": 0.17597, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:38.084993+00:00", "epoch": 0, "step": 22770, "train_loss": 3.5053231716156006, "perplexity": 33.292201208525505, "lr": 0.0026291804804649314, "grad_norm": 0.147452, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:38.389325+00:00", "epoch": 0, "step": 22771, "train_loss": 3.5853729248046875, "perplexity": 36.06680541757923, "lr": 0.0026291804804649314, "grad_norm": 0.18907, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:38.693891+00:00", "epoch": 0, "step": 22772, "train_loss": 3.3977580070495605, "perplexity": 29.896995997835905, "lr": 0.0026291804804649314, "grad_norm": 0.194986, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:38.998653+00:00", "epoch": 0, "step": 22773, "train_loss": 3.5125482082366943, "perplexity": 33.53360962362488, "lr": 0.0026291804804649314, "grad_norm": 0.160153, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:39.303210+00:00", "epoch": 0, "step": 22774, "train_loss": 3.3911595344543457, "perplexity": 29.700370915093153, "lr": 0.0026291804804649314, "grad_norm": 0.175864, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:39.608024+00:00", "epoch": 0, "step": 22775, "train_loss": 3.4865329265594482, "perplexity": 32.67247324353341, "lr": 0.0026291804804649314, "grad_norm": 0.185373, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:39.912007+00:00", "epoch": 0, "step": 22776, "train_loss": 3.4447405338287354, "perplexity": 31.335151979884508, "lr": 0.0026291804804649314, "grad_norm": 0.181511, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:40.217669+00:00", "epoch": 0, "step": 22777, "train_loss": 3.4515504837036133, "perplexity": 31.549271036723624, "lr": 0.0026291804804649314, "grad_norm": 0.166279, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:40.523419+00:00", "epoch": 0, "step": 22778, "train_loss": 3.472034454345703, "perplexity": 32.202189732261985, "lr": 0.0026291804804649314, "grad_norm": 0.172575, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:40.829368+00:00", "epoch": 0, "step": 22779, "train_loss": 3.5878446102142334, "perplexity": 36.156061475306124, "lr": 0.0026291804804649314, "grad_norm": 0.167048, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:41.133639+00:00", "epoch": 0, "step": 22780, "train_loss": 3.579357862472534, "perplexity": 35.85051249471769, "lr": 0.0026291804804649314, "grad_norm": 0.16613, "tokens_per_sec": 107762} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:41.437689+00:00", "epoch": 0, "step": 22781, "train_loss": 3.5183942317962646, "perplexity": 33.730222035967664, "lr": 0.0026291804804649314, "grad_norm": 0.163012, "tokens_per_sec": 107703} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:41.741587+00:00", "epoch": 0, "step": 22782, "train_loss": 3.5420687198638916, "perplexity": 34.53829539446462, "lr": 0.0026291804804649314, "grad_norm": 0.164332, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:42.045952+00:00", "epoch": 0, "step": 22783, "train_loss": 3.509351968765259, "perplexity": 33.42659928336382, "lr": 0.0026291804804649314, "grad_norm": 0.165, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:42.350718+00:00", "epoch": 0, "step": 22784, "train_loss": 3.5079245567321777, "perplexity": 33.37891979055825, "lr": 0.0026291804804649314, "grad_norm": 0.165161, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:42.655532+00:00", "epoch": 0, "step": 22785, "train_loss": 3.6108131408691406, "perplexity": 36.99612364739161, "lr": 0.0026291804804649314, "grad_norm": 0.174115, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:42.959393+00:00", "epoch": 0, "step": 22786, "train_loss": 3.5181853771209717, "perplexity": 33.72317805700636, "lr": 0.0026291804804649314, "grad_norm": 0.171078, "tokens_per_sec": 107838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:43.263541+00:00", "epoch": 0, "step": 22787, "train_loss": 3.390979766845703, "perplexity": 29.695032230313625, "lr": 0.0026291804804649314, "grad_norm": 0.159579, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:43.568110+00:00", "epoch": 0, "step": 22788, "train_loss": 3.460026264190674, "perplexity": 31.81781217277943, "lr": 0.0026291804804649314, "grad_norm": 0.170718, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:43.874905+00:00", "epoch": 0, "step": 22789, "train_loss": 3.46838641166687, "perplexity": 32.08492878627903, "lr": 0.0026291804804649314, "grad_norm": 0.187745, "tokens_per_sec": 106808} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:44.179435+00:00", "epoch": 0, "step": 22790, "train_loss": 3.5617549419403076, "perplexity": 35.22496069440166, "lr": 0.0026291804804649314, "grad_norm": 0.192127, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:44.483415+00:00", "epoch": 0, "step": 22791, "train_loss": 3.4680376052856445, "perplexity": 32.073739309971025, "lr": 0.0026291804804649314, "grad_norm": 0.186666, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:44.787997+00:00", "epoch": 0, "step": 22792, "train_loss": 3.5287258625030518, "perplexity": 34.080516675185734, "lr": 0.0026291804804649314, "grad_norm": 0.173543, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:45.092744+00:00", "epoch": 0, "step": 22793, "train_loss": 3.5640769004821777, "perplexity": 35.30684662384913, "lr": 0.0026291804804649314, "grad_norm": 0.199044, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:45.399287+00:00", "epoch": 0, "step": 22794, "train_loss": 3.55485463142395, "perplexity": 34.98273420778958, "lr": 0.0026291804804649314, "grad_norm": 0.201654, "tokens_per_sec": 106896} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:45.704206+00:00", "epoch": 0, "step": 22795, "train_loss": 3.430820941925049, "perplexity": 30.902001087652224, "lr": 0.0026291804804649314, "grad_norm": 0.188192, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:46.008372+00:00", "epoch": 0, "step": 22796, "train_loss": 3.3941526412963867, "perplexity": 29.78940046950885, "lr": 0.0026291804804649314, "grad_norm": 0.232469, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:46.313252+00:00", "epoch": 0, "step": 22797, "train_loss": 3.550093173980713, "perplexity": 34.81656133409286, "lr": 0.0026291804804649314, "grad_norm": 0.212292, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:46.617023+00:00", "epoch": 0, "step": 22798, "train_loss": 3.489687204360962, "perplexity": 32.775694008507216, "lr": 0.0026291804804649314, "grad_norm": 0.174839, "tokens_per_sec": 107872} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:46.921237+00:00", "epoch": 0, "step": 22799, "train_loss": 3.432504177093506, "perplexity": 30.95406022425193, "lr": 0.0026291804804649314, "grad_norm": 0.185994, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:47.226792+00:00", "epoch": 0, "step": 22800, "train_loss": 3.5052638053894043, "perplexity": 33.29022483484349, "lr": 0.0026291804804649314, "grad_norm": 0.18844, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:47.531197+00:00", "epoch": 0, "step": 22801, "train_loss": 3.493802547454834, "perplexity": 32.91085516099177, "lr": 0.0026291804804649314, "grad_norm": 0.180267, "tokens_per_sec": 107584} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:47.835202+00:00", "epoch": 0, "step": 22802, "train_loss": 3.3803391456604004, "perplexity": 29.38073377215589, "lr": 0.0026291804804649314, "grad_norm": 0.198328, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:48.139295+00:00", "epoch": 0, "step": 22803, "train_loss": 3.534635543823242, "perplexity": 34.282517961681904, "lr": 0.0026291804804649314, "grad_norm": 0.215404, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:48.443961+00:00", "epoch": 0, "step": 22804, "train_loss": 3.521989345550537, "perplexity": 33.85170426150895, "lr": 0.0026291804804649314, "grad_norm": 0.200959, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:48.749635+00:00", "epoch": 0, "step": 22805, "train_loss": 3.4591429233551025, "perplexity": 31.78971860991109, "lr": 0.0026291804804649314, "grad_norm": 0.203812, "tokens_per_sec": 107198} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:49.054669+00:00", "epoch": 0, "step": 22806, "train_loss": 3.433044910430908, "perplexity": 30.970802642727104, "lr": 0.0026291804804649314, "grad_norm": 0.177003, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:49.359512+00:00", "epoch": 0, "step": 22807, "train_loss": 3.3999485969543457, "perplexity": 29.962559840980276, "lr": 0.0026291804804649314, "grad_norm": 0.167902, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:49.664150+00:00", "epoch": 0, "step": 22808, "train_loss": 3.615304946899414, "perplexity": 37.16267684097155, "lr": 0.0026291804804649314, "grad_norm": 0.179629, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:49.968178+00:00", "epoch": 0, "step": 22809, "train_loss": 3.549102544784546, "perplexity": 34.78208810983557, "lr": 0.0026291804804649314, "grad_norm": 0.175481, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:50.273877+00:00", "epoch": 0, "step": 22810, "train_loss": 3.554847240447998, "perplexity": 34.9824756521978, "lr": 0.0026291804804649314, "grad_norm": 0.162286, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:50.578864+00:00", "epoch": 0, "step": 22811, "train_loss": 3.5224051475524902, "perplexity": 33.8657827946487, "lr": 0.0026291804804649314, "grad_norm": 0.174964, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:50.883233+00:00", "epoch": 0, "step": 22812, "train_loss": 3.4945566654205322, "perplexity": 32.93568318858937, "lr": 0.0026291804804649314, "grad_norm": 0.185005, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:51.187877+00:00", "epoch": 0, "step": 22813, "train_loss": 3.504899740219116, "perplexity": 33.278107229402956, "lr": 0.0026291804804649314, "grad_norm": 0.18031, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:51.492752+00:00", "epoch": 0, "step": 22814, "train_loss": 3.4621899127960205, "perplexity": 31.886729267014097, "lr": 0.0026291804804649314, "grad_norm": 0.158075, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:51.798764+00:00", "epoch": 0, "step": 22815, "train_loss": 3.42681884765625, "perplexity": 30.778575511340467, "lr": 0.0026291804804649314, "grad_norm": 0.174749, "tokens_per_sec": 107081} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:52.103836+00:00", "epoch": 0, "step": 22816, "train_loss": 3.4408986568450928, "perplexity": 31.214997138564325, "lr": 0.0026291804804649314, "grad_norm": 0.215353, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:52.410400+00:00", "epoch": 0, "step": 22817, "train_loss": 3.5286214351654053, "perplexity": 34.07695792338245, "lr": 0.0026291804804649314, "grad_norm": 0.172239, "tokens_per_sec": 106894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:52.715081+00:00", "epoch": 0, "step": 22818, "train_loss": 3.527068853378296, "perplexity": 34.024091709330605, "lr": 0.0026291804804649314, "grad_norm": 0.181523, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:53.019187+00:00", "epoch": 0, "step": 22819, "train_loss": 3.481762647628784, "perplexity": 32.516987582640034, "lr": 0.0026291804804649314, "grad_norm": 0.180439, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:53.324217+00:00", "epoch": 0, "step": 22820, "train_loss": 3.501131296157837, "perplexity": 33.152936541342555, "lr": 0.0026291804804649314, "grad_norm": 0.168386, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:53.629302+00:00", "epoch": 0, "step": 22821, "train_loss": 3.5365803241729736, "perplexity": 34.34925480216691, "lr": 0.0026291804804649314, "grad_norm": 0.195736, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:53.933784+00:00", "epoch": 0, "step": 22822, "train_loss": 3.509460687637329, "perplexity": 33.43023358308996, "lr": 0.0026291804804649314, "grad_norm": 0.150347, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:54.240685+00:00", "epoch": 0, "step": 22823, "train_loss": 3.428395986557007, "perplexity": 30.827155899029133, "lr": 0.0026291804804649314, "grad_norm": 0.166606, "tokens_per_sec": 106775} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:54.546556+00:00", "epoch": 0, "step": 22824, "train_loss": 3.3442673683166504, "perplexity": 28.33980541976164, "lr": 0.0026291804804649314, "grad_norm": 0.171145, "tokens_per_sec": 107126} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:54.850612+00:00", "epoch": 0, "step": 22825, "train_loss": 3.6081061363220215, "perplexity": 36.8961104016815, "lr": 0.0026291804804649314, "grad_norm": 0.159345, "tokens_per_sec": 107769} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:55.156471+00:00", "epoch": 0, "step": 22826, "train_loss": 3.5770533084869385, "perplexity": 35.767988180654456, "lr": 0.0026291804804649314, "grad_norm": 0.165034, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:55.461158+00:00", "epoch": 0, "step": 22827, "train_loss": 3.4385743141174316, "perplexity": 31.142527042269045, "lr": 0.0026291804804649314, "grad_norm": 0.159183, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:55.766210+00:00", "epoch": 0, "step": 22828, "train_loss": 3.403679847717285, "perplexity": 30.074566497759808, "lr": 0.0026291804804649314, "grad_norm": 0.15714, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:56.070541+00:00", "epoch": 0, "step": 22829, "train_loss": 3.481645107269287, "perplexity": 32.51316574884407, "lr": 0.0026291804804649314, "grad_norm": 0.164152, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:56.374501+00:00", "epoch": 0, "step": 22830, "train_loss": 3.4773709774017334, "perplexity": 32.374496811653714, "lr": 0.0026291804804649314, "grad_norm": 0.196165, "tokens_per_sec": 107894} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:56.679428+00:00", "epoch": 0, "step": 22831, "train_loss": 3.6091198921203613, "perplexity": 36.933533013027066, "lr": 0.0026291804804649314, "grad_norm": 0.171847, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:56.985410+00:00", "epoch": 0, "step": 22832, "train_loss": 3.4843943119049072, "perplexity": 32.60267407681831, "lr": 0.0026291804804649314, "grad_norm": 0.159982, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:57.290463+00:00", "epoch": 0, "step": 22833, "train_loss": 3.55375599861145, "perplexity": 34.94432213236476, "lr": 0.0026291804804649314, "grad_norm": 0.154084, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:57.595219+00:00", "epoch": 0, "step": 22834, "train_loss": 3.4070076942443848, "perplexity": 30.174816755650724, "lr": 0.0026291804804649314, "grad_norm": 0.156216, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:57.899503+00:00", "epoch": 0, "step": 22835, "train_loss": 3.438086986541748, "perplexity": 31.127354127455053, "lr": 0.0026291804804649314, "grad_norm": 0.150935, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:58.203409+00:00", "epoch": 0, "step": 22836, "train_loss": 3.5487194061279297, "perplexity": 34.76876429991916, "lr": 0.0026291804804649314, "grad_norm": 0.156591, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:58.507830+00:00", "epoch": 0, "step": 22837, "train_loss": 3.561403274536133, "perplexity": 35.21257540179159, "lr": 0.0026291804804649314, "grad_norm": 0.167458, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:58.812249+00:00", "epoch": 0, "step": 22838, "train_loss": 3.5476627349853516, "perplexity": 34.73204455378502, "lr": 0.0026291804804649314, "grad_norm": 0.162806, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:59.116718+00:00", "epoch": 0, "step": 22839, "train_loss": 3.4949848651885986, "perplexity": 32.949789260384634, "lr": 0.0026291804804649314, "grad_norm": 0.19806, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:59.421504+00:00", "epoch": 0, "step": 22840, "train_loss": 3.513976812362671, "perplexity": 33.58155011252869, "lr": 0.0026291804804649314, "grad_norm": 0.159738, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:57:59.725040+00:00", "epoch": 0, "step": 22841, "train_loss": 3.5108859539031982, "perplexity": 33.47791453823298, "lr": 0.0026291804804649314, "grad_norm": 0.199353, "tokens_per_sec": 107885} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:00.029967+00:00", "epoch": 0, "step": 22842, "train_loss": 3.4597294330596924, "perplexity": 31.808369057181505, "lr": 0.0026291804804649314, "grad_norm": 0.198947, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:00.335834+00:00", "epoch": 0, "step": 22843, "train_loss": 3.504706859588623, "perplexity": 33.2716891460808, "lr": 0.0026291804804649314, "grad_norm": 0.148359, "tokens_per_sec": 107131} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:00.640525+00:00", "epoch": 0, "step": 22844, "train_loss": 3.451857566833496, "perplexity": 31.558960773320276, "lr": 0.0026291804804649314, "grad_norm": 0.181517, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:00.945297+00:00", "epoch": 0, "step": 22845, "train_loss": 3.3893449306488037, "perplexity": 29.64652537793809, "lr": 0.0026291804804649314, "grad_norm": 0.182473, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:01.249971+00:00", "epoch": 0, "step": 22846, "train_loss": 3.6073620319366455, "perplexity": 36.86866605612354, "lr": 0.0026291804804649314, "grad_norm": 0.190383, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:01.554089+00:00", "epoch": 0, "step": 22847, "train_loss": 3.4734396934509277, "perplexity": 32.24747331822875, "lr": 0.0026291804804649314, "grad_norm": 0.168584, "tokens_per_sec": 107749} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:01.858920+00:00", "epoch": 0, "step": 22848, "train_loss": 3.5318522453308105, "perplexity": 34.18723214706742, "lr": 0.0026291804804649314, "grad_norm": 0.172139, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:02.163717+00:00", "epoch": 0, "step": 22849, "train_loss": 3.4854164123535156, "perplexity": 32.63601432025443, "lr": 0.0026291804804649314, "grad_norm": 0.181688, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:02.467764+00:00", "epoch": 0, "step": 22850, "train_loss": 3.4794139862060547, "perplexity": 32.440705803422894, "lr": 0.0026291804804649314, "grad_norm": 0.14602, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:02.772784+00:00", "epoch": 0, "step": 22851, "train_loss": 3.5042386054992676, "perplexity": 33.25611318861588, "lr": 0.0026291804804649314, "grad_norm": 0.169901, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:03.076896+00:00", "epoch": 0, "step": 22852, "train_loss": 3.484368324279785, "perplexity": 32.601826821755544, "lr": 0.0026291804804649314, "grad_norm": 0.167031, "tokens_per_sec": 107751} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:03.381660+00:00", "epoch": 0, "step": 22853, "train_loss": 3.4958009719848633, "perplexity": 32.97669078310018, "lr": 0.0026291804804649314, "grad_norm": 0.1987, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:03.686628+00:00", "epoch": 0, "step": 22854, "train_loss": 3.5958733558654785, "perplexity": 36.44751774514903, "lr": 0.0026291804804649314, "grad_norm": 0.181893, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:03.991371+00:00", "epoch": 0, "step": 22855, "train_loss": 3.535881996154785, "perplexity": 34.325276128598915, "lr": 0.0026291804804649314, "grad_norm": 0.16025, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:04.295185+00:00", "epoch": 0, "step": 22856, "train_loss": 3.5367789268493652, "perplexity": 34.35607733356675, "lr": 0.0026291804804649314, "grad_norm": 0.218001, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:04.599536+00:00", "epoch": 0, "step": 22857, "train_loss": 3.402095079421997, "perplexity": 30.026943024332184, "lr": 0.0026291804804649314, "grad_norm": 0.18367, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:04.904577+00:00", "epoch": 0, "step": 22858, "train_loss": 3.5162854194641113, "perplexity": 33.659166275736276, "lr": 0.0026291804804649314, "grad_norm": 0.196495, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:05.209761+00:00", "epoch": 0, "step": 22859, "train_loss": 3.430593729019165, "perplexity": 30.894980551795854, "lr": 0.0026291804804649314, "grad_norm": 0.169979, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:05.514193+00:00", "epoch": 0, "step": 22860, "train_loss": 3.3576087951660156, "perplexity": 28.720432267570715, "lr": 0.0026291804804649314, "grad_norm": 0.174495, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:05.818130+00:00", "epoch": 0, "step": 22861, "train_loss": 3.509831666946411, "perplexity": 33.44263780876033, "lr": 0.0026291804804649314, "grad_norm": 0.173464, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:06.122480+00:00", "epoch": 0, "step": 22862, "train_loss": 3.492678165435791, "perplexity": 32.87387158294883, "lr": 0.0026291804804649314, "grad_norm": 0.192912, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:06.427146+00:00", "epoch": 0, "step": 22863, "train_loss": 3.4415812492370605, "perplexity": 31.236311531820153, "lr": 0.0026291804804649314, "grad_norm": 0.180823, "tokens_per_sec": 107554} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:06.732836+00:00", "epoch": 0, "step": 22864, "train_loss": 3.5549957752227783, "perplexity": 34.987672152260984, "lr": 0.0026291804804649314, "grad_norm": 0.196095, "tokens_per_sec": 107194} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:07.037598+00:00", "epoch": 0, "step": 22865, "train_loss": 3.5624501705169678, "perplexity": 35.24945860852638, "lr": 0.0026291804804649314, "grad_norm": 0.193635, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:07.341543+00:00", "epoch": 0, "step": 22866, "train_loss": 3.524836540222168, "perplexity": 33.94822399353808, "lr": 0.0026291804804649314, "grad_norm": 0.162143, "tokens_per_sec": 107810} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:07.645938+00:00", "epoch": 0, "step": 22867, "train_loss": 3.403836965560913, "perplexity": 30.07929212002605, "lr": 0.0026291804804649314, "grad_norm": 0.166837, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:07.950208+00:00", "epoch": 0, "step": 22868, "train_loss": 3.4694244861602783, "perplexity": 32.11825262579632, "lr": 0.0026291804804649314, "grad_norm": 0.188008, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:08.255277+00:00", "epoch": 0, "step": 22869, "train_loss": 3.493321418762207, "perplexity": 32.89502461284577, "lr": 0.0026291804804649314, "grad_norm": 0.157572, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:08.560034+00:00", "epoch": 0, "step": 22870, "train_loss": 3.4631025791168213, "perplexity": 31.915844495119668, "lr": 0.0026291804804649314, "grad_norm": 0.170585, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:08.865291+00:00", "epoch": 0, "step": 22871, "train_loss": 3.522798776626587, "perplexity": 33.87911597536535, "lr": 0.0026291804804649314, "grad_norm": 0.211225, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:09.169637+00:00", "epoch": 0, "step": 22872, "train_loss": 3.5807735919952393, "perplexity": 35.90130306803438, "lr": 0.0026291804804649314, "grad_norm": 0.199796, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:09.474565+00:00", "epoch": 0, "step": 22873, "train_loss": 3.4184696674346924, "perplexity": 30.52266942701637, "lr": 0.0026291804804649314, "grad_norm": 0.197776, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:09.779734+00:00", "epoch": 0, "step": 22874, "train_loss": 3.5386621952056885, "perplexity": 34.42084001046534, "lr": 0.0026291804804649314, "grad_norm": 0.180241, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:10.084460+00:00", "epoch": 0, "step": 22875, "train_loss": 3.563007116317749, "perplexity": 35.269096114483844, "lr": 0.0026291804804649314, "grad_norm": 0.157505, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:10.389987+00:00", "epoch": 0, "step": 22876, "train_loss": 3.501115322113037, "perplexity": 33.152406959078796, "lr": 0.0026291804804649314, "grad_norm": 0.195734, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:10.694658+00:00", "epoch": 0, "step": 22877, "train_loss": 3.5675203800201416, "perplexity": 35.42863459459877, "lr": 0.0026291804804649314, "grad_norm": 0.172084, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:10.999401+00:00", "epoch": 0, "step": 22878, "train_loss": 3.3787758350372314, "perplexity": 29.334838442608618, "lr": 0.0026291804804649314, "grad_norm": 0.172165, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:11.304035+00:00", "epoch": 0, "step": 22879, "train_loss": 3.4928717613220215, "perplexity": 32.88023644533747, "lr": 0.0026291804804649314, "grad_norm": 0.183724, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:11.609528+00:00", "epoch": 0, "step": 22880, "train_loss": 3.4575583934783936, "perplexity": 31.739386737709975, "lr": 0.0026291804804649314, "grad_norm": 0.170786, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:11.913909+00:00", "epoch": 0, "step": 22881, "train_loss": 3.45849871635437, "perplexity": 31.769246045623905, "lr": 0.0026291804804649314, "grad_norm": 0.170812, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:12.218451+00:00", "epoch": 0, "step": 22882, "train_loss": 3.5714824199676514, "perplexity": 35.569282702871746, "lr": 0.0026291804804649314, "grad_norm": 0.177123, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:12.523733+00:00", "epoch": 0, "step": 22883, "train_loss": 3.5071051120758057, "perplexity": 33.351578816837744, "lr": 0.0026291804804649314, "grad_norm": 0.208659, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:12.828222+00:00", "epoch": 0, "step": 22884, "train_loss": 3.5255558490753174, "perplexity": 33.97265203620574, "lr": 0.0026291804804649314, "grad_norm": 0.157423, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:13.133635+00:00", "epoch": 0, "step": 22885, "train_loss": 3.58516263961792, "perplexity": 36.05922190004453, "lr": 0.0026291804804649314, "grad_norm": 0.200777, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:13.439614+00:00", "epoch": 0, "step": 22886, "train_loss": 3.472780704498291, "perplexity": 32.22622959002148, "lr": 0.0026291804804649314, "grad_norm": 0.205337, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:13.744131+00:00", "epoch": 0, "step": 22887, "train_loss": 3.528752088546753, "perplexity": 34.0814104840259, "lr": 0.0026291804804649314, "grad_norm": 0.15861, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:14.049019+00:00", "epoch": 0, "step": 22888, "train_loss": 3.4315414428710938, "perplexity": 30.9242740315456, "lr": 0.0026291804804649314, "grad_norm": 0.194895, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:14.353590+00:00", "epoch": 0, "step": 22889, "train_loss": 3.503833293914795, "perplexity": 33.24263683193223, "lr": 0.0026291804804649314, "grad_norm": 0.188157, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:14.659432+00:00", "epoch": 0, "step": 22890, "train_loss": 3.4272866249084473, "perplexity": 30.79297639676042, "lr": 0.0026291804804649314, "grad_norm": 0.175121, "tokens_per_sec": 107230} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:14.965943+00:00", "epoch": 0, "step": 22891, "train_loss": 3.5547218322753906, "perplexity": 34.97808883892984, "lr": 0.0026291804804649314, "grad_norm": 0.177965, "tokens_per_sec": 106814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:15.271135+00:00", "epoch": 0, "step": 22892, "train_loss": 3.5120794773101807, "perplexity": 33.51789506695373, "lr": 0.0026291804804649314, "grad_norm": 0.177228, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:15.575931+00:00", "epoch": 0, "step": 22893, "train_loss": 3.436718225479126, "perplexity": 31.08477736249929, "lr": 0.0026291804804649314, "grad_norm": 0.191649, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:15.880919+00:00", "epoch": 0, "step": 22894, "train_loss": 3.4485809803009033, "perplexity": 31.455724331397782, "lr": 0.0026291804804649314, "grad_norm": 0.188137, "tokens_per_sec": 107441} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:16.185717+00:00", "epoch": 0, "step": 22895, "train_loss": 3.392800807952881, "perplexity": 29.749157371780324, "lr": 0.0026291804804649314, "grad_norm": 0.169359, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:16.492057+00:00", "epoch": 0, "step": 22896, "train_loss": 3.5263452529907227, "perplexity": 33.99948076869985, "lr": 0.0026291804804649314, "grad_norm": 0.203937, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:16.797743+00:00", "epoch": 0, "step": 22897, "train_loss": 3.4980311393737793, "perplexity": 33.050316391679, "lr": 0.0026291804804649314, "grad_norm": 0.160005, "tokens_per_sec": 107195} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:17.103091+00:00", "epoch": 0, "step": 22898, "train_loss": 3.5054280757904053, "perplexity": 33.295693882615566, "lr": 0.0026291804804649314, "grad_norm": 0.191883, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:17.406957+00:00", "epoch": 0, "step": 22899, "train_loss": 3.4479634761810303, "perplexity": 31.43630628800748, "lr": 0.0026291804804649314, "grad_norm": 0.174012, "tokens_per_sec": 107837} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:17.712063+00:00", "epoch": 0, "step": 22900, "train_loss": 3.4113874435424805, "perplexity": 30.307264720877566, "lr": 0.0026291804804649314, "grad_norm": 0.196649, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:18.018309+00:00", "epoch": 0, "step": 22901, "train_loss": 3.458324432373047, "perplexity": 31.76370965740534, "lr": 0.0026291804804649314, "grad_norm": 0.170367, "tokens_per_sec": 106941} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:18.322606+00:00", "epoch": 0, "step": 22902, "train_loss": 3.5178585052490234, "perplexity": 33.71215670005074, "lr": 0.0026291804804649314, "grad_norm": 0.177701, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:18.627748+00:00", "epoch": 0, "step": 22903, "train_loss": 3.413116216659546, "perplexity": 30.359704420489926, "lr": 0.0026291804804649314, "grad_norm": 0.190175, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:18.932706+00:00", "epoch": 0, "step": 22904, "train_loss": 3.4489636421203613, "perplexity": 31.46776353942852, "lr": 0.0026291804804649314, "grad_norm": 0.177127, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:19.237642+00:00", "epoch": 0, "step": 22905, "train_loss": 3.3715004920959473, "perplexity": 29.122191908998605, "lr": 0.0026291804804649314, "grad_norm": 0.178675, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:19.543984+00:00", "epoch": 0, "step": 22906, "train_loss": 3.4449105262756348, "perplexity": 31.3404791718219, "lr": 0.0026291804804649314, "grad_norm": 0.183996, "tokens_per_sec": 106965} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:19.848077+00:00", "epoch": 0, "step": 22907, "train_loss": 3.380296230316162, "perplexity": 29.37947291490734, "lr": 0.0026291804804649314, "grad_norm": 0.222997, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:20.152694+00:00", "epoch": 0, "step": 22908, "train_loss": 3.598973512649536, "perplexity": 36.56068609372235, "lr": 0.0026291804804649314, "grad_norm": 0.202757, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:20.458838+00:00", "epoch": 0, "step": 22909, "train_loss": 3.50117826461792, "perplexity": 33.154493720287995, "lr": 0.0026291804804649314, "grad_norm": 0.164728, "tokens_per_sec": 107035} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:20.765715+00:00", "epoch": 0, "step": 22910, "train_loss": 3.4230620861053467, "perplexity": 30.66316466352093, "lr": 0.0026291804804649314, "grad_norm": 0.171552, "tokens_per_sec": 106838} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:21.071247+00:00", "epoch": 0, "step": 22911, "train_loss": 3.5107288360595703, "perplexity": 33.47265497368811, "lr": 0.0026291804804649314, "grad_norm": 0.198921, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:21.375353+00:00", "epoch": 0, "step": 22912, "train_loss": 3.5773396492004395, "perplexity": 35.7782314783772, "lr": 0.0026291804804649314, "grad_norm": 0.177613, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:21.680512+00:00", "epoch": 0, "step": 22913, "train_loss": 3.3060710430145264, "perplexity": 27.277741583359553, "lr": 0.0026291804804649314, "grad_norm": 0.178267, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:21.986330+00:00", "epoch": 0, "step": 22914, "train_loss": 3.416349411010742, "perplexity": 30.458022099761518, "lr": 0.0026291804804649314, "grad_norm": 0.170845, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:22.293386+00:00", "epoch": 0, "step": 22915, "train_loss": 3.5164434909820557, "perplexity": 33.664487251778525, "lr": 0.0026291804804649314, "grad_norm": 0.188456, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:22.598492+00:00", "epoch": 0, "step": 22916, "train_loss": 3.4787862300872803, "perplexity": 32.420347342600174, "lr": 0.0026291804804649314, "grad_norm": 0.167956, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:22.903999+00:00", "epoch": 0, "step": 22917, "train_loss": 3.515413761138916, "perplexity": 33.62983976643772, "lr": 0.0026291804804649314, "grad_norm": 0.185701, "tokens_per_sec": 107258} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:23.209117+00:00", "epoch": 0, "step": 22918, "train_loss": 3.5820040702819824, "perplexity": 35.94550603174169, "lr": 0.0026291804804649314, "grad_norm": 0.155373, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:23.514656+00:00", "epoch": 0, "step": 22919, "train_loss": 3.518666982650757, "perplexity": 33.73942323761095, "lr": 0.0026291804804649314, "grad_norm": 0.168941, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:23.820983+00:00", "epoch": 0, "step": 22920, "train_loss": 3.436560869216919, "perplexity": 31.079886362946784, "lr": 0.0026291804804649314, "grad_norm": 0.173229, "tokens_per_sec": 107026} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:24.124669+00:00", "epoch": 0, "step": 22921, "train_loss": 3.5443356037139893, "perplexity": 34.61667850764492, "lr": 0.0026291804804649314, "grad_norm": 0.153058, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:24.429351+00:00", "epoch": 0, "step": 22922, "train_loss": 3.521404981613159, "perplexity": 33.83192832505454, "lr": 0.0026291804804649314, "grad_norm": 0.158513, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:24.733645+00:00", "epoch": 0, "step": 22923, "train_loss": 3.489042043685913, "perplexity": 32.75455523932001, "lr": 0.0026291804804649314, "grad_norm": 0.177793, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:25.038536+00:00", "epoch": 0, "step": 22924, "train_loss": 3.540863513946533, "perplexity": 34.49669471021065, "lr": 0.0026291804804649314, "grad_norm": 0.177958, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:25.343255+00:00", "epoch": 0, "step": 22925, "train_loss": 3.4893276691436768, "perplexity": 32.76391211037077, "lr": 0.0026291804804649314, "grad_norm": 0.177473, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:25.648054+00:00", "epoch": 0, "step": 22926, "train_loss": 3.4973161220550537, "perplexity": 33.0266932895396, "lr": 0.0026291804804649314, "grad_norm": 0.178323, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:25.953420+00:00", "epoch": 0, "step": 22927, "train_loss": 3.5731332302093506, "perplexity": 35.62804933197837, "lr": 0.0026291804804649314, "grad_norm": 0.183668, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:26.257126+00:00", "epoch": 0, "step": 22928, "train_loss": 3.4916772842407227, "perplexity": 32.84098520349711, "lr": 0.0026291804804649314, "grad_norm": 0.165086, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:26.561838+00:00", "epoch": 0, "step": 22929, "train_loss": 3.4567251205444336, "perplexity": 31.71295018176389, "lr": 0.0026291804804649314, "grad_norm": 0.160878, "tokens_per_sec": 107539} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:26.867473+00:00", "epoch": 0, "step": 22930, "train_loss": 3.4864869117736816, "perplexity": 32.67096986126582, "lr": 0.0026291804804649314, "grad_norm": 0.181937, "tokens_per_sec": 107275} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:27.171484+00:00", "epoch": 0, "step": 22931, "train_loss": 3.460905075073242, "perplexity": 31.84578630256084, "lr": 0.0026291804804649314, "grad_norm": 0.1569, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:27.475354+00:00", "epoch": 0, "step": 22932, "train_loss": 3.580901861190796, "perplexity": 35.905908394652805, "lr": 0.0026291804804649314, "grad_norm": 0.175655, "tokens_per_sec": 107835} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:27.779809+00:00", "epoch": 0, "step": 22933, "train_loss": 3.466989755630493, "perplexity": 32.04014845545185, "lr": 0.0026291804804649314, "grad_norm": 0.184676, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:28.085942+00:00", "epoch": 0, "step": 22934, "train_loss": 3.4442601203918457, "perplexity": 31.320101767278036, "lr": 0.0026291804804649314, "grad_norm": 0.185443, "tokens_per_sec": 107038} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:28.391873+00:00", "epoch": 0, "step": 22935, "train_loss": 3.5498924255371094, "perplexity": 34.8095726650991, "lr": 0.0026291804804649314, "grad_norm": 0.174146, "tokens_per_sec": 107109} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:28.696261+00:00", "epoch": 0, "step": 22936, "train_loss": 3.496358633041382, "perplexity": 32.99508572791824, "lr": 0.0026291804804649314, "grad_norm": 0.16746, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:29.000422+00:00", "epoch": 0, "step": 22937, "train_loss": 3.4853110313415527, "perplexity": 32.6325752852466, "lr": 0.0026291804804649314, "grad_norm": 0.170117, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:29.305250+00:00", "epoch": 0, "step": 22938, "train_loss": 3.425576686859131, "perplexity": 30.740367306728878, "lr": 0.0026291804804649314, "grad_norm": 0.176403, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:29.610176+00:00", "epoch": 0, "step": 22939, "train_loss": 3.4644696712493896, "perplexity": 31.959506233042323, "lr": 0.0026291804804649314, "grad_norm": 0.16713, "tokens_per_sec": 107462} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:29.915657+00:00", "epoch": 0, "step": 22940, "train_loss": 3.489840269088745, "perplexity": 32.780711195155895, "lr": 0.0026291804804649314, "grad_norm": 0.168212, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:30.222673+00:00", "epoch": 0, "step": 22941, "train_loss": 3.50118088722229, "perplexity": 33.15458067152213, "lr": 0.0026291804804649314, "grad_norm": 0.149358, "tokens_per_sec": 106676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:30.526396+00:00", "epoch": 0, "step": 22942, "train_loss": 3.552027940750122, "perplexity": 34.8839884668416, "lr": 0.0026291804804649314, "grad_norm": 0.162734, "tokens_per_sec": 107881} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:30.831018+00:00", "epoch": 0, "step": 22943, "train_loss": 3.573729991912842, "perplexity": 35.64931713266311, "lr": 0.0026291804804649314, "grad_norm": 0.154464, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:31.136280+00:00", "epoch": 0, "step": 22944, "train_loss": 3.546642303466797, "perplexity": 34.69662095756936, "lr": 0.0026291804804649314, "grad_norm": 0.177363, "tokens_per_sec": 107343} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:31.442522+00:00", "epoch": 0, "step": 22945, "train_loss": 3.4965009689331055, "perplexity": 32.99978244711575, "lr": 0.0026291804804649314, "grad_norm": 0.183536, "tokens_per_sec": 107001} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:31.747320+00:00", "epoch": 0, "step": 22946, "train_loss": 3.5341038703918457, "perplexity": 34.26429570229514, "lr": 0.0026291804804649314, "grad_norm": 0.1814, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:32.051463+00:00", "epoch": 0, "step": 22947, "train_loss": 3.419459104537964, "perplexity": 30.552884634196865, "lr": 0.0026291804804649314, "grad_norm": 0.179489, "tokens_per_sec": 107744} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:32.355635+00:00", "epoch": 0, "step": 22948, "train_loss": 3.3960280418395996, "perplexity": 29.845319946650594, "lr": 0.0026291804804649314, "grad_norm": 0.193543, "tokens_per_sec": 107723} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:32.659742+00:00", "epoch": 0, "step": 22949, "train_loss": 3.5970048904418945, "perplexity": 36.48878271366866, "lr": 0.0026291804804649314, "grad_norm": 0.159641, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:32.964364+00:00", "epoch": 0, "step": 22950, "train_loss": 3.491852283477783, "perplexity": 32.846732853754574, "lr": 0.0026291804804649314, "grad_norm": 0.169963, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:33.269368+00:00", "epoch": 0, "step": 22951, "train_loss": 3.519747257232666, "perplexity": 33.775890772872074, "lr": 0.0026291804804649314, "grad_norm": 0.163623, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:33.573246+00:00", "epoch": 0, "step": 22952, "train_loss": 3.567875385284424, "perplexity": 35.441214179163445, "lr": 0.0026291804804649314, "grad_norm": 0.16664, "tokens_per_sec": 107834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:33.877707+00:00", "epoch": 0, "step": 22953, "train_loss": 3.413658380508423, "perplexity": 30.37616881748643, "lr": 0.0026291804804649314, "grad_norm": 0.166312, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:34.181033+00:00", "epoch": 0, "step": 22954, "train_loss": 3.4842031002044678, "perplexity": 32.59644066003931, "lr": 0.0026291804804649314, "grad_norm": 0.193954, "tokens_per_sec": 108024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:34.485928+00:00", "epoch": 0, "step": 22955, "train_loss": 3.563619375228882, "perplexity": 35.290696544720724, "lr": 0.0026291804804649314, "grad_norm": 0.208344, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:34.792280+00:00", "epoch": 0, "step": 22956, "train_loss": 3.4168524742126465, "perplexity": 30.47334826457425, "lr": 0.0026291804804649314, "grad_norm": 0.193034, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:35.097097+00:00", "epoch": 0, "step": 22957, "train_loss": 3.503486156463623, "perplexity": 33.23109907042258, "lr": 0.0026291804804649314, "grad_norm": 0.174275, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:35.401085+00:00", "epoch": 0, "step": 22958, "train_loss": 3.5134856700897217, "perplexity": 33.56506084329731, "lr": 0.0026291804804649314, "grad_norm": 0.167426, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:35.705508+00:00", "epoch": 0, "step": 22959, "train_loss": 3.430713415145874, "perplexity": 30.898678473643407, "lr": 0.0026291804804649314, "grad_norm": 0.170709, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:36.010286+00:00", "epoch": 0, "step": 22960, "train_loss": 3.5087311267852783, "perplexity": 33.4058530880016, "lr": 0.0026291804804649314, "grad_norm": 0.159553, "tokens_per_sec": 107582} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:36.315361+00:00", "epoch": 0, "step": 22961, "train_loss": 3.5071840286254883, "perplexity": 33.35421091222101, "lr": 0.0026291804804649314, "grad_norm": 0.170307, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:36.620662+00:00", "epoch": 0, "step": 22962, "train_loss": 3.4630110263824463, "perplexity": 31.912922646039927, "lr": 0.0026291804804649314, "grad_norm": 0.173036, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:36.924732+00:00", "epoch": 0, "step": 22963, "train_loss": 3.508453130722046, "perplexity": 33.39656768306702, "lr": 0.0026291804804649314, "grad_norm": 0.164343, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:37.228869+00:00", "epoch": 0, "step": 22964, "train_loss": 3.5296998023986816, "perplexity": 34.113725218977194, "lr": 0.0026291804804649314, "grad_norm": 0.179823, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:37.532827+00:00", "epoch": 0, "step": 22965, "train_loss": 3.4622089862823486, "perplexity": 31.88733746390902, "lr": 0.0026291804804649314, "grad_norm": 0.185686, "tokens_per_sec": 107804} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:37.836878+00:00", "epoch": 0, "step": 22966, "train_loss": 3.455552577972412, "perplexity": 31.67578718944449, "lr": 0.0026291804804649314, "grad_norm": 0.18518, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:38.141167+00:00", "epoch": 0, "step": 22967, "train_loss": 3.582307815551758, "perplexity": 35.95642596752397, "lr": 0.0026291804804649314, "grad_norm": 0.163327, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:38.445049+00:00", "epoch": 0, "step": 22968, "train_loss": 3.537966728210449, "perplexity": 34.396909774598804, "lr": 0.0026291804804649314, "grad_norm": 0.169949, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:38.748606+00:00", "epoch": 0, "step": 22969, "train_loss": 3.4590721130371094, "perplexity": 31.78746764952396, "lr": 0.0026291804804649314, "grad_norm": 0.207414, "tokens_per_sec": 107946} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:39.053667+00:00", "epoch": 0, "step": 22970, "train_loss": 3.522531747817993, "perplexity": 33.870070483147494, "lr": 0.0026291804804649314, "grad_norm": 0.23701, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:39.358638+00:00", "epoch": 0, "step": 22971, "train_loss": 3.4486725330352783, "perplexity": 31.45860432080551, "lr": 0.0026291804804649314, "grad_norm": 0.171858, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:39.663252+00:00", "epoch": 0, "step": 22972, "train_loss": 3.4430484771728516, "perplexity": 31.2821759592674, "lr": 0.0026291804804649314, "grad_norm": 0.224637, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:39.967907+00:00", "epoch": 0, "step": 22973, "train_loss": 3.560567855834961, "perplexity": 35.18317044222128, "lr": 0.0026291804804649314, "grad_norm": 0.242588, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:40.272300+00:00", "epoch": 0, "step": 22974, "train_loss": 3.463297128677368, "perplexity": 31.922054312679727, "lr": 0.0026291804804649314, "grad_norm": 0.181055, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:40.577177+00:00", "epoch": 0, "step": 22975, "train_loss": 3.4413115978240967, "perplexity": 31.22788975180329, "lr": 0.0026291804804649314, "grad_norm": 0.178409, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:40.882741+00:00", "epoch": 0, "step": 22976, "train_loss": 3.5268807411193848, "perplexity": 34.01769196253605, "lr": 0.0026291804804649314, "grad_norm": 0.162834, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:41.187112+00:00", "epoch": 0, "step": 22977, "train_loss": 3.4916303157806396, "perplexity": 32.83944274921813, "lr": 0.0026291804804649314, "grad_norm": 0.159131, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:41.491762+00:00", "epoch": 0, "step": 22978, "train_loss": 3.451723337173462, "perplexity": 31.554724909040313, "lr": 0.0026291804804649314, "grad_norm": 0.152259, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:41.795484+00:00", "epoch": 0, "step": 22979, "train_loss": 3.547849178314209, "perplexity": 34.738520715489464, "lr": 0.0026291804804649314, "grad_norm": 0.158528, "tokens_per_sec": 107888} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:42.100211+00:00", "epoch": 0, "step": 22980, "train_loss": 3.5061800479888916, "perplexity": 33.32074073483261, "lr": 0.0026291804804649314, "grad_norm": 0.154369, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:42.405725+00:00", "epoch": 0, "step": 22981, "train_loss": 3.5694780349731445, "perplexity": 35.49805956949191, "lr": 0.0026291804804649314, "grad_norm": 0.168288, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:42.711502+00:00", "epoch": 0, "step": 22982, "train_loss": 3.3859975337982178, "perplexity": 29.547452602714774, "lr": 0.0026291804804649314, "grad_norm": 0.162918, "tokens_per_sec": 107161} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:43.018932+00:00", "epoch": 0, "step": 22983, "train_loss": 3.522294282913208, "perplexity": 33.8620284849691, "lr": 0.0026291804804649314, "grad_norm": 0.163935, "tokens_per_sec": 106588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:43.323500+00:00", "epoch": 0, "step": 22984, "train_loss": 3.5354974269866943, "perplexity": 34.312078223629975, "lr": 0.0026291804804649314, "grad_norm": 0.142661, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:43.626973+00:00", "epoch": 0, "step": 22985, "train_loss": 3.5413637161254883, "perplexity": 34.51395434836588, "lr": 0.0026291804804649314, "grad_norm": 0.161719, "tokens_per_sec": 107977} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:43.931772+00:00", "epoch": 0, "step": 22986, "train_loss": 3.491848945617676, "perplexity": 32.846623216138305, "lr": 0.0026291804804649314, "grad_norm": 0.162698, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:44.237329+00:00", "epoch": 0, "step": 22987, "train_loss": 3.5024795532226562, "perplexity": 33.197665368456846, "lr": 0.0026291804804649314, "grad_norm": 0.198216, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:44.543450+00:00", "epoch": 0, "step": 22988, "train_loss": 3.542234182357788, "perplexity": 34.54401065977351, "lr": 0.0026291804804649314, "grad_norm": 0.170272, "tokens_per_sec": 107042} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:44.848187+00:00", "epoch": 0, "step": 22989, "train_loss": 3.3837218284606934, "perplexity": 29.480287759798586, "lr": 0.0026291804804649314, "grad_norm": 0.187918, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:45.152163+00:00", "epoch": 0, "step": 22990, "train_loss": 3.520237684249878, "perplexity": 33.792459444767566, "lr": 0.0026291804804649314, "grad_norm": 0.156044, "tokens_per_sec": 107856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:45.455346+00:00", "epoch": 0, "step": 22991, "train_loss": 3.371486186981201, "perplexity": 29.121775315681404, "lr": 0.0026291804804649314, "grad_norm": 0.18707, "tokens_per_sec": 108021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:45.759752+00:00", "epoch": 0, "step": 22992, "train_loss": 3.597785472869873, "perplexity": 36.5172763356368, "lr": 0.0026291804804649314, "grad_norm": 0.174997, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:46.064655+00:00", "epoch": 0, "step": 22993, "train_loss": 3.5864267349243164, "perplexity": 36.10483301552554, "lr": 0.0026291804804649314, "grad_norm": 0.23049, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:46.370185+00:00", "epoch": 0, "step": 22994, "train_loss": 3.507359266281128, "perplexity": 33.36005633810139, "lr": 0.0026291804804649314, "grad_norm": 0.219094, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:46.674181+00:00", "epoch": 0, "step": 22995, "train_loss": 3.5116074085235596, "perplexity": 33.50207604902545, "lr": 0.0026291804804649314, "grad_norm": 0.17023, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:46.978352+00:00", "epoch": 0, "step": 22996, "train_loss": 3.457533836364746, "perplexity": 31.738607319552923, "lr": 0.0026291804804649314, "grad_norm": 0.182223, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:47.283576+00:00", "epoch": 0, "step": 22997, "train_loss": 3.4655675888061523, "perplexity": 31.994614405453802, "lr": 0.0026291804804649314, "grad_norm": 0.191417, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:47.589317+00:00", "epoch": 0, "step": 22998, "train_loss": 3.5195844173431396, "perplexity": 33.77039115834043, "lr": 0.0026291804804649314, "grad_norm": 0.185609, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:47.893873+00:00", "epoch": 0, "step": 22999, "train_loss": 3.361471652984619, "perplexity": 28.831589768498684, "lr": 0.0026291804804649314, "grad_norm": 0.165577, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:48.198028+00:00", "epoch": 0, "step": 23000, "train_loss": 3.5790553092956543, "perplexity": 35.8396674489539, "lr": 0.0026291804804649314, "grad_norm": 0.165592, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T22:58:51.270796+00:00", "step": 23000, "epoch": 0, "val_loss": 3.4519365787506104, "val_ppl": 31.56145440582516, "eval_train_loss": 3.5790553092956543, "eval_train_ppl": 35.8396674489539} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T22:58:52.255080+00:00", "step": 23000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4519_epoch_0000_step_0023000.pt", "val_loss": 3.4519365787506104} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:53.651615+00:00", "epoch": 0, "step": 23001, "train_loss": 3.5538761615753174, "perplexity": 34.94852139797557, "lr": 0.0026291804804649314, "grad_norm": 0.167755, "tokens_per_sec": 6008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:53.955234+00:00", "epoch": 0, "step": 23002, "train_loss": 3.541808843612671, "perplexity": 34.52932087791632, "lr": 0.0026291804804649314, "grad_norm": 0.160293, "tokens_per_sec": 107919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:54.259603+00:00", "epoch": 0, "step": 23003, "train_loss": 3.397449493408203, "perplexity": 29.887773789396665, "lr": 0.0026291804804649314, "grad_norm": 0.165772, "tokens_per_sec": 107659} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:54.564440+00:00", "epoch": 0, "step": 23004, "train_loss": 3.639660358428955, "perplexity": 38.078901350949415, "lr": 0.0026291804804649314, "grad_norm": 0.183487, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:54.869182+00:00", "epoch": 0, "step": 23005, "train_loss": 3.540879011154175, "perplexity": 34.49722931679396, "lr": 0.0026291804804649314, "grad_norm": 0.16632, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:55.173737+00:00", "epoch": 0, "step": 23006, "train_loss": 3.399773359298706, "perplexity": 29.957309732258587, "lr": 0.0026291804804649314, "grad_norm": 0.161368, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:55.478693+00:00", "epoch": 0, "step": 23007, "train_loss": 3.4343154430389404, "perplexity": 31.01017706532885, "lr": 0.0026291804804649314, "grad_norm": 0.157564, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:55.783908+00:00", "epoch": 0, "step": 23008, "train_loss": 3.522390842437744, "perplexity": 33.86529834420492, "lr": 0.0026291804804649314, "grad_norm": 0.178941, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:56.087861+00:00", "epoch": 0, "step": 23009, "train_loss": 3.49216628074646, "perplexity": 32.85704825757517, "lr": 0.0026291804804649314, "grad_norm": 0.175502, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:56.391789+00:00", "epoch": 0, "step": 23010, "train_loss": 3.581299066543579, "perplexity": 35.92017324651392, "lr": 0.0026291804804649314, "grad_norm": 0.172036, "tokens_per_sec": 107873} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:56.696478+00:00", "epoch": 0, "step": 23011, "train_loss": 3.517760753631592, "perplexity": 33.70886144326717, "lr": 0.0026291804804649314, "grad_norm": 0.164313, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:57.001365+00:00", "epoch": 0, "step": 23012, "train_loss": 3.54473876953125, "perplexity": 34.63063758284821, "lr": 0.0026291804804649314, "grad_norm": 0.187003, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:57.305446+00:00", "epoch": 0, "step": 23013, "train_loss": 3.5170888900756836, "perplexity": 33.68622129414947, "lr": 0.0026291804804649314, "grad_norm": 0.179999, "tokens_per_sec": 107761} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:57.610493+00:00", "epoch": 0, "step": 23014, "train_loss": 3.506948709487915, "perplexity": 33.3463629514986, "lr": 0.0026291804804649314, "grad_norm": 0.186093, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:57.916738+00:00", "epoch": 0, "step": 23015, "train_loss": 3.4151644706726074, "perplexity": 30.421952535117164, "lr": 0.0026291804804649314, "grad_norm": 0.189312, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:58.221067+00:00", "epoch": 0, "step": 23016, "train_loss": 3.5049517154693604, "perplexity": 33.279836912303814, "lr": 0.0026291804804649314, "grad_norm": 0.194069, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:58.525490+00:00", "epoch": 0, "step": 23017, "train_loss": 3.4156582355499268, "perplexity": 30.436977535882047, "lr": 0.0026291804804649314, "grad_norm": 0.171943, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:58.830596+00:00", "epoch": 0, "step": 23018, "train_loss": 3.47145676612854, "perplexity": 32.183592278968824, "lr": 0.0026291804804649314, "grad_norm": 0.173508, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:59.135432+00:00", "epoch": 0, "step": 23019, "train_loss": 3.5041966438293457, "perplexity": 33.25471773584934, "lr": 0.0026291804804649314, "grad_norm": 0.182682, "tokens_per_sec": 107494} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:59.440193+00:00", "epoch": 0, "step": 23020, "train_loss": 3.5169270038604736, "perplexity": 33.68076840066549, "lr": 0.0026291804804649314, "grad_norm": 0.189601, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:58:59.744725+00:00", "epoch": 0, "step": 23021, "train_loss": 3.518296241760254, "perplexity": 33.726916972230015, "lr": 0.0026291804804649314, "grad_norm": 0.17046, "tokens_per_sec": 107542} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:00.049866+00:00", "epoch": 0, "step": 23022, "train_loss": 3.514622211456299, "perplexity": 33.603230610068096, "lr": 0.0026291804804649314, "grad_norm": 0.17167, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:00.354714+00:00", "epoch": 0, "step": 23023, "train_loss": 3.4950695037841797, "perplexity": 32.95257820229673, "lr": 0.0026291804804649314, "grad_norm": 0.194353, "tokens_per_sec": 107488} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:00.659311+00:00", "epoch": 0, "step": 23024, "train_loss": 3.50732159614563, "perplexity": 33.35879968392822, "lr": 0.0026291804804649314, "grad_norm": 0.156031, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:00.964043+00:00", "epoch": 0, "step": 23025, "train_loss": 3.4826438426971436, "perplexity": 32.54565402024251, "lr": 0.0026291804804649314, "grad_norm": 0.189592, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:01.268607+00:00", "epoch": 0, "step": 23026, "train_loss": 3.4078147411346436, "perplexity": 30.199179077123077, "lr": 0.0026291804804649314, "grad_norm": 0.159687, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:01.572693+00:00", "epoch": 0, "step": 23027, "train_loss": 3.447972059249878, "perplexity": 31.43657610914661, "lr": 0.0026291804804649314, "grad_norm": 0.182714, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:01.878012+00:00", "epoch": 0, "step": 23028, "train_loss": 3.445249557495117, "perplexity": 31.351106374069563, "lr": 0.0026291804804649314, "grad_norm": 0.208625, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:02.182762+00:00", "epoch": 0, "step": 23029, "train_loss": 3.4865312576293945, "perplexity": 32.67241871550639, "lr": 0.0026291804804649314, "grad_norm": 0.193964, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:02.488059+00:00", "epoch": 0, "step": 23030, "train_loss": 3.43241548538208, "perplexity": 30.951314977417184, "lr": 0.0026291804804649314, "grad_norm": 0.172853, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:02.792226+00:00", "epoch": 0, "step": 23031, "train_loss": 3.5715904235839844, "perplexity": 35.573124521495444, "lr": 0.0026291804804649314, "grad_norm": 0.182219, "tokens_per_sec": 107673} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:03.095760+00:00", "epoch": 0, "step": 23032, "train_loss": 3.479567527770996, "perplexity": 32.44568718257434, "lr": 0.0026291804804649314, "grad_norm": 0.178178, "tokens_per_sec": 107956} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:03.400697+00:00", "epoch": 0, "step": 23033, "train_loss": 3.529705762863159, "perplexity": 34.11392855323054, "lr": 0.0026291804804649314, "grad_norm": 0.189317, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:03.705585+00:00", "epoch": 0, "step": 23034, "train_loss": 3.4763851165771484, "perplexity": 32.34259579109845, "lr": 0.0026291804804649314, "grad_norm": 0.191651, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:04.011012+00:00", "epoch": 0, "step": 23035, "train_loss": 3.436600685119629, "perplexity": 31.08112386131434, "lr": 0.0026291804804649314, "grad_norm": 0.16962, "tokens_per_sec": 107278} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:04.315515+00:00", "epoch": 0, "step": 23036, "train_loss": 3.4198853969573975, "perplexity": 30.56591187381772, "lr": 0.0026291804804649314, "grad_norm": 0.16083, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:04.619646+00:00", "epoch": 0, "step": 23037, "train_loss": 3.553713321685791, "perplexity": 34.942830847948855, "lr": 0.0026291804804649314, "grad_norm": 0.161144, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:04.923812+00:00", "epoch": 0, "step": 23038, "train_loss": 3.5410072803497314, "perplexity": 34.50165453245072, "lr": 0.0026291804804649314, "grad_norm": 0.197945, "tokens_per_sec": 107731} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:05.229430+00:00", "epoch": 0, "step": 23039, "train_loss": 3.491375207901001, "perplexity": 32.83106621711496, "lr": 0.0026291804804649314, "grad_norm": 0.182624, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:05.534841+00:00", "epoch": 0, "step": 23040, "train_loss": 3.486163377761841, "perplexity": 32.66040140103623, "lr": 0.0026291804804649314, "grad_norm": 0.185062, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:05.838774+00:00", "epoch": 0, "step": 23041, "train_loss": 3.3758633136749268, "perplexity": 29.24952439878069, "lr": 0.0026291804804649314, "grad_norm": 0.202576, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:06.142771+00:00", "epoch": 0, "step": 23042, "train_loss": 3.5351502895355225, "perplexity": 34.30016928339018, "lr": 0.0026291804804649314, "grad_norm": 0.207638, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:06.446890+00:00", "epoch": 0, "step": 23043, "train_loss": 3.5447349548339844, "perplexity": 34.63050547770168, "lr": 0.0026291804804649314, "grad_norm": 0.197233, "tokens_per_sec": 107746} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:06.750907+00:00", "epoch": 0, "step": 23044, "train_loss": 3.506648302078247, "perplexity": 33.33634696149662, "lr": 0.0026291804804649314, "grad_norm": 0.164852, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:07.054543+00:00", "epoch": 0, "step": 23045, "train_loss": 3.4093995094299316, "perplexity": 30.24707572118538, "lr": 0.0026291804804649314, "grad_norm": 0.174895, "tokens_per_sec": 107923} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:07.359171+00:00", "epoch": 0, "step": 23046, "train_loss": 3.5334556102752686, "perplexity": 34.24209072404725, "lr": 0.0026291804804649314, "grad_norm": 0.197086, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:07.663381+00:00", "epoch": 0, "step": 23047, "train_loss": 3.5750057697296143, "perplexity": 35.694826764560794, "lr": 0.0026291804804649314, "grad_norm": 0.186102, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:07.967217+00:00", "epoch": 0, "step": 23048, "train_loss": 3.5510411262512207, "perplexity": 34.849581420721144, "lr": 0.0026291804804649314, "grad_norm": 0.143988, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:08.271272+00:00", "epoch": 0, "step": 23049, "train_loss": 3.5121004581451416, "perplexity": 33.518598307755624, "lr": 0.0026291804804649314, "grad_norm": 0.196849, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:08.576194+00:00", "epoch": 0, "step": 23050, "train_loss": 3.602585792541504, "perplexity": 36.692992344672994, "lr": 0.0026291804804649314, "grad_norm": 0.195636, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:08.880498+00:00", "epoch": 0, "step": 23051, "train_loss": 3.5042765140533447, "perplexity": 33.25737390367688, "lr": 0.0026291804804649314, "grad_norm": 0.16404, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:09.184346+00:00", "epoch": 0, "step": 23052, "train_loss": 3.4518349170684814, "perplexity": 31.558245978369655, "lr": 0.0026291804804649314, "grad_norm": 0.186703, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:09.487667+00:00", "epoch": 0, "step": 23053, "train_loss": 3.5100677013397217, "perplexity": 33.450532353141575, "lr": 0.0026291804804649314, "grad_norm": 0.165754, "tokens_per_sec": 108031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:09.791749+00:00", "epoch": 0, "step": 23054, "train_loss": 3.6049795150756836, "perplexity": 36.78093039495146, "lr": 0.0026291804804649314, "grad_norm": 0.168614, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:10.097087+00:00", "epoch": 0, "step": 23055, "train_loss": 3.536439895629883, "perplexity": 34.344431525029606, "lr": 0.0026291804804649314, "grad_norm": 0.16138, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:10.401861+00:00", "epoch": 0, "step": 23056, "train_loss": 3.567011594772339, "perplexity": 35.41061361276085, "lr": 0.0026291804804649314, "grad_norm": 0.18332, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:10.706205+00:00", "epoch": 0, "step": 23057, "train_loss": 3.4546852111816406, "perplexity": 31.648324575366274, "lr": 0.0026291804804649314, "grad_norm": 0.152652, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:11.010567+00:00", "epoch": 0, "step": 23058, "train_loss": 3.4744396209716797, "perplexity": 32.27973458104859, "lr": 0.0026291804804649314, "grad_norm": 0.182352, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:11.314265+00:00", "epoch": 0, "step": 23059, "train_loss": 3.5139992237091064, "perplexity": 33.58230272871564, "lr": 0.0026291804804649314, "grad_norm": 0.17745, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:11.619501+00:00", "epoch": 0, "step": 23060, "train_loss": 3.4438722133636475, "perplexity": 31.30795483577099, "lr": 0.0026291804804649314, "grad_norm": 0.187644, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:11.923770+00:00", "epoch": 0, "step": 23061, "train_loss": 3.4458253383636475, "perplexity": 31.369162939140523, "lr": 0.0026291804804649314, "grad_norm": 0.17086, "tokens_per_sec": 107636} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:12.228644+00:00", "epoch": 0, "step": 23062, "train_loss": 3.524531126022339, "perplexity": 33.93785730701869, "lr": 0.0026291804804649314, "grad_norm": 0.171273, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:12.532688+00:00", "epoch": 0, "step": 23063, "train_loss": 3.4337213039398193, "perplexity": 30.99175817889579, "lr": 0.0026291804804649314, "grad_norm": 0.17652, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:12.836188+00:00", "epoch": 0, "step": 23064, "train_loss": 3.5216448307037354, "perplexity": 33.84004385550814, "lr": 0.0026291804804649314, "grad_norm": 0.183429, "tokens_per_sec": 107967} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:13.140795+00:00", "epoch": 0, "step": 23065, "train_loss": 3.555426836013794, "perplexity": 35.00275721695111, "lr": 0.0026291804804649314, "grad_norm": 0.159345, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:13.446150+00:00", "epoch": 0, "step": 23066, "train_loss": 3.517829418182373, "perplexity": 33.711176126562954, "lr": 0.0026291804804649314, "grad_norm": 0.19556, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:13.750749+00:00", "epoch": 0, "step": 23067, "train_loss": 3.5537545680999756, "perplexity": 34.94427214414674, "lr": 0.0026291804804649314, "grad_norm": 0.156568, "tokens_per_sec": 107579} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:14.055188+00:00", "epoch": 0, "step": 23068, "train_loss": 3.4928526878356934, "perplexity": 32.879609310578005, "lr": 0.0026291804804649314, "grad_norm": 0.187613, "tokens_per_sec": 107633} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:14.359035+00:00", "epoch": 0, "step": 23069, "train_loss": 3.6888015270233154, "perplexity": 39.99688303782464, "lr": 0.0026291804804649314, "grad_norm": 0.193255, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:14.664003+00:00", "epoch": 0, "step": 23070, "train_loss": 3.4572038650512695, "perplexity": 31.72813621728447, "lr": 0.0026291804804649314, "grad_norm": 0.232106, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:14.970079+00:00", "epoch": 0, "step": 23071, "train_loss": 3.4787158966064453, "perplexity": 32.418067186908274, "lr": 0.0026291804804649314, "grad_norm": 0.222184, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:15.274447+00:00", "epoch": 0, "step": 23072, "train_loss": 3.4510490894317627, "perplexity": 31.533456377965717, "lr": 0.0026291804804649314, "grad_norm": 0.201215, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:15.579192+00:00", "epoch": 0, "step": 23073, "train_loss": 3.4048378467559814, "perplexity": 30.109412989061152, "lr": 0.0026291804804649314, "grad_norm": 0.181201, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:15.882850+00:00", "epoch": 0, "step": 23074, "train_loss": 3.5544559955596924, "perplexity": 34.96879161449806, "lr": 0.0026291804804649314, "grad_norm": 0.184686, "tokens_per_sec": 107908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:16.186823+00:00", "epoch": 0, "step": 23075, "train_loss": 3.460273265838623, "perplexity": 31.825672195499568, "lr": 0.0026291804804649314, "grad_norm": 0.184478, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:16.491242+00:00", "epoch": 0, "step": 23076, "train_loss": 3.4741272926330566, "perplexity": 32.26965427943957, "lr": 0.0026291804804649314, "grad_norm": 0.165848, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:16.795128+00:00", "epoch": 0, "step": 23077, "train_loss": 3.4622631072998047, "perplexity": 31.88906328575773, "lr": 0.0026291804804649314, "grad_norm": 0.19967, "tokens_per_sec": 107830} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:17.098978+00:00", "epoch": 0, "step": 23078, "train_loss": 3.510120153427124, "perplexity": 33.452286949403934, "lr": 0.0026291804804649314, "grad_norm": 0.186701, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:17.404514+00:00", "epoch": 0, "step": 23079, "train_loss": 3.5227818489074707, "perplexity": 33.878542484060176, "lr": 0.0026291804804649314, "grad_norm": 0.178895, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:17.708802+00:00", "epoch": 0, "step": 23080, "train_loss": 3.4193663597106934, "perplexity": 30.55005114358668, "lr": 0.0026291804804649314, "grad_norm": 0.161643, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:18.013851+00:00", "epoch": 0, "step": 23081, "train_loss": 3.5028738975524902, "perplexity": 33.210759261140105, "lr": 0.0026291804804649314, "grad_norm": 0.181583, "tokens_per_sec": 107360} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:18.320084+00:00", "epoch": 0, "step": 23082, "train_loss": 3.5866873264312744, "perplexity": 36.11424285437832, "lr": 0.0026291804804649314, "grad_norm": 0.169406, "tokens_per_sec": 107004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:18.623966+00:00", "epoch": 0, "step": 23083, "train_loss": 3.63649320602417, "perplexity": 37.95849044745746, "lr": 0.0026291804804649314, "grad_norm": 0.153808, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:18.928011+00:00", "epoch": 0, "step": 23084, "train_loss": 3.5588929653167725, "perplexity": 35.124291805047505, "lr": 0.0026291804804649314, "grad_norm": 0.162215, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:19.232659+00:00", "epoch": 0, "step": 23085, "train_loss": 3.653909206390381, "perplexity": 38.6253658257691, "lr": 0.0026291804804649314, "grad_norm": 0.169312, "tokens_per_sec": 107563} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:19.538505+00:00", "epoch": 0, "step": 23086, "train_loss": 3.5877490043640137, "perplexity": 36.15260490954476, "lr": 0.0026291804804649314, "grad_norm": 0.158621, "tokens_per_sec": 107137} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:19.844861+00:00", "epoch": 0, "step": 23087, "train_loss": 3.506150484085083, "perplexity": 33.319755658220124, "lr": 0.0026291804804649314, "grad_norm": 0.157987, "tokens_per_sec": 106961} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:20.149701+00:00", "epoch": 0, "step": 23088, "train_loss": 3.4403250217437744, "perplexity": 31.197096255296454, "lr": 0.0026291804804649314, "grad_norm": 0.15855, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:20.454905+00:00", "epoch": 0, "step": 23089, "train_loss": 3.455655813217163, "perplexity": 31.67905741588608, "lr": 0.0026291804804649314, "grad_norm": 0.182671, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:20.760249+00:00", "epoch": 0, "step": 23090, "train_loss": 3.454409599304199, "perplexity": 31.639603123135302, "lr": 0.0026291804804649314, "grad_norm": 0.155659, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:21.065903+00:00", "epoch": 0, "step": 23091, "train_loss": 3.5290513038635254, "perplexity": 34.091609689863134, "lr": 0.0026291804804649314, "grad_norm": 0.185293, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:21.370516+00:00", "epoch": 0, "step": 23092, "train_loss": 3.4159843921661377, "perplexity": 30.446906376573363, "lr": 0.0026291804804649314, "grad_norm": 0.183374, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:21.675709+00:00", "epoch": 0, "step": 23093, "train_loss": 3.514177083969116, "perplexity": 33.58827621701804, "lr": 0.0026291804804649314, "grad_norm": 0.175126, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:21.980735+00:00", "epoch": 0, "step": 23094, "train_loss": 3.5850937366485596, "perplexity": 36.056737398178555, "lr": 0.0026291804804649314, "grad_norm": 0.17213, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:22.284564+00:00", "epoch": 0, "step": 23095, "train_loss": 3.5558853149414062, "perplexity": 35.01880892294671, "lr": 0.0026291804804649314, "grad_norm": 0.1569, "tokens_per_sec": 107850} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:22.590139+00:00", "epoch": 0, "step": 23096, "train_loss": 3.4583427906036377, "perplexity": 31.76429278826426, "lr": 0.0026291804804649314, "grad_norm": 0.174551, "tokens_per_sec": 107238} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:22.895072+00:00", "epoch": 0, "step": 23097, "train_loss": 3.4823246002197266, "perplexity": 32.53526572330499, "lr": 0.0026291804804649314, "grad_norm": 0.183639, "tokens_per_sec": 107455} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:23.200880+00:00", "epoch": 0, "step": 23098, "train_loss": 3.4883806705474854, "perplexity": 32.732899418402084, "lr": 0.0026291804804649314, "grad_norm": 0.193207, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:23.505390+00:00", "epoch": 0, "step": 23099, "train_loss": 3.50386381149292, "perplexity": 33.24365133217879, "lr": 0.0026291804804649314, "grad_norm": 0.169284, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:23.810349+00:00", "epoch": 0, "step": 23100, "train_loss": 3.4054172039031982, "perplexity": 30.126862146833925, "lr": 0.0026291804804649314, "grad_norm": 0.165454, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:24.114831+00:00", "epoch": 0, "step": 23101, "train_loss": 3.5131258964538574, "perplexity": 33.552987191340776, "lr": 0.0026291804804649314, "grad_norm": 0.186961, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:24.420328+00:00", "epoch": 0, "step": 23102, "train_loss": 3.531428098678589, "perplexity": 34.17273482171686, "lr": 0.0026291804804649314, "grad_norm": 0.168809, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:24.725468+00:00", "epoch": 0, "step": 23103, "train_loss": 3.5167253017425537, "perplexity": 33.67397560342973, "lr": 0.0026291804804649314, "grad_norm": 0.179634, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:25.029399+00:00", "epoch": 0, "step": 23104, "train_loss": 3.512610673904419, "perplexity": 33.535704388366, "lr": 0.0026291804804649314, "grad_norm": 0.176521, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:25.334178+00:00", "epoch": 0, "step": 23105, "train_loss": 3.4365923404693604, "perplexity": 31.0808645012879, "lr": 0.0026291804804649314, "grad_norm": 0.181665, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:25.639330+00:00", "epoch": 0, "step": 23106, "train_loss": 3.4940571784973145, "perplexity": 32.91923635336093, "lr": 0.0026291804804649314, "grad_norm": 0.182988, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:25.944549+00:00", "epoch": 0, "step": 23107, "train_loss": 3.5049026012420654, "perplexity": 33.27820243896765, "lr": 0.0026291804804649314, "grad_norm": 0.171567, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:26.249737+00:00", "epoch": 0, "step": 23108, "train_loss": 3.453486919403076, "perplexity": 31.610423361114197, "lr": 0.0026291804804649314, "grad_norm": 0.166614, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:26.554356+00:00", "epoch": 0, "step": 23109, "train_loss": 3.4071896076202393, "perplexity": 30.180306457742514, "lr": 0.0026291804804649314, "grad_norm": 0.195513, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:26.859830+00:00", "epoch": 0, "step": 23110, "train_loss": 3.572871685028076, "perplexity": 35.618732205835215, "lr": 0.0026291804804649314, "grad_norm": 0.185691, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:27.165768+00:00", "epoch": 0, "step": 23111, "train_loss": 3.4297473430633545, "perplexity": 30.86884253713641, "lr": 0.0026291804804649314, "grad_norm": 0.173563, "tokens_per_sec": 107011} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:27.471095+00:00", "epoch": 0, "step": 23112, "train_loss": 3.5355923175811768, "perplexity": 34.315334271612265, "lr": 0.0026291804804649314, "grad_norm": 0.172945, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:27.775969+00:00", "epoch": 0, "step": 23113, "train_loss": 3.4999611377716064, "perplexity": 33.1141650434413, "lr": 0.0026291804804649314, "grad_norm": 0.158244, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:28.080900+00:00", "epoch": 0, "step": 23114, "train_loss": 3.4350740909576416, "perplexity": 31.033711797777134, "lr": 0.0026291804804649314, "grad_norm": 0.188312, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:28.385073+00:00", "epoch": 0, "step": 23115, "train_loss": 3.576024293899536, "perplexity": 35.73120132940385, "lr": 0.0026291804804649314, "grad_norm": 0.176085, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:28.690500+00:00", "epoch": 0, "step": 23116, "train_loss": 3.517014741897583, "perplexity": 33.68372361481375, "lr": 0.0026291804804649314, "grad_norm": 0.222154, "tokens_per_sec": 107285} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:28.996137+00:00", "epoch": 0, "step": 23117, "train_loss": 3.4362969398498535, "perplexity": 31.071684550607724, "lr": 0.0026291804804649314, "grad_norm": 0.159646, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:29.300757+00:00", "epoch": 0, "step": 23118, "train_loss": 3.461932897567749, "perplexity": 31.87853494509057, "lr": 0.0026291804804649314, "grad_norm": 0.190949, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:29.605262+00:00", "epoch": 0, "step": 23119, "train_loss": 3.4506969451904297, "perplexity": 31.522354007825477, "lr": 0.0026291804804649314, "grad_norm": 0.167381, "tokens_per_sec": 107611} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:29.909900+00:00", "epoch": 0, "step": 23120, "train_loss": 3.4682886600494385, "perplexity": 32.08179258588182, "lr": 0.0026291804804649314, "grad_norm": 0.181976, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:30.214275+00:00", "epoch": 0, "step": 23121, "train_loss": 3.4026830196380615, "perplexity": 30.044602262481114, "lr": 0.0026291804804649314, "grad_norm": 0.189653, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:30.518537+00:00", "epoch": 0, "step": 23122, "train_loss": 3.515153169631958, "perplexity": 33.621077257581504, "lr": 0.0026291804804649314, "grad_norm": 0.183739, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:30.822592+00:00", "epoch": 0, "step": 23123, "train_loss": 3.45469069480896, "perplexity": 31.648498123459362, "lr": 0.0026291804804649314, "grad_norm": 0.188262, "tokens_per_sec": 107770} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:31.127591+00:00", "epoch": 0, "step": 23124, "train_loss": 3.502938747406006, "perplexity": 33.21291304384882, "lr": 0.0026291804804649314, "grad_norm": 0.16479, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:31.431773+00:00", "epoch": 0, "step": 23125, "train_loss": 3.4754083156585693, "perplexity": 32.311018938480515, "lr": 0.0026291804804649314, "grad_norm": 0.183679, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:31.736592+00:00", "epoch": 0, "step": 23126, "train_loss": 3.499340772628784, "perplexity": 33.09362854043862, "lr": 0.0026291804804649314, "grad_norm": 0.186449, "tokens_per_sec": 107501} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:32.041846+00:00", "epoch": 0, "step": 23127, "train_loss": 3.6157467365264893, "perplexity": 37.17909855331823, "lr": 0.0026291804804649314, "grad_norm": 0.180956, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:32.345695+00:00", "epoch": 0, "step": 23128, "train_loss": 3.4506423473358154, "perplexity": 31.5206330019063, "lr": 0.0026291804804649314, "grad_norm": 0.179587, "tokens_per_sec": 107843} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:32.650937+00:00", "epoch": 0, "step": 23129, "train_loss": 3.4864020347595215, "perplexity": 32.66819696457354, "lr": 0.0026291804804649314, "grad_norm": 0.249561, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:32.956551+00:00", "epoch": 0, "step": 23130, "train_loss": 3.599264144897461, "perplexity": 36.571313352344674, "lr": 0.0026291804804649314, "grad_norm": 0.227708, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:33.261937+00:00", "epoch": 0, "step": 23131, "train_loss": 3.527219295501709, "perplexity": 34.02921075098465, "lr": 0.0026291804804649314, "grad_norm": 0.167583, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:33.565998+00:00", "epoch": 0, "step": 23132, "train_loss": 3.5007853507995605, "perplexity": 33.14146942044583, "lr": 0.0026291804804649314, "grad_norm": 0.26569, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:33.870403+00:00", "epoch": 0, "step": 23133, "train_loss": 3.4164719581604004, "perplexity": 30.461754872269715, "lr": 0.0026291804804649314, "grad_norm": 0.199169, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:34.176022+00:00", "epoch": 0, "step": 23134, "train_loss": 3.3991031646728516, "perplexity": 29.937239230593388, "lr": 0.0026291804804649314, "grad_norm": 0.192775, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:34.480661+00:00", "epoch": 0, "step": 23135, "train_loss": 3.51796293258667, "perplexity": 33.71567735464434, "lr": 0.0026291804804649314, "grad_norm": 0.229757, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:34.786021+00:00", "epoch": 0, "step": 23136, "train_loss": 3.512193441390991, "perplexity": 33.52171512072606, "lr": 0.0026291804804649314, "grad_norm": 0.169158, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:35.090063+00:00", "epoch": 0, "step": 23137, "train_loss": 3.5071561336517334, "perplexity": 33.35328051035983, "lr": 0.0026291804804649314, "grad_norm": 0.19593, "tokens_per_sec": 107773} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:35.393478+00:00", "epoch": 0, "step": 23138, "train_loss": 3.574532985687256, "perplexity": 35.67795480878177, "lr": 0.0026291804804649314, "grad_norm": 0.202026, "tokens_per_sec": 107998} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:35.697651+00:00", "epoch": 0, "step": 23139, "train_loss": 3.4385640621185303, "perplexity": 31.14220777075261, "lr": 0.0026291804804649314, "grad_norm": 0.164332, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:36.003130+00:00", "epoch": 0, "step": 23140, "train_loss": 3.4919111728668213, "perplexity": 32.84866723474093, "lr": 0.0026291804804649314, "grad_norm": 0.204074, "tokens_per_sec": 107323} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:36.308746+00:00", "epoch": 0, "step": 23141, "train_loss": 3.518026828765869, "perplexity": 33.71783172643375, "lr": 0.0026291804804649314, "grad_norm": 0.182249, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:36.614319+00:00", "epoch": 0, "step": 23142, "train_loss": 3.509381055831909, "perplexity": 33.42757157922562, "lr": 0.0026291804804649314, "grad_norm": 0.189121, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:36.919057+00:00", "epoch": 0, "step": 23143, "train_loss": 3.4192240238189697, "perplexity": 30.545703084264744, "lr": 0.0026291804804649314, "grad_norm": 0.179458, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:37.223374+00:00", "epoch": 0, "step": 23144, "train_loss": 3.479424238204956, "perplexity": 32.44103838720797, "lr": 0.0026291804804649314, "grad_norm": 0.167888, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:37.528852+00:00", "epoch": 0, "step": 23145, "train_loss": 3.5272743701934814, "perplexity": 34.03108495088804, "lr": 0.0026291804804649314, "grad_norm": 0.171287, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:37.833828+00:00", "epoch": 0, "step": 23146, "train_loss": 3.4654273986816406, "perplexity": 31.99012939086132, "lr": 0.0026291804804649314, "grad_norm": 0.173403, "tokens_per_sec": 107445} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:38.138476+00:00", "epoch": 0, "step": 23147, "train_loss": 3.554145336151123, "perplexity": 34.95792991760823, "lr": 0.0026291804804649314, "grad_norm": 0.176507, "tokens_per_sec": 107561} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:38.442324+00:00", "epoch": 0, "step": 23148, "train_loss": 3.469252824783325, "perplexity": 32.11273963552254, "lr": 0.0026291804804649314, "grad_norm": 0.197107, "tokens_per_sec": 107842} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:38.746758+00:00", "epoch": 0, "step": 23149, "train_loss": 3.5288662910461426, "perplexity": 34.085302888542834, "lr": 0.0026291804804649314, "grad_norm": 0.185753, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:39.051668+00:00", "epoch": 0, "step": 23150, "train_loss": 3.499376058578491, "perplexity": 33.09479630115357, "lr": 0.0026291804804649314, "grad_norm": 0.165444, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:39.357148+00:00", "epoch": 0, "step": 23151, "train_loss": 3.573983907699585, "perplexity": 35.658370206380226, "lr": 0.0026291804804649314, "grad_norm": 0.175391, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:39.661992+00:00", "epoch": 0, "step": 23152, "train_loss": 3.410489797592163, "perplexity": 30.280071734098836, "lr": 0.0026291804804649314, "grad_norm": 0.148711, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:39.966730+00:00", "epoch": 0, "step": 23153, "train_loss": 3.5470316410064697, "perplexity": 34.71013228467069, "lr": 0.0026291804804649314, "grad_norm": 0.156988, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:40.270987+00:00", "epoch": 0, "step": 23154, "train_loss": 3.437039375305176, "perplexity": 31.094761836512284, "lr": 0.0026291804804649314, "grad_norm": 0.184361, "tokens_per_sec": 107692} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:40.575084+00:00", "epoch": 0, "step": 23155, "train_loss": 3.4490561485290527, "perplexity": 31.47067464386894, "lr": 0.0026291804804649314, "grad_norm": 0.195635, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:40.880491+00:00", "epoch": 0, "step": 23156, "train_loss": 3.4113659858703613, "perplexity": 30.306614404505517, "lr": 0.0026291804804649314, "grad_norm": 0.208406, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:41.186095+00:00", "epoch": 0, "step": 23157, "train_loss": 3.5088255405426025, "perplexity": 33.40900720900212, "lr": 0.0026291804804649314, "grad_norm": 0.161694, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:41.491014+00:00", "epoch": 0, "step": 23158, "train_loss": 3.441195011138916, "perplexity": 31.224249207875538, "lr": 0.0026291804804649314, "grad_norm": 0.19728, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:41.795266+00:00", "epoch": 0, "step": 23159, "train_loss": 3.455885410308838, "perplexity": 31.68633167037765, "lr": 0.0026291804804649314, "grad_norm": 0.165043, "tokens_per_sec": 107696} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:42.099193+00:00", "epoch": 0, "step": 23160, "train_loss": 3.519254446029663, "perplexity": 33.75924973628922, "lr": 0.0026291804804649314, "grad_norm": 0.197456, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:42.403459+00:00", "epoch": 0, "step": 23161, "train_loss": 3.467459201812744, "perplexity": 32.055193111874985, "lr": 0.0026291804804649314, "grad_norm": 0.171402, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:42.708711+00:00", "epoch": 0, "step": 23162, "train_loss": 3.408369779586792, "perplexity": 30.215945435290507, "lr": 0.0026291804804649314, "grad_norm": 0.180855, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:43.014152+00:00", "epoch": 0, "step": 23163, "train_loss": 3.451343536376953, "perplexity": 31.542742674961303, "lr": 0.0026291804804649314, "grad_norm": 0.182423, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:43.319168+00:00", "epoch": 0, "step": 23164, "train_loss": 3.5205273628234863, "perplexity": 33.80224981417991, "lr": 0.0026291804804649314, "grad_norm": 0.167204, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:43.623223+00:00", "epoch": 0, "step": 23165, "train_loss": 3.537323236465454, "perplexity": 34.374782767142364, "lr": 0.0026291804804649314, "grad_norm": 0.16217, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:43.927698+00:00", "epoch": 0, "step": 23166, "train_loss": 3.3744943141937256, "perplexity": 29.209509211688136, "lr": 0.0026291804804649314, "grad_norm": 0.175763, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:44.233507+00:00", "epoch": 0, "step": 23167, "train_loss": 3.5254251956939697, "perplexity": 33.96821368429302, "lr": 0.0026291804804649314, "grad_norm": 0.182685, "tokens_per_sec": 107151} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:44.538579+00:00", "epoch": 0, "step": 23168, "train_loss": 3.5107078552246094, "perplexity": 33.47195269680561, "lr": 0.0026291804804649314, "grad_norm": 0.170021, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:44.843764+00:00", "epoch": 0, "step": 23169, "train_loss": 3.59956431388855, "perplexity": 36.582292574304994, "lr": 0.0026291804804649314, "grad_norm": 0.153964, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:45.149777+00:00", "epoch": 0, "step": 23170, "train_loss": 3.4775149822235107, "perplexity": 32.37915923099458, "lr": 0.0026291804804649314, "grad_norm": 0.178786, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:45.454351+00:00", "epoch": 0, "step": 23171, "train_loss": 3.5348479747772217, "perplexity": 34.28980140326415, "lr": 0.0026291804804649314, "grad_norm": 0.182996, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:45.759487+00:00", "epoch": 0, "step": 23172, "train_loss": 3.4469752311706543, "perplexity": 31.405254860909793, "lr": 0.0026291804804649314, "grad_norm": 0.165983, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:46.064411+00:00", "epoch": 0, "step": 23173, "train_loss": 3.4665720462799072, "perplexity": 32.02676778065889, "lr": 0.0026291804804649314, "grad_norm": 0.153635, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:46.368882+00:00", "epoch": 0, "step": 23174, "train_loss": 3.5353479385375977, "perplexity": 34.30694934763445, "lr": 0.0026291804804649314, "grad_norm": 0.167569, "tokens_per_sec": 107622} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:46.672863+00:00", "epoch": 0, "step": 23175, "train_loss": 3.430866003036499, "perplexity": 30.903393597541054, "lr": 0.0026291804804649314, "grad_norm": 0.179715, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:46.976401+00:00", "epoch": 0, "step": 23176, "train_loss": 3.5078580379486084, "perplexity": 33.37669953926195, "lr": 0.0026291804804649314, "grad_norm": 0.184188, "tokens_per_sec": 107953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:47.281230+00:00", "epoch": 0, "step": 23177, "train_loss": 3.5360450744628906, "perplexity": 34.33087429301241, "lr": 0.0026291804804649314, "grad_norm": 0.195001, "tokens_per_sec": 107497} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:47.586593+00:00", "epoch": 0, "step": 23178, "train_loss": 3.4632909297943115, "perplexity": 31.921856432211435, "lr": 0.0026291804804649314, "grad_norm": 0.194645, "tokens_per_sec": 107309} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:47.891776+00:00", "epoch": 0, "step": 23179, "train_loss": 3.579958438873291, "perplexity": 35.87204993326836, "lr": 0.0026291804804649314, "grad_norm": 0.186235, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:48.196681+00:00", "epoch": 0, "step": 23180, "train_loss": 3.503033399581909, "perplexity": 33.21605686711902, "lr": 0.0026291804804649314, "grad_norm": 0.192234, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:48.501913+00:00", "epoch": 0, "step": 23181, "train_loss": 3.522226333618164, "perplexity": 33.85972766217532, "lr": 0.0026291804804649314, "grad_norm": 0.17044, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:48.807080+00:00", "epoch": 0, "step": 23182, "train_loss": 3.464332342147827, "perplexity": 31.955117564117835, "lr": 0.0026291804804649314, "grad_norm": 0.162838, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:49.112467+00:00", "epoch": 0, "step": 23183, "train_loss": 3.4964590072631836, "perplexity": 32.998397750189504, "lr": 0.0026291804804649314, "grad_norm": 0.153576, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:49.416999+00:00", "epoch": 0, "step": 23184, "train_loss": 3.4545679092407227, "perplexity": 31.644612383194712, "lr": 0.0026291804804649314, "grad_norm": 0.151207, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:49.722998+00:00", "epoch": 0, "step": 23185, "train_loss": 3.4423410892486572, "perplexity": 31.260055150670162, "lr": 0.0026291804804649314, "grad_norm": 0.151708, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:50.026792+00:00", "epoch": 0, "step": 23186, "train_loss": 3.495858669281006, "perplexity": 32.97859350388449, "lr": 0.0026291804804649314, "grad_norm": 0.159923, "tokens_per_sec": 107863} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:50.331280+00:00", "epoch": 0, "step": 23187, "train_loss": 3.477417469024658, "perplexity": 32.376001989540676, "lr": 0.0026291804804649314, "grad_norm": 0.172288, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:50.636813+00:00", "epoch": 0, "step": 23188, "train_loss": 3.4447758197784424, "perplexity": 31.336257689989232, "lr": 0.0026291804804649314, "grad_norm": 0.153979, "tokens_per_sec": 107249} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:50.942075+00:00", "epoch": 0, "step": 23189, "train_loss": 3.540828227996826, "perplexity": 34.49547748305168, "lr": 0.0026291804804649314, "grad_norm": 0.157578, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:51.246894+00:00", "epoch": 0, "step": 23190, "train_loss": 3.5241527557373047, "perplexity": 33.925018659310545, "lr": 0.0026291804804649314, "grad_norm": 0.187926, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:51.551353+00:00", "epoch": 0, "step": 23191, "train_loss": 3.4019997119903564, "perplexity": 30.02407956843832, "lr": 0.0026291804804649314, "grad_norm": 0.173493, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:51.855559+00:00", "epoch": 0, "step": 23192, "train_loss": 3.506838321685791, "perplexity": 33.34268212294658, "lr": 0.0026291804804649314, "grad_norm": 0.172614, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:52.160095+00:00", "epoch": 0, "step": 23193, "train_loss": 3.5112783908843994, "perplexity": 33.491055088201506, "lr": 0.0026291804804649314, "grad_norm": 0.158269, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:52.466384+00:00", "epoch": 0, "step": 23194, "train_loss": 3.3820314407348633, "perplexity": 29.430496738137535, "lr": 0.0026291804804649314, "grad_norm": 0.181833, "tokens_per_sec": 106984} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:52.770947+00:00", "epoch": 0, "step": 23195, "train_loss": 3.3941924571990967, "perplexity": 29.790586584992703, "lr": 0.0026291804804649314, "grad_norm": 0.168077, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:53.076145+00:00", "epoch": 0, "step": 23196, "train_loss": 3.511587381362915, "perplexity": 33.50140510428507, "lr": 0.0026291804804649314, "grad_norm": 0.151914, "tokens_per_sec": 107366} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:53.381092+00:00", "epoch": 0, "step": 23197, "train_loss": 3.4432621002197266, "perplexity": 31.288859266837544, "lr": 0.0026291804804649314, "grad_norm": 0.202944, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:53.684697+00:00", "epoch": 0, "step": 23198, "train_loss": 3.5067102909088135, "perplexity": 33.33841350671082, "lr": 0.0026291804804649314, "grad_norm": 0.170553, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:53.988495+00:00", "epoch": 0, "step": 23199, "train_loss": 3.427572727203369, "perplexity": 30.801787598367405, "lr": 0.0026291804804649314, "grad_norm": 0.176612, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:54.294214+00:00", "epoch": 0, "step": 23200, "train_loss": 3.4815385341644287, "perplexity": 32.50970090445433, "lr": 0.0026291804804649314, "grad_norm": 0.186754, "tokens_per_sec": 107240} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:54.599561+00:00", "epoch": 0, "step": 23201, "train_loss": 3.4890296459198, "perplexity": 32.75414915852226, "lr": 0.0026291804804649314, "grad_norm": 0.181177, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:54.903553+00:00", "epoch": 0, "step": 23202, "train_loss": 3.4788239002227783, "perplexity": 32.42156864448062, "lr": 0.0026291804804649314, "grad_norm": 0.173668, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:55.207622+00:00", "epoch": 0, "step": 23203, "train_loss": 3.5178589820861816, "perplexity": 33.712172775263575, "lr": 0.0026291804804649314, "grad_norm": 0.202465, "tokens_per_sec": 107767} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:55.513561+00:00", "epoch": 0, "step": 23204, "train_loss": 3.4461779594421387, "perplexity": 31.38022631768309, "lr": 0.0026291804804649314, "grad_norm": 0.185974, "tokens_per_sec": 107105} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:55.819078+00:00", "epoch": 0, "step": 23205, "train_loss": 3.5779833793640137, "perplexity": 35.80127041981435, "lr": 0.0026291804804649314, "grad_norm": 0.196689, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:56.123869+00:00", "epoch": 0, "step": 23206, "train_loss": 3.441141128540039, "perplexity": 31.22256680950664, "lr": 0.0026291804804649314, "grad_norm": 0.178573, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:56.429121+00:00", "epoch": 0, "step": 23207, "train_loss": 3.4806602001190186, "perplexity": 32.481159063813955, "lr": 0.0026291804804649314, "grad_norm": 0.195713, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:56.733535+00:00", "epoch": 0, "step": 23208, "train_loss": 3.5046403408050537, "perplexity": 33.26947602739931, "lr": 0.0026291804804649314, "grad_norm": 0.194019, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:57.037477+00:00", "epoch": 0, "step": 23209, "train_loss": 3.4723570346832275, "perplexity": 32.21257920112394, "lr": 0.0026291804804649314, "grad_norm": 0.183491, "tokens_per_sec": 107807} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:57.342769+00:00", "epoch": 0, "step": 23210, "train_loss": 3.5087525844573975, "perplexity": 33.40656990753463, "lr": 0.0026291804804649314, "grad_norm": 0.161394, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:57.649358+00:00", "epoch": 0, "step": 23211, "train_loss": 3.4191126823425293, "perplexity": 30.54230226991379, "lr": 0.0026291804804649314, "grad_norm": 0.190802, "tokens_per_sec": 106815} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:57.955819+00:00", "epoch": 0, "step": 23212, "train_loss": 3.4794023036956787, "perplexity": 32.4403268167545, "lr": 0.0026291804804649314, "grad_norm": 0.170236, "tokens_per_sec": 106926} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:58.260932+00:00", "epoch": 0, "step": 23213, "train_loss": 3.433652877807617, "perplexity": 30.989637605305614, "lr": 0.0026291804804649314, "grad_norm": 0.195812, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:58.566062+00:00", "epoch": 0, "step": 23214, "train_loss": 3.4552903175354004, "perplexity": 31.667480972897096, "lr": 0.0026291804804649314, "grad_norm": 0.160697, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:58.870999+00:00", "epoch": 0, "step": 23215, "train_loss": 3.598482370376587, "perplexity": 36.54273400412961, "lr": 0.0026291804804649314, "grad_norm": 0.18387, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:59.175740+00:00", "epoch": 0, "step": 23216, "train_loss": 3.4178502559661865, "perplexity": 30.503769189640046, "lr": 0.0026291804804649314, "grad_norm": 0.190158, "tokens_per_sec": 107528} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:59.480753+00:00", "epoch": 0, "step": 23217, "train_loss": 3.5010595321655273, "perplexity": 33.1505574396275, "lr": 0.0026291804804649314, "grad_norm": 0.18834, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T22:59:59.784564+00:00", "epoch": 0, "step": 23218, "train_loss": 3.546257495880127, "perplexity": 34.683272003147266, "lr": 0.0026291804804649314, "grad_norm": 0.166491, "tokens_per_sec": 107856} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:00.088362+00:00", "epoch": 0, "step": 23219, "train_loss": 3.4606540203094482, "perplexity": 31.837792269714818, "lr": 0.0026291804804649314, "grad_norm": 0.203513, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:00.392683+00:00", "epoch": 0, "step": 23220, "train_loss": 3.5104193687438965, "perplexity": 33.462297883678005, "lr": 0.0026291804804649314, "grad_norm": 0.148156, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:00.697059+00:00", "epoch": 0, "step": 23221, "train_loss": 3.5547523498535156, "perplexity": 34.979156301776754, "lr": 0.0026291804804649314, "grad_norm": 0.177834, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:01.001748+00:00", "epoch": 0, "step": 23222, "train_loss": 3.5244131088256836, "perplexity": 33.93385229257376, "lr": 0.0026291804804649314, "grad_norm": 0.192877, "tokens_per_sec": 107546} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:01.306282+00:00", "epoch": 0, "step": 23223, "train_loss": 3.403047561645508, "perplexity": 30.055556778672155, "lr": 0.0026291804804649314, "grad_norm": 0.16968, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:01.611520+00:00", "epoch": 0, "step": 23224, "train_loss": 3.523517608642578, "perplexity": 33.90347812369899, "lr": 0.0026291804804649314, "grad_norm": 0.188014, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:01.917160+00:00", "epoch": 0, "step": 23225, "train_loss": 3.489619016647339, "perplexity": 32.773459185064986, "lr": 0.0026291804804649314, "grad_norm": 0.211932, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:02.221402+00:00", "epoch": 0, "step": 23226, "train_loss": 3.439587116241455, "perplexity": 31.174084237704662, "lr": 0.0026291804804649314, "grad_norm": 0.197724, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:02.526929+00:00", "epoch": 0, "step": 23227, "train_loss": 3.5247890949249268, "perplexity": 33.94661334816897, "lr": 0.0026291804804649314, "grad_norm": 0.160824, "tokens_per_sec": 107246} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:02.831649+00:00", "epoch": 0, "step": 23228, "train_loss": 3.50980544090271, "perplexity": 33.44176075218059, "lr": 0.0026291804804649314, "grad_norm": 0.176565, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:03.135781+00:00", "epoch": 0, "step": 23229, "train_loss": 3.4671475887298584, "perplexity": 32.045205850488365, "lr": 0.0026291804804649314, "grad_norm": 0.197964, "tokens_per_sec": 107743} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:03.440425+00:00", "epoch": 0, "step": 23230, "train_loss": 3.532740354537964, "perplexity": 34.21760762908042, "lr": 0.0026291804804649314, "grad_norm": 0.174305, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:03.744662+00:00", "epoch": 0, "step": 23231, "train_loss": 3.456664562225342, "perplexity": 31.71102975695688, "lr": 0.0026291804804649314, "grad_norm": 0.184722, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:04.049147+00:00", "epoch": 0, "step": 23232, "train_loss": 3.425539970397949, "perplexity": 30.73923864994622, "lr": 0.0026291804804649314, "grad_norm": 0.188126, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:04.353820+00:00", "epoch": 0, "step": 23233, "train_loss": 3.5629637241363525, "perplexity": 35.26756574467083, "lr": 0.0026291804804649314, "grad_norm": 0.196922, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:04.657784+00:00", "epoch": 0, "step": 23234, "train_loss": 3.4798686504364014, "perplexity": 32.45545878553139, "lr": 0.0026291804804649314, "grad_norm": 0.169192, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:04.961943+00:00", "epoch": 0, "step": 23235, "train_loss": 3.4738168716430664, "perplexity": 32.2596386560218, "lr": 0.0026291804804649314, "grad_norm": 0.155449, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:05.267348+00:00", "epoch": 0, "step": 23236, "train_loss": 3.437366008758545, "perplexity": 31.10492008487419, "lr": 0.0026291804804649314, "grad_norm": 0.190688, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:05.572368+00:00", "epoch": 0, "step": 23237, "train_loss": 3.4811878204345703, "perplexity": 32.4983013051084, "lr": 0.0026291804804649314, "grad_norm": 0.17478, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:05.877613+00:00", "epoch": 0, "step": 23238, "train_loss": 3.502276659011841, "perplexity": 33.19093043760334, "lr": 0.0026291804804649314, "grad_norm": 0.180361, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:06.182865+00:00", "epoch": 0, "step": 23239, "train_loss": 3.5539968013763428, "perplexity": 34.95273783497309, "lr": 0.0026291804804649314, "grad_norm": 0.170645, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:06.487377+00:00", "epoch": 0, "step": 23240, "train_loss": 3.435439348220825, "perplexity": 31.04504915681514, "lr": 0.0026291804804649314, "grad_norm": 0.162403, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:06.790742+00:00", "epoch": 0, "step": 23241, "train_loss": 3.512904167175293, "perplexity": 33.5455483364336, "lr": 0.0026291804804649314, "grad_norm": 0.167598, "tokens_per_sec": 107933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:07.095380+00:00", "epoch": 0, "step": 23242, "train_loss": 3.5533316135406494, "perplexity": 34.92949543008061, "lr": 0.0026291804804649314, "grad_norm": 0.184641, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:07.402028+00:00", "epoch": 0, "step": 23243, "train_loss": 3.4414775371551514, "perplexity": 31.23307211690616, "lr": 0.0026291804804649314, "grad_norm": 0.177555, "tokens_per_sec": 106859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:07.707058+00:00", "epoch": 0, "step": 23244, "train_loss": 3.5333077907562256, "perplexity": 34.23702944875225, "lr": 0.0026291804804649314, "grad_norm": 0.149045, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:08.011880+00:00", "epoch": 0, "step": 23245, "train_loss": 3.4843790531158447, "perplexity": 32.60217660328713, "lr": 0.0026291804804649314, "grad_norm": 0.171223, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:08.316911+00:00", "epoch": 0, "step": 23246, "train_loss": 3.5469987392425537, "perplexity": 34.708990278879874, "lr": 0.0026291804804649314, "grad_norm": 0.174223, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:08.621370+00:00", "epoch": 0, "step": 23247, "train_loss": 3.474071502685547, "perplexity": 32.26785400734017, "lr": 0.0026291804804649314, "grad_norm": 0.178162, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:08.927852+00:00", "epoch": 0, "step": 23248, "train_loss": 3.5036768913269043, "perplexity": 33.23743800406907, "lr": 0.0026291804804649314, "grad_norm": 0.186852, "tokens_per_sec": 106917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:09.232651+00:00", "epoch": 0, "step": 23249, "train_loss": 3.4858107566833496, "perplexity": 32.648886685355286, "lr": 0.0026291804804649314, "grad_norm": 0.191257, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:09.537611+00:00", "epoch": 0, "step": 23250, "train_loss": 3.494117259979248, "perplexity": 32.92121424928203, "lr": 0.0026291804804649314, "grad_norm": 0.14404, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:09.842127+00:00", "epoch": 0, "step": 23251, "train_loss": 3.4504330158233643, "perplexity": 31.514035430690466, "lr": 0.0026291804804649314, "grad_norm": 0.182127, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:10.146415+00:00", "epoch": 0, "step": 23252, "train_loss": 3.483851909637451, "perplexity": 32.584995107463875, "lr": 0.0026291804804649314, "grad_norm": 0.173304, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:10.450706+00:00", "epoch": 0, "step": 23253, "train_loss": 3.475550651550293, "perplexity": 32.315618283491794, "lr": 0.0026291804804649314, "grad_norm": 0.190718, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:10.756942+00:00", "epoch": 0, "step": 23254, "train_loss": 3.462615966796875, "perplexity": 31.900317630075147, "lr": 0.0026291804804649314, "grad_norm": 0.215596, "tokens_per_sec": 107002} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:11.061692+00:00", "epoch": 0, "step": 23255, "train_loss": 3.5009329319000244, "perplexity": 33.14636083590525, "lr": 0.0026291804804649314, "grad_norm": 0.202041, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:11.365902+00:00", "epoch": 0, "step": 23256, "train_loss": 3.462264060974121, "perplexity": 31.88909369755286, "lr": 0.0026291804804649314, "grad_norm": 0.193762, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:11.671135+00:00", "epoch": 0, "step": 23257, "train_loss": 3.5239667892456055, "perplexity": 33.918710329196834, "lr": 0.0026291804804649314, "grad_norm": 0.188435, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:11.975832+00:00", "epoch": 0, "step": 23258, "train_loss": 3.5580854415893555, "perplexity": 35.09593955510419, "lr": 0.0026291804804649314, "grad_norm": 0.177078, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:12.281385+00:00", "epoch": 0, "step": 23259, "train_loss": 3.403433322906494, "perplexity": 30.067153284752337, "lr": 0.0026291804804649314, "grad_norm": 0.157192, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:12.587676+00:00", "epoch": 0, "step": 23260, "train_loss": 3.524596691131592, "perplexity": 33.940082519288175, "lr": 0.0026291804804649314, "grad_norm": 0.162188, "tokens_per_sec": 107041} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:12.891793+00:00", "epoch": 0, "step": 23261, "train_loss": 3.400198221206665, "perplexity": 29.970040156173354, "lr": 0.0026291804804649314, "grad_norm": 0.160031, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:13.195442+00:00", "epoch": 0, "step": 23262, "train_loss": 3.47114896774292, "perplexity": 32.173687745599786, "lr": 0.0026291804804649314, "grad_norm": 0.196418, "tokens_per_sec": 107914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:13.499792+00:00", "epoch": 0, "step": 23263, "train_loss": 3.4893317222595215, "perplexity": 32.76404490657119, "lr": 0.0026291804804649314, "grad_norm": 0.159655, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:13.805040+00:00", "epoch": 0, "step": 23264, "train_loss": 3.486941337585449, "perplexity": 32.68581976710097, "lr": 0.0026291804804649314, "grad_norm": 0.185696, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:14.111032+00:00", "epoch": 0, "step": 23265, "train_loss": 3.489219903945923, "perplexity": 32.76038149114535, "lr": 0.0026291804804649314, "grad_norm": 0.16352, "tokens_per_sec": 107089} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:14.416895+00:00", "epoch": 0, "step": 23266, "train_loss": 3.460742712020874, "perplexity": 31.840616143224473, "lr": 0.0026291804804649314, "grad_norm": 0.188384, "tokens_per_sec": 107134} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:14.721798+00:00", "epoch": 0, "step": 23267, "train_loss": 3.5366716384887695, "perplexity": 34.3523915240789, "lr": 0.0026291804804649314, "grad_norm": 0.191492, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:15.026218+00:00", "epoch": 0, "step": 23268, "train_loss": 3.44527006149292, "perplexity": 31.351749203676036, "lr": 0.0026291804804649314, "grad_norm": 0.161861, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:15.331419+00:00", "epoch": 0, "step": 23269, "train_loss": 3.5346875190734863, "perplexity": 34.284299850438614, "lr": 0.0026291804804649314, "grad_norm": 0.171531, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:15.637358+00:00", "epoch": 0, "step": 23270, "train_loss": 3.4480814933776855, "perplexity": 31.440016531680758, "lr": 0.0026291804804649314, "grad_norm": 0.161373, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:15.942083+00:00", "epoch": 0, "step": 23271, "train_loss": 3.520258903503418, "perplexity": 33.79317650313996, "lr": 0.0026291804804649314, "grad_norm": 0.158538, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:16.246400+00:00", "epoch": 0, "step": 23272, "train_loss": 3.4597649574279785, "perplexity": 31.809499049469487, "lr": 0.0026291804804649314, "grad_norm": 0.173332, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:16.550111+00:00", "epoch": 0, "step": 23273, "train_loss": 3.512399435043335, "perplexity": 33.52862109252536, "lr": 0.0026291804804649314, "grad_norm": 0.175156, "tokens_per_sec": 107893} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:16.854847+00:00", "epoch": 0, "step": 23274, "train_loss": 3.6087048053741455, "perplexity": 36.91820557431084, "lr": 0.0026291804804649314, "grad_norm": 0.174226, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:17.160337+00:00", "epoch": 0, "step": 23275, "train_loss": 3.4310996532440186, "perplexity": 30.910615025479395, "lr": 0.0026291804804649314, "grad_norm": 0.144004, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:17.464823+00:00", "epoch": 0, "step": 23276, "train_loss": 3.607485055923462, "perplexity": 36.87320206542362, "lr": 0.0026291804804649314, "grad_norm": 0.170614, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:17.770109+00:00", "epoch": 0, "step": 23277, "train_loss": 3.4813125133514404, "perplexity": 32.50235386574901, "lr": 0.0026291804804649314, "grad_norm": 0.184331, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:18.074381+00:00", "epoch": 0, "step": 23278, "train_loss": 3.4133141040802, "perplexity": 30.365712818563118, "lr": 0.0026291804804649314, "grad_norm": 0.147101, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:18.378157+00:00", "epoch": 0, "step": 23279, "train_loss": 3.542599678039551, "perplexity": 34.55663865409854, "lr": 0.0026291804804649314, "grad_norm": 0.172306, "tokens_per_sec": 107870} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:18.683470+00:00", "epoch": 0, "step": 23280, "train_loss": 3.5576560497283936, "perplexity": 35.08087287929189, "lr": 0.0026291804804649314, "grad_norm": 0.165671, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:18.988792+00:00", "epoch": 0, "step": 23281, "train_loss": 3.5288941860198975, "perplexity": 34.08625371043385, "lr": 0.0026291804804649314, "grad_norm": 0.166587, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:19.294389+00:00", "epoch": 0, "step": 23282, "train_loss": 3.5018465518951416, "perplexity": 33.176657851802645, "lr": 0.0026291804804649314, "grad_norm": 0.172825, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:19.598923+00:00", "epoch": 0, "step": 23283, "train_loss": 3.457393169403076, "perplexity": 31.734143060087916, "lr": 0.0026291804804649314, "grad_norm": 0.153594, "tokens_per_sec": 107596} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:19.902729+00:00", "epoch": 0, "step": 23284, "train_loss": 3.41654896736145, "perplexity": 30.46410079800278, "lr": 0.0026291804804649314, "grad_norm": 0.177702, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:20.207377+00:00", "epoch": 0, "step": 23285, "train_loss": 3.444103240966797, "perplexity": 31.315188673112086, "lr": 0.0026291804804649314, "grad_norm": 0.175587, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:20.512843+00:00", "epoch": 0, "step": 23286, "train_loss": 3.460787057876587, "perplexity": 31.842028173902392, "lr": 0.0026291804804649314, "grad_norm": 0.162874, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:20.818030+00:00", "epoch": 0, "step": 23287, "train_loss": 3.457709550857544, "perplexity": 31.74418474284289, "lr": 0.0026291804804649314, "grad_norm": 0.160221, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:21.122201+00:00", "epoch": 0, "step": 23288, "train_loss": 3.4390132427215576, "perplexity": 31.156199388565092, "lr": 0.0026291804804649314, "grad_norm": 0.162745, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:21.425579+00:00", "epoch": 0, "step": 23289, "train_loss": 3.3963897228240967, "perplexity": 29.856116383666823, "lr": 0.0026291804804649314, "grad_norm": 0.157594, "tokens_per_sec": 108006} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:21.730988+00:00", "epoch": 0, "step": 23290, "train_loss": 3.4984562397003174, "perplexity": 33.06436907865857, "lr": 0.0026291804804649314, "grad_norm": 0.173573, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:22.035715+00:00", "epoch": 0, "step": 23291, "train_loss": 3.505824565887451, "perplexity": 33.308897912974956, "lr": 0.0026291804804649314, "grad_norm": 0.192315, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:22.341196+00:00", "epoch": 0, "step": 23292, "train_loss": 3.4288439750671387, "perplexity": 30.840969204542, "lr": 0.0026291804804649314, "grad_norm": 0.216624, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:22.645596+00:00", "epoch": 0, "step": 23293, "train_loss": 3.5584304332733154, "perplexity": 35.10804945117807, "lr": 0.0026291804804649314, "grad_norm": 0.183485, "tokens_per_sec": 107648} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:22.951114+00:00", "epoch": 0, "step": 23294, "train_loss": 3.607805013656616, "perplexity": 36.88500181918118, "lr": 0.0026291804804649314, "grad_norm": 0.165551, "tokens_per_sec": 107260} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:23.255229+00:00", "epoch": 0, "step": 23295, "train_loss": 3.3675637245178223, "perplexity": 29.00776998213049, "lr": 0.0026291804804649314, "grad_norm": 0.195448, "tokens_per_sec": 107742} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:23.560452+00:00", "epoch": 0, "step": 23296, "train_loss": 3.50527286529541, "perplexity": 33.29052644251768, "lr": 0.0026291804804649314, "grad_norm": 0.179258, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:23.865579+00:00", "epoch": 0, "step": 23297, "train_loss": 3.429797410964966, "perplexity": 30.870388113998985, "lr": 0.0026291804804649314, "grad_norm": 0.167042, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:24.170668+00:00", "epoch": 0, "step": 23298, "train_loss": 3.4891533851623535, "perplexity": 32.75820238289592, "lr": 0.0026291804804649314, "grad_norm": 0.182494, "tokens_per_sec": 107409} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:24.475222+00:00", "epoch": 0, "step": 23299, "train_loss": 3.4687163829803467, "perplexity": 32.095517639286236, "lr": 0.0026291804804649314, "grad_norm": 0.179193, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:24.779219+00:00", "epoch": 0, "step": 23300, "train_loss": 3.5591070652008057, "perplexity": 35.13181271693417, "lr": 0.0026291804804649314, "grad_norm": 0.203002, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:25.084209+00:00", "epoch": 0, "step": 23301, "train_loss": 3.4882454872131348, "perplexity": 32.728474774991504, "lr": 0.0026291804804649314, "grad_norm": 0.205823, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:25.389812+00:00", "epoch": 0, "step": 23302, "train_loss": 3.3792724609375, "perplexity": 29.34941050129079, "lr": 0.0026291804804649314, "grad_norm": 0.171396, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:25.694732+00:00", "epoch": 0, "step": 23303, "train_loss": 3.576540231704712, "perplexity": 35.74964116348876, "lr": 0.0026291804804649314, "grad_norm": 0.183605, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:25.998632+00:00", "epoch": 0, "step": 23304, "train_loss": 3.5123090744018555, "perplexity": 33.52559156169276, "lr": 0.0026291804804649314, "grad_norm": 0.182271, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:26.303571+00:00", "epoch": 0, "step": 23305, "train_loss": 3.54473876953125, "perplexity": 34.63063758284821, "lr": 0.0026291804804649314, "grad_norm": 0.201421, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:26.607449+00:00", "epoch": 0, "step": 23306, "train_loss": 3.376962423324585, "perplexity": 29.281690507097156, "lr": 0.0026291804804649314, "grad_norm": 0.178862, "tokens_per_sec": 107833} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:26.911820+00:00", "epoch": 0, "step": 23307, "train_loss": 3.51763916015625, "perplexity": 33.70476291483667, "lr": 0.0026291804804649314, "grad_norm": 0.176927, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:27.217220+00:00", "epoch": 0, "step": 23308, "train_loss": 3.510939359664917, "perplexity": 33.47970249950294, "lr": 0.0026291804804649314, "grad_norm": 0.162234, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:27.521494+00:00", "epoch": 0, "step": 23309, "train_loss": 3.5101284980773926, "perplexity": 33.4525660982039, "lr": 0.0026291804804649314, "grad_norm": 0.164322, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:27.827524+00:00", "epoch": 0, "step": 23310, "train_loss": 3.4531993865966797, "perplexity": 31.601335633946256, "lr": 0.0026291804804649314, "grad_norm": 0.149682, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:28.132865+00:00", "epoch": 0, "step": 23311, "train_loss": 3.436424493789673, "perplexity": 31.075648119167997, "lr": 0.0026291804804649314, "grad_norm": 0.173105, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:28.437868+00:00", "epoch": 0, "step": 23312, "train_loss": 3.5324747562408447, "perplexity": 34.20852069755283, "lr": 0.0026291804804649314, "grad_norm": 0.198138, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:28.742296+00:00", "epoch": 0, "step": 23313, "train_loss": 3.4740195274353027, "perplexity": 32.26617692113716, "lr": 0.0026291804804649314, "grad_norm": 0.184676, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:29.046583+00:00", "epoch": 0, "step": 23314, "train_loss": 3.4571423530578613, "perplexity": 31.726184616402662, "lr": 0.0026291804804649314, "grad_norm": 0.184017, "tokens_per_sec": 107687} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:29.351264+00:00", "epoch": 0, "step": 23315, "train_loss": 3.5139660835266113, "perplexity": 33.58118982351565, "lr": 0.0026291804804649314, "grad_norm": 0.197144, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:29.657009+00:00", "epoch": 0, "step": 23316, "train_loss": 3.4701311588287354, "perplexity": 32.14095773868204, "lr": 0.0026291804804649314, "grad_norm": 0.185035, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:29.964400+00:00", "epoch": 0, "step": 23317, "train_loss": 3.4678919315338135, "perplexity": 32.069067348329845, "lr": 0.0026291804804649314, "grad_norm": 0.207326, "tokens_per_sec": 106600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:30.269951+00:00", "epoch": 0, "step": 23318, "train_loss": 3.3851418495178223, "perplexity": 29.522180126169385, "lr": 0.0026291804804649314, "grad_norm": 0.182283, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:30.574864+00:00", "epoch": 0, "step": 23319, "train_loss": 3.435494899749756, "perplexity": 31.046773804664497, "lr": 0.0026291804804649314, "grad_norm": 0.207134, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:30.879434+00:00", "epoch": 0, "step": 23320, "train_loss": 3.467024564743042, "perplexity": 32.04126376399686, "lr": 0.0026291804804649314, "grad_norm": 0.206825, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:31.183675+00:00", "epoch": 0, "step": 23321, "train_loss": 3.6053342819213867, "perplexity": 36.79398136449847, "lr": 0.0026291804804649314, "grad_norm": 0.16692, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:31.489011+00:00", "epoch": 0, "step": 23322, "train_loss": 3.4808907508850098, "perplexity": 32.488648483228914, "lr": 0.0026291804804649314, "grad_norm": 0.197414, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:31.793398+00:00", "epoch": 0, "step": 23323, "train_loss": 3.537482738494873, "perplexity": 34.38026605204047, "lr": 0.0026291804804649314, "grad_norm": 0.184109, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:32.097178+00:00", "epoch": 0, "step": 23324, "train_loss": 3.4753873348236084, "perplexity": 32.31034103343628, "lr": 0.0026291804804649314, "grad_norm": 0.174371, "tokens_per_sec": 107867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:32.402486+00:00", "epoch": 0, "step": 23325, "train_loss": 3.5117366313934326, "perplexity": 33.5064055631693, "lr": 0.0026291804804649314, "grad_norm": 0.192958, "tokens_per_sec": 107328} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:32.707567+00:00", "epoch": 0, "step": 23326, "train_loss": 3.4803011417388916, "perplexity": 32.469498524988225, "lr": 0.0026291804804649314, "grad_norm": 0.177334, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:33.011535+00:00", "epoch": 0, "step": 23327, "train_loss": 3.519212484359741, "perplexity": 33.75783317151589, "lr": 0.0026291804804649314, "grad_norm": 0.173437, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:33.315843+00:00", "epoch": 0, "step": 23328, "train_loss": 3.6570775508880615, "perplexity": 38.747938364532395, "lr": 0.0026291804804649314, "grad_norm": 0.194434, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:33.620572+00:00", "epoch": 0, "step": 23329, "train_loss": 3.5330395698547363, "perplexity": 34.22784759328702, "lr": 0.0026291804804649314, "grad_norm": 0.176959, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:33.925480+00:00", "epoch": 0, "step": 23330, "train_loss": 3.456294536590576, "perplexity": 31.69929803369502, "lr": 0.0026291804804649314, "grad_norm": 0.179607, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:34.230681+00:00", "epoch": 0, "step": 23331, "train_loss": 3.4557156562805176, "perplexity": 31.68095324445156, "lr": 0.0026291804804649314, "grad_norm": 0.184773, "tokens_per_sec": 107308} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:34.535822+00:00", "epoch": 0, "step": 23332, "train_loss": 3.458714246749878, "perplexity": 31.776094021736093, "lr": 0.0026291804804649314, "grad_norm": 0.170212, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:34.842143+00:00", "epoch": 0, "step": 23333, "train_loss": 3.4898152351379395, "perplexity": 32.779890574716184, "lr": 0.0026291804804649314, "grad_norm": 0.150682, "tokens_per_sec": 106974} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:35.146933+00:00", "epoch": 0, "step": 23334, "train_loss": 3.44779896736145, "perplexity": 31.431135163727518, "lr": 0.0026291804804649314, "grad_norm": 0.167539, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:35.451588+00:00", "epoch": 0, "step": 23335, "train_loss": 3.4698917865753174, "perplexity": 32.13326500595124, "lr": 0.0026291804804649314, "grad_norm": 0.152628, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:35.756144+00:00", "epoch": 0, "step": 23336, "train_loss": 3.473215341567993, "perplexity": 32.24023934837765, "lr": 0.0026291804804649314, "grad_norm": 0.159114, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:36.060966+00:00", "epoch": 0, "step": 23337, "train_loss": 3.575698137283325, "perplexity": 35.71954926198784, "lr": 0.0026291804804649314, "grad_norm": 0.139048, "tokens_per_sec": 107500} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:36.366938+00:00", "epoch": 0, "step": 23338, "train_loss": 3.443452835083008, "perplexity": 31.294827712309253, "lr": 0.0026291804804649314, "grad_norm": 0.168339, "tokens_per_sec": 107095} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:36.672442+00:00", "epoch": 0, "step": 23339, "train_loss": 3.584657907485962, "perplexity": 36.04102624445048, "lr": 0.0026291804804649314, "grad_norm": 0.168498, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:36.976812+00:00", "epoch": 0, "step": 23340, "train_loss": 3.475850820541382, "perplexity": 32.32531988601564, "lr": 0.0026291804804649314, "grad_norm": 0.161435, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:37.281793+00:00", "epoch": 0, "step": 23341, "train_loss": 3.4071624279022217, "perplexity": 30.179486176670864, "lr": 0.0026291804804649314, "grad_norm": 0.157778, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:37.587374+00:00", "epoch": 0, "step": 23342, "train_loss": 3.5359630584716797, "perplexity": 34.32805872779039, "lr": 0.0026291804804649314, "grad_norm": 0.15362, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:37.891990+00:00", "epoch": 0, "step": 23343, "train_loss": 3.5253472328186035, "perplexity": 33.96556552791342, "lr": 0.0026291804804649314, "grad_norm": 0.171581, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:38.196335+00:00", "epoch": 0, "step": 23344, "train_loss": 3.559788942337036, "perplexity": 35.15577646601797, "lr": 0.0026291804804649314, "grad_norm": 0.193907, "tokens_per_sec": 107668} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:38.501480+00:00", "epoch": 0, "step": 23345, "train_loss": 3.5189175605773926, "perplexity": 33.74787865165754, "lr": 0.0026291804804649314, "grad_norm": 0.165592, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:38.806954+00:00", "epoch": 0, "step": 23346, "train_loss": 3.503844976425171, "perplexity": 33.243025191650446, "lr": 0.0026291804804649314, "grad_norm": 0.189476, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:39.111735+00:00", "epoch": 0, "step": 23347, "train_loss": 3.5567193031311035, "perplexity": 35.04802637781838, "lr": 0.0026291804804649314, "grad_norm": 0.182444, "tokens_per_sec": 107513} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:39.416540+00:00", "epoch": 0, "step": 23348, "train_loss": 3.4936740398406982, "perplexity": 32.906626137252566, "lr": 0.0026291804804649314, "grad_norm": 0.165275, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:39.721777+00:00", "epoch": 0, "step": 23349, "train_loss": 3.5288970470428467, "perplexity": 34.08635123212747, "lr": 0.0026291804804649314, "grad_norm": 0.179904, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:40.026827+00:00", "epoch": 0, "step": 23350, "train_loss": 3.568877935409546, "perplexity": 35.476763589928765, "lr": 0.0026291804804649314, "grad_norm": 0.168973, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:40.332515+00:00", "epoch": 0, "step": 23351, "train_loss": 3.5695178508758545, "perplexity": 35.49947298491613, "lr": 0.0026291804804649314, "grad_norm": 0.17509, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:40.638222+00:00", "epoch": 0, "step": 23352, "train_loss": 3.5802512168884277, "perplexity": 35.88255401845428, "lr": 0.0026291804804649314, "grad_norm": 0.178267, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:40.943072+00:00", "epoch": 0, "step": 23353, "train_loss": 3.6142749786376953, "perplexity": 37.124420168266184, "lr": 0.0026291804804649314, "grad_norm": 0.164862, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:41.247604+00:00", "epoch": 0, "step": 23354, "train_loss": 3.4054079055786133, "perplexity": 30.12658201879332, "lr": 0.0026291804804649314, "grad_norm": 0.160282, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:41.551508+00:00", "epoch": 0, "step": 23355, "train_loss": 3.3275086879730225, "perplexity": 27.86882520613537, "lr": 0.0026291804804649314, "grad_norm": 0.168001, "tokens_per_sec": 107823} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:41.856872+00:00", "epoch": 0, "step": 23356, "train_loss": 3.6363792419433594, "perplexity": 37.954164789474156, "lr": 0.0026291804804649314, "grad_norm": 0.156701, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:42.161227+00:00", "epoch": 0, "step": 23357, "train_loss": 3.555783987045288, "perplexity": 35.015260720482964, "lr": 0.0026291804804649314, "grad_norm": 0.181053, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:42.465278+00:00", "epoch": 0, "step": 23358, "train_loss": 3.582695484161377, "perplexity": 35.97036784742566, "lr": 0.0026291804804649314, "grad_norm": 0.212837, "tokens_per_sec": 107772} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:42.769629+00:00", "epoch": 0, "step": 23359, "train_loss": 3.499396562576294, "perplexity": 33.09547488374102, "lr": 0.0026291804804649314, "grad_norm": 0.233226, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:43.074640+00:00", "epoch": 0, "step": 23360, "train_loss": 3.5792534351348877, "perplexity": 35.84676891661396, "lr": 0.0026291804804649314, "grad_norm": 0.217378, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:43.380192+00:00", "epoch": 0, "step": 23361, "train_loss": 3.4600746631622314, "perplexity": 31.81935215943239, "lr": 0.0026291804804649314, "grad_norm": 0.167059, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:43.684874+00:00", "epoch": 0, "step": 23362, "train_loss": 3.527420997619629, "perplexity": 34.03607520712867, "lr": 0.0026291804804649314, "grad_norm": 0.186896, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:43.988828+00:00", "epoch": 0, "step": 23363, "train_loss": 3.495558977127075, "perplexity": 32.96871155900834, "lr": 0.0026291804804649314, "grad_norm": 0.169113, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:44.292498+00:00", "epoch": 0, "step": 23364, "train_loss": 3.5279736518859863, "perplexity": 34.05489058802332, "lr": 0.0026291804804649314, "grad_norm": 0.174097, "tokens_per_sec": 107907} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:44.597117+00:00", "epoch": 0, "step": 23365, "train_loss": 3.4307427406311035, "perplexity": 30.89958460566891, "lr": 0.0026291804804649314, "grad_norm": 0.192587, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:44.901889+00:00", "epoch": 0, "step": 23366, "train_loss": 3.4030375480651855, "perplexity": 30.055255816447076, "lr": 0.0026291804804649314, "grad_norm": 0.226148, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:45.208003+00:00", "epoch": 0, "step": 23367, "train_loss": 3.451998472213745, "perplexity": 31.563407913993974, "lr": 0.0026291804804649314, "grad_norm": 0.212044, "tokens_per_sec": 107045} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:45.513472+00:00", "epoch": 0, "step": 23368, "train_loss": 3.611290454864502, "perplexity": 37.0137866300417, "lr": 0.0026291804804649314, "grad_norm": 0.168878, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:45.817803+00:00", "epoch": 0, "step": 23369, "train_loss": 3.519939422607422, "perplexity": 33.78238195325024, "lr": 0.0026291804804649314, "grad_norm": 0.162206, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:46.122400+00:00", "epoch": 0, "step": 23370, "train_loss": 3.486452341079712, "perplexity": 32.669840422687905, "lr": 0.0026291804804649314, "grad_norm": 0.173811, "tokens_per_sec": 107634} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:46.427495+00:00", "epoch": 0, "step": 23371, "train_loss": 3.4609923362731934, "perplexity": 31.848565325335656, "lr": 0.0026291804804649314, "grad_norm": 0.154542, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:46.732481+00:00", "epoch": 0, "step": 23372, "train_loss": 3.514193534851074, "perplexity": 33.58882877833031, "lr": 0.0026291804804649314, "grad_norm": 0.166145, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:47.037243+00:00", "epoch": 0, "step": 23373, "train_loss": 3.5297584533691406, "perplexity": 34.115726080742924, "lr": 0.0026291804804649314, "grad_norm": 0.152708, "tokens_per_sec": 107521} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:47.342024+00:00", "epoch": 0, "step": 23374, "train_loss": 3.485137939453125, "perplexity": 32.62692733998708, "lr": 0.0026291804804649314, "grad_norm": 0.160842, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:47.646710+00:00", "epoch": 0, "step": 23375, "train_loss": 3.452702760696411, "perplexity": 31.585645488576123, "lr": 0.0026291804804649314, "grad_norm": 0.177881, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:47.951212+00:00", "epoch": 0, "step": 23376, "train_loss": 3.532118797302246, "perplexity": 34.19634603580269, "lr": 0.0026291804804649314, "grad_norm": 0.172522, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:48.256280+00:00", "epoch": 0, "step": 23377, "train_loss": 3.497445821762085, "perplexity": 33.03097711978322, "lr": 0.0026291804804649314, "grad_norm": 0.163611, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:48.561678+00:00", "epoch": 0, "step": 23378, "train_loss": 3.4925100803375244, "perplexity": 32.86834643937357, "lr": 0.0026291804804649314, "grad_norm": 0.169147, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:48.866696+00:00", "epoch": 0, "step": 23379, "train_loss": 3.411147356033325, "perplexity": 30.299989198597377, "lr": 0.0026291804804649314, "grad_norm": 0.160514, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:49.171738+00:00", "epoch": 0, "step": 23380, "train_loss": 3.4198813438415527, "perplexity": 30.56578798688706, "lr": 0.0026291804804649314, "grad_norm": 0.165517, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:49.476088+00:00", "epoch": 0, "step": 23381, "train_loss": 3.4427857398986816, "perplexity": 31.273958045249195, "lr": 0.0026291804804649314, "grad_norm": 0.162843, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:49.780139+00:00", "epoch": 0, "step": 23382, "train_loss": 3.5294570922851562, "perplexity": 34.10544647756429, "lr": 0.0026291804804649314, "grad_norm": 0.174281, "tokens_per_sec": 107771} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:50.085429+00:00", "epoch": 0, "step": 23383, "train_loss": 3.50286602973938, "perplexity": 33.2104979661209, "lr": 0.0026291804804649314, "grad_norm": 0.169348, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:50.391627+00:00", "epoch": 0, "step": 23384, "train_loss": 3.4706170558929443, "perplexity": 32.15657873047782, "lr": 0.0026291804804649314, "grad_norm": 0.171061, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:50.695808+00:00", "epoch": 0, "step": 23385, "train_loss": 3.5065927505493164, "perplexity": 33.33449512789052, "lr": 0.0026291804804649314, "grad_norm": 0.156067, "tokens_per_sec": 107726} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:51.000027+00:00", "epoch": 0, "step": 23386, "train_loss": 3.446645498275757, "perplexity": 31.394901222370923, "lr": 0.0026291804804649314, "grad_norm": 0.162154, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:51.304294+00:00", "epoch": 0, "step": 23387, "train_loss": 3.434251070022583, "perplexity": 31.008180910943306, "lr": 0.0026291804804649314, "grad_norm": 0.163372, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:51.609365+00:00", "epoch": 0, "step": 23388, "train_loss": 3.4144811630249023, "perplexity": 30.401172082821336, "lr": 0.0026291804804649314, "grad_norm": 0.152456, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:51.913969+00:00", "epoch": 0, "step": 23389, "train_loss": 3.4461188316345215, "perplexity": 31.37837092855148, "lr": 0.0026291804804649314, "grad_norm": 0.153842, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:52.218346+00:00", "epoch": 0, "step": 23390, "train_loss": 3.5397040843963623, "perplexity": 34.45672140057654, "lr": 0.0026291804804649314, "grad_norm": 0.169461, "tokens_per_sec": 107712} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:52.522501+00:00", "epoch": 0, "step": 23391, "train_loss": 3.4780755043029785, "perplexity": 32.397313552131756, "lr": 0.0026291804804649314, "grad_norm": 0.200773, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:52.827798+00:00", "epoch": 0, "step": 23392, "train_loss": 3.4743192195892334, "perplexity": 32.27584829034211, "lr": 0.0026291804804649314, "grad_norm": 0.183403, "tokens_per_sec": 107333} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:53.132280+00:00", "epoch": 0, "step": 23393, "train_loss": 3.411046028137207, "perplexity": 30.296919119984413, "lr": 0.0026291804804649314, "grad_norm": 0.174149, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:53.437371+00:00", "epoch": 0, "step": 23394, "train_loss": 3.5468881130218506, "perplexity": 34.70515076684017, "lr": 0.0026291804804649314, "grad_norm": 0.160214, "tokens_per_sec": 107406} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:53.741796+00:00", "epoch": 0, "step": 23395, "train_loss": 3.5857601165771484, "perplexity": 36.08077289176691, "lr": 0.0026291804804649314, "grad_norm": 0.151449, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:54.045747+00:00", "epoch": 0, "step": 23396, "train_loss": 3.4553143978118896, "perplexity": 31.668243543776065, "lr": 0.0026291804804649314, "grad_norm": 0.179741, "tokens_per_sec": 107806} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:54.350475+00:00", "epoch": 0, "step": 23397, "train_loss": 3.526411533355713, "perplexity": 34.001734341377656, "lr": 0.0026291804804649314, "grad_norm": 0.15659, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:54.655898+00:00", "epoch": 0, "step": 23398, "train_loss": 3.4177606105804443, "perplexity": 30.501034790049435, "lr": 0.0026291804804649314, "grad_norm": 0.174703, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:54.961375+00:00", "epoch": 0, "step": 23399, "train_loss": 3.5736162662506104, "perplexity": 35.645263120991046, "lr": 0.0026291804804649314, "grad_norm": 0.16043, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:55.267142+00:00", "epoch": 0, "step": 23400, "train_loss": 3.412796974182129, "perplexity": 30.35001386013863, "lr": 0.0026291804804649314, "grad_norm": 0.171112, "tokens_per_sec": 107234} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:55.571738+00:00", "epoch": 0, "step": 23401, "train_loss": 3.5915629863739014, "perplexity": 36.29075357507842, "lr": 0.0026291804804649314, "grad_norm": 0.205266, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:55.876623+00:00", "epoch": 0, "step": 23402, "train_loss": 3.511101245880127, "perplexity": 33.48512284055435, "lr": 0.0026291804804649314, "grad_norm": 0.188193, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:56.181586+00:00", "epoch": 0, "step": 23403, "train_loss": 3.4818928241729736, "perplexity": 32.52122080723829, "lr": 0.0026291804804649314, "grad_norm": 0.169713, "tokens_per_sec": 107449} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:56.487625+00:00", "epoch": 0, "step": 23404, "train_loss": 3.4470674991607666, "perplexity": 31.40815269434135, "lr": 0.0026291804804649314, "grad_norm": 0.202799, "tokens_per_sec": 107070} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:56.793470+00:00", "epoch": 0, "step": 23405, "train_loss": 3.513662099838257, "perplexity": 33.57098324096966, "lr": 0.0026291804804649314, "grad_norm": 0.187306, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:57.097765+00:00", "epoch": 0, "step": 23406, "train_loss": 3.4432168006896973, "perplexity": 31.28744192832023, "lr": 0.0026291804804649314, "grad_norm": 0.194435, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:57.401062+00:00", "epoch": 0, "step": 23407, "train_loss": 3.629995584487915, "perplexity": 37.712650096151854, "lr": 0.0026291804804649314, "grad_norm": 0.214312, "tokens_per_sec": 108034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:57.705488+00:00", "epoch": 0, "step": 23408, "train_loss": 3.511394500732422, "perplexity": 33.49494395528143, "lr": 0.0026291804804649314, "grad_norm": 0.201641, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:58.010858+00:00", "epoch": 0, "step": 23409, "train_loss": 3.4796173572540283, "perplexity": 32.447303974674895, "lr": 0.0026291804804649314, "grad_norm": 0.188414, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:58.316188+00:00", "epoch": 0, "step": 23410, "train_loss": 3.4366960525512695, "perplexity": 31.084088129614578, "lr": 0.0026291804804649314, "grad_norm": 0.175286, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:58.620672+00:00", "epoch": 0, "step": 23411, "train_loss": 3.539238452911377, "perplexity": 34.44068100097042, "lr": 0.0026291804804649314, "grad_norm": 0.209833, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:58.924968+00:00", "epoch": 0, "step": 23412, "train_loss": 3.540553569793701, "perplexity": 34.486004318191014, "lr": 0.0026291804804649314, "grad_norm": 0.209429, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:59.229354+00:00", "epoch": 0, "step": 23413, "train_loss": 3.517390727996826, "perplexity": 33.69639060782109, "lr": 0.0026291804804649314, "grad_norm": 0.186784, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:59.534038+00:00", "epoch": 0, "step": 23414, "train_loss": 3.4314510822296143, "perplexity": 30.921479820552076, "lr": 0.0026291804804649314, "grad_norm": 0.174463, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:00:59.838805+00:00", "epoch": 0, "step": 23415, "train_loss": 3.355595827102661, "perplexity": 28.662677103802626, "lr": 0.0026291804804649314, "grad_norm": 0.175543, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:00.144020+00:00", "epoch": 0, "step": 23416, "train_loss": 3.4687235355377197, "perplexity": 32.09574720513856, "lr": 0.0026291804804649314, "grad_norm": 0.183228, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:00.447957+00:00", "epoch": 0, "step": 23417, "train_loss": 3.5375378131866455, "perplexity": 34.38215958673878, "lr": 0.0026291804804649314, "grad_norm": 0.191243, "tokens_per_sec": 107812} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:00.752721+00:00", "epoch": 0, "step": 23418, "train_loss": 3.452819585800171, "perplexity": 31.58933570043814, "lr": 0.0026291804804649314, "grad_norm": 0.167076, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:01.057210+00:00", "epoch": 0, "step": 23419, "train_loss": 3.397479295730591, "perplexity": 29.888664527739557, "lr": 0.0026291804804649314, "grad_norm": 0.173621, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:01.363389+00:00", "epoch": 0, "step": 23420, "train_loss": 3.41607666015625, "perplexity": 30.449715781036925, "lr": 0.0026291804804649314, "grad_norm": 0.171114, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:01.667851+00:00", "epoch": 0, "step": 23421, "train_loss": 3.512855291366577, "perplexity": 33.54390881069674, "lr": 0.0026291804804649314, "grad_norm": 0.178644, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:01.972067+00:00", "epoch": 0, "step": 23422, "train_loss": 3.6059632301330566, "perplexity": 36.81713015220912, "lr": 0.0026291804804649314, "grad_norm": 0.167174, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:02.276507+00:00", "epoch": 0, "step": 23423, "train_loss": 3.513451337814331, "perplexity": 33.56390849816636, "lr": 0.0026291804804649314, "grad_norm": 0.18564, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:02.581110+00:00", "epoch": 0, "step": 23424, "train_loss": 3.478924036026001, "perplexity": 32.42481536685244, "lr": 0.0026291804804649314, "grad_norm": 0.167089, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:02.885599+00:00", "epoch": 0, "step": 23425, "train_loss": 3.4803950786590576, "perplexity": 32.472548752941265, "lr": 0.0026291804804649314, "grad_norm": 0.164211, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:03.190464+00:00", "epoch": 0, "step": 23426, "train_loss": 3.3802356719970703, "perplexity": 29.37769379728254, "lr": 0.0026291804804649314, "grad_norm": 0.182172, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:03.494458+00:00", "epoch": 0, "step": 23427, "train_loss": 3.631333351135254, "perplexity": 37.76313458233516, "lr": 0.0026291804804649314, "grad_norm": 0.169101, "tokens_per_sec": 107792} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:03.799328+00:00", "epoch": 0, "step": 23428, "train_loss": 3.4446308612823486, "perplexity": 31.33171556241923, "lr": 0.0026291804804649314, "grad_norm": 0.158512, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:04.103594+00:00", "epoch": 0, "step": 23429, "train_loss": 3.562403678894043, "perplexity": 35.2478198420832, "lr": 0.0026291804804649314, "grad_norm": 0.176958, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:04.408243+00:00", "epoch": 0, "step": 23430, "train_loss": 3.5711588859558105, "perplexity": 35.55777669153395, "lr": 0.0026291804804649314, "grad_norm": 0.154844, "tokens_per_sec": 107617} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:04.712613+00:00", "epoch": 0, "step": 23431, "train_loss": 3.5773024559020996, "perplexity": 35.7769007926862, "lr": 0.0026291804804649314, "grad_norm": 0.168118, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:05.017390+00:00", "epoch": 0, "step": 23432, "train_loss": 3.5249521732330322, "perplexity": 33.95214975586143, "lr": 0.0026291804804649314, "grad_norm": 0.156902, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:05.321405+00:00", "epoch": 0, "step": 23433, "train_loss": 3.571181297302246, "perplexity": 35.55857359811571, "lr": 0.0026291804804649314, "grad_norm": 0.147696, "tokens_per_sec": 107783} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:05.625675+00:00", "epoch": 0, "step": 23434, "train_loss": 3.413158893585205, "perplexity": 30.36100010698628, "lr": 0.0026291804804649314, "grad_norm": 0.162962, "tokens_per_sec": 107693} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:05.931036+00:00", "epoch": 0, "step": 23435, "train_loss": 3.521451950073242, "perplexity": 33.833517395947425, "lr": 0.0026291804804649314, "grad_norm": 0.160543, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:06.236315+00:00", "epoch": 0, "step": 23436, "train_loss": 3.5183866024017334, "perplexity": 33.7299646957778, "lr": 0.0026291804804649314, "grad_norm": 0.178699, "tokens_per_sec": 107338} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:06.540934+00:00", "epoch": 0, "step": 23437, "train_loss": 3.4427993297576904, "perplexity": 31.274383056817594, "lr": 0.0026291804804649314, "grad_norm": 0.172469, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:06.845462+00:00", "epoch": 0, "step": 23438, "train_loss": 3.5197157859802246, "perplexity": 33.77482782001344, "lr": 0.0026291804804649314, "grad_norm": 0.166965, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:07.149605+00:00", "epoch": 0, "step": 23439, "train_loss": 3.4714877605438232, "perplexity": 32.18458980605203, "lr": 0.0026291804804649314, "grad_norm": 0.1631, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:07.455555+00:00", "epoch": 0, "step": 23440, "train_loss": 3.4828834533691406, "perplexity": 32.55345324062379, "lr": 0.0026291804804649314, "grad_norm": 0.173516, "tokens_per_sec": 107163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:07.761078+00:00", "epoch": 0, "step": 23441, "train_loss": 3.4617063999176025, "perplexity": 31.871315349476983, "lr": 0.0026291804804649314, "grad_norm": 0.192135, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:08.065389+00:00", "epoch": 0, "step": 23442, "train_loss": 3.368168830871582, "perplexity": 29.025328079777555, "lr": 0.0026291804804649314, "grad_norm": 0.167611, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:08.370305+00:00", "epoch": 0, "step": 23443, "train_loss": 3.490213394165039, "perplexity": 32.792944782708766, "lr": 0.0026291804804649314, "grad_norm": 0.205955, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:08.674509+00:00", "epoch": 0, "step": 23444, "train_loss": 3.443354606628418, "perplexity": 31.2917538207207, "lr": 0.0026291804804649314, "grad_norm": 0.190724, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:08.978337+00:00", "epoch": 0, "step": 23445, "train_loss": 3.4763331413269043, "perplexity": 32.340914820273476, "lr": 0.0026291804804649314, "grad_norm": 0.189685, "tokens_per_sec": 107851} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:09.284289+00:00", "epoch": 0, "step": 23446, "train_loss": 3.4765243530273438, "perplexity": 32.34709937285059, "lr": 0.0026291804804649314, "grad_norm": 0.16886, "tokens_per_sec": 107102} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:09.590176+00:00", "epoch": 0, "step": 23447, "train_loss": 3.4790256023406982, "perplexity": 32.428108803102255, "lr": 0.0026291804804649314, "grad_norm": 0.154876, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:09.895554+00:00", "epoch": 0, "step": 23448, "train_loss": 3.4352993965148926, "perplexity": 31.040704653242315, "lr": 0.0026291804804649314, "grad_norm": 0.180193, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:10.199750+00:00", "epoch": 0, "step": 23449, "train_loss": 3.6420302391052246, "perplexity": 38.1692508198788, "lr": 0.0026291804804649314, "grad_norm": 0.170419, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:10.504203+00:00", "epoch": 0, "step": 23450, "train_loss": 3.4780194759368896, "perplexity": 32.39549843443728, "lr": 0.0026291804804649314, "grad_norm": 0.164554, "tokens_per_sec": 107689} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:10.808591+00:00", "epoch": 0, "step": 23451, "train_loss": 3.5557379722595215, "perplexity": 35.013649537831746, "lr": 0.0026291804804649314, "grad_norm": 0.15839, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:11.114261+00:00", "epoch": 0, "step": 23452, "train_loss": 3.5588059425354004, "perplexity": 35.121235324474554, "lr": 0.0026291804804649314, "grad_norm": 0.173849, "tokens_per_sec": 107201} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:11.419198+00:00", "epoch": 0, "step": 23453, "train_loss": 3.534721612930298, "perplexity": 34.285468754374705, "lr": 0.0026291804804649314, "grad_norm": 0.191893, "tokens_per_sec": 107459} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:11.723052+00:00", "epoch": 0, "step": 23454, "train_loss": 3.5713891983032227, "perplexity": 35.56596702968434, "lr": 0.0026291804804649314, "grad_norm": 0.158526, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:12.027461+00:00", "epoch": 0, "step": 23455, "train_loss": 3.5412328243255615, "perplexity": 34.50943705040371, "lr": 0.0026291804804649314, "grad_norm": 0.179352, "tokens_per_sec": 107645} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:12.331578+00:00", "epoch": 0, "step": 23456, "train_loss": 3.395397186279297, "perplexity": 29.826497798265954, "lr": 0.0026291804804649314, "grad_norm": 0.19278, "tokens_per_sec": 107748} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:12.636901+00:00", "epoch": 0, "step": 23457, "train_loss": 3.5051076412200928, "perplexity": 33.285026500443315, "lr": 0.0026291804804649314, "grad_norm": 0.178859, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:12.941965+00:00", "epoch": 0, "step": 23458, "train_loss": 3.6380021572113037, "perplexity": 38.01581119290711, "lr": 0.0026291804804649314, "grad_norm": 0.185285, "tokens_per_sec": 107417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:13.247098+00:00", "epoch": 0, "step": 23459, "train_loss": 3.546529531478882, "perplexity": 34.692708371269255, "lr": 0.0026291804804649314, "grad_norm": 0.214435, "tokens_per_sec": 107388} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:13.551859+00:00", "epoch": 0, "step": 23460, "train_loss": 3.5027084350585938, "perplexity": 33.205264580682865, "lr": 0.0026291804804649314, "grad_norm": 0.200042, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:13.856009+00:00", "epoch": 0, "step": 23461, "train_loss": 3.4776434898376465, "perplexity": 32.3833204668646, "lr": 0.0026291804804649314, "grad_norm": 0.16696, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:14.160429+00:00", "epoch": 0, "step": 23462, "train_loss": 3.506481409072876, "perplexity": 33.330783822601525, "lr": 0.0026291804804649314, "grad_norm": 0.208171, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:14.466454+00:00", "epoch": 0, "step": 23463, "train_loss": 3.510542869567871, "perplexity": 33.4664307602403, "lr": 0.0026291804804649314, "grad_norm": 0.189339, "tokens_per_sec": 107084} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:14.771592+00:00", "epoch": 0, "step": 23464, "train_loss": 3.491048812866211, "perplexity": 32.82035206873237, "lr": 0.0026291804804649314, "grad_norm": 0.172789, "tokens_per_sec": 107381} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:15.076239+00:00", "epoch": 0, "step": 23465, "train_loss": 3.526942253112793, "perplexity": 34.01978452293793, "lr": 0.0026291804804649314, "grad_norm": 0.155136, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:15.380268+00:00", "epoch": 0, "step": 23466, "train_loss": 3.502253293991089, "perplexity": 33.190154939884685, "lr": 0.0026291804804649314, "grad_norm": 0.165525, "tokens_per_sec": 107779} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:15.685594+00:00", "epoch": 0, "step": 23467, "train_loss": 3.440431594848633, "perplexity": 31.200421203878847, "lr": 0.0026291804804649314, "grad_norm": 0.187283, "tokens_per_sec": 107322} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:15.991032+00:00", "epoch": 0, "step": 23468, "train_loss": 3.4584977626800537, "perplexity": 31.769215748124346, "lr": 0.0026291804804649314, "grad_norm": 0.177485, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:16.296851+00:00", "epoch": 0, "step": 23469, "train_loss": 3.4619951248168945, "perplexity": 31.880518720348793, "lr": 0.0026291804804649314, "grad_norm": 0.20439, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:16.600922+00:00", "epoch": 0, "step": 23470, "train_loss": 3.532632350921631, "perplexity": 34.21391220327748, "lr": 0.0026291804804649314, "grad_norm": 0.183951, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:16.905255+00:00", "epoch": 0, "step": 23471, "train_loss": 3.441406011581421, "perplexity": 31.230838233394486, "lr": 0.0026291804804649314, "grad_norm": 0.164748, "tokens_per_sec": 107606} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:17.209089+00:00", "epoch": 0, "step": 23472, "train_loss": 3.543583869934082, "perplexity": 34.590665759618226, "lr": 0.0026291804804649314, "grad_norm": 0.172482, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:17.513477+00:00", "epoch": 0, "step": 23473, "train_loss": 3.342212438583374, "perplexity": 28.2816289057655, "lr": 0.0026291804804649314, "grad_norm": 0.226951, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:17.817681+00:00", "epoch": 0, "step": 23474, "train_loss": 3.551652431488037, "perplexity": 34.870891665212895, "lr": 0.0026291804804649314, "grad_norm": 0.158267, "tokens_per_sec": 107718} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:18.122214+00:00", "epoch": 0, "step": 23475, "train_loss": 3.533308506011963, "perplexity": 34.23705393699275, "lr": 0.0026291804804649314, "grad_norm": 0.174994, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:18.426134+00:00", "epoch": 0, "step": 23476, "train_loss": 3.5029947757720947, "perplexity": 33.214773961231316, "lr": 0.0026291804804649314, "grad_norm": 0.183967, "tokens_per_sec": 107813} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:18.730339+00:00", "epoch": 0, "step": 23477, "train_loss": 3.4438204765319824, "perplexity": 31.30633510328215, "lr": 0.0026291804804649314, "grad_norm": 0.148094, "tokens_per_sec": 107717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:19.034928+00:00", "epoch": 0, "step": 23478, "train_loss": 3.472450017929077, "perplexity": 32.21557457055267, "lr": 0.0026291804804649314, "grad_norm": 0.156486, "tokens_per_sec": 107581} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:19.340680+00:00", "epoch": 0, "step": 23479, "train_loss": 3.525090456008911, "perplexity": 33.95684507801033, "lr": 0.0026291804804649314, "grad_norm": 0.146628, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:19.646647+00:00", "epoch": 0, "step": 23480, "train_loss": 3.4682419300079346, "perplexity": 32.080293437410674, "lr": 0.0026291804804649314, "grad_norm": 0.155483, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:19.951380+00:00", "epoch": 0, "step": 23481, "train_loss": 3.5365583896636963, "perplexity": 34.348501376381826, "lr": 0.0026291804804649314, "grad_norm": 0.162476, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:20.256428+00:00", "epoch": 0, "step": 23482, "train_loss": 3.5203678607940674, "perplexity": 33.79685871669255, "lr": 0.0026291804804649314, "grad_norm": 0.169528, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:20.561587+00:00", "epoch": 0, "step": 23483, "train_loss": 3.4827685356140137, "perplexity": 32.54971248579944, "lr": 0.0026291804804649314, "grad_norm": 0.187793, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:20.867832+00:00", "epoch": 0, "step": 23484, "train_loss": 3.453640937805176, "perplexity": 31.615292322955206, "lr": 0.0026291804804649314, "grad_norm": 0.17326, "tokens_per_sec": 107000} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:21.172703+00:00", "epoch": 0, "step": 23485, "train_loss": 3.6006627082824707, "perplexity": 36.622496435191024, "lr": 0.0026291804804649314, "grad_norm": 0.172866, "tokens_per_sec": 107483} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:21.476870+00:00", "epoch": 0, "step": 23486, "train_loss": 3.5732100009918213, "perplexity": 35.63078463019764, "lr": 0.0026291804804649314, "grad_norm": 0.173873, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:21.781318+00:00", "epoch": 0, "step": 23487, "train_loss": 3.472233772277832, "perplexity": 32.20860884583041, "lr": 0.0026291804804649314, "grad_norm": 0.172553, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:22.086299+00:00", "epoch": 0, "step": 23488, "train_loss": 3.5349297523498535, "perplexity": 34.29260565464981, "lr": 0.0026291804804649314, "grad_norm": 0.177563, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:22.391633+00:00", "epoch": 0, "step": 23489, "train_loss": 3.5216381549835205, "perplexity": 33.83981794959735, "lr": 0.0026291804804649314, "grad_norm": 0.195273, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:22.698108+00:00", "epoch": 0, "step": 23490, "train_loss": 3.406324863433838, "perplexity": 30.154219494090064, "lr": 0.0026291804804649314, "grad_norm": 0.168225, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:23.003713+00:00", "epoch": 0, "step": 23491, "train_loss": 3.487522602081299, "perplexity": 32.70482439647239, "lr": 0.0026291804804649314, "grad_norm": 0.162496, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:23.308407+00:00", "epoch": 0, "step": 23492, "train_loss": 3.3614964485168457, "perplexity": 28.832304671975102, "lr": 0.0026291804804649314, "grad_norm": 0.163981, "tokens_per_sec": 107544} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:23.613988+00:00", "epoch": 0, "step": 23493, "train_loss": 3.42348313331604, "perplexity": 30.676078021849513, "lr": 0.0026291804804649314, "grad_norm": 0.157962, "tokens_per_sec": 107232} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:23.919408+00:00", "epoch": 0, "step": 23494, "train_loss": 3.5170388221740723, "perplexity": 33.68453473795758, "lr": 0.0026291804804649314, "grad_norm": 0.153758, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:24.225686+00:00", "epoch": 0, "step": 23495, "train_loss": 3.4009156227111816, "perplexity": 29.99154842218055, "lr": 0.0026291804804649314, "grad_norm": 0.174197, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:24.530874+00:00", "epoch": 0, "step": 23496, "train_loss": 3.4717960357666016, "perplexity": 32.194513047110604, "lr": 0.0026291804804649314, "grad_norm": 0.177566, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:24.836182+00:00", "epoch": 0, "step": 23497, "train_loss": 3.5400381088256836, "perplexity": 34.46823270970085, "lr": 0.0026291804804649314, "grad_norm": 0.154706, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:25.140647+00:00", "epoch": 0, "step": 23498, "train_loss": 3.5608439445495605, "perplexity": 35.19288545956583, "lr": 0.0026291804804649314, "grad_norm": 0.164135, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:25.445128+00:00", "epoch": 0, "step": 23499, "train_loss": 3.5615153312683105, "perplexity": 35.21652142900801, "lr": 0.0026291804804649314, "grad_norm": 0.182267, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:25.751061+00:00", "epoch": 0, "step": 23500, "train_loss": 3.5056910514831543, "perplexity": 33.304450992182915, "lr": 0.0026291804804649314, "grad_norm": 0.161713, "tokens_per_sec": 107165} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T23:01:28.823524+00:00", "step": 23500, "epoch": 0, "val_loss": 3.4510040163993834, "val_ppl": 31.532035101496223, "eval_train_loss": 3.5056910514831543, "eval_train_ppl": 33.304450992182915} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T23:01:29.795692+00:00", "step": 23500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4510_epoch_0000_step_0023500.pt", "val_loss": 3.4510040163993834} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:31.204323+00:00", "epoch": 0, "step": 23501, "train_loss": 3.638503313064575, "perplexity": 38.034867813973044, "lr": 0.0026291804804649314, "grad_norm": 0.175022, "tokens_per_sec": 6009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:31.508789+00:00", "epoch": 0, "step": 23502, "train_loss": 3.6080093383789062, "perplexity": 36.89253910693544, "lr": 0.0026291804804649314, "grad_norm": 0.211493, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:31.812770+00:00", "epoch": 0, "step": 23503, "train_loss": 3.4650559425354004, "perplexity": 31.97824866740055, "lr": 0.0026291804804649314, "grad_norm": 0.196688, "tokens_per_sec": 107797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:32.118277+00:00", "epoch": 0, "step": 23504, "train_loss": 3.536672353744507, "perplexity": 34.35241609483281, "lr": 0.0026291804804649314, "grad_norm": 0.175234, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:32.423902+00:00", "epoch": 0, "step": 23505, "train_loss": 3.4413623809814453, "perplexity": 31.22947564291016, "lr": 0.0026291804804649314, "grad_norm": 0.177759, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:32.729205+00:00", "epoch": 0, "step": 23506, "train_loss": 3.458672285079956, "perplexity": 31.77476067174234, "lr": 0.0026291804804649314, "grad_norm": 0.216058, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:33.033658+00:00", "epoch": 0, "step": 23507, "train_loss": 3.5448148250579834, "perplexity": 34.6332715343927, "lr": 0.0026291804804649314, "grad_norm": 0.216767, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:33.338466+00:00", "epoch": 0, "step": 23508, "train_loss": 3.4572479724884033, "perplexity": 31.729535694921502, "lr": 0.0026291804804649314, "grad_norm": 0.189175, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:33.644284+00:00", "epoch": 0, "step": 23509, "train_loss": 3.488218307495117, "perplexity": 32.72758523636473, "lr": 0.0026291804804649314, "grad_norm": 0.172476, "tokens_per_sec": 107154} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:33.949769+00:00", "epoch": 0, "step": 23510, "train_loss": 3.5012173652648926, "perplexity": 33.15579010778713, "lr": 0.0026291804804649314, "grad_norm": 0.172055, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:34.254130+00:00", "epoch": 0, "step": 23511, "train_loss": 3.4666738510131836, "perplexity": 32.03002842318211, "lr": 0.0026291804804649314, "grad_norm": 0.180146, "tokens_per_sec": 107566} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:34.558694+00:00", "epoch": 0, "step": 23512, "train_loss": 3.564568281173706, "perplexity": 35.324199989764146, "lr": 0.0026291804804649314, "grad_norm": 0.18046, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:34.861645+00:00", "epoch": 0, "step": 23513, "train_loss": 3.5120911598205566, "perplexity": 33.51828664239791, "lr": 0.0026291804804649314, "grad_norm": 0.173602, "tokens_per_sec": 108163} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:35.165716+00:00", "epoch": 0, "step": 23514, "train_loss": 3.541543960571289, "perplexity": 34.52017585761919, "lr": 0.0026291804804649314, "grad_norm": 0.191371, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:35.470371+00:00", "epoch": 0, "step": 23515, "train_loss": 3.3779470920562744, "perplexity": 29.31053747217199, "lr": 0.0026291804804649314, "grad_norm": 0.172685, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:35.775184+00:00", "epoch": 0, "step": 23516, "train_loss": 3.577315330505371, "perplexity": 35.77736140905531, "lr": 0.0026291804804649314, "grad_norm": 0.178839, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:36.079390+00:00", "epoch": 0, "step": 23517, "train_loss": 3.501554489135742, "perplexity": 33.16696960042039, "lr": 0.0026291804804649314, "grad_norm": 0.179637, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:36.383070+00:00", "epoch": 0, "step": 23518, "train_loss": 3.494236946105957, "perplexity": 32.92515469770588, "lr": 0.0026291804804649314, "grad_norm": 0.19914, "tokens_per_sec": 107903} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:36.688580+00:00", "epoch": 0, "step": 23519, "train_loss": 3.464031457901001, "perplexity": 31.9455042189619, "lr": 0.0026291804804649314, "grad_norm": 0.181328, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:36.993210+00:00", "epoch": 0, "step": 23520, "train_loss": 3.5721933841705322, "perplexity": 35.59458018133085, "lr": 0.0026291804804649314, "grad_norm": 0.172442, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:37.297217+00:00", "epoch": 0, "step": 23521, "train_loss": 3.3809945583343506, "perplexity": 29.39999658929781, "lr": 0.0026291804804649314, "grad_norm": 0.167995, "tokens_per_sec": 107725} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:37.601343+00:00", "epoch": 0, "step": 23522, "train_loss": 3.4835567474365234, "perplexity": 32.57537866786529, "lr": 0.0026291804804649314, "grad_norm": 0.167443, "tokens_per_sec": 107745} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:37.905159+00:00", "epoch": 0, "step": 23523, "train_loss": 3.527228832244873, "perplexity": 34.029535280375136, "lr": 0.0026291804804649314, "grad_norm": 0.170608, "tokens_per_sec": 107855} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:38.209789+00:00", "epoch": 0, "step": 23524, "train_loss": 3.4603145122528076, "perplexity": 31.826984917428994, "lr": 0.0026291804804649314, "grad_norm": 0.165809, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:38.514508+00:00", "epoch": 0, "step": 23525, "train_loss": 3.433220863342285, "perplexity": 30.976252525067046, "lr": 0.0026291804804649314, "grad_norm": 0.157571, "tokens_per_sec": 107534} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:38.819543+00:00", "epoch": 0, "step": 23526, "train_loss": 3.5314087867736816, "perplexity": 34.17207488748388, "lr": 0.0026291804804649314, "grad_norm": 0.170713, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:39.122935+00:00", "epoch": 0, "step": 23527, "train_loss": 3.512063503265381, "perplexity": 33.51735965487269, "lr": 0.0026291804804649314, "grad_norm": 0.171193, "tokens_per_sec": 108004} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:39.427465+00:00", "epoch": 0, "step": 23528, "train_loss": 3.5566565990448, "perplexity": 35.04582879224713, "lr": 0.0026291804804649314, "grad_norm": 0.157865, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:39.732112+00:00", "epoch": 0, "step": 23529, "train_loss": 3.483494520187378, "perplexity": 32.573351654729294, "lr": 0.0026291804804649314, "grad_norm": 0.180538, "tokens_per_sec": 107562} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:40.038025+00:00", "epoch": 0, "step": 23530, "train_loss": 3.4922971725463867, "perplexity": 32.86134925723854, "lr": 0.0026291804804649314, "grad_norm": 0.158917, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:40.341836+00:00", "epoch": 0, "step": 23531, "train_loss": 3.580756425857544, "perplexity": 35.90068678661208, "lr": 0.0026291804804649314, "grad_norm": 0.171568, "tokens_per_sec": 107789} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:40.646747+00:00", "epoch": 0, "step": 23532, "train_loss": 3.470367193222046, "perplexity": 32.14854500553505, "lr": 0.0026291804804649314, "grad_norm": 0.164783, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:40.951394+00:00", "epoch": 0, "step": 23533, "train_loss": 3.4212663173675537, "perplexity": 30.608150122497, "lr": 0.0026291804804649314, "grad_norm": 0.155128, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:41.257202+00:00", "epoch": 0, "step": 23534, "train_loss": 3.5404255390167236, "perplexity": 34.48158933089646, "lr": 0.0026291804804649314, "grad_norm": 0.173134, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:41.562254+00:00", "epoch": 0, "step": 23535, "train_loss": 3.4383764266967773, "perplexity": 31.136364937641595, "lr": 0.0026291804804649314, "grad_norm": 0.160561, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:41.866324+00:00", "epoch": 0, "step": 23536, "train_loss": 3.4562418460845947, "perplexity": 31.69762782564481, "lr": 0.0026291804804649314, "grad_norm": 0.165028, "tokens_per_sec": 107765} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:42.170485+00:00", "epoch": 0, "step": 23537, "train_loss": 3.5672590732574463, "perplexity": 35.41937806223606, "lr": 0.0026291804804649314, "grad_norm": 0.150328, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:42.474574+00:00", "epoch": 0, "step": 23538, "train_loss": 3.363783121109009, "perplexity": 28.898310150535604, "lr": 0.0026291804804649314, "grad_norm": 0.160196, "tokens_per_sec": 107760} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:42.779592+00:00", "epoch": 0, "step": 23539, "train_loss": 3.5983169078826904, "perplexity": 36.536688052430414, "lr": 0.0026291804804649314, "grad_norm": 0.166415, "tokens_per_sec": 107429} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:43.084687+00:00", "epoch": 0, "step": 23540, "train_loss": 3.5883803367614746, "perplexity": 36.17543642665663, "lr": 0.0026291804804649314, "grad_norm": 0.177735, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:43.388543+00:00", "epoch": 0, "step": 23541, "train_loss": 3.47322416305542, "perplexity": 32.24052375649814, "lr": 0.0026291804804649314, "grad_norm": 0.180984, "tokens_per_sec": 107782} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:43.692345+00:00", "epoch": 0, "step": 23542, "train_loss": 3.368479013442993, "perplexity": 29.034332627132024, "lr": 0.0026291804804649314, "grad_norm": 0.194146, "tokens_per_sec": 107860} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:43.997929+00:00", "epoch": 0, "step": 23543, "train_loss": 3.4808990955352783, "perplexity": 32.488919590769356, "lr": 0.0026291804804649314, "grad_norm": 0.177143, "tokens_per_sec": 107233} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:44.303347+00:00", "epoch": 0, "step": 23544, "train_loss": 3.433873176574707, "perplexity": 30.996465336305373, "lr": 0.0026291804804649314, "grad_norm": 0.153087, "tokens_per_sec": 107288} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:44.608537+00:00", "epoch": 0, "step": 23545, "train_loss": 3.5173404216766357, "perplexity": 33.694695509043356, "lr": 0.0026291804804649314, "grad_norm": 0.222163, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:44.913174+00:00", "epoch": 0, "step": 23546, "train_loss": 3.424919366836548, "perplexity": 30.720167687331053, "lr": 0.0026291804804649314, "grad_norm": 0.205842, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:45.216815+00:00", "epoch": 0, "step": 23547, "train_loss": 3.4752156734466553, "perplexity": 32.304795071831975, "lr": 0.0026291804804649314, "grad_norm": 0.248539, "tokens_per_sec": 107917} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:45.521317+00:00", "epoch": 0, "step": 23548, "train_loss": 3.548429250717163, "perplexity": 34.758677418284286, "lr": 0.0026291804804649314, "grad_norm": 0.205967, "tokens_per_sec": 107612} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:45.825545+00:00", "epoch": 0, "step": 23549, "train_loss": 3.4471898078918457, "perplexity": 31.411994420576477, "lr": 0.0026291804804649314, "grad_norm": 0.193115, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:46.129288+00:00", "epoch": 0, "step": 23550, "train_loss": 3.4044978618621826, "perplexity": 30.099177983455956, "lr": 0.0026291804804649314, "grad_norm": 0.220801, "tokens_per_sec": 107937} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:46.432936+00:00", "epoch": 0, "step": 23551, "train_loss": 3.4845120906829834, "perplexity": 32.606514206071516, "lr": 0.0026291804804649314, "grad_norm": 0.197853, "tokens_per_sec": 107857} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:46.737966+00:00", "epoch": 0, "step": 23552, "train_loss": 3.4918179512023926, "perplexity": 32.845605170034645, "lr": 0.0026291804804649314, "grad_norm": 0.219267, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:47.042502+00:00", "epoch": 0, "step": 23553, "train_loss": 3.516244649887085, "perplexity": 33.65779403373721, "lr": 0.0026291804804649314, "grad_norm": 0.192336, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:47.347387+00:00", "epoch": 0, "step": 23554, "train_loss": 3.5309031009674072, "perplexity": 34.154798922715564, "lr": 0.0026291804804649314, "grad_norm": 0.175819, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:47.652230+00:00", "epoch": 0, "step": 23555, "train_loss": 3.453312397003174, "perplexity": 31.604907115535497, "lr": 0.0026291804804649314, "grad_norm": 0.211914, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:47.955834+00:00", "epoch": 0, "step": 23556, "train_loss": 3.5282933712005615, "perplexity": 34.06578033503852, "lr": 0.0026291804804649314, "grad_norm": 0.149832, "tokens_per_sec": 107928} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:48.261201+00:00", "epoch": 0, "step": 23557, "train_loss": 3.5007338523864746, "perplexity": 33.13976273130961, "lr": 0.0026291804804649314, "grad_norm": 0.186284, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:48.565197+00:00", "epoch": 0, "step": 23558, "train_loss": 3.4699697494506836, "perplexity": 32.135770305344906, "lr": 0.0026291804804649314, "grad_norm": 0.174665, "tokens_per_sec": 107791} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:48.870435+00:00", "epoch": 0, "step": 23559, "train_loss": 3.486832857131958, "perplexity": 32.68227418686646, "lr": 0.0026291804804649314, "grad_norm": 0.177138, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:49.176095+00:00", "epoch": 0, "step": 23560, "train_loss": 3.4532785415649414, "perplexity": 31.603837135667227, "lr": 0.0026291804804649314, "grad_norm": 0.171849, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:49.479728+00:00", "epoch": 0, "step": 23561, "train_loss": 3.404068946838379, "perplexity": 30.086270762063172, "lr": 0.0026291804804649314, "grad_norm": 0.174312, "tokens_per_sec": 107844} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:49.784731+00:00", "epoch": 0, "step": 23562, "train_loss": 3.484161138534546, "perplexity": 32.59507288765291, "lr": 0.0026291804804649314, "grad_norm": 0.180722, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:50.089489+00:00", "epoch": 0, "step": 23563, "train_loss": 3.5360183715820312, "perplexity": 34.32995757200596, "lr": 0.0026291804804649314, "grad_norm": 0.206934, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:50.393702+00:00", "epoch": 0, "step": 23564, "train_loss": 3.519312858581543, "perplexity": 33.76122175781071, "lr": 0.0026291804804649314, "grad_norm": 0.162981, "tokens_per_sec": 107713} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:50.697049+00:00", "epoch": 0, "step": 23565, "train_loss": 3.5724997520446777, "perplexity": 35.605486887839035, "lr": 0.0026291804804649314, "grad_norm": 0.200888, "tokens_per_sec": 108021} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:51.001028+00:00", "epoch": 0, "step": 23566, "train_loss": 3.5311315059661865, "perplexity": 34.16260094049783, "lr": 0.0026291804804649314, "grad_norm": 0.251154, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:51.305929+00:00", "epoch": 0, "step": 23567, "train_loss": 3.4230003356933594, "perplexity": 30.661271258929983, "lr": 0.0026291804804649314, "grad_norm": 0.157232, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:51.610618+00:00", "epoch": 0, "step": 23568, "train_loss": 3.52121639251709, "perplexity": 33.825548593866216, "lr": 0.0026291804804649314, "grad_norm": 0.205624, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:51.916316+00:00", "epoch": 0, "step": 23569, "train_loss": 3.454681396484375, "perplexity": 31.648203846819328, "lr": 0.0026291804804649314, "grad_norm": 0.186878, "tokens_per_sec": 107191} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:52.221502+00:00", "epoch": 0, "step": 23570, "train_loss": 3.526268482208252, "perplexity": 33.99687070214734, "lr": 0.0026291804804649314, "grad_norm": 0.174524, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:52.526790+00:00", "epoch": 0, "step": 23571, "train_loss": 3.553074359893799, "perplexity": 34.920510845706666, "lr": 0.0026291804804649314, "grad_norm": 0.175472, "tokens_per_sec": 107250} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:52.832186+00:00", "epoch": 0, "step": 23572, "train_loss": 3.436840295791626, "perplexity": 31.088572122594982, "lr": 0.0026291804804649314, "grad_norm": 0.19189, "tokens_per_sec": 107303} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:53.136858+00:00", "epoch": 0, "step": 23573, "train_loss": 3.4692938327789307, "perplexity": 32.11405654161005, "lr": 0.0026291804804649314, "grad_norm": 0.143882, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:53.442550+00:00", "epoch": 0, "step": 23574, "train_loss": 3.4463906288146973, "perplexity": 31.386900640411675, "lr": 0.0026291804804649314, "grad_norm": 0.167712, "tokens_per_sec": 107193} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:53.747694+00:00", "epoch": 0, "step": 23575, "train_loss": 3.5536820888519287, "perplexity": 34.94173950136131, "lr": 0.0026291804804649314, "grad_norm": 0.159937, "tokens_per_sec": 107386} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:54.052110+00:00", "epoch": 0, "step": 23576, "train_loss": 3.459859848022461, "perplexity": 31.812517614958953, "lr": 0.0026291804804649314, "grad_norm": 0.163582, "tokens_per_sec": 107643} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:54.356202+00:00", "epoch": 0, "step": 23577, "train_loss": 3.4602646827697754, "perplexity": 31.82539903473627, "lr": 0.0026291804804649314, "grad_norm": 0.161147, "tokens_per_sec": 107757} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:54.660575+00:00", "epoch": 0, "step": 23578, "train_loss": 3.4801816940307617, "perplexity": 32.465620349429486, "lr": 0.0026291804804649314, "grad_norm": 0.181441, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:54.965945+00:00", "epoch": 0, "step": 23579, "train_loss": 3.4717793464660645, "perplexity": 32.1939757476903, "lr": 0.0026291804804649314, "grad_norm": 0.160916, "tokens_per_sec": 107307} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:55.271630+00:00", "epoch": 0, "step": 23580, "train_loss": 3.532205820083618, "perplexity": 34.1993220264351, "lr": 0.0026291804804649314, "grad_norm": 0.15134, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:55.576502+00:00", "epoch": 0, "step": 23581, "train_loss": 3.5471248626708984, "perplexity": 34.71336817179986, "lr": 0.0026291804804649314, "grad_norm": 0.164864, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:55.880259+00:00", "epoch": 0, "step": 23582, "train_loss": 3.475996494293213, "perplexity": 32.330029179644484, "lr": 0.0026291804804649314, "grad_norm": 0.167447, "tokens_per_sec": 107876} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:56.185721+00:00", "epoch": 0, "step": 23583, "train_loss": 3.584991693496704, "perplexity": 36.053058242768316, "lr": 0.0026291804804649314, "grad_norm": 0.164641, "tokens_per_sec": 107273} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:56.491337+00:00", "epoch": 0, "step": 23584, "train_loss": 3.4798083305358887, "perplexity": 32.4535011345296, "lr": 0.0026291804804649314, "grad_norm": 0.171268, "tokens_per_sec": 107221} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:56.797816+00:00", "epoch": 0, "step": 23585, "train_loss": 3.412813901901245, "perplexity": 30.350527620996818, "lr": 0.0026291804804649314, "grad_norm": 0.175358, "tokens_per_sec": 106918} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:57.103361+00:00", "epoch": 0, "step": 23586, "train_loss": 3.5211896896362305, "perplexity": 33.82464536633155, "lr": 0.0026291804804649314, "grad_norm": 0.165455, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:57.408258+00:00", "epoch": 0, "step": 23587, "train_loss": 3.507654905319214, "perplexity": 33.369920331086526, "lr": 0.0026291804804649314, "grad_norm": 0.163882, "tokens_per_sec": 107472} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:57.711594+00:00", "epoch": 0, "step": 23588, "train_loss": 3.459700345993042, "perplexity": 31.807443858486423, "lr": 0.0026291804804649314, "grad_norm": 0.17014, "tokens_per_sec": 108025} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:58.017943+00:00", "epoch": 0, "step": 23589, "train_loss": 3.511756420135498, "perplexity": 33.50706861934703, "lr": 0.0026291804804649314, "grad_norm": 0.195584, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:58.322818+00:00", "epoch": 0, "step": 23590, "train_loss": 3.4482736587524414, "perplexity": 31.446058794778182, "lr": 0.0026291804804649314, "grad_norm": 0.170241, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:58.627747+00:00", "epoch": 0, "step": 23591, "train_loss": 3.5138208866119385, "perplexity": 33.57631429232684, "lr": 0.0026291804804649314, "grad_norm": 0.176083, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:58.931844+00:00", "epoch": 0, "step": 23592, "train_loss": 3.457714796066284, "perplexity": 31.744351248154832, "lr": 0.0026291804804649314, "grad_norm": 0.192796, "tokens_per_sec": 107754} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:59.236280+00:00", "epoch": 0, "step": 23593, "train_loss": 3.547585964202881, "perplexity": 34.72937824989625, "lr": 0.0026291804804649314, "grad_norm": 0.212094, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:59.540956+00:00", "epoch": 0, "step": 23594, "train_loss": 3.4949169158935547, "perplexity": 32.94755042149717, "lr": 0.0026291804804649314, "grad_norm": 0.17205, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:01:59.845872+00:00", "epoch": 0, "step": 23595, "train_loss": 3.5232317447662354, "perplexity": 33.893787729153914, "lr": 0.0026291804804649314, "grad_norm": 0.175083, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:00.150741+00:00", "epoch": 0, "step": 23596, "train_loss": 3.5092241764068604, "perplexity": 33.42232789233954, "lr": 0.0026291804804649314, "grad_norm": 0.163014, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:00.455513+00:00", "epoch": 0, "step": 23597, "train_loss": 3.5557148456573486, "perplexity": 35.01283980045154, "lr": 0.0026291804804649314, "grad_norm": 0.177844, "tokens_per_sec": 107516} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:00.759894+00:00", "epoch": 0, "step": 23598, "train_loss": 3.5242016315460205, "perplexity": 33.92667681255468, "lr": 0.0026291804804649314, "grad_norm": 0.165436, "tokens_per_sec": 107654} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:01.064593+00:00", "epoch": 0, "step": 23599, "train_loss": 3.5092811584472656, "perplexity": 33.42423241903933, "lr": 0.0026291804804649314, "grad_norm": 0.175761, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:01.369194+00:00", "epoch": 0, "step": 23600, "train_loss": 3.493396520614624, "perplexity": 32.897495182900535, "lr": 0.0026291804804649314, "grad_norm": 0.152099, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:01.673456+00:00", "epoch": 0, "step": 23601, "train_loss": 3.5532777309417725, "perplexity": 34.92761338879447, "lr": 0.0026291804804649314, "grad_norm": 0.181657, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:01.978662+00:00", "epoch": 0, "step": 23602, "train_loss": 3.466925621032715, "perplexity": 32.038093639311015, "lr": 0.0026291804804649314, "grad_norm": 0.182787, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:02.283232+00:00", "epoch": 0, "step": 23603, "train_loss": 3.512080192565918, "perplexity": 33.51791904082905, "lr": 0.0026291804804649314, "grad_norm": 0.149822, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:02.587743+00:00", "epoch": 0, "step": 23604, "train_loss": 3.5809624195098877, "perplexity": 35.90808286195103, "lr": 0.0026291804804649314, "grad_norm": 0.17986, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:02.892964+00:00", "epoch": 0, "step": 23605, "train_loss": 3.3995940685272217, "perplexity": 29.95193914454789, "lr": 0.0026291804804649314, "grad_norm": 0.174176, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:03.197916+00:00", "epoch": 0, "step": 23606, "train_loss": 3.5516409873962402, "perplexity": 34.8704926018111, "lr": 0.0026291804804649314, "grad_norm": 0.162098, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:03.502439+00:00", "epoch": 0, "step": 23607, "train_loss": 3.5475540161132812, "perplexity": 34.72826873033179, "lr": 0.0026291804804649314, "grad_norm": 0.165317, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:03.807144+00:00", "epoch": 0, "step": 23608, "train_loss": 3.454439163208008, "perplexity": 31.640538527145605, "lr": 0.0026291804804649314, "grad_norm": 0.165416, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:04.111783+00:00", "epoch": 0, "step": 23609, "train_loss": 3.53922963142395, "perplexity": 34.44037718427606, "lr": 0.0026291804804649314, "grad_norm": 0.183953, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:04.418043+00:00", "epoch": 0, "step": 23610, "train_loss": 3.4734878540039062, "perplexity": 32.249026411774544, "lr": 0.0026291804804649314, "grad_norm": 0.169157, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:04.722334+00:00", "epoch": 0, "step": 23611, "train_loss": 3.6397364139556885, "perplexity": 38.0817975719845, "lr": 0.0026291804804649314, "grad_norm": 0.171971, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:05.026881+00:00", "epoch": 0, "step": 23612, "train_loss": 3.514665126800537, "perplexity": 33.604672735221676, "lr": 0.0026291804804649314, "grad_norm": 0.187668, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:05.330696+00:00", "epoch": 0, "step": 23613, "train_loss": 3.5237672328948975, "perplexity": 33.91194231046584, "lr": 0.0026291804804649314, "grad_norm": 0.206146, "tokens_per_sec": 107859} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:05.634479+00:00", "epoch": 0, "step": 23614, "train_loss": 3.574767827987671, "perplexity": 35.68633448567595, "lr": 0.0026291804804649314, "grad_norm": 0.187841, "tokens_per_sec": 107861} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:05.938362+00:00", "epoch": 0, "step": 23615, "train_loss": 3.4125895500183105, "perplexity": 30.343719186748068, "lr": 0.0026291804804649314, "grad_norm": 0.161925, "tokens_per_sec": 107841} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:06.243157+00:00", "epoch": 0, "step": 23616, "train_loss": 3.455165147781372, "perplexity": 31.663517410157283, "lr": 0.0026291804804649314, "grad_norm": 0.218041, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:06.547686+00:00", "epoch": 0, "step": 23617, "train_loss": 3.4283483028411865, "perplexity": 30.825685980733557, "lr": 0.0026291804804649314, "grad_norm": 0.188938, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:06.853445+00:00", "epoch": 0, "step": 23618, "train_loss": 3.48928165435791, "perplexity": 32.762404520660105, "lr": 0.0026291804804649314, "grad_norm": 0.185296, "tokens_per_sec": 107171} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:07.158559+00:00", "epoch": 0, "step": 23619, "train_loss": 3.431211233139038, "perplexity": 30.91406422108579, "lr": 0.0026291804804649314, "grad_norm": 0.19365, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:07.462729+00:00", "epoch": 0, "step": 23620, "train_loss": 3.5404551029205322, "perplexity": 34.48260875635563, "lr": 0.0026291804804649314, "grad_norm": 0.175881, "tokens_per_sec": 107793} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:07.768058+00:00", "epoch": 0, "step": 23621, "train_loss": 3.4430923461914062, "perplexity": 31.28354830772656, "lr": 0.0026291804804649314, "grad_norm": 0.196413, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:08.073366+00:00", "epoch": 0, "step": 23622, "train_loss": 3.465222120285034, "perplexity": 31.983563182366165, "lr": 0.0026291804804649314, "grad_norm": 0.174824, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:08.378676+00:00", "epoch": 0, "step": 23623, "train_loss": 3.600546360015869, "perplexity": 36.618235719080445, "lr": 0.0026291804804649314, "grad_norm": 0.198814, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:08.682945+00:00", "epoch": 0, "step": 23624, "train_loss": 3.4762520790100098, "perplexity": 32.33829329704246, "lr": 0.0026291804804649314, "grad_norm": 0.171386, "tokens_per_sec": 107694} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:08.986215+00:00", "epoch": 0, "step": 23625, "train_loss": 3.444786548614502, "perplexity": 31.336593893364242, "lr": 0.0026291804804649314, "grad_norm": 0.192663, "tokens_per_sec": 108050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:09.291276+00:00", "epoch": 0, "step": 23626, "train_loss": 3.4904754161834717, "perplexity": 32.801538382098045, "lr": 0.0026291804804649314, "grad_norm": 0.173263, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:09.595649+00:00", "epoch": 0, "step": 23627, "train_loss": 3.4014172554016113, "perplexity": 30.00659693740375, "lr": 0.0026291804804649314, "grad_norm": 0.164544, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:09.900572+00:00", "epoch": 0, "step": 23628, "train_loss": 3.5031938552856445, "perplexity": 33.221387000512664, "lr": 0.0026291804804649314, "grad_norm": 0.194801, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:10.205341+00:00", "epoch": 0, "step": 23629, "train_loss": 3.521663188934326, "perplexity": 33.84066510453894, "lr": 0.0026291804804649314, "grad_norm": 0.16718, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:10.509958+00:00", "epoch": 0, "step": 23630, "train_loss": 3.3903462886810303, "perplexity": 29.67622703276635, "lr": 0.0026291804804649314, "grad_norm": 0.16258, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:10.814283+00:00", "epoch": 0, "step": 23631, "train_loss": 3.415630578994751, "perplexity": 30.43613576557373, "lr": 0.0026291804804649314, "grad_norm": 0.180815, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:11.119520+00:00", "epoch": 0, "step": 23632, "train_loss": 3.4630072116851807, "perplexity": 31.91280090813337, "lr": 0.0026291804804649314, "grad_norm": 0.187841, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:11.423714+00:00", "epoch": 0, "step": 23633, "train_loss": 3.4664969444274902, "perplexity": 32.02436260138948, "lr": 0.0026291804804649314, "grad_norm": 0.182425, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:11.728225+00:00", "epoch": 0, "step": 23634, "train_loss": 3.5415050983428955, "perplexity": 34.518834352727936, "lr": 0.0026291804804649314, "grad_norm": 0.160543, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:12.032195+00:00", "epoch": 0, "step": 23635, "train_loss": 3.5129122734069824, "perplexity": 33.54582026552272, "lr": 0.0026291804804649314, "grad_norm": 0.176289, "tokens_per_sec": 107800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:12.335353+00:00", "epoch": 0, "step": 23636, "train_loss": 3.493486166000366, "perplexity": 32.9004444237371, "lr": 0.0026291804804649314, "grad_norm": 0.164951, "tokens_per_sec": 108090} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:12.639926+00:00", "epoch": 0, "step": 23637, "train_loss": 3.518150806427002, "perplexity": 33.72201224348967, "lr": 0.0026291804804649314, "grad_norm": 0.183252, "tokens_per_sec": 107586} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:12.944848+00:00", "epoch": 0, "step": 23638, "train_loss": 3.4359638690948486, "perplexity": 31.061337204470032, "lr": 0.0026291804804649314, "grad_norm": 0.156102, "tokens_per_sec": 107463} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:13.249818+00:00", "epoch": 0, "step": 23639, "train_loss": 3.574058771133423, "perplexity": 35.661039814345706, "lr": 0.0026291804804649314, "grad_norm": 0.16361, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:13.553860+00:00", "epoch": 0, "step": 23640, "train_loss": 3.4577975273132324, "perplexity": 31.74697760655661, "lr": 0.0026291804804649314, "grad_norm": 0.156865, "tokens_per_sec": 107832} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:13.857859+00:00", "epoch": 0, "step": 23641, "train_loss": 3.5591394901275635, "perplexity": 35.132951881856954, "lr": 0.0026291804804649314, "grad_norm": 0.151088, "tokens_per_sec": 107734} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:14.162311+00:00", "epoch": 0, "step": 23642, "train_loss": 3.511272430419922, "perplexity": 33.49085546655226, "lr": 0.0026291804804649314, "grad_norm": 0.170612, "tokens_per_sec": 107629} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:14.467563+00:00", "epoch": 0, "step": 23643, "train_loss": 3.4628286361694336, "perplexity": 31.90710257205906, "lr": 0.0026291804804649314, "grad_norm": 0.180467, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:14.773197+00:00", "epoch": 0, "step": 23644, "train_loss": 3.4592320919036865, "perplexity": 31.79255337936369, "lr": 0.0026291804804649314, "grad_norm": 0.168178, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:15.077562+00:00", "epoch": 0, "step": 23645, "train_loss": 3.3921597003936768, "perplexity": 29.73009107453602, "lr": 0.0026291804804649314, "grad_norm": 0.192368, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:15.381990+00:00", "epoch": 0, "step": 23646, "train_loss": 3.533806324005127, "perplexity": 34.25410200154008, "lr": 0.0026291804804649314, "grad_norm": 0.158131, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:15.685753+00:00", "epoch": 0, "step": 23647, "train_loss": 3.477238655090332, "perplexity": 32.370213226818315, "lr": 0.0026291804804649314, "grad_norm": 0.165706, "tokens_per_sec": 107874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:15.990080+00:00", "epoch": 0, "step": 23648, "train_loss": 3.33711576461792, "perplexity": 28.13785336377872, "lr": 0.0026291804804649314, "grad_norm": 0.186226, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:16.295063+00:00", "epoch": 0, "step": 23649, "train_loss": 3.6042051315307617, "perplexity": 36.75245887304852, "lr": 0.0026291804804649314, "grad_norm": 0.172248, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:16.600499+00:00", "epoch": 0, "step": 23650, "train_loss": 3.4368820190429688, "perplexity": 31.089869265963873, "lr": 0.0026291804804649314, "grad_norm": 0.155567, "tokens_per_sec": 107340} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:16.904418+00:00", "epoch": 0, "step": 23651, "train_loss": 3.568032741546631, "perplexity": 35.4467915149576, "lr": 0.0026291804804649314, "grad_norm": 0.17363, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:17.209579+00:00", "epoch": 0, "step": 23652, "train_loss": 3.496169090270996, "perplexity": 32.98883234062115, "lr": 0.0026291804804649314, "grad_norm": 0.167475, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:17.514453+00:00", "epoch": 0, "step": 23653, "train_loss": 3.4557392597198486, "perplexity": 31.68170103273457, "lr": 0.0026291804804649314, "grad_norm": 0.191772, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:17.818902+00:00", "epoch": 0, "step": 23654, "train_loss": 3.4966752529144287, "perplexity": 33.00553428179519, "lr": 0.0026291804804649314, "grad_norm": 0.194422, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:18.123304+00:00", "epoch": 0, "step": 23655, "train_loss": 3.5030641555786133, "perplexity": 33.217078475764765, "lr": 0.0026291804804649314, "grad_norm": 0.19399, "tokens_per_sec": 107647} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:18.427875+00:00", "epoch": 0, "step": 23656, "train_loss": 3.4682068824768066, "perplexity": 32.07916912203019, "lr": 0.0026291804804649314, "grad_norm": 0.166267, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:18.731601+00:00", "epoch": 0, "step": 23657, "train_loss": 3.4481828212738037, "perplexity": 31.443202443817995, "lr": 0.0026291804804649314, "grad_norm": 0.183243, "tokens_per_sec": 107883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:19.035832+00:00", "epoch": 0, "step": 23658, "train_loss": 3.4997239112854004, "perplexity": 33.10631041812381, "lr": 0.0026291804804649314, "grad_norm": 0.172627, "tokens_per_sec": 107708} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:19.340942+00:00", "epoch": 0, "step": 23659, "train_loss": 3.519448757171631, "perplexity": 33.76581017201969, "lr": 0.0026291804804649314, "grad_norm": 0.177198, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:19.646633+00:00", "epoch": 0, "step": 23660, "train_loss": 3.4303314685821533, "perplexity": 30.88687908308867, "lr": 0.0026291804804649314, "grad_norm": 0.165802, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:19.950968+00:00", "epoch": 0, "step": 23661, "train_loss": 3.486049175262451, "perplexity": 32.65667171453897, "lr": 0.0026291804804649314, "grad_norm": 0.204429, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:20.255177+00:00", "epoch": 0, "step": 23662, "train_loss": 3.501403331756592, "perplexity": 33.16195654711077, "lr": 0.0026291804804649314, "grad_norm": 0.239176, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:20.559015+00:00", "epoch": 0, "step": 23663, "train_loss": 3.473351240158081, "perplexity": 32.244621049175926, "lr": 0.0026291804804649314, "grad_norm": 0.1495, "tokens_per_sec": 107847} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:20.864161+00:00", "epoch": 0, "step": 23664, "train_loss": 3.526667594909668, "perplexity": 34.01044199311012, "lr": 0.0026291804804649314, "grad_norm": 0.205668, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:21.169515+00:00", "epoch": 0, "step": 23665, "train_loss": 3.4239048957824707, "perplexity": 30.689018768945306, "lr": 0.0026291804804649314, "grad_norm": 0.188391, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:21.473866+00:00", "epoch": 0, "step": 23666, "train_loss": 3.490875244140625, "perplexity": 32.81465597639656, "lr": 0.0026291804804649314, "grad_norm": 0.169426, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:21.778016+00:00", "epoch": 0, "step": 23667, "train_loss": 3.513932943344116, "perplexity": 33.58007695519692, "lr": 0.0026291804804649314, "grad_norm": 0.188433, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:22.081706+00:00", "epoch": 0, "step": 23668, "train_loss": 3.4852747917175293, "perplexity": 32.63139271441544, "lr": 0.0026291804804649314, "grad_norm": 0.214206, "tokens_per_sec": 107897} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:22.385943+00:00", "epoch": 0, "step": 23669, "train_loss": 3.492356538772583, "perplexity": 32.86330016944023, "lr": 0.0026291804804649314, "grad_norm": 0.172827, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:22.691633+00:00", "epoch": 0, "step": 23670, "train_loss": 3.4661827087402344, "perplexity": 32.01430098474081, "lr": 0.0026291804804649314, "grad_norm": 0.192532, "tokens_per_sec": 107251} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:22.996503+00:00", "epoch": 0, "step": 23671, "train_loss": 3.573029041290283, "perplexity": 35.6243374774004, "lr": 0.0026291804804649314, "grad_norm": 0.207746, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:23.300368+00:00", "epoch": 0, "step": 23672, "train_loss": 3.4730844497680664, "perplexity": 32.2360196415878, "lr": 0.0026291804804649314, "grad_norm": 0.189229, "tokens_per_sec": 107836} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:23.604309+00:00", "epoch": 0, "step": 23673, "train_loss": 3.48112154006958, "perplexity": 32.49614737721868, "lr": 0.0026291804804649314, "grad_norm": 0.200752, "tokens_per_sec": 107820} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:23.908707+00:00", "epoch": 0, "step": 23674, "train_loss": 3.422678232192993, "perplexity": 30.651396746524682, "lr": 0.0026291804804649314, "grad_norm": 0.176467, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:24.214191+00:00", "epoch": 0, "step": 23675, "train_loss": 3.448071241378784, "perplexity": 31.439694210318038, "lr": 0.0026291804804649314, "grad_norm": 0.169851, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:24.518557+00:00", "epoch": 0, "step": 23676, "train_loss": 3.5287981033325195, "perplexity": 34.08297876890981, "lr": 0.0026291804804649314, "grad_norm": 0.170224, "tokens_per_sec": 107660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:24.823407+00:00", "epoch": 0, "step": 23677, "train_loss": 3.418755531311035, "perplexity": 30.53139600286208, "lr": 0.0026291804804649314, "grad_norm": 0.206502, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:25.128133+00:00", "epoch": 0, "step": 23678, "train_loss": 3.4893548488616943, "perplexity": 32.76480263636515, "lr": 0.0026291804804649314, "grad_norm": 0.202338, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:25.432656+00:00", "epoch": 0, "step": 23679, "train_loss": 3.573269844055176, "perplexity": 35.63291694930125, "lr": 0.0026291804804649314, "grad_norm": 0.164263, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:25.737004+00:00", "epoch": 0, "step": 23680, "train_loss": 3.44295072555542, "perplexity": 31.27911822542225, "lr": 0.0026291804804649314, "grad_norm": 0.164294, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:26.042143+00:00", "epoch": 0, "step": 23681, "train_loss": 3.503492832183838, "perplexity": 33.23132091268289, "lr": 0.0026291804804649314, "grad_norm": 0.161576, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:26.346501+00:00", "epoch": 0, "step": 23682, "train_loss": 3.5461580753326416, "perplexity": 34.6798239446629, "lr": 0.0026291804804649314, "grad_norm": 0.175297, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:26.650499+00:00", "epoch": 0, "step": 23683, "train_loss": 3.4671199321746826, "perplexity": 32.04431960273998, "lr": 0.0026291804804649314, "grad_norm": 0.173483, "tokens_per_sec": 107790} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:26.954733+00:00", "epoch": 0, "step": 23684, "train_loss": 3.485264778137207, "perplexity": 32.63106595897946, "lr": 0.0026291804804649314, "grad_norm": 0.170312, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:27.259705+00:00", "epoch": 0, "step": 23685, "train_loss": 3.4640378952026367, "perplexity": 31.945709862470355, "lr": 0.0026291804804649314, "grad_norm": 0.152823, "tokens_per_sec": 107451} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:27.565516+00:00", "epoch": 0, "step": 23686, "train_loss": 3.569483995437622, "perplexity": 35.498271155045565, "lr": 0.0026291804804649314, "grad_norm": 0.157985, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:27.870551+00:00", "epoch": 0, "step": 23687, "train_loss": 3.5660181045532227, "perplexity": 35.37545098423962, "lr": 0.0026291804804649314, "grad_norm": 0.14897, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:28.175087+00:00", "epoch": 0, "step": 23688, "train_loss": 3.4336109161376953, "perplexity": 30.988337255644034, "lr": 0.0026291804804649314, "grad_norm": 0.156719, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:28.481012+00:00", "epoch": 0, "step": 23689, "train_loss": 3.4903857707977295, "perplexity": 32.79859800733433, "lr": 0.0026291804804649314, "grad_norm": 0.157522, "tokens_per_sec": 107111} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:28.785669+00:00", "epoch": 0, "step": 23690, "train_loss": 3.5927867889404297, "perplexity": 36.335193479733086, "lr": 0.0026291804804649314, "grad_norm": 0.153542, "tokens_per_sec": 107624} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:29.089864+00:00", "epoch": 0, "step": 23691, "train_loss": 3.562157392501831, "perplexity": 35.23913985262648, "lr": 0.0026291804804649314, "grad_norm": 0.144207, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:29.394423+00:00", "epoch": 0, "step": 23692, "train_loss": 3.4805636405944824, "perplexity": 32.47802284995647, "lr": 0.0026291804804649314, "grad_norm": 0.166871, "tokens_per_sec": 107591} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:29.699403+00:00", "epoch": 0, "step": 23693, "train_loss": 3.487278699874878, "perplexity": 32.69684859033938, "lr": 0.0026291804804649314, "grad_norm": 0.160326, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:30.004432+00:00", "epoch": 0, "step": 23694, "train_loss": 3.4511468410491943, "perplexity": 31.536538974991842, "lr": 0.0026291804804649314, "grad_norm": 0.171369, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:30.308957+00:00", "epoch": 0, "step": 23695, "train_loss": 3.4590976238250732, "perplexity": 31.78827858321481, "lr": 0.0026291804804649314, "grad_norm": 0.174926, "tokens_per_sec": 107603} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:30.614423+00:00", "epoch": 0, "step": 23696, "train_loss": 3.4160919189453125, "perplexity": 30.450180410371875, "lr": 0.0026291804804649314, "grad_norm": 0.171595, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:30.919665+00:00", "epoch": 0, "step": 23697, "train_loss": 3.512328624725342, "perplexity": 33.52624700425989, "lr": 0.0026291804804649314, "grad_norm": 0.17183, "tokens_per_sec": 107352} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:31.224645+00:00", "epoch": 0, "step": 23698, "train_loss": 3.4509806632995605, "perplexity": 31.53129873933108, "lr": 0.0026291804804649314, "grad_norm": 0.171067, "tokens_per_sec": 107443} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:31.529160+00:00", "epoch": 0, "step": 23699, "train_loss": 3.4362733364105225, "perplexity": 31.07095116064182, "lr": 0.0026291804804649314, "grad_norm": 0.170977, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:31.833936+00:00", "epoch": 0, "step": 23700, "train_loss": 3.5656442642211914, "perplexity": 35.362228685566166, "lr": 0.0026291804804649314, "grad_norm": 0.181562, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:32.137918+00:00", "epoch": 0, "step": 23701, "train_loss": 3.4064009189605713, "perplexity": 30.156512976351816, "lr": 0.0026291804804649314, "grad_norm": 0.199662, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:32.443268+00:00", "epoch": 0, "step": 23702, "train_loss": 3.452646017074585, "perplexity": 31.583853255502763, "lr": 0.0026291804804649314, "grad_norm": 0.185609, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:32.747940+00:00", "epoch": 0, "step": 23703, "train_loss": 3.3989696502685547, "perplexity": 29.9332424447514, "lr": 0.0026291804804649314, "grad_norm": 0.179322, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:33.052613+00:00", "epoch": 0, "step": 23704, "train_loss": 3.4326629638671875, "perplexity": 30.95897570985401, "lr": 0.0026291804804649314, "grad_norm": 0.200194, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:33.358162+00:00", "epoch": 0, "step": 23705, "train_loss": 3.4902994632720947, "perplexity": 32.79576736365071, "lr": 0.0026291804804649314, "grad_norm": 0.203432, "tokens_per_sec": 107242} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:33.662986+00:00", "epoch": 0, "step": 23706, "train_loss": 3.4020583629608154, "perplexity": 30.025840561483616, "lr": 0.0026291804804649314, "grad_norm": 0.216788, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:33.969204+00:00", "epoch": 0, "step": 23707, "train_loss": 3.600904703140259, "perplexity": 36.63135996342858, "lr": 0.0026291804804649314, "grad_norm": 0.184399, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:34.274883+00:00", "epoch": 0, "step": 23708, "train_loss": 3.472278356552124, "perplexity": 32.21004487529368, "lr": 0.0026291804804649314, "grad_norm": 0.180139, "tokens_per_sec": 107197} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:34.579455+00:00", "epoch": 0, "step": 23709, "train_loss": 3.4444522857666016, "perplexity": 31.3261209846956, "lr": 0.0026291804804649314, "grad_norm": 0.190094, "tokens_per_sec": 107589} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:34.884466+00:00", "epoch": 0, "step": 23710, "train_loss": 3.510960578918457, "perplexity": 33.480412921336004, "lr": 0.0026291804804649314, "grad_norm": 0.172837, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:35.189928+00:00", "epoch": 0, "step": 23711, "train_loss": 3.4186980724334717, "perplexity": 30.52964175351639, "lr": 0.0026291804804649314, "grad_norm": 0.158384, "tokens_per_sec": 107215} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:35.494234+00:00", "epoch": 0, "step": 23712, "train_loss": 3.5721023082733154, "perplexity": 35.5913385206256, "lr": 0.0026291804804649314, "grad_norm": 0.161456, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:35.799826+00:00", "epoch": 0, "step": 23713, "train_loss": 3.4573352336883545, "perplexity": 31.732304573086193, "lr": 0.0026291804804649314, "grad_norm": 0.164769, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:36.105105+00:00", "epoch": 0, "step": 23714, "train_loss": 3.5925862789154053, "perplexity": 36.32790863954535, "lr": 0.0026291804804649314, "grad_norm": 0.152805, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:36.409741+00:00", "epoch": 0, "step": 23715, "train_loss": 3.576519250869751, "perplexity": 35.748891114035956, "lr": 0.0026291804804649314, "grad_norm": 0.162836, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:36.714422+00:00", "epoch": 0, "step": 23716, "train_loss": 3.488325357437134, "perplexity": 32.73108890999731, "lr": 0.0026291804804649314, "grad_norm": 0.193553, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:37.018976+00:00", "epoch": 0, "step": 23717, "train_loss": 3.468376398086548, "perplexity": 32.084607502876096, "lr": 0.0026291804804649314, "grad_norm": 0.204385, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:37.324485+00:00", "epoch": 0, "step": 23718, "train_loss": 3.499377727508545, "perplexity": 33.09485153409983, "lr": 0.0026291804804649314, "grad_norm": 0.184675, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:37.630943+00:00", "epoch": 0, "step": 23719, "train_loss": 3.5413615703582764, "perplexity": 34.513880289533745, "lr": 0.0026291804804649314, "grad_norm": 0.164839, "tokens_per_sec": 106927} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:37.937211+00:00", "epoch": 0, "step": 23720, "train_loss": 3.4354546070098877, "perplexity": 31.045522870285794, "lr": 0.0026291804804649314, "grad_norm": 0.151239, "tokens_per_sec": 107050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:38.241953+00:00", "epoch": 0, "step": 23721, "train_loss": 3.464005470275879, "perplexity": 31.944674041961132, "lr": 0.0026291804804649314, "grad_norm": 0.154552, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:38.545968+00:00", "epoch": 0, "step": 23722, "train_loss": 3.5194549560546875, "perplexity": 33.76601948297701, "lr": 0.0026291804804649314, "grad_norm": 0.179079, "tokens_per_sec": 107784} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:38.850771+00:00", "epoch": 0, "step": 23723, "train_loss": 3.5473453998565674, "perplexity": 34.72102460455459, "lr": 0.0026291804804649314, "grad_norm": 0.178972, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:39.156407+00:00", "epoch": 0, "step": 23724, "train_loss": 3.5171890258789062, "perplexity": 33.68959465987082, "lr": 0.0026291804804649314, "grad_norm": 0.17409, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:39.461625+00:00", "epoch": 0, "step": 23725, "train_loss": 3.442422389984131, "perplexity": 31.262596719459165, "lr": 0.0026291804804649314, "grad_norm": 0.163127, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:39.766307+00:00", "epoch": 0, "step": 23726, "train_loss": 3.4890713691711426, "perplexity": 32.75551579663022, "lr": 0.0026291804804649314, "grad_norm": 0.217084, "tokens_per_sec": 107548} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:40.070927+00:00", "epoch": 0, "step": 23727, "train_loss": 3.472776412963867, "perplexity": 32.226091290344606, "lr": 0.0026291804804649314, "grad_norm": 0.173505, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:40.375406+00:00", "epoch": 0, "step": 23728, "train_loss": 3.4748480319976807, "perplexity": 32.292920673057274, "lr": 0.0026291804804649314, "grad_norm": 0.217013, "tokens_per_sec": 107619} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:40.680246+00:00", "epoch": 0, "step": 23729, "train_loss": 3.476518392562866, "perplexity": 32.34690656968843, "lr": 0.0026291804804649314, "grad_norm": 0.159578, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:40.985822+00:00", "epoch": 0, "step": 23730, "train_loss": 3.41316819190979, "perplexity": 30.36128241473249, "lr": 0.0026291804804649314, "grad_norm": 0.17424, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:41.290159+00:00", "epoch": 0, "step": 23731, "train_loss": 3.416579008102417, "perplexity": 30.46501597590988, "lr": 0.0026291804804649314, "grad_norm": 0.179132, "tokens_per_sec": 107598} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:41.594297+00:00", "epoch": 0, "step": 23732, "train_loss": 3.51177716255188, "perplexity": 33.507763644124296, "lr": 0.0026291804804649314, "grad_norm": 0.166861, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:41.898106+00:00", "epoch": 0, "step": 23733, "train_loss": 3.4330077171325684, "perplexity": 30.969650757845923, "lr": 0.0026291804804649314, "grad_norm": 0.164059, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:42.202529+00:00", "epoch": 0, "step": 23734, "train_loss": 3.387486696243286, "perplexity": 29.5914863380332, "lr": 0.0026291804804649314, "grad_norm": 0.177154, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:42.507432+00:00", "epoch": 0, "step": 23735, "train_loss": 3.5199756622314453, "perplexity": 33.78360623625448, "lr": 0.0026291804804649314, "grad_norm": 0.199389, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:42.813010+00:00", "epoch": 0, "step": 23736, "train_loss": 3.494600296020508, "perplexity": 32.93712022355662, "lr": 0.0026291804804649314, "grad_norm": 0.187939, "tokens_per_sec": 107239} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:43.118323+00:00", "epoch": 0, "step": 23737, "train_loss": 3.5468697547912598, "perplexity": 34.70451364752791, "lr": 0.0026291804804649314, "grad_norm": 0.185943, "tokens_per_sec": 107320} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:43.422699+00:00", "epoch": 0, "step": 23738, "train_loss": 3.4221177101135254, "perplexity": 30.634220776086764, "lr": 0.0026291804804649314, "grad_norm": 0.175234, "tokens_per_sec": 107657} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:43.728353+00:00", "epoch": 0, "step": 23739, "train_loss": 3.44174861907959, "perplexity": 31.24153998589299, "lr": 0.0026291804804649314, "grad_norm": 0.187163, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:44.033498+00:00", "epoch": 0, "step": 23740, "train_loss": 3.5256199836730957, "perplexity": 33.97483092844998, "lr": 0.0026291804804649314, "grad_norm": 0.200131, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:44.339288+00:00", "epoch": 0, "step": 23741, "train_loss": 3.537116050720215, "perplexity": 34.367661539891174, "lr": 0.0026291804804649314, "grad_norm": 0.179087, "tokens_per_sec": 107076} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:44.644149+00:00", "epoch": 0, "step": 23742, "train_loss": 3.393439769744873, "perplexity": 29.76817202085505, "lr": 0.0026291804804649314, "grad_norm": 0.196171, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:44.948635+00:00", "epoch": 0, "step": 23743, "train_loss": 3.473804235458374, "perplexity": 32.25923101984512, "lr": 0.0026291804804649314, "grad_norm": 0.216068, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:45.253795+00:00", "epoch": 0, "step": 23744, "train_loss": 3.566978931427002, "perplexity": 35.4094570025493, "lr": 0.0026291804804649314, "grad_norm": 0.222636, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:45.558586+00:00", "epoch": 0, "step": 23745, "train_loss": 3.453200578689575, "perplexity": 31.601373305696406, "lr": 0.0026291804804649314, "grad_norm": 0.182931, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:45.863254+00:00", "epoch": 0, "step": 23746, "train_loss": 3.575349807739258, "perplexity": 35.70710925441598, "lr": 0.0026291804804649314, "grad_norm": 0.175822, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:46.167912+00:00", "epoch": 0, "step": 23747, "train_loss": 3.5112664699554443, "perplexity": 33.49065584609284, "lr": 0.0026291804804649314, "grad_norm": 0.177955, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:46.472446+00:00", "epoch": 0, "step": 23748, "train_loss": 3.424060583114624, "perplexity": 30.693797032351736, "lr": 0.0026291804804649314, "grad_norm": 0.180078, "tokens_per_sec": 107600} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:46.776857+00:00", "epoch": 0, "step": 23749, "train_loss": 3.478278160095215, "perplexity": 32.40387972068957, "lr": 0.0026291804804649314, "grad_norm": 0.213541, "tokens_per_sec": 107644} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:47.082070+00:00", "epoch": 0, "step": 23750, "train_loss": 3.5540947914123535, "perplexity": 34.956163022826615, "lr": 0.0026291804804649314, "grad_norm": 0.177414, "tokens_per_sec": 107431} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:47.386752+00:00", "epoch": 0, "step": 23751, "train_loss": 3.5275068283081055, "perplexity": 34.03899667227062, "lr": 0.0026291804804649314, "grad_norm": 0.167814, "tokens_per_sec": 107479} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:47.691210+00:00", "epoch": 0, "step": 23752, "train_loss": 3.483783483505249, "perplexity": 32.58276551856279, "lr": 0.0026291804804649314, "grad_norm": 0.185469, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:47.995951+00:00", "epoch": 0, "step": 23753, "train_loss": 3.4868390560150146, "perplexity": 32.682476781090095, "lr": 0.0026291804804649314, "grad_norm": 0.161007, "tokens_per_sec": 107529} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:48.300720+00:00", "epoch": 0, "step": 23754, "train_loss": 3.482701301574707, "perplexity": 32.54752411071822, "lr": 0.0026291804804649314, "grad_norm": 0.166375, "tokens_per_sec": 107517} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:48.605135+00:00", "epoch": 0, "step": 23755, "train_loss": 3.5082998275756836, "perplexity": 33.39144827658409, "lr": 0.0026291804804649314, "grad_norm": 0.173936, "tokens_per_sec": 107642} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:48.910051+00:00", "epoch": 0, "step": 23756, "train_loss": 3.553184986114502, "perplexity": 34.924374183535846, "lr": 0.0026291804804649314, "grad_norm": 0.172541, "tokens_per_sec": 107466} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:49.213766+00:00", "epoch": 0, "step": 23757, "train_loss": 3.5097289085388184, "perplexity": 33.43920147311262, "lr": 0.0026291804804649314, "grad_norm": 0.160981, "tokens_per_sec": 107891} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:49.518677+00:00", "epoch": 0, "step": 23758, "train_loss": 3.385894298553467, "perplexity": 29.544402421659367, "lr": 0.0026291804804649314, "grad_norm": 0.174071, "tokens_per_sec": 107468} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:49.823291+00:00", "epoch": 0, "step": 23759, "train_loss": 3.484064817428589, "perplexity": 32.591933445383326, "lr": 0.0026291804804649314, "grad_norm": 0.164508, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:50.129380+00:00", "epoch": 0, "step": 23760, "train_loss": 3.5187816619873047, "perplexity": 33.74329267415131, "lr": 0.0026291804804649314, "grad_norm": 0.163493, "tokens_per_sec": 107110} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:50.435177+00:00", "epoch": 0, "step": 23761, "train_loss": 3.492889165878296, "perplexity": 32.88080871624304, "lr": 0.0026291804804649314, "grad_norm": 0.167176, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:50.739707+00:00", "epoch": 0, "step": 23762, "train_loss": 3.394535541534424, "perplexity": 29.800809022068965, "lr": 0.0026291804804649314, "grad_norm": 0.156287, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:51.044062+00:00", "epoch": 0, "step": 23763, "train_loss": 3.4566352367401123, "perplexity": 31.710099829257487, "lr": 0.0026291804804649314, "grad_norm": 0.169089, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:51.349021+00:00", "epoch": 0, "step": 23764, "train_loss": 3.4612510204315186, "perplexity": 31.856805110355683, "lr": 0.0026291804804649314, "grad_norm": 0.187431, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:51.654674+00:00", "epoch": 0, "step": 23765, "train_loss": 3.498246669769287, "perplexity": 33.05744050714669, "lr": 0.0026291804804649314, "grad_norm": 0.181084, "tokens_per_sec": 107206} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:51.959791+00:00", "epoch": 0, "step": 23766, "train_loss": 3.452763795852661, "perplexity": 31.587573382217823, "lr": 0.0026291804804649314, "grad_norm": 0.170149, "tokens_per_sec": 107396} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:52.264103+00:00", "epoch": 0, "step": 23767, "train_loss": 3.3523201942443848, "perplexity": 28.56894230108784, "lr": 0.0026291804804649314, "grad_norm": 0.156697, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:52.569439+00:00", "epoch": 0, "step": 23768, "train_loss": 3.4146921634674072, "perplexity": 30.40758742037921, "lr": 0.0026291804804649314, "grad_norm": 0.167656, "tokens_per_sec": 107319} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:52.875451+00:00", "epoch": 0, "step": 23769, "train_loss": 3.5052976608276367, "perplexity": 33.29135190907282, "lr": 0.0026291804804649314, "grad_norm": 0.167135, "tokens_per_sec": 107080} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:53.181186+00:00", "epoch": 0, "step": 23770, "train_loss": 3.4569835662841797, "perplexity": 31.72114731784497, "lr": 0.0026291804804649314, "grad_norm": 0.168587, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:53.486819+00:00", "epoch": 0, "step": 23771, "train_loss": 3.5133004188537598, "perplexity": 33.55884345019861, "lr": 0.0026291804804649314, "grad_norm": 0.177485, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:53.791158+00:00", "epoch": 0, "step": 23772, "train_loss": 3.4425413608551025, "perplexity": 31.26631627907495, "lr": 0.0026291804804649314, "grad_norm": 0.18582, "tokens_per_sec": 107669} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:54.096267+00:00", "epoch": 0, "step": 23773, "train_loss": 3.4363839626312256, "perplexity": 31.07438861267503, "lr": 0.0026291804804649314, "grad_norm": 0.172008, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:54.401721+00:00", "epoch": 0, "step": 23774, "train_loss": 3.3638973236083984, "perplexity": 28.90161059823903, "lr": 0.0026291804804649314, "grad_norm": 0.213796, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:54.706677+00:00", "epoch": 0, "step": 23775, "train_loss": 3.4863975048065186, "perplexity": 32.66804897951178, "lr": 0.0026291804804649314, "grad_norm": 0.155642, "tokens_per_sec": 107453} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:55.011281+00:00", "epoch": 0, "step": 23776, "train_loss": 3.542085647583008, "perplexity": 34.53888005397627, "lr": 0.0026291804804649314, "grad_norm": 0.195471, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:55.315098+00:00", "epoch": 0, "step": 23777, "train_loss": 3.4934327602386475, "perplexity": 32.89868739735984, "lr": 0.0026291804804649314, "grad_norm": 0.160595, "tokens_per_sec": 107854} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:55.620969+00:00", "epoch": 0, "step": 23778, "train_loss": 3.4277594089508057, "perplexity": 30.807538266656138, "lr": 0.0026291804804649314, "grad_norm": 0.189649, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:55.926043+00:00", "epoch": 0, "step": 23779, "train_loss": 3.415710210800171, "perplexity": 30.438559546518494, "lr": 0.0026291804804649314, "grad_norm": 0.17411, "tokens_per_sec": 107411} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:56.230745+00:00", "epoch": 0, "step": 23780, "train_loss": 3.552683115005493, "perplexity": 34.90685104667878, "lr": 0.0026291804804649314, "grad_norm": 0.177714, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:56.535239+00:00", "epoch": 0, "step": 23781, "train_loss": 3.47213077545166, "perplexity": 32.20529163217805, "lr": 0.0026291804804649314, "grad_norm": 0.176543, "tokens_per_sec": 107556} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:56.839791+00:00", "epoch": 0, "step": 23782, "train_loss": 3.43656063079834, "perplexity": 31.07987895292532, "lr": 0.0026291804804649314, "grad_norm": 0.172661, "tokens_per_sec": 107592} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:57.145183+00:00", "epoch": 0, "step": 23783, "train_loss": 3.455441951751709, "perplexity": 31.672283210639463, "lr": 0.0026291804804649314, "grad_norm": 0.185299, "tokens_per_sec": 107298} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:57.451822+00:00", "epoch": 0, "step": 23784, "train_loss": 3.413041353225708, "perplexity": 30.357431673840576, "lr": 0.0026291804804649314, "grad_norm": 0.158315, "tokens_per_sec": 106862} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:57.757996+00:00", "epoch": 0, "step": 23785, "train_loss": 3.611668586730957, "perplexity": 37.02778536878249, "lr": 0.0026291804804649314, "grad_norm": 0.167875, "tokens_per_sec": 107024} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:58.061689+00:00", "epoch": 0, "step": 23786, "train_loss": 3.6005330085754395, "perplexity": 36.61774681615139, "lr": 0.0026291804804649314, "grad_norm": 0.170518, "tokens_per_sec": 107899} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:58.365676+00:00", "epoch": 0, "step": 23787, "train_loss": 3.4215738773345947, "perplexity": 30.6175654119487, "lr": 0.0026291804804649314, "grad_norm": 0.175711, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:58.670852+00:00", "epoch": 0, "step": 23788, "train_loss": 3.482603073120117, "perplexity": 32.54432717474171, "lr": 0.0026291804804649314, "grad_norm": 0.153092, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:58.976112+00:00", "epoch": 0, "step": 23789, "train_loss": 3.374619483947754, "perplexity": 29.21316558760051, "lr": 0.0026291804804649314, "grad_norm": 0.151911, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:59.281040+00:00", "epoch": 0, "step": 23790, "train_loss": 3.5017073154449463, "perplexity": 33.172038773313524, "lr": 0.0026291804804649314, "grad_norm": 0.146936, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:59.585317+00:00", "epoch": 0, "step": 23791, "train_loss": 3.5037221908569336, "perplexity": 33.238943678492944, "lr": 0.0026291804804649314, "grad_norm": 0.166619, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:02:59.890982+00:00", "epoch": 0, "step": 23792, "train_loss": 3.4402997493743896, "perplexity": 31.196307840718752, "lr": 0.0026291804804649314, "grad_norm": 0.176207, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:00.195352+00:00", "epoch": 0, "step": 23793, "train_loss": 3.448103904724121, "perplexity": 31.44072115267891, "lr": 0.0026291804804649314, "grad_norm": 0.192729, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:00.501160+00:00", "epoch": 0, "step": 23794, "train_loss": 3.303114175796509, "perplexity": 27.19720405156268, "lr": 0.0026291804804649314, "grad_norm": 0.161513, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:00.806587+00:00", "epoch": 0, "step": 23795, "train_loss": 3.4988176822662354, "perplexity": 33.07632210909569, "lr": 0.0026291804804649314, "grad_norm": 0.157373, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:01.110893+00:00", "epoch": 0, "step": 23796, "train_loss": 3.4608209133148193, "perplexity": 31.843106217969154, "lr": 0.0026291804804649314, "grad_norm": 0.166434, "tokens_per_sec": 107681} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:01.415924+00:00", "epoch": 0, "step": 23797, "train_loss": 3.4583828449249268, "perplexity": 31.76556511093396, "lr": 0.0026291804804649314, "grad_norm": 0.158581, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:01.721208+00:00", "epoch": 0, "step": 23798, "train_loss": 3.5038275718688965, "perplexity": 33.242446616582704, "lr": 0.0026291804804649314, "grad_norm": 0.153001, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:02.026311+00:00", "epoch": 0, "step": 23799, "train_loss": 3.4187533855438232, "perplexity": 30.53133048966389, "lr": 0.0026291804804649314, "grad_norm": 0.164973, "tokens_per_sec": 107395} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:02.331446+00:00", "epoch": 0, "step": 23800, "train_loss": 3.631227970123291, "perplexity": 37.7591552746734, "lr": 0.0026291804804649314, "grad_norm": 0.185937, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:02.635805+00:00", "epoch": 0, "step": 23801, "train_loss": 3.4327707290649414, "perplexity": 30.962312189768618, "lr": 0.0026291804804649314, "grad_norm": 0.172194, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:02.939407+00:00", "epoch": 0, "step": 23802, "train_loss": 3.5091001987457275, "perplexity": 33.4181845271455, "lr": 0.0026291804804649314, "grad_norm": 0.175101, "tokens_per_sec": 107930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:03.243976+00:00", "epoch": 0, "step": 23803, "train_loss": 3.4757163524627686, "perplexity": 32.3209734545947, "lr": 0.0026291804804649314, "grad_norm": 0.165764, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:03.549890+00:00", "epoch": 0, "step": 23804, "train_loss": 3.4380080699920654, "perplexity": 31.12489776099182, "lr": 0.0026291804804649314, "grad_norm": 0.168746, "tokens_per_sec": 107115} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:03.856468+00:00", "epoch": 0, "step": 23805, "train_loss": 3.5559637546539307, "perplexity": 35.021555895986054, "lr": 0.0026291804804649314, "grad_norm": 0.165661, "tokens_per_sec": 106883} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:04.161108+00:00", "epoch": 0, "step": 23806, "train_loss": 3.5969979763031006, "perplexity": 36.48853042603274, "lr": 0.0026291804804649314, "grad_norm": 0.168416, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:04.465127+00:00", "epoch": 0, "step": 23807, "train_loss": 3.4535369873046875, "perplexity": 31.61200606830202, "lr": 0.0026291804804649314, "grad_norm": 0.175709, "tokens_per_sec": 107781} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:04.769136+00:00", "epoch": 0, "step": 23808, "train_loss": 3.445580005645752, "perplexity": 31.361468001087186, "lr": 0.0026291804804649314, "grad_norm": 0.185885, "tokens_per_sec": 107786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:05.073661+00:00", "epoch": 0, "step": 23809, "train_loss": 3.5600810050964355, "perplexity": 35.166045658652834, "lr": 0.0026291804804649314, "grad_norm": 0.206844, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:05.378709+00:00", "epoch": 0, "step": 23810, "train_loss": 3.501350164413452, "perplexity": 33.16019346085757, "lr": 0.0026291804804649314, "grad_norm": 0.184101, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:05.683529+00:00", "epoch": 0, "step": 23811, "train_loss": 3.5427346229553223, "perplexity": 34.561302211445664, "lr": 0.0026291804804649314, "grad_norm": 0.156022, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:05.988241+00:00", "epoch": 0, "step": 23812, "train_loss": 3.4872801303863525, "perplexity": 32.696895363589924, "lr": 0.0026291804804649314, "grad_norm": 0.173702, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:06.292339+00:00", "epoch": 0, "step": 23813, "train_loss": 3.4948642253875732, "perplexity": 32.94581444412977, "lr": 0.0026291804804649314, "grad_norm": 0.170188, "tokens_per_sec": 107755} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:06.596308+00:00", "epoch": 0, "step": 23814, "train_loss": 3.469219207763672, "perplexity": 32.111660119068254, "lr": 0.0026291804804649314, "grad_norm": 0.153791, "tokens_per_sec": 107801} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:06.902913+00:00", "epoch": 0, "step": 23815, "train_loss": 3.5651984214782715, "perplexity": 35.34646620658776, "lr": 0.0026291804804649314, "grad_norm": 0.170332, "tokens_per_sec": 106874} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:07.208564+00:00", "epoch": 0, "step": 23816, "train_loss": 3.5399606227874756, "perplexity": 34.465562006376594, "lr": 0.0026291804804649314, "grad_norm": 0.161299, "tokens_per_sec": 107207} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:07.515041+00:00", "epoch": 0, "step": 23817, "train_loss": 3.456803321838379, "perplexity": 31.715430272474816, "lr": 0.0026291804804649314, "grad_norm": 0.15357, "tokens_per_sec": 106919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:07.819818+00:00", "epoch": 0, "step": 23818, "train_loss": 3.4946064949035645, "perplexity": 32.93732439754593, "lr": 0.0026291804804649314, "grad_norm": 0.162607, "tokens_per_sec": 107514} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:08.125466+00:00", "epoch": 0, "step": 23819, "train_loss": 3.5152010917663574, "perplexity": 33.62268848997103, "lr": 0.0026291804804649314, "grad_norm": 0.159178, "tokens_per_sec": 107209} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:08.430306+00:00", "epoch": 0, "step": 23820, "train_loss": 3.5076851844787598, "perplexity": 33.37093075952565, "lr": 0.0026291804804649314, "grad_norm": 0.150964, "tokens_per_sec": 107551} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:08.735551+00:00", "epoch": 0, "step": 23821, "train_loss": 3.6359002590179443, "perplexity": 37.93598974570691, "lr": 0.0026291804804649314, "grad_norm": 0.158541, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:09.041176+00:00", "epoch": 0, "step": 23822, "train_loss": 3.432558059692383, "perplexity": 30.955728154398408, "lr": 0.0026291804804649314, "grad_norm": 0.159588, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:09.346389+00:00", "epoch": 0, "step": 23823, "train_loss": 3.4772000312805176, "perplexity": 32.368962990003595, "lr": 0.0026291804804649314, "grad_norm": 0.174949, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:09.651221+00:00", "epoch": 0, "step": 23824, "train_loss": 3.501692056655884, "perplexity": 33.17153261203283, "lr": 0.0026291804804649314, "grad_norm": 0.171807, "tokens_per_sec": 107495} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:09.956310+00:00", "epoch": 0, "step": 23825, "train_loss": 3.5277280807495117, "perplexity": 34.04652871659825, "lr": 0.0026291804804649314, "grad_norm": 0.183585, "tokens_per_sec": 107405} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:10.261780+00:00", "epoch": 0, "step": 23826, "train_loss": 3.5408012866973877, "perplexity": 34.49454814258241, "lr": 0.0026291804804649314, "grad_norm": 0.158059, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:10.567548+00:00", "epoch": 0, "step": 23827, "train_loss": 3.362943410873413, "perplexity": 28.874054129150764, "lr": 0.0026291804804649314, "grad_norm": 0.171767, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:10.872590+00:00", "epoch": 0, "step": 23828, "train_loss": 3.4150171279907227, "perplexity": 30.417470413254506, "lr": 0.0026291804804649314, "grad_norm": 0.174425, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:11.176226+00:00", "epoch": 0, "step": 23829, "train_loss": 3.562492847442627, "perplexity": 35.25096297915167, "lr": 0.0026291804804649314, "grad_norm": 0.169906, "tokens_per_sec": 107919} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:11.480580+00:00", "epoch": 0, "step": 23830, "train_loss": 3.5806961059570312, "perplexity": 35.89852132606761, "lr": 0.0026291804804649314, "grad_norm": 0.199825, "tokens_per_sec": 107721} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:11.785654+00:00", "epoch": 0, "step": 23831, "train_loss": 3.4293010234832764, "perplexity": 30.855068242395113, "lr": 0.0026291804804649314, "grad_norm": 0.186788, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:12.091618+00:00", "epoch": 0, "step": 23832, "train_loss": 3.426851987838745, "perplexity": 30.77959553585166, "lr": 0.0026291804804649314, "grad_norm": 0.174009, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:12.397433+00:00", "epoch": 0, "step": 23833, "train_loss": 3.5444462299346924, "perplexity": 34.62050823179156, "lr": 0.0026291804804649314, "grad_norm": 0.179029, "tokens_per_sec": 107148} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:12.701852+00:00", "epoch": 0, "step": 23834, "train_loss": 3.584308624267578, "perplexity": 36.0284399170344, "lr": 0.0026291804804649314, "grad_norm": 0.183693, "tokens_per_sec": 107641} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:13.006848+00:00", "epoch": 0, "step": 23835, "train_loss": 3.4961111545562744, "perplexity": 32.986921164404876, "lr": 0.0026291804804649314, "grad_norm": 0.192692, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:13.311641+00:00", "epoch": 0, "step": 23836, "train_loss": 3.455393075942993, "perplexity": 31.670735240013133, "lr": 0.0026291804804649314, "grad_norm": 0.184543, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:13.617245+00:00", "epoch": 0, "step": 23837, "train_loss": 3.4826653003692627, "perplexity": 32.54635238170796, "lr": 0.0026291804804649314, "grad_norm": 0.182126, "tokens_per_sec": 107224} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:13.922825+00:00", "epoch": 0, "step": 23838, "train_loss": 3.450697898864746, "perplexity": 31.52238406989922, "lr": 0.0026291804804649314, "grad_norm": 0.196771, "tokens_per_sec": 107235} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:14.228077+00:00", "epoch": 0, "step": 23839, "train_loss": 3.564971446990967, "perplexity": 35.33844437095295, "lr": 0.0026291804804649314, "grad_norm": 0.18128, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:14.533476+00:00", "epoch": 0, "step": 23840, "train_loss": 3.504544734954834, "perplexity": 33.26629542290196, "lr": 0.0026291804804649314, "grad_norm": 0.173238, "tokens_per_sec": 107355} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:14.838064+00:00", "epoch": 0, "step": 23841, "train_loss": 3.5158073902130127, "perplexity": 33.64308005483727, "lr": 0.0026291804804649314, "grad_norm": 0.198979, "tokens_per_sec": 107522} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:15.143671+00:00", "epoch": 0, "step": 23842, "train_loss": 3.4949207305908203, "perplexity": 32.947676106667394, "lr": 0.0026291804804649314, "grad_norm": 0.178783, "tokens_per_sec": 107223} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:15.448799+00:00", "epoch": 0, "step": 23843, "train_loss": 3.353543519973755, "perplexity": 28.60391280906555, "lr": 0.0026291804804649314, "grad_norm": 0.161541, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:15.754230+00:00", "epoch": 0, "step": 23844, "train_loss": 3.4575083255767822, "perplexity": 31.737797652998985, "lr": 0.0026291804804649314, "grad_norm": 0.172182, "tokens_per_sec": 107293} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:16.058703+00:00", "epoch": 0, "step": 23845, "train_loss": 3.525453567504883, "perplexity": 33.969177437700374, "lr": 0.0026291804804649314, "grad_norm": 0.16806, "tokens_per_sec": 107614} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:16.363931+00:00", "epoch": 0, "step": 23846, "train_loss": 3.519860029220581, "perplexity": 33.77969996199901, "lr": 0.0026291804804649314, "grad_norm": 0.164944, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:16.668377+00:00", "epoch": 0, "step": 23847, "train_loss": 3.381721019744873, "perplexity": 29.421362312036326, "lr": 0.0026291804804649314, "grad_norm": 0.177027, "tokens_per_sec": 107637} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:16.973721+00:00", "epoch": 0, "step": 23848, "train_loss": 3.5774683952331543, "perplexity": 35.78283808027209, "lr": 0.0026291804804649314, "grad_norm": 0.188593, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:17.278393+00:00", "epoch": 0, "step": 23849, "train_loss": 3.4660515785217285, "perplexity": 32.010103217690435, "lr": 0.0026291804804649314, "grad_norm": 0.187885, "tokens_per_sec": 107552} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:17.584102+00:00", "epoch": 0, "step": 23850, "train_loss": 3.5229694843292236, "perplexity": 33.88489989508665, "lr": 0.0026291804804649314, "grad_norm": 0.161268, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:17.889019+00:00", "epoch": 0, "step": 23851, "train_loss": 3.42332124710083, "perplexity": 30.67111238962562, "lr": 0.0026291804804649314, "grad_norm": 0.174033, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:18.192937+00:00", "epoch": 0, "step": 23852, "train_loss": 3.346386432647705, "perplexity": 28.39992296455195, "lr": 0.0026291804804649314, "grad_norm": 0.164234, "tokens_per_sec": 107818} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:18.498076+00:00", "epoch": 0, "step": 23853, "train_loss": 3.558126211166382, "perplexity": 35.09737043088309, "lr": 0.0026291804804649314, "grad_norm": 0.19293, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:18.802676+00:00", "epoch": 0, "step": 23854, "train_loss": 3.5175085067749023, "perplexity": 33.70035956125711, "lr": 0.0026291804804649314, "grad_norm": 0.170818, "tokens_per_sec": 107571} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:19.107285+00:00", "epoch": 0, "step": 23855, "train_loss": 3.5071630477905273, "perplexity": 33.35351112036774, "lr": 0.0026291804804649314, "grad_norm": 0.170929, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:19.411656+00:00", "epoch": 0, "step": 23856, "train_loss": 3.5110249519348145, "perplexity": 33.482568225875625, "lr": 0.0026291804804649314, "grad_norm": 0.160147, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:19.715687+00:00", "epoch": 0, "step": 23857, "train_loss": 3.477480411529541, "perplexity": 32.378039880338285, "lr": 0.0026291804804649314, "grad_norm": 0.180158, "tokens_per_sec": 107780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:20.020944+00:00", "epoch": 0, "step": 23858, "train_loss": 3.4837303161621094, "perplexity": 32.58103322553914, "lr": 0.0026291804804649314, "grad_norm": 0.24693, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:20.325590+00:00", "epoch": 0, "step": 23859, "train_loss": 3.52042293548584, "perplexity": 33.79872011952687, "lr": 0.0026291804804649314, "grad_norm": 0.250988, "tokens_per_sec": 107560} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:20.630194+00:00", "epoch": 0, "step": 23860, "train_loss": 3.4768733978271484, "perplexity": 32.35839193036504, "lr": 0.0026291804804649314, "grad_norm": 0.201261, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:20.934724+00:00", "epoch": 0, "step": 23861, "train_loss": 3.465392827987671, "perplexity": 31.989023489004104, "lr": 0.0026291804804649314, "grad_norm": 0.155864, "tokens_per_sec": 107532} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:21.238958+00:00", "epoch": 0, "step": 23862, "train_loss": 3.4994051456451416, "perplexity": 33.09575894569955, "lr": 0.0026291804804649314, "grad_norm": 0.203733, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:21.543188+00:00", "epoch": 0, "step": 23863, "train_loss": 3.506702423095703, "perplexity": 33.33815120733582, "lr": 0.0026291804804649314, "grad_norm": 0.170756, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:21.848202+00:00", "epoch": 0, "step": 23864, "train_loss": 3.413834571838379, "perplexity": 30.381521306586595, "lr": 0.0026291804804649314, "grad_norm": 0.171575, "tokens_per_sec": 107433} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:22.153509+00:00", "epoch": 0, "step": 23865, "train_loss": 3.4474499225616455, "perplexity": 31.42016620388722, "lr": 0.0026291804804649314, "grad_norm": 0.160595, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:22.457708+00:00", "epoch": 0, "step": 23866, "train_loss": 3.4983603954315186, "perplexity": 33.06120020024278, "lr": 0.0026291804804649314, "grad_norm": 0.171025, "tokens_per_sec": 107720} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:22.762711+00:00", "epoch": 0, "step": 23867, "train_loss": 3.4709131717681885, "perplexity": 32.16610221389118, "lr": 0.0026291804804649314, "grad_norm": 0.148988, "tokens_per_sec": 107435} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:23.067968+00:00", "epoch": 0, "step": 23868, "train_loss": 3.5134267807006836, "perplexity": 33.56308427557123, "lr": 0.0026291804804649314, "grad_norm": 0.164119, "tokens_per_sec": 107345} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:23.373738+00:00", "epoch": 0, "step": 23869, "train_loss": 3.3933658599853516, "perplexity": 29.765971943724182, "lr": 0.0026291804804649314, "grad_norm": 0.183115, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:23.680110+00:00", "epoch": 0, "step": 23870, "train_loss": 3.4858784675598145, "perplexity": 32.65109744493374, "lr": 0.0026291804804649314, "grad_norm": 0.152642, "tokens_per_sec": 107016} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:23.986006+00:00", "epoch": 0, "step": 23871, "train_loss": 3.522597074508667, "perplexity": 33.87228317503818, "lr": 0.0026291804804649314, "grad_norm": 0.179922, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:24.290369+00:00", "epoch": 0, "step": 23872, "train_loss": 3.5286002159118652, "perplexity": 34.076234843444034, "lr": 0.0026291804804649314, "grad_norm": 0.167557, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:24.595097+00:00", "epoch": 0, "step": 23873, "train_loss": 3.4654369354248047, "perplexity": 31.99043447396385, "lr": 0.0026291804804649314, "grad_norm": 0.184826, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:24.899709+00:00", "epoch": 0, "step": 23874, "train_loss": 3.4747984409332275, "perplexity": 32.2913192724547, "lr": 0.0026291804804649314, "grad_norm": 0.199539, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:25.204906+00:00", "epoch": 0, "step": 23875, "train_loss": 3.4727580547332764, "perplexity": 32.22549968176011, "lr": 0.0026291804804649314, "grad_norm": 0.198043, "tokens_per_sec": 107367} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:25.509812+00:00", "epoch": 0, "step": 23876, "train_loss": 3.4434146881103516, "perplexity": 31.293633932141923, "lr": 0.0026291804804649314, "grad_norm": 0.164433, "tokens_per_sec": 107469} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:25.813973+00:00", "epoch": 0, "step": 23877, "train_loss": 3.5470309257507324, "perplexity": 34.71010745805831, "lr": 0.0026291804804649314, "grad_norm": 0.16521, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:26.118584+00:00", "epoch": 0, "step": 23878, "train_loss": 3.458972215652466, "perplexity": 31.784292323247367, "lr": 0.0026291804804649314, "grad_norm": 0.163318, "tokens_per_sec": 107573} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:26.422770+00:00", "epoch": 0, "step": 23879, "train_loss": 3.4594764709472656, "perplexity": 31.800323762572212, "lr": 0.0026291804804649314, "grad_norm": 0.165809, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:26.727744+00:00", "epoch": 0, "step": 23880, "train_loss": 3.5065276622772217, "perplexity": 33.332325513810474, "lr": 0.0026291804804649314, "grad_norm": 0.170113, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:27.033018+00:00", "epoch": 0, "step": 23881, "train_loss": 3.4728362560272217, "perplexity": 32.22801985607242, "lr": 0.0026291804804649314, "grad_norm": 0.174623, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:27.337576+00:00", "epoch": 0, "step": 23882, "train_loss": 3.4857161045074463, "perplexity": 32.64579654343633, "lr": 0.0026291804804649314, "grad_norm": 0.179811, "tokens_per_sec": 107594} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:27.642009+00:00", "epoch": 0, "step": 23883, "train_loss": 3.3890273571014404, "perplexity": 29.63711192051847, "lr": 0.0026291804804649314, "grad_norm": 0.195715, "tokens_per_sec": 107635} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:27.947107+00:00", "epoch": 0, "step": 23884, "train_loss": 3.434596061706543, "perplexity": 31.018880320989933, "lr": 0.0026291804804649314, "grad_norm": 0.206644, "tokens_per_sec": 107402} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:28.251494+00:00", "epoch": 0, "step": 23885, "train_loss": 3.526820421218872, "perplexity": 34.015640080626476, "lr": 0.0026291804804649314, "grad_norm": 0.187583, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:28.557254+00:00", "epoch": 0, "step": 23886, "train_loss": 3.528613567352295, "perplexity": 34.07668981330086, "lr": 0.0026291804804649314, "grad_norm": 0.183205, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:28.861723+00:00", "epoch": 0, "step": 23887, "train_loss": 3.5517683029174805, "perplexity": 34.87493243937672, "lr": 0.0026291804804649314, "grad_norm": 0.200564, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:29.166414+00:00", "epoch": 0, "step": 23888, "train_loss": 3.3628923892974854, "perplexity": 28.87258096698752, "lr": 0.0026291804804649314, "grad_norm": 0.161698, "tokens_per_sec": 107545} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:29.470400+00:00", "epoch": 0, "step": 23889, "train_loss": 3.4488000869750977, "perplexity": 31.46261724565459, "lr": 0.0026291804804649314, "grad_norm": 0.180738, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:29.774761+00:00", "epoch": 0, "step": 23890, "train_loss": 3.5005133152008057, "perplexity": 33.1324549871474, "lr": 0.0026291804804649314, "grad_norm": 0.173433, "tokens_per_sec": 107739} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:30.079365+00:00", "epoch": 0, "step": 23891, "train_loss": 3.482455015182495, "perplexity": 32.5395090854659, "lr": 0.0026291804804649314, "grad_norm": 0.189934, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:30.384260+00:00", "epoch": 0, "step": 23892, "train_loss": 3.417046546936035, "perplexity": 30.47926288417733, "lr": 0.0026291804804649314, "grad_norm": 0.187726, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:30.688268+00:00", "epoch": 0, "step": 23893, "train_loss": 3.435431957244873, "perplexity": 31.044819704451324, "lr": 0.0026291804804649314, "grad_norm": 0.172982, "tokens_per_sec": 107787} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:30.993453+00:00", "epoch": 0, "step": 23894, "train_loss": 3.505765676498413, "perplexity": 33.306936430083155, "lr": 0.0026291804804649314, "grad_norm": 0.190634, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:31.298478+00:00", "epoch": 0, "step": 23895, "train_loss": 3.530226230621338, "perplexity": 34.13168837445357, "lr": 0.0026291804804649314, "grad_norm": 0.185828, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:31.604642+00:00", "epoch": 0, "step": 23896, "train_loss": 3.5354580879211426, "perplexity": 34.3107284450852, "lr": 0.0026291804804649314, "grad_norm": 0.153996, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:31.910212+00:00", "epoch": 0, "step": 23897, "train_loss": 3.4763858318328857, "perplexity": 32.34261892433391, "lr": 0.0026291804804649314, "grad_norm": 0.203865, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:32.214953+00:00", "epoch": 0, "step": 23898, "train_loss": 3.3859052658081055, "perplexity": 29.54472644442069, "lr": 0.0026291804804649314, "grad_norm": 0.192162, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:32.519909+00:00", "epoch": 0, "step": 23899, "train_loss": 3.529935836791992, "perplexity": 34.121778181763524, "lr": 0.0026291804804649314, "grad_norm": 0.176393, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:32.824842+00:00", "epoch": 0, "step": 23900, "train_loss": 3.4614758491516113, "perplexity": 31.863968240283384, "lr": 0.0026291804804649314, "grad_norm": 0.205819, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:33.130415+00:00", "epoch": 0, "step": 23901, "train_loss": 3.4727587699890137, "perplexity": 32.225522731241895, "lr": 0.0026291804804649314, "grad_norm": 0.211946, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:33.436406+00:00", "epoch": 0, "step": 23902, "train_loss": 3.47159743309021, "perplexity": 32.188119765536754, "lr": 0.0026291804804649314, "grad_norm": 0.208173, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:33.741439+00:00", "epoch": 0, "step": 23903, "train_loss": 3.550729513168335, "perplexity": 34.83872352703608, "lr": 0.0026291804804649314, "grad_norm": 0.187045, "tokens_per_sec": 107424} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:34.046245+00:00", "epoch": 0, "step": 23904, "train_loss": 3.500080108642578, "perplexity": 33.11810489885738, "lr": 0.0026291804804649314, "grad_norm": 0.186025, "tokens_per_sec": 107505} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:34.350788+00:00", "epoch": 0, "step": 23905, "train_loss": 3.5727431774139404, "perplexity": 35.61415522163586, "lr": 0.0026291804804649314, "grad_norm": 0.183623, "tokens_per_sec": 107602} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:34.656005+00:00", "epoch": 0, "step": 23906, "train_loss": 3.4679183959960938, "perplexity": 32.06991605018322, "lr": 0.0026291804804649314, "grad_norm": 0.178563, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:34.962142+00:00", "epoch": 0, "step": 23907, "train_loss": 3.6185107231140137, "perplexity": 37.28200323111774, "lr": 0.0026291804804649314, "grad_norm": 0.172188, "tokens_per_sec": 107037} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:35.267293+00:00", "epoch": 0, "step": 23908, "train_loss": 3.4725024700164795, "perplexity": 32.21726438900263, "lr": 0.0026291804804649314, "grad_norm": 0.158529, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:35.572482+00:00", "epoch": 0, "step": 23909, "train_loss": 3.4812912940979004, "perplexity": 32.50166419737883, "lr": 0.0026291804804649314, "grad_norm": 0.151964, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:35.877863+00:00", "epoch": 0, "step": 23910, "train_loss": 3.473438262939453, "perplexity": 32.24742718788114, "lr": 0.0026291804804649314, "grad_norm": 0.186227, "tokens_per_sec": 107391} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:36.182776+00:00", "epoch": 0, "step": 23911, "train_loss": 3.483896493911743, "perplexity": 32.586447918209565, "lr": 0.0026291804804649314, "grad_norm": 0.162658, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:36.487869+00:00", "epoch": 0, "step": 23912, "train_loss": 3.481536388397217, "perplexity": 32.5096311462789, "lr": 0.0026291804804649314, "grad_norm": 0.155851, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:36.793322+00:00", "epoch": 0, "step": 23913, "train_loss": 3.5586330890655518, "perplexity": 35.11516502173491, "lr": 0.0026291804804649314, "grad_norm": 0.1665, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:37.098831+00:00", "epoch": 0, "step": 23914, "train_loss": 3.535252571105957, "perplexity": 34.30367773799273, "lr": 0.0026291804804649314, "grad_norm": 0.197067, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:37.402486+00:00", "epoch": 0, "step": 23915, "train_loss": 3.587350845336914, "perplexity": 36.138213288813844, "lr": 0.0026291804804649314, "grad_norm": 0.226006, "tokens_per_sec": 107913} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:37.706572+00:00", "epoch": 0, "step": 23916, "train_loss": 3.6870338916778564, "perplexity": 39.926245582682434, "lr": 0.0026291804804649314, "grad_norm": 0.225988, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:38.011593+00:00", "epoch": 0, "step": 23917, "train_loss": 3.505809783935547, "perplexity": 33.3084055460871, "lr": 0.0026291804804649314, "grad_norm": 0.158674, "tokens_per_sec": 107428} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:38.316871+00:00", "epoch": 0, "step": 23918, "train_loss": 3.5974011421203613, "perplexity": 36.50324432009285, "lr": 0.0026291804804649314, "grad_norm": 0.200987, "tokens_per_sec": 107339} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:38.622287+00:00", "epoch": 0, "step": 23919, "train_loss": 3.585787534713745, "perplexity": 36.08176217288863, "lr": 0.0026291804804649314, "grad_norm": 0.169681, "tokens_per_sec": 107290} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:38.926494+00:00", "epoch": 0, "step": 23920, "train_loss": 3.449171543121338, "perplexity": 31.47430639907683, "lr": 0.0026291804804649314, "grad_norm": 0.175043, "tokens_per_sec": 107798} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:39.230834+00:00", "epoch": 0, "step": 23921, "train_loss": 3.511695623397827, "perplexity": 33.50503156080995, "lr": 0.0026291804804649314, "grad_norm": 0.175513, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:39.535752+00:00", "epoch": 0, "step": 23922, "train_loss": 3.551933526992798, "perplexity": 34.8806950938923, "lr": 0.0026291804804649314, "grad_norm": 0.186591, "tokens_per_sec": 107465} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:39.841008+00:00", "epoch": 0, "step": 23923, "train_loss": 3.514528274536133, "perplexity": 33.60007417433221, "lr": 0.0026291804804649314, "grad_norm": 0.188116, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:40.146518+00:00", "epoch": 0, "step": 23924, "train_loss": 3.5284481048583984, "perplexity": 34.07105186566792, "lr": 0.0026291804804649314, "grad_norm": 0.172734, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:40.450922+00:00", "epoch": 0, "step": 23925, "train_loss": 3.468223810195923, "perplexity": 32.079712153790695, "lr": 0.0026291804804649314, "grad_norm": 0.191577, "tokens_per_sec": 107646} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:40.754463+00:00", "epoch": 0, "step": 23926, "train_loss": 3.4739787578582764, "perplexity": 32.264861469567215, "lr": 0.0026291804804649314, "grad_norm": 0.199831, "tokens_per_sec": 107953} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:41.058613+00:00", "epoch": 0, "step": 23927, "train_loss": 3.5300769805908203, "perplexity": 34.126594599054584, "lr": 0.0026291804804649314, "grad_norm": 0.164095, "tokens_per_sec": 107737} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:41.364382+00:00", "epoch": 0, "step": 23928, "train_loss": 3.500443935394287, "perplexity": 33.13015634357155, "lr": 0.0026291804804649314, "grad_norm": 0.184495, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:41.669134+00:00", "epoch": 0, "step": 23929, "train_loss": 3.5177221298217773, "perplexity": 33.70755950375681, "lr": 0.0026291804804649314, "grad_norm": 0.193038, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:41.974609+00:00", "epoch": 0, "step": 23930, "train_loss": 3.5395524501800537, "perplexity": 34.45149697874096, "lr": 0.0026291804804649314, "grad_norm": 0.16969, "tokens_per_sec": 107327} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:42.279489+00:00", "epoch": 0, "step": 23931, "train_loss": 3.5289735794067383, "perplexity": 34.08896004099159, "lr": 0.0026291804804649314, "grad_norm": 0.208753, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:42.584462+00:00", "epoch": 0, "step": 23932, "train_loss": 3.512728214263916, "perplexity": 33.53964641878509, "lr": 0.0026291804804649314, "grad_norm": 0.205574, "tokens_per_sec": 107446} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:42.889250+00:00", "epoch": 0, "step": 23933, "train_loss": 3.5021209716796875, "perplexity": 33.185763432420885, "lr": 0.0026291804804649314, "grad_norm": 0.18348, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:43.194980+00:00", "epoch": 0, "step": 23934, "train_loss": 3.4713237285614014, "perplexity": 32.17931093694622, "lr": 0.0026291804804649314, "grad_norm": 0.182227, "tokens_per_sec": 107180} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:43.499916+00:00", "epoch": 0, "step": 23935, "train_loss": 3.4488162994384766, "perplexity": 31.463127336319396, "lr": 0.0026291804804649314, "grad_norm": 0.191745, "tokens_per_sec": 107458} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:43.803338+00:00", "epoch": 0, "step": 23936, "train_loss": 3.6047544479370117, "perplexity": 36.7726531476929, "lr": 0.0026291804804649314, "grad_norm": 0.177038, "tokens_per_sec": 107995} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:44.107847+00:00", "epoch": 0, "step": 23937, "train_loss": 3.5471673011779785, "perplexity": 34.714841386581085, "lr": 0.0026291804804649314, "grad_norm": 0.193329, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:44.413136+00:00", "epoch": 0, "step": 23938, "train_loss": 3.40395450592041, "perplexity": 30.082827858627176, "lr": 0.0026291804804649314, "grad_norm": 0.181457, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:44.718341+00:00", "epoch": 0, "step": 23939, "train_loss": 3.5164501667022705, "perplexity": 33.66471198722673, "lr": 0.0026291804804649314, "grad_norm": 0.17343, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:45.024126+00:00", "epoch": 0, "step": 23940, "train_loss": 3.5166189670562744, "perplexity": 33.67039508216834, "lr": 0.0026291804804649314, "grad_norm": 0.1776, "tokens_per_sec": 107217} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:45.329009+00:00", "epoch": 0, "step": 23941, "train_loss": 3.4907283782958984, "perplexity": 32.80983697810896, "lr": 0.0026291804804649314, "grad_norm": 0.182546, "tokens_per_sec": 107418} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:45.633149+00:00", "epoch": 0, "step": 23942, "train_loss": 3.478564500808716, "perplexity": 32.413159599269115, "lr": 0.0026291804804649314, "grad_norm": 0.193769, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:45.938494+00:00", "epoch": 0, "step": 23943, "train_loss": 3.5152575969696045, "perplexity": 33.62458840049479, "lr": 0.0026291804804649314, "grad_norm": 0.185918, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:46.243793+00:00", "epoch": 0, "step": 23944, "train_loss": 3.487987518310547, "perplexity": 32.72003293518331, "lr": 0.0026291804804649314, "grad_norm": 0.212013, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:46.548575+00:00", "epoch": 0, "step": 23945, "train_loss": 3.4008870124816895, "perplexity": 29.99069036937197, "lr": 0.0026291804804649314, "grad_norm": 0.182702, "tokens_per_sec": 107515} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:46.853770+00:00", "epoch": 0, "step": 23946, "train_loss": 3.5066909790039062, "perplexity": 33.33776968465616, "lr": 0.0026291804804649314, "grad_norm": 0.17847, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:47.159106+00:00", "epoch": 0, "step": 23947, "train_loss": 3.517179250717163, "perplexity": 33.68926534024354, "lr": 0.0026291804804649314, "grad_norm": 0.199847, "tokens_per_sec": 107318} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:47.463811+00:00", "epoch": 0, "step": 23948, "train_loss": 3.460080146789551, "perplexity": 31.819526645379582, "lr": 0.0026291804804649314, "grad_norm": 0.193371, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:47.768844+00:00", "epoch": 0, "step": 23949, "train_loss": 3.4512290954589844, "perplexity": 31.53913310107974, "lr": 0.0026291804804649314, "grad_norm": 0.179242, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:48.074851+00:00", "epoch": 0, "step": 23950, "train_loss": 3.4423024654388428, "perplexity": 31.258847791561788, "lr": 0.0026291804804649314, "grad_norm": 0.162471, "tokens_per_sec": 107141} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:48.380431+00:00", "epoch": 0, "step": 23951, "train_loss": 3.522963047027588, "perplexity": 33.884681768467196, "lr": 0.0026291804804649314, "grad_norm": 0.173347, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:48.685420+00:00", "epoch": 0, "step": 23952, "train_loss": 3.4953372478485107, "perplexity": 32.96140224075404, "lr": 0.0026291804804649314, "grad_norm": 0.161854, "tokens_per_sec": 107439} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:48.991064+00:00", "epoch": 0, "step": 23953, "train_loss": 3.50993275642395, "perplexity": 33.44601867842624, "lr": 0.0026291804804649314, "grad_norm": 0.178196, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:49.296498+00:00", "epoch": 0, "step": 23954, "train_loss": 3.522143602371216, "perplexity": 33.85692652055707, "lr": 0.0026291804804649314, "grad_norm": 0.190164, "tokens_per_sec": 107283} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:49.603329+00:00", "epoch": 0, "step": 23955, "train_loss": 3.4982521533966064, "perplexity": 33.05762178232758, "lr": 0.0026291804804649314, "grad_norm": 0.169601, "tokens_per_sec": 106795} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:49.908515+00:00", "epoch": 0, "step": 23956, "train_loss": 3.478936195373535, "perplexity": 32.425209633848226, "lr": 0.0026291804804649314, "grad_norm": 0.179725, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:50.212885+00:00", "epoch": 0, "step": 23957, "train_loss": 3.42476224899292, "perplexity": 30.71534137998756, "lr": 0.0026291804804649314, "grad_norm": 0.18036, "tokens_per_sec": 107649} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:50.517881+00:00", "epoch": 0, "step": 23958, "train_loss": 3.455893039703369, "perplexity": 31.686573418825407, "lr": 0.0026291804804649314, "grad_norm": 0.211237, "tokens_per_sec": 107438} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:50.822918+00:00", "epoch": 0, "step": 23959, "train_loss": 3.480097532272339, "perplexity": 32.462888100709634, "lr": 0.0026291804804649314, "grad_norm": 0.181231, "tokens_per_sec": 107423} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:51.128565+00:00", "epoch": 0, "step": 23960, "train_loss": 3.6084611415863037, "perplexity": 36.90921104036589, "lr": 0.0026291804804649314, "grad_norm": 0.180893, "tokens_per_sec": 107271} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:51.433534+00:00", "epoch": 0, "step": 23961, "train_loss": 3.484743118286133, "perplexity": 32.614048081128644, "lr": 0.0026291804804649314, "grad_norm": 0.169857, "tokens_per_sec": 107385} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:51.738183+00:00", "epoch": 0, "step": 23962, "train_loss": 3.50555682182312, "perplexity": 33.29998084706835, "lr": 0.0026291804804649314, "grad_norm": 0.187385, "tokens_per_sec": 107565} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:52.043273+00:00", "epoch": 0, "step": 23963, "train_loss": 3.565356731414795, "perplexity": 35.35206234635984, "lr": 0.0026291804804649314, "grad_norm": 0.194006, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:52.348421+00:00", "epoch": 0, "step": 23964, "train_loss": 3.476351499557495, "perplexity": 32.34150854769514, "lr": 0.0026291804804649314, "grad_norm": 0.1611, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:52.653762+00:00", "epoch": 0, "step": 23965, "train_loss": 3.473325729370117, "perplexity": 32.24379847397768, "lr": 0.0026291804804649314, "grad_norm": 0.181056, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:53.265312+00:00", "epoch": 0, "step": 23966, "train_loss": 3.462007761001587, "perplexity": 31.880921571016675, "lr": 0.0026291804804649314, "grad_norm": 0.212263, "tokens_per_sec": 107130} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:53.570570+00:00", "epoch": 0, "step": 23967, "train_loss": 3.404754638671875, "perplexity": 30.106907746722413, "lr": 0.0026291804804649314, "grad_norm": 0.164603, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:53.874768+00:00", "epoch": 0, "step": 23968, "train_loss": 3.517697334289551, "perplexity": 33.70672371724079, "lr": 0.0026291804804649314, "grad_norm": 0.197871, "tokens_per_sec": 107715} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:54.180151+00:00", "epoch": 0, "step": 23969, "train_loss": 3.492271900177002, "perplexity": 32.86051878357569, "lr": 0.0026291804804649314, "grad_norm": 0.195657, "tokens_per_sec": 107301} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:54.485533+00:00", "epoch": 0, "step": 23970, "train_loss": 3.5035018920898438, "perplexity": 33.231621986690655, "lr": 0.0026291804804649314, "grad_norm": 0.166551, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:54.790239+00:00", "epoch": 0, "step": 23971, "train_loss": 3.451780080795288, "perplexity": 31.556515489218903, "lr": 0.0026291804804649314, "grad_norm": 0.186357, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:55.098154+00:00", "epoch": 0, "step": 23972, "train_loss": 3.4310545921325684, "perplexity": 30.909222190192324, "lr": 0.0026291804804649314, "grad_norm": 0.177805, "tokens_per_sec": 106417} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:55.405374+00:00", "epoch": 0, "step": 23973, "train_loss": 3.5704357624053955, "perplexity": 35.53207332028449, "lr": 0.0026291804804649314, "grad_norm": 0.167928, "tokens_per_sec": 106660} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:55.712004+00:00", "epoch": 0, "step": 23974, "train_loss": 3.504106044769287, "perplexity": 33.25170502615636, "lr": 0.0026291804804649314, "grad_norm": 0.174681, "tokens_per_sec": 106866} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:56.016857+00:00", "epoch": 0, "step": 23975, "train_loss": 3.485901355743408, "perplexity": 32.65184477779911, "lr": 0.0026291804804649314, "grad_norm": 0.177726, "tokens_per_sec": 107487} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:56.322422+00:00", "epoch": 0, "step": 23976, "train_loss": 3.4625208377838135, "perplexity": 31.8972831286795, "lr": 0.0026291804804649314, "grad_norm": 0.167966, "tokens_per_sec": 107237} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:56.627401+00:00", "epoch": 0, "step": 23977, "train_loss": 3.5754075050354004, "perplexity": 35.709169517508265, "lr": 0.0026291804804649314, "grad_norm": 0.175671, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:56.932428+00:00", "epoch": 0, "step": 23978, "train_loss": 3.5447840690612793, "perplexity": 34.63220636998771, "lr": 0.0026291804804649314, "grad_norm": 0.18856, "tokens_per_sec": 107427} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:57.236880+00:00", "epoch": 0, "step": 23979, "train_loss": 3.4430806636810303, "perplexity": 31.28318283948365, "lr": 0.0026291804804649314, "grad_norm": 0.187956, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:57.543029+00:00", "epoch": 0, "step": 23980, "train_loss": 3.4528558254241943, "perplexity": 31.590480506830627, "lr": 0.0026291804804649314, "grad_norm": 0.183896, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:57.848846+00:00", "epoch": 0, "step": 23981, "train_loss": 3.6339380741119385, "perplexity": 37.861625301488125, "lr": 0.0026291804804649314, "grad_norm": 0.174265, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:58.154988+00:00", "epoch": 0, "step": 23982, "train_loss": 3.372396230697632, "perplexity": 29.148289467007277, "lr": 0.0026291804804649314, "grad_norm": 0.177503, "tokens_per_sec": 107036} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:58.459223+00:00", "epoch": 0, "step": 23983, "train_loss": 3.5618507862091064, "perplexity": 35.22833696679848, "lr": 0.0026291804804649314, "grad_norm": 0.169133, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:58.764457+00:00", "epoch": 0, "step": 23984, "train_loss": 3.507704734802246, "perplexity": 33.371583178394516, "lr": 0.0026291804804649314, "grad_norm": 0.172899, "tokens_per_sec": 107354} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:59.069803+00:00", "epoch": 0, "step": 23985, "train_loss": 3.5787441730499268, "perplexity": 35.828518163938995, "lr": 0.0026291804804649314, "grad_norm": 0.163892, "tokens_per_sec": 107314} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:59.375108+00:00", "epoch": 0, "step": 23986, "train_loss": 3.5965328216552734, "perplexity": 36.47156156339131, "lr": 0.0026291804804649314, "grad_norm": 0.165696, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:59.679721+00:00", "epoch": 0, "step": 23987, "train_loss": 3.5450758934020996, "perplexity": 34.642314365590636, "lr": 0.0026291804804649314, "grad_norm": 0.152357, "tokens_per_sec": 107572} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:03:59.985761+00:00", "epoch": 0, "step": 23988, "train_loss": 3.4554238319396973, "perplexity": 31.671709320021115, "lr": 0.0026291804804649314, "grad_norm": 0.149003, "tokens_per_sec": 107071} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:00.291670+00:00", "epoch": 0, "step": 23989, "train_loss": 3.4728481769561768, "perplexity": 32.22840404629744, "lr": 0.0026291804804649314, "grad_norm": 0.1676, "tokens_per_sec": 107117} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:00.596033+00:00", "epoch": 0, "step": 23990, "train_loss": 3.494313955307007, "perplexity": 32.92769033519605, "lr": 0.0026291804804649314, "grad_norm": 0.176256, "tokens_per_sec": 107730} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:00.900592+00:00", "epoch": 0, "step": 23991, "train_loss": 3.4703450202941895, "perplexity": 32.14783218606863, "lr": 0.0026291804804649314, "grad_norm": 0.163763, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:01.206840+00:00", "epoch": 0, "step": 23992, "train_loss": 3.4586141109466553, "perplexity": 31.772912256344924, "lr": 0.0026291804804649314, "grad_norm": 0.1823, "tokens_per_sec": 106997} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:01.512932+00:00", "epoch": 0, "step": 23993, "train_loss": 3.5264675617218018, "perplexity": 34.003639456366734, "lr": 0.0026291804804649314, "grad_norm": 0.161609, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:01.818334+00:00", "epoch": 0, "step": 23994, "train_loss": 3.529839038848877, "perplexity": 34.11847542367276, "lr": 0.0026291804804649314, "grad_norm": 0.173992, "tokens_per_sec": 107295} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:02.123779+00:00", "epoch": 0, "step": 23995, "train_loss": 3.4633872509002686, "perplexity": 31.92493132881353, "lr": 0.0026291804804649314, "grad_norm": 0.161561, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:02.428162+00:00", "epoch": 0, "step": 23996, "train_loss": 3.6398136615753174, "perplexity": 38.0847394138218, "lr": 0.0026291804804649314, "grad_norm": 0.185655, "tokens_per_sec": 107652} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:02.732680+00:00", "epoch": 0, "step": 23997, "train_loss": 3.48397159576416, "perplexity": 32.588895312712694, "lr": 0.0026291804804649314, "grad_norm": 0.184026, "tokens_per_sec": 107607} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:03.038404+00:00", "epoch": 0, "step": 23998, "train_loss": 3.577139377593994, "perplexity": 35.77106683194483, "lr": 0.0026291804804649314, "grad_norm": 0.198909, "tokens_per_sec": 107182} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:03.343278+00:00", "epoch": 0, "step": 23999, "train_loss": 3.4311444759368896, "perplexity": 30.912000553534455, "lr": 0.0026291804804649314, "grad_norm": 0.16788, "tokens_per_sec": 107480} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:03.647980+00:00", "epoch": 0, "step": 24000, "train_loss": 3.5342319011688232, "perplexity": 34.26868286753659, "lr": 0.0026291804804649314, "grad_norm": 0.189206, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T23:04:06.721936+00:00", "step": 24000, "epoch": 0, "val_loss": 3.4516000866889955, "val_ppl": 31.55083601356712, "eval_train_loss": 3.5342319011688232, "eval_train_ppl": 34.26868286753659} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T23:04:07.710874+00:00", "step": 24000, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4516_epoch_0000_step_0024000.pt", "val_loss": 3.4516000866889955} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:09.133494+00:00", "epoch": 0, "step": 24001, "train_loss": 3.453900098800659, "perplexity": 31.623486835389254, "lr": 0.0026291804804649314, "grad_norm": 0.186822, "tokens_per_sec": 5973} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:09.437802+00:00", "epoch": 0, "step": 24002, "train_loss": 3.4377601146698, "perplexity": 31.117181133665955, "lr": 0.0026291804804649314, "grad_norm": 0.176546, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:09.741240+00:00", "epoch": 0, "step": 24003, "train_loss": 3.532440662384033, "perplexity": 34.20735441702805, "lr": 0.0026291804804649314, "grad_norm": 0.176239, "tokens_per_sec": 107988} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:10.046415+00:00", "epoch": 0, "step": 24004, "train_loss": 3.421854257583618, "perplexity": 30.6261511761465, "lr": 0.0026291804804649314, "grad_norm": 0.162769, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:10.350787+00:00", "epoch": 0, "step": 24005, "train_loss": 3.3972764015197754, "perplexity": 29.88260090589556, "lr": 0.0026291804804649314, "grad_norm": 0.176629, "tokens_per_sec": 107658} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:10.654721+00:00", "epoch": 0, "step": 24006, "train_loss": 3.5362513065338135, "perplexity": 34.33795515043929, "lr": 0.0026291804804649314, "grad_norm": 0.162375, "tokens_per_sec": 107814} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:10.959607+00:00", "epoch": 0, "step": 24007, "train_loss": 3.505810499191284, "perplexity": 33.30842937012379, "lr": 0.0026291804804649314, "grad_norm": 0.163593, "tokens_per_sec": 107474} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:11.264299+00:00", "epoch": 0, "step": 24008, "train_loss": 3.5244383811950684, "perplexity": 33.934709892260294, "lr": 0.0026291804804649314, "grad_norm": 0.159339, "tokens_per_sec": 107547} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:11.568725+00:00", "epoch": 0, "step": 24009, "train_loss": 3.3734068870544434, "perplexity": 29.17776326247789, "lr": 0.0026291804804649314, "grad_norm": 0.177027, "tokens_per_sec": 107638} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:11.872860+00:00", "epoch": 0, "step": 24010, "train_loss": 3.3657209873199463, "perplexity": 28.95436550567668, "lr": 0.0026291804804649314, "grad_norm": 0.159722, "tokens_per_sec": 107803} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:12.176905+00:00", "epoch": 0, "step": 24011, "train_loss": 3.591939687728882, "perplexity": 36.30442692634669, "lr": 0.0026291804804649314, "grad_norm": 0.165217, "tokens_per_sec": 107711} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:12.482417+00:00", "epoch": 0, "step": 24012, "train_loss": 3.4573891162872314, "perplexity": 31.73401443819052, "lr": 0.0026291804804649314, "grad_norm": 0.158398, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:12.787183+00:00", "epoch": 0, "step": 24013, "train_loss": 3.481177568435669, "perplexity": 32.497968134266955, "lr": 0.0026291804804649314, "grad_norm": 0.176515, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:13.092366+00:00", "epoch": 0, "step": 24014, "train_loss": 3.4703116416931152, "perplexity": 32.146759154310935, "lr": 0.0026291804804649314, "grad_norm": 0.186847, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:13.396609+00:00", "epoch": 0, "step": 24015, "train_loss": 3.487077236175537, "perplexity": 32.690262025764696, "lr": 0.0026291804804649314, "grad_norm": 0.179315, "tokens_per_sec": 107705} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:13.701059+00:00", "epoch": 0, "step": 24016, "train_loss": 3.5282368659973145, "perplexity": 34.06385549557915, "lr": 0.0026291804804649314, "grad_norm": 0.168154, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:14.005339+00:00", "epoch": 0, "step": 24017, "train_loss": 3.3881471157073975, "perplexity": 29.611035586224286, "lr": 0.0026291804804649314, "grad_norm": 0.191903, "tokens_per_sec": 107690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:14.311362+00:00", "epoch": 0, "step": 24018, "train_loss": 3.4803860187530518, "perplexity": 32.47225455603449, "lr": 0.0026291804804649314, "grad_norm": 0.171381, "tokens_per_sec": 107078} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:14.615709+00:00", "epoch": 0, "step": 24019, "train_loss": 3.4412617683410645, "perplexity": 31.226333720969198, "lr": 0.0026291804804649314, "grad_norm": 0.197867, "tokens_per_sec": 107665} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:14.920745+00:00", "epoch": 0, "step": 24020, "train_loss": 3.5579047203063965, "perplexity": 35.0895975449665, "lr": 0.0026291804804649314, "grad_norm": 0.179505, "tokens_per_sec": 107489} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:15.224783+00:00", "epoch": 0, "step": 24021, "train_loss": 3.415060043334961, "perplexity": 30.418775817478885, "lr": 0.0026291804804649314, "grad_norm": 0.18993, "tokens_per_sec": 107710} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:15.528667+00:00", "epoch": 0, "step": 24022, "train_loss": 3.443615674972534, "perplexity": 31.299924173539, "lr": 0.0026291804804649314, "grad_norm": 0.19715, "tokens_per_sec": 107831} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:15.833024+00:00", "epoch": 0, "step": 24023, "train_loss": 3.530910015106201, "perplexity": 34.15503507455219, "lr": 0.0026291804804649314, "grad_norm": 0.190335, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:16.137489+00:00", "epoch": 0, "step": 24024, "train_loss": 3.4180941581726074, "perplexity": 30.5112100336318, "lr": 0.0026291804804649314, "grad_norm": 0.174172, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:16.442183+00:00", "epoch": 0, "step": 24025, "train_loss": 3.5026955604553223, "perplexity": 33.20483707882683, "lr": 0.0026291804804649314, "grad_norm": 0.161777, "tokens_per_sec": 107543} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:16.746503+00:00", "epoch": 0, "step": 24026, "train_loss": 3.45080828666687, "perplexity": 31.525863948658955, "lr": 0.0026291804804649314, "grad_norm": 0.192069, "tokens_per_sec": 107677} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:17.050699+00:00", "epoch": 0, "step": 24027, "train_loss": 3.5146944522857666, "perplexity": 33.6056582230055, "lr": 0.0026291804804649314, "grad_norm": 0.197481, "tokens_per_sec": 107719} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:17.356178+00:00", "epoch": 0, "step": 24028, "train_loss": 3.473013401031494, "perplexity": 32.23372939448069, "lr": 0.0026291804804649314, "grad_norm": 0.217062, "tokens_per_sec": 107268} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:17.661448+00:00", "epoch": 0, "step": 24029, "train_loss": 3.485563278198242, "perplexity": 32.64080778805582, "lr": 0.0026291804804649314, "grad_norm": 0.264663, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:17.967581+00:00", "epoch": 0, "step": 24030, "train_loss": 3.461493492126465, "perplexity": 31.86453042043303, "lr": 0.0026291804804649314, "grad_norm": 0.180639, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:18.271697+00:00", "epoch": 0, "step": 24031, "train_loss": 3.4804399013519287, "perplexity": 32.47400429264112, "lr": 0.0026291804804649314, "grad_norm": 0.185347, "tokens_per_sec": 107683} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:18.575792+00:00", "epoch": 0, "step": 24032, "train_loss": 3.5065720081329346, "perplexity": 33.333803697083695, "lr": 0.0026291804804649314, "grad_norm": 0.183356, "tokens_per_sec": 107756} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:18.879852+00:00", "epoch": 0, "step": 24033, "train_loss": 3.472203254699707, "perplexity": 32.20762593209181, "lr": 0.0026291804804649314, "grad_norm": 0.154587, "tokens_per_sec": 107768} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:19.185149+00:00", "epoch": 0, "step": 24034, "train_loss": 3.5737974643707275, "perplexity": 35.65172256086116, "lr": 0.0026291804804649314, "grad_norm": 0.178586, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:19.489475+00:00", "epoch": 0, "step": 24035, "train_loss": 3.5157887935638428, "perplexity": 33.64245441209795, "lr": 0.0026291804804649314, "grad_norm": 0.162563, "tokens_per_sec": 107676} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:19.794205+00:00", "epoch": 0, "step": 24036, "train_loss": 3.416728973388672, "perplexity": 30.469585013341405, "lr": 0.0026291804804649314, "grad_norm": 0.171837, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:20.097470+00:00", "epoch": 0, "step": 24037, "train_loss": 3.49067759513855, "perplexity": 32.808170833301524, "lr": 0.0026291804804649314, "grad_norm": 0.155129, "tokens_per_sec": 108050} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:20.401701+00:00", "epoch": 0, "step": 24038, "train_loss": 3.5568840503692627, "perplexity": 35.05380091902385, "lr": 0.0026291804804649314, "grad_norm": 0.174751, "tokens_per_sec": 107709} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:20.705688+00:00", "epoch": 0, "step": 24039, "train_loss": 3.4924521446228027, "perplexity": 32.86644224339189, "lr": 0.0026291804804649314, "grad_norm": 0.154796, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:21.010202+00:00", "epoch": 0, "step": 24040, "train_loss": 3.5299060344696045, "perplexity": 34.12076128868267, "lr": 0.0026291804804649314, "grad_norm": 0.16892, "tokens_per_sec": 107666} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:21.314033+00:00", "epoch": 0, "step": 24041, "train_loss": 3.501006603240967, "perplexity": 33.14880286270798, "lr": 0.0026291804804649314, "grad_norm": 0.185277, "tokens_per_sec": 107799} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:21.618690+00:00", "epoch": 0, "step": 24042, "train_loss": 3.4367740154266357, "perplexity": 31.08651162897356, "lr": 0.0026291804804649314, "grad_norm": 0.197587, "tokens_per_sec": 107550} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:21.922702+00:00", "epoch": 0, "step": 24043, "train_loss": 3.487508773803711, "perplexity": 32.70437214820909, "lr": 0.0026291804804649314, "grad_norm": 0.18271, "tokens_per_sec": 107785} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:22.227193+00:00", "epoch": 0, "step": 24044, "train_loss": 3.4772748947143555, "perplexity": 32.371386332431534, "lr": 0.0026291804804649314, "grad_norm": 0.189593, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:22.531335+00:00", "epoch": 0, "step": 24045, "train_loss": 3.4086387157440186, "perplexity": 30.22407268834996, "lr": 0.0026291804804649314, "grad_norm": 0.168283, "tokens_per_sec": 107740} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:22.835547+00:00", "epoch": 0, "step": 24046, "train_loss": 3.4935522079467773, "perplexity": 32.9026173048745, "lr": 0.0026291804804649314, "grad_norm": 0.178692, "tokens_per_sec": 107724} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:23.140155+00:00", "epoch": 0, "step": 24047, "train_loss": 3.5005369186401367, "perplexity": 33.13323703626806, "lr": 0.0026291804804649314, "grad_norm": 0.159484, "tokens_per_sec": 107567} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:23.445189+00:00", "epoch": 0, "step": 24048, "train_loss": 3.410301685333252, "perplexity": 30.27437621712002, "lr": 0.0026291804804649314, "grad_norm": 0.170965, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:23.750132+00:00", "epoch": 0, "step": 24049, "train_loss": 3.5113108158111572, "perplexity": 33.49214105081582, "lr": 0.0026291804804649314, "grad_norm": 0.169552, "tokens_per_sec": 107454} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:24.055229+00:00", "epoch": 0, "step": 24050, "train_loss": 3.59108567237854, "perplexity": 36.27343562387386, "lr": 0.0026291804804649314, "grad_norm": 0.155209, "tokens_per_sec": 107499} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:24.360880+00:00", "epoch": 0, "step": 24051, "train_loss": 3.4608020782470703, "perplexity": 31.842506456554492, "lr": 0.0026291804804649314, "grad_norm": 0.160283, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:24.665493+00:00", "epoch": 0, "step": 24052, "train_loss": 3.523660182952881, "perplexity": 33.90831223331107, "lr": 0.0026291804804649314, "grad_norm": 0.158586, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:24.969639+00:00", "epoch": 0, "step": 24053, "train_loss": 3.5857176780700684, "perplexity": 36.07924171012185, "lr": 0.0026291804804649314, "grad_norm": 0.159796, "tokens_per_sec": 107732} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:25.275136+00:00", "epoch": 0, "step": 24054, "train_loss": 3.3912596702575684, "perplexity": 29.703345134501188, "lr": 0.0026291804804649314, "grad_norm": 0.173001, "tokens_per_sec": 107262} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:25.579524+00:00", "epoch": 0, "step": 24055, "train_loss": 3.425295829772949, "perplexity": 30.73173486903635, "lr": 0.0026291804804649314, "grad_norm": 0.166992, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:25.884056+00:00", "epoch": 0, "step": 24056, "train_loss": 3.5421154499053955, "perplexity": 34.53990940815305, "lr": 0.0026291804804649314, "grad_norm": 0.16655, "tokens_per_sec": 107601} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:26.203132+00:00", "epoch": 0, "step": 24057, "train_loss": 3.4773976802825928, "perplexity": 32.37536131552728, "lr": 0.0026291804804649314, "grad_norm": 0.166651, "tokens_per_sec": 102697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:26.507449+00:00", "epoch": 0, "step": 24058, "train_loss": 3.4302966594696045, "perplexity": 30.885803956950586, "lr": 0.0026291804804649314, "grad_norm": 0.162822, "tokens_per_sec": 107678} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:26.812071+00:00", "epoch": 0, "step": 24059, "train_loss": 3.5198168754577637, "perplexity": 33.77824227229143, "lr": 0.0026291804804649314, "grad_norm": 0.199569, "tokens_per_sec": 107570} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:27.117160+00:00", "epoch": 0, "step": 24060, "train_loss": 3.4970643520355225, "perplexity": 33.018379204987575, "lr": 0.0026291804804649314, "grad_norm": 0.191184, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:27.422593+00:00", "epoch": 0, "step": 24061, "train_loss": 3.4628775119781494, "perplexity": 31.908662095612225, "lr": 0.0026291804804649314, "grad_norm": 0.157728, "tokens_per_sec": 107213} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:27.727718+00:00", "epoch": 0, "step": 24062, "train_loss": 3.438011884689331, "perplexity": 31.12501649328067, "lr": 0.0026291804804649314, "grad_norm": 0.168402, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:28.033206+00:00", "epoch": 0, "step": 24063, "train_loss": 3.41520619392395, "perplexity": 30.42322186436921, "lr": 0.0026291804804649314, "grad_norm": 0.179666, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:28.338461+00:00", "epoch": 0, "step": 24064, "train_loss": 3.477247953414917, "perplexity": 32.37051421696713, "lr": 0.0026291804804649314, "grad_norm": 0.151292, "tokens_per_sec": 107346} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:28.642269+00:00", "epoch": 0, "step": 24065, "train_loss": 3.5442397594451904, "perplexity": 34.613360856396575, "lr": 0.0026291804804649314, "grad_norm": 0.178813, "tokens_per_sec": 107858} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:28.948196+00:00", "epoch": 0, "step": 24066, "train_loss": 3.4404656887054443, "perplexity": 31.201484964705582, "lr": 0.0026291804804649314, "grad_norm": 0.18728, "tokens_per_sec": 107112} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:29.254819+00:00", "epoch": 0, "step": 24067, "train_loss": 3.471984386444092, "perplexity": 32.20057747655627, "lr": 0.0026291804804649314, "grad_norm": 0.16382, "tokens_per_sec": 106867} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:29.559582+00:00", "epoch": 0, "step": 24068, "train_loss": 3.5303194522857666, "perplexity": 34.13487033556515, "lr": 0.0026291804804649314, "grad_norm": 0.215573, "tokens_per_sec": 107520} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:29.865139+00:00", "epoch": 0, "step": 24069, "train_loss": 3.4577345848083496, "perplexity": 31.74497943514921, "lr": 0.0026291804804649314, "grad_norm": 0.223777, "tokens_per_sec": 107241} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:30.169564+00:00", "epoch": 0, "step": 24070, "train_loss": 3.4894919395446777, "perplexity": 32.76929469343893, "lr": 0.0026291804804649314, "grad_norm": 0.24041, "tokens_per_sec": 107697} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:30.474260+00:00", "epoch": 0, "step": 24071, "train_loss": 3.462221384048462, "perplexity": 31.887732798111497, "lr": 0.0026291804804649314, "grad_norm": 0.240789, "tokens_per_sec": 107485} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:30.779560+00:00", "epoch": 0, "step": 24072, "train_loss": 3.5181145668029785, "perplexity": 33.72079019258812, "lr": 0.0026291804804649314, "grad_norm": 0.198212, "tokens_per_sec": 107331} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:31.083547+00:00", "epoch": 0, "step": 24073, "train_loss": 3.460613965988159, "perplexity": 31.836517054093232, "lr": 0.0026291804804649314, "grad_norm": 0.173907, "tokens_per_sec": 107794} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:31.387719+00:00", "epoch": 0, "step": 24074, "train_loss": 3.647892475128174, "perplexity": 38.39366511911736, "lr": 0.0026291804804649314, "grad_norm": 0.15523, "tokens_per_sec": 107729} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:31.692226+00:00", "epoch": 0, "step": 24075, "train_loss": 3.5348637104034424, "perplexity": 34.29034097900749, "lr": 0.0026291804804649314, "grad_norm": 0.174167, "tokens_per_sec": 107609} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:31.996749+00:00", "epoch": 0, "step": 24076, "train_loss": 3.4298412799835205, "perplexity": 30.87174239733327, "lr": 0.0026291804804649314, "grad_norm": 0.176131, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:32.302039+00:00", "epoch": 0, "step": 24077, "train_loss": 3.4413487911224365, "perplexity": 31.229051241623036, "lr": 0.0026291804804649314, "grad_norm": 0.148636, "tokens_per_sec": 107334} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:32.606915+00:00", "epoch": 0, "step": 24078, "train_loss": 3.506868839263916, "perplexity": 33.34369967637972, "lr": 0.0026291804804649314, "grad_norm": 0.166769, "tokens_per_sec": 107482} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:32.911377+00:00", "epoch": 0, "step": 24079, "train_loss": 3.443298816680908, "perplexity": 31.290008104114722, "lr": 0.0026291804804649314, "grad_norm": 0.176162, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:33.216578+00:00", "epoch": 0, "step": 24080, "train_loss": 3.468656063079834, "perplexity": 32.09358169924377, "lr": 0.0026291804804649314, "grad_norm": 0.216243, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:33.521007+00:00", "epoch": 0, "step": 24081, "train_loss": 3.4657070636749268, "perplexity": 31.9990771613134, "lr": 0.0026291804804649314, "grad_norm": 0.171748, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:33.825613+00:00", "epoch": 0, "step": 24082, "train_loss": 3.5826590061187744, "perplexity": 35.96905574274655, "lr": 0.0026291804804649314, "grad_norm": 0.169134, "tokens_per_sec": 107575} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:34.131783+00:00", "epoch": 0, "step": 24083, "train_loss": 3.5413389205932617, "perplexity": 34.51309856710839, "lr": 0.0026291804804649314, "grad_norm": 0.179106, "tokens_per_sec": 107028} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:34.437096+00:00", "epoch": 0, "step": 24084, "train_loss": 3.5116515159606934, "perplexity": 33.503553772327685, "lr": 0.0026291804804649314, "grad_norm": 0.161111, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:34.742222+00:00", "epoch": 0, "step": 24085, "train_loss": 3.518906831741333, "perplexity": 33.74751657814244, "lr": 0.0026291804804649314, "grad_norm": 0.162326, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:35.048195+00:00", "epoch": 0, "step": 24086, "train_loss": 3.507713794708252, "perplexity": 33.37188552317099, "lr": 0.0026291804804649314, "grad_norm": 0.147438, "tokens_per_sec": 107093} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:35.352428+00:00", "epoch": 0, "step": 24087, "train_loss": 3.508206367492676, "perplexity": 33.388327654885465, "lr": 0.0026291804804649314, "grad_norm": 0.165021, "tokens_per_sec": 107707} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:35.657803+00:00", "epoch": 0, "step": 24088, "train_loss": 3.4645421504974365, "perplexity": 31.96182271796957, "lr": 0.0026291804804649314, "grad_norm": 0.151643, "tokens_per_sec": 107304} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:35.963554+00:00", "epoch": 0, "step": 24089, "train_loss": 3.589423656463623, "perplexity": 36.21319866783409, "lr": 0.0026291804804649314, "grad_norm": 0.160973, "tokens_per_sec": 107173} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:36.268765+00:00", "epoch": 0, "step": 24090, "train_loss": 3.596447229385376, "perplexity": 36.46844001324258, "lr": 0.0026291804804649314, "grad_norm": 0.15316, "tokens_per_sec": 107425} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:36.573928+00:00", "epoch": 0, "step": 24091, "train_loss": 3.5662410259246826, "perplexity": 35.38333780732709, "lr": 0.0026291804804649314, "grad_norm": 0.175663, "tokens_per_sec": 107316} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:36.878818+00:00", "epoch": 0, "step": 24092, "train_loss": 3.4081263542175293, "perplexity": 30.208591002779343, "lr": 0.0026291804804649314, "grad_norm": 0.179277, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:37.184124+00:00", "epoch": 0, "step": 24093, "train_loss": 3.5302093029022217, "perplexity": 34.13111060770996, "lr": 0.0026291804804649314, "grad_norm": 0.199542, "tokens_per_sec": 107329} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:37.489390+00:00", "epoch": 0, "step": 24094, "train_loss": 3.4578006267547607, "perplexity": 31.74707600460989, "lr": 0.0026291804804649314, "grad_norm": 0.162714, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:37.794058+00:00", "epoch": 0, "step": 24095, "train_loss": 3.433232307434082, "perplexity": 30.97660702217291, "lr": 0.0026291804804649314, "grad_norm": 0.162262, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:38.099720+00:00", "epoch": 0, "step": 24096, "train_loss": 3.5183475017547607, "perplexity": 33.728645858119705, "lr": 0.0026291804804649314, "grad_norm": 0.171477, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:38.404608+00:00", "epoch": 0, "step": 24097, "train_loss": 3.531550168991089, "perplexity": 34.176906552752605, "lr": 0.0026291804804649314, "grad_norm": 0.167631, "tokens_per_sec": 107467} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:38.710094+00:00", "epoch": 0, "step": 24098, "train_loss": 3.427166223526001, "perplexity": 30.789269103018736, "lr": 0.0026291804804649314, "grad_norm": 0.167247, "tokens_per_sec": 107265} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:39.015852+00:00", "epoch": 0, "step": 24099, "train_loss": 3.4050848484039307, "perplexity": 30.116850982248703, "lr": 0.0026291804804649314, "grad_norm": 0.16886, "tokens_per_sec": 107170} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:39.320689+00:00", "epoch": 0, "step": 24100, "train_loss": 3.604105234146118, "perplexity": 36.74878758190713, "lr": 0.0026291804804649314, "grad_norm": 0.193582, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:39.625300+00:00", "epoch": 0, "step": 24101, "train_loss": 3.5149974822998047, "perplexity": 33.61584328920101, "lr": 0.0026291804804649314, "grad_norm": 0.178934, "tokens_per_sec": 107503} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:39.931719+00:00", "epoch": 0, "step": 24102, "train_loss": 3.470754623413086, "perplexity": 32.16100273556274, "lr": 0.0026291804804649314, "grad_norm": 0.169707, "tokens_per_sec": 106944} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:40.236971+00:00", "epoch": 0, "step": 24103, "train_loss": 3.6650352478027344, "perplexity": 39.05751283034451, "lr": 0.0026291804804649314, "grad_norm": 0.173805, "tokens_per_sec": 107342} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:40.542743+00:00", "epoch": 0, "step": 24104, "train_loss": 3.433598756790161, "perplexity": 30.987960459972637, "lr": 0.0026291804804649314, "grad_norm": 0.167046, "tokens_per_sec": 107166} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:40.847725+00:00", "epoch": 0, "step": 24105, "train_loss": 3.3729281425476074, "perplexity": 29.16379791177842, "lr": 0.0026291804804649314, "grad_norm": 0.172108, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:41.152522+00:00", "epoch": 0, "step": 24106, "train_loss": 3.4016594886779785, "perplexity": 30.01386641411156, "lr": 0.0026291804804649314, "grad_norm": 0.15689, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:41.456973+00:00", "epoch": 0, "step": 24107, "train_loss": 3.4636313915252686, "perplexity": 31.932726453015757, "lr": 0.0026291804804649314, "grad_norm": 0.1488, "tokens_per_sec": 107630} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:41.761585+00:00", "epoch": 0, "step": 24108, "train_loss": 3.5133566856384277, "perplexity": 33.56073175154055, "lr": 0.0026291804804649314, "grad_norm": 0.169062, "tokens_per_sec": 107578} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:42.066286+00:00", "epoch": 0, "step": 24109, "train_loss": 3.465216636657715, "perplexity": 31.983387796906204, "lr": 0.0026291804804649314, "grad_norm": 0.183957, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:42.371723+00:00", "epoch": 0, "step": 24110, "train_loss": 3.3971195220947266, "perplexity": 29.8779133083499, "lr": 0.0026291804804649314, "grad_norm": 0.181849, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:42.675370+00:00", "epoch": 0, "step": 24111, "train_loss": 3.524705648422241, "perplexity": 33.943780740194306, "lr": 0.0026291804804649314, "grad_norm": 0.159208, "tokens_per_sec": 107816} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:42.979915+00:00", "epoch": 0, "step": 24112, "train_loss": 3.429027557373047, "perplexity": 30.846631580525127, "lr": 0.0026291804804649314, "grad_norm": 0.167896, "tokens_per_sec": 107597} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:43.285041+00:00", "epoch": 0, "step": 24113, "train_loss": 3.471088409423828, "perplexity": 32.17173942014519, "lr": 0.0026291804804649314, "grad_norm": 0.170017, "tokens_per_sec": 107392} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:43.590994+00:00", "epoch": 0, "step": 24114, "train_loss": 3.459742784500122, "perplexity": 31.808793747561246, "lr": 0.0026291804804649314, "grad_norm": 0.208338, "tokens_per_sec": 107101} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:43.895282+00:00", "epoch": 0, "step": 24115, "train_loss": 3.6147398948669434, "perplexity": 37.14168392649356, "lr": 0.0026291804804649314, "grad_norm": 0.215045, "tokens_per_sec": 107688} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:44.201348+00:00", "epoch": 0, "step": 24116, "train_loss": 3.4852030277252197, "perplexity": 32.62905103942461, "lr": 0.0026291804804649314, "grad_norm": 0.225463, "tokens_per_sec": 107062} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:44.506415+00:00", "epoch": 0, "step": 24117, "train_loss": 3.4769136905670166, "perplexity": 32.359695764901, "lr": 0.0026291804804649314, "grad_norm": 0.199107, "tokens_per_sec": 107412} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:44.811079+00:00", "epoch": 0, "step": 24118, "train_loss": 3.4901864528656006, "perplexity": 32.79206131006491, "lr": 0.0026291804804649314, "grad_norm": 0.149115, "tokens_per_sec": 107555} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:45.116332+00:00", "epoch": 0, "step": 24119, "train_loss": 3.5296061038970947, "perplexity": 34.11052896378513, "lr": 0.0026291804804649314, "grad_norm": 0.188089, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:45.421516+00:00", "epoch": 0, "step": 24120, "train_loss": 3.5441670417785645, "perplexity": 34.61084394507409, "lr": 0.0026291804804649314, "grad_norm": 0.171773, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:45.726792+00:00", "epoch": 0, "step": 24121, "train_loss": 3.503359317779541, "perplexity": 33.22688434884641, "lr": 0.0026291804804649314, "grad_norm": 0.177481, "tokens_per_sec": 107280} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:46.032408+00:00", "epoch": 0, "step": 24122, "train_loss": 3.495549201965332, "perplexity": 32.96838928609553, "lr": 0.0026291804804649314, "grad_norm": 0.169494, "tokens_per_sec": 107220} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:46.337892+00:00", "epoch": 0, "step": 24123, "train_loss": 3.4455454349517822, "perplexity": 31.36038383211482, "lr": 0.0026291804804649314, "grad_norm": 0.176634, "tokens_per_sec": 107266} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:46.643244+00:00", "epoch": 0, "step": 24124, "train_loss": 3.5131518840789795, "perplexity": 33.553859165123846, "lr": 0.0026291804804649314, "grad_norm": 0.164294, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:46.949842+00:00", "epoch": 0, "step": 24125, "train_loss": 3.6007285118103027, "perplexity": 36.62490640394583, "lr": 0.0026291804804649314, "grad_norm": 0.150622, "tokens_per_sec": 106877} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:47.254561+00:00", "epoch": 0, "step": 24126, "train_loss": 3.440248966217041, "perplexity": 31.19472363393483, "lr": 0.0026291804804649314, "grad_norm": 0.193468, "tokens_per_sec": 107535} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:47.559798+00:00", "epoch": 0, "step": 24127, "train_loss": 3.5854082107543945, "perplexity": 36.06807809151491, "lr": 0.0026291804804649314, "grad_norm": 0.187811, "tokens_per_sec": 107353} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:47.864467+00:00", "epoch": 0, "step": 24128, "train_loss": 3.3701746463775635, "perplexity": 29.083605960705686, "lr": 0.0026291804804649314, "grad_norm": 0.182914, "tokens_per_sec": 107553} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:48.168553+00:00", "epoch": 0, "step": 24129, "train_loss": 3.532996416091919, "perplexity": 34.22637056474012, "lr": 0.0026291804804649314, "grad_norm": 0.172815, "tokens_per_sec": 107759} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:48.473516+00:00", "epoch": 0, "step": 24130, "train_loss": 3.4704270362854004, "perplexity": 32.15046893051678, "lr": 0.0026291804804649314, "grad_norm": 0.166909, "tokens_per_sec": 107509} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:48.778465+00:00", "epoch": 0, "step": 24131, "train_loss": 3.5463151931762695, "perplexity": 34.68527319189426, "lr": 0.0026291804804649314, "grad_norm": 0.166688, "tokens_per_sec": 107394} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:49.083258+00:00", "epoch": 0, "step": 24132, "train_loss": 3.459841251373291, "perplexity": 31.811926014230565, "lr": 0.0026291804804649314, "grad_norm": 0.177077, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:49.388387+00:00", "epoch": 0, "step": 24133, "train_loss": 3.4086642265319824, "perplexity": 30.22484373809472, "lr": 0.0026291804804649314, "grad_norm": 0.172341, "tokens_per_sec": 107390} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:49.693557+00:00", "epoch": 0, "step": 24134, "train_loss": 3.6249613761901855, "perplexity": 37.52327383981965, "lr": 0.0026291804804649314, "grad_norm": 0.167261, "tokens_per_sec": 107376} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:49.998499+00:00", "epoch": 0, "step": 24135, "train_loss": 3.551983118057251, "perplexity": 34.88242490758216, "lr": 0.0026291804804649314, "grad_norm": 0.170975, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:50.304333+00:00", "epoch": 0, "step": 24136, "train_loss": 3.4839959144592285, "perplexity": 32.58968784175701, "lr": 0.0026291804804649314, "grad_norm": 0.164257, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:50.609425+00:00", "epoch": 0, "step": 24137, "train_loss": 3.431671380996704, "perplexity": 30.92829253482186, "lr": 0.0026291804804649314, "grad_norm": 0.184154, "tokens_per_sec": 107403} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:50.914768+00:00", "epoch": 0, "step": 24138, "train_loss": 3.4668188095092773, "perplexity": 32.03467178447138, "lr": 0.0026291804804649314, "grad_norm": 0.198317, "tokens_per_sec": 107315} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:51.220144+00:00", "epoch": 0, "step": 24139, "train_loss": 3.4818928241729736, "perplexity": 32.52122080723829, "lr": 0.0026291804804649314, "grad_norm": 0.180835, "tokens_per_sec": 107305} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:51.524844+00:00", "epoch": 0, "step": 24140, "train_loss": 3.4738588333129883, "perplexity": 32.260992352732366, "lr": 0.0026291804804649314, "grad_norm": 0.180962, "tokens_per_sec": 107599} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:51.828284+00:00", "epoch": 0, "step": 24141, "train_loss": 3.517728567123413, "perplexity": 33.70777649018315, "lr": 0.0026291804804649314, "grad_norm": 0.166369, "tokens_per_sec": 107930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:52.134154+00:00", "epoch": 0, "step": 24142, "train_loss": 3.526496648788452, "perplexity": 34.00462853687861, "lr": 0.0026291804804649314, "grad_norm": 0.161911, "tokens_per_sec": 107133} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:52.438758+00:00", "epoch": 0, "step": 24143, "train_loss": 3.4784605503082275, "perplexity": 32.40979041022384, "lr": 0.0026291804804649314, "grad_norm": 0.197594, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:52.744903+00:00", "epoch": 0, "step": 24144, "train_loss": 3.547433376312256, "perplexity": 34.724079371608994, "lr": 0.0026291804804649314, "grad_norm": 0.201264, "tokens_per_sec": 107034} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:53.051084+00:00", "epoch": 0, "step": 24145, "train_loss": 3.4201221466064453, "perplexity": 30.573149199410015, "lr": 0.0026291804804649314, "grad_norm": 0.14699, "tokens_per_sec": 107023} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:53.355868+00:00", "epoch": 0, "step": 24146, "train_loss": 3.508979082107544, "perplexity": 33.41413727408116, "lr": 0.0026291804804649314, "grad_norm": 0.181358, "tokens_per_sec": 107512} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:53.662365+00:00", "epoch": 0, "step": 24147, "train_loss": 3.50368595123291, "perplexity": 33.237739133497364, "lr": 0.0026291804804649314, "grad_norm": 0.175064, "tokens_per_sec": 106914} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:53.967465+00:00", "epoch": 0, "step": 24148, "train_loss": 3.4356369972229004, "perplexity": 31.05118578622972, "lr": 0.0026291804804649314, "grad_norm": 0.212793, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:54.271959+00:00", "epoch": 0, "step": 24149, "train_loss": 3.609551429748535, "perplexity": 36.94947466172683, "lr": 0.0026291804804649314, "grad_norm": 0.180795, "tokens_per_sec": 107615} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:54.577442+00:00", "epoch": 0, "step": 24150, "train_loss": 3.582132339477539, "perplexity": 35.95011702860236, "lr": 0.0026291804804649314, "grad_norm": 0.189465, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:54.881743+00:00", "epoch": 0, "step": 24151, "train_loss": 3.4656054973602295, "perplexity": 31.995827298013527, "lr": 0.0026291804804649314, "grad_norm": 0.205082, "tokens_per_sec": 107623} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:55.187285+00:00", "epoch": 0, "step": 24152, "train_loss": 3.376309871673584, "perplexity": 29.26258892467505, "lr": 0.0026291804804649314, "grad_norm": 0.220021, "tokens_per_sec": 107247} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:55.493528+00:00", "epoch": 0, "step": 24153, "train_loss": 3.532515287399292, "perplexity": 34.209907236624296, "lr": 0.0026291804804649314, "grad_norm": 0.177802, "tokens_per_sec": 107009} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:55.798055+00:00", "epoch": 0, "step": 24154, "train_loss": 3.4327330589294434, "perplexity": 30.96114585724119, "lr": 0.0026291804804649314, "grad_norm": 0.189796, "tokens_per_sec": 107593} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:56.102862+00:00", "epoch": 0, "step": 24155, "train_loss": 3.523318290710449, "perplexity": 33.89672122595521, "lr": 0.0026291804804649314, "grad_norm": 0.199541, "tokens_per_sec": 107504} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:56.408450+00:00", "epoch": 0, "step": 24156, "train_loss": 3.462663412094116, "perplexity": 31.901831186032467, "lr": 0.0026291804804649314, "grad_norm": 0.179045, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:56.713383+00:00", "epoch": 0, "step": 24157, "train_loss": 3.4643867015838623, "perplexity": 31.956854673500775, "lr": 0.0026291804804649314, "grad_norm": 0.178551, "tokens_per_sec": 107460} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:57.018838+00:00", "epoch": 0, "step": 24158, "train_loss": 3.4532055854797363, "perplexity": 31.601531527537443, "lr": 0.0026291804804649314, "grad_norm": 0.163014, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:57.323652+00:00", "epoch": 0, "step": 24159, "train_loss": 3.560946226119995, "perplexity": 35.19648522725056, "lr": 0.0026291804804649314, "grad_norm": 0.155711, "tokens_per_sec": 107507} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:57.628643+00:00", "epoch": 0, "step": 24160, "train_loss": 3.4714558124542236, "perplexity": 32.18356158631809, "lr": 0.0026291804804649314, "grad_norm": 0.174846, "tokens_per_sec": 107536} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:57.932665+00:00", "epoch": 0, "step": 24161, "train_loss": 3.45105242729187, "perplexity": 31.533561632407473, "lr": 0.0026291804804649314, "grad_norm": 0.177934, "tokens_per_sec": 107679} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:58.239108+00:00", "epoch": 0, "step": 24162, "train_loss": 3.5691962242126465, "perplexity": 35.488057243776126, "lr": 0.0026291804804649314, "grad_norm": 0.172632, "tokens_per_sec": 106930} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:58.544717+00:00", "epoch": 0, "step": 24163, "train_loss": 3.435176134109497, "perplexity": 31.03687873712224, "lr": 0.0026291804804649314, "grad_norm": 0.148413, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:58.850279+00:00", "epoch": 0, "step": 24164, "train_loss": 3.541365623474121, "perplexity": 34.5140201785723, "lr": 0.0026291804804649314, "grad_norm": 0.177146, "tokens_per_sec": 107245} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:59.154348+00:00", "epoch": 0, "step": 24165, "train_loss": 3.4398903846740723, "perplexity": 31.183539787082058, "lr": 0.0026291804804649314, "grad_norm": 0.156625, "tokens_per_sec": 107758} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:59.459569+00:00", "epoch": 0, "step": 24166, "train_loss": 3.515106201171875, "perplexity": 33.619498164440465, "lr": 0.0026291804804649314, "grad_norm": 0.173655, "tokens_per_sec": 107359} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:04:59.765071+00:00", "epoch": 0, "step": 24167, "train_loss": 3.460089921951294, "perplexity": 31.819837687919378, "lr": 0.0026291804804649314, "grad_norm": 0.189348, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:00.070821+00:00", "epoch": 0, "step": 24168, "train_loss": 3.5618844032287598, "perplexity": 35.22952125840071, "lr": 0.0026291804804649314, "grad_norm": 0.195642, "tokens_per_sec": 107172} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:00.376460+00:00", "epoch": 0, "step": 24169, "train_loss": 3.555288314819336, "perplexity": 34.99790892901517, "lr": 0.0026291804804649314, "grad_norm": 0.174769, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:00.681444+00:00", "epoch": 0, "step": 24170, "train_loss": 3.5610034465789795, "perplexity": 35.19849924391083, "lr": 0.0026291804804649314, "grad_norm": 0.149022, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:00.985570+00:00", "epoch": 0, "step": 24171, "train_loss": 3.513190746307373, "perplexity": 33.555163168200274, "lr": 0.0026291804804649314, "grad_norm": 0.17296, "tokens_per_sec": 107675} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:01.291119+00:00", "epoch": 0, "step": 24172, "train_loss": 3.5394513607025146, "perplexity": 34.44801447093636, "lr": 0.0026291804804649314, "grad_norm": 0.16758, "tokens_per_sec": 107244} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:01.596964+00:00", "epoch": 0, "step": 24173, "train_loss": 3.5123379230499268, "perplexity": 33.526558743635974, "lr": 0.0026291804804649314, "grad_norm": 0.173403, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:01.902921+00:00", "epoch": 0, "step": 24174, "train_loss": 3.523717164993286, "perplexity": 33.91024445317922, "lr": 0.0026291804804649314, "grad_norm": 0.222387, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:02.207770+00:00", "epoch": 0, "step": 24175, "train_loss": 3.610234260559082, "perplexity": 36.97471351741287, "lr": 0.0026291804804649314, "grad_norm": 0.209432, "tokens_per_sec": 107491} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:02.513128+00:00", "epoch": 0, "step": 24176, "train_loss": 3.6078550815582275, "perplexity": 36.88684862005553, "lr": 0.0026291804804649314, "grad_norm": 0.191196, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:02.817513+00:00", "epoch": 0, "step": 24177, "train_loss": 3.4519124031066895, "perplexity": 31.560691396564973, "lr": 0.0026291804804649314, "grad_norm": 0.207129, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:03.122938+00:00", "epoch": 0, "step": 24178, "train_loss": 3.505064010620117, "perplexity": 33.28357428644762, "lr": 0.0026291804804649314, "grad_norm": 0.237085, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:03.428683+00:00", "epoch": 0, "step": 24179, "train_loss": 3.404970169067383, "perplexity": 30.113397399790205, "lr": 0.0026291804804649314, "grad_norm": 0.184904, "tokens_per_sec": 107175} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:03.734112+00:00", "epoch": 0, "step": 24180, "train_loss": 3.5261054039001465, "perplexity": 33.99132700203326, "lr": 0.0026291804804649314, "grad_norm": 0.189185, "tokens_per_sec": 107348} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:04.039869+00:00", "epoch": 0, "step": 24181, "train_loss": 3.473379135131836, "perplexity": 32.24552052457919, "lr": 0.0026291804804649314, "grad_norm": 0.184551, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:04.344472+00:00", "epoch": 0, "step": 24182, "train_loss": 3.550973892211914, "perplexity": 34.84723842135962, "lr": 0.0026291804804649314, "grad_norm": 0.184576, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:04.649991+00:00", "epoch": 0, "step": 24183, "train_loss": 3.477409839630127, "perplexity": 32.37575498119041, "lr": 0.0026291804804649314, "grad_norm": 0.191992, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:04.955835+00:00", "epoch": 0, "step": 24184, "train_loss": 3.521846294403076, "perplexity": 33.8468620827187, "lr": 0.0026291804804649314, "grad_norm": 0.178034, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:05.260722+00:00", "epoch": 0, "step": 24185, "train_loss": 3.5298171043395996, "perplexity": 34.117727059864585, "lr": 0.0026291804804649314, "grad_norm": 0.173931, "tokens_per_sec": 107476} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:05.566568+00:00", "epoch": 0, "step": 24186, "train_loss": 3.412144899368286, "perplexity": 30.330229831535057, "lr": 0.0026291804804649314, "grad_norm": 0.185185, "tokens_per_sec": 107139} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:05.872370+00:00", "epoch": 0, "step": 24187, "train_loss": 3.5379562377929688, "perplexity": 34.396548938547895, "lr": 0.0026291804804649314, "grad_norm": 0.181848, "tokens_per_sec": 107155} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:06.178528+00:00", "epoch": 0, "step": 24188, "train_loss": 3.562495708465576, "perplexity": 35.25106383311001, "lr": 0.0026291804804649314, "grad_norm": 0.18121, "tokens_per_sec": 107030} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:06.483918+00:00", "epoch": 0, "step": 24189, "train_loss": 3.523820161819458, "perplexity": 33.91373728060459, "lr": 0.0026291804804649314, "grad_norm": 0.15529, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:06.789351+00:00", "epoch": 0, "step": 24190, "train_loss": 3.5710983276367188, "perplexity": 35.55562343754625, "lr": 0.0026291804804649314, "grad_norm": 0.16873, "tokens_per_sec": 107344} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:07.094463+00:00", "epoch": 0, "step": 24191, "train_loss": 3.502285957336426, "perplexity": 33.191239059082655, "lr": 0.0026291804804649314, "grad_norm": 0.186192, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:07.400828+00:00", "epoch": 0, "step": 24192, "train_loss": 3.5232741832733154, "perplexity": 33.89522616142667, "lr": 0.0026291804804649314, "grad_norm": 0.189896, "tokens_per_sec": 106957} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:07.707650+00:00", "epoch": 0, "step": 24193, "train_loss": 3.499469757080078, "perplexity": 33.097897379258235, "lr": 0.0026291804804649314, "grad_norm": 0.201241, "tokens_per_sec": 106800} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:08.011882+00:00", "epoch": 0, "step": 24194, "train_loss": 3.4766147136688232, "perplexity": 32.35002240956167, "lr": 0.0026291804804649314, "grad_norm": 0.175135, "tokens_per_sec": 107706} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:08.317072+00:00", "epoch": 0, "step": 24195, "train_loss": 3.4484498500823975, "perplexity": 31.45159980582383, "lr": 0.0026291804804649314, "grad_norm": 0.154677, "tokens_per_sec": 107369} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:08.622776+00:00", "epoch": 0, "step": 24196, "train_loss": 3.503722906112671, "perplexity": 33.238967452846616, "lr": 0.0026291804804649314, "grad_norm": 0.180093, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:08.928833+00:00", "epoch": 0, "step": 24197, "train_loss": 3.4847962856292725, "perplexity": 32.61578212951111, "lr": 0.0026291804804649314, "grad_norm": 0.163909, "tokens_per_sec": 107065} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:09.236183+00:00", "epoch": 0, "step": 24198, "train_loss": 3.502462863922119, "perplexity": 33.197111327265674, "lr": 0.0026291804804649314, "grad_norm": 0.16739, "tokens_per_sec": 106618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:09.540069+00:00", "epoch": 0, "step": 24199, "train_loss": 3.4300408363342285, "perplexity": 30.87790366432561, "lr": 0.0026291804804649314, "grad_norm": 0.184611, "tokens_per_sec": 107827} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:09.844806+00:00", "epoch": 0, "step": 24200, "train_loss": 3.40334415435791, "perplexity": 30.06447235986275, "lr": 0.0026291804804649314, "grad_norm": 0.157917, "tokens_per_sec": 107588} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:10.149511+00:00", "epoch": 0, "step": 24201, "train_loss": 3.438943386077881, "perplexity": 31.154022997064672, "lr": 0.0026291804804649314, "grad_norm": 0.180806, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:10.454934+00:00", "epoch": 0, "step": 24202, "train_loss": 3.449084758758545, "perplexity": 31.471575039972983, "lr": 0.0026291804804649314, "grad_norm": 0.164262, "tokens_per_sec": 107287} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:10.760446+00:00", "epoch": 0, "step": 24203, "train_loss": 3.409799337387085, "perplexity": 30.25917176568808, "lr": 0.0026291804804649314, "grad_norm": 0.170056, "tokens_per_sec": 107256} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:11.065936+00:00", "epoch": 0, "step": 24204, "train_loss": 3.46854567527771, "perplexity": 32.09003915482815, "lr": 0.0026291804804649314, "grad_norm": 0.16391, "tokens_per_sec": 107269} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:11.370371+00:00", "epoch": 0, "step": 24205, "train_loss": 3.350386142730713, "perplexity": 28.513741892410078, "lr": 0.0026291804804649314, "grad_norm": 0.165173, "tokens_per_sec": 107632} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:11.675403+00:00", "epoch": 0, "step": 24206, "train_loss": 3.477828025817871, "perplexity": 32.38929690606695, "lr": 0.0026291804804649314, "grad_norm": 0.161291, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:11.980577+00:00", "epoch": 0, "step": 24207, "train_loss": 3.537957191467285, "perplexity": 34.39658174166883, "lr": 0.0026291804804649314, "grad_norm": 0.165435, "tokens_per_sec": 107365} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:12.286018+00:00", "epoch": 0, "step": 24208, "train_loss": 3.4840171337127686, "perplexity": 32.59037937794303, "lr": 0.0026291804804649314, "grad_norm": 0.172346, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:12.591299+00:00", "epoch": 0, "step": 24209, "train_loss": 3.4039692878723145, "perplexity": 30.083272544828386, "lr": 0.0026291804804649314, "grad_norm": 0.179888, "tokens_per_sec": 107337} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:12.896732+00:00", "epoch": 0, "step": 24210, "train_loss": 3.523744583129883, "perplexity": 33.91117422163986, "lr": 0.0026291804804649314, "grad_norm": 0.193347, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:13.201805+00:00", "epoch": 0, "step": 24211, "train_loss": 3.489159345626831, "perplexity": 32.75839763757947, "lr": 0.0026291804804649314, "grad_norm": 0.194268, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:13.507073+00:00", "epoch": 0, "step": 24212, "train_loss": 3.459378480911255, "perplexity": 31.797207800370682, "lr": 0.0026291804804649314, "grad_norm": 0.165074, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:13.813055+00:00", "epoch": 0, "step": 24213, "train_loss": 3.596198320388794, "perplexity": 36.459363820051905, "lr": 0.0026291804804649314, "grad_norm": 0.165058, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:14.118755+00:00", "epoch": 0, "step": 24214, "train_loss": 3.4064197540283203, "perplexity": 30.157080981665995, "lr": 0.0026291804804649314, "grad_norm": 0.177135, "tokens_per_sec": 107190} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:14.424737+00:00", "epoch": 0, "step": 24215, "train_loss": 3.5015830993652344, "perplexity": 33.16791852860668, "lr": 0.0026291804804649314, "grad_norm": 0.170982, "tokens_per_sec": 107097} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:14.729444+00:00", "epoch": 0, "step": 24216, "train_loss": 3.491281747817993, "perplexity": 32.8279979663228, "lr": 0.0026291804804649314, "grad_norm": 0.198204, "tokens_per_sec": 107533} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:15.035095+00:00", "epoch": 0, "step": 24217, "train_loss": 3.479032039642334, "perplexity": 32.42831755329199, "lr": 0.0026291804804649314, "grad_norm": 0.185947, "tokens_per_sec": 107208} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:15.340302+00:00", "epoch": 0, "step": 24218, "train_loss": 3.496753454208374, "perplexity": 33.00811545820774, "lr": 0.0026291804804649314, "grad_norm": 0.176433, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:15.645103+00:00", "epoch": 0, "step": 24219, "train_loss": 3.455380916595459, "perplexity": 31.670350146877933, "lr": 0.0026291804804649314, "grad_norm": 0.210859, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:15.950047+00:00", "epoch": 0, "step": 24220, "train_loss": 3.5268702507019043, "perplexity": 34.01733510461744, "lr": 0.0026291804804649314, "grad_norm": 0.219456, "tokens_per_sec": 107524} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:16.254454+00:00", "epoch": 0, "step": 24221, "train_loss": 3.495537281036377, "perplexity": 32.96799627461162, "lr": 0.0026291804804649314, "grad_norm": 0.173772, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:16.558901+00:00", "epoch": 0, "step": 24222, "train_loss": 3.475590705871582, "perplexity": 32.316912689572284, "lr": 0.0026291804804649314, "grad_norm": 0.186391, "tokens_per_sec": 107631} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:16.863473+00:00", "epoch": 0, "step": 24223, "train_loss": 3.546140193939209, "perplexity": 34.679203826631074, "lr": 0.0026291804804649314, "grad_norm": 0.182754, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:17.168519+00:00", "epoch": 0, "step": 24224, "train_loss": 3.50765323638916, "perplexity": 33.36986463907007, "lr": 0.0026291804804649314, "grad_norm": 0.188742, "tokens_per_sec": 107420} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:17.473790+00:00", "epoch": 0, "step": 24225, "train_loss": 3.417015790939331, "perplexity": 30.478325478484017, "lr": 0.0026291804804649314, "grad_norm": 0.171629, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:17.778274+00:00", "epoch": 0, "step": 24226, "train_loss": 3.590775489807129, "perplexity": 36.26218598114988, "lr": 0.0026291804804649314, "grad_norm": 0.170951, "tokens_per_sec": 107618} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:18.082628+00:00", "epoch": 0, "step": 24227, "train_loss": 3.489602565765381, "perplexity": 32.77292003719131, "lr": 0.0026291804804649314, "grad_norm": 0.166248, "tokens_per_sec": 107667} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:18.387822+00:00", "epoch": 0, "step": 24228, "train_loss": 3.5508272647857666, "perplexity": 34.84212923506412, "lr": 0.0026291804804649314, "grad_norm": 0.169887, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:18.693673+00:00", "epoch": 0, "step": 24229, "train_loss": 3.590069055557251, "perplexity": 36.23657817717721, "lr": 0.0026291804804649314, "grad_norm": 0.161543, "tokens_per_sec": 107138} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:18.999669+00:00", "epoch": 0, "step": 24230, "train_loss": 3.5661513805389404, "perplexity": 35.38016599653173, "lr": 0.0026291804804649314, "grad_norm": 0.160528, "tokens_per_sec": 107146} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:19.303826+00:00", "epoch": 0, "step": 24231, "train_loss": 3.487795829772949, "perplexity": 32.71376148102196, "lr": 0.0026291804804649314, "grad_norm": 0.167287, "tokens_per_sec": 107674} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:19.609808+00:00", "epoch": 0, "step": 24232, "train_loss": 3.4292426109313965, "perplexity": 30.853265971758773, "lr": 0.0026291804804649314, "grad_norm": 0.176509, "tokens_per_sec": 107092} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:19.914264+00:00", "epoch": 0, "step": 24233, "train_loss": 3.416205644607544, "perplexity": 30.45364357422568, "lr": 0.0026291804804649314, "grad_norm": 0.168166, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:20.220211+00:00", "epoch": 0, "step": 24234, "train_loss": 3.5398828983306885, "perplexity": 34.46288329339389, "lr": 0.0026291804804649314, "grad_norm": 0.183783, "tokens_per_sec": 107103} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:20.526561+00:00", "epoch": 0, "step": 24235, "train_loss": 3.4537782669067383, "perplexity": 31.619634320780058, "lr": 0.0026291804804649314, "grad_norm": 0.200608, "tokens_per_sec": 106963} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:20.831052+00:00", "epoch": 0, "step": 24236, "train_loss": 3.4726696014404297, "perplexity": 32.22264935626183, "lr": 0.0026291804804649314, "grad_norm": 0.197513, "tokens_per_sec": 107616} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:21.134664+00:00", "epoch": 0, "step": 24237, "train_loss": 3.432805299758911, "perplexity": 30.963382596890177, "lr": 0.0026291804804649314, "grad_norm": 0.172696, "tokens_per_sec": 107929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:21.439879+00:00", "epoch": 0, "step": 24238, "train_loss": 3.358271598815918, "perplexity": 28.739474584866016, "lr": 0.0026291804804649314, "grad_norm": 0.159005, "tokens_per_sec": 107358} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:21.744887+00:00", "epoch": 0, "step": 24239, "train_loss": 3.45559024810791, "perplexity": 31.676980443114797, "lr": 0.0026291804804649314, "grad_norm": 0.17146, "tokens_per_sec": 107434} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:22.051137+00:00", "epoch": 0, "step": 24240, "train_loss": 3.5347163677215576, "perplexity": 34.28528892040597, "lr": 0.0026291804804649314, "grad_norm": 0.166178, "tokens_per_sec": 107053} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:22.356392+00:00", "epoch": 0, "step": 24241, "train_loss": 3.564929485321045, "perplexity": 35.33696154192591, "lr": 0.0026291804804649314, "grad_norm": 0.166908, "tokens_per_sec": 107291} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:22.661760+00:00", "epoch": 0, "step": 24242, "train_loss": 3.553375482559204, "perplexity": 34.931027786374976, "lr": 0.0026291804804649314, "grad_norm": 0.156172, "tokens_per_sec": 107306} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:22.967204+00:00", "epoch": 0, "step": 24243, "train_loss": 3.505589723587036, "perplexity": 33.30107649320083, "lr": 0.0026291804804649314, "grad_norm": 0.200113, "tokens_per_sec": 107282} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:23.271576+00:00", "epoch": 0, "step": 24244, "train_loss": 3.4347083568573, "perplexity": 31.022363786416363, "lr": 0.0026291804804649314, "grad_norm": 0.182802, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:23.577186+00:00", "epoch": 0, "step": 24245, "train_loss": 3.522162914276123, "perplexity": 33.857580368615984, "lr": 0.0026291804804649314, "grad_norm": 0.166176, "tokens_per_sec": 107216} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:23.884056+00:00", "epoch": 0, "step": 24246, "train_loss": 3.493169069290161, "perplexity": 32.89001345494639, "lr": 0.0026291804804649314, "grad_norm": 0.161382, "tokens_per_sec": 106780} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:24.188852+00:00", "epoch": 0, "step": 24247, "train_loss": 3.481843948364258, "perplexity": 32.51963134511435, "lr": 0.0026291804804649314, "grad_norm": 0.191627, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:24.494017+00:00", "epoch": 0, "step": 24248, "train_loss": 3.4908640384674072, "perplexity": 32.81428826814515, "lr": 0.0026291804804649314, "grad_norm": 0.192838, "tokens_per_sec": 107375} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:24.798474+00:00", "epoch": 0, "step": 24249, "train_loss": 3.491393804550171, "perplexity": 32.83167677061238, "lr": 0.0026291804804649314, "grad_norm": 0.162304, "tokens_per_sec": 107628} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:25.103037+00:00", "epoch": 0, "step": 24250, "train_loss": 3.569042444229126, "perplexity": 35.48260031051231, "lr": 0.0026291804804649314, "grad_norm": 0.168906, "tokens_per_sec": 107650} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:25.408000+00:00", "epoch": 0, "step": 24251, "train_loss": 3.4907758235931396, "perplexity": 32.811393687505806, "lr": 0.0026291804804649314, "grad_norm": 0.19567, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:25.713938+00:00", "epoch": 0, "step": 24252, "train_loss": 3.459092617034912, "perplexity": 31.78811942637279, "lr": 0.0026291804804649314, "grad_norm": 0.229809, "tokens_per_sec": 107108} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:26.018535+00:00", "epoch": 0, "step": 24253, "train_loss": 3.445344924926758, "perplexity": 31.35409639113641, "lr": 0.0026291804804649314, "grad_norm": 0.206155, "tokens_per_sec": 107580} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:26.323708+00:00", "epoch": 0, "step": 24254, "train_loss": 3.375558376312256, "perplexity": 29.240606485722697, "lr": 0.0026291804804649314, "grad_norm": 0.261804, "tokens_per_sec": 107374} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:26.628162+00:00", "epoch": 0, "step": 24255, "train_loss": 3.4784700870513916, "perplexity": 32.41009949554481, "lr": 0.0026291804804649314, "grad_norm": 0.168441, "tokens_per_sec": 107627} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:26.933404+00:00", "epoch": 0, "step": 24256, "train_loss": 3.415127992630005, "perplexity": 30.42084282207674, "lr": 0.0026291804804649314, "grad_norm": 0.190194, "tokens_per_sec": 107351} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:27.239312+00:00", "epoch": 0, "step": 24257, "train_loss": 3.524533987045288, "perplexity": 33.93795440414619, "lr": 0.0026291804804649314, "grad_norm": 0.167819, "tokens_per_sec": 107119} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:27.544926+00:00", "epoch": 0, "step": 24258, "train_loss": 3.6032614707946777, "perplexity": 36.71779337946054, "lr": 0.0026291804804649314, "grad_norm": 0.173794, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:27.851028+00:00", "epoch": 0, "step": 24259, "train_loss": 3.531527280807495, "perplexity": 34.1761243143928, "lr": 0.0026291804804649314, "grad_norm": 0.191663, "tokens_per_sec": 107058} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:28.157000+00:00", "epoch": 0, "step": 24260, "train_loss": 3.4386181831359863, "perplexity": 31.143893264332892, "lr": 0.0026291804804649314, "grad_norm": 0.181358, "tokens_per_sec": 107153} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:28.463029+00:00", "epoch": 0, "step": 24261, "train_loss": 3.527269124984741, "perplexity": 34.03090645121195, "lr": 0.0026291804804649314, "grad_norm": 0.157267, "tokens_per_sec": 107008} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:28.768167+00:00", "epoch": 0, "step": 24262, "train_loss": 3.475994348526001, "perplexity": 32.32995980700234, "lr": 0.0026291804804649314, "grad_norm": 0.18069, "tokens_per_sec": 107387} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:29.072899+00:00", "epoch": 0, "step": 24263, "train_loss": 3.4896605014801025, "perplexity": 32.77481881474017, "lr": 0.0026291804804649314, "grad_norm": 0.160128, "tokens_per_sec": 107531} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:29.377961+00:00", "epoch": 0, "step": 24264, "train_loss": 3.5396041870117188, "perplexity": 34.453279436149714, "lr": 0.0026291804804649314, "grad_norm": 0.16577, "tokens_per_sec": 107419} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:29.682960+00:00", "epoch": 0, "step": 24265, "train_loss": 3.499952554702759, "perplexity": 33.11388082350264, "lr": 0.0026291804804649314, "grad_norm": 0.171709, "tokens_per_sec": 107432} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:29.988619+00:00", "epoch": 0, "step": 24266, "train_loss": 3.486903190612793, "perplexity": 32.68457292580983, "lr": 0.0026291804804649314, "grad_norm": 0.185248, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:30.294705+00:00", "epoch": 0, "step": 24267, "train_loss": 3.4574661254882812, "perplexity": 31.736458343388893, "lr": 0.0026291804804649314, "grad_norm": 0.149129, "tokens_per_sec": 107055} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:30.600061+00:00", "epoch": 0, "step": 24268, "train_loss": 3.4936065673828125, "perplexity": 32.90440592120891, "lr": 0.0026291804804649314, "grad_norm": 0.185679, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:30.904809+00:00", "epoch": 0, "step": 24269, "train_loss": 3.498863935470581, "perplexity": 33.07785203036281, "lr": 0.0026291804804649314, "grad_norm": 0.203927, "tokens_per_sec": 107525} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:31.209909+00:00", "epoch": 0, "step": 24270, "train_loss": 3.4701144695281982, "perplexity": 32.14042133305492, "lr": 0.0026291804804649314, "grad_norm": 0.189445, "tokens_per_sec": 107506} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:31.514594+00:00", "epoch": 0, "step": 24271, "train_loss": 3.550771951675415, "perplexity": 34.84020206182432, "lr": 0.0026291804804649314, "grad_norm": 0.177586, "tokens_per_sec": 107440} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:31.819994+00:00", "epoch": 0, "step": 24272, "train_loss": 3.525315999984741, "perplexity": 33.96450470361461, "lr": 0.0026291804804649314, "grad_norm": 0.181287, "tokens_per_sec": 107296} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:32.125604+00:00", "epoch": 0, "step": 24273, "train_loss": 3.521571159362793, "perplexity": 33.83755090593035, "lr": 0.0026291804804649314, "grad_norm": 0.202879, "tokens_per_sec": 107222} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:32.430112+00:00", "epoch": 0, "step": 24274, "train_loss": 3.607314348220825, "perplexity": 36.8669080630428, "lr": 0.0026291804804649314, "grad_norm": 0.189541, "tokens_per_sec": 107610} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:32.736069+00:00", "epoch": 0, "step": 24275, "train_loss": 3.5038743019104004, "perplexity": 33.24400007378906, "lr": 0.0026291804804649314, "grad_norm": 0.190992, "tokens_per_sec": 107100} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:33.041364+00:00", "epoch": 0, "step": 24276, "train_loss": 3.4587481021881104, "perplexity": 31.77716983353545, "lr": 0.0026291804804649314, "grad_norm": 0.179002, "tokens_per_sec": 107332} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:33.346888+00:00", "epoch": 0, "step": 24277, "train_loss": 3.6385154724121094, "perplexity": 38.03533029596094, "lr": 0.0026291804804649314, "grad_norm": 0.161766, "tokens_per_sec": 107252} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:33.652134+00:00", "epoch": 0, "step": 24278, "train_loss": 3.4932568073272705, "perplexity": 32.89289928676419, "lr": 0.0026291804804649314, "grad_norm": 0.181589, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:33.957389+00:00", "epoch": 0, "step": 24279, "train_loss": 3.496943235397339, "perplexity": 33.014380372067485, "lr": 0.0026291804804649314, "grad_norm": 0.176489, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:34.262209+00:00", "epoch": 0, "step": 24280, "train_loss": 3.5976288318634033, "perplexity": 36.511556680696, "lr": 0.0026291804804649314, "grad_norm": 0.167312, "tokens_per_sec": 107559} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:34.566913+00:00", "epoch": 0, "step": 24281, "train_loss": 3.55458927154541, "perplexity": 34.97345242525002, "lr": 0.0026291804804649314, "grad_norm": 0.193244, "tokens_per_sec": 107481} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:34.871771+00:00", "epoch": 0, "step": 24282, "train_loss": 3.529879331588745, "perplexity": 34.119850178223835, "lr": 0.0026291804804649314, "grad_norm": 0.205272, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:35.177087+00:00", "epoch": 0, "step": 24283, "train_loss": 3.6453819274902344, "perplexity": 38.29739688739527, "lr": 0.0026291804804649314, "grad_norm": 0.209733, "tokens_per_sec": 107325} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:35.482408+00:00", "epoch": 0, "step": 24284, "train_loss": 3.460278034210205, "perplexity": 31.82582395249226, "lr": 0.0026291804804649314, "grad_norm": 0.188678, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:35.786769+00:00", "epoch": 0, "step": 24285, "train_loss": 3.426757574081421, "perplexity": 30.776689655767886, "lr": 0.0026291804804649314, "grad_norm": 0.211397, "tokens_per_sec": 107661} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:36.091194+00:00", "epoch": 0, "step": 24286, "train_loss": 3.4004266262054443, "perplexity": 29.97688624496933, "lr": 0.0026291804804649314, "grad_norm": 0.20148, "tokens_per_sec": 107639} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:36.395984+00:00", "epoch": 0, "step": 24287, "train_loss": 3.5326013565063477, "perplexity": 34.21285177950788, "lr": 0.0026291804804649314, "grad_norm": 0.231314, "tokens_per_sec": 107510} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:36.701765+00:00", "epoch": 0, "step": 24288, "train_loss": 3.570420026779175, "perplexity": 35.531514205258894, "lr": 0.0026291804804649314, "grad_norm": 0.183174, "tokens_per_sec": 107162} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:37.007660+00:00", "epoch": 0, "step": 24289, "train_loss": 3.527406692504883, "perplexity": 34.035588320649815, "lr": 0.0026291804804649314, "grad_norm": 0.204458, "tokens_per_sec": 107122} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:37.313478+00:00", "epoch": 0, "step": 24290, "train_loss": 3.45689058303833, "perplexity": 31.718197919729715, "lr": 0.0026291804804649314, "grad_norm": 0.165142, "tokens_per_sec": 107205} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:37.618424+00:00", "epoch": 0, "step": 24291, "train_loss": 3.509692430496216, "perplexity": 33.43798169874431, "lr": 0.0026291804804649314, "grad_norm": 0.186455, "tokens_per_sec": 107398} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:37.924266+00:00", "epoch": 0, "step": 24292, "train_loss": 3.5529873371124268, "perplexity": 34.917472097948, "lr": 0.0026291804804649314, "grad_norm": 0.161036, "tokens_per_sec": 107140} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:38.230372+00:00", "epoch": 0, "step": 24293, "train_loss": 3.486799478530884, "perplexity": 32.68118331648014, "lr": 0.0026291804804649314, "grad_norm": 0.20801, "tokens_per_sec": 107048} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:38.535876+00:00", "epoch": 0, "step": 24294, "train_loss": 3.559420108795166, "perplexity": 35.1428122274368, "lr": 0.0026291804804649314, "grad_norm": 0.198025, "tokens_per_sec": 107259} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:38.841342+00:00", "epoch": 0, "step": 24295, "train_loss": 3.465114116668701, "perplexity": 31.98010902841317, "lr": 0.0026291804804649314, "grad_norm": 0.167845, "tokens_per_sec": 107272} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:39.146670+00:00", "epoch": 0, "step": 24296, "train_loss": 3.543635845184326, "perplexity": 34.59246366485007, "lr": 0.0026291804804649314, "grad_norm": 0.17488, "tokens_per_sec": 107321} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:39.451030+00:00", "epoch": 0, "step": 24297, "train_loss": 3.509223699569702, "perplexity": 33.42231195533548, "lr": 0.0026291804804649314, "grad_norm": 0.188851, "tokens_per_sec": 107662} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:39.755544+00:00", "epoch": 0, "step": 24298, "train_loss": 3.515953540802002, "perplexity": 33.64799737012821, "lr": 0.0026291804804649314, "grad_norm": 0.157836, "tokens_per_sec": 107608} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:40.061036+00:00", "epoch": 0, "step": 24299, "train_loss": 3.422248601913452, "perplexity": 30.638230806817845, "lr": 0.0026291804804649314, "grad_norm": 0.187701, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:40.367483+00:00", "epoch": 0, "step": 24300, "train_loss": 3.526261806488037, "perplexity": 33.99664374930789, "lr": 0.0026291804804649314, "grad_norm": 0.190573, "tokens_per_sec": 106989} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:40.672598+00:00", "epoch": 0, "step": 24301, "train_loss": 3.47342586517334, "perplexity": 32.24702739429939, "lr": 0.0026291804804649314, "grad_norm": 0.182676, "tokens_per_sec": 107335} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:40.977823+00:00", "epoch": 0, "step": 24302, "train_loss": 3.535470485687256, "perplexity": 34.311153824108516, "lr": 0.0026291804804649314, "grad_norm": 0.178324, "tokens_per_sec": 107357} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:41.283009+00:00", "epoch": 0, "step": 24303, "train_loss": 3.5305962562561035, "perplexity": 34.144320311036, "lr": 0.0026291804804649314, "grad_norm": 0.180074, "tokens_per_sec": 107372} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:41.587968+00:00", "epoch": 0, "step": 24304, "train_loss": 3.373584032058716, "perplexity": 29.182932415306965, "lr": 0.0026291804804649314, "grad_norm": 0.164371, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:41.893576+00:00", "epoch": 0, "step": 24305, "train_loss": 3.5461935997009277, "perplexity": 34.681055945383704, "lr": 0.0026291804804649314, "grad_norm": 0.195742, "tokens_per_sec": 107225} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:42.198493+00:00", "epoch": 0, "step": 24306, "train_loss": 3.5552618503570557, "perplexity": 34.99698274043002, "lr": 0.0026291804804649314, "grad_norm": 0.195458, "tokens_per_sec": 107464} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:42.503520+00:00", "epoch": 0, "step": 24307, "train_loss": 3.4306414127349854, "perplexity": 30.896453774392846, "lr": 0.0026291804804649314, "grad_norm": 0.161916, "tokens_per_sec": 107426} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:42.808772+00:00", "epoch": 0, "step": 24308, "train_loss": 3.4896047115325928, "perplexity": 32.772990360324016, "lr": 0.0026291804804649314, "grad_norm": 0.162529, "tokens_per_sec": 107347} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:43.114420+00:00", "epoch": 0, "step": 24309, "train_loss": 3.3769969940185547, "perplexity": 29.28270281295655, "lr": 0.0026291804804649314, "grad_norm": 0.155664, "tokens_per_sec": 107210} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:43.420253+00:00", "epoch": 0, "step": 24310, "train_loss": 3.418055534362793, "perplexity": 30.51003159721625, "lr": 0.0026291804804649314, "grad_norm": 0.176411, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:43.725346+00:00", "epoch": 0, "step": 24311, "train_loss": 3.5822973251342773, "perplexity": 35.95604877158294, "lr": 0.0026291804804649314, "grad_norm": 0.175897, "tokens_per_sec": 107349} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:44.030975+00:00", "epoch": 0, "step": 24312, "train_loss": 3.5342860221862793, "perplexity": 34.2705375737091, "lr": 0.0026291804804649314, "grad_norm": 0.171181, "tokens_per_sec": 107211} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:44.336083+00:00", "epoch": 0, "step": 24313, "train_loss": 3.472928047180176, "perplexity": 32.23097823894705, "lr": 0.0026291804804649314, "grad_norm": 0.152919, "tokens_per_sec": 107399} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:44.640548+00:00", "epoch": 0, "step": 24314, "train_loss": 3.5007107257843018, "perplexity": 33.13899633006299, "lr": 0.0026291804804649314, "grad_norm": 0.191237, "tokens_per_sec": 107625} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:44.945069+00:00", "epoch": 0, "step": 24315, "train_loss": 3.513793468475342, "perplexity": 33.57539370497561, "lr": 0.0026291804804649314, "grad_norm": 0.209544, "tokens_per_sec": 107604} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:45.250517+00:00", "epoch": 0, "step": 24316, "train_loss": 3.475736141204834, "perplexity": 32.32161305233009, "lr": 0.0026291804804649314, "grad_norm": 0.196222, "tokens_per_sec": 107289} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:45.555982+00:00", "epoch": 0, "step": 24317, "train_loss": 3.431605815887451, "perplexity": 30.926264784418372, "lr": 0.0026291804804649314, "grad_norm": 0.182061, "tokens_per_sec": 107264} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:45.861881+00:00", "epoch": 0, "step": 24318, "train_loss": 3.4779274463653564, "perplexity": 32.39251722777844, "lr": 0.0026291804804649314, "grad_norm": 0.198054, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:46.167760+00:00", "epoch": 0, "step": 24319, "train_loss": 3.5739636421203613, "perplexity": 35.65764757617611, "lr": 0.0026291804804649314, "grad_norm": 0.172902, "tokens_per_sec": 107127} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:46.472826+00:00", "epoch": 0, "step": 24320, "train_loss": 3.514500856399536, "perplexity": 33.59915293553822, "lr": 0.0026291804804649314, "grad_norm": 0.17173, "tokens_per_sec": 107478} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:46.777795+00:00", "epoch": 0, "step": 24321, "train_loss": 3.522000789642334, "perplexity": 33.85209166573674, "lr": 0.0026291804804649314, "grad_norm": 0.170738, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:47.083965+00:00", "epoch": 0, "step": 24322, "train_loss": 3.6391265392303467, "perplexity": 38.05857952691885, "lr": 0.0026291804804649314, "grad_norm": 0.184435, "tokens_per_sec": 107031} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:47.389457+00:00", "epoch": 0, "step": 24323, "train_loss": 3.4757940769195557, "perplexity": 32.323485682328794, "lr": 0.0026291804804649314, "grad_norm": 0.163052, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:47.695508+00:00", "epoch": 0, "step": 24324, "train_loss": 3.547001600265503, "perplexity": 34.709089582239656, "lr": 0.0026291804804649314, "grad_norm": 0.174886, "tokens_per_sec": 107068} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:48.000333+00:00", "epoch": 0, "step": 24325, "train_loss": 3.49100399017334, "perplexity": 32.81888100514043, "lr": 0.0026291804804649314, "grad_norm": 0.17488, "tokens_per_sec": 107498} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:48.304970+00:00", "epoch": 0, "step": 24326, "train_loss": 3.37642240524292, "perplexity": 29.265882133549553, "lr": 0.0026291804804649314, "grad_norm": 0.183176, "tokens_per_sec": 107564} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:48.610128+00:00", "epoch": 0, "step": 24327, "train_loss": 3.564943313598633, "perplexity": 35.33745019461783, "lr": 0.0026291804804649314, "grad_norm": 0.169382, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:48.917094+00:00", "epoch": 0, "step": 24328, "train_loss": 3.5114803314208984, "perplexity": 33.49781897276219, "lr": 0.0026291804804649314, "grad_norm": 0.197344, "tokens_per_sec": 106753} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:49.221286+00:00", "epoch": 0, "step": 24329, "train_loss": 3.538936138153076, "perplexity": 34.430270648498784, "lr": 0.0026291804804649314, "grad_norm": 0.19697, "tokens_per_sec": 107716} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:49.526455+00:00", "epoch": 0, "step": 24330, "train_loss": 3.535289764404297, "perplexity": 34.304953628640135, "lr": 0.0026291804804649314, "grad_norm": 0.152942, "tokens_per_sec": 107475} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:49.830985+00:00", "epoch": 0, "step": 24331, "train_loss": 3.447474479675293, "perplexity": 31.420937801953585, "lr": 0.0026291804804649314, "grad_norm": 0.174753, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:50.136418+00:00", "epoch": 0, "step": 24332, "train_loss": 3.4525575637817383, "perplexity": 31.581059683233683, "lr": 0.0026291804804649314, "grad_norm": 0.161002, "tokens_per_sec": 107284} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:50.442733+00:00", "epoch": 0, "step": 24333, "train_loss": 3.469815969467163, "perplexity": 32.13082884707536, "lr": 0.0026291804804649314, "grad_norm": 0.168159, "tokens_per_sec": 106976} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:50.747900+00:00", "epoch": 0, "step": 24334, "train_loss": 3.4965481758117676, "perplexity": 33.001340300612014, "lr": 0.0026291804804649314, "grad_norm": 0.166709, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:51.052958+00:00", "epoch": 0, "step": 24335, "train_loss": 3.525670289993286, "perplexity": 33.9765401201643, "lr": 0.0026291804804649314, "grad_norm": 0.154176, "tokens_per_sec": 107416} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:51.359107+00:00", "epoch": 0, "step": 24336, "train_loss": 3.4436275959014893, "perplexity": 31.300297299935377, "lr": 0.0026291804804649314, "grad_norm": 0.193922, "tokens_per_sec": 107032} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:51.663245+00:00", "epoch": 0, "step": 24337, "train_loss": 3.5054256916046143, "perplexity": 33.29561449958994, "lr": 0.0026291804804649314, "grad_norm": 0.195334, "tokens_per_sec": 107741} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:51.969493+00:00", "epoch": 0, "step": 24338, "train_loss": 3.516591787338257, "perplexity": 33.669479942761136, "lr": 0.0026291804804649314, "grad_norm": 0.190796, "tokens_per_sec": 106999} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:52.275843+00:00", "epoch": 0, "step": 24339, "train_loss": 3.472771167755127, "perplexity": 32.22592225821221, "lr": 0.0026291804804649314, "grad_norm": 0.179768, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:52.580724+00:00", "epoch": 0, "step": 24340, "train_loss": 3.5206480026245117, "perplexity": 33.80632795685992, "lr": 0.0026291804804649314, "grad_norm": 0.183754, "tokens_per_sec": 107540} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:52.885720+00:00", "epoch": 0, "step": 24341, "train_loss": 3.4757940769195557, "perplexity": 32.323485682328794, "lr": 0.0026291804804649314, "grad_norm": 0.194459, "tokens_per_sec": 107377} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:53.190561+00:00", "epoch": 0, "step": 24342, "train_loss": 3.5503318309783936, "perplexity": 34.82487154169557, "lr": 0.0026291804804649314, "grad_norm": 0.163982, "tokens_per_sec": 107492} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:53.494457+00:00", "epoch": 0, "step": 24343, "train_loss": 3.5003981590270996, "perplexity": 33.128639800080975, "lr": 0.0026291804804649314, "grad_norm": 0.203149, "tokens_per_sec": 107826} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:53.799610+00:00", "epoch": 0, "step": 24344, "train_loss": 3.4994394779205322, "perplexity": 33.09689521791523, "lr": 0.0026291804804649314, "grad_norm": 0.166359, "tokens_per_sec": 107383} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:54.106745+00:00", "epoch": 0, "step": 24345, "train_loss": 3.440077781677246, "perplexity": 31.189384036566825, "lr": 0.0026291804804649314, "grad_norm": 0.199931, "tokens_per_sec": 106690} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:54.411127+00:00", "epoch": 0, "step": 24346, "train_loss": 3.568244218826294, "perplexity": 35.45428849869289, "lr": 0.0026291804804649314, "grad_norm": 0.151198, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:54.715784+00:00", "epoch": 0, "step": 24347, "train_loss": 3.5777206420898438, "perplexity": 35.79186532720073, "lr": 0.0026291804804649314, "grad_norm": 0.160334, "tokens_per_sec": 107557} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:55.021290+00:00", "epoch": 0, "step": 24348, "train_loss": 3.5326380729675293, "perplexity": 34.21410797741358, "lr": 0.0026291804804649314, "grad_norm": 0.166422, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:55.326001+00:00", "epoch": 0, "step": 24349, "train_loss": 3.3892712593078613, "perplexity": 29.64434135911, "lr": 0.0026291804804649314, "grad_norm": 0.157855, "tokens_per_sec": 107538} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:55.631252+00:00", "epoch": 0, "step": 24350, "train_loss": 3.5150389671325684, "perplexity": 33.617237865764764, "lr": 0.0026291804804649314, "grad_norm": 0.164466, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:55.935603+00:00", "epoch": 0, "step": 24351, "train_loss": 3.4254579544067383, "perplexity": 30.73671764420208, "lr": 0.0026291804804649314, "grad_norm": 0.167159, "tokens_per_sec": 107605} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:56.240203+00:00", "epoch": 0, "step": 24352, "train_loss": 3.4380226135253906, "perplexity": 31.12535043027135, "lr": 0.0026291804804649314, "grad_norm": 0.158335, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:56.545245+00:00", "epoch": 0, "step": 24353, "train_loss": 3.4312820434570312, "perplexity": 30.916253333308696, "lr": 0.0026291804804649314, "grad_norm": 0.159053, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:56.852102+00:00", "epoch": 0, "step": 24354, "train_loss": 3.4738101959228516, "perplexity": 32.259423300418725, "lr": 0.0026291804804649314, "grad_norm": 0.160769, "tokens_per_sec": 106786} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:57.157581+00:00", "epoch": 0, "step": 24355, "train_loss": 3.473963737487793, "perplexity": 32.264376843033986, "lr": 0.0026291804804649314, "grad_norm": 0.191787, "tokens_per_sec": 107267} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:57.462268+00:00", "epoch": 0, "step": 24356, "train_loss": 3.4774813652038574, "perplexity": 32.37807075845806, "lr": 0.0026291804804649314, "grad_norm": 0.154207, "tokens_per_sec": 107549} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:57.766407+00:00", "epoch": 0, "step": 24357, "train_loss": 3.6319689750671387, "perplexity": 37.78714536452417, "lr": 0.0026291804804649314, "grad_norm": 0.14696, "tokens_per_sec": 107738} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:58.071149+00:00", "epoch": 0, "step": 24358, "train_loss": 3.4110729694366455, "perplexity": 30.297735369349844, "lr": 0.0026291804804649314, "grad_norm": 0.171922, "tokens_per_sec": 107527} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:58.376046+00:00", "epoch": 0, "step": 24359, "train_loss": 3.508333444595337, "perplexity": 33.39257081642518, "lr": 0.0026291804804649314, "grad_norm": 0.153794, "tokens_per_sec": 107473} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:58.682689+00:00", "epoch": 0, "step": 24360, "train_loss": 3.536412477493286, "perplexity": 34.34348987762388, "lr": 0.0026291804804649314, "grad_norm": 0.152135, "tokens_per_sec": 106929} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:58.988488+00:00", "epoch": 0, "step": 24361, "train_loss": 3.4244277477264404, "perplexity": 30.705068777590647, "lr": 0.0026291804804649314, "grad_norm": 0.168619, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:59.294229+00:00", "epoch": 0, "step": 24362, "train_loss": 3.456380844116211, "perplexity": 31.702034039739925, "lr": 0.0026291804804649314, "grad_norm": 0.18446, "tokens_per_sec": 107179} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:59.598771+00:00", "epoch": 0, "step": 24363, "train_loss": 3.4289779663085938, "perplexity": 30.845101901159783, "lr": 0.0026291804804649314, "grad_norm": 0.176855, "tokens_per_sec": 107595} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:05:59.903628+00:00", "epoch": 0, "step": 24364, "train_loss": 3.4669606685638428, "perplexity": 32.03921651507201, "lr": 0.0026291804804649314, "grad_norm": 0.178611, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:00.209813+00:00", "epoch": 0, "step": 24365, "train_loss": 3.4374799728393555, "perplexity": 31.108465130500445, "lr": 0.0026291804804649314, "grad_norm": 0.163985, "tokens_per_sec": 107020} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:00.515805+00:00", "epoch": 0, "step": 24366, "train_loss": 3.4666709899902344, "perplexity": 32.02993678466682, "lr": 0.0026291804804649314, "grad_norm": 0.177238, "tokens_per_sec": 107088} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:00.821310+00:00", "epoch": 0, "step": 24367, "train_loss": 3.4970266819000244, "perplexity": 33.01713542159589, "lr": 0.0026291804804649314, "grad_norm": 0.170525, "tokens_per_sec": 107261} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:01.126114+00:00", "epoch": 0, "step": 24368, "train_loss": 3.5781846046447754, "perplexity": 35.80847526538043, "lr": 0.0026291804804649314, "grad_norm": 0.192853, "tokens_per_sec": 107502} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:01.430356+00:00", "epoch": 0, "step": 24369, "train_loss": 3.4902830123901367, "perplexity": 32.795227848790844, "lr": 0.0026291804804649314, "grad_norm": 0.167272, "tokens_per_sec": 107704} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:01.736294+00:00", "epoch": 0, "step": 24370, "train_loss": 3.6167757511138916, "perplexity": 37.21737607876451, "lr": 0.0026291804804649314, "grad_norm": 0.194219, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:02.041267+00:00", "epoch": 0, "step": 24371, "train_loss": 3.573038339614868, "perplexity": 35.624668725593416, "lr": 0.0026291804804649314, "grad_norm": 0.168911, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:02.347038+00:00", "epoch": 0, "step": 24372, "train_loss": 3.492656946182251, "perplexity": 32.873174031333654, "lr": 0.0026291804804649314, "grad_norm": 0.187256, "tokens_per_sec": 107167} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:02.652703+00:00", "epoch": 0, "step": 24373, "train_loss": 3.576331377029419, "perplexity": 35.74217546344203, "lr": 0.0026291804804649314, "grad_norm": 0.164473, "tokens_per_sec": 107212} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:02.958061+00:00", "epoch": 0, "step": 24374, "train_loss": 3.50748610496521, "perplexity": 33.36428795210872, "lr": 0.0026291804804649314, "grad_norm": 0.17649, "tokens_per_sec": 107299} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:03.263241+00:00", "epoch": 0, "step": 24375, "train_loss": 3.4395482540130615, "perplexity": 31.17287276686344, "lr": 0.0026291804804649314, "grad_norm": 0.200053, "tokens_per_sec": 107373} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:03.568694+00:00", "epoch": 0, "step": 24376, "train_loss": 3.5165743827819824, "perplexity": 33.66889394550227, "lr": 0.0026291804804649314, "grad_norm": 0.200506, "tokens_per_sec": 107277} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:03.873995+00:00", "epoch": 0, "step": 24377, "train_loss": 3.452841281890869, "perplexity": 31.59002107296552, "lr": 0.0026291804804649314, "grad_norm": 0.250766, "tokens_per_sec": 107330} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:04.179896+00:00", "epoch": 0, "step": 24378, "train_loss": 3.472607135772705, "perplexity": 32.220636609818364, "lr": 0.0026291804804649314, "grad_norm": 0.198397, "tokens_per_sec": 107129} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:04.484526+00:00", "epoch": 0, "step": 24379, "train_loss": 3.602320909500122, "perplexity": 36.683274280395366, "lr": 0.0026291804804649314, "grad_norm": 0.205755, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:04.789398+00:00", "epoch": 0, "step": 24380, "train_loss": 3.4135589599609375, "perplexity": 30.373148952272903, "lr": 0.0026291804804649314, "grad_norm": 0.197458, "tokens_per_sec": 107576} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:05.094171+00:00", "epoch": 0, "step": 24381, "train_loss": 3.553903341293335, "perplexity": 34.94947130184129, "lr": 0.0026291804804649314, "grad_norm": 0.177735, "tokens_per_sec": 107421} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:05.400516+00:00", "epoch": 0, "step": 24382, "train_loss": 3.4879608154296875, "perplexity": 32.719159227707436, "lr": 0.0026291804804649314, "grad_norm": 0.176919, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:05.705584+00:00", "epoch": 0, "step": 24383, "train_loss": 3.497706174850464, "perplexity": 33.039577956267124, "lr": 0.0026291804804649314, "grad_norm": 0.166522, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:06.010500+00:00", "epoch": 0, "step": 24384, "train_loss": 3.4745006561279297, "perplexity": 32.28170483981937, "lr": 0.0026291804804649314, "grad_norm": 0.156445, "tokens_per_sec": 107471} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:06.314841+00:00", "epoch": 0, "step": 24385, "train_loss": 3.5129623413085938, "perplexity": 33.54749987639819, "lr": 0.0026291804804649314, "grad_norm": 0.193325, "tokens_per_sec": 107663} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:06.619858+00:00", "epoch": 0, "step": 24386, "train_loss": 3.5923409461975098, "perplexity": 36.318997308148624, "lr": 0.0026291804804649314, "grad_norm": 0.190265, "tokens_per_sec": 107430} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:06.924837+00:00", "epoch": 0, "step": 24387, "train_loss": 3.503134250640869, "perplexity": 33.21940691055342, "lr": 0.0026291804804649314, "grad_norm": 0.178775, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:07.230730+00:00", "epoch": 0, "step": 24388, "train_loss": 3.5775232315063477, "perplexity": 35.78480033155748, "lr": 0.0026291804804649314, "grad_norm": 0.176718, "tokens_per_sec": 107123} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:07.536068+00:00", "epoch": 0, "step": 24389, "train_loss": 3.419386863708496, "perplexity": 30.550677548190077, "lr": 0.0026291804804649314, "grad_norm": 0.18447, "tokens_per_sec": 107317} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:07.843087+00:00", "epoch": 0, "step": 24390, "train_loss": 3.5797438621520996, "perplexity": 35.86435345218363, "lr": 0.0026291804804649314, "grad_norm": 0.179048, "tokens_per_sec": 106797} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:08.148208+00:00", "epoch": 0, "step": 24391, "train_loss": 3.4920380115509033, "perplexity": 32.852833980713534, "lr": 0.0026291804804649314, "grad_norm": 0.174628, "tokens_per_sec": 107326} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:08.453984+00:00", "epoch": 0, "step": 24392, "train_loss": 3.453772783279419, "perplexity": 31.619460930964873, "lr": 0.0026291804804649314, "grad_norm": 0.158481, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:08.759761+00:00", "epoch": 0, "step": 24393, "train_loss": 3.4116289615631104, "perplexity": 30.314585355460657, "lr": 0.0026291804804649314, "grad_norm": 0.184001, "tokens_per_sec": 107164} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:09.064744+00:00", "epoch": 0, "step": 24394, "train_loss": 3.549664258956909, "perplexity": 34.80163118996994, "lr": 0.0026291804804649314, "grad_norm": 0.167989, "tokens_per_sec": 107442} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:09.370473+00:00", "epoch": 0, "step": 24395, "train_loss": 3.5137839317321777, "perplexity": 33.57507350659604, "lr": 0.0026291804804649314, "grad_norm": 0.154623, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:09.676821+00:00", "epoch": 0, "step": 24396, "train_loss": 3.4962353706359863, "perplexity": 32.99101892493229, "lr": 0.0026291804804649314, "grad_norm": 0.155436, "tokens_per_sec": 106962} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:09.982548+00:00", "epoch": 0, "step": 24397, "train_loss": 3.5119576454162598, "perplexity": 33.51381176706061, "lr": 0.0026291804804649314, "grad_norm": 0.15761, "tokens_per_sec": 107181} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:10.286937+00:00", "epoch": 0, "step": 24398, "train_loss": 3.504640817642212, "perplexity": 33.2694918915255, "lr": 0.0026291804804649314, "grad_norm": 0.146565, "tokens_per_sec": 107651} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:10.591694+00:00", "epoch": 0, "step": 24399, "train_loss": 3.474973201751709, "perplexity": 32.29696302297947, "lr": 0.0026291804804649314, "grad_norm": 0.163056, "tokens_per_sec": 107523} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:10.897245+00:00", "epoch": 0, "step": 24400, "train_loss": 3.491826295852661, "perplexity": 32.84587925626622, "lr": 0.0026291804804649314, "grad_norm": 0.149102, "tokens_per_sec": 107313} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:11.204098+00:00", "epoch": 0, "step": 24401, "train_loss": 3.4813084602355957, "perplexity": 32.502222130210534, "lr": 0.0026291804804649314, "grad_norm": 0.16687, "tokens_per_sec": 106717} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:11.508709+00:00", "epoch": 0, "step": 24402, "train_loss": 3.3913819789886475, "perplexity": 29.706978335134945, "lr": 0.0026291804804649314, "grad_norm": 0.172613, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:11.813306+00:00", "epoch": 0, "step": 24403, "train_loss": 3.4843790531158447, "perplexity": 32.60217660328713, "lr": 0.0026291804804649314, "grad_norm": 0.179285, "tokens_per_sec": 107574} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:12.118521+00:00", "epoch": 0, "step": 24404, "train_loss": 3.4576706886291504, "perplexity": 31.742951117056126, "lr": 0.0026291804804649314, "grad_norm": 0.173427, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:12.424233+00:00", "epoch": 0, "step": 24405, "train_loss": 3.455860137939453, "perplexity": 31.685530891818058, "lr": 0.0026291804804649314, "grad_norm": 0.187185, "tokens_per_sec": 107187} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:12.729987+00:00", "epoch": 0, "step": 24406, "train_loss": 3.4459125995635986, "perplexity": 31.371900369374043, "lr": 0.0026291804804649314, "grad_norm": 0.264279, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:13.035110+00:00", "epoch": 0, "step": 24407, "train_loss": 3.4536266326904297, "perplexity": 31.6148400658056, "lr": 0.0026291804804649314, "grad_norm": 0.269044, "tokens_per_sec": 107393} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:13.340782+00:00", "epoch": 0, "step": 24408, "train_loss": 3.514646291732788, "perplexity": 33.60403979489478, "lr": 0.0026291804804649314, "grad_norm": 0.191511, "tokens_per_sec": 107200} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:13.645850+00:00", "epoch": 0, "step": 24409, "train_loss": 3.482123613357544, "perplexity": 32.52872721943568, "lr": 0.0026291804804649314, "grad_norm": 0.223139, "tokens_per_sec": 107413} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:13.951382+00:00", "epoch": 0, "step": 24410, "train_loss": 3.463115692138672, "perplexity": 31.91626301102991, "lr": 0.0026291804804649314, "grad_norm": 0.229859, "tokens_per_sec": 107311} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:14.256787+00:00", "epoch": 0, "step": 24411, "train_loss": 3.434028148651123, "perplexity": 31.001269295130218, "lr": 0.0026291804804649314, "grad_norm": 0.181174, "tokens_per_sec": 107231} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:14.561090+00:00", "epoch": 0, "step": 24412, "train_loss": 3.4274537563323975, "perplexity": 30.798123300845997, "lr": 0.0026291804804649314, "grad_norm": 0.20729, "tokens_per_sec": 107684} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:14.866659+00:00", "epoch": 0, "step": 24413, "train_loss": 3.430875301361084, "perplexity": 30.903680948661442, "lr": 0.0026291804804649314, "grad_norm": 0.2048, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:15.172459+00:00", "epoch": 0, "step": 24414, "train_loss": 3.4794914722442627, "perplexity": 32.443219602583184, "lr": 0.0026291804804649314, "grad_norm": 0.150768, "tokens_per_sec": 107152} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:15.477822+00:00", "epoch": 0, "step": 24415, "train_loss": 3.4407243728637695, "perplexity": 31.209557338634784, "lr": 0.0026291804804649314, "grad_norm": 0.188224, "tokens_per_sec": 107310} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:15.782558+00:00", "epoch": 0, "step": 24416, "train_loss": 3.5012311935424805, "perplexity": 33.15624859842645, "lr": 0.0026291804804649314, "grad_norm": 0.176077, "tokens_per_sec": 107530} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:16.088185+00:00", "epoch": 0, "step": 24417, "train_loss": 3.445563793182373, "perplexity": 31.360959558557273, "lr": 0.0026291804804649314, "grad_norm": 0.166645, "tokens_per_sec": 107214} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:16.393042+00:00", "epoch": 0, "step": 24418, "train_loss": 3.418030023574829, "perplexity": 30.509253272197288, "lr": 0.0026291804804649314, "grad_norm": 0.167022, "tokens_per_sec": 107486} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:16.698145+00:00", "epoch": 0, "step": 24419, "train_loss": 3.572894334793091, "perplexity": 35.61953897088628, "lr": 0.0026291804804649314, "grad_norm": 0.162523, "tokens_per_sec": 107400} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:17.003756+00:00", "epoch": 0, "step": 24420, "train_loss": 3.5903663635253906, "perplexity": 36.24735320227844, "lr": 0.0026291804804649314, "grad_norm": 0.181886, "tokens_per_sec": 107281} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:17.308905+00:00", "epoch": 0, "step": 24421, "train_loss": 3.4478118419647217, "perplexity": 31.431539829728074, "lr": 0.0026291804804649314, "grad_norm": 0.182615, "tokens_per_sec": 107324} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:17.613341+00:00", "epoch": 0, "step": 24422, "train_loss": 3.528123140335083, "perplexity": 34.05998178132929, "lr": 0.0026291804804649314, "grad_norm": 0.164709, "tokens_per_sec": 107640} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:17.918329+00:00", "epoch": 0, "step": 24423, "train_loss": 3.418006658554077, "perplexity": 30.508540431189264, "lr": 0.0026291804804649314, "grad_norm": 0.167888, "tokens_per_sec": 107437} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:18.224249+00:00", "epoch": 0, "step": 24424, "train_loss": 3.54345703125, "perplexity": 34.586278603328196, "lr": 0.0026291804804649314, "grad_norm": 0.179469, "tokens_per_sec": 107113} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:18.530077+00:00", "epoch": 0, "step": 24425, "train_loss": 3.5020813941955566, "perplexity": 33.18445004938564, "lr": 0.0026291804804649314, "grad_norm": 0.169397, "tokens_per_sec": 107144} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:18.834843+00:00", "epoch": 0, "step": 24426, "train_loss": 3.449904680252075, "perplexity": 31.49738984239297, "lr": 0.0026291804804649314, "grad_norm": 0.182848, "tokens_per_sec": 107519} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:19.140260+00:00", "epoch": 0, "step": 24427, "train_loss": 3.4849233627319336, "perplexity": 32.61992711196576, "lr": 0.0026291804804649314, "grad_norm": 0.200696, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:19.445395+00:00", "epoch": 0, "step": 24428, "train_loss": 3.513461112976074, "perplexity": 33.56423659240425, "lr": 0.0026291804804649314, "grad_norm": 0.184428, "tokens_per_sec": 107380} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:19.750339+00:00", "epoch": 0, "step": 24429, "train_loss": 3.6844942569732666, "perplexity": 39.82497615188731, "lr": 0.0026291804804649314, "grad_norm": 0.165781, "tokens_per_sec": 107456} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:20.055911+00:00", "epoch": 0, "step": 24430, "train_loss": 3.482461929321289, "perplexity": 32.53973406892579, "lr": 0.0026291804804649314, "grad_norm": 0.172309, "tokens_per_sec": 107300} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:20.362068+00:00", "epoch": 0, "step": 24431, "train_loss": 3.4394192695617676, "perplexity": 31.168852210274565, "lr": 0.0026291804804649314, "grad_norm": 0.19172, "tokens_per_sec": 106966} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:20.667028+00:00", "epoch": 0, "step": 24432, "train_loss": 3.4404330253601074, "perplexity": 31.200465836471317, "lr": 0.0026291804804649314, "grad_norm": 0.179835, "tokens_per_sec": 107450} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:20.972660+00:00", "epoch": 0, "step": 24433, "train_loss": 3.484971523284912, "perplexity": 32.62149814352416, "lr": 0.0026291804804649314, "grad_norm": 0.190811, "tokens_per_sec": 107219} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:21.277784+00:00", "epoch": 0, "step": 24434, "train_loss": 3.5326921939849854, "perplexity": 34.21595972985759, "lr": 0.0026291804804649314, "grad_norm": 0.207888, "tokens_per_sec": 107389} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:21.583543+00:00", "epoch": 0, "step": 24435, "train_loss": 3.461951494216919, "perplexity": 31.879127784533395, "lr": 0.0026291804804649314, "grad_norm": 0.170641, "tokens_per_sec": 107169} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:21.889542+00:00", "epoch": 0, "step": 24436, "train_loss": 3.4754841327667236, "perplexity": 32.313468759365946, "lr": 0.0026291804804649314, "grad_norm": 0.217918, "tokens_per_sec": 107085} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:22.194970+00:00", "epoch": 0, "step": 24437, "train_loss": 3.507469892501831, "perplexity": 33.3637470391969, "lr": 0.0026291804804649314, "grad_norm": 0.16405, "tokens_per_sec": 107286} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:22.499433+00:00", "epoch": 0, "step": 24438, "train_loss": 3.414024829864502, "perplexity": 30.3873021847728, "lr": 0.0026291804804649314, "grad_norm": 0.177911, "tokens_per_sec": 107626} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:22.804376+00:00", "epoch": 0, "step": 24439, "train_loss": 3.52329421043396, "perplexity": 33.8959049933636, "lr": 0.0026291804804649314, "grad_norm": 0.182478, "tokens_per_sec": 107461} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:23.109748+00:00", "epoch": 0, "step": 24440, "train_loss": 3.5136866569519043, "perplexity": 33.571807657542976, "lr": 0.0026291804804649314, "grad_norm": 0.193388, "tokens_per_sec": 107404} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:23.414944+00:00", "epoch": 0, "step": 24441, "train_loss": 3.4471750259399414, "perplexity": 31.41153009341757, "lr": 0.0026291804804649314, "grad_norm": 0.169203, "tokens_per_sec": 107263} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:23.720025+00:00", "epoch": 0, "step": 24442, "train_loss": 3.4596753120422363, "perplexity": 31.806647602468374, "lr": 0.0026291804804649314, "grad_norm": 0.192478, "tokens_per_sec": 107407} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:24.025539+00:00", "epoch": 0, "step": 24443, "train_loss": 3.559030771255493, "perplexity": 35.129132474580445, "lr": 0.0026291804804649314, "grad_norm": 0.19686, "tokens_per_sec": 107257} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:24.331199+00:00", "epoch": 0, "step": 24444, "train_loss": 3.4672458171844482, "perplexity": 32.04835375614049, "lr": 0.0026291804804649314, "grad_norm": 0.202407, "tokens_per_sec": 107204} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:24.636485+00:00", "epoch": 0, "step": 24445, "train_loss": 3.438746213912964, "perplexity": 31.14788089645, "lr": 0.0026291804804649314, "grad_norm": 0.186653, "tokens_per_sec": 107336} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:24.941345+00:00", "epoch": 0, "step": 24446, "train_loss": 3.4295506477355957, "perplexity": 30.862771377140007, "lr": 0.0026291804804649314, "grad_norm": 0.174556, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:25.246800+00:00", "epoch": 0, "step": 24447, "train_loss": 3.475131034851074, "perplexity": 32.302060955054095, "lr": 0.0026291804804649314, "grad_norm": 0.201998, "tokens_per_sec": 107276} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:25.550845+00:00", "epoch": 0, "step": 24448, "train_loss": 3.389296293258667, "perplexity": 29.645083483382365, "lr": 0.0026291804804649314, "grad_norm": 0.165808, "tokens_per_sec": 107774} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:25.855417+00:00", "epoch": 0, "step": 24449, "train_loss": 3.4891090393066406, "perplexity": 32.756749724589554, "lr": 0.0026291804804649314, "grad_norm": 0.169653, "tokens_per_sec": 107587} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:26.160590+00:00", "epoch": 0, "step": 24450, "train_loss": 3.5137884616851807, "perplexity": 33.575225600445584, "lr": 0.0026291804804649314, "grad_norm": 0.174794, "tokens_per_sec": 107444} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:26.465237+00:00", "epoch": 0, "step": 24451, "train_loss": 3.508842945098877, "perplexity": 33.4095886830083, "lr": 0.0026291804804649314, "grad_norm": 0.187535, "tokens_per_sec": 107490} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:26.770304+00:00", "epoch": 0, "step": 24452, "train_loss": 3.4643592834472656, "perplexity": 31.955978488105877, "lr": 0.0026291804804649314, "grad_norm": 0.172553, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:27.075545+00:00", "epoch": 0, "step": 24453, "train_loss": 3.3964993953704834, "perplexity": 29.859390959538093, "lr": 0.0026291804804649314, "grad_norm": 0.186856, "tokens_per_sec": 107350} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:27.380386+00:00", "epoch": 0, "step": 24454, "train_loss": 3.483482599258423, "perplexity": 32.57296335243286, "lr": 0.0026291804804649314, "grad_norm": 0.177621, "tokens_per_sec": 107493} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:27.685006+00:00", "epoch": 0, "step": 24455, "train_loss": 3.435142755508423, "perplexity": 31.035842786817668, "lr": 0.0026291804804649314, "grad_norm": 0.18768, "tokens_per_sec": 107569} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:27.989633+00:00", "epoch": 0, "step": 24456, "train_loss": 3.484095811843872, "perplexity": 32.59294362895836, "lr": 0.0026291804804649314, "grad_norm": 0.17439, "tokens_per_sec": 107568} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:28.295182+00:00", "epoch": 0, "step": 24457, "train_loss": 3.4893484115600586, "perplexity": 32.764591720126404, "lr": 0.0026291804804649314, "grad_norm": 0.197996, "tokens_per_sec": 107243} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:28.600294+00:00", "epoch": 0, "step": 24458, "train_loss": 3.4464457035064697, "perplexity": 31.388629311892725, "lr": 0.0026291804804649314, "grad_norm": 0.170997, "tokens_per_sec": 107397} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:28.905083+00:00", "epoch": 0, "step": 24459, "train_loss": 3.402374744415283, "perplexity": 30.035341683502224, "lr": 0.0026291804804649314, "grad_norm": 0.169617, "tokens_per_sec": 107511} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:29.210999+00:00", "epoch": 0, "step": 24460, "train_loss": 3.547682285308838, "perplexity": 34.732723583128994, "lr": 0.0026291804804649314, "grad_norm": 0.177258, "tokens_per_sec": 107174} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:29.516059+00:00", "epoch": 0, "step": 24461, "train_loss": 3.591200590133667, "perplexity": 36.27760432519081, "lr": 0.0026291804804649314, "grad_norm": 0.146371, "tokens_per_sec": 107356} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:29.821726+00:00", "epoch": 0, "step": 24462, "train_loss": 3.5190656185150146, "perplexity": 33.75287566288428, "lr": 0.0026291804804649314, "grad_norm": 0.152696, "tokens_per_sec": 107203} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:30.127080+00:00", "epoch": 0, "step": 24463, "train_loss": 3.471832513809204, "perplexity": 32.195687461349145, "lr": 0.0026291804804649314, "grad_norm": 0.187643, "tokens_per_sec": 107312} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:30.431942+00:00", "epoch": 0, "step": 24464, "train_loss": 3.456871509552002, "perplexity": 31.71759294888481, "lr": 0.0026291804804649314, "grad_norm": 0.183586, "tokens_per_sec": 107484} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:30.736543+00:00", "epoch": 0, "step": 24465, "train_loss": 3.501535415649414, "perplexity": 33.16633699671216, "lr": 0.0026291804804649314, "grad_norm": 0.163902, "tokens_per_sec": 107577} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:31.040880+00:00", "epoch": 0, "step": 24466, "train_loss": 3.4476563930511475, "perplexity": 31.426654210751497, "lr": 0.0026291804804649314, "grad_norm": 0.190456, "tokens_per_sec": 107670} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:31.346772+00:00", "epoch": 0, "step": 24467, "train_loss": 3.434584856033325, "perplexity": 31.01853273550094, "lr": 0.0026291804804649314, "grad_norm": 0.167889, "tokens_per_sec": 107125} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:31.652674+00:00", "epoch": 0, "step": 24468, "train_loss": 3.563323974609375, "perplexity": 35.28027319070699, "lr": 0.0026291804804649314, "grad_norm": 0.158562, "tokens_per_sec": 107120} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:31.958073+00:00", "epoch": 0, "step": 24469, "train_loss": 3.4561097621917725, "perplexity": 31.69344135605705, "lr": 0.0026291804804649314, "grad_norm": 0.196293, "tokens_per_sec": 107294} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:32.262642+00:00", "epoch": 0, "step": 24470, "train_loss": 3.547764539718628, "perplexity": 34.73558062030806, "lr": 0.0026291804804649314, "grad_norm": 0.175041, "tokens_per_sec": 107653} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:32.567643+00:00", "epoch": 0, "step": 24471, "train_loss": 3.6139862537384033, "perplexity": 37.113702971027124, "lr": 0.0026291804804649314, "grad_norm": 0.159247, "tokens_per_sec": 107370} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:32.873164+00:00", "epoch": 0, "step": 24472, "train_loss": 3.5381076335906982, "perplexity": 34.401756825729755, "lr": 0.0026291804804649314, "grad_norm": 0.169822, "tokens_per_sec": 107254} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:33.178731+00:00", "epoch": 0, "step": 24473, "train_loss": 3.5964696407318115, "perplexity": 36.46925732924422, "lr": 0.0026291804804649314, "grad_norm": 0.172628, "tokens_per_sec": 107236} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:33.485553+00:00", "epoch": 0, "step": 24474, "train_loss": 3.4782233238220215, "perplexity": 32.40210286140729, "lr": 0.0026291804804649314, "grad_norm": 0.171217, "tokens_per_sec": 106802} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:33.791450+00:00", "epoch": 0, "step": 24475, "train_loss": 3.522761821746826, "perplexity": 33.87786399984152, "lr": 0.0026291804804649314, "grad_norm": 0.162671, "tokens_per_sec": 107118} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:34.095571+00:00", "epoch": 0, "step": 24476, "train_loss": 3.48777437210083, "perplexity": 32.71305952738549, "lr": 0.0026291804804649314, "grad_norm": 0.16934, "tokens_per_sec": 107747} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:34.400775+00:00", "epoch": 0, "step": 24477, "train_loss": 3.552549123764038, "perplexity": 34.9021741477105, "lr": 0.0026291804804649314, "grad_norm": 0.170512, "tokens_per_sec": 107364} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:34.706046+00:00", "epoch": 0, "step": 24478, "train_loss": 3.5137760639190674, "perplexity": 33.574809345231714, "lr": 0.0026291804804649314, "grad_norm": 0.176697, "tokens_per_sec": 107341} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:35.011214+00:00", "epoch": 0, "step": 24479, "train_loss": 3.6175575256347656, "perplexity": 37.246483051178934, "lr": 0.0026291804804649314, "grad_norm": 0.168894, "tokens_per_sec": 107382} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:35.316592+00:00", "epoch": 0, "step": 24480, "train_loss": 3.4320223331451416, "perplexity": 30.939148790436146, "lr": 0.0026291804804649314, "grad_norm": 0.167947, "tokens_per_sec": 107361} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:35.620709+00:00", "epoch": 0, "step": 24481, "train_loss": 3.592700242996216, "perplexity": 36.33204895218021, "lr": 0.0026291804804649314, "grad_norm": 0.174697, "tokens_per_sec": 107685} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:35.925455+00:00", "epoch": 0, "step": 24482, "train_loss": 3.4468255043029785, "perplexity": 31.40055300247696, "lr": 0.0026291804804649314, "grad_norm": 0.159172, "tokens_per_sec": 107526} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:36.230394+00:00", "epoch": 0, "step": 24483, "train_loss": 3.4528815746307373, "perplexity": 31.591293947110664, "lr": 0.0026291804804649314, "grad_norm": 0.167786, "tokens_per_sec": 107457} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:36.535580+00:00", "epoch": 0, "step": 24484, "train_loss": 3.5038790702819824, "perplexity": 33.24415859391223, "lr": 0.0026291804804649314, "grad_norm": 0.19778, "tokens_per_sec": 107371} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:36.840754+00:00", "epoch": 0, "step": 24485, "train_loss": 3.4361355304718018, "perplexity": 31.066669694062718, "lr": 0.0026291804804649314, "grad_norm": 0.214531, "tokens_per_sec": 107384} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:37.144940+00:00", "epoch": 0, "step": 24486, "train_loss": 3.524928092956543, "perplexity": 33.95133218855157, "lr": 0.0026291804804649314, "grad_norm": 0.182918, "tokens_per_sec": 107714} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:37.449505+00:00", "epoch": 0, "step": 24487, "train_loss": 3.3479583263397217, "perplexity": 28.44459972867641, "lr": 0.0026291804804649314, "grad_norm": 0.174133, "tokens_per_sec": 107590} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:37.753985+00:00", "epoch": 0, "step": 24488, "train_loss": 3.460538387298584, "perplexity": 31.834110982778647, "lr": 0.0026291804804649314, "grad_norm": 0.171368, "tokens_per_sec": 107621} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:38.058567+00:00", "epoch": 0, "step": 24489, "train_loss": 3.488135576248169, "perplexity": 32.72487775442675, "lr": 0.0026291804804649314, "grad_norm": 0.17295, "tokens_per_sec": 107583} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:38.364765+00:00", "epoch": 0, "step": 24490, "train_loss": 3.5098748207092285, "perplexity": 33.44408101556, "lr": 0.0026291804804649314, "grad_norm": 0.159692, "tokens_per_sec": 107079} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:38.671104+00:00", "epoch": 0, "step": 24491, "train_loss": 3.407099962234497, "perplexity": 30.177601053793595, "lr": 0.0026291804804649314, "grad_norm": 0.169294, "tokens_per_sec": 106908} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:38.974832+00:00", "epoch": 0, "step": 24492, "train_loss": 3.5376131534576416, "perplexity": 34.38475004554118, "lr": 0.0026291804804649314, "grad_norm": 0.1638, "tokens_per_sec": 107882} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:39.280538+00:00", "epoch": 0, "step": 24493, "train_loss": 3.456634759902954, "perplexity": 31.7100847087072, "lr": 0.0026291804804649314, "grad_norm": 0.177359, "tokens_per_sec": 107188} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:39.585193+00:00", "epoch": 0, "step": 24494, "train_loss": 3.4959704875946045, "perplexity": 32.982281320773836, "lr": 0.0026291804804649314, "grad_norm": 0.182452, "tokens_per_sec": 107558} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:39.890258+00:00", "epoch": 0, "step": 24495, "train_loss": 3.50842022895813, "perplexity": 33.39546889515764, "lr": 0.0026291804804649314, "grad_norm": 0.178503, "tokens_per_sec": 107414} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:40.195753+00:00", "epoch": 0, "step": 24496, "train_loss": 3.5277509689331055, "perplexity": 34.04730798871628, "lr": 0.0026291804804649314, "grad_norm": 0.160465, "tokens_per_sec": 107270} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:40.500525+00:00", "epoch": 0, "step": 24497, "train_loss": 3.5556681156158447, "perplexity": 35.01120368722261, "lr": 0.0026291804804649314, "grad_norm": 0.163466, "tokens_per_sec": 107508} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:40.806964+00:00", "epoch": 0, "step": 24498, "train_loss": 3.4498395919799805, "perplexity": 31.49533979843027, "lr": 0.0026291804804649314, "grad_norm": 0.18358, "tokens_per_sec": 106933} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:41.112946+00:00", "epoch": 0, "step": 24499, "train_loss": 3.50484037399292, "perplexity": 33.27613169240246, "lr": 0.0026291804804649314, "grad_norm": 0.193802, "tokens_per_sec": 107091} +{"stage": "pretraining", "type": "step", "timestamp": "2026-04-09T23:06:41.417326+00:00", "epoch": 0, "step": 24500, "train_loss": 3.488776206970215, "perplexity": 32.745849033196926, "lr": 0.0026291804804649314, "grad_norm": 0.167863, "tokens_per_sec": 107752} +{"stage": "pretraining", "type": "eval", "timestamp": "2026-04-09T23:06:44.490560+00:00", "step": 24500, "epoch": 0, "val_loss": 3.4519856452941893, "val_ppl": 31.563003055296303, "eval_train_loss": 3.488776206970215, "eval_train_ppl": 32.745849033196926} +{"stage": "pretraining", "type": "best_checkpoint", "timestamp": "2026-04-09T23:06:45.479491+00:00", "step": 24500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4520_epoch_0000_step_0024500.pt", "val_loss": 3.4519856452941893} +{"stage": "pretraining", "type": "early_stopping", "timestamp": "2026-04-09T23:06:46.622172+00:00", "step": 24500, "epoch": 0, "val_loss": 3.4519856452941893, "best_val_loss": 3.4509812951087953, "patience": 5, "min_delta": 0.001} +{"stage": "pretraining", "type": "checkpoint", "timestamp": "2026-04-09T23:06:47.575417+00:00", "step": 24500, "epoch": 0, "path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/last_epoch_0000_step_0024500.pt", "category": "last"} +{"stage": "pretraining", "type": "training_complete", "timestamp": "2026-04-09T23:06:47.596565+00:00", "epochs": 1, "total_steps": 24500, "total_time_hours": 2.16, "best_val_loss": 3.4509812951087953, "best_checkpoint_path": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints/best_loss_3p4510_epoch_0000_step_0022000.pt", "checkpoint_dir": "/home/user/code/ParrotLLM/runs/big_run/exp_c_8b/run_20260409_205717/checkpoints", "stopped_early": true, "stop_reason": "validation loss stopped improving", "run_dir": "runs/big_run/exp_c_8b/run_20260409_205717"}