| { |
| "best_metric": 0.7813047170639038, |
| "best_model_checkpoint": "model/checkpoints/run1-python-codegen/checkpoint-10000", |
| "epoch": 4.999356582164458, |
| "eval_steps": 1000, |
| "global_step": 19425, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0025736713421696047, |
| "grad_norm": 1.7038387060165405, |
| "learning_rate": 4.997425997425998e-05, |
| "loss": 0.9335, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0051473426843392095, |
| "grad_norm": 1.1648929119110107, |
| "learning_rate": 4.994851994851995e-05, |
| "loss": 0.9433, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.007721014026508815, |
| "grad_norm": 3.9555516242980957, |
| "learning_rate": 4.992277992277993e-05, |
| "loss": 0.9713, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.010294685368678419, |
| "grad_norm": 0.9799448847770691, |
| "learning_rate": 4.9897039897039896e-05, |
| "loss": 0.9518, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.012868356710848025, |
| "grad_norm": 1.26755952835083, |
| "learning_rate": 4.9871299871299874e-05, |
| "loss": 1.0754, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.01544202805301763, |
| "grad_norm": 4.542232513427734, |
| "learning_rate": 4.9845559845559846e-05, |
| "loss": 0.9471, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.018015699395187233, |
| "grad_norm": 1.2068519592285156, |
| "learning_rate": 4.9819819819819824e-05, |
| "loss": 0.9576, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.020589370737356838, |
| "grad_norm": 1.019250750541687, |
| "learning_rate": 4.9794079794079796e-05, |
| "loss": 0.9516, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.023163042079526446, |
| "grad_norm": 1.2534639835357666, |
| "learning_rate": 4.9768339768339774e-05, |
| "loss": 1.0046, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.02573671342169605, |
| "grad_norm": 0.8234790563583374, |
| "learning_rate": 4.9742599742599746e-05, |
| "loss": 0.8847, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.028310384763865656, |
| "grad_norm": 0.9790653586387634, |
| "learning_rate": 4.971685971685972e-05, |
| "loss": 0.8495, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.03088405610603526, |
| "grad_norm": 4.180076599121094, |
| "learning_rate": 4.969111969111969e-05, |
| "loss": 0.8801, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.03345772744820486, |
| "grad_norm": 0.9933150410652161, |
| "learning_rate": 4.966537966537967e-05, |
| "loss": 0.9608, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.036031398790374466, |
| "grad_norm": 1.1884509325027466, |
| "learning_rate": 4.963963963963964e-05, |
| "loss": 0.9133, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.03860507013254407, |
| "grad_norm": 4.013946056365967, |
| "learning_rate": 4.961389961389962e-05, |
| "loss": 0.9142, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.041178741474713676, |
| "grad_norm": 2.0833263397216797, |
| "learning_rate": 4.958815958815959e-05, |
| "loss": 0.9864, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.04375241281688328, |
| "grad_norm": 3.602749824523926, |
| "learning_rate": 4.956241956241957e-05, |
| "loss": 0.876, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.04632608415905289, |
| "grad_norm": 1.2331393957138062, |
| "learning_rate": 4.953667953667954e-05, |
| "loss": 0.9233, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.0488997555012225, |
| "grad_norm": 1.0604982376098633, |
| "learning_rate": 4.951093951093951e-05, |
| "loss": 0.9497, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.0514734268433921, |
| "grad_norm": 1.0822978019714355, |
| "learning_rate": 4.948519948519948e-05, |
| "loss": 0.8719, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.05404709818556171, |
| "grad_norm": 0.9496659636497498, |
| "learning_rate": 4.945945945945946e-05, |
| "loss": 0.9186, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.05662076952773131, |
| "grad_norm": 1.0661075115203857, |
| "learning_rate": 4.943371943371944e-05, |
| "loss": 0.8779, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.059194440869900916, |
| "grad_norm": 1.1287846565246582, |
| "learning_rate": 4.940797940797941e-05, |
| "loss": 0.8501, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.06176811221207052, |
| "grad_norm": 1.16519296169281, |
| "learning_rate": 4.938223938223939e-05, |
| "loss": 0.9317, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.06434178355424013, |
| "grad_norm": 1.5323387384414673, |
| "learning_rate": 4.935649935649936e-05, |
| "loss": 0.9511, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.06691545489640972, |
| "grad_norm": 1.1095296144485474, |
| "learning_rate": 4.933075933075933e-05, |
| "loss": 0.9372, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.06948912623857934, |
| "grad_norm": 1.1270581483840942, |
| "learning_rate": 4.9305019305019304e-05, |
| "loss": 0.9549, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.07206279758074893, |
| "grad_norm": 1.0347909927368164, |
| "learning_rate": 4.927927927927928e-05, |
| "loss": 0.882, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.07463646892291854, |
| "grad_norm": 0.9273091554641724, |
| "learning_rate": 4.9253539253539254e-05, |
| "loss": 0.9191, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.07721014026508814, |
| "grad_norm": 0.9906871318817139, |
| "learning_rate": 4.922779922779923e-05, |
| "loss": 0.9259, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.07978381160725775, |
| "grad_norm": 0.8945180773735046, |
| "learning_rate": 4.9202059202059204e-05, |
| "loss": 0.8202, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.08235748294942735, |
| "grad_norm": 1.2765549421310425, |
| "learning_rate": 4.917631917631918e-05, |
| "loss": 0.9122, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.08493115429159696, |
| "grad_norm": 0.8597398400306702, |
| "learning_rate": 4.9150579150579154e-05, |
| "loss": 0.9114, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.08750482563376656, |
| "grad_norm": 3.5975146293640137, |
| "learning_rate": 4.9124839124839125e-05, |
| "loss": 0.8598, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.09007849697593617, |
| "grad_norm": 4.273248672485352, |
| "learning_rate": 4.90990990990991e-05, |
| "loss": 0.9325, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.09265216831810578, |
| "grad_norm": 0.9573097229003906, |
| "learning_rate": 4.9073359073359075e-05, |
| "loss": 0.992, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.09522583966027538, |
| "grad_norm": 1.2998064756393433, |
| "learning_rate": 4.904761904761905e-05, |
| "loss": 0.8984, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.097799511002445, |
| "grad_norm": 1.3857780694961548, |
| "learning_rate": 4.9021879021879025e-05, |
| "loss": 1.0102, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.10037318234461459, |
| "grad_norm": 1.2418707609176636, |
| "learning_rate": 4.8996138996139e-05, |
| "loss": 0.9023, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.1029468536867842, |
| "grad_norm": 1.2056580781936646, |
| "learning_rate": 4.8970398970398975e-05, |
| "loss": 0.8691, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.1055205250289538, |
| "grad_norm": 0.9754079580307007, |
| "learning_rate": 4.894465894465895e-05, |
| "loss": 0.9484, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.10809419637112341, |
| "grad_norm": 0.9066883325576782, |
| "learning_rate": 4.891891891891892e-05, |
| "loss": 0.9009, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.11066786771329301, |
| "grad_norm": 1.480106234550476, |
| "learning_rate": 4.88931788931789e-05, |
| "loss": 0.8764, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.11324153905546262, |
| "grad_norm": 1.1571600437164307, |
| "learning_rate": 4.886743886743887e-05, |
| "loss": 0.8711, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.11581521039763222, |
| "grad_norm": 0.9223948121070862, |
| "learning_rate": 4.884169884169885e-05, |
| "loss": 0.8466, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.11838888173980183, |
| "grad_norm": 1.7378956079483032, |
| "learning_rate": 4.881595881595882e-05, |
| "loss": 0.98, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.12096255308197143, |
| "grad_norm": 1.114259958267212, |
| "learning_rate": 4.87902187902188e-05, |
| "loss": 0.9181, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.12353622442414104, |
| "grad_norm": 1.0394957065582275, |
| "learning_rate": 4.876447876447877e-05, |
| "loss": 0.9139, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.12610989576631065, |
| "grad_norm": 1.100290298461914, |
| "learning_rate": 4.873873873873874e-05, |
| "loss": 0.8707, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.12868356710848025, |
| "grad_norm": 0.9463397264480591, |
| "learning_rate": 4.871299871299871e-05, |
| "loss": 0.8608, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.13125723845064985, |
| "grad_norm": 0.9843112230300903, |
| "learning_rate": 4.868725868725869e-05, |
| "loss": 0.931, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.13383090979281945, |
| "grad_norm": 1.050361156463623, |
| "learning_rate": 4.866151866151866e-05, |
| "loss": 0.9192, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.13640458113498907, |
| "grad_norm": 1.1644479036331177, |
| "learning_rate": 4.863577863577864e-05, |
| "loss": 0.8797, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.13897825247715867, |
| "grad_norm": 1.2110307216644287, |
| "learning_rate": 4.861003861003861e-05, |
| "loss": 0.9662, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.14155192381932827, |
| "grad_norm": 1.7485480308532715, |
| "learning_rate": 4.858429858429859e-05, |
| "loss": 0.9391, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.14412559516149787, |
| "grad_norm": 4.363419532775879, |
| "learning_rate": 4.855855855855856e-05, |
| "loss": 0.9487, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.1466992665036675, |
| "grad_norm": 1.028351068496704, |
| "learning_rate": 4.853281853281853e-05, |
| "loss": 0.7843, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.1492729378458371, |
| "grad_norm": 1.3750572204589844, |
| "learning_rate": 4.8507078507078505e-05, |
| "loss": 0.9074, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.1518466091880067, |
| "grad_norm": 4.166394233703613, |
| "learning_rate": 4.848133848133848e-05, |
| "loss": 0.9494, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.15442028053017628, |
| "grad_norm": 1.1192243099212646, |
| "learning_rate": 4.8455598455598455e-05, |
| "loss": 0.9937, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.1569939518723459, |
| "grad_norm": 1.243043065071106, |
| "learning_rate": 4.842985842985843e-05, |
| "loss": 0.8259, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.1595676232145155, |
| "grad_norm": 1.0789657831192017, |
| "learning_rate": 4.8404118404118405e-05, |
| "loss": 0.8466, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.1621412945566851, |
| "grad_norm": 1.2137137651443481, |
| "learning_rate": 4.837837837837838e-05, |
| "loss": 0.8358, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.1647149658988547, |
| "grad_norm": 5.4415974617004395, |
| "learning_rate": 4.8352638352638355e-05, |
| "loss": 0.9067, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.16728863724102433, |
| "grad_norm": 1.1608994007110596, |
| "learning_rate": 4.8326898326898327e-05, |
| "loss": 0.8828, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.16986230858319393, |
| "grad_norm": 1.287034034729004, |
| "learning_rate": 4.8301158301158305e-05, |
| "loss": 0.8905, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.17243597992536353, |
| "grad_norm": 1.0879331827163696, |
| "learning_rate": 4.8275418275418277e-05, |
| "loss": 0.8821, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.17500965126753312, |
| "grad_norm": 1.1660597324371338, |
| "learning_rate": 4.8249678249678255e-05, |
| "loss": 0.8955, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.17758332260970275, |
| "grad_norm": 0.8420181274414062, |
| "learning_rate": 4.8223938223938227e-05, |
| "loss": 0.8927, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.18015699395187235, |
| "grad_norm": 0.7901207208633423, |
| "learning_rate": 4.8198198198198205e-05, |
| "loss": 0.8926, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.18273066529404194, |
| "grad_norm": 1.2035242319107056, |
| "learning_rate": 4.8172458172458177e-05, |
| "loss": 0.8742, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.18530433663621157, |
| "grad_norm": 0.8758509159088135, |
| "learning_rate": 4.814671814671815e-05, |
| "loss": 0.9302, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.18787800797838117, |
| "grad_norm": 0.9391271471977234, |
| "learning_rate": 4.812097812097812e-05, |
| "loss": 0.8675, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.19045167932055077, |
| "grad_norm": 4.989656448364258, |
| "learning_rate": 4.80952380952381e-05, |
| "loss": 0.9314, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.19302535066272036, |
| "grad_norm": 2.094479560852051, |
| "learning_rate": 4.806949806949807e-05, |
| "loss": 0.8705, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.19559902200489, |
| "grad_norm": 1.1212397813796997, |
| "learning_rate": 4.804375804375805e-05, |
| "loss": 0.9013, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.1981726933470596, |
| "grad_norm": 3.019252300262451, |
| "learning_rate": 4.801801801801802e-05, |
| "loss": 0.8546, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.20074636468922918, |
| "grad_norm": 2.4471728801727295, |
| "learning_rate": 4.7992277992278e-05, |
| "loss": 0.8738, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.20332003603139878, |
| "grad_norm": 1.226980209350586, |
| "learning_rate": 4.796653796653797e-05, |
| "loss": 0.921, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.2058937073735684, |
| "grad_norm": 0.9777143001556396, |
| "learning_rate": 4.794079794079794e-05, |
| "loss": 0.893, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.208467378715738, |
| "grad_norm": 1.05388343334198, |
| "learning_rate": 4.791505791505791e-05, |
| "loss": 0.8911, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.2110410500579076, |
| "grad_norm": 3.421910285949707, |
| "learning_rate": 4.788931788931789e-05, |
| "loss": 0.9105, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.2136147214000772, |
| "grad_norm": 1.2801098823547363, |
| "learning_rate": 4.786357786357786e-05, |
| "loss": 0.8692, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.21618839274224683, |
| "grad_norm": 0.8514150381088257, |
| "learning_rate": 4.783783783783784e-05, |
| "loss": 0.8643, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.21876206408441642, |
| "grad_norm": 3.815889596939087, |
| "learning_rate": 4.781209781209782e-05, |
| "loss": 0.8395, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.22133573542658602, |
| "grad_norm": 1.0590102672576904, |
| "learning_rate": 4.778635778635779e-05, |
| "loss": 0.9168, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.22390940676875562, |
| "grad_norm": 0.9641560316085815, |
| "learning_rate": 4.776061776061776e-05, |
| "loss": 0.8777, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.22648307811092525, |
| "grad_norm": 1.1582986116409302, |
| "learning_rate": 4.7734877734877735e-05, |
| "loss": 0.9626, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.22905674945309484, |
| "grad_norm": 1.3446133136749268, |
| "learning_rate": 4.770913770913771e-05, |
| "loss": 0.8254, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.23163042079526444, |
| "grad_norm": 0.9514161944389343, |
| "learning_rate": 4.7683397683397685e-05, |
| "loss": 0.9177, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.23420409213743404, |
| "grad_norm": 1.3230582475662231, |
| "learning_rate": 4.765765765765766e-05, |
| "loss": 0.9428, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.23677776347960366, |
| "grad_norm": 3.5208494663238525, |
| "learning_rate": 4.7631917631917635e-05, |
| "loss": 0.9657, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.23935143482177326, |
| "grad_norm": 1.2036529779434204, |
| "learning_rate": 4.760617760617761e-05, |
| "loss": 0.9157, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.24192510616394286, |
| "grad_norm": 0.9241206645965576, |
| "learning_rate": 4.7580437580437585e-05, |
| "loss": 0.931, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.24449877750611246, |
| "grad_norm": 0.9766173958778381, |
| "learning_rate": 4.7554697554697556e-05, |
| "loss": 0.8657, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.24707244884828208, |
| "grad_norm": 1.0403867959976196, |
| "learning_rate": 4.752895752895753e-05, |
| "loss": 1.002, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.24964612019045168, |
| "grad_norm": 1.6326649188995361, |
| "learning_rate": 4.7503217503217506e-05, |
| "loss": 0.8585, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.2522197915326213, |
| "grad_norm": 0.8411473035812378, |
| "learning_rate": 4.747747747747748e-05, |
| "loss": 0.8956, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.2547934628747909, |
| "grad_norm": 1.342337727546692, |
| "learning_rate": 4.7451737451737456e-05, |
| "loss": 0.9417, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.2573671342169605, |
| "grad_norm": 1.0992860794067383, |
| "learning_rate": 4.742599742599743e-05, |
| "loss": 0.8915, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.2573671342169605, |
| "eval_loss": 0.8595815300941467, |
| "eval_runtime": 395.6739, |
| "eval_samples_per_second": 49.096, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.0003721938368800249, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.2599408055591301, |
| "grad_norm": 0.776602566242218, |
| "learning_rate": 4.7400257400257406e-05, |
| "loss": 0.8822, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.2625144769012997, |
| "grad_norm": 1.099664568901062, |
| "learning_rate": 4.737451737451737e-05, |
| "loss": 0.8394, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.2650881482434693, |
| "grad_norm": 0.80992192029953, |
| "learning_rate": 4.734877734877735e-05, |
| "loss": 0.8998, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.2676618195856389, |
| "grad_norm": 0.8013533353805542, |
| "learning_rate": 4.732303732303733e-05, |
| "loss": 0.8744, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.2702354909278085, |
| "grad_norm": 0.8478692770004272, |
| "learning_rate": 4.72972972972973e-05, |
| "loss": 0.8574, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.27280916226997814, |
| "grad_norm": 1.3109067678451538, |
| "learning_rate": 4.727155727155728e-05, |
| "loss": 0.914, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.2753828336121477, |
| "grad_norm": 1.043544888496399, |
| "learning_rate": 4.724581724581725e-05, |
| "loss": 0.8326, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.27795650495431734, |
| "grad_norm": 0.9514428973197937, |
| "learning_rate": 4.722007722007723e-05, |
| "loss": 0.9688, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.28053017629648697, |
| "grad_norm": 2.1324360370635986, |
| "learning_rate": 4.719433719433719e-05, |
| "loss": 0.9619, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.28310384763865654, |
| "grad_norm": 1.0498496294021606, |
| "learning_rate": 4.716859716859717e-05, |
| "loss": 0.8134, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.28567751898082616, |
| "grad_norm": 1.7257237434387207, |
| "learning_rate": 4.714285714285714e-05, |
| "loss": 0.8972, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.28825119032299573, |
| "grad_norm": 3.86918568611145, |
| "learning_rate": 4.711711711711712e-05, |
| "loss": 0.9396, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.29082486166516536, |
| "grad_norm": 0.8664824962615967, |
| "learning_rate": 4.709137709137709e-05, |
| "loss": 0.8619, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.293398533007335, |
| "grad_norm": 1.012129783630371, |
| "learning_rate": 4.706563706563707e-05, |
| "loss": 0.858, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.29597220434950455, |
| "grad_norm": 1.0971994400024414, |
| "learning_rate": 4.703989703989704e-05, |
| "loss": 0.8932, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.2985458756916742, |
| "grad_norm": 1.346057415008545, |
| "learning_rate": 4.7014157014157014e-05, |
| "loss": 0.9042, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.3011195470338438, |
| "grad_norm": 1.1237205266952515, |
| "learning_rate": 4.6988416988416986e-05, |
| "loss": 0.7794, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.3036932183760134, |
| "grad_norm": 0.975764274597168, |
| "learning_rate": 4.6962676962676964e-05, |
| "loss": 0.8831, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.306266889718183, |
| "grad_norm": 1.2752641439437866, |
| "learning_rate": 4.6936936936936936e-05, |
| "loss": 0.8896, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.30884056106035257, |
| "grad_norm": 4.031650543212891, |
| "learning_rate": 4.6911196911196914e-05, |
| "loss": 0.9143, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.3114142324025222, |
| "grad_norm": 1.1821297407150269, |
| "learning_rate": 4.6885456885456886e-05, |
| "loss": 0.8443, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.3139879037446918, |
| "grad_norm": 1.1812553405761719, |
| "learning_rate": 4.6859716859716864e-05, |
| "loss": 0.9018, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.3165615750868614, |
| "grad_norm": 0.8742527365684509, |
| "learning_rate": 4.6833976833976836e-05, |
| "loss": 0.7639, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.319135246429031, |
| "grad_norm": 1.2843698263168335, |
| "learning_rate": 4.680823680823681e-05, |
| "loss": 0.8889, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.32170891777120064, |
| "grad_norm": 0.6974703073501587, |
| "learning_rate": 4.6782496782496786e-05, |
| "loss": 0.9007, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.3242825891133702, |
| "grad_norm": 3.426180839538574, |
| "learning_rate": 4.675675675675676e-05, |
| "loss": 0.8939, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.32685626045553984, |
| "grad_norm": 1.3513213396072388, |
| "learning_rate": 4.6731016731016736e-05, |
| "loss": 0.8406, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.3294299317977094, |
| "grad_norm": 1.0384771823883057, |
| "learning_rate": 4.670527670527671e-05, |
| "loss": 0.8806, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.33200360313987903, |
| "grad_norm": 1.1321779489517212, |
| "learning_rate": 4.6679536679536686e-05, |
| "loss": 0.7905, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.33457727448204866, |
| "grad_norm": 0.8733064532279968, |
| "learning_rate": 4.665379665379666e-05, |
| "loss": 0.8215, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.33715094582421823, |
| "grad_norm": 0.7575940489768982, |
| "learning_rate": 4.662805662805663e-05, |
| "loss": 0.8211, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.33972461716638785, |
| "grad_norm": 0.6808631420135498, |
| "learning_rate": 4.66023166023166e-05, |
| "loss": 0.8255, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.3422982885085575, |
| "grad_norm": 1.1921888589859009, |
| "learning_rate": 4.657657657657658e-05, |
| "loss": 0.8856, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.34487195985072705, |
| "grad_norm": 1.2014636993408203, |
| "learning_rate": 4.655083655083655e-05, |
| "loss": 0.8523, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.3474456311928967, |
| "grad_norm": 1.0800600051879883, |
| "learning_rate": 4.652509652509653e-05, |
| "loss": 0.8697, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.35001930253506625, |
| "grad_norm": 1.1528334617614746, |
| "learning_rate": 4.64993564993565e-05, |
| "loss": 0.8628, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.35259297387723587, |
| "grad_norm": 1.2347609996795654, |
| "learning_rate": 4.647361647361648e-05, |
| "loss": 0.8431, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.3551666452194055, |
| "grad_norm": 1.086850881576538, |
| "learning_rate": 4.644787644787645e-05, |
| "loss": 0.8429, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.35774031656157507, |
| "grad_norm": 1.0318641662597656, |
| "learning_rate": 4.642213642213642e-05, |
| "loss": 0.9399, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.3603139879037447, |
| "grad_norm": 0.9806049466133118, |
| "learning_rate": 4.6396396396396394e-05, |
| "loss": 0.8605, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.3628876592459143, |
| "grad_norm": 1.040371298789978, |
| "learning_rate": 4.637065637065637e-05, |
| "loss": 0.8756, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.3654613305880839, |
| "grad_norm": 3.93825626373291, |
| "learning_rate": 4.6344916344916344e-05, |
| "loss": 0.8493, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.3680350019302535, |
| "grad_norm": 1.530876874923706, |
| "learning_rate": 4.631917631917632e-05, |
| "loss": 0.8755, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.37060867327242314, |
| "grad_norm": 1.0033499002456665, |
| "learning_rate": 4.6293436293436294e-05, |
| "loss": 0.8141, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.3731823446145927, |
| "grad_norm": 1.1045700311660767, |
| "learning_rate": 4.626769626769627e-05, |
| "loss": 0.8565, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.37575601595676233, |
| "grad_norm": 0.9272418022155762, |
| "learning_rate": 4.6241956241956244e-05, |
| "loss": 0.8335, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.3783296872989319, |
| "grad_norm": 2.702624559402466, |
| "learning_rate": 4.6216216216216215e-05, |
| "loss": 0.8326, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.38090335864110153, |
| "grad_norm": 1.0378378629684448, |
| "learning_rate": 4.6190476190476194e-05, |
| "loss": 0.8696, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.38347702998327116, |
| "grad_norm": 0.9696075320243835, |
| "learning_rate": 4.6164736164736165e-05, |
| "loss": 0.8559, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.3860507013254407, |
| "grad_norm": 0.9262058734893799, |
| "learning_rate": 4.6138996138996144e-05, |
| "loss": 0.9104, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.38862437266761035, |
| "grad_norm": 3.3203704357147217, |
| "learning_rate": 4.6113256113256115e-05, |
| "loss": 0.803, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.39119804400978, |
| "grad_norm": 1.3920564651489258, |
| "learning_rate": 4.6087516087516094e-05, |
| "loss": 0.9708, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.39377171535194955, |
| "grad_norm": 0.9011679291725159, |
| "learning_rate": 4.6061776061776065e-05, |
| "loss": 0.872, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.3963453866941192, |
| "grad_norm": 1.1053688526153564, |
| "learning_rate": 4.603603603603604e-05, |
| "loss": 0.882, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.39891905803628874, |
| "grad_norm": 0.8040546774864197, |
| "learning_rate": 4.601029601029601e-05, |
| "loss": 0.8539, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.40149272937845837, |
| "grad_norm": 3.191498041152954, |
| "learning_rate": 4.598455598455599e-05, |
| "loss": 0.9703, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.404066400720628, |
| "grad_norm": 1.2510138750076294, |
| "learning_rate": 4.595881595881596e-05, |
| "loss": 0.8436, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.40664007206279756, |
| "grad_norm": 0.8919886350631714, |
| "learning_rate": 4.593307593307594e-05, |
| "loss": 0.8437, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.4092137434049672, |
| "grad_norm": 0.8489089608192444, |
| "learning_rate": 4.590733590733591e-05, |
| "loss": 0.9111, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.4117874147471368, |
| "grad_norm": 1.2358871698379517, |
| "learning_rate": 4.588159588159589e-05, |
| "loss": 0.8806, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.4143610860893064, |
| "grad_norm": 1.3173274993896484, |
| "learning_rate": 4.585585585585586e-05, |
| "loss": 0.833, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.416934757431476, |
| "grad_norm": 1.3292251825332642, |
| "learning_rate": 4.583011583011583e-05, |
| "loss": 0.8926, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.4195084287736456, |
| "grad_norm": 1.155118465423584, |
| "learning_rate": 4.58043758043758e-05, |
| "loss": 0.8289, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.4220821001158152, |
| "grad_norm": 0.9190046787261963, |
| "learning_rate": 4.577863577863578e-05, |
| "loss": 0.8332, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.42465577145798483, |
| "grad_norm": 0.9945672750473022, |
| "learning_rate": 4.575289575289575e-05, |
| "loss": 0.8378, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.4272294428001544, |
| "grad_norm": 1.0910547971725464, |
| "learning_rate": 4.572715572715573e-05, |
| "loss": 0.8843, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.429803114142324, |
| "grad_norm": 1.24397611618042, |
| "learning_rate": 4.570141570141571e-05, |
| "loss": 0.8926, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.43237678548449365, |
| "grad_norm": 1.0651038885116577, |
| "learning_rate": 4.567567567567568e-05, |
| "loss": 0.8542, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.4349504568266632, |
| "grad_norm": 1.0350909233093262, |
| "learning_rate": 4.564993564993565e-05, |
| "loss": 0.8258, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.43752412816883285, |
| "grad_norm": 1.0753473043441772, |
| "learning_rate": 4.5624195624195624e-05, |
| "loss": 0.8557, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.4400977995110024, |
| "grad_norm": 1.2361218929290771, |
| "learning_rate": 4.55984555984556e-05, |
| "loss": 0.9459, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.44267147085317204, |
| "grad_norm": 1.2962043285369873, |
| "learning_rate": 4.5572715572715574e-05, |
| "loss": 0.8822, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.44524514219534167, |
| "grad_norm": 2.557123899459839, |
| "learning_rate": 4.554697554697555e-05, |
| "loss": 0.8018, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.44781881353751124, |
| "grad_norm": 0.8217394351959229, |
| "learning_rate": 4.5521235521235524e-05, |
| "loss": 0.8012, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.45039248487968087, |
| "grad_norm": 0.9458732008934021, |
| "learning_rate": 4.54954954954955e-05, |
| "loss": 0.835, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.4529661562218505, |
| "grad_norm": 0.9185703992843628, |
| "learning_rate": 4.5469755469755473e-05, |
| "loss": 0.8571, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.45553982756402006, |
| "grad_norm": 1.190473198890686, |
| "learning_rate": 4.5444015444015445e-05, |
| "loss": 0.81, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.4581134989061897, |
| "grad_norm": 1.1734472513198853, |
| "learning_rate": 4.541827541827542e-05, |
| "loss": 0.8926, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.46068717024835926, |
| "grad_norm": 1.0630743503570557, |
| "learning_rate": 4.5392535392535395e-05, |
| "loss": 0.8848, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.4632608415905289, |
| "grad_norm": 3.2704575061798096, |
| "learning_rate": 4.536679536679537e-05, |
| "loss": 0.923, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.4658345129326985, |
| "grad_norm": 0.8586927056312561, |
| "learning_rate": 4.5341055341055345e-05, |
| "loss": 0.8056, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.4684081842748681, |
| "grad_norm": 1.2519656419754028, |
| "learning_rate": 4.531531531531532e-05, |
| "loss": 0.8505, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.4709818556170377, |
| "grad_norm": 0.8403356075286865, |
| "learning_rate": 4.5289575289575295e-05, |
| "loss": 0.8093, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.47355552695920733, |
| "grad_norm": 1.1714375019073486, |
| "learning_rate": 4.526383526383527e-05, |
| "loss": 0.9739, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.4761291983013769, |
| "grad_norm": 0.9783056378364563, |
| "learning_rate": 4.523809523809524e-05, |
| "loss": 0.8602, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.4787028696435465, |
| "grad_norm": 2.869663715362549, |
| "learning_rate": 4.521235521235521e-05, |
| "loss": 0.8444, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.48127654098571615, |
| "grad_norm": 0.8355293273925781, |
| "learning_rate": 4.518661518661519e-05, |
| "loss": 0.9036, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.4838502123278857, |
| "grad_norm": 1.140135407447815, |
| "learning_rate": 4.516087516087517e-05, |
| "loss": 0.8022, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.48642388367005535, |
| "grad_norm": 0.9020267724990845, |
| "learning_rate": 4.513513513513514e-05, |
| "loss": 0.8493, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.4889975550122249, |
| "grad_norm": 3.613842248916626, |
| "learning_rate": 4.510939510939512e-05, |
| "loss": 0.8602, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.49157122635439454, |
| "grad_norm": 3.0518836975097656, |
| "learning_rate": 4.508365508365509e-05, |
| "loss": 0.8379, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.49414489769656417, |
| "grad_norm": 0.9278186559677124, |
| "learning_rate": 4.505791505791506e-05, |
| "loss": 0.9156, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.49671856903873374, |
| "grad_norm": 4.282588481903076, |
| "learning_rate": 4.503217503217503e-05, |
| "loss": 0.9301, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.49929224038090336, |
| "grad_norm": 3.971865653991699, |
| "learning_rate": 4.500643500643501e-05, |
| "loss": 0.8885, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.5018659117230729, |
| "grad_norm": 1.102675199508667, |
| "learning_rate": 4.498069498069498e-05, |
| "loss": 0.8936, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.5044395830652426, |
| "grad_norm": 0.9979203939437866, |
| "learning_rate": 4.495495495495496e-05, |
| "loss": 0.8014, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.5070132544074122, |
| "grad_norm": 4.093165397644043, |
| "learning_rate": 4.492921492921493e-05, |
| "loss": 0.8699, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.5095869257495818, |
| "grad_norm": 1.107493281364441, |
| "learning_rate": 4.490347490347491e-05, |
| "loss": 0.8319, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.5121605970917514, |
| "grad_norm": 0.8577767014503479, |
| "learning_rate": 4.487773487773488e-05, |
| "loss": 0.8874, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.514734268433921, |
| "grad_norm": 1.2690497636795044, |
| "learning_rate": 4.485199485199485e-05, |
| "loss": 0.8586, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.514734268433921, |
| "eval_loss": 0.8306575417518616, |
| "eval_runtime": 395.638, |
| "eval_samples_per_second": 49.1, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.00039326141255247916, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.5173079397760906, |
| "grad_norm": 1.1434204578399658, |
| "learning_rate": 4.4826254826254825e-05, |
| "loss": 0.9541, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.5198816111182601, |
| "grad_norm": 3.3350868225097656, |
| "learning_rate": 4.48005148005148e-05, |
| "loss": 0.8295, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.5224552824604298, |
| "grad_norm": 1.2531917095184326, |
| "learning_rate": 4.4774774774774775e-05, |
| "loss": 0.8997, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.5250289538025994, |
| "grad_norm": 0.9255910515785217, |
| "learning_rate": 4.474903474903475e-05, |
| "loss": 0.8375, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.527602625144769, |
| "grad_norm": 1.0736703872680664, |
| "learning_rate": 4.4723294723294725e-05, |
| "loss": 0.8293, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.5301762964869386, |
| "grad_norm": 1.5314749479293823, |
| "learning_rate": 4.46975546975547e-05, |
| "loss": 0.8036, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.5327499678291082, |
| "grad_norm": 1.0924575328826904, |
| "learning_rate": 4.467181467181467e-05, |
| "loss": 0.7548, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.5353236391712778, |
| "grad_norm": 0.9415216445922852, |
| "learning_rate": 4.4646074646074646e-05, |
| "loss": 0.8181, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.5378973105134475, |
| "grad_norm": 1.0533065795898438, |
| "learning_rate": 4.4620334620334625e-05, |
| "loss": 0.8045, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.540470981855617, |
| "grad_norm": 1.1450499296188354, |
| "learning_rate": 4.4594594594594596e-05, |
| "loss": 0.8571, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.5430446531977866, |
| "grad_norm": 3.1443707942962646, |
| "learning_rate": 4.4568854568854575e-05, |
| "loss": 0.8687, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.5456183245399563, |
| "grad_norm": 0.8034738302230835, |
| "learning_rate": 4.4543114543114546e-05, |
| "loss": 0.8746, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.5481919958821259, |
| "grad_norm": 0.8792498707771301, |
| "learning_rate": 4.4517374517374525e-05, |
| "loss": 0.8886, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.5507656672242954, |
| "grad_norm": 3.1286137104034424, |
| "learning_rate": 4.449163449163449e-05, |
| "loss": 0.8574, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.5533393385664651, |
| "grad_norm": 0.9840642213821411, |
| "learning_rate": 4.446589446589447e-05, |
| "loss": 0.815, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.5559130099086347, |
| "grad_norm": 1.1499487161636353, |
| "learning_rate": 4.444015444015444e-05, |
| "loss": 0.8691, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.5584866812508043, |
| "grad_norm": 0.8926144242286682, |
| "learning_rate": 4.441441441441442e-05, |
| "loss": 0.8082, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.5610603525929739, |
| "grad_norm": 0.7535387873649597, |
| "learning_rate": 4.438867438867439e-05, |
| "loss": 0.9033, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.5636340239351435, |
| "grad_norm": 1.196241021156311, |
| "learning_rate": 4.436293436293437e-05, |
| "loss": 0.9039, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.5662076952773131, |
| "grad_norm": 2.5870370864868164, |
| "learning_rate": 4.433719433719434e-05, |
| "loss": 0.8692, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.5687813666194826, |
| "grad_norm": 1.209121823310852, |
| "learning_rate": 4.431145431145432e-05, |
| "loss": 0.8467, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.5713550379616523, |
| "grad_norm": 0.988710343837738, |
| "learning_rate": 4.428571428571428e-05, |
| "loss": 0.7887, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.5739287093038219, |
| "grad_norm": 0.9768078327178955, |
| "learning_rate": 4.425997425997426e-05, |
| "loss": 0.8597, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.5765023806459915, |
| "grad_norm": 1.3332276344299316, |
| "learning_rate": 4.423423423423423e-05, |
| "loss": 0.9351, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.5790760519881611, |
| "grad_norm": 2.8126509189605713, |
| "learning_rate": 4.420849420849421e-05, |
| "loss": 0.8706, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.5816497233303307, |
| "grad_norm": 1.1589819192886353, |
| "learning_rate": 4.418275418275418e-05, |
| "loss": 0.874, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.5842233946725003, |
| "grad_norm": 1.085998296737671, |
| "learning_rate": 4.415701415701416e-05, |
| "loss": 0.8321, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.58679706601467, |
| "grad_norm": 0.8814729452133179, |
| "learning_rate": 4.413127413127413e-05, |
| "loss": 0.8407, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.5893707373568395, |
| "grad_norm": 3.6216423511505127, |
| "learning_rate": 4.4105534105534104e-05, |
| "loss": 0.7986, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.5919444086990091, |
| "grad_norm": 3.6593005657196045, |
| "learning_rate": 4.407979407979408e-05, |
| "loss": 0.8252, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.5945180800411788, |
| "grad_norm": 0.9884054660797119, |
| "learning_rate": 4.4054054054054054e-05, |
| "loss": 0.8606, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.5970917513833484, |
| "grad_norm": 1.0792969465255737, |
| "learning_rate": 4.402831402831403e-05, |
| "loss": 0.7866, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.5996654227255179, |
| "grad_norm": 1.0255166292190552, |
| "learning_rate": 4.4002574002574004e-05, |
| "loss": 0.7659, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.6022390940676876, |
| "grad_norm": 0.7235389947891235, |
| "learning_rate": 4.397683397683398e-05, |
| "loss": 0.8475, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.6048127654098572, |
| "grad_norm": 0.8555417060852051, |
| "learning_rate": 4.3951093951093954e-05, |
| "loss": 0.8234, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.6073864367520267, |
| "grad_norm": 1.1346522569656372, |
| "learning_rate": 4.3925353925353926e-05, |
| "loss": 0.8572, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.6099601080941963, |
| "grad_norm": 3.7354023456573486, |
| "learning_rate": 4.38996138996139e-05, |
| "loss": 0.842, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.612533779436366, |
| "grad_norm": 1.0269885063171387, |
| "learning_rate": 4.3873873873873876e-05, |
| "loss": 0.8483, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.6151074507785356, |
| "grad_norm": 4.06434965133667, |
| "learning_rate": 4.384813384813385e-05, |
| "loss": 0.9082, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.6176811221207051, |
| "grad_norm": 1.0140035152435303, |
| "learning_rate": 4.3822393822393826e-05, |
| "loss": 0.8993, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.6202547934628748, |
| "grad_norm": 0.9860438108444214, |
| "learning_rate": 4.37966537966538e-05, |
| "loss": 0.8143, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.6228284648050444, |
| "grad_norm": 0.9462414979934692, |
| "learning_rate": 4.3770913770913776e-05, |
| "loss": 0.8752, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.625402136147214, |
| "grad_norm": 0.889310896396637, |
| "learning_rate": 4.374517374517375e-05, |
| "loss": 0.8576, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.6279758074893836, |
| "grad_norm": 1.0256364345550537, |
| "learning_rate": 4.371943371943372e-05, |
| "loss": 0.8331, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.6305494788315532, |
| "grad_norm": 1.242521047592163, |
| "learning_rate": 4.369369369369369e-05, |
| "loss": 0.917, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.6331231501737228, |
| "grad_norm": 0.8450289368629456, |
| "learning_rate": 4.366795366795367e-05, |
| "loss": 0.8463, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.6356968215158925, |
| "grad_norm": 2.8590588569641113, |
| "learning_rate": 4.364221364221364e-05, |
| "loss": 0.8509, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.638270492858062, |
| "grad_norm": 0.9202858805656433, |
| "learning_rate": 4.361647361647362e-05, |
| "loss": 0.8259, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.6408441642002316, |
| "grad_norm": 1.0481261014938354, |
| "learning_rate": 4.359073359073359e-05, |
| "loss": 0.8764, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.6434178355424013, |
| "grad_norm": 1.0081082582473755, |
| "learning_rate": 4.356499356499357e-05, |
| "loss": 0.8097, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.6459915068845709, |
| "grad_norm": 0.9049886465072632, |
| "learning_rate": 4.353925353925354e-05, |
| "loss": 0.8617, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.6485651782267404, |
| "grad_norm": 1.0273019075393677, |
| "learning_rate": 4.351351351351351e-05, |
| "loss": 0.8457, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.6511388495689101, |
| "grad_norm": 3.1183676719665527, |
| "learning_rate": 4.348777348777349e-05, |
| "loss": 0.8202, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.6537125209110797, |
| "grad_norm": 0.9991865158081055, |
| "learning_rate": 4.346203346203346e-05, |
| "loss": 0.8147, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.6562861922532492, |
| "grad_norm": 1.0085402727127075, |
| "learning_rate": 4.343629343629344e-05, |
| "loss": 0.8073, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.6588598635954188, |
| "grad_norm": 0.8995606899261475, |
| "learning_rate": 4.341055341055341e-05, |
| "loss": 0.8394, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.6614335349375885, |
| "grad_norm": 0.9390326142311096, |
| "learning_rate": 4.338481338481339e-05, |
| "loss": 0.8784, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.6640072062797581, |
| "grad_norm": 0.7828226089477539, |
| "learning_rate": 4.335907335907336e-05, |
| "loss": 0.861, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.6665808776219276, |
| "grad_norm": 0.752167284488678, |
| "learning_rate": 4.3333333333333334e-05, |
| "loss": 0.8099, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.6691545489640973, |
| "grad_norm": 0.9056263566017151, |
| "learning_rate": 4.3307593307593306e-05, |
| "loss": 0.9575, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.6717282203062669, |
| "grad_norm": 1.1399377584457397, |
| "learning_rate": 4.3281853281853284e-05, |
| "loss": 0.8491, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.6743018916484365, |
| "grad_norm": 1.1799904108047485, |
| "learning_rate": 4.3256113256113256e-05, |
| "loss": 0.8375, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.6768755629906061, |
| "grad_norm": 3.550177812576294, |
| "learning_rate": 4.3230373230373234e-05, |
| "loss": 0.8797, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.6794492343327757, |
| "grad_norm": 0.8482165336608887, |
| "learning_rate": 4.3204633204633206e-05, |
| "loss": 0.8424, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.6820229056749453, |
| "grad_norm": 0.8729701042175293, |
| "learning_rate": 4.3178893178893184e-05, |
| "loss": 0.7896, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.684596577017115, |
| "grad_norm": 1.0203555822372437, |
| "learning_rate": 4.3153153153153156e-05, |
| "loss": 0.8836, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.6871702483592845, |
| "grad_norm": 1.1605058908462524, |
| "learning_rate": 4.312741312741313e-05, |
| "loss": 0.7561, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.6897439197014541, |
| "grad_norm": 1.3112170696258545, |
| "learning_rate": 4.31016731016731e-05, |
| "loss": 0.8504, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.6923175910436238, |
| "grad_norm": 2.7593069076538086, |
| "learning_rate": 4.307593307593308e-05, |
| "loss": 0.9337, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.6948912623857934, |
| "grad_norm": 0.7685155272483826, |
| "learning_rate": 4.305019305019305e-05, |
| "loss": 0.9312, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.6974649337279629, |
| "grad_norm": 0.9489961862564087, |
| "learning_rate": 4.302445302445303e-05, |
| "loss": 0.839, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.7000386050701325, |
| "grad_norm": 2.002380132675171, |
| "learning_rate": 4.2998712998713006e-05, |
| "loss": 0.7451, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.7026122764123022, |
| "grad_norm": 1.3577697277069092, |
| "learning_rate": 4.297297297297298e-05, |
| "loss": 0.823, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.7051859477544717, |
| "grad_norm": 0.991814911365509, |
| "learning_rate": 4.294723294723295e-05, |
| "loss": 0.7832, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.7077596190966413, |
| "grad_norm": 1.302040696144104, |
| "learning_rate": 4.292149292149292e-05, |
| "loss": 0.8151, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.710333290438811, |
| "grad_norm": 1.1003100872039795, |
| "learning_rate": 4.28957528957529e-05, |
| "loss": 0.8792, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.7129069617809806, |
| "grad_norm": 1.0496855974197388, |
| "learning_rate": 4.287001287001287e-05, |
| "loss": 0.8354, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.7154806331231501, |
| "grad_norm": 1.375705599784851, |
| "learning_rate": 4.284427284427285e-05, |
| "loss": 0.7915, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.7180543044653198, |
| "grad_norm": 1.030092716217041, |
| "learning_rate": 4.281853281853282e-05, |
| "loss": 0.828, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.7206279758074894, |
| "grad_norm": 2.520185708999634, |
| "learning_rate": 4.27927927927928e-05, |
| "loss": 0.8099, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.723201647149659, |
| "grad_norm": 0.7944628596305847, |
| "learning_rate": 4.276705276705277e-05, |
| "loss": 0.8974, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.7257753184918286, |
| "grad_norm": 1.2783660888671875, |
| "learning_rate": 4.274131274131274e-05, |
| "loss": 0.7787, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.7283489898339982, |
| "grad_norm": 1.3568511009216309, |
| "learning_rate": 4.2715572715572714e-05, |
| "loss": 0.8208, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.7309226611761678, |
| "grad_norm": 2.9468472003936768, |
| "learning_rate": 4.268983268983269e-05, |
| "loss": 0.8063, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.7334963325183375, |
| "grad_norm": 0.871306300163269, |
| "learning_rate": 4.2664092664092664e-05, |
| "loss": 0.7988, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.736070003860507, |
| "grad_norm": 1.0941967964172363, |
| "learning_rate": 4.263835263835264e-05, |
| "loss": 0.8288, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.7386436752026766, |
| "grad_norm": 0.992978036403656, |
| "learning_rate": 4.2612612612612614e-05, |
| "loss": 0.8742, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.7412173465448463, |
| "grad_norm": 0.8720837831497192, |
| "learning_rate": 4.258687258687259e-05, |
| "loss": 0.8032, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.7437910178870158, |
| "grad_norm": 3.578805446624756, |
| "learning_rate": 4.2561132561132564e-05, |
| "loss": 0.8462, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.7463646892291854, |
| "grad_norm": 1.0471611022949219, |
| "learning_rate": 4.2535392535392535e-05, |
| "loss": 0.8461, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.748938360571355, |
| "grad_norm": 0.9668796062469482, |
| "learning_rate": 4.2509652509652514e-05, |
| "loss": 0.8373, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.7515120319135247, |
| "grad_norm": 0.8879560232162476, |
| "learning_rate": 4.2483912483912485e-05, |
| "loss": 0.8707, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.7540857032556942, |
| "grad_norm": 0.8577817678451538, |
| "learning_rate": 4.2458172458172464e-05, |
| "loss": 0.8934, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.7566593745978638, |
| "grad_norm": 2.37251877784729, |
| "learning_rate": 4.2432432432432435e-05, |
| "loss": 0.8708, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.7592330459400335, |
| "grad_norm": 0.7796174883842468, |
| "learning_rate": 4.2406692406692414e-05, |
| "loss": 0.8832, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.7618067172822031, |
| "grad_norm": 1.032285213470459, |
| "learning_rate": 4.2380952380952385e-05, |
| "loss": 0.8615, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.7643803886243726, |
| "grad_norm": 4.347442150115967, |
| "learning_rate": 4.235521235521236e-05, |
| "loss": 0.8413, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.7669540599665423, |
| "grad_norm": 1.1776065826416016, |
| "learning_rate": 4.232947232947233e-05, |
| "loss": 0.8533, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.7695277313087119, |
| "grad_norm": 0.956089973449707, |
| "learning_rate": 4.230373230373231e-05, |
| "loss": 0.8051, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.7721014026508815, |
| "grad_norm": 2.812290906906128, |
| "learning_rate": 4.227799227799228e-05, |
| "loss": 0.816, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.7721014026508815, |
| "eval_loss": 0.8144821524620056, |
| "eval_runtime": 395.614, |
| "eval_samples_per_second": 49.103, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.00037428161464936723, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.7746750739930511, |
| "grad_norm": 0.971863865852356, |
| "learning_rate": 4.225225225225226e-05, |
| "loss": 0.8316, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.7772487453352207, |
| "grad_norm": 0.9860455393791199, |
| "learning_rate": 4.222651222651223e-05, |
| "loss": 0.8641, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.7798224166773903, |
| "grad_norm": 1.0967594385147095, |
| "learning_rate": 4.220077220077221e-05, |
| "loss": 0.8423, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.78239608801956, |
| "grad_norm": 0.7346376776695251, |
| "learning_rate": 4.217503217503218e-05, |
| "loss": 0.8773, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.7849697593617295, |
| "grad_norm": 0.7431369423866272, |
| "learning_rate": 4.214929214929215e-05, |
| "loss": 0.8448, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.7875434307038991, |
| "grad_norm": 1.0294193029403687, |
| "learning_rate": 4.212355212355212e-05, |
| "loss": 0.7999, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.7901171020460687, |
| "grad_norm": 3.059105157852173, |
| "learning_rate": 4.20978120978121e-05, |
| "loss": 0.8155, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.7926907733882383, |
| "grad_norm": 1.1290513277053833, |
| "learning_rate": 4.207207207207207e-05, |
| "loss": 0.9172, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.7952644447304079, |
| "grad_norm": 0.7203499674797058, |
| "learning_rate": 4.204633204633205e-05, |
| "loss": 0.8059, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.7978381160725775, |
| "grad_norm": 3.0688018798828125, |
| "learning_rate": 4.202059202059202e-05, |
| "loss": 0.9509, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.8004117874147472, |
| "grad_norm": 4.001659393310547, |
| "learning_rate": 4.1994851994852e-05, |
| "loss": 0.7508, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.8029854587569167, |
| "grad_norm": 1.1222361326217651, |
| "learning_rate": 4.196911196911197e-05, |
| "loss": 0.9271, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.8055591300990863, |
| "grad_norm": 0.8683832287788391, |
| "learning_rate": 4.194337194337194e-05, |
| "loss": 0.8042, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.808132801441256, |
| "grad_norm": 0.89637291431427, |
| "learning_rate": 4.191763191763192e-05, |
| "loss": 0.7162, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.8107064727834256, |
| "grad_norm": 0.8857401609420776, |
| "learning_rate": 4.189189189189189e-05, |
| "loss": 0.7789, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.8132801441255951, |
| "grad_norm": 1.2026958465576172, |
| "learning_rate": 4.186615186615187e-05, |
| "loss": 0.8782, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.8158538154677648, |
| "grad_norm": 0.8439406156539917, |
| "learning_rate": 4.184041184041184e-05, |
| "loss": 0.8504, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.8184274868099344, |
| "grad_norm": 0.9551829099655151, |
| "learning_rate": 4.181467181467182e-05, |
| "loss": 0.8152, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.821001158152104, |
| "grad_norm": 1.0437098741531372, |
| "learning_rate": 4.1788931788931787e-05, |
| "loss": 0.8451, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.8235748294942736, |
| "grad_norm": 0.9579282402992249, |
| "learning_rate": 4.1763191763191765e-05, |
| "loss": 0.8714, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.8261485008364432, |
| "grad_norm": 3.2837677001953125, |
| "learning_rate": 4.1737451737451737e-05, |
| "loss": 0.9142, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.8287221721786128, |
| "grad_norm": 0.8237940073013306, |
| "learning_rate": 4.1711711711711715e-05, |
| "loss": 0.8592, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.8312958435207825, |
| "grad_norm": 0.9125580787658691, |
| "learning_rate": 4.1685971685971687e-05, |
| "loss": 0.8598, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.833869514862952, |
| "grad_norm": 3.1465914249420166, |
| "learning_rate": 4.1660231660231665e-05, |
| "loss": 0.756, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.8364431862051216, |
| "grad_norm": 0.9401727318763733, |
| "learning_rate": 4.1634491634491637e-05, |
| "loss": 0.765, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.8390168575472912, |
| "grad_norm": 1.0945508480072021, |
| "learning_rate": 4.1608751608751615e-05, |
| "loss": 0.8335, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.8415905288894608, |
| "grad_norm": 1.5504162311553955, |
| "learning_rate": 4.158301158301158e-05, |
| "loss": 0.8269, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.8441642002316304, |
| "grad_norm": 0.7810274362564087, |
| "learning_rate": 4.155727155727156e-05, |
| "loss": 0.8326, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.8467378715738, |
| "grad_norm": 1.2022150754928589, |
| "learning_rate": 4.153153153153153e-05, |
| "loss": 0.7965, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.8493115429159697, |
| "grad_norm": 2.462810754776001, |
| "learning_rate": 4.150579150579151e-05, |
| "loss": 0.8781, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.8518852142581392, |
| "grad_norm": 1.1842944622039795, |
| "learning_rate": 4.148005148005148e-05, |
| "loss": 0.7906, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.8544588856003088, |
| "grad_norm": 0.8692466020584106, |
| "learning_rate": 4.145431145431146e-05, |
| "loss": 0.8167, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.8570325569424785, |
| "grad_norm": 1.1452484130859375, |
| "learning_rate": 4.1428571428571437e-05, |
| "loss": 0.795, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.859606228284648, |
| "grad_norm": 1.0664327144622803, |
| "learning_rate": 4.14028314028314e-05, |
| "loss": 0.8176, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.8621798996268176, |
| "grad_norm": 1.0237557888031006, |
| "learning_rate": 4.137709137709138e-05, |
| "loss": 0.8243, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.8647535709689873, |
| "grad_norm": 1.301072597503662, |
| "learning_rate": 4.135135135135135e-05, |
| "loss": 0.7853, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.8673272423111569, |
| "grad_norm": 1.061145544052124, |
| "learning_rate": 4.132561132561133e-05, |
| "loss": 0.8669, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.8699009136533264, |
| "grad_norm": 4.076564788818359, |
| "learning_rate": 4.12998712998713e-05, |
| "loss": 0.8657, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.8724745849954961, |
| "grad_norm": 0.8901606798171997, |
| "learning_rate": 4.127413127413128e-05, |
| "loss": 0.8867, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.8750482563376657, |
| "grad_norm": 1.0653889179229736, |
| "learning_rate": 4.124839124839125e-05, |
| "loss": 0.8954, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.8776219276798353, |
| "grad_norm": 0.9028094410896301, |
| "learning_rate": 4.122265122265122e-05, |
| "loss": 0.8455, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.8801955990220048, |
| "grad_norm": 0.7900987267494202, |
| "learning_rate": 4.1196911196911195e-05, |
| "loss": 0.8002, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.8827692703641745, |
| "grad_norm": 3.244790554046631, |
| "learning_rate": 4.117117117117117e-05, |
| "loss": 0.7287, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.8853429417063441, |
| "grad_norm": 1.0797080993652344, |
| "learning_rate": 4.1145431145431145e-05, |
| "loss": 0.8955, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.8879166130485137, |
| "grad_norm": 1.0693559646606445, |
| "learning_rate": 4.111969111969112e-05, |
| "loss": 0.8617, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.8904902843906833, |
| "grad_norm": 0.8851631283760071, |
| "learning_rate": 4.1093951093951095e-05, |
| "loss": 0.8405, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.8930639557328529, |
| "grad_norm": 2.6843698024749756, |
| "learning_rate": 4.106821106821107e-05, |
| "loss": 0.8352, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.8956376270750225, |
| "grad_norm": 3.100241184234619, |
| "learning_rate": 4.1042471042471045e-05, |
| "loss": 0.7919, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.8982112984171922, |
| "grad_norm": 0.8943722248077393, |
| "learning_rate": 4.1016731016731016e-05, |
| "loss": 0.8122, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.9007849697593617, |
| "grad_norm": 0.827441394329071, |
| "learning_rate": 4.099099099099099e-05, |
| "loss": 0.8883, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.9033586411015313, |
| "grad_norm": 1.1157348155975342, |
| "learning_rate": 4.0965250965250966e-05, |
| "loss": 0.8779, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.905932312443701, |
| "grad_norm": 1.1023356914520264, |
| "learning_rate": 4.093951093951094e-05, |
| "loss": 0.8149, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.9085059837858706, |
| "grad_norm": 0.661834180355072, |
| "learning_rate": 4.0913770913770916e-05, |
| "loss": 0.7732, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.9110796551280401, |
| "grad_norm": 1.1258097887039185, |
| "learning_rate": 4.0888030888030895e-05, |
| "loss": 0.7748, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.9136533264702098, |
| "grad_norm": 1.0704106092453003, |
| "learning_rate": 4.0862290862290866e-05, |
| "loss": 0.8873, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.9162269978123794, |
| "grad_norm": 0.9617460370063782, |
| "learning_rate": 4.083655083655084e-05, |
| "loss": 0.7, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.9188006691545489, |
| "grad_norm": 0.8261712789535522, |
| "learning_rate": 4.081081081081081e-05, |
| "loss": 0.7724, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.9213743404967185, |
| "grad_norm": 0.8264958262443542, |
| "learning_rate": 4.078507078507079e-05, |
| "loss": 0.8189, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.9239480118388882, |
| "grad_norm": 0.9358187913894653, |
| "learning_rate": 4.075933075933076e-05, |
| "loss": 0.823, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.9265216831810578, |
| "grad_norm": 0.9269980788230896, |
| "learning_rate": 4.073359073359074e-05, |
| "loss": 0.7901, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.9290953545232273, |
| "grad_norm": 1.0300779342651367, |
| "learning_rate": 4.070785070785071e-05, |
| "loss": 0.8129, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.931669025865397, |
| "grad_norm": 2.253840446472168, |
| "learning_rate": 4.068211068211069e-05, |
| "loss": 0.785, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.9342426972075666, |
| "grad_norm": 0.7381809949874878, |
| "learning_rate": 4.065637065637066e-05, |
| "loss": 0.8316, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.9368163685497362, |
| "grad_norm": 2.6727185249328613, |
| "learning_rate": 4.063063063063063e-05, |
| "loss": 0.88, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.9393900398919058, |
| "grad_norm": 0.7248988747596741, |
| "learning_rate": 4.06048906048906e-05, |
| "loss": 0.8239, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.9419637112340754, |
| "grad_norm": 0.8684216141700745, |
| "learning_rate": 4.057915057915058e-05, |
| "loss": 0.8899, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.944537382576245, |
| "grad_norm": 2.2277228832244873, |
| "learning_rate": 4.055341055341055e-05, |
| "loss": 0.8437, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.9471110539184147, |
| "grad_norm": 0.946140706539154, |
| "learning_rate": 4.052767052767053e-05, |
| "loss": 0.7992, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.9496847252605842, |
| "grad_norm": 0.7557298541069031, |
| "learning_rate": 4.05019305019305e-05, |
| "loss": 0.849, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.9522583966027538, |
| "grad_norm": 1.0458942651748657, |
| "learning_rate": 4.047619047619048e-05, |
| "loss": 0.8492, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.9548320679449235, |
| "grad_norm": 1.1995244026184082, |
| "learning_rate": 4.045045045045045e-05, |
| "loss": 0.7823, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.957405739287093, |
| "grad_norm": 3.7634544372558594, |
| "learning_rate": 4.0424710424710424e-05, |
| "loss": 0.7751, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.9599794106292626, |
| "grad_norm": 1.2692762613296509, |
| "learning_rate": 4.0398970398970396e-05, |
| "loss": 0.8069, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.9625530819714323, |
| "grad_norm": 1.2810945510864258, |
| "learning_rate": 4.0373230373230374e-05, |
| "loss": 0.9275, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.9651267533136019, |
| "grad_norm": 1.7654497623443604, |
| "learning_rate": 4.034749034749035e-05, |
| "loss": 0.8789, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.9677004246557714, |
| "grad_norm": 3.197882652282715, |
| "learning_rate": 4.0321750321750324e-05, |
| "loss": 0.8377, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.970274095997941, |
| "grad_norm": 0.8221775889396667, |
| "learning_rate": 4.02960102960103e-05, |
| "loss": 0.7834, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.9728477673401107, |
| "grad_norm": 0.9172367453575134, |
| "learning_rate": 4.0270270270270274e-05, |
| "loss": 0.8551, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.9754214386822803, |
| "grad_norm": 0.8287565112113953, |
| "learning_rate": 4.0244530244530246e-05, |
| "loss": 0.8427, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.9779951100244498, |
| "grad_norm": 0.9844699501991272, |
| "learning_rate": 4.021879021879022e-05, |
| "loss": 0.8414, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.9805687813666195, |
| "grad_norm": 0.8020986318588257, |
| "learning_rate": 4.0193050193050196e-05, |
| "loss": 0.9036, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.9831424527087891, |
| "grad_norm": 0.9322710633277893, |
| "learning_rate": 4.016731016731017e-05, |
| "loss": 0.773, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.9857161240509587, |
| "grad_norm": 1.5379607677459717, |
| "learning_rate": 4.0141570141570146e-05, |
| "loss": 0.8547, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.9882897953931283, |
| "grad_norm": 3.0354602336883545, |
| "learning_rate": 4.011583011583012e-05, |
| "loss": 0.8487, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.9908634667352979, |
| "grad_norm": 0.9367321133613586, |
| "learning_rate": 4.0090090090090096e-05, |
| "loss": 0.8564, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.9934371380774675, |
| "grad_norm": 0.9189753532409668, |
| "learning_rate": 4.006435006435007e-05, |
| "loss": 0.8674, |
| "step": 3860 |
| }, |
| { |
| "epoch": 0.9960108094196372, |
| "grad_norm": 1.0535331964492798, |
| "learning_rate": 4.003861003861004e-05, |
| "loss": 0.8038, |
| "step": 3870 |
| }, |
| { |
| "epoch": 0.9985844807618067, |
| "grad_norm": 0.8121357560157776, |
| "learning_rate": 4.001287001287001e-05, |
| "loss": 0.8706, |
| "step": 3880 |
| }, |
| { |
| "epoch": 1.0011581521039763, |
| "grad_norm": 1.0434505939483643, |
| "learning_rate": 3.998712998712999e-05, |
| "loss": 0.7571, |
| "step": 3890 |
| }, |
| { |
| "epoch": 1.0037318234461459, |
| "grad_norm": 0.9326075315475464, |
| "learning_rate": 3.996138996138996e-05, |
| "loss": 0.736, |
| "step": 3900 |
| }, |
| { |
| "epoch": 1.0063054947883154, |
| "grad_norm": 0.7963811755180359, |
| "learning_rate": 3.993564993564994e-05, |
| "loss": 0.6516, |
| "step": 3910 |
| }, |
| { |
| "epoch": 1.0088791661304852, |
| "grad_norm": 3.7952184677124023, |
| "learning_rate": 3.990990990990991e-05, |
| "loss": 0.7163, |
| "step": 3920 |
| }, |
| { |
| "epoch": 1.0114528374726548, |
| "grad_norm": 3.2275214195251465, |
| "learning_rate": 3.988416988416989e-05, |
| "loss": 0.7229, |
| "step": 3930 |
| }, |
| { |
| "epoch": 1.0140265088148244, |
| "grad_norm": 0.9668965339660645, |
| "learning_rate": 3.985842985842986e-05, |
| "loss": 0.6849, |
| "step": 3940 |
| }, |
| { |
| "epoch": 1.016600180156994, |
| "grad_norm": 1.117182731628418, |
| "learning_rate": 3.983268983268983e-05, |
| "loss": 0.7061, |
| "step": 3950 |
| }, |
| { |
| "epoch": 1.0191738514991635, |
| "grad_norm": 2.518076181411743, |
| "learning_rate": 3.980694980694981e-05, |
| "loss": 0.7505, |
| "step": 3960 |
| }, |
| { |
| "epoch": 1.021747522841333, |
| "grad_norm": 1.0619583129882812, |
| "learning_rate": 3.978120978120978e-05, |
| "loss": 0.7439, |
| "step": 3970 |
| }, |
| { |
| "epoch": 1.0243211941835029, |
| "grad_norm": 0.724909245967865, |
| "learning_rate": 3.975546975546976e-05, |
| "loss": 0.6356, |
| "step": 3980 |
| }, |
| { |
| "epoch": 1.0268948655256724, |
| "grad_norm": 0.8852310180664062, |
| "learning_rate": 3.972972972972973e-05, |
| "loss": 0.6923, |
| "step": 3990 |
| }, |
| { |
| "epoch": 1.029468536867842, |
| "grad_norm": 1.0579869747161865, |
| "learning_rate": 3.970398970398971e-05, |
| "loss": 0.6591, |
| "step": 4000 |
| }, |
| { |
| "epoch": 1.029468536867842, |
| "eval_loss": 0.8084650039672852, |
| "eval_runtime": 395.6791, |
| "eval_samples_per_second": 49.095, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.00037788777625095846, |
| "step": 4000 |
| }, |
| { |
| "epoch": 1.0320422082100116, |
| "grad_norm": 1.5254931449890137, |
| "learning_rate": 3.967824967824968e-05, |
| "loss": 0.7753, |
| "step": 4010 |
| }, |
| { |
| "epoch": 1.0346158795521812, |
| "grad_norm": 0.9235827922821045, |
| "learning_rate": 3.9652509652509654e-05, |
| "loss": 0.6961, |
| "step": 4020 |
| }, |
| { |
| "epoch": 1.0371895508943507, |
| "grad_norm": 1.1861381530761719, |
| "learning_rate": 3.9626769626769626e-05, |
| "loss": 0.6979, |
| "step": 4030 |
| }, |
| { |
| "epoch": 1.0397632222365203, |
| "grad_norm": 1.0676695108413696, |
| "learning_rate": 3.9601029601029604e-05, |
| "loss": 0.7373, |
| "step": 4040 |
| }, |
| { |
| "epoch": 1.04233689357869, |
| "grad_norm": 0.9689359664916992, |
| "learning_rate": 3.9575289575289576e-05, |
| "loss": 0.7406, |
| "step": 4050 |
| }, |
| { |
| "epoch": 1.0449105649208597, |
| "grad_norm": 1.1008409261703491, |
| "learning_rate": 3.9549549549549554e-05, |
| "loss": 0.6522, |
| "step": 4060 |
| }, |
| { |
| "epoch": 1.0474842362630292, |
| "grad_norm": 1.293125033378601, |
| "learning_rate": 3.9523809523809526e-05, |
| "loss": 0.6762, |
| "step": 4070 |
| }, |
| { |
| "epoch": 1.0500579076051988, |
| "grad_norm": 0.9956673979759216, |
| "learning_rate": 3.9498069498069504e-05, |
| "loss": 0.6625, |
| "step": 4080 |
| }, |
| { |
| "epoch": 1.0526315789473684, |
| "grad_norm": 0.8510293960571289, |
| "learning_rate": 3.9472329472329476e-05, |
| "loss": 0.7396, |
| "step": 4090 |
| }, |
| { |
| "epoch": 1.055205250289538, |
| "grad_norm": 1.1811062097549438, |
| "learning_rate": 3.944658944658945e-05, |
| "loss": 0.6562, |
| "step": 4100 |
| }, |
| { |
| "epoch": 1.0577789216317077, |
| "grad_norm": 1.1089203357696533, |
| "learning_rate": 3.942084942084942e-05, |
| "loss": 0.6791, |
| "step": 4110 |
| }, |
| { |
| "epoch": 1.0603525929738773, |
| "grad_norm": 1.135820984840393, |
| "learning_rate": 3.93951093951094e-05, |
| "loss": 0.7083, |
| "step": 4120 |
| }, |
| { |
| "epoch": 1.0629262643160469, |
| "grad_norm": 0.8153538107872009, |
| "learning_rate": 3.936936936936937e-05, |
| "loss": 0.6713, |
| "step": 4130 |
| }, |
| { |
| "epoch": 1.0654999356582164, |
| "grad_norm": 1.1000064611434937, |
| "learning_rate": 3.934362934362935e-05, |
| "loss": 0.7594, |
| "step": 4140 |
| }, |
| { |
| "epoch": 1.068073607000386, |
| "grad_norm": 0.9205760955810547, |
| "learning_rate": 3.931788931788932e-05, |
| "loss": 0.7066, |
| "step": 4150 |
| }, |
| { |
| "epoch": 1.0706472783425556, |
| "grad_norm": 1.1012518405914307, |
| "learning_rate": 3.92921492921493e-05, |
| "loss": 0.691, |
| "step": 4160 |
| }, |
| { |
| "epoch": 1.0732209496847251, |
| "grad_norm": 0.9490509629249573, |
| "learning_rate": 3.926640926640927e-05, |
| "loss": 0.6361, |
| "step": 4170 |
| }, |
| { |
| "epoch": 1.075794621026895, |
| "grad_norm": 0.9934425950050354, |
| "learning_rate": 3.924066924066924e-05, |
| "loss": 0.6546, |
| "step": 4180 |
| }, |
| { |
| "epoch": 1.0783682923690645, |
| "grad_norm": 0.9373430609703064, |
| "learning_rate": 3.921492921492922e-05, |
| "loss": 0.7585, |
| "step": 4190 |
| }, |
| { |
| "epoch": 1.080941963711234, |
| "grad_norm": 0.7465323209762573, |
| "learning_rate": 3.918918918918919e-05, |
| "loss": 0.7026, |
| "step": 4200 |
| }, |
| { |
| "epoch": 1.0835156350534036, |
| "grad_norm": 0.8605757355690002, |
| "learning_rate": 3.916344916344917e-05, |
| "loss": 0.7215, |
| "step": 4210 |
| }, |
| { |
| "epoch": 1.0860893063955732, |
| "grad_norm": 1.27156400680542, |
| "learning_rate": 3.913770913770914e-05, |
| "loss": 0.7393, |
| "step": 4220 |
| }, |
| { |
| "epoch": 1.0886629777377428, |
| "grad_norm": 3.140281915664673, |
| "learning_rate": 3.911196911196912e-05, |
| "loss": 0.6781, |
| "step": 4230 |
| }, |
| { |
| "epoch": 1.0912366490799126, |
| "grad_norm": 1.438792109489441, |
| "learning_rate": 3.908622908622909e-05, |
| "loss": 0.7344, |
| "step": 4240 |
| }, |
| { |
| "epoch": 1.0938103204220821, |
| "grad_norm": 1.1474467515945435, |
| "learning_rate": 3.906048906048906e-05, |
| "loss": 0.6722, |
| "step": 4250 |
| }, |
| { |
| "epoch": 1.0963839917642517, |
| "grad_norm": 0.7898831963539124, |
| "learning_rate": 3.9034749034749034e-05, |
| "loss": 0.709, |
| "step": 4260 |
| }, |
| { |
| "epoch": 1.0989576631064213, |
| "grad_norm": 4.05173397064209, |
| "learning_rate": 3.900900900900901e-05, |
| "loss": 0.682, |
| "step": 4270 |
| }, |
| { |
| "epoch": 1.1015313344485909, |
| "grad_norm": 1.2369790077209473, |
| "learning_rate": 3.8983268983268984e-05, |
| "loss": 0.6313, |
| "step": 4280 |
| }, |
| { |
| "epoch": 1.1041050057907604, |
| "grad_norm": 1.0742212533950806, |
| "learning_rate": 3.895752895752896e-05, |
| "loss": 0.6878, |
| "step": 4290 |
| }, |
| { |
| "epoch": 1.1066786771329302, |
| "grad_norm": 4.394085884094238, |
| "learning_rate": 3.8931788931788934e-05, |
| "loss": 0.6974, |
| "step": 4300 |
| }, |
| { |
| "epoch": 1.1092523484750998, |
| "grad_norm": 1.285081148147583, |
| "learning_rate": 3.890604890604891e-05, |
| "loss": 0.6822, |
| "step": 4310 |
| }, |
| { |
| "epoch": 1.1118260198172694, |
| "grad_norm": 1.0450791120529175, |
| "learning_rate": 3.888030888030888e-05, |
| "loss": 0.6848, |
| "step": 4320 |
| }, |
| { |
| "epoch": 1.114399691159439, |
| "grad_norm": 0.808032214641571, |
| "learning_rate": 3.8854568854568855e-05, |
| "loss": 0.6127, |
| "step": 4330 |
| }, |
| { |
| "epoch": 1.1169733625016085, |
| "grad_norm": 0.6634959578514099, |
| "learning_rate": 3.882882882882883e-05, |
| "loss": 0.6332, |
| "step": 4340 |
| }, |
| { |
| "epoch": 1.119547033843778, |
| "grad_norm": 0.7773045301437378, |
| "learning_rate": 3.8803088803088805e-05, |
| "loss": 0.6558, |
| "step": 4350 |
| }, |
| { |
| "epoch": 1.1221207051859476, |
| "grad_norm": 0.9749086499214172, |
| "learning_rate": 3.877734877734878e-05, |
| "loss": 0.7365, |
| "step": 4360 |
| }, |
| { |
| "epoch": 1.1246943765281174, |
| "grad_norm": 1.0104619264602661, |
| "learning_rate": 3.8751608751608755e-05, |
| "loss": 0.731, |
| "step": 4370 |
| }, |
| { |
| "epoch": 1.127268047870287, |
| "grad_norm": 1.246193766593933, |
| "learning_rate": 3.8725868725868734e-05, |
| "loss": 0.6772, |
| "step": 4380 |
| }, |
| { |
| "epoch": 1.1298417192124566, |
| "grad_norm": 0.9654109477996826, |
| "learning_rate": 3.87001287001287e-05, |
| "loss": 0.7359, |
| "step": 4390 |
| }, |
| { |
| "epoch": 1.1324153905546261, |
| "grad_norm": 0.8749091029167175, |
| "learning_rate": 3.867438867438868e-05, |
| "loss": 0.6908, |
| "step": 4400 |
| }, |
| { |
| "epoch": 1.1349890618967957, |
| "grad_norm": 2.6671011447906494, |
| "learning_rate": 3.864864864864865e-05, |
| "loss": 0.6935, |
| "step": 4410 |
| }, |
| { |
| "epoch": 1.1375627332389655, |
| "grad_norm": 0.834434986114502, |
| "learning_rate": 3.862290862290863e-05, |
| "loss": 0.6978, |
| "step": 4420 |
| }, |
| { |
| "epoch": 1.140136404581135, |
| "grad_norm": 1.0454275608062744, |
| "learning_rate": 3.85971685971686e-05, |
| "loss": 0.7106, |
| "step": 4430 |
| }, |
| { |
| "epoch": 1.1427100759233046, |
| "grad_norm": 1.1424705982208252, |
| "learning_rate": 3.857142857142858e-05, |
| "loss": 0.6781, |
| "step": 4440 |
| }, |
| { |
| "epoch": 1.1452837472654742, |
| "grad_norm": 0.8291309475898743, |
| "learning_rate": 3.854568854568855e-05, |
| "loss": 0.6827, |
| "step": 4450 |
| }, |
| { |
| "epoch": 1.1478574186076438, |
| "grad_norm": 1.2139686346054077, |
| "learning_rate": 3.851994851994852e-05, |
| "loss": 0.6321, |
| "step": 4460 |
| }, |
| { |
| "epoch": 1.1504310899498134, |
| "grad_norm": 0.935051441192627, |
| "learning_rate": 3.849420849420849e-05, |
| "loss": 0.7246, |
| "step": 4470 |
| }, |
| { |
| "epoch": 1.153004761291983, |
| "grad_norm": 2.4891180992126465, |
| "learning_rate": 3.846846846846847e-05, |
| "loss": 0.6742, |
| "step": 4480 |
| }, |
| { |
| "epoch": 1.1555784326341527, |
| "grad_norm": 1.248016357421875, |
| "learning_rate": 3.844272844272844e-05, |
| "loss": 0.7204, |
| "step": 4490 |
| }, |
| { |
| "epoch": 1.1581521039763223, |
| "grad_norm": 3.4102649688720703, |
| "learning_rate": 3.841698841698842e-05, |
| "loss": 0.6985, |
| "step": 4500 |
| }, |
| { |
| "epoch": 1.1607257753184919, |
| "grad_norm": 1.1724501848220825, |
| "learning_rate": 3.839124839124839e-05, |
| "loss": 0.6828, |
| "step": 4510 |
| }, |
| { |
| "epoch": 1.1632994466606614, |
| "grad_norm": 0.8048924207687378, |
| "learning_rate": 3.836550836550837e-05, |
| "loss": 0.6608, |
| "step": 4520 |
| }, |
| { |
| "epoch": 1.165873118002831, |
| "grad_norm": 0.8188682794570923, |
| "learning_rate": 3.833976833976834e-05, |
| "loss": 0.671, |
| "step": 4530 |
| }, |
| { |
| "epoch": 1.1684467893450006, |
| "grad_norm": 0.879068911075592, |
| "learning_rate": 3.831402831402831e-05, |
| "loss": 0.7447, |
| "step": 4540 |
| }, |
| { |
| "epoch": 1.1710204606871701, |
| "grad_norm": 0.9891297817230225, |
| "learning_rate": 3.8288288288288285e-05, |
| "loss": 0.7087, |
| "step": 4550 |
| }, |
| { |
| "epoch": 1.17359413202934, |
| "grad_norm": 1.1570535898208618, |
| "learning_rate": 3.826254826254826e-05, |
| "loss": 0.6537, |
| "step": 4560 |
| }, |
| { |
| "epoch": 1.1761678033715095, |
| "grad_norm": 1.1008327007293701, |
| "learning_rate": 3.8236808236808235e-05, |
| "loss": 0.7138, |
| "step": 4570 |
| }, |
| { |
| "epoch": 1.178741474713679, |
| "grad_norm": 0.8416620492935181, |
| "learning_rate": 3.821106821106821e-05, |
| "loss": 0.6743, |
| "step": 4580 |
| }, |
| { |
| "epoch": 1.1813151460558486, |
| "grad_norm": 0.8778625726699829, |
| "learning_rate": 3.818532818532819e-05, |
| "loss": 0.6572, |
| "step": 4590 |
| }, |
| { |
| "epoch": 1.1838888173980182, |
| "grad_norm": 2.828808307647705, |
| "learning_rate": 3.815958815958816e-05, |
| "loss": 0.7036, |
| "step": 4600 |
| }, |
| { |
| "epoch": 1.1864624887401878, |
| "grad_norm": 1.3439350128173828, |
| "learning_rate": 3.8133848133848135e-05, |
| "loss": 0.759, |
| "step": 4610 |
| }, |
| { |
| "epoch": 1.1890361600823576, |
| "grad_norm": 0.7776147723197937, |
| "learning_rate": 3.8108108108108106e-05, |
| "loss": 0.7424, |
| "step": 4620 |
| }, |
| { |
| "epoch": 1.1916098314245271, |
| "grad_norm": 0.7925708293914795, |
| "learning_rate": 3.8082368082368085e-05, |
| "loss": 0.7157, |
| "step": 4630 |
| }, |
| { |
| "epoch": 1.1941835027666967, |
| "grad_norm": 0.9618962407112122, |
| "learning_rate": 3.8056628056628056e-05, |
| "loss": 0.6598, |
| "step": 4640 |
| }, |
| { |
| "epoch": 1.1967571741088663, |
| "grad_norm": 1.325016975402832, |
| "learning_rate": 3.8030888030888035e-05, |
| "loss": 0.7199, |
| "step": 4650 |
| }, |
| { |
| "epoch": 1.1993308454510359, |
| "grad_norm": 1.2265050411224365, |
| "learning_rate": 3.8005148005148006e-05, |
| "loss": 0.6643, |
| "step": 4660 |
| }, |
| { |
| "epoch": 1.2019045167932054, |
| "grad_norm": 3.7649102210998535, |
| "learning_rate": 3.7979407979407985e-05, |
| "loss": 0.6875, |
| "step": 4670 |
| }, |
| { |
| "epoch": 1.2044781881353752, |
| "grad_norm": 0.7999979257583618, |
| "learning_rate": 3.7953667953667956e-05, |
| "loss": 0.5883, |
| "step": 4680 |
| }, |
| { |
| "epoch": 1.2070518594775448, |
| "grad_norm": 1.012054681777954, |
| "learning_rate": 3.792792792792793e-05, |
| "loss": 0.707, |
| "step": 4690 |
| }, |
| { |
| "epoch": 1.2096255308197144, |
| "grad_norm": 3.4517581462860107, |
| "learning_rate": 3.79021879021879e-05, |
| "loss": 0.6276, |
| "step": 4700 |
| }, |
| { |
| "epoch": 1.212199202161884, |
| "grad_norm": 1.0213309526443481, |
| "learning_rate": 3.787644787644788e-05, |
| "loss": 0.6809, |
| "step": 4710 |
| }, |
| { |
| "epoch": 1.2147728735040535, |
| "grad_norm": 1.3979082107543945, |
| "learning_rate": 3.785070785070785e-05, |
| "loss": 0.7044, |
| "step": 4720 |
| }, |
| { |
| "epoch": 1.217346544846223, |
| "grad_norm": 0.9082636833190918, |
| "learning_rate": 3.782496782496783e-05, |
| "loss": 0.6827, |
| "step": 4730 |
| }, |
| { |
| "epoch": 1.2199202161883926, |
| "grad_norm": 0.9929207563400269, |
| "learning_rate": 3.77992277992278e-05, |
| "loss": 0.6552, |
| "step": 4740 |
| }, |
| { |
| "epoch": 1.2224938875305624, |
| "grad_norm": 0.873182475566864, |
| "learning_rate": 3.777348777348778e-05, |
| "loss": 0.5661, |
| "step": 4750 |
| }, |
| { |
| "epoch": 1.225067558872732, |
| "grad_norm": 1.0508198738098145, |
| "learning_rate": 3.774774774774775e-05, |
| "loss": 0.7024, |
| "step": 4760 |
| }, |
| { |
| "epoch": 1.2276412302149016, |
| "grad_norm": 1.1543952226638794, |
| "learning_rate": 3.772200772200772e-05, |
| "loss": 0.6994, |
| "step": 4770 |
| }, |
| { |
| "epoch": 1.2302149015570711, |
| "grad_norm": 3.8903653621673584, |
| "learning_rate": 3.76962676962677e-05, |
| "loss": 0.7234, |
| "step": 4780 |
| }, |
| { |
| "epoch": 1.2327885728992407, |
| "grad_norm": 1.0679028034210205, |
| "learning_rate": 3.767052767052767e-05, |
| "loss": 0.6768, |
| "step": 4790 |
| }, |
| { |
| "epoch": 1.2353622442414103, |
| "grad_norm": 2.860356330871582, |
| "learning_rate": 3.764478764478765e-05, |
| "loss": 0.7189, |
| "step": 4800 |
| }, |
| { |
| "epoch": 1.23793591558358, |
| "grad_norm": 1.192459225654602, |
| "learning_rate": 3.761904761904762e-05, |
| "loss": 0.7076, |
| "step": 4810 |
| }, |
| { |
| "epoch": 1.2405095869257496, |
| "grad_norm": 0.8065189719200134, |
| "learning_rate": 3.75933075933076e-05, |
| "loss": 0.6556, |
| "step": 4820 |
| }, |
| { |
| "epoch": 1.2430832582679192, |
| "grad_norm": 1.1386466026306152, |
| "learning_rate": 3.756756756756757e-05, |
| "loss": 0.7085, |
| "step": 4830 |
| }, |
| { |
| "epoch": 1.2456569296100888, |
| "grad_norm": 1.2543619871139526, |
| "learning_rate": 3.754182754182754e-05, |
| "loss": 0.7253, |
| "step": 4840 |
| }, |
| { |
| "epoch": 1.2482306009522584, |
| "grad_norm": 1.2314350605010986, |
| "learning_rate": 3.7516087516087514e-05, |
| "loss": 0.6282, |
| "step": 4850 |
| }, |
| { |
| "epoch": 1.250804272294428, |
| "grad_norm": 1.4268747568130493, |
| "learning_rate": 3.749034749034749e-05, |
| "loss": 0.705, |
| "step": 4860 |
| }, |
| { |
| "epoch": 1.2533779436365977, |
| "grad_norm": 0.9429617524147034, |
| "learning_rate": 3.7464607464607464e-05, |
| "loss": 0.702, |
| "step": 4870 |
| }, |
| { |
| "epoch": 1.2559516149787673, |
| "grad_norm": 1.0535614490509033, |
| "learning_rate": 3.743886743886744e-05, |
| "loss": 0.7376, |
| "step": 4880 |
| }, |
| { |
| "epoch": 1.2585252863209369, |
| "grad_norm": 4.010643005371094, |
| "learning_rate": 3.7413127413127414e-05, |
| "loss": 0.6941, |
| "step": 4890 |
| }, |
| { |
| "epoch": 1.2610989576631064, |
| "grad_norm": 0.8559678196907043, |
| "learning_rate": 3.738738738738739e-05, |
| "loss": 0.711, |
| "step": 4900 |
| }, |
| { |
| "epoch": 1.263672629005276, |
| "grad_norm": 0.8388733863830566, |
| "learning_rate": 3.7361647361647364e-05, |
| "loss": 0.7521, |
| "step": 4910 |
| }, |
| { |
| "epoch": 1.2662463003474456, |
| "grad_norm": 3.9231414794921875, |
| "learning_rate": 3.7335907335907336e-05, |
| "loss": 0.6587, |
| "step": 4920 |
| }, |
| { |
| "epoch": 1.2688199716896151, |
| "grad_norm": 1.1947028636932373, |
| "learning_rate": 3.731016731016731e-05, |
| "loss": 0.7029, |
| "step": 4930 |
| }, |
| { |
| "epoch": 1.271393643031785, |
| "grad_norm": 1.23485267162323, |
| "learning_rate": 3.7284427284427286e-05, |
| "loss": 0.7298, |
| "step": 4940 |
| }, |
| { |
| "epoch": 1.2739673143739545, |
| "grad_norm": 1.133660912513733, |
| "learning_rate": 3.725868725868726e-05, |
| "loss": 0.6935, |
| "step": 4950 |
| }, |
| { |
| "epoch": 1.276540985716124, |
| "grad_norm": 3.402822256088257, |
| "learning_rate": 3.7232947232947236e-05, |
| "loss": 0.649, |
| "step": 4960 |
| }, |
| { |
| "epoch": 1.2791146570582936, |
| "grad_norm": 1.2218265533447266, |
| "learning_rate": 3.720720720720721e-05, |
| "loss": 0.734, |
| "step": 4970 |
| }, |
| { |
| "epoch": 1.2816883284004632, |
| "grad_norm": 2.865407705307007, |
| "learning_rate": 3.7181467181467186e-05, |
| "loss": 0.6461, |
| "step": 4980 |
| }, |
| { |
| "epoch": 1.284261999742633, |
| "grad_norm": 1.4591176509857178, |
| "learning_rate": 3.715572715572716e-05, |
| "loss": 0.7582, |
| "step": 4990 |
| }, |
| { |
| "epoch": 1.2868356710848023, |
| "grad_norm": 3.178297519683838, |
| "learning_rate": 3.712998712998713e-05, |
| "loss": 0.7112, |
| "step": 5000 |
| }, |
| { |
| "epoch": 1.2868356710848023, |
| "eval_loss": 0.7959006428718567, |
| "eval_runtime": 395.735, |
| "eval_samples_per_second": 49.088, |
| "eval_steps_per_second": 2.456, |
| "eval_token_accuracy": 0.0003792163621041763, |
| "step": 5000 |
| }, |
| { |
| "epoch": 1.2894093424269721, |
| "grad_norm": 0.7849406003952026, |
| "learning_rate": 3.710424710424711e-05, |
| "loss": 0.6349, |
| "step": 5010 |
| }, |
| { |
| "epoch": 1.2919830137691417, |
| "grad_norm": 1.164996862411499, |
| "learning_rate": 3.707850707850708e-05, |
| "loss": 0.6836, |
| "step": 5020 |
| }, |
| { |
| "epoch": 1.2945566851113113, |
| "grad_norm": 4.410824298858643, |
| "learning_rate": 3.705276705276706e-05, |
| "loss": 0.7274, |
| "step": 5030 |
| }, |
| { |
| "epoch": 1.2971303564534808, |
| "grad_norm": 1.3747788667678833, |
| "learning_rate": 3.702702702702703e-05, |
| "loss": 0.713, |
| "step": 5040 |
| }, |
| { |
| "epoch": 1.2997040277956504, |
| "grad_norm": 3.5286660194396973, |
| "learning_rate": 3.700128700128701e-05, |
| "loss": 0.691, |
| "step": 5050 |
| }, |
| { |
| "epoch": 1.3022776991378202, |
| "grad_norm": 1.0159038305282593, |
| "learning_rate": 3.697554697554698e-05, |
| "loss": 0.7158, |
| "step": 5060 |
| }, |
| { |
| "epoch": 1.3048513704799898, |
| "grad_norm": 1.5588384866714478, |
| "learning_rate": 3.694980694980695e-05, |
| "loss": 0.7042, |
| "step": 5070 |
| }, |
| { |
| "epoch": 1.3074250418221594, |
| "grad_norm": 1.0354678630828857, |
| "learning_rate": 3.692406692406692e-05, |
| "loss": 0.6915, |
| "step": 5080 |
| }, |
| { |
| "epoch": 1.309998713164329, |
| "grad_norm": 3.5660855770111084, |
| "learning_rate": 3.68983268983269e-05, |
| "loss": 0.6111, |
| "step": 5090 |
| }, |
| { |
| "epoch": 1.3125723845064985, |
| "grad_norm": 1.010709285736084, |
| "learning_rate": 3.687258687258687e-05, |
| "loss": 0.6168, |
| "step": 5100 |
| }, |
| { |
| "epoch": 1.315146055848668, |
| "grad_norm": 2.2821505069732666, |
| "learning_rate": 3.684684684684685e-05, |
| "loss": 0.7227, |
| "step": 5110 |
| }, |
| { |
| "epoch": 1.3177197271908376, |
| "grad_norm": 0.8663123250007629, |
| "learning_rate": 3.682110682110682e-05, |
| "loss": 0.645, |
| "step": 5120 |
| }, |
| { |
| "epoch": 1.3202933985330074, |
| "grad_norm": 0.9988008141517639, |
| "learning_rate": 3.67953667953668e-05, |
| "loss": 0.7149, |
| "step": 5130 |
| }, |
| { |
| "epoch": 1.322867069875177, |
| "grad_norm": 1.1684341430664062, |
| "learning_rate": 3.676962676962677e-05, |
| "loss": 0.6919, |
| "step": 5140 |
| }, |
| { |
| "epoch": 1.3254407412173466, |
| "grad_norm": 0.9359066486358643, |
| "learning_rate": 3.6743886743886744e-05, |
| "loss": 0.6732, |
| "step": 5150 |
| }, |
| { |
| "epoch": 1.3280144125595161, |
| "grad_norm": 0.785476565361023, |
| "learning_rate": 3.6718146718146716e-05, |
| "loss": 0.7868, |
| "step": 5160 |
| }, |
| { |
| "epoch": 1.3305880839016857, |
| "grad_norm": 3.040825366973877, |
| "learning_rate": 3.6692406692406694e-05, |
| "loss": 0.704, |
| "step": 5170 |
| }, |
| { |
| "epoch": 1.3331617552438555, |
| "grad_norm": 1.1442739963531494, |
| "learning_rate": 3.6666666666666666e-05, |
| "loss": 0.7099, |
| "step": 5180 |
| }, |
| { |
| "epoch": 1.3357354265860248, |
| "grad_norm": 1.0111069679260254, |
| "learning_rate": 3.6640926640926644e-05, |
| "loss": 0.6733, |
| "step": 5190 |
| }, |
| { |
| "epoch": 1.3383090979281946, |
| "grad_norm": 0.922022819519043, |
| "learning_rate": 3.661518661518662e-05, |
| "loss": 0.7364, |
| "step": 5200 |
| }, |
| { |
| "epoch": 1.3408827692703642, |
| "grad_norm": 1.0119050741195679, |
| "learning_rate": 3.6589446589446594e-05, |
| "loss": 0.7414, |
| "step": 5210 |
| }, |
| { |
| "epoch": 1.3434564406125338, |
| "grad_norm": 1.0058611631393433, |
| "learning_rate": 3.6563706563706566e-05, |
| "loss": 0.7387, |
| "step": 5220 |
| }, |
| { |
| "epoch": 1.3460301119547033, |
| "grad_norm": 2.5699520111083984, |
| "learning_rate": 3.653796653796654e-05, |
| "loss": 0.6831, |
| "step": 5230 |
| }, |
| { |
| "epoch": 1.348603783296873, |
| "grad_norm": 0.723125159740448, |
| "learning_rate": 3.6512226512226516e-05, |
| "loss": 0.6642, |
| "step": 5240 |
| }, |
| { |
| "epoch": 1.3511774546390427, |
| "grad_norm": 0.9540688991546631, |
| "learning_rate": 3.648648648648649e-05, |
| "loss": 0.7312, |
| "step": 5250 |
| }, |
| { |
| "epoch": 1.3537511259812123, |
| "grad_norm": 1.160875916481018, |
| "learning_rate": 3.6460746460746466e-05, |
| "loss": 0.7023, |
| "step": 5260 |
| }, |
| { |
| "epoch": 1.3563247973233818, |
| "grad_norm": 1.0321364402770996, |
| "learning_rate": 3.643500643500644e-05, |
| "loss": 0.6955, |
| "step": 5270 |
| }, |
| { |
| "epoch": 1.3588984686655514, |
| "grad_norm": 0.8098282814025879, |
| "learning_rate": 3.6409266409266416e-05, |
| "loss": 0.6461, |
| "step": 5280 |
| }, |
| { |
| "epoch": 1.361472140007721, |
| "grad_norm": 1.080040454864502, |
| "learning_rate": 3.638352638352639e-05, |
| "loss": 0.7128, |
| "step": 5290 |
| }, |
| { |
| "epoch": 1.3640458113498906, |
| "grad_norm": 0.9302416443824768, |
| "learning_rate": 3.635778635778636e-05, |
| "loss": 0.7007, |
| "step": 5300 |
| }, |
| { |
| "epoch": 1.3666194826920601, |
| "grad_norm": 1.0521005392074585, |
| "learning_rate": 3.633204633204633e-05, |
| "loss": 0.6293, |
| "step": 5310 |
| }, |
| { |
| "epoch": 1.36919315403423, |
| "grad_norm": 1.0561240911483765, |
| "learning_rate": 3.630630630630631e-05, |
| "loss": 0.702, |
| "step": 5320 |
| }, |
| { |
| "epoch": 1.3717668253763995, |
| "grad_norm": 3.3522236347198486, |
| "learning_rate": 3.628056628056628e-05, |
| "loss": 0.6907, |
| "step": 5330 |
| }, |
| { |
| "epoch": 1.374340496718569, |
| "grad_norm": 1.2843527793884277, |
| "learning_rate": 3.625482625482626e-05, |
| "loss": 0.7478, |
| "step": 5340 |
| }, |
| { |
| "epoch": 1.3769141680607386, |
| "grad_norm": 0.7206987142562866, |
| "learning_rate": 3.622908622908623e-05, |
| "loss": 0.7065, |
| "step": 5350 |
| }, |
| { |
| "epoch": 1.3794878394029082, |
| "grad_norm": 2.9926466941833496, |
| "learning_rate": 3.620334620334621e-05, |
| "loss": 0.7363, |
| "step": 5360 |
| }, |
| { |
| "epoch": 1.3820615107450778, |
| "grad_norm": 3.599290609359741, |
| "learning_rate": 3.6177606177606174e-05, |
| "loss": 0.7904, |
| "step": 5370 |
| }, |
| { |
| "epoch": 1.3846351820872473, |
| "grad_norm": 1.1965876817703247, |
| "learning_rate": 3.615186615186615e-05, |
| "loss": 0.6879, |
| "step": 5380 |
| }, |
| { |
| "epoch": 1.3872088534294171, |
| "grad_norm": 0.8417690396308899, |
| "learning_rate": 3.6126126126126124e-05, |
| "loss": 0.7154, |
| "step": 5390 |
| }, |
| { |
| "epoch": 1.3897825247715867, |
| "grad_norm": 2.286024570465088, |
| "learning_rate": 3.61003861003861e-05, |
| "loss": 0.7073, |
| "step": 5400 |
| }, |
| { |
| "epoch": 1.3923561961137563, |
| "grad_norm": 0.7775086164474487, |
| "learning_rate": 3.607464607464608e-05, |
| "loss": 0.6656, |
| "step": 5410 |
| }, |
| { |
| "epoch": 1.3949298674559258, |
| "grad_norm": 0.9922925233840942, |
| "learning_rate": 3.604890604890605e-05, |
| "loss": 0.7027, |
| "step": 5420 |
| }, |
| { |
| "epoch": 1.3975035387980954, |
| "grad_norm": 3.689131021499634, |
| "learning_rate": 3.602316602316603e-05, |
| "loss": 0.6914, |
| "step": 5430 |
| }, |
| { |
| "epoch": 1.4000772101402652, |
| "grad_norm": 1.1228481531143188, |
| "learning_rate": 3.5997425997425995e-05, |
| "loss": 0.6495, |
| "step": 5440 |
| }, |
| { |
| "epoch": 1.4026508814824348, |
| "grad_norm": 1.0339312553405762, |
| "learning_rate": 3.5971685971685974e-05, |
| "loss": 0.6677, |
| "step": 5450 |
| }, |
| { |
| "epoch": 1.4052245528246043, |
| "grad_norm": 1.3286689519882202, |
| "learning_rate": 3.5945945945945945e-05, |
| "loss": 0.7425, |
| "step": 5460 |
| }, |
| { |
| "epoch": 1.407798224166774, |
| "grad_norm": 1.022578477859497, |
| "learning_rate": 3.5920205920205924e-05, |
| "loss": 0.6938, |
| "step": 5470 |
| }, |
| { |
| "epoch": 1.4103718955089435, |
| "grad_norm": 0.8292086124420166, |
| "learning_rate": 3.5894465894465895e-05, |
| "loss": 0.6519, |
| "step": 5480 |
| }, |
| { |
| "epoch": 1.412945566851113, |
| "grad_norm": 0.8953461647033691, |
| "learning_rate": 3.5868725868725874e-05, |
| "loss": 0.6347, |
| "step": 5490 |
| }, |
| { |
| "epoch": 1.4155192381932826, |
| "grad_norm": 1.0440467596054077, |
| "learning_rate": 3.5842985842985845e-05, |
| "loss": 0.6647, |
| "step": 5500 |
| }, |
| { |
| "epoch": 1.4180929095354524, |
| "grad_norm": 1.0310571193695068, |
| "learning_rate": 3.581724581724582e-05, |
| "loss": 0.7084, |
| "step": 5510 |
| }, |
| { |
| "epoch": 1.420666580877622, |
| "grad_norm": 1.1693068742752075, |
| "learning_rate": 3.579150579150579e-05, |
| "loss": 0.6517, |
| "step": 5520 |
| }, |
| { |
| "epoch": 1.4232402522197916, |
| "grad_norm": 0.7255993485450745, |
| "learning_rate": 3.576576576576577e-05, |
| "loss": 0.7003, |
| "step": 5530 |
| }, |
| { |
| "epoch": 1.4258139235619611, |
| "grad_norm": 0.8931599259376526, |
| "learning_rate": 3.574002574002574e-05, |
| "loss": 0.6894, |
| "step": 5540 |
| }, |
| { |
| "epoch": 1.4283875949041307, |
| "grad_norm": 1.551103115081787, |
| "learning_rate": 3.571428571428572e-05, |
| "loss": 0.7459, |
| "step": 5550 |
| }, |
| { |
| "epoch": 1.4309612662463003, |
| "grad_norm": 0.9363420605659485, |
| "learning_rate": 3.568854568854569e-05, |
| "loss": 0.7155, |
| "step": 5560 |
| }, |
| { |
| "epoch": 1.4335349375884698, |
| "grad_norm": 0.7762523889541626, |
| "learning_rate": 3.566280566280567e-05, |
| "loss": 0.6855, |
| "step": 5570 |
| }, |
| { |
| "epoch": 1.4361086089306396, |
| "grad_norm": 3.491201400756836, |
| "learning_rate": 3.563706563706564e-05, |
| "loss": 0.7322, |
| "step": 5580 |
| }, |
| { |
| "epoch": 1.4386822802728092, |
| "grad_norm": 1.4872835874557495, |
| "learning_rate": 3.561132561132561e-05, |
| "loss": 0.7379, |
| "step": 5590 |
| }, |
| { |
| "epoch": 1.4412559516149788, |
| "grad_norm": 3.8220291137695312, |
| "learning_rate": 3.558558558558558e-05, |
| "loss": 0.7969, |
| "step": 5600 |
| }, |
| { |
| "epoch": 1.4438296229571483, |
| "grad_norm": 1.0501987934112549, |
| "learning_rate": 3.555984555984556e-05, |
| "loss": 0.7007, |
| "step": 5610 |
| }, |
| { |
| "epoch": 1.446403294299318, |
| "grad_norm": 0.7753943204879761, |
| "learning_rate": 3.553410553410554e-05, |
| "loss": 0.6857, |
| "step": 5620 |
| }, |
| { |
| "epoch": 1.4489769656414877, |
| "grad_norm": 0.8087158799171448, |
| "learning_rate": 3.550836550836551e-05, |
| "loss": 0.6824, |
| "step": 5630 |
| }, |
| { |
| "epoch": 1.451550636983657, |
| "grad_norm": 1.105362892150879, |
| "learning_rate": 3.548262548262549e-05, |
| "loss": 0.6828, |
| "step": 5640 |
| }, |
| { |
| "epoch": 1.4541243083258268, |
| "grad_norm": 4.095690727233887, |
| "learning_rate": 3.545688545688546e-05, |
| "loss": 0.6885, |
| "step": 5650 |
| }, |
| { |
| "epoch": 1.4566979796679964, |
| "grad_norm": 1.0908024311065674, |
| "learning_rate": 3.543114543114543e-05, |
| "loss": 0.7319, |
| "step": 5660 |
| }, |
| { |
| "epoch": 1.459271651010166, |
| "grad_norm": 1.1036632061004639, |
| "learning_rate": 3.5405405405405403e-05, |
| "loss": 0.713, |
| "step": 5670 |
| }, |
| { |
| "epoch": 1.4618453223523356, |
| "grad_norm": 0.9031351208686829, |
| "learning_rate": 3.537966537966538e-05, |
| "loss": 0.6322, |
| "step": 5680 |
| }, |
| { |
| "epoch": 1.4644189936945051, |
| "grad_norm": 1.1566755771636963, |
| "learning_rate": 3.5353925353925353e-05, |
| "loss": 0.6899, |
| "step": 5690 |
| }, |
| { |
| "epoch": 1.466992665036675, |
| "grad_norm": 0.8092819452285767, |
| "learning_rate": 3.532818532818533e-05, |
| "loss": 0.6701, |
| "step": 5700 |
| }, |
| { |
| "epoch": 1.4695663363788445, |
| "grad_norm": 0.9954015612602234, |
| "learning_rate": 3.5302445302445303e-05, |
| "loss": 0.7205, |
| "step": 5710 |
| }, |
| { |
| "epoch": 1.472140007721014, |
| "grad_norm": 0.8541660308837891, |
| "learning_rate": 3.527670527670528e-05, |
| "loss": 0.6193, |
| "step": 5720 |
| }, |
| { |
| "epoch": 1.4747136790631836, |
| "grad_norm": 3.6179423332214355, |
| "learning_rate": 3.5250965250965253e-05, |
| "loss": 0.7084, |
| "step": 5730 |
| }, |
| { |
| "epoch": 1.4772873504053532, |
| "grad_norm": 1.010477900505066, |
| "learning_rate": 3.5225225225225225e-05, |
| "loss": 0.7267, |
| "step": 5740 |
| }, |
| { |
| "epoch": 1.4798610217475228, |
| "grad_norm": 0.8246733546257019, |
| "learning_rate": 3.51994851994852e-05, |
| "loss": 0.6874, |
| "step": 5750 |
| }, |
| { |
| "epoch": 1.4824346930896923, |
| "grad_norm": 0.7221471071243286, |
| "learning_rate": 3.5173745173745175e-05, |
| "loss": 0.7386, |
| "step": 5760 |
| }, |
| { |
| "epoch": 1.4850083644318621, |
| "grad_norm": 1.1807109117507935, |
| "learning_rate": 3.514800514800515e-05, |
| "loss": 0.7059, |
| "step": 5770 |
| }, |
| { |
| "epoch": 1.4875820357740317, |
| "grad_norm": 1.179556131362915, |
| "learning_rate": 3.5122265122265125e-05, |
| "loss": 0.675, |
| "step": 5780 |
| }, |
| { |
| "epoch": 1.4901557071162013, |
| "grad_norm": 3.564592123031616, |
| "learning_rate": 3.50965250965251e-05, |
| "loss": 0.628, |
| "step": 5790 |
| }, |
| { |
| "epoch": 1.4927293784583708, |
| "grad_norm": 0.8945732712745667, |
| "learning_rate": 3.5070785070785075e-05, |
| "loss": 0.7223, |
| "step": 5800 |
| }, |
| { |
| "epoch": 1.4953030498005404, |
| "grad_norm": 1.0145196914672852, |
| "learning_rate": 3.504504504504505e-05, |
| "loss": 0.6822, |
| "step": 5810 |
| }, |
| { |
| "epoch": 1.4978767211427102, |
| "grad_norm": 1.1235549449920654, |
| "learning_rate": 3.501930501930502e-05, |
| "loss": 0.6738, |
| "step": 5820 |
| }, |
| { |
| "epoch": 1.5004503924848795, |
| "grad_norm": 1.2400606870651245, |
| "learning_rate": 3.4993564993565e-05, |
| "loss": 0.6637, |
| "step": 5830 |
| }, |
| { |
| "epoch": 1.5030240638270493, |
| "grad_norm": 4.3333821296691895, |
| "learning_rate": 3.496782496782497e-05, |
| "loss": 0.6351, |
| "step": 5840 |
| }, |
| { |
| "epoch": 1.505597735169219, |
| "grad_norm": 1.126704216003418, |
| "learning_rate": 3.4942084942084947e-05, |
| "loss": 0.654, |
| "step": 5850 |
| }, |
| { |
| "epoch": 1.5081714065113885, |
| "grad_norm": 1.162214994430542, |
| "learning_rate": 3.491634491634492e-05, |
| "loss": 0.6906, |
| "step": 5860 |
| }, |
| { |
| "epoch": 1.510745077853558, |
| "grad_norm": 1.1057708263397217, |
| "learning_rate": 3.4890604890604897e-05, |
| "loss": 0.6567, |
| "step": 5870 |
| }, |
| { |
| "epoch": 1.5133187491957276, |
| "grad_norm": 2.668668270111084, |
| "learning_rate": 3.486486486486487e-05, |
| "loss": 0.7647, |
| "step": 5880 |
| }, |
| { |
| "epoch": 1.5158924205378974, |
| "grad_norm": 4.099859714508057, |
| "learning_rate": 3.483912483912484e-05, |
| "loss": 0.75, |
| "step": 5890 |
| }, |
| { |
| "epoch": 1.5184660918800668, |
| "grad_norm": 0.9541329145431519, |
| "learning_rate": 3.481338481338481e-05, |
| "loss": 0.7209, |
| "step": 5900 |
| }, |
| { |
| "epoch": 1.5210397632222366, |
| "grad_norm": 0.9891238212585449, |
| "learning_rate": 3.478764478764479e-05, |
| "loss": 0.6793, |
| "step": 5910 |
| }, |
| { |
| "epoch": 1.5236134345644061, |
| "grad_norm": 1.2427529096603394, |
| "learning_rate": 3.476190476190476e-05, |
| "loss": 0.7109, |
| "step": 5920 |
| }, |
| { |
| "epoch": 1.5261871059065757, |
| "grad_norm": 0.8798409104347229, |
| "learning_rate": 3.473616473616474e-05, |
| "loss": 0.7161, |
| "step": 5930 |
| }, |
| { |
| "epoch": 1.5287607772487455, |
| "grad_norm": 0.9166936278343201, |
| "learning_rate": 3.471042471042471e-05, |
| "loss": 0.6958, |
| "step": 5940 |
| }, |
| { |
| "epoch": 1.5313344485909148, |
| "grad_norm": 1.1786912679672241, |
| "learning_rate": 3.468468468468469e-05, |
| "loss": 0.6598, |
| "step": 5950 |
| }, |
| { |
| "epoch": 1.5339081199330846, |
| "grad_norm": 0.8544349670410156, |
| "learning_rate": 3.465894465894466e-05, |
| "loss": 0.6384, |
| "step": 5960 |
| }, |
| { |
| "epoch": 1.5364817912752542, |
| "grad_norm": 1.3337987661361694, |
| "learning_rate": 3.463320463320463e-05, |
| "loss": 0.72, |
| "step": 5970 |
| }, |
| { |
| "epoch": 1.5390554626174238, |
| "grad_norm": 1.6757917404174805, |
| "learning_rate": 3.4607464607464605e-05, |
| "loss": 0.6767, |
| "step": 5980 |
| }, |
| { |
| "epoch": 1.5416291339595933, |
| "grad_norm": 1.042203426361084, |
| "learning_rate": 3.458172458172458e-05, |
| "loss": 0.6199, |
| "step": 5990 |
| }, |
| { |
| "epoch": 1.544202805301763, |
| "grad_norm": 0.9380660057067871, |
| "learning_rate": 3.4555984555984555e-05, |
| "loss": 0.635, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.544202805301763, |
| "eval_loss": 0.7921908497810364, |
| "eval_runtime": 395.6711, |
| "eval_samples_per_second": 49.096, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.0003712448469848693, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.5467764766439327, |
| "grad_norm": 1.0668463706970215, |
| "learning_rate": 3.453024453024453e-05, |
| "loss": 0.7029, |
| "step": 6010 |
| }, |
| { |
| "epoch": 1.549350147986102, |
| "grad_norm": 0.9166046380996704, |
| "learning_rate": 3.4504504504504505e-05, |
| "loss": 0.7513, |
| "step": 6020 |
| }, |
| { |
| "epoch": 1.5519238193282718, |
| "grad_norm": 1.067365050315857, |
| "learning_rate": 3.447876447876448e-05, |
| "loss": 0.6937, |
| "step": 6030 |
| }, |
| { |
| "epoch": 1.5544974906704414, |
| "grad_norm": 1.0622658729553223, |
| "learning_rate": 3.4453024453024455e-05, |
| "loss": 0.7011, |
| "step": 6040 |
| }, |
| { |
| "epoch": 1.557071162012611, |
| "grad_norm": 0.8257264494895935, |
| "learning_rate": 3.4427284427284426e-05, |
| "loss": 0.6929, |
| "step": 6050 |
| }, |
| { |
| "epoch": 1.5596448333547805, |
| "grad_norm": 1.0909701585769653, |
| "learning_rate": 3.440411840411841e-05, |
| "loss": 0.6587, |
| "step": 6060 |
| }, |
| { |
| "epoch": 1.5622185046969501, |
| "grad_norm": 1.0562666654586792, |
| "learning_rate": 3.437837837837838e-05, |
| "loss": 0.7101, |
| "step": 6070 |
| }, |
| { |
| "epoch": 1.56479217603912, |
| "grad_norm": 0.9798877239227295, |
| "learning_rate": 3.4352638352638353e-05, |
| "loss": 0.7049, |
| "step": 6080 |
| }, |
| { |
| "epoch": 1.5673658473812893, |
| "grad_norm": 1.297472357749939, |
| "learning_rate": 3.4326898326898325e-05, |
| "loss": 0.7213, |
| "step": 6090 |
| }, |
| { |
| "epoch": 1.569939518723459, |
| "grad_norm": 0.9715744256973267, |
| "learning_rate": 3.4301158301158303e-05, |
| "loss": 0.667, |
| "step": 6100 |
| }, |
| { |
| "epoch": 1.5725131900656286, |
| "grad_norm": 4.2070817947387695, |
| "learning_rate": 3.4275418275418275e-05, |
| "loss": 0.6737, |
| "step": 6110 |
| }, |
| { |
| "epoch": 1.5750868614077982, |
| "grad_norm": 0.8832525014877319, |
| "learning_rate": 3.4249678249678253e-05, |
| "loss": 0.8282, |
| "step": 6120 |
| }, |
| { |
| "epoch": 1.577660532749968, |
| "grad_norm": 1.0799152851104736, |
| "learning_rate": 3.4223938223938225e-05, |
| "loss": 0.6991, |
| "step": 6130 |
| }, |
| { |
| "epoch": 1.5802342040921373, |
| "grad_norm": 0.986268162727356, |
| "learning_rate": 3.4198198198198203e-05, |
| "loss": 0.6998, |
| "step": 6140 |
| }, |
| { |
| "epoch": 1.5828078754343071, |
| "grad_norm": 1.0623130798339844, |
| "learning_rate": 3.417245817245817e-05, |
| "loss": 0.7329, |
| "step": 6150 |
| }, |
| { |
| "epoch": 1.5853815467764767, |
| "grad_norm": 0.9408344626426697, |
| "learning_rate": 3.414671814671815e-05, |
| "loss": 0.6361, |
| "step": 6160 |
| }, |
| { |
| "epoch": 1.5879552181186463, |
| "grad_norm": 1.0099743604660034, |
| "learning_rate": 3.412097812097812e-05, |
| "loss": 0.6459, |
| "step": 6170 |
| }, |
| { |
| "epoch": 1.5905288894608158, |
| "grad_norm": 1.1268500089645386, |
| "learning_rate": 3.40952380952381e-05, |
| "loss": 0.663, |
| "step": 6180 |
| }, |
| { |
| "epoch": 1.5931025608029854, |
| "grad_norm": 1.2413302659988403, |
| "learning_rate": 3.4069498069498075e-05, |
| "loss": 0.6333, |
| "step": 6190 |
| }, |
| { |
| "epoch": 1.5956762321451552, |
| "grad_norm": 1.3237611055374146, |
| "learning_rate": 3.404375804375805e-05, |
| "loss": 0.6517, |
| "step": 6200 |
| }, |
| { |
| "epoch": 1.5982499034873245, |
| "grad_norm": 0.8348039388656616, |
| "learning_rate": 3.4018018018018025e-05, |
| "loss": 0.6988, |
| "step": 6210 |
| }, |
| { |
| "epoch": 1.6008235748294943, |
| "grad_norm": 1.1034482717514038, |
| "learning_rate": 3.3992277992278e-05, |
| "loss": 0.6542, |
| "step": 6220 |
| }, |
| { |
| "epoch": 1.603397246171664, |
| "grad_norm": 0.817833662033081, |
| "learning_rate": 3.396653796653797e-05, |
| "loss": 0.6705, |
| "step": 6230 |
| }, |
| { |
| "epoch": 1.6059709175138335, |
| "grad_norm": 0.8758569955825806, |
| "learning_rate": 3.394079794079794e-05, |
| "loss": 0.6661, |
| "step": 6240 |
| }, |
| { |
| "epoch": 1.608544588856003, |
| "grad_norm": 1.172839641571045, |
| "learning_rate": 3.391505791505792e-05, |
| "loss": 0.6228, |
| "step": 6250 |
| }, |
| { |
| "epoch": 1.6111182601981726, |
| "grad_norm": 0.9346746802330017, |
| "learning_rate": 3.388931788931789e-05, |
| "loss": 0.6691, |
| "step": 6260 |
| }, |
| { |
| "epoch": 1.6136919315403424, |
| "grad_norm": 0.942145586013794, |
| "learning_rate": 3.386357786357787e-05, |
| "loss": 0.6838, |
| "step": 6270 |
| }, |
| { |
| "epoch": 1.6162656028825118, |
| "grad_norm": 4.026924133300781, |
| "learning_rate": 3.383783783783784e-05, |
| "loss": 0.6771, |
| "step": 6280 |
| }, |
| { |
| "epoch": 1.6188392742246815, |
| "grad_norm": 0.9714291095733643, |
| "learning_rate": 3.381209781209782e-05, |
| "loss": 0.6274, |
| "step": 6290 |
| }, |
| { |
| "epoch": 1.6214129455668511, |
| "grad_norm": 3.0900719165802, |
| "learning_rate": 3.378635778635778e-05, |
| "loss": 0.6962, |
| "step": 6300 |
| }, |
| { |
| "epoch": 1.6239866169090207, |
| "grad_norm": 1.203955888748169, |
| "learning_rate": 3.376061776061776e-05, |
| "loss": 0.7344, |
| "step": 6310 |
| }, |
| { |
| "epoch": 1.6265602882511905, |
| "grad_norm": 1.0462515354156494, |
| "learning_rate": 3.373487773487773e-05, |
| "loss": 0.698, |
| "step": 6320 |
| }, |
| { |
| "epoch": 1.6291339595933598, |
| "grad_norm": 0.8859325647354126, |
| "learning_rate": 3.370913770913771e-05, |
| "loss": 0.6949, |
| "step": 6330 |
| }, |
| { |
| "epoch": 1.6317076309355296, |
| "grad_norm": 1.0180076360702515, |
| "learning_rate": 3.368339768339768e-05, |
| "loss": 0.6681, |
| "step": 6340 |
| }, |
| { |
| "epoch": 1.6342813022776992, |
| "grad_norm": 3.5328500270843506, |
| "learning_rate": 3.365765765765766e-05, |
| "loss": 0.6147, |
| "step": 6350 |
| }, |
| { |
| "epoch": 1.6368549736198688, |
| "grad_norm": 0.9282681345939636, |
| "learning_rate": 3.363191763191763e-05, |
| "loss": 0.7505, |
| "step": 6360 |
| }, |
| { |
| "epoch": 1.6394286449620383, |
| "grad_norm": 0.9516205191612244, |
| "learning_rate": 3.3606177606177605e-05, |
| "loss": 0.7205, |
| "step": 6370 |
| }, |
| { |
| "epoch": 1.642002316304208, |
| "grad_norm": 1.07736337184906, |
| "learning_rate": 3.358043758043758e-05, |
| "loss": 0.6584, |
| "step": 6380 |
| }, |
| { |
| "epoch": 1.6445759876463777, |
| "grad_norm": 0.8120790123939514, |
| "learning_rate": 3.3554697554697555e-05, |
| "loss": 0.6473, |
| "step": 6390 |
| }, |
| { |
| "epoch": 1.647149658988547, |
| "grad_norm": 0.781129777431488, |
| "learning_rate": 3.352895752895753e-05, |
| "loss": 0.673, |
| "step": 6400 |
| }, |
| { |
| "epoch": 1.6497233303307168, |
| "grad_norm": 1.0196880102157593, |
| "learning_rate": 3.3503217503217505e-05, |
| "loss": 0.6674, |
| "step": 6410 |
| }, |
| { |
| "epoch": 1.6522970016728864, |
| "grad_norm": 1.0872300863265991, |
| "learning_rate": 3.347747747747748e-05, |
| "loss": 0.674, |
| "step": 6420 |
| }, |
| { |
| "epoch": 1.654870673015056, |
| "grad_norm": 0.8442071676254272, |
| "learning_rate": 3.3451737451737455e-05, |
| "loss": 0.6944, |
| "step": 6430 |
| }, |
| { |
| "epoch": 1.6574443443572255, |
| "grad_norm": 0.9583492279052734, |
| "learning_rate": 3.3425997425997426e-05, |
| "loss": 0.7546, |
| "step": 6440 |
| }, |
| { |
| "epoch": 1.660018015699395, |
| "grad_norm": 1.018364667892456, |
| "learning_rate": 3.34002574002574e-05, |
| "loss": 0.6987, |
| "step": 6450 |
| }, |
| { |
| "epoch": 1.662591687041565, |
| "grad_norm": 0.833738386631012, |
| "learning_rate": 3.3374517374517376e-05, |
| "loss": 0.7055, |
| "step": 6460 |
| }, |
| { |
| "epoch": 1.6651653583837343, |
| "grad_norm": 1.014647126197815, |
| "learning_rate": 3.334877734877735e-05, |
| "loss": 0.7545, |
| "step": 6470 |
| }, |
| { |
| "epoch": 1.667739029725904, |
| "grad_norm": 1.2637064456939697, |
| "learning_rate": 3.3323037323037326e-05, |
| "loss": 0.728, |
| "step": 6480 |
| }, |
| { |
| "epoch": 1.6703127010680736, |
| "grad_norm": 0.9318623542785645, |
| "learning_rate": 3.32972972972973e-05, |
| "loss": 0.663, |
| "step": 6490 |
| }, |
| { |
| "epoch": 1.6728863724102432, |
| "grad_norm": 1.4659100770950317, |
| "learning_rate": 3.3271557271557276e-05, |
| "loss": 0.7003, |
| "step": 6500 |
| }, |
| { |
| "epoch": 1.675460043752413, |
| "grad_norm": 0.7381783127784729, |
| "learning_rate": 3.324581724581725e-05, |
| "loss": 0.7492, |
| "step": 6510 |
| }, |
| { |
| "epoch": 1.6780337150945823, |
| "grad_norm": 0.9546041488647461, |
| "learning_rate": 3.322007722007722e-05, |
| "loss": 0.6817, |
| "step": 6520 |
| }, |
| { |
| "epoch": 1.6806073864367521, |
| "grad_norm": 0.8316722512245178, |
| "learning_rate": 3.319433719433719e-05, |
| "loss": 0.7269, |
| "step": 6530 |
| }, |
| { |
| "epoch": 1.6831810577789217, |
| "grad_norm": 0.8137519955635071, |
| "learning_rate": 3.316859716859717e-05, |
| "loss": 0.7524, |
| "step": 6540 |
| }, |
| { |
| "epoch": 1.6857547291210913, |
| "grad_norm": 0.772885262966156, |
| "learning_rate": 3.314285714285714e-05, |
| "loss": 0.661, |
| "step": 6550 |
| }, |
| { |
| "epoch": 1.6883284004632608, |
| "grad_norm": 0.9286116361618042, |
| "learning_rate": 3.311711711711712e-05, |
| "loss": 0.7038, |
| "step": 6560 |
| }, |
| { |
| "epoch": 1.6909020718054304, |
| "grad_norm": 1.0171113014221191, |
| "learning_rate": 3.309137709137709e-05, |
| "loss": 0.715, |
| "step": 6570 |
| }, |
| { |
| "epoch": 1.6934757431476002, |
| "grad_norm": 1.098702311515808, |
| "learning_rate": 3.306563706563707e-05, |
| "loss": 0.673, |
| "step": 6580 |
| }, |
| { |
| "epoch": 1.6960494144897695, |
| "grad_norm": 0.9981555342674255, |
| "learning_rate": 3.303989703989704e-05, |
| "loss": 0.6945, |
| "step": 6590 |
| }, |
| { |
| "epoch": 1.6986230858319393, |
| "grad_norm": 1.3243064880371094, |
| "learning_rate": 3.301415701415701e-05, |
| "loss": 0.7054, |
| "step": 6600 |
| }, |
| { |
| "epoch": 1.701196757174109, |
| "grad_norm": 2.013225555419922, |
| "learning_rate": 3.298841698841699e-05, |
| "loss": 0.6962, |
| "step": 6610 |
| }, |
| { |
| "epoch": 1.7037704285162785, |
| "grad_norm": 0.7733390927314758, |
| "learning_rate": 3.296267696267696e-05, |
| "loss": 0.6986, |
| "step": 6620 |
| }, |
| { |
| "epoch": 1.706344099858448, |
| "grad_norm": 0.9102844595909119, |
| "learning_rate": 3.293693693693694e-05, |
| "loss": 0.7269, |
| "step": 6630 |
| }, |
| { |
| "epoch": 1.7089177712006176, |
| "grad_norm": 0.9313980340957642, |
| "learning_rate": 3.291119691119691e-05, |
| "loss": 0.6498, |
| "step": 6640 |
| }, |
| { |
| "epoch": 1.7114914425427874, |
| "grad_norm": 0.9392831325531006, |
| "learning_rate": 3.288545688545689e-05, |
| "loss": 0.7894, |
| "step": 6650 |
| }, |
| { |
| "epoch": 1.7140651138849567, |
| "grad_norm": 1.059606909751892, |
| "learning_rate": 3.285971685971686e-05, |
| "loss": 0.6854, |
| "step": 6660 |
| }, |
| { |
| "epoch": 1.7166387852271265, |
| "grad_norm": 0.7358580827713013, |
| "learning_rate": 3.2833976833976834e-05, |
| "loss": 0.6555, |
| "step": 6670 |
| }, |
| { |
| "epoch": 1.719212456569296, |
| "grad_norm": 0.9238846302032471, |
| "learning_rate": 3.2808236808236806e-05, |
| "loss": 0.7385, |
| "step": 6680 |
| }, |
| { |
| "epoch": 1.7217861279114657, |
| "grad_norm": 2.5383057594299316, |
| "learning_rate": 3.2782496782496784e-05, |
| "loss": 0.6411, |
| "step": 6690 |
| }, |
| { |
| "epoch": 1.7243597992536355, |
| "grad_norm": 1.0750601291656494, |
| "learning_rate": 3.2756756756756756e-05, |
| "loss": 0.6677, |
| "step": 6700 |
| }, |
| { |
| "epoch": 1.7269334705958048, |
| "grad_norm": 3.5669894218444824, |
| "learning_rate": 3.2731016731016734e-05, |
| "loss": 0.7254, |
| "step": 6710 |
| }, |
| { |
| "epoch": 1.7295071419379746, |
| "grad_norm": 1.3590166568756104, |
| "learning_rate": 3.2705276705276706e-05, |
| "loss": 0.5707, |
| "step": 6720 |
| }, |
| { |
| "epoch": 1.732080813280144, |
| "grad_norm": 3.761894702911377, |
| "learning_rate": 3.2679536679536684e-05, |
| "loss": 0.654, |
| "step": 6730 |
| }, |
| { |
| "epoch": 1.7346544846223138, |
| "grad_norm": 1.5816675424575806, |
| "learning_rate": 3.2653796653796656e-05, |
| "loss": 0.7373, |
| "step": 6740 |
| }, |
| { |
| "epoch": 1.7372281559644833, |
| "grad_norm": 1.1728434562683105, |
| "learning_rate": 3.262805662805663e-05, |
| "loss": 0.6679, |
| "step": 6750 |
| }, |
| { |
| "epoch": 1.739801827306653, |
| "grad_norm": 1.1865681409835815, |
| "learning_rate": 3.26023166023166e-05, |
| "loss": 0.6444, |
| "step": 6760 |
| }, |
| { |
| "epoch": 1.7423754986488227, |
| "grad_norm": 3.5320911407470703, |
| "learning_rate": 3.257657657657658e-05, |
| "loss": 0.7412, |
| "step": 6770 |
| }, |
| { |
| "epoch": 1.744949169990992, |
| "grad_norm": 0.9932678937911987, |
| "learning_rate": 3.255083655083655e-05, |
| "loss": 0.6619, |
| "step": 6780 |
| }, |
| { |
| "epoch": 1.7475228413331618, |
| "grad_norm": 0.9329503774642944, |
| "learning_rate": 3.252509652509653e-05, |
| "loss": 0.668, |
| "step": 6790 |
| }, |
| { |
| "epoch": 1.7500965126753314, |
| "grad_norm": 0.91359943151474, |
| "learning_rate": 3.2499356499356506e-05, |
| "loss": 0.7856, |
| "step": 6800 |
| }, |
| { |
| "epoch": 1.752670184017501, |
| "grad_norm": 2.197328567504883, |
| "learning_rate": 3.247361647361648e-05, |
| "loss": 0.7182, |
| "step": 6810 |
| }, |
| { |
| "epoch": 1.7552438553596705, |
| "grad_norm": 1.2823790311813354, |
| "learning_rate": 3.244787644787645e-05, |
| "loss": 0.7202, |
| "step": 6820 |
| }, |
| { |
| "epoch": 1.75781752670184, |
| "grad_norm": 1.1787232160568237, |
| "learning_rate": 3.242213642213642e-05, |
| "loss": 0.6817, |
| "step": 6830 |
| }, |
| { |
| "epoch": 1.76039119804401, |
| "grad_norm": 0.9441166520118713, |
| "learning_rate": 3.23963963963964e-05, |
| "loss": 0.629, |
| "step": 6840 |
| }, |
| { |
| "epoch": 1.7629648693861792, |
| "grad_norm": 2.959947109222412, |
| "learning_rate": 3.237065637065637e-05, |
| "loss": 0.6758, |
| "step": 6850 |
| }, |
| { |
| "epoch": 1.765538540728349, |
| "grad_norm": 0.9195284247398376, |
| "learning_rate": 3.234491634491635e-05, |
| "loss": 0.6811, |
| "step": 6860 |
| }, |
| { |
| "epoch": 1.7681122120705186, |
| "grad_norm": 1.753348708152771, |
| "learning_rate": 3.231917631917632e-05, |
| "loss": 0.6048, |
| "step": 6870 |
| }, |
| { |
| "epoch": 1.7706858834126882, |
| "grad_norm": 0.8950250744819641, |
| "learning_rate": 3.22934362934363e-05, |
| "loss": 0.7056, |
| "step": 6880 |
| }, |
| { |
| "epoch": 1.773259554754858, |
| "grad_norm": 0.9784029722213745, |
| "learning_rate": 3.226769626769627e-05, |
| "loss": 0.664, |
| "step": 6890 |
| }, |
| { |
| "epoch": 1.7758332260970273, |
| "grad_norm": 3.3027687072753906, |
| "learning_rate": 3.224195624195624e-05, |
| "loss": 0.6592, |
| "step": 6900 |
| }, |
| { |
| "epoch": 1.778406897439197, |
| "grad_norm": 2.5597102642059326, |
| "learning_rate": 3.2216216216216214e-05, |
| "loss": 0.6791, |
| "step": 6910 |
| }, |
| { |
| "epoch": 1.7809805687813665, |
| "grad_norm": 1.0470184087753296, |
| "learning_rate": 3.219047619047619e-05, |
| "loss": 0.6964, |
| "step": 6920 |
| }, |
| { |
| "epoch": 1.7835542401235362, |
| "grad_norm": 2.2413344383239746, |
| "learning_rate": 3.2164736164736164e-05, |
| "loss": 0.6795, |
| "step": 6930 |
| }, |
| { |
| "epoch": 1.7861279114657058, |
| "grad_norm": 1.297616958618164, |
| "learning_rate": 3.213899613899614e-05, |
| "loss": 0.6991, |
| "step": 6940 |
| }, |
| { |
| "epoch": 1.7887015828078754, |
| "grad_norm": 0.9208124279975891, |
| "learning_rate": 3.2113256113256114e-05, |
| "loss": 0.706, |
| "step": 6950 |
| }, |
| { |
| "epoch": 1.7912752541500452, |
| "grad_norm": 1.298047423362732, |
| "learning_rate": 3.208751608751609e-05, |
| "loss": 0.6466, |
| "step": 6960 |
| }, |
| { |
| "epoch": 1.7938489254922145, |
| "grad_norm": 0.9276260733604431, |
| "learning_rate": 3.2061776061776064e-05, |
| "loss": 0.6572, |
| "step": 6970 |
| }, |
| { |
| "epoch": 1.7964225968343843, |
| "grad_norm": 0.9074581861495972, |
| "learning_rate": 3.2036036036036036e-05, |
| "loss": 0.743, |
| "step": 6980 |
| }, |
| { |
| "epoch": 1.798996268176554, |
| "grad_norm": 0.9442815184593201, |
| "learning_rate": 3.201029601029601e-05, |
| "loss": 0.7033, |
| "step": 6990 |
| }, |
| { |
| "epoch": 1.8015699395187235, |
| "grad_norm": 1.1410044431686401, |
| "learning_rate": 3.1984555984555986e-05, |
| "loss": 0.7442, |
| "step": 7000 |
| }, |
| { |
| "epoch": 1.8015699395187235, |
| "eval_loss": 0.7747899889945984, |
| "eval_runtime": 395.6047, |
| "eval_samples_per_second": 49.105, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.00037655919039774065, |
| "step": 7000 |
| }, |
| { |
| "epoch": 1.804143610860893, |
| "grad_norm": 0.8645033836364746, |
| "learning_rate": 3.1958815958815964e-05, |
| "loss": 0.6423, |
| "step": 7010 |
| }, |
| { |
| "epoch": 1.8067172822030626, |
| "grad_norm": 1.2555421590805054, |
| "learning_rate": 3.1933075933075936e-05, |
| "loss": 0.5997, |
| "step": 7020 |
| }, |
| { |
| "epoch": 1.8092909535452324, |
| "grad_norm": 2.7066619396209717, |
| "learning_rate": 3.1907335907335914e-05, |
| "loss": 0.659, |
| "step": 7030 |
| }, |
| { |
| "epoch": 1.8118646248874017, |
| "grad_norm": 0.806398868560791, |
| "learning_rate": 3.1881595881595886e-05, |
| "loss": 0.6708, |
| "step": 7040 |
| }, |
| { |
| "epoch": 1.8144382962295715, |
| "grad_norm": 3.4134740829467773, |
| "learning_rate": 3.185585585585586e-05, |
| "loss": 0.6872, |
| "step": 7050 |
| }, |
| { |
| "epoch": 1.817011967571741, |
| "grad_norm": 1.0480446815490723, |
| "learning_rate": 3.183011583011583e-05, |
| "loss": 0.6778, |
| "step": 7060 |
| }, |
| { |
| "epoch": 1.8195856389139107, |
| "grad_norm": 0.9471696019172668, |
| "learning_rate": 3.180437580437581e-05, |
| "loss": 0.6911, |
| "step": 7070 |
| }, |
| { |
| "epoch": 1.8221593102560802, |
| "grad_norm": 4.224599361419678, |
| "learning_rate": 3.177863577863578e-05, |
| "loss": 0.7447, |
| "step": 7080 |
| }, |
| { |
| "epoch": 1.8247329815982498, |
| "grad_norm": 1.0805798768997192, |
| "learning_rate": 3.175289575289576e-05, |
| "loss": 0.704, |
| "step": 7090 |
| }, |
| { |
| "epoch": 1.8273066529404196, |
| "grad_norm": 1.3605204820632935, |
| "learning_rate": 3.172715572715573e-05, |
| "loss": 0.7009, |
| "step": 7100 |
| }, |
| { |
| "epoch": 1.829880324282589, |
| "grad_norm": 1.358485460281372, |
| "learning_rate": 3.170141570141571e-05, |
| "loss": 0.6511, |
| "step": 7110 |
| }, |
| { |
| "epoch": 1.8324539956247587, |
| "grad_norm": 1.068803071975708, |
| "learning_rate": 3.167567567567568e-05, |
| "loss": 0.662, |
| "step": 7120 |
| }, |
| { |
| "epoch": 1.8350276669669283, |
| "grad_norm": 1.6438531875610352, |
| "learning_rate": 3.164993564993565e-05, |
| "loss": 0.7509, |
| "step": 7130 |
| }, |
| { |
| "epoch": 1.8376013383090979, |
| "grad_norm": 0.9765356183052063, |
| "learning_rate": 3.162419562419562e-05, |
| "loss": 0.6796, |
| "step": 7140 |
| }, |
| { |
| "epoch": 1.8401750096512677, |
| "grad_norm": 0.8123814463615417, |
| "learning_rate": 3.15984555984556e-05, |
| "loss": 0.6197, |
| "step": 7150 |
| }, |
| { |
| "epoch": 1.842748680993437, |
| "grad_norm": 0.9948647618293762, |
| "learning_rate": 3.157271557271557e-05, |
| "loss": 0.7236, |
| "step": 7160 |
| }, |
| { |
| "epoch": 1.8453223523356068, |
| "grad_norm": 0.8912683725357056, |
| "learning_rate": 3.154697554697555e-05, |
| "loss": 0.666, |
| "step": 7170 |
| }, |
| { |
| "epoch": 1.8478960236777764, |
| "grad_norm": 1.1189854145050049, |
| "learning_rate": 3.152123552123552e-05, |
| "loss": 0.7255, |
| "step": 7180 |
| }, |
| { |
| "epoch": 1.850469695019946, |
| "grad_norm": 1.395694375038147, |
| "learning_rate": 3.14954954954955e-05, |
| "loss": 0.6849, |
| "step": 7190 |
| }, |
| { |
| "epoch": 1.8530433663621155, |
| "grad_norm": 0.9775906205177307, |
| "learning_rate": 3.1469755469755465e-05, |
| "loss": 0.7228, |
| "step": 7200 |
| }, |
| { |
| "epoch": 1.855617037704285, |
| "grad_norm": 1.1532434225082397, |
| "learning_rate": 3.1444015444015444e-05, |
| "loss": 0.7402, |
| "step": 7210 |
| }, |
| { |
| "epoch": 1.858190709046455, |
| "grad_norm": 0.9169228076934814, |
| "learning_rate": 3.141827541827542e-05, |
| "loss": 0.6879, |
| "step": 7220 |
| }, |
| { |
| "epoch": 1.8607643803886242, |
| "grad_norm": 1.0046635866165161, |
| "learning_rate": 3.1392535392535394e-05, |
| "loss": 0.6767, |
| "step": 7230 |
| }, |
| { |
| "epoch": 1.863338051730794, |
| "grad_norm": 0.9030658602714539, |
| "learning_rate": 3.136679536679537e-05, |
| "loss": 0.6526, |
| "step": 7240 |
| }, |
| { |
| "epoch": 1.8659117230729636, |
| "grad_norm": 0.8485135436058044, |
| "learning_rate": 3.1341055341055344e-05, |
| "loss": 0.6785, |
| "step": 7250 |
| }, |
| { |
| "epoch": 1.8684853944151332, |
| "grad_norm": 1.1969527006149292, |
| "learning_rate": 3.131531531531532e-05, |
| "loss": 0.7079, |
| "step": 7260 |
| }, |
| { |
| "epoch": 1.8710590657573027, |
| "grad_norm": 1.0521395206451416, |
| "learning_rate": 3.1289575289575294e-05, |
| "loss": 0.6408, |
| "step": 7270 |
| }, |
| { |
| "epoch": 1.8736327370994723, |
| "grad_norm": 1.203535556793213, |
| "learning_rate": 3.1263835263835265e-05, |
| "loss": 0.6577, |
| "step": 7280 |
| }, |
| { |
| "epoch": 1.876206408441642, |
| "grad_norm": 1.180917501449585, |
| "learning_rate": 3.123809523809524e-05, |
| "loss": 0.7297, |
| "step": 7290 |
| }, |
| { |
| "epoch": 1.8787800797838115, |
| "grad_norm": 1.5537686347961426, |
| "learning_rate": 3.1212355212355215e-05, |
| "loss": 0.7638, |
| "step": 7300 |
| }, |
| { |
| "epoch": 1.8813537511259812, |
| "grad_norm": 1.0814871788024902, |
| "learning_rate": 3.118661518661519e-05, |
| "loss": 0.6378, |
| "step": 7310 |
| }, |
| { |
| "epoch": 1.8839274224681508, |
| "grad_norm": 7.449766159057617, |
| "learning_rate": 3.1160875160875165e-05, |
| "loss": 0.7747, |
| "step": 7320 |
| }, |
| { |
| "epoch": 1.8865010938103204, |
| "grad_norm": 1.0792584419250488, |
| "learning_rate": 3.113513513513514e-05, |
| "loss": 0.7314, |
| "step": 7330 |
| }, |
| { |
| "epoch": 1.8890747651524902, |
| "grad_norm": 0.8132847547531128, |
| "learning_rate": 3.1109395109395115e-05, |
| "loss": 0.7088, |
| "step": 7340 |
| }, |
| { |
| "epoch": 1.8916484364946595, |
| "grad_norm": 0.9090826511383057, |
| "learning_rate": 3.108365508365508e-05, |
| "loss": 0.6871, |
| "step": 7350 |
| }, |
| { |
| "epoch": 1.8942221078368293, |
| "grad_norm": 0.9906513094902039, |
| "learning_rate": 3.105791505791506e-05, |
| "loss": 0.6722, |
| "step": 7360 |
| }, |
| { |
| "epoch": 1.8967957791789989, |
| "grad_norm": 1.0088332891464233, |
| "learning_rate": 3.103217503217503e-05, |
| "loss": 0.6669, |
| "step": 7370 |
| }, |
| { |
| "epoch": 1.8993694505211685, |
| "grad_norm": 3.6802520751953125, |
| "learning_rate": 3.100643500643501e-05, |
| "loss": 0.6908, |
| "step": 7380 |
| }, |
| { |
| "epoch": 1.901943121863338, |
| "grad_norm": 1.407848596572876, |
| "learning_rate": 3.098069498069498e-05, |
| "loss": 0.6203, |
| "step": 7390 |
| }, |
| { |
| "epoch": 1.9045167932055076, |
| "grad_norm": 1.0382202863693237, |
| "learning_rate": 3.095495495495496e-05, |
| "loss": 0.6618, |
| "step": 7400 |
| }, |
| { |
| "epoch": 1.9070904645476774, |
| "grad_norm": 1.39858877658844, |
| "learning_rate": 3.092921492921493e-05, |
| "loss": 0.6262, |
| "step": 7410 |
| }, |
| { |
| "epoch": 1.9096641358898467, |
| "grad_norm": 0.9530224204063416, |
| "learning_rate": 3.09034749034749e-05, |
| "loss": 0.7161, |
| "step": 7420 |
| }, |
| { |
| "epoch": 1.9122378072320165, |
| "grad_norm": 1.1237123012542725, |
| "learning_rate": 3.087773487773488e-05, |
| "loss": 0.7333, |
| "step": 7430 |
| }, |
| { |
| "epoch": 1.914811478574186, |
| "grad_norm": 2.556382417678833, |
| "learning_rate": 3.085199485199485e-05, |
| "loss": 0.6714, |
| "step": 7440 |
| }, |
| { |
| "epoch": 1.9173851499163557, |
| "grad_norm": 0.9731000065803528, |
| "learning_rate": 3.082625482625483e-05, |
| "loss": 0.6965, |
| "step": 7450 |
| }, |
| { |
| "epoch": 1.9199588212585252, |
| "grad_norm": 3.8856775760650635, |
| "learning_rate": 3.08005148005148e-05, |
| "loss": 0.6913, |
| "step": 7460 |
| }, |
| { |
| "epoch": 1.9225324926006948, |
| "grad_norm": 0.8493612408638, |
| "learning_rate": 3.077477477477478e-05, |
| "loss": 0.6683, |
| "step": 7470 |
| }, |
| { |
| "epoch": 1.9251061639428646, |
| "grad_norm": 3.1507325172424316, |
| "learning_rate": 3.074903474903475e-05, |
| "loss": 0.6804, |
| "step": 7480 |
| }, |
| { |
| "epoch": 1.927679835285034, |
| "grad_norm": 1.0596544742584229, |
| "learning_rate": 3.072329472329472e-05, |
| "loss": 0.6675, |
| "step": 7490 |
| }, |
| { |
| "epoch": 1.9302535066272037, |
| "grad_norm": 0.8844677209854126, |
| "learning_rate": 3.0697554697554695e-05, |
| "loss": 0.7052, |
| "step": 7500 |
| }, |
| { |
| "epoch": 1.9328271779693733, |
| "grad_norm": 3.017733335494995, |
| "learning_rate": 3.067181467181467e-05, |
| "loss": 0.6967, |
| "step": 7510 |
| }, |
| { |
| "epoch": 1.9354008493115429, |
| "grad_norm": 2.120694875717163, |
| "learning_rate": 3.0646074646074645e-05, |
| "loss": 0.7005, |
| "step": 7520 |
| }, |
| { |
| "epoch": 1.9379745206537127, |
| "grad_norm": 0.8085142970085144, |
| "learning_rate": 3.062033462033462e-05, |
| "loss": 0.7189, |
| "step": 7530 |
| }, |
| { |
| "epoch": 1.940548191995882, |
| "grad_norm": 0.9313369393348694, |
| "learning_rate": 3.0594594594594595e-05, |
| "loss": 0.6743, |
| "step": 7540 |
| }, |
| { |
| "epoch": 1.9431218633380518, |
| "grad_norm": 1.2053954601287842, |
| "learning_rate": 3.056885456885457e-05, |
| "loss": 0.6897, |
| "step": 7550 |
| }, |
| { |
| "epoch": 1.9456955346802214, |
| "grad_norm": 0.8504372239112854, |
| "learning_rate": 3.0543114543114545e-05, |
| "loss": 0.6518, |
| "step": 7560 |
| }, |
| { |
| "epoch": 1.948269206022391, |
| "grad_norm": 3.263662576675415, |
| "learning_rate": 3.051737451737452e-05, |
| "loss": 0.6482, |
| "step": 7570 |
| }, |
| { |
| "epoch": 1.9508428773645605, |
| "grad_norm": 0.7058959007263184, |
| "learning_rate": 3.049163449163449e-05, |
| "loss": 0.6701, |
| "step": 7580 |
| }, |
| { |
| "epoch": 1.95341654870673, |
| "grad_norm": 0.9414685964584351, |
| "learning_rate": 3.0465894465894466e-05, |
| "loss": 0.6468, |
| "step": 7590 |
| }, |
| { |
| "epoch": 1.9559902200488999, |
| "grad_norm": 0.9003808498382568, |
| "learning_rate": 3.0440154440154438e-05, |
| "loss": 0.7726, |
| "step": 7600 |
| }, |
| { |
| "epoch": 1.9585638913910692, |
| "grad_norm": 0.8774452209472656, |
| "learning_rate": 3.0414414414414416e-05, |
| "loss": 0.6717, |
| "step": 7610 |
| }, |
| { |
| "epoch": 1.961137562733239, |
| "grad_norm": 1.403334379196167, |
| "learning_rate": 3.0388674388674388e-05, |
| "loss": 0.7249, |
| "step": 7620 |
| }, |
| { |
| "epoch": 1.9637112340754086, |
| "grad_norm": 0.7909294962882996, |
| "learning_rate": 3.0362934362934363e-05, |
| "loss": 0.6854, |
| "step": 7630 |
| }, |
| { |
| "epoch": 1.9662849054175782, |
| "grad_norm": 0.72877037525177, |
| "learning_rate": 3.033719433719434e-05, |
| "loss": 0.6771, |
| "step": 7640 |
| }, |
| { |
| "epoch": 1.9688585767597477, |
| "grad_norm": 1.0888760089874268, |
| "learning_rate": 3.0311454311454313e-05, |
| "loss": 0.723, |
| "step": 7650 |
| }, |
| { |
| "epoch": 1.9714322481019173, |
| "grad_norm": 1.091776728630066, |
| "learning_rate": 3.0285714285714288e-05, |
| "loss": 0.7481, |
| "step": 7660 |
| }, |
| { |
| "epoch": 1.974005919444087, |
| "grad_norm": 1.1127774715423584, |
| "learning_rate": 3.025997425997426e-05, |
| "loss": 0.6342, |
| "step": 7670 |
| }, |
| { |
| "epoch": 1.9765795907862564, |
| "grad_norm": 1.1457735300064087, |
| "learning_rate": 3.0234234234234238e-05, |
| "loss": 0.6916, |
| "step": 7680 |
| }, |
| { |
| "epoch": 1.9791532621284262, |
| "grad_norm": 1.0062847137451172, |
| "learning_rate": 3.020849420849421e-05, |
| "loss": 0.6477, |
| "step": 7690 |
| }, |
| { |
| "epoch": 1.9817269334705958, |
| "grad_norm": 0.9393907785415649, |
| "learning_rate": 3.0182754182754185e-05, |
| "loss": 0.6596, |
| "step": 7700 |
| }, |
| { |
| "epoch": 1.9843006048127654, |
| "grad_norm": 0.8132648468017578, |
| "learning_rate": 3.0157014157014156e-05, |
| "loss": 0.7459, |
| "step": 7710 |
| }, |
| { |
| "epoch": 1.9868742761549352, |
| "grad_norm": 0.8783884644508362, |
| "learning_rate": 3.0131274131274135e-05, |
| "loss": 0.6531, |
| "step": 7720 |
| }, |
| { |
| "epoch": 1.9894479474971045, |
| "grad_norm": 0.982463538646698, |
| "learning_rate": 3.0105534105534106e-05, |
| "loss": 0.6378, |
| "step": 7730 |
| }, |
| { |
| "epoch": 1.9920216188392743, |
| "grad_norm": 2.940614700317383, |
| "learning_rate": 3.007979407979408e-05, |
| "loss": 0.6651, |
| "step": 7740 |
| }, |
| { |
| "epoch": 1.9945952901814439, |
| "grad_norm": 1.3412425518035889, |
| "learning_rate": 3.0054054054054053e-05, |
| "loss": 0.7025, |
| "step": 7750 |
| }, |
| { |
| "epoch": 1.9971689615236135, |
| "grad_norm": 4.413862228393555, |
| "learning_rate": 3.002831402831403e-05, |
| "loss": 0.7143, |
| "step": 7760 |
| }, |
| { |
| "epoch": 1.999742632865783, |
| "grad_norm": 1.016687035560608, |
| "learning_rate": 3.0002574002574003e-05, |
| "loss": 0.6285, |
| "step": 7770 |
| }, |
| { |
| "epoch": 2.0023163042079526, |
| "grad_norm": 2.836697816848755, |
| "learning_rate": 2.9976833976833978e-05, |
| "loss": 0.584, |
| "step": 7780 |
| }, |
| { |
| "epoch": 2.0048899755501224, |
| "grad_norm": 1.2601174116134644, |
| "learning_rate": 2.995109395109395e-05, |
| "loss": 0.509, |
| "step": 7790 |
| }, |
| { |
| "epoch": 2.0074636468922917, |
| "grad_norm": 0.9443026781082153, |
| "learning_rate": 2.9925353925353928e-05, |
| "loss": 0.5627, |
| "step": 7800 |
| }, |
| { |
| "epoch": 2.0100373182344615, |
| "grad_norm": 1.8750293254852295, |
| "learning_rate": 2.98996138996139e-05, |
| "loss": 0.544, |
| "step": 7810 |
| }, |
| { |
| "epoch": 2.012610989576631, |
| "grad_norm": 1.145896553993225, |
| "learning_rate": 2.9873873873873875e-05, |
| "loss": 0.5443, |
| "step": 7820 |
| }, |
| { |
| "epoch": 2.0151846609188007, |
| "grad_norm": 1.1090469360351562, |
| "learning_rate": 2.9848133848133846e-05, |
| "loss": 0.5265, |
| "step": 7830 |
| }, |
| { |
| "epoch": 2.0177583322609705, |
| "grad_norm": 0.9929465651512146, |
| "learning_rate": 2.9822393822393825e-05, |
| "loss": 0.5974, |
| "step": 7840 |
| }, |
| { |
| "epoch": 2.02033200360314, |
| "grad_norm": 0.9841685891151428, |
| "learning_rate": 2.97966537966538e-05, |
| "loss": 0.5093, |
| "step": 7850 |
| }, |
| { |
| "epoch": 2.0229056749453096, |
| "grad_norm": 0.946467399597168, |
| "learning_rate": 2.977091377091377e-05, |
| "loss": 0.4556, |
| "step": 7860 |
| }, |
| { |
| "epoch": 2.025479346287479, |
| "grad_norm": 3.8299388885498047, |
| "learning_rate": 2.974517374517375e-05, |
| "loss": 0.5295, |
| "step": 7870 |
| }, |
| { |
| "epoch": 2.0280530176296487, |
| "grad_norm": 0.9343544840812683, |
| "learning_rate": 2.971943371943372e-05, |
| "loss": 0.4939, |
| "step": 7880 |
| }, |
| { |
| "epoch": 2.030626688971818, |
| "grad_norm": 1.0280673503875732, |
| "learning_rate": 2.9693693693693696e-05, |
| "loss": 0.55, |
| "step": 7890 |
| }, |
| { |
| "epoch": 2.033200360313988, |
| "grad_norm": 1.1726477146148682, |
| "learning_rate": 2.9667953667953668e-05, |
| "loss": 0.5473, |
| "step": 7900 |
| }, |
| { |
| "epoch": 2.0357740316561577, |
| "grad_norm": 1.583446741104126, |
| "learning_rate": 2.9642213642213646e-05, |
| "loss": 0.5337, |
| "step": 7910 |
| }, |
| { |
| "epoch": 2.038347702998327, |
| "grad_norm": 1.0653187036514282, |
| "learning_rate": 2.9616473616473618e-05, |
| "loss": 0.5612, |
| "step": 7920 |
| }, |
| { |
| "epoch": 2.040921374340497, |
| "grad_norm": 1.5087721347808838, |
| "learning_rate": 2.9590733590733593e-05, |
| "loss": 0.5698, |
| "step": 7930 |
| }, |
| { |
| "epoch": 2.043495045682666, |
| "grad_norm": 2.7015230655670166, |
| "learning_rate": 2.9564993564993564e-05, |
| "loss": 0.5721, |
| "step": 7940 |
| }, |
| { |
| "epoch": 2.046068717024836, |
| "grad_norm": 2.5739264488220215, |
| "learning_rate": 2.9539253539253543e-05, |
| "loss": 0.5572, |
| "step": 7950 |
| }, |
| { |
| "epoch": 2.0486423883670057, |
| "grad_norm": 1.3727151155471802, |
| "learning_rate": 2.9513513513513514e-05, |
| "loss": 0.5949, |
| "step": 7960 |
| }, |
| { |
| "epoch": 2.051216059709175, |
| "grad_norm": 1.1039259433746338, |
| "learning_rate": 2.948777348777349e-05, |
| "loss": 0.5954, |
| "step": 7970 |
| }, |
| { |
| "epoch": 2.053789731051345, |
| "grad_norm": 3.637061834335327, |
| "learning_rate": 2.946203346203346e-05, |
| "loss": 0.6065, |
| "step": 7980 |
| }, |
| { |
| "epoch": 2.0563634023935142, |
| "grad_norm": 1.178566336631775, |
| "learning_rate": 2.943629343629344e-05, |
| "loss": 0.5253, |
| "step": 7990 |
| }, |
| { |
| "epoch": 2.058937073735684, |
| "grad_norm": 0.8954353332519531, |
| "learning_rate": 2.941055341055341e-05, |
| "loss": 0.5421, |
| "step": 8000 |
| }, |
| { |
| "epoch": 2.058937073735684, |
| "eval_loss": 0.7958057522773743, |
| "eval_runtime": 395.6991, |
| "eval_samples_per_second": 49.093, |
| "eval_steps_per_second": 2.456, |
| "eval_token_accuracy": 0.0003653611096349046, |
| "step": 8000 |
| }, |
| { |
| "epoch": 2.0615107450778534, |
| "grad_norm": 1.133931040763855, |
| "learning_rate": 2.9384813384813386e-05, |
| "loss": 0.5644, |
| "step": 8010 |
| }, |
| { |
| "epoch": 2.064084416420023, |
| "grad_norm": 1.2688968181610107, |
| "learning_rate": 2.9359073359073358e-05, |
| "loss": 0.5866, |
| "step": 8020 |
| }, |
| { |
| "epoch": 2.066658087762193, |
| "grad_norm": 0.861214816570282, |
| "learning_rate": 2.9333333333333336e-05, |
| "loss": 0.5449, |
| "step": 8030 |
| }, |
| { |
| "epoch": 2.0692317591043623, |
| "grad_norm": 0.9040514230728149, |
| "learning_rate": 2.9307593307593308e-05, |
| "loss": 0.5398, |
| "step": 8040 |
| }, |
| { |
| "epoch": 2.071805430446532, |
| "grad_norm": 1.010221004486084, |
| "learning_rate": 2.9281853281853283e-05, |
| "loss": 0.5111, |
| "step": 8050 |
| }, |
| { |
| "epoch": 2.0743791017887014, |
| "grad_norm": 0.7724061012268066, |
| "learning_rate": 2.925611325611326e-05, |
| "loss": 0.5378, |
| "step": 8060 |
| }, |
| { |
| "epoch": 2.0769527731308712, |
| "grad_norm": 1.2511149644851685, |
| "learning_rate": 2.9230373230373233e-05, |
| "loss": 0.5746, |
| "step": 8070 |
| }, |
| { |
| "epoch": 2.0795264444730406, |
| "grad_norm": 1.4768840074539185, |
| "learning_rate": 2.9204633204633208e-05, |
| "loss": 0.5502, |
| "step": 8080 |
| }, |
| { |
| "epoch": 2.0821001158152104, |
| "grad_norm": 1.175214171409607, |
| "learning_rate": 2.917889317889318e-05, |
| "loss": 0.5351, |
| "step": 8090 |
| }, |
| { |
| "epoch": 2.08467378715738, |
| "grad_norm": 0.9883387088775635, |
| "learning_rate": 2.9153153153153158e-05, |
| "loss": 0.5447, |
| "step": 8100 |
| }, |
| { |
| "epoch": 2.0872474584995495, |
| "grad_norm": 1.3396533727645874, |
| "learning_rate": 2.912998712998713e-05, |
| "loss": 0.641, |
| "step": 8110 |
| }, |
| { |
| "epoch": 2.0898211298417193, |
| "grad_norm": 2.0846426486968994, |
| "learning_rate": 2.910682110682111e-05, |
| "loss": 0.5135, |
| "step": 8120 |
| }, |
| { |
| "epoch": 2.0923948011838887, |
| "grad_norm": 2.5317373275756836, |
| "learning_rate": 2.9081081081081087e-05, |
| "loss": 0.5579, |
| "step": 8130 |
| }, |
| { |
| "epoch": 2.0949684725260584, |
| "grad_norm": 3.6536707878112793, |
| "learning_rate": 2.9055341055341055e-05, |
| "loss": 0.5584, |
| "step": 8140 |
| }, |
| { |
| "epoch": 2.0975421438682282, |
| "grad_norm": 0.9044039249420166, |
| "learning_rate": 2.9029601029601033e-05, |
| "loss": 0.546, |
| "step": 8150 |
| }, |
| { |
| "epoch": 2.1001158152103976, |
| "grad_norm": 1.1351265907287598, |
| "learning_rate": 2.9003861003861005e-05, |
| "loss": 0.6126, |
| "step": 8160 |
| }, |
| { |
| "epoch": 2.1026894865525674, |
| "grad_norm": 1.3066338300704956, |
| "learning_rate": 2.897812097812098e-05, |
| "loss": 0.5059, |
| "step": 8170 |
| }, |
| { |
| "epoch": 2.1052631578947367, |
| "grad_norm": 0.9377551674842834, |
| "learning_rate": 2.8952380952380952e-05, |
| "loss": 0.5245, |
| "step": 8180 |
| }, |
| { |
| "epoch": 2.1078368292369065, |
| "grad_norm": 1.2370498180389404, |
| "learning_rate": 2.892664092664093e-05, |
| "loss": 0.5056, |
| "step": 8190 |
| }, |
| { |
| "epoch": 2.110410500579076, |
| "grad_norm": 2.860896348953247, |
| "learning_rate": 2.8900900900900902e-05, |
| "loss": 0.575, |
| "step": 8200 |
| }, |
| { |
| "epoch": 2.1129841719212457, |
| "grad_norm": 1.3423675298690796, |
| "learning_rate": 2.8875160875160877e-05, |
| "loss": 0.5974, |
| "step": 8210 |
| }, |
| { |
| "epoch": 2.1155578432634154, |
| "grad_norm": 1.3402422666549683, |
| "learning_rate": 2.884942084942085e-05, |
| "loss": 0.582, |
| "step": 8220 |
| }, |
| { |
| "epoch": 2.118131514605585, |
| "grad_norm": 1.225522756576538, |
| "learning_rate": 2.8823680823680827e-05, |
| "loss": 0.5424, |
| "step": 8230 |
| }, |
| { |
| "epoch": 2.1207051859477546, |
| "grad_norm": 2.7710585594177246, |
| "learning_rate": 2.87979407979408e-05, |
| "loss": 0.5898, |
| "step": 8240 |
| }, |
| { |
| "epoch": 2.123278857289924, |
| "grad_norm": 1.2279800176620483, |
| "learning_rate": 2.8772200772200773e-05, |
| "loss": 0.5657, |
| "step": 8250 |
| }, |
| { |
| "epoch": 2.1258525286320937, |
| "grad_norm": 3.662959575653076, |
| "learning_rate": 2.8746460746460745e-05, |
| "loss": 0.4772, |
| "step": 8260 |
| }, |
| { |
| "epoch": 2.128426199974263, |
| "grad_norm": 1.2062429189682007, |
| "learning_rate": 2.8720720720720723e-05, |
| "loss": 0.5685, |
| "step": 8270 |
| }, |
| { |
| "epoch": 2.130999871316433, |
| "grad_norm": 3.6382572650909424, |
| "learning_rate": 2.8694980694980695e-05, |
| "loss": 0.5208, |
| "step": 8280 |
| }, |
| { |
| "epoch": 2.1335735426586027, |
| "grad_norm": 1.2735857963562012, |
| "learning_rate": 2.866924066924067e-05, |
| "loss": 0.6405, |
| "step": 8290 |
| }, |
| { |
| "epoch": 2.136147214000772, |
| "grad_norm": 1.1604303121566772, |
| "learning_rate": 2.864350064350064e-05, |
| "loss": 0.5217, |
| "step": 8300 |
| }, |
| { |
| "epoch": 2.138720885342942, |
| "grad_norm": 5.992644786834717, |
| "learning_rate": 2.861776061776062e-05, |
| "loss": 0.5425, |
| "step": 8310 |
| }, |
| { |
| "epoch": 2.141294556685111, |
| "grad_norm": 1.644361972808838, |
| "learning_rate": 2.859202059202059e-05, |
| "loss": 0.601, |
| "step": 8320 |
| }, |
| { |
| "epoch": 2.143868228027281, |
| "grad_norm": 1.2567894458770752, |
| "learning_rate": 2.8566280566280567e-05, |
| "loss": 0.614, |
| "step": 8330 |
| }, |
| { |
| "epoch": 2.1464418993694503, |
| "grad_norm": 1.1754887104034424, |
| "learning_rate": 2.8540540540540545e-05, |
| "loss": 0.5085, |
| "step": 8340 |
| }, |
| { |
| "epoch": 2.14901557071162, |
| "grad_norm": 0.9711121320724487, |
| "learning_rate": 2.8514800514800517e-05, |
| "loss": 0.6065, |
| "step": 8350 |
| }, |
| { |
| "epoch": 2.15158924205379, |
| "grad_norm": 0.790381669998169, |
| "learning_rate": 2.848906048906049e-05, |
| "loss": 0.5743, |
| "step": 8360 |
| }, |
| { |
| "epoch": 2.1541629133959592, |
| "grad_norm": 1.308838129043579, |
| "learning_rate": 2.8463320463320463e-05, |
| "loss": 0.5701, |
| "step": 8370 |
| }, |
| { |
| "epoch": 2.156736584738129, |
| "grad_norm": 1.1394035816192627, |
| "learning_rate": 2.843758043758044e-05, |
| "loss": 0.5446, |
| "step": 8380 |
| }, |
| { |
| "epoch": 2.1593102560802984, |
| "grad_norm": 0.9981347918510437, |
| "learning_rate": 2.8411840411840413e-05, |
| "loss": 0.5563, |
| "step": 8390 |
| }, |
| { |
| "epoch": 2.161883927422468, |
| "grad_norm": 1.1559290885925293, |
| "learning_rate": 2.8386100386100388e-05, |
| "loss": 0.554, |
| "step": 8400 |
| }, |
| { |
| "epoch": 2.164457598764638, |
| "grad_norm": 1.4176164865493774, |
| "learning_rate": 2.836036036036036e-05, |
| "loss": 0.5751, |
| "step": 8410 |
| }, |
| { |
| "epoch": 2.1670312701068073, |
| "grad_norm": 1.236528992652893, |
| "learning_rate": 2.8334620334620338e-05, |
| "loss": 0.6102, |
| "step": 8420 |
| }, |
| { |
| "epoch": 2.169604941448977, |
| "grad_norm": 1.4529129266738892, |
| "learning_rate": 2.830888030888031e-05, |
| "loss": 0.5281, |
| "step": 8430 |
| }, |
| { |
| "epoch": 2.1721786127911464, |
| "grad_norm": 2.4904539585113525, |
| "learning_rate": 2.8283140283140285e-05, |
| "loss": 0.5257, |
| "step": 8440 |
| }, |
| { |
| "epoch": 2.1747522841333162, |
| "grad_norm": 2.638392448425293, |
| "learning_rate": 2.8257400257400256e-05, |
| "loss": 0.5261, |
| "step": 8450 |
| }, |
| { |
| "epoch": 2.1773259554754856, |
| "grad_norm": 0.9878402352333069, |
| "learning_rate": 2.8231660231660235e-05, |
| "loss": 0.6199, |
| "step": 8460 |
| }, |
| { |
| "epoch": 2.1798996268176554, |
| "grad_norm": 1.2372921705245972, |
| "learning_rate": 2.8205920205920206e-05, |
| "loss": 0.5189, |
| "step": 8470 |
| }, |
| { |
| "epoch": 2.182473298159825, |
| "grad_norm": 1.3234797716140747, |
| "learning_rate": 2.818018018018018e-05, |
| "loss": 0.5475, |
| "step": 8480 |
| }, |
| { |
| "epoch": 2.1850469695019945, |
| "grad_norm": 1.4524154663085938, |
| "learning_rate": 2.8154440154440153e-05, |
| "loss": 0.5446, |
| "step": 8490 |
| }, |
| { |
| "epoch": 2.1876206408441643, |
| "grad_norm": 2.357632637023926, |
| "learning_rate": 2.812870012870013e-05, |
| "loss": 0.5848, |
| "step": 8500 |
| }, |
| { |
| "epoch": 2.1901943121863336, |
| "grad_norm": 0.8181639909744263, |
| "learning_rate": 2.8102960102960103e-05, |
| "loss": 0.4966, |
| "step": 8510 |
| }, |
| { |
| "epoch": 2.1927679835285034, |
| "grad_norm": 1.1121617555618286, |
| "learning_rate": 2.8077220077220078e-05, |
| "loss": 0.5481, |
| "step": 8520 |
| }, |
| { |
| "epoch": 2.1953416548706732, |
| "grad_norm": 1.105263113975525, |
| "learning_rate": 2.805148005148005e-05, |
| "loss": 0.5626, |
| "step": 8530 |
| }, |
| { |
| "epoch": 2.1979153262128426, |
| "grad_norm": 3.460721731185913, |
| "learning_rate": 2.8025740025740028e-05, |
| "loss": 0.55, |
| "step": 8540 |
| }, |
| { |
| "epoch": 2.2004889975550124, |
| "grad_norm": 1.0655298233032227, |
| "learning_rate": 2.8000000000000003e-05, |
| "loss": 0.555, |
| "step": 8550 |
| }, |
| { |
| "epoch": 2.2030626688971817, |
| "grad_norm": 1.5254201889038086, |
| "learning_rate": 2.7974259974259975e-05, |
| "loss": 0.5404, |
| "step": 8560 |
| }, |
| { |
| "epoch": 2.2056363402393515, |
| "grad_norm": 0.9580132961273193, |
| "learning_rate": 2.7948519948519953e-05, |
| "loss": 0.574, |
| "step": 8570 |
| }, |
| { |
| "epoch": 2.208210011581521, |
| "grad_norm": 0.87156081199646, |
| "learning_rate": 2.7922779922779925e-05, |
| "loss": 0.5618, |
| "step": 8580 |
| }, |
| { |
| "epoch": 2.2107836829236907, |
| "grad_norm": 1.130436897277832, |
| "learning_rate": 2.78970398970399e-05, |
| "loss": 0.555, |
| "step": 8590 |
| }, |
| { |
| "epoch": 2.2133573542658604, |
| "grad_norm": 1.059656023979187, |
| "learning_rate": 2.787129987129987e-05, |
| "loss": 0.6065, |
| "step": 8600 |
| }, |
| { |
| "epoch": 2.21593102560803, |
| "grad_norm": 2.916320562362671, |
| "learning_rate": 2.784555984555985e-05, |
| "loss": 0.4926, |
| "step": 8610 |
| }, |
| { |
| "epoch": 2.2185046969501996, |
| "grad_norm": 1.392271876335144, |
| "learning_rate": 2.781981981981982e-05, |
| "loss": 0.5869, |
| "step": 8620 |
| }, |
| { |
| "epoch": 2.221078368292369, |
| "grad_norm": 0.8909618258476257, |
| "learning_rate": 2.7794079794079796e-05, |
| "loss": 0.522, |
| "step": 8630 |
| }, |
| { |
| "epoch": 2.2236520396345387, |
| "grad_norm": 1.329585075378418, |
| "learning_rate": 2.7768339768339768e-05, |
| "loss": 0.6073, |
| "step": 8640 |
| }, |
| { |
| "epoch": 2.226225710976708, |
| "grad_norm": 0.8332410454750061, |
| "learning_rate": 2.7742599742599746e-05, |
| "loss": 0.5942, |
| "step": 8650 |
| }, |
| { |
| "epoch": 2.228799382318878, |
| "grad_norm": 1.180031180381775, |
| "learning_rate": 2.7716859716859718e-05, |
| "loss": 0.5056, |
| "step": 8660 |
| }, |
| { |
| "epoch": 2.2313730536610477, |
| "grad_norm": 0.7509261965751648, |
| "learning_rate": 2.7691119691119693e-05, |
| "loss": 0.5434, |
| "step": 8670 |
| }, |
| { |
| "epoch": 2.233946725003217, |
| "grad_norm": 1.0426994562149048, |
| "learning_rate": 2.7665379665379664e-05, |
| "loss": 0.5641, |
| "step": 8680 |
| }, |
| { |
| "epoch": 2.236520396345387, |
| "grad_norm": 0.997373640537262, |
| "learning_rate": 2.7639639639639643e-05, |
| "loss": 0.5676, |
| "step": 8690 |
| }, |
| { |
| "epoch": 2.239094067687556, |
| "grad_norm": 1.2469213008880615, |
| "learning_rate": 2.7613899613899614e-05, |
| "loss": 0.5751, |
| "step": 8700 |
| }, |
| { |
| "epoch": 2.241667739029726, |
| "grad_norm": 1.0776273012161255, |
| "learning_rate": 2.758815958815959e-05, |
| "loss": 0.529, |
| "step": 8710 |
| }, |
| { |
| "epoch": 2.2442414103718953, |
| "grad_norm": 1.02994966506958, |
| "learning_rate": 2.756241956241956e-05, |
| "loss": 0.5762, |
| "step": 8720 |
| }, |
| { |
| "epoch": 2.246815081714065, |
| "grad_norm": 3.1107001304626465, |
| "learning_rate": 2.753667953667954e-05, |
| "loss": 0.5354, |
| "step": 8730 |
| }, |
| { |
| "epoch": 2.249388753056235, |
| "grad_norm": 5.489597797393799, |
| "learning_rate": 2.751093951093951e-05, |
| "loss": 0.554, |
| "step": 8740 |
| }, |
| { |
| "epoch": 2.251962424398404, |
| "grad_norm": 1.1937365531921387, |
| "learning_rate": 2.7485199485199486e-05, |
| "loss": 0.5603, |
| "step": 8750 |
| }, |
| { |
| "epoch": 2.254536095740574, |
| "grad_norm": 0.7883516550064087, |
| "learning_rate": 2.7459459459459464e-05, |
| "loss": 0.5309, |
| "step": 8760 |
| }, |
| { |
| "epoch": 2.2571097670827434, |
| "grad_norm": 0.9160059094429016, |
| "learning_rate": 2.7433719433719436e-05, |
| "loss": 0.5477, |
| "step": 8770 |
| }, |
| { |
| "epoch": 2.259683438424913, |
| "grad_norm": 1.0978742837905884, |
| "learning_rate": 2.740797940797941e-05, |
| "loss": 0.5444, |
| "step": 8780 |
| }, |
| { |
| "epoch": 2.2622571097670825, |
| "grad_norm": 5.506754398345947, |
| "learning_rate": 2.7382239382239383e-05, |
| "loss": 0.5292, |
| "step": 8790 |
| }, |
| { |
| "epoch": 2.2648307811092523, |
| "grad_norm": 1.0431208610534668, |
| "learning_rate": 2.735649935649936e-05, |
| "loss": 0.522, |
| "step": 8800 |
| }, |
| { |
| "epoch": 2.267404452451422, |
| "grad_norm": 1.238174319267273, |
| "learning_rate": 2.7330759330759333e-05, |
| "loss": 0.5596, |
| "step": 8810 |
| }, |
| { |
| "epoch": 2.2699781237935914, |
| "grad_norm": 3.145578384399414, |
| "learning_rate": 2.7305019305019308e-05, |
| "loss": 0.5143, |
| "step": 8820 |
| }, |
| { |
| "epoch": 2.272551795135761, |
| "grad_norm": 0.794948399066925, |
| "learning_rate": 2.727927927927928e-05, |
| "loss": 0.5643, |
| "step": 8830 |
| }, |
| { |
| "epoch": 2.275125466477931, |
| "grad_norm": 0.9516599774360657, |
| "learning_rate": 2.7253539253539258e-05, |
| "loss": 0.55, |
| "step": 8840 |
| }, |
| { |
| "epoch": 2.2776991378201004, |
| "grad_norm": 1.01409912109375, |
| "learning_rate": 2.722779922779923e-05, |
| "loss": 0.4978, |
| "step": 8850 |
| }, |
| { |
| "epoch": 2.28027280916227, |
| "grad_norm": 2.1791932582855225, |
| "learning_rate": 2.7202059202059204e-05, |
| "loss": 0.568, |
| "step": 8860 |
| }, |
| { |
| "epoch": 2.2828464805044395, |
| "grad_norm": 0.9807432889938354, |
| "learning_rate": 2.7176319176319176e-05, |
| "loss": 0.6153, |
| "step": 8870 |
| }, |
| { |
| "epoch": 2.2854201518466093, |
| "grad_norm": 0.947486400604248, |
| "learning_rate": 2.7150579150579154e-05, |
| "loss": 0.5095, |
| "step": 8880 |
| }, |
| { |
| "epoch": 2.2879938231887786, |
| "grad_norm": 1.0243611335754395, |
| "learning_rate": 2.7124839124839126e-05, |
| "loss": 0.5296, |
| "step": 8890 |
| }, |
| { |
| "epoch": 2.2905674945309484, |
| "grad_norm": 0.9641122817993164, |
| "learning_rate": 2.70990990990991e-05, |
| "loss": 0.5597, |
| "step": 8900 |
| }, |
| { |
| "epoch": 2.2931411658731182, |
| "grad_norm": 0.9149156212806702, |
| "learning_rate": 2.7073359073359072e-05, |
| "loss": 0.5361, |
| "step": 8910 |
| }, |
| { |
| "epoch": 2.2957148372152876, |
| "grad_norm": 1.1571760177612305, |
| "learning_rate": 2.704761904761905e-05, |
| "loss": 0.5116, |
| "step": 8920 |
| }, |
| { |
| "epoch": 2.2982885085574574, |
| "grad_norm": 1.1743327379226685, |
| "learning_rate": 2.7021879021879022e-05, |
| "loss": 0.5913, |
| "step": 8930 |
| }, |
| { |
| "epoch": 2.3008621798996267, |
| "grad_norm": 0.9054014682769775, |
| "learning_rate": 2.6996138996138997e-05, |
| "loss": 0.5819, |
| "step": 8940 |
| }, |
| { |
| "epoch": 2.3034358512417965, |
| "grad_norm": 1.0355494022369385, |
| "learning_rate": 2.697039897039897e-05, |
| "loss": 0.5633, |
| "step": 8950 |
| }, |
| { |
| "epoch": 2.306009522583966, |
| "grad_norm": 2.785370111465454, |
| "learning_rate": 2.6944658944658947e-05, |
| "loss": 0.5018, |
| "step": 8960 |
| }, |
| { |
| "epoch": 2.3085831939261356, |
| "grad_norm": 1.2618205547332764, |
| "learning_rate": 2.6918918918918922e-05, |
| "loss": 0.5516, |
| "step": 8970 |
| }, |
| { |
| "epoch": 2.3111568652683054, |
| "grad_norm": 1.069515347480774, |
| "learning_rate": 2.6893178893178894e-05, |
| "loss": 0.5578, |
| "step": 8980 |
| }, |
| { |
| "epoch": 2.313730536610475, |
| "grad_norm": 2.962970018386841, |
| "learning_rate": 2.6867438867438872e-05, |
| "loss": 0.565, |
| "step": 8990 |
| }, |
| { |
| "epoch": 2.3163042079526446, |
| "grad_norm": 1.353009581565857, |
| "learning_rate": 2.6841698841698844e-05, |
| "loss": 0.507, |
| "step": 9000 |
| }, |
| { |
| "epoch": 2.3163042079526446, |
| "eval_loss": 0.7905660271644592, |
| "eval_runtime": 395.6161, |
| "eval_samples_per_second": 49.103, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.00036479171569781123, |
| "step": 9000 |
| }, |
| { |
| "epoch": 2.318877879294814, |
| "grad_norm": 1.2040934562683105, |
| "learning_rate": 2.681595881595882e-05, |
| "loss": 0.5755, |
| "step": 9010 |
| }, |
| { |
| "epoch": 2.3214515506369837, |
| "grad_norm": 0.8578842282295227, |
| "learning_rate": 2.679021879021879e-05, |
| "loss": 0.5335, |
| "step": 9020 |
| }, |
| { |
| "epoch": 2.324025221979153, |
| "grad_norm": 0.989274799823761, |
| "learning_rate": 2.676447876447877e-05, |
| "loss": 0.5647, |
| "step": 9030 |
| }, |
| { |
| "epoch": 2.326598893321323, |
| "grad_norm": 1.0703078508377075, |
| "learning_rate": 2.6738738738738737e-05, |
| "loss": 0.5929, |
| "step": 9040 |
| }, |
| { |
| "epoch": 2.3291725646634927, |
| "grad_norm": 0.8797844052314758, |
| "learning_rate": 2.6712998712998716e-05, |
| "loss": 0.548, |
| "step": 9050 |
| }, |
| { |
| "epoch": 2.331746236005662, |
| "grad_norm": 1.1639105081558228, |
| "learning_rate": 2.6687258687258687e-05, |
| "loss": 0.5716, |
| "step": 9060 |
| }, |
| { |
| "epoch": 2.334319907347832, |
| "grad_norm": 1.3019888401031494, |
| "learning_rate": 2.6661518661518666e-05, |
| "loss": 0.5555, |
| "step": 9070 |
| }, |
| { |
| "epoch": 2.336893578690001, |
| "grad_norm": 4.5147271156311035, |
| "learning_rate": 2.6635778635778634e-05, |
| "loss": 0.6358, |
| "step": 9080 |
| }, |
| { |
| "epoch": 2.339467250032171, |
| "grad_norm": 1.306016206741333, |
| "learning_rate": 2.6610038610038612e-05, |
| "loss": 0.5318, |
| "step": 9090 |
| }, |
| { |
| "epoch": 2.3420409213743403, |
| "grad_norm": 1.211195468902588, |
| "learning_rate": 2.6584298584298584e-05, |
| "loss": 0.5253, |
| "step": 9100 |
| }, |
| { |
| "epoch": 2.34461459271651, |
| "grad_norm": 1.1488116979599, |
| "learning_rate": 2.655855855855856e-05, |
| "loss": 0.6381, |
| "step": 9110 |
| }, |
| { |
| "epoch": 2.34718826405868, |
| "grad_norm": 1.1285436153411865, |
| "learning_rate": 2.653281853281853e-05, |
| "loss": 0.5281, |
| "step": 9120 |
| }, |
| { |
| "epoch": 2.349761935400849, |
| "grad_norm": 1.651699423789978, |
| "learning_rate": 2.650707850707851e-05, |
| "loss": 0.5664, |
| "step": 9130 |
| }, |
| { |
| "epoch": 2.352335606743019, |
| "grad_norm": 1.1540205478668213, |
| "learning_rate": 2.648133848133848e-05, |
| "loss": 0.5808, |
| "step": 9140 |
| }, |
| { |
| "epoch": 2.3549092780851884, |
| "grad_norm": 1.0008643865585327, |
| "learning_rate": 2.6455598455598455e-05, |
| "loss": 0.5613, |
| "step": 9150 |
| }, |
| { |
| "epoch": 2.357482949427358, |
| "grad_norm": 0.9590166211128235, |
| "learning_rate": 2.6429858429858427e-05, |
| "loss": 0.6062, |
| "step": 9160 |
| }, |
| { |
| "epoch": 2.3600566207695275, |
| "grad_norm": 0.9757869243621826, |
| "learning_rate": 2.6404118404118405e-05, |
| "loss": 0.5632, |
| "step": 9170 |
| }, |
| { |
| "epoch": 2.3626302921116973, |
| "grad_norm": 0.9153413772583008, |
| "learning_rate": 2.6378378378378384e-05, |
| "loss": 0.5501, |
| "step": 9180 |
| }, |
| { |
| "epoch": 2.365203963453867, |
| "grad_norm": 1.4172649383544922, |
| "learning_rate": 2.6352638352638352e-05, |
| "loss": 0.5803, |
| "step": 9190 |
| }, |
| { |
| "epoch": 2.3677776347960364, |
| "grad_norm": 1.612728238105774, |
| "learning_rate": 2.632689832689833e-05, |
| "loss": 0.6375, |
| "step": 9200 |
| }, |
| { |
| "epoch": 2.370351306138206, |
| "grad_norm": 1.0975571870803833, |
| "learning_rate": 2.6301158301158302e-05, |
| "loss": 0.5791, |
| "step": 9210 |
| }, |
| { |
| "epoch": 2.3729249774803756, |
| "grad_norm": 3.388500452041626, |
| "learning_rate": 2.6275418275418277e-05, |
| "loss": 0.5247, |
| "step": 9220 |
| }, |
| { |
| "epoch": 2.3754986488225454, |
| "grad_norm": 0.9185710549354553, |
| "learning_rate": 2.624967824967825e-05, |
| "loss": 0.508, |
| "step": 9230 |
| }, |
| { |
| "epoch": 2.378072320164715, |
| "grad_norm": 1.2001655101776123, |
| "learning_rate": 2.6223938223938227e-05, |
| "loss": 0.5771, |
| "step": 9240 |
| }, |
| { |
| "epoch": 2.3806459915068845, |
| "grad_norm": 1.817859411239624, |
| "learning_rate": 2.61981981981982e-05, |
| "loss": 0.5443, |
| "step": 9250 |
| }, |
| { |
| "epoch": 2.3832196628490543, |
| "grad_norm": 1.0994833707809448, |
| "learning_rate": 2.6172458172458174e-05, |
| "loss": 0.5538, |
| "step": 9260 |
| }, |
| { |
| "epoch": 2.3857933341912236, |
| "grad_norm": 1.3039990663528442, |
| "learning_rate": 2.6146718146718145e-05, |
| "loss": 0.5196, |
| "step": 9270 |
| }, |
| { |
| "epoch": 2.3883670055333934, |
| "grad_norm": 2.651137113571167, |
| "learning_rate": 2.6120978120978124e-05, |
| "loss": 0.5538, |
| "step": 9280 |
| }, |
| { |
| "epoch": 2.390940676875563, |
| "grad_norm": 1.00489342212677, |
| "learning_rate": 2.6095238095238095e-05, |
| "loss": 0.5978, |
| "step": 9290 |
| }, |
| { |
| "epoch": 2.3935143482177326, |
| "grad_norm": 2.647886276245117, |
| "learning_rate": 2.606949806949807e-05, |
| "loss": 0.5674, |
| "step": 9300 |
| }, |
| { |
| "epoch": 2.3960880195599024, |
| "grad_norm": 1.094959020614624, |
| "learning_rate": 2.6043758043758042e-05, |
| "loss": 0.5768, |
| "step": 9310 |
| }, |
| { |
| "epoch": 2.3986616909020717, |
| "grad_norm": 3.2877941131591797, |
| "learning_rate": 2.601801801801802e-05, |
| "loss": 0.5772, |
| "step": 9320 |
| }, |
| { |
| "epoch": 2.4012353622442415, |
| "grad_norm": 0.7123093008995056, |
| "learning_rate": 2.5992277992277992e-05, |
| "loss": 0.5598, |
| "step": 9330 |
| }, |
| { |
| "epoch": 2.403809033586411, |
| "grad_norm": 0.902870237827301, |
| "learning_rate": 2.5966537966537967e-05, |
| "loss": 0.5089, |
| "step": 9340 |
| }, |
| { |
| "epoch": 2.4063827049285806, |
| "grad_norm": 1.3696403503417969, |
| "learning_rate": 2.594079794079794e-05, |
| "loss": 0.5726, |
| "step": 9350 |
| }, |
| { |
| "epoch": 2.4089563762707504, |
| "grad_norm": 3.460519790649414, |
| "learning_rate": 2.5915057915057917e-05, |
| "loss": 0.5474, |
| "step": 9360 |
| }, |
| { |
| "epoch": 2.41153004761292, |
| "grad_norm": 1.307590365409851, |
| "learning_rate": 2.588931788931789e-05, |
| "loss": 0.6438, |
| "step": 9370 |
| }, |
| { |
| "epoch": 2.4141037189550896, |
| "grad_norm": 3.9630610942840576, |
| "learning_rate": 2.5863577863577864e-05, |
| "loss": 0.5163, |
| "step": 9380 |
| }, |
| { |
| "epoch": 2.416677390297259, |
| "grad_norm": 1.028784990310669, |
| "learning_rate": 2.5837837837837842e-05, |
| "loss": 0.5345, |
| "step": 9390 |
| }, |
| { |
| "epoch": 2.4192510616394287, |
| "grad_norm": 0.9117390513420105, |
| "learning_rate": 2.5812097812097814e-05, |
| "loss": 0.5966, |
| "step": 9400 |
| }, |
| { |
| "epoch": 2.421824732981598, |
| "grad_norm": 3.028024911880493, |
| "learning_rate": 2.578635778635779e-05, |
| "loss": 0.5833, |
| "step": 9410 |
| }, |
| { |
| "epoch": 2.424398404323768, |
| "grad_norm": 1.1407766342163086, |
| "learning_rate": 2.576061776061776e-05, |
| "loss": 0.5501, |
| "step": 9420 |
| }, |
| { |
| "epoch": 2.4269720756659376, |
| "grad_norm": 1.0818943977355957, |
| "learning_rate": 2.573487773487774e-05, |
| "loss": 0.5323, |
| "step": 9430 |
| }, |
| { |
| "epoch": 2.429545747008107, |
| "grad_norm": 1.1713372468948364, |
| "learning_rate": 2.570913770913771e-05, |
| "loss": 0.6247, |
| "step": 9440 |
| }, |
| { |
| "epoch": 2.432119418350277, |
| "grad_norm": 1.3347368240356445, |
| "learning_rate": 2.5683397683397685e-05, |
| "loss": 0.6161, |
| "step": 9450 |
| }, |
| { |
| "epoch": 2.434693089692446, |
| "grad_norm": 1.3403719663619995, |
| "learning_rate": 2.5657657657657657e-05, |
| "loss": 0.543, |
| "step": 9460 |
| }, |
| { |
| "epoch": 2.437266761034616, |
| "grad_norm": 1.3469754457473755, |
| "learning_rate": 2.5631917631917635e-05, |
| "loss": 0.5616, |
| "step": 9470 |
| }, |
| { |
| "epoch": 2.4398404323767853, |
| "grad_norm": 1.171751856803894, |
| "learning_rate": 2.5606177606177607e-05, |
| "loss": 0.5982, |
| "step": 9480 |
| }, |
| { |
| "epoch": 2.442414103718955, |
| "grad_norm": 2.609652042388916, |
| "learning_rate": 2.5580437580437582e-05, |
| "loss": 0.5782, |
| "step": 9490 |
| }, |
| { |
| "epoch": 2.444987775061125, |
| "grad_norm": 1.0340471267700195, |
| "learning_rate": 2.5554697554697553e-05, |
| "loss": 0.5698, |
| "step": 9500 |
| }, |
| { |
| "epoch": 2.447561446403294, |
| "grad_norm": 0.9706631302833557, |
| "learning_rate": 2.5528957528957532e-05, |
| "loss": 0.5886, |
| "step": 9510 |
| }, |
| { |
| "epoch": 2.450135117745464, |
| "grad_norm": 1.287369966506958, |
| "learning_rate": 2.5503217503217503e-05, |
| "loss": 0.5721, |
| "step": 9520 |
| }, |
| { |
| "epoch": 2.4527087890876333, |
| "grad_norm": 0.9612070918083191, |
| "learning_rate": 2.547747747747748e-05, |
| "loss": 0.6088, |
| "step": 9530 |
| }, |
| { |
| "epoch": 2.455282460429803, |
| "grad_norm": 2.4824914932250977, |
| "learning_rate": 2.545173745173745e-05, |
| "loss": 0.5057, |
| "step": 9540 |
| }, |
| { |
| "epoch": 2.4578561317719725, |
| "grad_norm": 2.4112958908081055, |
| "learning_rate": 2.542599742599743e-05, |
| "loss": 0.5689, |
| "step": 9550 |
| }, |
| { |
| "epoch": 2.4604298031141423, |
| "grad_norm": 0.7653863430023193, |
| "learning_rate": 2.54002574002574e-05, |
| "loss": 0.5119, |
| "step": 9560 |
| }, |
| { |
| "epoch": 2.463003474456312, |
| "grad_norm": 1.1979918479919434, |
| "learning_rate": 2.5374517374517375e-05, |
| "loss": 0.5162, |
| "step": 9570 |
| }, |
| { |
| "epoch": 2.4655771457984814, |
| "grad_norm": 1.1979789733886719, |
| "learning_rate": 2.5348777348777347e-05, |
| "loss": 0.5627, |
| "step": 9580 |
| }, |
| { |
| "epoch": 2.468150817140651, |
| "grad_norm": 2.0756313800811768, |
| "learning_rate": 2.5323037323037325e-05, |
| "loss": 0.5502, |
| "step": 9590 |
| }, |
| { |
| "epoch": 2.4707244884828206, |
| "grad_norm": 3.4498813152313232, |
| "learning_rate": 2.52972972972973e-05, |
| "loss": 0.5584, |
| "step": 9600 |
| }, |
| { |
| "epoch": 2.4732981598249903, |
| "grad_norm": 1.145731806755066, |
| "learning_rate": 2.527155727155727e-05, |
| "loss": 0.5082, |
| "step": 9610 |
| }, |
| { |
| "epoch": 2.47587183116716, |
| "grad_norm": 1.515875220298767, |
| "learning_rate": 2.524581724581725e-05, |
| "loss": 0.5051, |
| "step": 9620 |
| }, |
| { |
| "epoch": 2.4784455025093295, |
| "grad_norm": 0.8977373838424683, |
| "learning_rate": 2.522007722007722e-05, |
| "loss": 0.6196, |
| "step": 9630 |
| }, |
| { |
| "epoch": 2.4810191738514993, |
| "grad_norm": 1.2553268671035767, |
| "learning_rate": 2.5194337194337197e-05, |
| "loss": 0.5503, |
| "step": 9640 |
| }, |
| { |
| "epoch": 2.4835928451936686, |
| "grad_norm": 0.8410692811012268, |
| "learning_rate": 2.5168597168597168e-05, |
| "loss": 0.5605, |
| "step": 9650 |
| }, |
| { |
| "epoch": 2.4861665165358384, |
| "grad_norm": 1.2535247802734375, |
| "learning_rate": 2.5142857142857147e-05, |
| "loss": 0.5164, |
| "step": 9660 |
| }, |
| { |
| "epoch": 2.488740187878008, |
| "grad_norm": 1.1014875173568726, |
| "learning_rate": 2.5117117117117118e-05, |
| "loss": 0.5934, |
| "step": 9670 |
| }, |
| { |
| "epoch": 2.4913138592201776, |
| "grad_norm": 1.0416326522827148, |
| "learning_rate": 2.5091377091377093e-05, |
| "loss": 0.6187, |
| "step": 9680 |
| }, |
| { |
| "epoch": 2.4938875305623474, |
| "grad_norm": 1.3807120323181152, |
| "learning_rate": 2.5065637065637065e-05, |
| "loss": 0.5462, |
| "step": 9690 |
| }, |
| { |
| "epoch": 2.4964612019045167, |
| "grad_norm": 1.4746110439300537, |
| "learning_rate": 2.5039897039897043e-05, |
| "loss": 0.5576, |
| "step": 9700 |
| }, |
| { |
| "epoch": 2.4990348732466865, |
| "grad_norm": 1.372180461883545, |
| "learning_rate": 2.5014157014157015e-05, |
| "loss": 0.5218, |
| "step": 9710 |
| }, |
| { |
| "epoch": 2.501608544588856, |
| "grad_norm": 0.8345748782157898, |
| "learning_rate": 2.498841698841699e-05, |
| "loss": 0.5393, |
| "step": 9720 |
| }, |
| { |
| "epoch": 2.5041822159310256, |
| "grad_norm": 1.1809195280075073, |
| "learning_rate": 2.4962676962676965e-05, |
| "loss": 0.5577, |
| "step": 9730 |
| }, |
| { |
| "epoch": 2.5067558872731954, |
| "grad_norm": 1.1558030843734741, |
| "learning_rate": 2.493693693693694e-05, |
| "loss": 0.6091, |
| "step": 9740 |
| }, |
| { |
| "epoch": 2.5093295586153648, |
| "grad_norm": 0.8379091620445251, |
| "learning_rate": 2.491119691119691e-05, |
| "loss": 0.5787, |
| "step": 9750 |
| }, |
| { |
| "epoch": 2.5119032299575346, |
| "grad_norm": 1.2100491523742676, |
| "learning_rate": 2.4885456885456886e-05, |
| "loss": 0.5179, |
| "step": 9760 |
| }, |
| { |
| "epoch": 2.514476901299704, |
| "grad_norm": 2.15959095954895, |
| "learning_rate": 2.485971685971686e-05, |
| "loss": 0.5448, |
| "step": 9770 |
| }, |
| { |
| "epoch": 2.5170505726418737, |
| "grad_norm": 1.1132477521896362, |
| "learning_rate": 2.4833976833976836e-05, |
| "loss": 0.5354, |
| "step": 9780 |
| }, |
| { |
| "epoch": 2.519624243984043, |
| "grad_norm": 1.1005936861038208, |
| "learning_rate": 2.4808236808236808e-05, |
| "loss": 0.6376, |
| "step": 9790 |
| }, |
| { |
| "epoch": 2.522197915326213, |
| "grad_norm": 1.0520378351211548, |
| "learning_rate": 2.4782496782496783e-05, |
| "loss": 0.5468, |
| "step": 9800 |
| }, |
| { |
| "epoch": 2.5247715866683826, |
| "grad_norm": 3.0992591381073, |
| "learning_rate": 2.4756756756756758e-05, |
| "loss": 0.6128, |
| "step": 9810 |
| }, |
| { |
| "epoch": 2.527345258010552, |
| "grad_norm": 0.9129619002342224, |
| "learning_rate": 2.4731016731016733e-05, |
| "loss": 0.5671, |
| "step": 9820 |
| }, |
| { |
| "epoch": 2.529918929352722, |
| "grad_norm": 4.400023460388184, |
| "learning_rate": 2.4705276705276705e-05, |
| "loss": 0.5863, |
| "step": 9830 |
| }, |
| { |
| "epoch": 2.532492600694891, |
| "grad_norm": 1.1237571239471436, |
| "learning_rate": 2.467953667953668e-05, |
| "loss": 0.5115, |
| "step": 9840 |
| }, |
| { |
| "epoch": 2.535066272037061, |
| "grad_norm": 1.2072199583053589, |
| "learning_rate": 2.4653796653796655e-05, |
| "loss": 0.5748, |
| "step": 9850 |
| }, |
| { |
| "epoch": 2.5376399433792303, |
| "grad_norm": 0.8158368468284607, |
| "learning_rate": 2.462805662805663e-05, |
| "loss": 0.6466, |
| "step": 9860 |
| }, |
| { |
| "epoch": 2.5402136147214, |
| "grad_norm": 1.3030297756195068, |
| "learning_rate": 2.46023166023166e-05, |
| "loss": 0.5114, |
| "step": 9870 |
| }, |
| { |
| "epoch": 2.54278728606357, |
| "grad_norm": 1.1183769702911377, |
| "learning_rate": 2.4576576576576576e-05, |
| "loss": 0.5518, |
| "step": 9880 |
| }, |
| { |
| "epoch": 2.545360957405739, |
| "grad_norm": 0.9065691828727722, |
| "learning_rate": 2.455083655083655e-05, |
| "loss": 0.522, |
| "step": 9890 |
| }, |
| { |
| "epoch": 2.547934628747909, |
| "grad_norm": 1.4766273498535156, |
| "learning_rate": 2.4525096525096526e-05, |
| "loss": 0.587, |
| "step": 9900 |
| }, |
| { |
| "epoch": 2.5505083000900783, |
| "grad_norm": 4.340095043182373, |
| "learning_rate": 2.44993564993565e-05, |
| "loss": 0.6021, |
| "step": 9910 |
| }, |
| { |
| "epoch": 2.553081971432248, |
| "grad_norm": 0.8867147564888, |
| "learning_rate": 2.4473616473616476e-05, |
| "loss": 0.5076, |
| "step": 9920 |
| }, |
| { |
| "epoch": 2.5556556427744175, |
| "grad_norm": 1.2805358171463013, |
| "learning_rate": 2.444787644787645e-05, |
| "loss": 0.5817, |
| "step": 9930 |
| }, |
| { |
| "epoch": 2.5582293141165873, |
| "grad_norm": 1.0419846773147583, |
| "learning_rate": 2.4422136422136423e-05, |
| "loss": 0.5846, |
| "step": 9940 |
| }, |
| { |
| "epoch": 2.560802985458757, |
| "grad_norm": 1.3988200426101685, |
| "learning_rate": 2.4396396396396398e-05, |
| "loss": 0.5483, |
| "step": 9950 |
| }, |
| { |
| "epoch": 2.5633766568009264, |
| "grad_norm": 1.3084689378738403, |
| "learning_rate": 2.4370656370656373e-05, |
| "loss": 0.5685, |
| "step": 9960 |
| }, |
| { |
| "epoch": 2.565950328143096, |
| "grad_norm": 1.160992980003357, |
| "learning_rate": 2.4344916344916348e-05, |
| "loss": 0.5852, |
| "step": 9970 |
| }, |
| { |
| "epoch": 2.568523999485266, |
| "grad_norm": 1.859298825263977, |
| "learning_rate": 2.431917631917632e-05, |
| "loss": 0.5561, |
| "step": 9980 |
| }, |
| { |
| "epoch": 2.5710976708274353, |
| "grad_norm": 0.9575828313827515, |
| "learning_rate": 2.4293436293436294e-05, |
| "loss": 0.5501, |
| "step": 9990 |
| }, |
| { |
| "epoch": 2.5736713421696047, |
| "grad_norm": 1.3349659442901611, |
| "learning_rate": 2.426769626769627e-05, |
| "loss": 0.5458, |
| "step": 10000 |
| }, |
| { |
| "epoch": 2.5736713421696047, |
| "eval_loss": 0.7813047170639038, |
| "eval_runtime": 395.5545, |
| "eval_samples_per_second": 49.111, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.0003636529278236245, |
| "step": 10000 |
| }, |
| { |
| "epoch": 2.5762450135117745, |
| "grad_norm": 0.8265610933303833, |
| "learning_rate": 2.4241956241956244e-05, |
| "loss": 0.5199, |
| "step": 10010 |
| }, |
| { |
| "epoch": 2.5788186848539443, |
| "grad_norm": 1.1226060390472412, |
| "learning_rate": 2.4216216216216216e-05, |
| "loss": 0.5416, |
| "step": 10020 |
| }, |
| { |
| "epoch": 2.5813923561961136, |
| "grad_norm": 1.006882905960083, |
| "learning_rate": 2.419047619047619e-05, |
| "loss": 0.6422, |
| "step": 10030 |
| }, |
| { |
| "epoch": 2.5839660275382834, |
| "grad_norm": 1.1571662425994873, |
| "learning_rate": 2.4164736164736166e-05, |
| "loss": 0.5548, |
| "step": 10040 |
| }, |
| { |
| "epoch": 2.586539698880453, |
| "grad_norm": 0.9538158774375916, |
| "learning_rate": 2.413899613899614e-05, |
| "loss": 0.5416, |
| "step": 10050 |
| }, |
| { |
| "epoch": 2.5891133702226226, |
| "grad_norm": 3.1586387157440186, |
| "learning_rate": 2.4113256113256113e-05, |
| "loss": 0.5445, |
| "step": 10060 |
| }, |
| { |
| "epoch": 2.591687041564792, |
| "grad_norm": 0.9174941778182983, |
| "learning_rate": 2.4087516087516088e-05, |
| "loss": 0.5539, |
| "step": 10070 |
| }, |
| { |
| "epoch": 2.5942607129069617, |
| "grad_norm": 0.8916310667991638, |
| "learning_rate": 2.4061776061776063e-05, |
| "loss": 0.5487, |
| "step": 10080 |
| }, |
| { |
| "epoch": 2.5968343842491315, |
| "grad_norm": 3.3344202041625977, |
| "learning_rate": 2.4036036036036034e-05, |
| "loss": 0.5935, |
| "step": 10090 |
| }, |
| { |
| "epoch": 2.599408055591301, |
| "grad_norm": 1.8310283422470093, |
| "learning_rate": 2.4010296010296013e-05, |
| "loss": 0.5507, |
| "step": 10100 |
| }, |
| { |
| "epoch": 2.6019817269334706, |
| "grad_norm": 0.95711749792099, |
| "learning_rate": 2.3984555984555988e-05, |
| "loss": 0.5899, |
| "step": 10110 |
| }, |
| { |
| "epoch": 2.6045553982756404, |
| "grad_norm": 1.1293368339538574, |
| "learning_rate": 2.3958815958815963e-05, |
| "loss": 0.5468, |
| "step": 10120 |
| }, |
| { |
| "epoch": 2.6071290696178098, |
| "grad_norm": 0.8633250594139099, |
| "learning_rate": 2.3933075933075934e-05, |
| "loss": 0.5763, |
| "step": 10130 |
| }, |
| { |
| "epoch": 2.6097027409599796, |
| "grad_norm": 1.343990445137024, |
| "learning_rate": 2.390733590733591e-05, |
| "loss": 0.5839, |
| "step": 10140 |
| }, |
| { |
| "epoch": 2.612276412302149, |
| "grad_norm": 1.5311845541000366, |
| "learning_rate": 2.3881595881595884e-05, |
| "loss": 0.531, |
| "step": 10150 |
| }, |
| { |
| "epoch": 2.6148500836443187, |
| "grad_norm": 1.0829716920852661, |
| "learning_rate": 2.385585585585586e-05, |
| "loss": 0.5402, |
| "step": 10160 |
| }, |
| { |
| "epoch": 2.617423754986488, |
| "grad_norm": 1.5562556982040405, |
| "learning_rate": 2.383011583011583e-05, |
| "loss": 0.5652, |
| "step": 10170 |
| }, |
| { |
| "epoch": 2.619997426328658, |
| "grad_norm": 1.0857690572738647, |
| "learning_rate": 2.3804375804375806e-05, |
| "loss": 0.6161, |
| "step": 10180 |
| }, |
| { |
| "epoch": 2.6225710976708276, |
| "grad_norm": 1.1553452014923096, |
| "learning_rate": 2.377863577863578e-05, |
| "loss": 0.5975, |
| "step": 10190 |
| }, |
| { |
| "epoch": 2.625144769012997, |
| "grad_norm": 0.9992988109588623, |
| "learning_rate": 2.3752895752895752e-05, |
| "loss": 0.5837, |
| "step": 10200 |
| }, |
| { |
| "epoch": 2.6277184403551668, |
| "grad_norm": 1.0627033710479736, |
| "learning_rate": 2.3727155727155727e-05, |
| "loss": 0.5527, |
| "step": 10210 |
| }, |
| { |
| "epoch": 2.630292111697336, |
| "grad_norm": 0.8239421248435974, |
| "learning_rate": 2.3701415701415702e-05, |
| "loss": 0.5822, |
| "step": 10220 |
| }, |
| { |
| "epoch": 2.632865783039506, |
| "grad_norm": 2.0370047092437744, |
| "learning_rate": 2.3675675675675677e-05, |
| "loss": 0.55, |
| "step": 10230 |
| }, |
| { |
| "epoch": 2.6354394543816753, |
| "grad_norm": 1.0203136205673218, |
| "learning_rate": 2.364993564993565e-05, |
| "loss": 0.5289, |
| "step": 10240 |
| }, |
| { |
| "epoch": 2.638013125723845, |
| "grad_norm": 1.1402500867843628, |
| "learning_rate": 2.3624195624195624e-05, |
| "loss": 0.5492, |
| "step": 10250 |
| }, |
| { |
| "epoch": 2.640586797066015, |
| "grad_norm": 1.1327241659164429, |
| "learning_rate": 2.35984555984556e-05, |
| "loss": 0.5416, |
| "step": 10260 |
| }, |
| { |
| "epoch": 2.643160468408184, |
| "grad_norm": 0.9209975004196167, |
| "learning_rate": 2.3572715572715574e-05, |
| "loss": 0.5442, |
| "step": 10270 |
| }, |
| { |
| "epoch": 2.645734139750354, |
| "grad_norm": 1.3790748119354248, |
| "learning_rate": 2.3546975546975546e-05, |
| "loss": 0.6103, |
| "step": 10280 |
| }, |
| { |
| "epoch": 2.6483078110925233, |
| "grad_norm": 1.2023276090621948, |
| "learning_rate": 2.352123552123552e-05, |
| "loss": 0.5886, |
| "step": 10290 |
| }, |
| { |
| "epoch": 2.650881482434693, |
| "grad_norm": 1.0166372060775757, |
| "learning_rate": 2.3495495495495496e-05, |
| "loss": 0.5741, |
| "step": 10300 |
| }, |
| { |
| "epoch": 2.6534551537768625, |
| "grad_norm": 1.1085914373397827, |
| "learning_rate": 2.346975546975547e-05, |
| "loss": 0.5302, |
| "step": 10310 |
| }, |
| { |
| "epoch": 2.6560288251190323, |
| "grad_norm": 1.1305118799209595, |
| "learning_rate": 2.3444015444015446e-05, |
| "loss": 0.5851, |
| "step": 10320 |
| }, |
| { |
| "epoch": 2.658602496461202, |
| "grad_norm": 4.382242202758789, |
| "learning_rate": 2.341827541827542e-05, |
| "loss": 0.6207, |
| "step": 10330 |
| }, |
| { |
| "epoch": 2.6611761678033714, |
| "grad_norm": 1.3495676517486572, |
| "learning_rate": 2.3392535392535396e-05, |
| "loss": 0.531, |
| "step": 10340 |
| }, |
| { |
| "epoch": 2.663749839145541, |
| "grad_norm": 1.2372089624404907, |
| "learning_rate": 2.3366795366795367e-05, |
| "loss": 0.5543, |
| "step": 10350 |
| }, |
| { |
| "epoch": 2.666323510487711, |
| "grad_norm": 1.2781168222427368, |
| "learning_rate": 2.3341055341055342e-05, |
| "loss": 0.5379, |
| "step": 10360 |
| }, |
| { |
| "epoch": 2.6688971818298803, |
| "grad_norm": 0.9602215886116028, |
| "learning_rate": 2.3315315315315317e-05, |
| "loss": 0.5677, |
| "step": 10370 |
| }, |
| { |
| "epoch": 2.6714708531720497, |
| "grad_norm": 1.4778882265090942, |
| "learning_rate": 2.3289575289575292e-05, |
| "loss": 0.5196, |
| "step": 10380 |
| }, |
| { |
| "epoch": 2.6740445245142195, |
| "grad_norm": 1.1406737565994263, |
| "learning_rate": 2.3263835263835264e-05, |
| "loss": 0.5886, |
| "step": 10390 |
| }, |
| { |
| "epoch": 2.6766181958563893, |
| "grad_norm": 1.3045039176940918, |
| "learning_rate": 2.323809523809524e-05, |
| "loss": 0.5468, |
| "step": 10400 |
| }, |
| { |
| "epoch": 2.6791918671985586, |
| "grad_norm": 1.1120688915252686, |
| "learning_rate": 2.3212355212355214e-05, |
| "loss": 0.6246, |
| "step": 10410 |
| }, |
| { |
| "epoch": 2.6817655385407284, |
| "grad_norm": 3.4695723056793213, |
| "learning_rate": 2.318661518661519e-05, |
| "loss": 0.5557, |
| "step": 10420 |
| }, |
| { |
| "epoch": 2.684339209882898, |
| "grad_norm": 1.4263606071472168, |
| "learning_rate": 2.316087516087516e-05, |
| "loss": 0.5656, |
| "step": 10430 |
| }, |
| { |
| "epoch": 2.6869128812250676, |
| "grad_norm": 0.7746553421020508, |
| "learning_rate": 2.3135135135135136e-05, |
| "loss": 0.5458, |
| "step": 10440 |
| }, |
| { |
| "epoch": 2.689486552567237, |
| "grad_norm": 1.4084376096725464, |
| "learning_rate": 2.310939510939511e-05, |
| "loss": 0.5506, |
| "step": 10450 |
| }, |
| { |
| "epoch": 2.6920602239094067, |
| "grad_norm": 0.9711858034133911, |
| "learning_rate": 2.3083655083655086e-05, |
| "loss": 0.5219, |
| "step": 10460 |
| }, |
| { |
| "epoch": 2.6946338952515765, |
| "grad_norm": 1.1122053861618042, |
| "learning_rate": 2.3057915057915057e-05, |
| "loss": 0.5287, |
| "step": 10470 |
| }, |
| { |
| "epoch": 2.697207566593746, |
| "grad_norm": 1.0087629556655884, |
| "learning_rate": 2.3032175032175032e-05, |
| "loss": 0.6167, |
| "step": 10480 |
| }, |
| { |
| "epoch": 2.6997812379359156, |
| "grad_norm": 1.0684638023376465, |
| "learning_rate": 2.3009009009009013e-05, |
| "loss": 0.5191, |
| "step": 10490 |
| }, |
| { |
| "epoch": 2.7023549092780854, |
| "grad_norm": 1.1910923719406128, |
| "learning_rate": 2.2983268983268984e-05, |
| "loss": 0.551, |
| "step": 10500 |
| }, |
| { |
| "epoch": 2.7049285806202548, |
| "grad_norm": 1.287802815437317, |
| "learning_rate": 2.295752895752896e-05, |
| "loss": 0.5438, |
| "step": 10510 |
| }, |
| { |
| "epoch": 2.7075022519624246, |
| "grad_norm": 0.9688395857810974, |
| "learning_rate": 2.2931788931788934e-05, |
| "loss": 0.5567, |
| "step": 10520 |
| }, |
| { |
| "epoch": 2.710075923304594, |
| "grad_norm": 1.0128886699676514, |
| "learning_rate": 2.2906048906048906e-05, |
| "loss": 0.6127, |
| "step": 10530 |
| }, |
| { |
| "epoch": 2.7126495946467637, |
| "grad_norm": 1.974613070487976, |
| "learning_rate": 2.288030888030888e-05, |
| "loss": 0.5642, |
| "step": 10540 |
| }, |
| { |
| "epoch": 2.715223265988933, |
| "grad_norm": 2.323258876800537, |
| "learning_rate": 2.2854568854568856e-05, |
| "loss": 0.5192, |
| "step": 10550 |
| }, |
| { |
| "epoch": 2.717796937331103, |
| "grad_norm": 2.6239235401153564, |
| "learning_rate": 2.282882882882883e-05, |
| "loss": 0.5786, |
| "step": 10560 |
| }, |
| { |
| "epoch": 2.7203706086732726, |
| "grad_norm": 2.535557508468628, |
| "learning_rate": 2.2803088803088802e-05, |
| "loss": 0.5732, |
| "step": 10570 |
| }, |
| { |
| "epoch": 2.722944280015442, |
| "grad_norm": 1.4049146175384521, |
| "learning_rate": 2.2777348777348777e-05, |
| "loss": 0.5613, |
| "step": 10580 |
| }, |
| { |
| "epoch": 2.7255179513576118, |
| "grad_norm": 0.9087219834327698, |
| "learning_rate": 2.2751608751608752e-05, |
| "loss": 0.5429, |
| "step": 10590 |
| }, |
| { |
| "epoch": 2.728091622699781, |
| "grad_norm": 1.2348556518554688, |
| "learning_rate": 2.2725868725868727e-05, |
| "loss": 0.6106, |
| "step": 10600 |
| }, |
| { |
| "epoch": 2.730665294041951, |
| "grad_norm": 0.954634964466095, |
| "learning_rate": 2.27001287001287e-05, |
| "loss": 0.5648, |
| "step": 10610 |
| }, |
| { |
| "epoch": 2.7332389653841203, |
| "grad_norm": 0.9417296051979065, |
| "learning_rate": 2.2674388674388674e-05, |
| "loss": 0.5641, |
| "step": 10620 |
| }, |
| { |
| "epoch": 2.73581263672629, |
| "grad_norm": 3.07540225982666, |
| "learning_rate": 2.264864864864865e-05, |
| "loss": 0.6197, |
| "step": 10630 |
| }, |
| { |
| "epoch": 2.73838630806846, |
| "grad_norm": 1.1706743240356445, |
| "learning_rate": 2.2622908622908624e-05, |
| "loss": 0.5237, |
| "step": 10640 |
| }, |
| { |
| "epoch": 2.740959979410629, |
| "grad_norm": 1.0390594005584717, |
| "learning_rate": 2.2597168597168596e-05, |
| "loss": 0.5377, |
| "step": 10650 |
| }, |
| { |
| "epoch": 2.743533650752799, |
| "grad_norm": 4.094557762145996, |
| "learning_rate": 2.257142857142857e-05, |
| "loss": 0.5717, |
| "step": 10660 |
| }, |
| { |
| "epoch": 2.7461073220949683, |
| "grad_norm": 3.71659255027771, |
| "learning_rate": 2.254568854568855e-05, |
| "loss": 0.5725, |
| "step": 10670 |
| }, |
| { |
| "epoch": 2.748680993437138, |
| "grad_norm": 1.1988261938095093, |
| "learning_rate": 2.251994851994852e-05, |
| "loss": 0.565, |
| "step": 10680 |
| }, |
| { |
| "epoch": 2.7512546647793075, |
| "grad_norm": 1.5038524866104126, |
| "learning_rate": 2.2494208494208496e-05, |
| "loss": 0.5142, |
| "step": 10690 |
| }, |
| { |
| "epoch": 2.7538283361214773, |
| "grad_norm": 1.0578770637512207, |
| "learning_rate": 2.246846846846847e-05, |
| "loss": 0.5922, |
| "step": 10700 |
| }, |
| { |
| "epoch": 2.756402007463647, |
| "grad_norm": 1.151419997215271, |
| "learning_rate": 2.2442728442728446e-05, |
| "loss": 0.6072, |
| "step": 10710 |
| }, |
| { |
| "epoch": 2.7589756788058164, |
| "grad_norm": 1.1115498542785645, |
| "learning_rate": 2.2416988416988417e-05, |
| "loss": 0.5549, |
| "step": 10720 |
| }, |
| { |
| "epoch": 2.761549350147986, |
| "grad_norm": 2.30755352973938, |
| "learning_rate": 2.2391248391248392e-05, |
| "loss": 0.5863, |
| "step": 10730 |
| }, |
| { |
| "epoch": 2.7641230214901555, |
| "grad_norm": 2.3531439304351807, |
| "learning_rate": 2.2365508365508367e-05, |
| "loss": 0.5236, |
| "step": 10740 |
| }, |
| { |
| "epoch": 2.7666966928323253, |
| "grad_norm": 1.0621081590652466, |
| "learning_rate": 2.2339768339768342e-05, |
| "loss": 0.5689, |
| "step": 10750 |
| }, |
| { |
| "epoch": 2.7692703641744947, |
| "grad_norm": 3.393990993499756, |
| "learning_rate": 2.2314028314028314e-05, |
| "loss": 0.5736, |
| "step": 10760 |
| }, |
| { |
| "epoch": 2.7718440355166645, |
| "grad_norm": 0.9856056571006775, |
| "learning_rate": 2.228828828828829e-05, |
| "loss": 0.5719, |
| "step": 10770 |
| }, |
| { |
| "epoch": 2.7744177068588343, |
| "grad_norm": 1.3142099380493164, |
| "learning_rate": 2.2262548262548264e-05, |
| "loss": 0.5546, |
| "step": 10780 |
| }, |
| { |
| "epoch": 2.7769913782010036, |
| "grad_norm": 0.969963550567627, |
| "learning_rate": 2.223680823680824e-05, |
| "loss": 0.5425, |
| "step": 10790 |
| }, |
| { |
| "epoch": 2.7795650495431734, |
| "grad_norm": 2.802344560623169, |
| "learning_rate": 2.221106821106821e-05, |
| "loss": 0.5235, |
| "step": 10800 |
| }, |
| { |
| "epoch": 2.782138720885343, |
| "grad_norm": 1.5331348180770874, |
| "learning_rate": 2.2185328185328186e-05, |
| "loss": 0.5436, |
| "step": 10810 |
| }, |
| { |
| "epoch": 2.7847123922275125, |
| "grad_norm": 0.9570822715759277, |
| "learning_rate": 2.215958815958816e-05, |
| "loss": 0.5615, |
| "step": 10820 |
| }, |
| { |
| "epoch": 2.787286063569682, |
| "grad_norm": 3.3715949058532715, |
| "learning_rate": 2.2133848133848136e-05, |
| "loss": 0.5473, |
| "step": 10830 |
| }, |
| { |
| "epoch": 2.7898597349118517, |
| "grad_norm": 1.2080599069595337, |
| "learning_rate": 2.2108108108108107e-05, |
| "loss": 0.6758, |
| "step": 10840 |
| }, |
| { |
| "epoch": 2.7924334062540215, |
| "grad_norm": 0.9372673034667969, |
| "learning_rate": 2.2082368082368082e-05, |
| "loss": 0.5913, |
| "step": 10850 |
| }, |
| { |
| "epoch": 2.795007077596191, |
| "grad_norm": 1.019864559173584, |
| "learning_rate": 2.2056628056628057e-05, |
| "loss": 0.5194, |
| "step": 10860 |
| }, |
| { |
| "epoch": 2.7975807489383606, |
| "grad_norm": 1.1062774658203125, |
| "learning_rate": 2.2030888030888032e-05, |
| "loss": 0.5451, |
| "step": 10870 |
| }, |
| { |
| "epoch": 2.8001544202805304, |
| "grad_norm": 0.8986897468566895, |
| "learning_rate": 2.2005148005148007e-05, |
| "loss": 0.5787, |
| "step": 10880 |
| }, |
| { |
| "epoch": 2.8027280916226998, |
| "grad_norm": 0.9067063331604004, |
| "learning_rate": 2.1979407979407982e-05, |
| "loss": 0.5901, |
| "step": 10890 |
| }, |
| { |
| "epoch": 2.8053017629648695, |
| "grad_norm": 1.141993522644043, |
| "learning_rate": 2.1953667953667957e-05, |
| "loss": 0.5681, |
| "step": 10900 |
| }, |
| { |
| "epoch": 2.807875434307039, |
| "grad_norm": 1.1285045146942139, |
| "learning_rate": 2.192792792792793e-05, |
| "loss": 0.5557, |
| "step": 10910 |
| }, |
| { |
| "epoch": 2.8104491056492087, |
| "grad_norm": 0.9529311656951904, |
| "learning_rate": 2.1902187902187904e-05, |
| "loss": 0.6106, |
| "step": 10920 |
| }, |
| { |
| "epoch": 2.813022776991378, |
| "grad_norm": 0.9633564949035645, |
| "learning_rate": 2.187644787644788e-05, |
| "loss": 0.521, |
| "step": 10930 |
| }, |
| { |
| "epoch": 2.815596448333548, |
| "grad_norm": 1.3736627101898193, |
| "learning_rate": 2.1850707850707854e-05, |
| "loss": 0.4961, |
| "step": 10940 |
| }, |
| { |
| "epoch": 2.8181701196757176, |
| "grad_norm": 1.1988080739974976, |
| "learning_rate": 2.1824967824967825e-05, |
| "loss": 0.5339, |
| "step": 10950 |
| }, |
| { |
| "epoch": 2.820743791017887, |
| "grad_norm": 3.6586461067199707, |
| "learning_rate": 2.17992277992278e-05, |
| "loss": 0.5535, |
| "step": 10960 |
| }, |
| { |
| "epoch": 2.8233174623600568, |
| "grad_norm": 1.6406865119934082, |
| "learning_rate": 2.1773487773487775e-05, |
| "loss": 0.5244, |
| "step": 10970 |
| }, |
| { |
| "epoch": 2.825891133702226, |
| "grad_norm": 1.2448114156723022, |
| "learning_rate": 2.174774774774775e-05, |
| "loss": 0.6124, |
| "step": 10980 |
| }, |
| { |
| "epoch": 2.828464805044396, |
| "grad_norm": 0.9966627955436707, |
| "learning_rate": 2.1722007722007722e-05, |
| "loss": 0.5213, |
| "step": 10990 |
| }, |
| { |
| "epoch": 2.8310384763865653, |
| "grad_norm": 1.5324152708053589, |
| "learning_rate": 2.1696267696267697e-05, |
| "loss": 0.5412, |
| "step": 11000 |
| }, |
| { |
| "epoch": 2.8310384763865653, |
| "eval_loss": 0.7792600989341736, |
| "eval_runtime": 395.6052, |
| "eval_samples_per_second": 49.105, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.00037447141262839833, |
| "step": 11000 |
| }, |
| { |
| "epoch": 2.833612147728735, |
| "grad_norm": 3.205904483795166, |
| "learning_rate": 2.1670527670527672e-05, |
| "loss": 0.4943, |
| "step": 11010 |
| }, |
| { |
| "epoch": 2.836185819070905, |
| "grad_norm": 1.0843708515167236, |
| "learning_rate": 2.1644787644787644e-05, |
| "loss": 0.5571, |
| "step": 11020 |
| }, |
| { |
| "epoch": 2.838759490413074, |
| "grad_norm": 0.9089128971099854, |
| "learning_rate": 2.161904761904762e-05, |
| "loss": 0.5604, |
| "step": 11030 |
| }, |
| { |
| "epoch": 2.841333161755244, |
| "grad_norm": 1.2132848501205444, |
| "learning_rate": 2.1593307593307594e-05, |
| "loss": 0.5426, |
| "step": 11040 |
| }, |
| { |
| "epoch": 2.8439068330974133, |
| "grad_norm": 1.2147984504699707, |
| "learning_rate": 2.156756756756757e-05, |
| "loss": 0.5833, |
| "step": 11050 |
| }, |
| { |
| "epoch": 2.846480504439583, |
| "grad_norm": 0.9623033404350281, |
| "learning_rate": 2.154182754182754e-05, |
| "loss": 0.5253, |
| "step": 11060 |
| }, |
| { |
| "epoch": 2.8490541757817525, |
| "grad_norm": 1.1936800479888916, |
| "learning_rate": 2.1516087516087515e-05, |
| "loss": 0.548, |
| "step": 11070 |
| }, |
| { |
| "epoch": 2.8516278471239223, |
| "grad_norm": 3.678395986557007, |
| "learning_rate": 2.1490347490347494e-05, |
| "loss": 0.5867, |
| "step": 11080 |
| }, |
| { |
| "epoch": 2.854201518466092, |
| "grad_norm": 3.4173741340637207, |
| "learning_rate": 2.1464607464607465e-05, |
| "loss": 0.4922, |
| "step": 11090 |
| }, |
| { |
| "epoch": 2.8567751898082614, |
| "grad_norm": 0.8987242579460144, |
| "learning_rate": 2.143886743886744e-05, |
| "loss": 0.5296, |
| "step": 11100 |
| }, |
| { |
| "epoch": 2.859348861150431, |
| "grad_norm": 1.1060086488723755, |
| "learning_rate": 2.1413127413127415e-05, |
| "loss": 0.52, |
| "step": 11110 |
| }, |
| { |
| "epoch": 2.8619225324926005, |
| "grad_norm": 0.9053888320922852, |
| "learning_rate": 2.138738738738739e-05, |
| "loss": 0.5979, |
| "step": 11120 |
| }, |
| { |
| "epoch": 2.8644962038347703, |
| "grad_norm": 0.9858607053756714, |
| "learning_rate": 2.1361647361647362e-05, |
| "loss": 0.5683, |
| "step": 11130 |
| }, |
| { |
| "epoch": 2.8670698751769397, |
| "grad_norm": 2.134103298187256, |
| "learning_rate": 2.1335907335907337e-05, |
| "loss": 0.5358, |
| "step": 11140 |
| }, |
| { |
| "epoch": 2.8696435465191095, |
| "grad_norm": 4.8553009033203125, |
| "learning_rate": 2.1310167310167312e-05, |
| "loss": 0.5975, |
| "step": 11150 |
| }, |
| { |
| "epoch": 2.8722172178612793, |
| "grad_norm": 1.1935124397277832, |
| "learning_rate": 2.1284427284427287e-05, |
| "loss": 0.6338, |
| "step": 11160 |
| }, |
| { |
| "epoch": 2.8747908892034486, |
| "grad_norm": 1.042869210243225, |
| "learning_rate": 2.125868725868726e-05, |
| "loss": 0.5529, |
| "step": 11170 |
| }, |
| { |
| "epoch": 2.8773645605456184, |
| "grad_norm": 0.7498083114624023, |
| "learning_rate": 2.1232947232947233e-05, |
| "loss": 0.6005, |
| "step": 11180 |
| }, |
| { |
| "epoch": 2.879938231887788, |
| "grad_norm": 1.3521268367767334, |
| "learning_rate": 2.120720720720721e-05, |
| "loss": 0.5645, |
| "step": 11190 |
| }, |
| { |
| "epoch": 2.8825119032299575, |
| "grad_norm": 1.2290796041488647, |
| "learning_rate": 2.1181467181467183e-05, |
| "loss": 0.561, |
| "step": 11200 |
| }, |
| { |
| "epoch": 2.885085574572127, |
| "grad_norm": 1.114356517791748, |
| "learning_rate": 2.1155727155727155e-05, |
| "loss": 0.5426, |
| "step": 11210 |
| }, |
| { |
| "epoch": 2.8876592459142967, |
| "grad_norm": 0.7884505391120911, |
| "learning_rate": 2.112998712998713e-05, |
| "loss": 0.5519, |
| "step": 11220 |
| }, |
| { |
| "epoch": 2.8902329172564665, |
| "grad_norm": 1.0232847929000854, |
| "learning_rate": 2.1104247104247105e-05, |
| "loss": 0.5406, |
| "step": 11230 |
| }, |
| { |
| "epoch": 2.892806588598636, |
| "grad_norm": 1.2134202718734741, |
| "learning_rate": 2.107850707850708e-05, |
| "loss": 0.537, |
| "step": 11240 |
| }, |
| { |
| "epoch": 2.8953802599408056, |
| "grad_norm": 1.1819850206375122, |
| "learning_rate": 2.105276705276705e-05, |
| "loss": 0.4655, |
| "step": 11250 |
| }, |
| { |
| "epoch": 2.8979539312829754, |
| "grad_norm": 0.9227721095085144, |
| "learning_rate": 2.1027027027027027e-05, |
| "loss": 0.5465, |
| "step": 11260 |
| }, |
| { |
| "epoch": 2.9005276026251448, |
| "grad_norm": 4.2423858642578125, |
| "learning_rate": 2.1001287001287e-05, |
| "loss": 0.5245, |
| "step": 11270 |
| }, |
| { |
| "epoch": 2.903101273967314, |
| "grad_norm": 1.0948383808135986, |
| "learning_rate": 2.0975546975546977e-05, |
| "loss": 0.5712, |
| "step": 11280 |
| }, |
| { |
| "epoch": 2.905674945309484, |
| "grad_norm": 0.783808171749115, |
| "learning_rate": 2.094980694980695e-05, |
| "loss": 0.5705, |
| "step": 11290 |
| }, |
| { |
| "epoch": 2.9082486166516537, |
| "grad_norm": 0.858773410320282, |
| "learning_rate": 2.0924066924066927e-05, |
| "loss": 0.5632, |
| "step": 11300 |
| }, |
| { |
| "epoch": 2.910822287993823, |
| "grad_norm": 4.24513578414917, |
| "learning_rate": 2.08983268983269e-05, |
| "loss": 0.5613, |
| "step": 11310 |
| }, |
| { |
| "epoch": 2.913395959335993, |
| "grad_norm": 1.336777925491333, |
| "learning_rate": 2.0872586872586873e-05, |
| "loss": 0.6024, |
| "step": 11320 |
| }, |
| { |
| "epoch": 2.9159696306781626, |
| "grad_norm": 0.7360126972198486, |
| "learning_rate": 2.0846846846846848e-05, |
| "loss": 0.5051, |
| "step": 11330 |
| }, |
| { |
| "epoch": 2.918543302020332, |
| "grad_norm": 0.9983276128768921, |
| "learning_rate": 2.0821106821106823e-05, |
| "loss": 0.5366, |
| "step": 11340 |
| }, |
| { |
| "epoch": 2.9211169733625018, |
| "grad_norm": 1.2790330648422241, |
| "learning_rate": 2.0795366795366798e-05, |
| "loss": 0.4972, |
| "step": 11350 |
| }, |
| { |
| "epoch": 2.923690644704671, |
| "grad_norm": 1.0925854444503784, |
| "learning_rate": 2.076962676962677e-05, |
| "loss": 0.4877, |
| "step": 11360 |
| }, |
| { |
| "epoch": 2.926264316046841, |
| "grad_norm": 1.262011170387268, |
| "learning_rate": 2.0743886743886745e-05, |
| "loss": 0.5177, |
| "step": 11370 |
| }, |
| { |
| "epoch": 2.9288379873890102, |
| "grad_norm": 0.7657809853553772, |
| "learning_rate": 2.071814671814672e-05, |
| "loss": 0.5845, |
| "step": 11380 |
| }, |
| { |
| "epoch": 2.93141165873118, |
| "grad_norm": 1.21810781955719, |
| "learning_rate": 2.0692406692406695e-05, |
| "loss": 0.4873, |
| "step": 11390 |
| }, |
| { |
| "epoch": 2.93398533007335, |
| "grad_norm": 1.1129614114761353, |
| "learning_rate": 2.0666666666666666e-05, |
| "loss": 0.5668, |
| "step": 11400 |
| }, |
| { |
| "epoch": 2.936559001415519, |
| "grad_norm": 1.1121175289154053, |
| "learning_rate": 2.064092664092664e-05, |
| "loss": 0.4534, |
| "step": 11410 |
| }, |
| { |
| "epoch": 2.939132672757689, |
| "grad_norm": 1.254805564880371, |
| "learning_rate": 2.0615186615186616e-05, |
| "loss": 0.5647, |
| "step": 11420 |
| }, |
| { |
| "epoch": 2.9417063440998583, |
| "grad_norm": 1.1753339767456055, |
| "learning_rate": 2.058944658944659e-05, |
| "loss": 0.5619, |
| "step": 11430 |
| }, |
| { |
| "epoch": 2.944280015442028, |
| "grad_norm": 1.9952584505081177, |
| "learning_rate": 2.0563706563706563e-05, |
| "loss": 0.5437, |
| "step": 11440 |
| }, |
| { |
| "epoch": 2.9468536867841975, |
| "grad_norm": 2.9970035552978516, |
| "learning_rate": 2.0537966537966538e-05, |
| "loss": 0.6106, |
| "step": 11450 |
| }, |
| { |
| "epoch": 2.9494273581263672, |
| "grad_norm": 1.2098712921142578, |
| "learning_rate": 2.0512226512226513e-05, |
| "loss": 0.5032, |
| "step": 11460 |
| }, |
| { |
| "epoch": 2.952001029468537, |
| "grad_norm": 2.842196464538574, |
| "learning_rate": 2.0486486486486488e-05, |
| "loss": 0.4807, |
| "step": 11470 |
| }, |
| { |
| "epoch": 2.9545747008107064, |
| "grad_norm": 0.9120272397994995, |
| "learning_rate": 2.046074646074646e-05, |
| "loss": 0.5058, |
| "step": 11480 |
| }, |
| { |
| "epoch": 2.957148372152876, |
| "grad_norm": 2.2939398288726807, |
| "learning_rate": 2.0435006435006435e-05, |
| "loss": 0.5666, |
| "step": 11490 |
| }, |
| { |
| "epoch": 2.9597220434950455, |
| "grad_norm": 1.4755898714065552, |
| "learning_rate": 2.0409266409266413e-05, |
| "loss": 0.4845, |
| "step": 11500 |
| }, |
| { |
| "epoch": 2.9622957148372153, |
| "grad_norm": 1.120060920715332, |
| "learning_rate": 2.0383526383526385e-05, |
| "loss": 0.4979, |
| "step": 11510 |
| }, |
| { |
| "epoch": 2.9648693861793847, |
| "grad_norm": 4.3705058097839355, |
| "learning_rate": 2.035778635778636e-05, |
| "loss": 0.6247, |
| "step": 11520 |
| }, |
| { |
| "epoch": 2.9674430575215545, |
| "grad_norm": 1.2147754430770874, |
| "learning_rate": 2.0332046332046335e-05, |
| "loss": 0.512, |
| "step": 11530 |
| }, |
| { |
| "epoch": 2.9700167288637243, |
| "grad_norm": 4.265937328338623, |
| "learning_rate": 2.030630630630631e-05, |
| "loss": 0.533, |
| "step": 11540 |
| }, |
| { |
| "epoch": 2.9725904002058936, |
| "grad_norm": 1.1540546417236328, |
| "learning_rate": 2.028056628056628e-05, |
| "loss": 0.5189, |
| "step": 11550 |
| }, |
| { |
| "epoch": 2.9751640715480634, |
| "grad_norm": 1.3933578729629517, |
| "learning_rate": 2.0254826254826256e-05, |
| "loss": 0.6381, |
| "step": 11560 |
| }, |
| { |
| "epoch": 2.977737742890233, |
| "grad_norm": 1.1027209758758545, |
| "learning_rate": 2.022908622908623e-05, |
| "loss": 0.5235, |
| "step": 11570 |
| }, |
| { |
| "epoch": 2.9803114142324025, |
| "grad_norm": 0.9615407586097717, |
| "learning_rate": 2.0203346203346203e-05, |
| "loss": 0.6184, |
| "step": 11580 |
| }, |
| { |
| "epoch": 2.982885085574572, |
| "grad_norm": 1.2538810968399048, |
| "learning_rate": 2.0177606177606178e-05, |
| "loss": 0.576, |
| "step": 11590 |
| }, |
| { |
| "epoch": 2.9854587569167417, |
| "grad_norm": 1.3477590084075928, |
| "learning_rate": 2.0151866151866153e-05, |
| "loss": 0.5825, |
| "step": 11600 |
| }, |
| { |
| "epoch": 2.9880324282589115, |
| "grad_norm": 4.23989725112915, |
| "learning_rate": 2.0126126126126128e-05, |
| "loss": 0.5581, |
| "step": 11610 |
| }, |
| { |
| "epoch": 2.990606099601081, |
| "grad_norm": 1.2521101236343384, |
| "learning_rate": 2.01003861003861e-05, |
| "loss": 0.6155, |
| "step": 11620 |
| }, |
| { |
| "epoch": 2.9931797709432506, |
| "grad_norm": 1.0629501342773438, |
| "learning_rate": 2.0074646074646074e-05, |
| "loss": 0.5349, |
| "step": 11630 |
| }, |
| { |
| "epoch": 2.9957534422854204, |
| "grad_norm": 1.3439209461212158, |
| "learning_rate": 2.004890604890605e-05, |
| "loss": 0.5551, |
| "step": 11640 |
| }, |
| { |
| "epoch": 2.9983271136275897, |
| "grad_norm": 0.9323076605796814, |
| "learning_rate": 2.0023166023166024e-05, |
| "loss": 0.5558, |
| "step": 11650 |
| }, |
| { |
| "epoch": 3.0009007849697595, |
| "grad_norm": 1.5501611232757568, |
| "learning_rate": 1.9997425997425996e-05, |
| "loss": 0.5076, |
| "step": 11660 |
| }, |
| { |
| "epoch": 3.003474456311929, |
| "grad_norm": 1.0962281227111816, |
| "learning_rate": 1.997168597168597e-05, |
| "loss": 0.4875, |
| "step": 11670 |
| }, |
| { |
| "epoch": 3.0060481276540987, |
| "grad_norm": 1.1923311948776245, |
| "learning_rate": 1.9945945945945946e-05, |
| "loss": 0.4643, |
| "step": 11680 |
| }, |
| { |
| "epoch": 3.008621798996268, |
| "grad_norm": 1.4216543436050415, |
| "learning_rate": 1.992020592020592e-05, |
| "loss": 0.4181, |
| "step": 11690 |
| }, |
| { |
| "epoch": 3.011195470338438, |
| "grad_norm": 1.0901702642440796, |
| "learning_rate": 1.9894465894465893e-05, |
| "loss": 0.4309, |
| "step": 11700 |
| }, |
| { |
| "epoch": 3.013769141680607, |
| "grad_norm": 0.9830016493797302, |
| "learning_rate": 1.986872586872587e-05, |
| "loss": 0.4027, |
| "step": 11710 |
| }, |
| { |
| "epoch": 3.016342813022777, |
| "grad_norm": 0.9537155032157898, |
| "learning_rate": 1.9842985842985846e-05, |
| "loss": 0.4108, |
| "step": 11720 |
| }, |
| { |
| "epoch": 3.0189164843649468, |
| "grad_norm": 1.4666250944137573, |
| "learning_rate": 1.9817245817245818e-05, |
| "loss": 0.4573, |
| "step": 11730 |
| }, |
| { |
| "epoch": 3.021490155707116, |
| "grad_norm": 1.1737228631973267, |
| "learning_rate": 1.9791505791505793e-05, |
| "loss": 0.4616, |
| "step": 11740 |
| }, |
| { |
| "epoch": 3.024063827049286, |
| "grad_norm": 0.9376081824302673, |
| "learning_rate": 1.9765765765765768e-05, |
| "loss": 0.4141, |
| "step": 11750 |
| }, |
| { |
| "epoch": 3.0266374983914552, |
| "grad_norm": 3.225501775741577, |
| "learning_rate": 1.9740025740025743e-05, |
| "loss": 0.4542, |
| "step": 11760 |
| }, |
| { |
| "epoch": 3.029211169733625, |
| "grad_norm": 0.9168078303337097, |
| "learning_rate": 1.9714285714285714e-05, |
| "loss": 0.4518, |
| "step": 11770 |
| }, |
| { |
| "epoch": 3.031784841075795, |
| "grad_norm": 1.606330394744873, |
| "learning_rate": 1.968854568854569e-05, |
| "loss": 0.3725, |
| "step": 11780 |
| }, |
| { |
| "epoch": 3.034358512417964, |
| "grad_norm": 3.959381341934204, |
| "learning_rate": 1.9662805662805664e-05, |
| "loss": 0.4397, |
| "step": 11790 |
| }, |
| { |
| "epoch": 3.036932183760134, |
| "grad_norm": 0.9928016066551208, |
| "learning_rate": 1.963706563706564e-05, |
| "loss": 0.4031, |
| "step": 11800 |
| }, |
| { |
| "epoch": 3.0395058551023033, |
| "grad_norm": 1.3497694730758667, |
| "learning_rate": 1.961132561132561e-05, |
| "loss": 0.4435, |
| "step": 11810 |
| }, |
| { |
| "epoch": 3.042079526444473, |
| "grad_norm": 1.0364716053009033, |
| "learning_rate": 1.9585585585585586e-05, |
| "loss": 0.4138, |
| "step": 11820 |
| }, |
| { |
| "epoch": 3.0446531977866425, |
| "grad_norm": 1.0680062770843506, |
| "learning_rate": 1.955984555984556e-05, |
| "loss": 0.448, |
| "step": 11830 |
| }, |
| { |
| "epoch": 3.0472268691288122, |
| "grad_norm": 0.7505292892456055, |
| "learning_rate": 1.9534105534105536e-05, |
| "loss": 0.4604, |
| "step": 11840 |
| }, |
| { |
| "epoch": 3.049800540470982, |
| "grad_norm": 1.549533724784851, |
| "learning_rate": 1.9508365508365508e-05, |
| "loss": 0.4082, |
| "step": 11850 |
| }, |
| { |
| "epoch": 3.0523742118131514, |
| "grad_norm": 1.0561308860778809, |
| "learning_rate": 1.9482625482625483e-05, |
| "loss": 0.4062, |
| "step": 11860 |
| }, |
| { |
| "epoch": 3.054947883155321, |
| "grad_norm": 1.0458426475524902, |
| "learning_rate": 1.9456885456885458e-05, |
| "loss": 0.4374, |
| "step": 11870 |
| }, |
| { |
| "epoch": 3.0575215544974905, |
| "grad_norm": 0.9172849059104919, |
| "learning_rate": 1.9431145431145433e-05, |
| "loss": 0.4426, |
| "step": 11880 |
| }, |
| { |
| "epoch": 3.0600952258396603, |
| "grad_norm": 1.1287096738815308, |
| "learning_rate": 1.9405405405405404e-05, |
| "loss": 0.4038, |
| "step": 11890 |
| }, |
| { |
| "epoch": 3.0626688971818297, |
| "grad_norm": 1.0460259914398193, |
| "learning_rate": 1.937966537966538e-05, |
| "loss": 0.3982, |
| "step": 11900 |
| }, |
| { |
| "epoch": 3.0652425685239995, |
| "grad_norm": 1.1148087978363037, |
| "learning_rate": 1.9353925353925358e-05, |
| "loss": 0.4112, |
| "step": 11910 |
| }, |
| { |
| "epoch": 3.0678162398661692, |
| "grad_norm": 1.347355604171753, |
| "learning_rate": 1.932818532818533e-05, |
| "loss": 0.4286, |
| "step": 11920 |
| }, |
| { |
| "epoch": 3.0703899112083386, |
| "grad_norm": 1.4332010746002197, |
| "learning_rate": 1.9302445302445304e-05, |
| "loss": 0.3971, |
| "step": 11930 |
| }, |
| { |
| "epoch": 3.0729635825505084, |
| "grad_norm": 2.455965042114258, |
| "learning_rate": 1.927670527670528e-05, |
| "loss": 0.4328, |
| "step": 11940 |
| }, |
| { |
| "epoch": 3.0755372538926777, |
| "grad_norm": 1.319761037826538, |
| "learning_rate": 1.9250965250965254e-05, |
| "loss": 0.4166, |
| "step": 11950 |
| }, |
| { |
| "epoch": 3.0781109252348475, |
| "grad_norm": 1.1820671558380127, |
| "learning_rate": 1.9225225225225226e-05, |
| "loss": 0.4594, |
| "step": 11960 |
| }, |
| { |
| "epoch": 3.0806845965770173, |
| "grad_norm": 0.905479907989502, |
| "learning_rate": 1.91994851994852e-05, |
| "loss": 0.4466, |
| "step": 11970 |
| }, |
| { |
| "epoch": 3.0832582679191867, |
| "grad_norm": 1.3796758651733398, |
| "learning_rate": 1.9173745173745176e-05, |
| "loss": 0.398, |
| "step": 11980 |
| }, |
| { |
| "epoch": 3.0858319392613565, |
| "grad_norm": 1.7083042860031128, |
| "learning_rate": 1.914800514800515e-05, |
| "loss": 0.4554, |
| "step": 11990 |
| }, |
| { |
| "epoch": 3.088405610603526, |
| "grad_norm": 1.0998810529708862, |
| "learning_rate": 1.9122265122265122e-05, |
| "loss": 0.4031, |
| "step": 12000 |
| }, |
| { |
| "epoch": 3.088405610603526, |
| "eval_loss": 0.8223614692687988, |
| "eval_runtime": 395.6102, |
| "eval_samples_per_second": 49.104, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.00035795898845269097, |
| "step": 12000 |
| }, |
| { |
| "epoch": 3.0909792819456956, |
| "grad_norm": 0.8669257760047913, |
| "learning_rate": 1.9096525096525097e-05, |
| "loss": 0.4385, |
| "step": 12010 |
| }, |
| { |
| "epoch": 3.093552953287865, |
| "grad_norm": 1.6666041612625122, |
| "learning_rate": 1.9070785070785072e-05, |
| "loss": 0.4202, |
| "step": 12020 |
| }, |
| { |
| "epoch": 3.0961266246300347, |
| "grad_norm": 1.2697246074676514, |
| "learning_rate": 1.9045045045045047e-05, |
| "loss": 0.4472, |
| "step": 12030 |
| }, |
| { |
| "epoch": 3.0987002959722045, |
| "grad_norm": 1.0847759246826172, |
| "learning_rate": 1.901930501930502e-05, |
| "loss": 0.4642, |
| "step": 12040 |
| }, |
| { |
| "epoch": 3.101273967314374, |
| "grad_norm": 2.803271532058716, |
| "learning_rate": 1.8993564993564994e-05, |
| "loss": 0.4555, |
| "step": 12050 |
| }, |
| { |
| "epoch": 3.1038476386565437, |
| "grad_norm": 1.1774652004241943, |
| "learning_rate": 1.896782496782497e-05, |
| "loss": 0.472, |
| "step": 12060 |
| }, |
| { |
| "epoch": 3.106421309998713, |
| "grad_norm": 3.2696588039398193, |
| "learning_rate": 1.894208494208494e-05, |
| "loss": 0.4239, |
| "step": 12070 |
| }, |
| { |
| "epoch": 3.108994981340883, |
| "grad_norm": 1.0834957361221313, |
| "learning_rate": 1.8916344916344916e-05, |
| "loss": 0.3855, |
| "step": 12080 |
| }, |
| { |
| "epoch": 3.111568652683052, |
| "grad_norm": 1.2990940809249878, |
| "learning_rate": 1.889060489060489e-05, |
| "loss": 0.4416, |
| "step": 12090 |
| }, |
| { |
| "epoch": 3.114142324025222, |
| "grad_norm": 5.679697036743164, |
| "learning_rate": 1.8864864864864866e-05, |
| "loss": 0.4118, |
| "step": 12100 |
| }, |
| { |
| "epoch": 3.1167159953673917, |
| "grad_norm": 1.1892821788787842, |
| "learning_rate": 1.8839124839124837e-05, |
| "loss": 0.3947, |
| "step": 12110 |
| }, |
| { |
| "epoch": 3.119289666709561, |
| "grad_norm": 3.293264389038086, |
| "learning_rate": 1.8813384813384816e-05, |
| "loss": 0.4016, |
| "step": 12120 |
| }, |
| { |
| "epoch": 3.121863338051731, |
| "grad_norm": 1.1677567958831787, |
| "learning_rate": 1.878764478764479e-05, |
| "loss": 0.4077, |
| "step": 12130 |
| }, |
| { |
| "epoch": 3.1244370093939002, |
| "grad_norm": 1.0109992027282715, |
| "learning_rate": 1.8761904761904766e-05, |
| "loss": 0.3985, |
| "step": 12140 |
| }, |
| { |
| "epoch": 3.12701068073607, |
| "grad_norm": 1.0192488431930542, |
| "learning_rate": 1.8736164736164737e-05, |
| "loss": 0.4088, |
| "step": 12150 |
| }, |
| { |
| "epoch": 3.12958435207824, |
| "grad_norm": 2.916017532348633, |
| "learning_rate": 1.8710424710424712e-05, |
| "loss": 0.4372, |
| "step": 12160 |
| }, |
| { |
| "epoch": 3.132158023420409, |
| "grad_norm": 1.1380913257598877, |
| "learning_rate": 1.8684684684684687e-05, |
| "loss": 0.4914, |
| "step": 12170 |
| }, |
| { |
| "epoch": 3.134731694762579, |
| "grad_norm": 1.2204816341400146, |
| "learning_rate": 1.865894465894466e-05, |
| "loss": 0.3692, |
| "step": 12180 |
| }, |
| { |
| "epoch": 3.1373053661047483, |
| "grad_norm": 5.1090803146362305, |
| "learning_rate": 1.8633204633204634e-05, |
| "loss": 0.4168, |
| "step": 12190 |
| }, |
| { |
| "epoch": 3.139879037446918, |
| "grad_norm": 5.300877094268799, |
| "learning_rate": 1.860746460746461e-05, |
| "loss": 0.4533, |
| "step": 12200 |
| }, |
| { |
| "epoch": 3.1424527087890874, |
| "grad_norm": 1.053403377532959, |
| "learning_rate": 1.8581724581724584e-05, |
| "loss": 0.4207, |
| "step": 12210 |
| }, |
| { |
| "epoch": 3.1450263801312572, |
| "grad_norm": 0.8829357028007507, |
| "learning_rate": 1.8555984555984555e-05, |
| "loss": 0.4247, |
| "step": 12220 |
| }, |
| { |
| "epoch": 3.147600051473427, |
| "grad_norm": 1.015411138534546, |
| "learning_rate": 1.853024453024453e-05, |
| "loss": 0.4154, |
| "step": 12230 |
| }, |
| { |
| "epoch": 3.1501737228155964, |
| "grad_norm": 4.939092636108398, |
| "learning_rate": 1.8504504504504505e-05, |
| "loss": 0.3905, |
| "step": 12240 |
| }, |
| { |
| "epoch": 3.152747394157766, |
| "grad_norm": 1.0195497274398804, |
| "learning_rate": 1.847876447876448e-05, |
| "loss": 0.4135, |
| "step": 12250 |
| }, |
| { |
| "epoch": 3.1553210654999355, |
| "grad_norm": 1.3973890542984009, |
| "learning_rate": 1.8453024453024452e-05, |
| "loss": 0.4462, |
| "step": 12260 |
| }, |
| { |
| "epoch": 3.1578947368421053, |
| "grad_norm": 1.9131966829299927, |
| "learning_rate": 1.8427284427284427e-05, |
| "loss": 0.4195, |
| "step": 12270 |
| }, |
| { |
| "epoch": 3.1604684081842747, |
| "grad_norm": 1.4273961782455444, |
| "learning_rate": 1.8401544401544402e-05, |
| "loss": 0.4078, |
| "step": 12280 |
| }, |
| { |
| "epoch": 3.1630420795264445, |
| "grad_norm": 0.850229024887085, |
| "learning_rate": 1.8375804375804377e-05, |
| "loss": 0.4249, |
| "step": 12290 |
| }, |
| { |
| "epoch": 3.1656157508686142, |
| "grad_norm": 1.3435311317443848, |
| "learning_rate": 1.835006435006435e-05, |
| "loss": 0.4236, |
| "step": 12300 |
| }, |
| { |
| "epoch": 3.1681894222107836, |
| "grad_norm": 1.1101402044296265, |
| "learning_rate": 1.8324324324324324e-05, |
| "loss": 0.4806, |
| "step": 12310 |
| }, |
| { |
| "epoch": 3.1707630935529534, |
| "grad_norm": 1.9284107685089111, |
| "learning_rate": 1.82985842985843e-05, |
| "loss": 0.4059, |
| "step": 12320 |
| }, |
| { |
| "epoch": 3.1733367648951227, |
| "grad_norm": 0.9383375644683838, |
| "learning_rate": 1.8272844272844274e-05, |
| "loss": 0.4593, |
| "step": 12330 |
| }, |
| { |
| "epoch": 3.1759104362372925, |
| "grad_norm": 1.1090718507766724, |
| "learning_rate": 1.824710424710425e-05, |
| "loss": 0.451, |
| "step": 12340 |
| }, |
| { |
| "epoch": 3.178484107579462, |
| "grad_norm": 3.4713733196258545, |
| "learning_rate": 1.8221364221364224e-05, |
| "loss": 0.4403, |
| "step": 12350 |
| }, |
| { |
| "epoch": 3.1810577789216317, |
| "grad_norm": 1.001657485961914, |
| "learning_rate": 1.81956241956242e-05, |
| "loss": 0.4593, |
| "step": 12360 |
| }, |
| { |
| "epoch": 3.1836314502638015, |
| "grad_norm": 0.9134547710418701, |
| "learning_rate": 1.816988416988417e-05, |
| "loss": 0.4513, |
| "step": 12370 |
| }, |
| { |
| "epoch": 3.186205121605971, |
| "grad_norm": 1.105063557624817, |
| "learning_rate": 1.8144144144144145e-05, |
| "loss": 0.4476, |
| "step": 12380 |
| }, |
| { |
| "epoch": 3.1887787929481406, |
| "grad_norm": 1.3801063299179077, |
| "learning_rate": 1.811840411840412e-05, |
| "loss": 0.4312, |
| "step": 12390 |
| }, |
| { |
| "epoch": 3.19135246429031, |
| "grad_norm": 5.133249282836914, |
| "learning_rate": 1.8092664092664095e-05, |
| "loss": 0.4124, |
| "step": 12400 |
| }, |
| { |
| "epoch": 3.1939261356324797, |
| "grad_norm": 1.1436964273452759, |
| "learning_rate": 1.8066924066924067e-05, |
| "loss": 0.4242, |
| "step": 12410 |
| }, |
| { |
| "epoch": 3.1964998069746495, |
| "grad_norm": 0.8527979850769043, |
| "learning_rate": 1.8041184041184042e-05, |
| "loss": 0.4159, |
| "step": 12420 |
| }, |
| { |
| "epoch": 3.199073478316819, |
| "grad_norm": 1.0988205671310425, |
| "learning_rate": 1.8015444015444017e-05, |
| "loss": 0.3943, |
| "step": 12430 |
| }, |
| { |
| "epoch": 3.2016471496589887, |
| "grad_norm": 2.3774611949920654, |
| "learning_rate": 1.7989703989703992e-05, |
| "loss": 0.4405, |
| "step": 12440 |
| }, |
| { |
| "epoch": 3.204220821001158, |
| "grad_norm": 1.3425320386886597, |
| "learning_rate": 1.7963963963963963e-05, |
| "loss": 0.4397, |
| "step": 12450 |
| }, |
| { |
| "epoch": 3.206794492343328, |
| "grad_norm": 1.1551008224487305, |
| "learning_rate": 1.793822393822394e-05, |
| "loss": 0.4177, |
| "step": 12460 |
| }, |
| { |
| "epoch": 3.209368163685497, |
| "grad_norm": 1.0739468336105347, |
| "learning_rate": 1.7912483912483913e-05, |
| "loss": 0.4226, |
| "step": 12470 |
| }, |
| { |
| "epoch": 3.211941835027667, |
| "grad_norm": 1.012554407119751, |
| "learning_rate": 1.788674388674389e-05, |
| "loss": 0.4394, |
| "step": 12480 |
| }, |
| { |
| "epoch": 3.2145155063698367, |
| "grad_norm": 1.0329439640045166, |
| "learning_rate": 1.786100386100386e-05, |
| "loss": 0.4219, |
| "step": 12490 |
| }, |
| { |
| "epoch": 3.217089177712006, |
| "grad_norm": 1.340021014213562, |
| "learning_rate": 1.7835263835263835e-05, |
| "loss": 0.4431, |
| "step": 12500 |
| }, |
| { |
| "epoch": 3.219662849054176, |
| "grad_norm": 0.9239644408226013, |
| "learning_rate": 1.780952380952381e-05, |
| "loss": 0.4128, |
| "step": 12510 |
| }, |
| { |
| "epoch": 3.2222365203963452, |
| "grad_norm": 1.1054936647415161, |
| "learning_rate": 1.7783783783783785e-05, |
| "loss": 0.4215, |
| "step": 12520 |
| }, |
| { |
| "epoch": 3.224810191738515, |
| "grad_norm": 4.010881423950195, |
| "learning_rate": 1.7758043758043757e-05, |
| "loss": 0.3813, |
| "step": 12530 |
| }, |
| { |
| "epoch": 3.227383863080685, |
| "grad_norm": 1.6302077770233154, |
| "learning_rate": 1.7732303732303735e-05, |
| "loss": 0.427, |
| "step": 12540 |
| }, |
| { |
| "epoch": 3.229957534422854, |
| "grad_norm": 4.041463851928711, |
| "learning_rate": 1.770656370656371e-05, |
| "loss": 0.4766, |
| "step": 12550 |
| }, |
| { |
| "epoch": 3.232531205765024, |
| "grad_norm": 1.1331210136413574, |
| "learning_rate": 1.768082368082368e-05, |
| "loss": 0.4566, |
| "step": 12560 |
| }, |
| { |
| "epoch": 3.2351048771071933, |
| "grad_norm": 1.1952770948410034, |
| "learning_rate": 1.7655083655083657e-05, |
| "loss": 0.4393, |
| "step": 12570 |
| }, |
| { |
| "epoch": 3.237678548449363, |
| "grad_norm": 1.2376477718353271, |
| "learning_rate": 1.762934362934363e-05, |
| "loss": 0.4346, |
| "step": 12580 |
| }, |
| { |
| "epoch": 3.2402522197915324, |
| "grad_norm": 0.931395947933197, |
| "learning_rate": 1.7603603603603607e-05, |
| "loss": 0.4355, |
| "step": 12590 |
| }, |
| { |
| "epoch": 3.2428258911337022, |
| "grad_norm": 0.7813390493392944, |
| "learning_rate": 1.7577863577863578e-05, |
| "loss": 0.459, |
| "step": 12600 |
| }, |
| { |
| "epoch": 3.245399562475872, |
| "grad_norm": 0.9730493426322937, |
| "learning_rate": 1.7552123552123553e-05, |
| "loss": 0.4541, |
| "step": 12610 |
| }, |
| { |
| "epoch": 3.2479732338180414, |
| "grad_norm": 1.0856324434280396, |
| "learning_rate": 1.7526383526383528e-05, |
| "loss": 0.4046, |
| "step": 12620 |
| }, |
| { |
| "epoch": 3.250546905160211, |
| "grad_norm": 1.2812646627426147, |
| "learning_rate": 1.7500643500643503e-05, |
| "loss": 0.4322, |
| "step": 12630 |
| }, |
| { |
| "epoch": 3.2531205765023805, |
| "grad_norm": 1.2020498514175415, |
| "learning_rate": 1.7474903474903475e-05, |
| "loss": 0.4366, |
| "step": 12640 |
| }, |
| { |
| "epoch": 3.2556942478445503, |
| "grad_norm": 1.462583065032959, |
| "learning_rate": 1.744916344916345e-05, |
| "loss": 0.3911, |
| "step": 12650 |
| }, |
| { |
| "epoch": 3.2582679191867197, |
| "grad_norm": 0.9804076552391052, |
| "learning_rate": 1.7423423423423425e-05, |
| "loss": 0.501, |
| "step": 12660 |
| }, |
| { |
| "epoch": 3.2608415905288894, |
| "grad_norm": 1.153205156326294, |
| "learning_rate": 1.7397683397683396e-05, |
| "loss": 0.4456, |
| "step": 12670 |
| }, |
| { |
| "epoch": 3.2634152618710592, |
| "grad_norm": 0.8798128366470337, |
| "learning_rate": 1.737194337194337e-05, |
| "loss": 0.4622, |
| "step": 12680 |
| }, |
| { |
| "epoch": 3.2659889332132286, |
| "grad_norm": 1.3517348766326904, |
| "learning_rate": 1.7346203346203346e-05, |
| "loss": 0.4705, |
| "step": 12690 |
| }, |
| { |
| "epoch": 3.2685626045553984, |
| "grad_norm": 1.0517302751541138, |
| "learning_rate": 1.732046332046332e-05, |
| "loss": 0.502, |
| "step": 12700 |
| }, |
| { |
| "epoch": 3.2711362758975677, |
| "grad_norm": 0.8883270621299744, |
| "learning_rate": 1.7294723294723293e-05, |
| "loss": 0.4055, |
| "step": 12710 |
| }, |
| { |
| "epoch": 3.2737099472397375, |
| "grad_norm": 0.9611861109733582, |
| "learning_rate": 1.7268983268983268e-05, |
| "loss": 0.4462, |
| "step": 12720 |
| }, |
| { |
| "epoch": 3.276283618581907, |
| "grad_norm": 0.9720549583435059, |
| "learning_rate": 1.7243243243243243e-05, |
| "loss": 0.4339, |
| "step": 12730 |
| }, |
| { |
| "epoch": 3.2788572899240767, |
| "grad_norm": 1.8205825090408325, |
| "learning_rate": 1.7217503217503218e-05, |
| "loss": 0.4815, |
| "step": 12740 |
| }, |
| { |
| "epoch": 3.2814309612662464, |
| "grad_norm": 1.316831111907959, |
| "learning_rate": 1.7191763191763193e-05, |
| "loss": 0.4655, |
| "step": 12750 |
| }, |
| { |
| "epoch": 3.284004632608416, |
| "grad_norm": 1.3125663995742798, |
| "learning_rate": 1.7166023166023168e-05, |
| "loss": 0.4719, |
| "step": 12760 |
| }, |
| { |
| "epoch": 3.2865783039505856, |
| "grad_norm": 1.2112400531768799, |
| "learning_rate": 1.7140283140283143e-05, |
| "loss": 0.4137, |
| "step": 12770 |
| }, |
| { |
| "epoch": 3.289151975292755, |
| "grad_norm": 1.2232820987701416, |
| "learning_rate": 1.7114543114543115e-05, |
| "loss": 0.4392, |
| "step": 12780 |
| }, |
| { |
| "epoch": 3.2917256466349247, |
| "grad_norm": 1.1535100936889648, |
| "learning_rate": 1.708880308880309e-05, |
| "loss": 0.4628, |
| "step": 12790 |
| }, |
| { |
| "epoch": 3.294299317977094, |
| "grad_norm": 1.443505048751831, |
| "learning_rate": 1.7063063063063065e-05, |
| "loss": 0.4385, |
| "step": 12800 |
| }, |
| { |
| "epoch": 3.296872989319264, |
| "grad_norm": 1.0268503427505493, |
| "learning_rate": 1.703732303732304e-05, |
| "loss": 0.4135, |
| "step": 12810 |
| }, |
| { |
| "epoch": 3.2994466606614337, |
| "grad_norm": 0.8816016316413879, |
| "learning_rate": 1.701158301158301e-05, |
| "loss": 0.4897, |
| "step": 12820 |
| }, |
| { |
| "epoch": 3.302020332003603, |
| "grad_norm": 1.4715735912322998, |
| "learning_rate": 1.6985842985842986e-05, |
| "loss": 0.4285, |
| "step": 12830 |
| }, |
| { |
| "epoch": 3.304594003345773, |
| "grad_norm": 1.63077974319458, |
| "learning_rate": 1.696010296010296e-05, |
| "loss": 0.4841, |
| "step": 12840 |
| }, |
| { |
| "epoch": 3.3071676746879426, |
| "grad_norm": 1.157626748085022, |
| "learning_rate": 1.6934362934362936e-05, |
| "loss": 0.4523, |
| "step": 12850 |
| }, |
| { |
| "epoch": 3.309741346030112, |
| "grad_norm": 2.7212443351745605, |
| "learning_rate": 1.6908622908622908e-05, |
| "loss": 0.4134, |
| "step": 12860 |
| }, |
| { |
| "epoch": 3.3123150173722817, |
| "grad_norm": 0.8014114499092102, |
| "learning_rate": 1.6882882882882883e-05, |
| "loss": 0.3669, |
| "step": 12870 |
| }, |
| { |
| "epoch": 3.314888688714451, |
| "grad_norm": 1.2325034141540527, |
| "learning_rate": 1.6857142857142858e-05, |
| "loss": 0.4272, |
| "step": 12880 |
| }, |
| { |
| "epoch": 3.317462360056621, |
| "grad_norm": 0.9247467517852783, |
| "learning_rate": 1.6831402831402833e-05, |
| "loss": 0.4294, |
| "step": 12890 |
| }, |
| { |
| "epoch": 3.32003603139879, |
| "grad_norm": 0.9387899041175842, |
| "learning_rate": 1.6805662805662805e-05, |
| "loss": 0.4372, |
| "step": 12900 |
| }, |
| { |
| "epoch": 3.32260970274096, |
| "grad_norm": 3.1787610054016113, |
| "learning_rate": 1.677992277992278e-05, |
| "loss": 0.4523, |
| "step": 12910 |
| }, |
| { |
| "epoch": 3.32518337408313, |
| "grad_norm": 1.0653231143951416, |
| "learning_rate": 1.6754182754182755e-05, |
| "loss": 0.4646, |
| "step": 12920 |
| }, |
| { |
| "epoch": 3.327757045425299, |
| "grad_norm": 1.1255438327789307, |
| "learning_rate": 1.672844272844273e-05, |
| "loss": 0.4399, |
| "step": 12930 |
| }, |
| { |
| "epoch": 3.330330716767469, |
| "grad_norm": 1.021522879600525, |
| "learning_rate": 1.67027027027027e-05, |
| "loss": 0.4362, |
| "step": 12940 |
| }, |
| { |
| "epoch": 3.3329043881096383, |
| "grad_norm": 1.3337011337280273, |
| "learning_rate": 1.667696267696268e-05, |
| "loss": 0.4598, |
| "step": 12950 |
| }, |
| { |
| "epoch": 3.335478059451808, |
| "grad_norm": 1.0208321809768677, |
| "learning_rate": 1.6651222651222654e-05, |
| "loss": 0.4132, |
| "step": 12960 |
| }, |
| { |
| "epoch": 3.3380517307939774, |
| "grad_norm": 1.224876046180725, |
| "learning_rate": 1.6625482625482626e-05, |
| "loss": 0.412, |
| "step": 12970 |
| }, |
| { |
| "epoch": 3.3406254021361472, |
| "grad_norm": 1.0516496896743774, |
| "learning_rate": 1.65997425997426e-05, |
| "loss": 0.4059, |
| "step": 12980 |
| }, |
| { |
| "epoch": 3.343199073478317, |
| "grad_norm": 1.2751177549362183, |
| "learning_rate": 1.6574002574002576e-05, |
| "loss": 0.4684, |
| "step": 12990 |
| }, |
| { |
| "epoch": 3.3457727448204864, |
| "grad_norm": 1.3460466861724854, |
| "learning_rate": 1.654826254826255e-05, |
| "loss": 0.4497, |
| "step": 13000 |
| }, |
| { |
| "epoch": 3.3457727448204864, |
| "eval_loss": 0.8037166595458984, |
| "eval_runtime": 395.705, |
| "eval_samples_per_second": 49.092, |
| "eval_steps_per_second": 2.456, |
| "eval_token_accuracy": 0.00035492222078819304, |
| "step": 13000 |
| }, |
| { |
| "epoch": 3.348346416162656, |
| "grad_norm": 1.3930424451828003, |
| "learning_rate": 1.6522522522522523e-05, |
| "loss": 0.3934, |
| "step": 13010 |
| }, |
| { |
| "epoch": 3.3509200875048255, |
| "grad_norm": 1.0718811750411987, |
| "learning_rate": 1.6496782496782498e-05, |
| "loss": 0.439, |
| "step": 13020 |
| }, |
| { |
| "epoch": 3.3534937588469953, |
| "grad_norm": 0.9547449946403503, |
| "learning_rate": 1.6471042471042473e-05, |
| "loss": 0.4055, |
| "step": 13030 |
| }, |
| { |
| "epoch": 3.3560674301891646, |
| "grad_norm": 1.0712281465530396, |
| "learning_rate": 1.6445302445302448e-05, |
| "loss": 0.4388, |
| "step": 13040 |
| }, |
| { |
| "epoch": 3.3586411015313344, |
| "grad_norm": 1.256049394607544, |
| "learning_rate": 1.641956241956242e-05, |
| "loss": 0.5076, |
| "step": 13050 |
| }, |
| { |
| "epoch": 3.3612147728735042, |
| "grad_norm": 0.8553959131240845, |
| "learning_rate": 1.6393822393822394e-05, |
| "loss": 0.4392, |
| "step": 13060 |
| }, |
| { |
| "epoch": 3.3637884442156736, |
| "grad_norm": 0.9777474403381348, |
| "learning_rate": 1.636808236808237e-05, |
| "loss": 0.4033, |
| "step": 13070 |
| }, |
| { |
| "epoch": 3.3663621155578434, |
| "grad_norm": 1.901464819908142, |
| "learning_rate": 1.6342342342342344e-05, |
| "loss": 0.446, |
| "step": 13080 |
| }, |
| { |
| "epoch": 3.3689357869000127, |
| "grad_norm": 1.1458752155303955, |
| "learning_rate": 1.6316602316602316e-05, |
| "loss": 0.4282, |
| "step": 13090 |
| }, |
| { |
| "epoch": 3.3715094582421825, |
| "grad_norm": 1.075851321220398, |
| "learning_rate": 1.629086229086229e-05, |
| "loss": 0.4591, |
| "step": 13100 |
| }, |
| { |
| "epoch": 3.374083129584352, |
| "grad_norm": 1.1775199174880981, |
| "learning_rate": 1.6265122265122266e-05, |
| "loss": 0.4184, |
| "step": 13110 |
| }, |
| { |
| "epoch": 3.3766568009265217, |
| "grad_norm": 1.3904964923858643, |
| "learning_rate": 1.623938223938224e-05, |
| "loss": 0.4494, |
| "step": 13120 |
| }, |
| { |
| "epoch": 3.3792304722686914, |
| "grad_norm": 1.2566157579421997, |
| "learning_rate": 1.6213642213642213e-05, |
| "loss": 0.483, |
| "step": 13130 |
| }, |
| { |
| "epoch": 3.381804143610861, |
| "grad_norm": 0.7696701288223267, |
| "learning_rate": 1.6187902187902188e-05, |
| "loss": 0.4492, |
| "step": 13140 |
| }, |
| { |
| "epoch": 3.3843778149530306, |
| "grad_norm": 1.138867974281311, |
| "learning_rate": 1.6162162162162163e-05, |
| "loss": 0.448, |
| "step": 13150 |
| }, |
| { |
| "epoch": 3.3869514862952, |
| "grad_norm": 1.0716248750686646, |
| "learning_rate": 1.6136422136422138e-05, |
| "loss": 0.403, |
| "step": 13160 |
| }, |
| { |
| "epoch": 3.3895251576373697, |
| "grad_norm": 1.0408307313919067, |
| "learning_rate": 1.6110682110682113e-05, |
| "loss": 0.4505, |
| "step": 13170 |
| }, |
| { |
| "epoch": 3.392098828979539, |
| "grad_norm": 1.176862120628357, |
| "learning_rate": 1.6084942084942088e-05, |
| "loss": 0.4501, |
| "step": 13180 |
| }, |
| { |
| "epoch": 3.394672500321709, |
| "grad_norm": 1.18982994556427, |
| "learning_rate": 1.6059202059202063e-05, |
| "loss": 0.4152, |
| "step": 13190 |
| }, |
| { |
| "epoch": 3.3972461716638787, |
| "grad_norm": 3.0527262687683105, |
| "learning_rate": 1.6033462033462034e-05, |
| "loss": 0.4094, |
| "step": 13200 |
| }, |
| { |
| "epoch": 3.399819843006048, |
| "grad_norm": 1.14426589012146, |
| "learning_rate": 1.600772200772201e-05, |
| "loss": 0.4099, |
| "step": 13210 |
| }, |
| { |
| "epoch": 3.402393514348218, |
| "grad_norm": 1.081701397895813, |
| "learning_rate": 1.5981981981981984e-05, |
| "loss": 0.428, |
| "step": 13220 |
| }, |
| { |
| "epoch": 3.404967185690387, |
| "grad_norm": 1.6688580513000488, |
| "learning_rate": 1.5956241956241956e-05, |
| "loss": 0.5, |
| "step": 13230 |
| }, |
| { |
| "epoch": 3.407540857032557, |
| "grad_norm": 1.1758854389190674, |
| "learning_rate": 1.593050193050193e-05, |
| "loss": 0.4582, |
| "step": 13240 |
| }, |
| { |
| "epoch": 3.4101145283747267, |
| "grad_norm": 1.9258641004562378, |
| "learning_rate": 1.5904761904761906e-05, |
| "loss": 0.4227, |
| "step": 13250 |
| }, |
| { |
| "epoch": 3.412688199716896, |
| "grad_norm": 0.9658583402633667, |
| "learning_rate": 1.587902187902188e-05, |
| "loss": 0.4391, |
| "step": 13260 |
| }, |
| { |
| "epoch": 3.415261871059066, |
| "grad_norm": 1.136678695678711, |
| "learning_rate": 1.5853281853281852e-05, |
| "loss": 0.4561, |
| "step": 13270 |
| }, |
| { |
| "epoch": 3.417835542401235, |
| "grad_norm": 1.336816668510437, |
| "learning_rate": 1.5827541827541827e-05, |
| "loss": 0.4605, |
| "step": 13280 |
| }, |
| { |
| "epoch": 3.420409213743405, |
| "grad_norm": 1.2746632099151611, |
| "learning_rate": 1.5801801801801802e-05, |
| "loss": 0.4618, |
| "step": 13290 |
| }, |
| { |
| "epoch": 3.422982885085575, |
| "grad_norm": 1.2792760133743286, |
| "learning_rate": 1.5776061776061777e-05, |
| "loss": 0.4101, |
| "step": 13300 |
| }, |
| { |
| "epoch": 3.425556556427744, |
| "grad_norm": 1.0193665027618408, |
| "learning_rate": 1.575032175032175e-05, |
| "loss": 0.4173, |
| "step": 13310 |
| }, |
| { |
| "epoch": 3.428130227769914, |
| "grad_norm": 1.210837721824646, |
| "learning_rate": 1.5724581724581724e-05, |
| "loss": 0.4466, |
| "step": 13320 |
| }, |
| { |
| "epoch": 3.4307038991120833, |
| "grad_norm": 0.9081295132637024, |
| "learning_rate": 1.56988416988417e-05, |
| "loss": 0.4679, |
| "step": 13330 |
| }, |
| { |
| "epoch": 3.433277570454253, |
| "grad_norm": 1.0284429788589478, |
| "learning_rate": 1.5673101673101674e-05, |
| "loss": 0.4082, |
| "step": 13340 |
| }, |
| { |
| "epoch": 3.4358512417964224, |
| "grad_norm": 1.0147459506988525, |
| "learning_rate": 1.5647361647361646e-05, |
| "loss": 0.4096, |
| "step": 13350 |
| }, |
| { |
| "epoch": 3.438424913138592, |
| "grad_norm": 3.9549002647399902, |
| "learning_rate": 1.562162162162162e-05, |
| "loss": 0.4907, |
| "step": 13360 |
| }, |
| { |
| "epoch": 3.440998584480762, |
| "grad_norm": 1.1533993482589722, |
| "learning_rate": 1.55958815958816e-05, |
| "loss": 0.4478, |
| "step": 13370 |
| }, |
| { |
| "epoch": 3.4435722558229314, |
| "grad_norm": 1.3518677949905396, |
| "learning_rate": 1.557014157014157e-05, |
| "loss": 0.4654, |
| "step": 13380 |
| }, |
| { |
| "epoch": 3.446145927165101, |
| "grad_norm": 1.235627293586731, |
| "learning_rate": 1.5544401544401546e-05, |
| "loss": 0.4478, |
| "step": 13390 |
| }, |
| { |
| "epoch": 3.4487195985072705, |
| "grad_norm": 1.4760185480117798, |
| "learning_rate": 1.551866151866152e-05, |
| "loss": 0.4625, |
| "step": 13400 |
| }, |
| { |
| "epoch": 3.4512932698494403, |
| "grad_norm": 1.0546268224716187, |
| "learning_rate": 1.5492921492921496e-05, |
| "loss": 0.3887, |
| "step": 13410 |
| }, |
| { |
| "epoch": 3.4538669411916096, |
| "grad_norm": 0.971731960773468, |
| "learning_rate": 1.5467181467181467e-05, |
| "loss": 0.4381, |
| "step": 13420 |
| }, |
| { |
| "epoch": 3.4564406125337794, |
| "grad_norm": 1.2670224905014038, |
| "learning_rate": 1.5441441441441442e-05, |
| "loss": 0.5249, |
| "step": 13430 |
| }, |
| { |
| "epoch": 3.4590142838759492, |
| "grad_norm": 1.4726381301879883, |
| "learning_rate": 1.5415701415701417e-05, |
| "loss": 0.4644, |
| "step": 13440 |
| }, |
| { |
| "epoch": 3.4615879552181186, |
| "grad_norm": 0.9591072797775269, |
| "learning_rate": 1.5389961389961392e-05, |
| "loss": 0.4573, |
| "step": 13450 |
| }, |
| { |
| "epoch": 3.4641616265602884, |
| "grad_norm": 1.0489442348480225, |
| "learning_rate": 1.5364221364221364e-05, |
| "loss": 0.4467, |
| "step": 13460 |
| }, |
| { |
| "epoch": 3.4667352979024577, |
| "grad_norm": 1.2839806079864502, |
| "learning_rate": 1.533848133848134e-05, |
| "loss": 0.4844, |
| "step": 13470 |
| }, |
| { |
| "epoch": 3.4693089692446275, |
| "grad_norm": 1.2316621541976929, |
| "learning_rate": 1.5312741312741314e-05, |
| "loss": 0.4506, |
| "step": 13480 |
| }, |
| { |
| "epoch": 3.471882640586797, |
| "grad_norm": 1.1504662036895752, |
| "learning_rate": 1.528700128700129e-05, |
| "loss": 0.3969, |
| "step": 13490 |
| }, |
| { |
| "epoch": 3.4744563119289666, |
| "grad_norm": 1.4414169788360596, |
| "learning_rate": 1.526126126126126e-05, |
| "loss": 0.433, |
| "step": 13500 |
| }, |
| { |
| "epoch": 3.4770299832711364, |
| "grad_norm": 1.1712054014205933, |
| "learning_rate": 1.5235521235521235e-05, |
| "loss": 0.4622, |
| "step": 13510 |
| }, |
| { |
| "epoch": 3.479603654613306, |
| "grad_norm": 0.9692360758781433, |
| "learning_rate": 1.520978120978121e-05, |
| "loss": 0.4323, |
| "step": 13520 |
| }, |
| { |
| "epoch": 3.4821773259554756, |
| "grad_norm": 3.664949655532837, |
| "learning_rate": 1.5184041184041184e-05, |
| "loss": 0.4761, |
| "step": 13530 |
| }, |
| { |
| "epoch": 3.484750997297645, |
| "grad_norm": 1.2114914655685425, |
| "learning_rate": 1.5158301158301159e-05, |
| "loss": 0.5089, |
| "step": 13540 |
| }, |
| { |
| "epoch": 3.4873246686398147, |
| "grad_norm": 1.1861014366149902, |
| "learning_rate": 1.5132561132561132e-05, |
| "loss": 0.4537, |
| "step": 13550 |
| }, |
| { |
| "epoch": 3.489898339981984, |
| "grad_norm": 1.2321648597717285, |
| "learning_rate": 1.5106821106821107e-05, |
| "loss": 0.5039, |
| "step": 13560 |
| }, |
| { |
| "epoch": 3.492472011324154, |
| "grad_norm": 1.0023382902145386, |
| "learning_rate": 1.508108108108108e-05, |
| "loss": 0.4347, |
| "step": 13570 |
| }, |
| { |
| "epoch": 3.4950456826663236, |
| "grad_norm": 1.5064911842346191, |
| "learning_rate": 1.5055341055341057e-05, |
| "loss": 0.4311, |
| "step": 13580 |
| }, |
| { |
| "epoch": 3.497619354008493, |
| "grad_norm": 2.946744203567505, |
| "learning_rate": 1.5029601029601032e-05, |
| "loss": 0.4479, |
| "step": 13590 |
| }, |
| { |
| "epoch": 3.500193025350663, |
| "grad_norm": 3.439570188522339, |
| "learning_rate": 1.5003861003861005e-05, |
| "loss": 0.4942, |
| "step": 13600 |
| }, |
| { |
| "epoch": 3.5027666966928326, |
| "grad_norm": 1.228228211402893, |
| "learning_rate": 1.497812097812098e-05, |
| "loss": 0.4554, |
| "step": 13610 |
| }, |
| { |
| "epoch": 3.505340368035002, |
| "grad_norm": 1.0095043182373047, |
| "learning_rate": 1.4952380952380954e-05, |
| "loss": 0.4988, |
| "step": 13620 |
| }, |
| { |
| "epoch": 3.5079140393771713, |
| "grad_norm": 1.0728389024734497, |
| "learning_rate": 1.4926640926640929e-05, |
| "loss": 0.5008, |
| "step": 13630 |
| }, |
| { |
| "epoch": 3.510487710719341, |
| "grad_norm": 1.4836064577102661, |
| "learning_rate": 1.4900900900900902e-05, |
| "loss": 0.4352, |
| "step": 13640 |
| }, |
| { |
| "epoch": 3.513061382061511, |
| "grad_norm": 1.051338791847229, |
| "learning_rate": 1.4875160875160877e-05, |
| "loss": 0.4109, |
| "step": 13650 |
| }, |
| { |
| "epoch": 3.51563505340368, |
| "grad_norm": 1.345146656036377, |
| "learning_rate": 1.484942084942085e-05, |
| "loss": 0.4174, |
| "step": 13660 |
| }, |
| { |
| "epoch": 3.51820872474585, |
| "grad_norm": 1.2961459159851074, |
| "learning_rate": 1.4823680823680825e-05, |
| "loss": 0.474, |
| "step": 13670 |
| }, |
| { |
| "epoch": 3.52078239608802, |
| "grad_norm": 1.370326042175293, |
| "learning_rate": 1.4797940797940799e-05, |
| "loss": 0.4244, |
| "step": 13680 |
| }, |
| { |
| "epoch": 3.523356067430189, |
| "grad_norm": 3.830646514892578, |
| "learning_rate": 1.4772200772200774e-05, |
| "loss": 0.4581, |
| "step": 13690 |
| }, |
| { |
| "epoch": 3.525929738772359, |
| "grad_norm": 1.1295220851898193, |
| "learning_rate": 1.4746460746460747e-05, |
| "loss": 0.435, |
| "step": 13700 |
| }, |
| { |
| "epoch": 3.5285034101145283, |
| "grad_norm": 1.1341050863265991, |
| "learning_rate": 1.472072072072072e-05, |
| "loss": 0.4484, |
| "step": 13710 |
| }, |
| { |
| "epoch": 3.531077081456698, |
| "grad_norm": 1.2032519578933716, |
| "learning_rate": 1.4694980694980695e-05, |
| "loss": 0.4368, |
| "step": 13720 |
| }, |
| { |
| "epoch": 3.5336507527988674, |
| "grad_norm": 2.008488655090332, |
| "learning_rate": 1.4669240669240668e-05, |
| "loss": 0.3941, |
| "step": 13730 |
| }, |
| { |
| "epoch": 3.536224424141037, |
| "grad_norm": 1.0747300386428833, |
| "learning_rate": 1.4643500643500643e-05, |
| "loss": 0.4884, |
| "step": 13740 |
| }, |
| { |
| "epoch": 3.538798095483207, |
| "grad_norm": 2.756009578704834, |
| "learning_rate": 1.4617760617760617e-05, |
| "loss": 0.4298, |
| "step": 13750 |
| }, |
| { |
| "epoch": 3.5413717668253764, |
| "grad_norm": 0.8556985259056091, |
| "learning_rate": 1.4592020592020592e-05, |
| "loss": 0.4827, |
| "step": 13760 |
| }, |
| { |
| "epoch": 3.543945438167546, |
| "grad_norm": 1.360617995262146, |
| "learning_rate": 1.4566280566280565e-05, |
| "loss": 0.4264, |
| "step": 13770 |
| }, |
| { |
| "epoch": 3.5465191095097155, |
| "grad_norm": 1.2899786233901978, |
| "learning_rate": 1.4540540540540543e-05, |
| "loss": 0.4293, |
| "step": 13780 |
| }, |
| { |
| "epoch": 3.5490927808518853, |
| "grad_norm": 1.2246886491775513, |
| "learning_rate": 1.4514800514800517e-05, |
| "loss": 0.4131, |
| "step": 13790 |
| }, |
| { |
| "epoch": 3.5516664521940546, |
| "grad_norm": 1.1932190656661987, |
| "learning_rate": 1.448906048906049e-05, |
| "loss": 0.476, |
| "step": 13800 |
| }, |
| { |
| "epoch": 3.5542401235362244, |
| "grad_norm": 4.828482151031494, |
| "learning_rate": 1.4463320463320465e-05, |
| "loss": 0.465, |
| "step": 13810 |
| }, |
| { |
| "epoch": 3.556813794878394, |
| "grad_norm": 1.4605960845947266, |
| "learning_rate": 1.4437580437580438e-05, |
| "loss": 0.4007, |
| "step": 13820 |
| }, |
| { |
| "epoch": 3.5593874662205636, |
| "grad_norm": 0.9224157929420471, |
| "learning_rate": 1.4411840411840413e-05, |
| "loss": 0.4044, |
| "step": 13830 |
| }, |
| { |
| "epoch": 3.5619611375627334, |
| "grad_norm": 1.017256498336792, |
| "learning_rate": 1.4386100386100387e-05, |
| "loss": 0.4502, |
| "step": 13840 |
| }, |
| { |
| "epoch": 3.5645348089049027, |
| "grad_norm": 1.9395209550857544, |
| "learning_rate": 1.4360360360360362e-05, |
| "loss": 0.3843, |
| "step": 13850 |
| }, |
| { |
| "epoch": 3.5671084802470725, |
| "grad_norm": 1.1960591077804565, |
| "learning_rate": 1.4334620334620335e-05, |
| "loss": 0.4528, |
| "step": 13860 |
| }, |
| { |
| "epoch": 3.569682151589242, |
| "grad_norm": 0.9504403471946716, |
| "learning_rate": 1.430888030888031e-05, |
| "loss": 0.4412, |
| "step": 13870 |
| }, |
| { |
| "epoch": 3.5722558229314116, |
| "grad_norm": 1.2557878494262695, |
| "learning_rate": 1.4283140283140283e-05, |
| "loss": 0.3845, |
| "step": 13880 |
| }, |
| { |
| "epoch": 3.5748294942735814, |
| "grad_norm": 4.165874004364014, |
| "learning_rate": 1.4257400257400258e-05, |
| "loss": 0.3932, |
| "step": 13890 |
| }, |
| { |
| "epoch": 3.577403165615751, |
| "grad_norm": 3.318126678466797, |
| "learning_rate": 1.4231660231660232e-05, |
| "loss": 0.3809, |
| "step": 13900 |
| }, |
| { |
| "epoch": 3.5799768369579206, |
| "grad_norm": 1.0943117141723633, |
| "learning_rate": 1.4205920205920207e-05, |
| "loss": 0.48, |
| "step": 13910 |
| }, |
| { |
| "epoch": 3.58255050830009, |
| "grad_norm": 1.2082642316818237, |
| "learning_rate": 1.418018018018018e-05, |
| "loss": 0.4894, |
| "step": 13920 |
| }, |
| { |
| "epoch": 3.5851241796422597, |
| "grad_norm": 0.6787092089653015, |
| "learning_rate": 1.4154440154440155e-05, |
| "loss": 0.4201, |
| "step": 13930 |
| }, |
| { |
| "epoch": 3.587697850984429, |
| "grad_norm": 1.3796510696411133, |
| "learning_rate": 1.4128700128700128e-05, |
| "loss": 0.5031, |
| "step": 13940 |
| }, |
| { |
| "epoch": 3.590271522326599, |
| "grad_norm": 0.9589558839797974, |
| "learning_rate": 1.4102960102960103e-05, |
| "loss": 0.4299, |
| "step": 13950 |
| }, |
| { |
| "epoch": 3.5928451936687686, |
| "grad_norm": 1.0614103078842163, |
| "learning_rate": 1.4077220077220076e-05, |
| "loss": 0.4265, |
| "step": 13960 |
| }, |
| { |
| "epoch": 3.595418865010938, |
| "grad_norm": 1.4390029907226562, |
| "learning_rate": 1.4051480051480051e-05, |
| "loss": 0.4648, |
| "step": 13970 |
| }, |
| { |
| "epoch": 3.597992536353108, |
| "grad_norm": 4.360158920288086, |
| "learning_rate": 1.4025740025740025e-05, |
| "loss": 0.4575, |
| "step": 13980 |
| }, |
| { |
| "epoch": 3.6005662076952776, |
| "grad_norm": 2.3131937980651855, |
| "learning_rate": 1.4000000000000001e-05, |
| "loss": 0.4087, |
| "step": 13990 |
| }, |
| { |
| "epoch": 3.603139879037447, |
| "grad_norm": 1.0350521802902222, |
| "learning_rate": 1.3974259974259976e-05, |
| "loss": 0.4096, |
| "step": 14000 |
| }, |
| { |
| "epoch": 3.603139879037447, |
| "eval_loss": 0.8108994960784912, |
| "eval_runtime": 395.6175, |
| "eval_samples_per_second": 49.103, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.0003543528268510997, |
| "step": 14000 |
| }, |
| { |
| "epoch": 3.6057135503796163, |
| "grad_norm": 1.2455912828445435, |
| "learning_rate": 1.394851994851995e-05, |
| "loss": 0.4473, |
| "step": 14010 |
| }, |
| { |
| "epoch": 3.608287221721786, |
| "grad_norm": 0.8037226796150208, |
| "learning_rate": 1.3922779922779925e-05, |
| "loss": 0.4257, |
| "step": 14020 |
| }, |
| { |
| "epoch": 3.610860893063956, |
| "grad_norm": 2.2132887840270996, |
| "learning_rate": 1.3897039897039898e-05, |
| "loss": 0.4057, |
| "step": 14030 |
| }, |
| { |
| "epoch": 3.613434564406125, |
| "grad_norm": 1.2261261940002441, |
| "learning_rate": 1.3871299871299873e-05, |
| "loss": 0.4251, |
| "step": 14040 |
| }, |
| { |
| "epoch": 3.616008235748295, |
| "grad_norm": 0.9967421293258667, |
| "learning_rate": 1.3845559845559846e-05, |
| "loss": 0.4117, |
| "step": 14050 |
| }, |
| { |
| "epoch": 3.618581907090465, |
| "grad_norm": 1.0189718008041382, |
| "learning_rate": 1.3819819819819821e-05, |
| "loss": 0.4292, |
| "step": 14060 |
| }, |
| { |
| "epoch": 3.621155578432634, |
| "grad_norm": 1.333105206489563, |
| "learning_rate": 1.3794079794079795e-05, |
| "loss": 0.4989, |
| "step": 14070 |
| }, |
| { |
| "epoch": 3.6237292497748035, |
| "grad_norm": 3.6797611713409424, |
| "learning_rate": 1.376833976833977e-05, |
| "loss": 0.4377, |
| "step": 14080 |
| }, |
| { |
| "epoch": 3.6263029211169733, |
| "grad_norm": 2.8054308891296387, |
| "learning_rate": 1.3742599742599743e-05, |
| "loss": 0.4932, |
| "step": 14090 |
| }, |
| { |
| "epoch": 3.628876592459143, |
| "grad_norm": 2.897303819656372, |
| "learning_rate": 1.3716859716859718e-05, |
| "loss": 0.4816, |
| "step": 14100 |
| }, |
| { |
| "epoch": 3.6314502638013124, |
| "grad_norm": 1.0359746217727661, |
| "learning_rate": 1.3691119691119691e-05, |
| "loss": 0.4421, |
| "step": 14110 |
| }, |
| { |
| "epoch": 3.634023935143482, |
| "grad_norm": 1.1515438556671143, |
| "learning_rate": 1.3665379665379666e-05, |
| "loss": 0.4766, |
| "step": 14120 |
| }, |
| { |
| "epoch": 3.636597606485652, |
| "grad_norm": 1.0731172561645508, |
| "learning_rate": 1.363963963963964e-05, |
| "loss": 0.4578, |
| "step": 14130 |
| }, |
| { |
| "epoch": 3.6391712778278213, |
| "grad_norm": 3.8715970516204834, |
| "learning_rate": 1.3613899613899615e-05, |
| "loss": 0.4549, |
| "step": 14140 |
| }, |
| { |
| "epoch": 3.641744949169991, |
| "grad_norm": 1.386406660079956, |
| "learning_rate": 1.3588159588159588e-05, |
| "loss": 0.5023, |
| "step": 14150 |
| }, |
| { |
| "epoch": 3.6443186205121605, |
| "grad_norm": 1.2222354412078857, |
| "learning_rate": 1.3562419562419563e-05, |
| "loss": 0.4375, |
| "step": 14160 |
| }, |
| { |
| "epoch": 3.6468922918543303, |
| "grad_norm": 1.2102971076965332, |
| "learning_rate": 1.3536679536679536e-05, |
| "loss": 0.4404, |
| "step": 14170 |
| }, |
| { |
| "epoch": 3.6494659631964996, |
| "grad_norm": 0.73658686876297, |
| "learning_rate": 1.3510939510939511e-05, |
| "loss": 0.4004, |
| "step": 14180 |
| }, |
| { |
| "epoch": 3.6520396345386694, |
| "grad_norm": 1.3987421989440918, |
| "learning_rate": 1.3485199485199485e-05, |
| "loss": 0.432, |
| "step": 14190 |
| }, |
| { |
| "epoch": 3.654613305880839, |
| "grad_norm": 1.2989593744277954, |
| "learning_rate": 1.3459459459459461e-05, |
| "loss": 0.4684, |
| "step": 14200 |
| }, |
| { |
| "epoch": 3.6571869772230086, |
| "grad_norm": 1.5131486654281616, |
| "learning_rate": 1.3433719433719436e-05, |
| "loss": 0.5014, |
| "step": 14210 |
| }, |
| { |
| "epoch": 3.6597606485651784, |
| "grad_norm": 0.9460519552230835, |
| "learning_rate": 1.340797940797941e-05, |
| "loss": 0.4686, |
| "step": 14220 |
| }, |
| { |
| "epoch": 3.6623343199073477, |
| "grad_norm": 1.3320332765579224, |
| "learning_rate": 1.3382239382239385e-05, |
| "loss": 0.4612, |
| "step": 14230 |
| }, |
| { |
| "epoch": 3.6649079912495175, |
| "grad_norm": 1.110107421875, |
| "learning_rate": 1.3356499356499358e-05, |
| "loss": 0.4824, |
| "step": 14240 |
| }, |
| { |
| "epoch": 3.667481662591687, |
| "grad_norm": 1.6514250040054321, |
| "learning_rate": 1.3330759330759333e-05, |
| "loss": 0.4678, |
| "step": 14250 |
| }, |
| { |
| "epoch": 3.6700553339338566, |
| "grad_norm": 2.858354330062866, |
| "learning_rate": 1.3305019305019306e-05, |
| "loss": 0.4576, |
| "step": 14260 |
| }, |
| { |
| "epoch": 3.6726290052760264, |
| "grad_norm": 1.5715115070343018, |
| "learning_rate": 1.327927927927928e-05, |
| "loss": 0.4547, |
| "step": 14270 |
| }, |
| { |
| "epoch": 3.6752026766181958, |
| "grad_norm": 0.938371479511261, |
| "learning_rate": 1.3253539253539254e-05, |
| "loss": 0.3987, |
| "step": 14280 |
| }, |
| { |
| "epoch": 3.6777763479603656, |
| "grad_norm": 0.9989467859268188, |
| "learning_rate": 1.3227799227799228e-05, |
| "loss": 0.4403, |
| "step": 14290 |
| }, |
| { |
| "epoch": 3.680350019302535, |
| "grad_norm": 1.117611289024353, |
| "learning_rate": 1.3202059202059203e-05, |
| "loss": 0.4009, |
| "step": 14300 |
| }, |
| { |
| "epoch": 3.6829236906447047, |
| "grad_norm": 1.238559603691101, |
| "learning_rate": 1.3176319176319176e-05, |
| "loss": 0.4198, |
| "step": 14310 |
| }, |
| { |
| "epoch": 3.685497361986874, |
| "grad_norm": 1.2747526168823242, |
| "learning_rate": 1.3150579150579151e-05, |
| "loss": 0.439, |
| "step": 14320 |
| }, |
| { |
| "epoch": 3.688071033329044, |
| "grad_norm": 3.8777308464050293, |
| "learning_rate": 1.3124839124839124e-05, |
| "loss": 0.4573, |
| "step": 14330 |
| }, |
| { |
| "epoch": 3.6906447046712136, |
| "grad_norm": 1.1357238292694092, |
| "learning_rate": 1.30990990990991e-05, |
| "loss": 0.4578, |
| "step": 14340 |
| }, |
| { |
| "epoch": 3.693218376013383, |
| "grad_norm": 1.4202942848205566, |
| "learning_rate": 1.3073359073359073e-05, |
| "loss": 0.4409, |
| "step": 14350 |
| }, |
| { |
| "epoch": 3.695792047355553, |
| "grad_norm": 1.4478346109390259, |
| "learning_rate": 1.3047619047619048e-05, |
| "loss": 0.4707, |
| "step": 14360 |
| }, |
| { |
| "epoch": 3.6983657186977226, |
| "grad_norm": 1.4722789525985718, |
| "learning_rate": 1.3021879021879021e-05, |
| "loss": 0.4445, |
| "step": 14370 |
| }, |
| { |
| "epoch": 3.700939390039892, |
| "grad_norm": 1.0790144205093384, |
| "learning_rate": 1.2996138996138996e-05, |
| "loss": 0.454, |
| "step": 14380 |
| }, |
| { |
| "epoch": 3.7035130613820613, |
| "grad_norm": 2.759157419204712, |
| "learning_rate": 1.297039897039897e-05, |
| "loss": 0.3881, |
| "step": 14390 |
| }, |
| { |
| "epoch": 3.706086732724231, |
| "grad_norm": 1.4952332973480225, |
| "learning_rate": 1.2944658944658944e-05, |
| "loss": 0.4109, |
| "step": 14400 |
| }, |
| { |
| "epoch": 3.708660404066401, |
| "grad_norm": 1.024506688117981, |
| "learning_rate": 1.2918918918918921e-05, |
| "loss": 0.4616, |
| "step": 14410 |
| }, |
| { |
| "epoch": 3.71123407540857, |
| "grad_norm": 5.463751316070557, |
| "learning_rate": 1.2893178893178894e-05, |
| "loss": 0.4293, |
| "step": 14420 |
| }, |
| { |
| "epoch": 3.71380774675074, |
| "grad_norm": 1.2399414777755737, |
| "learning_rate": 1.286743886743887e-05, |
| "loss": 0.4671, |
| "step": 14430 |
| }, |
| { |
| "epoch": 3.71638141809291, |
| "grad_norm": 1.1366684436798096, |
| "learning_rate": 1.2841698841698843e-05, |
| "loss": 0.3873, |
| "step": 14440 |
| }, |
| { |
| "epoch": 3.718955089435079, |
| "grad_norm": 1.025657296180725, |
| "learning_rate": 1.2815958815958818e-05, |
| "loss": 0.4927, |
| "step": 14450 |
| }, |
| { |
| "epoch": 3.7215287607772485, |
| "grad_norm": 1.4381457567214966, |
| "learning_rate": 1.2790218790218791e-05, |
| "loss": 0.399, |
| "step": 14460 |
| }, |
| { |
| "epoch": 3.7241024321194183, |
| "grad_norm": 4.138847351074219, |
| "learning_rate": 1.2764478764478766e-05, |
| "loss": 0.493, |
| "step": 14470 |
| }, |
| { |
| "epoch": 3.726676103461588, |
| "grad_norm": 3.5278961658477783, |
| "learning_rate": 1.273873873873874e-05, |
| "loss": 0.4522, |
| "step": 14480 |
| }, |
| { |
| "epoch": 3.7292497748037574, |
| "grad_norm": 1.029931902885437, |
| "learning_rate": 1.2712998712998714e-05, |
| "loss": 0.3838, |
| "step": 14490 |
| }, |
| { |
| "epoch": 3.731823446145927, |
| "grad_norm": 0.8327284455299377, |
| "learning_rate": 1.268983268983269e-05, |
| "loss": 0.4587, |
| "step": 14500 |
| }, |
| { |
| "epoch": 3.734397117488097, |
| "grad_norm": 2.9695467948913574, |
| "learning_rate": 1.2664092664092665e-05, |
| "loss": 0.4544, |
| "step": 14510 |
| }, |
| { |
| "epoch": 3.7369707888302663, |
| "grad_norm": 1.0291829109191895, |
| "learning_rate": 1.2638352638352638e-05, |
| "loss": 0.4691, |
| "step": 14520 |
| }, |
| { |
| "epoch": 3.739544460172436, |
| "grad_norm": 1.1941965818405151, |
| "learning_rate": 1.2612612612612611e-05, |
| "loss": 0.4585, |
| "step": 14530 |
| }, |
| { |
| "epoch": 3.7421181315146055, |
| "grad_norm": 1.7099372148513794, |
| "learning_rate": 1.2586872586872586e-05, |
| "loss": 0.4164, |
| "step": 14540 |
| }, |
| { |
| "epoch": 3.7446918028567753, |
| "grad_norm": 2.46553373336792, |
| "learning_rate": 1.2561132561132563e-05, |
| "loss": 0.4529, |
| "step": 14550 |
| }, |
| { |
| "epoch": 3.7472654741989446, |
| "grad_norm": 1.0007997751235962, |
| "learning_rate": 1.2535392535392538e-05, |
| "loss": 0.403, |
| "step": 14560 |
| }, |
| { |
| "epoch": 3.7498391455411144, |
| "grad_norm": 1.1691094636917114, |
| "learning_rate": 1.2509652509652511e-05, |
| "loss": 0.4435, |
| "step": 14570 |
| }, |
| { |
| "epoch": 3.752412816883284, |
| "grad_norm": 1.2735754251480103, |
| "learning_rate": 1.2483912483912485e-05, |
| "loss": 0.5126, |
| "step": 14580 |
| }, |
| { |
| "epoch": 3.7549864882254536, |
| "grad_norm": 3.02443265914917, |
| "learning_rate": 1.2458172458172458e-05, |
| "loss": 0.4898, |
| "step": 14590 |
| }, |
| { |
| "epoch": 3.7575601595676233, |
| "grad_norm": 1.1990752220153809, |
| "learning_rate": 1.2432432432432435e-05, |
| "loss": 0.4059, |
| "step": 14600 |
| }, |
| { |
| "epoch": 3.7601338309097927, |
| "grad_norm": 1.180837631225586, |
| "learning_rate": 1.2406692406692408e-05, |
| "loss": 0.4384, |
| "step": 14610 |
| }, |
| { |
| "epoch": 3.7627075022519625, |
| "grad_norm": 0.8962213397026062, |
| "learning_rate": 1.2380952380952381e-05, |
| "loss": 0.4297, |
| "step": 14620 |
| }, |
| { |
| "epoch": 3.765281173594132, |
| "grad_norm": 0.9834825396537781, |
| "learning_rate": 1.2355212355212356e-05, |
| "loss": 0.4057, |
| "step": 14630 |
| }, |
| { |
| "epoch": 3.7678548449363016, |
| "grad_norm": 1.174039363861084, |
| "learning_rate": 1.232947232947233e-05, |
| "loss": 0.4586, |
| "step": 14640 |
| }, |
| { |
| "epoch": 3.7704285162784714, |
| "grad_norm": 0.9779637455940247, |
| "learning_rate": 1.2303732303732304e-05, |
| "loss": 0.4011, |
| "step": 14650 |
| }, |
| { |
| "epoch": 3.7730021876206408, |
| "grad_norm": 1.003485918045044, |
| "learning_rate": 1.2277992277992278e-05, |
| "loss": 0.4899, |
| "step": 14660 |
| }, |
| { |
| "epoch": 3.7755758589628106, |
| "grad_norm": 1.3274792432785034, |
| "learning_rate": 1.2252252252252253e-05, |
| "loss": 0.4584, |
| "step": 14670 |
| }, |
| { |
| "epoch": 3.77814953030498, |
| "grad_norm": 1.5608062744140625, |
| "learning_rate": 1.2226512226512226e-05, |
| "loss": 0.5178, |
| "step": 14680 |
| }, |
| { |
| "epoch": 3.7807232016471497, |
| "grad_norm": 3.1873466968536377, |
| "learning_rate": 1.2200772200772201e-05, |
| "loss": 0.448, |
| "step": 14690 |
| }, |
| { |
| "epoch": 3.783296872989319, |
| "grad_norm": 0.8975459337234497, |
| "learning_rate": 1.2175032175032174e-05, |
| "loss": 0.3988, |
| "step": 14700 |
| }, |
| { |
| "epoch": 3.785870544331489, |
| "grad_norm": 1.3727668523788452, |
| "learning_rate": 1.2149292149292151e-05, |
| "loss": 0.4206, |
| "step": 14710 |
| }, |
| { |
| "epoch": 3.7884442156736586, |
| "grad_norm": 1.0184861421585083, |
| "learning_rate": 1.2123552123552124e-05, |
| "loss": 0.407, |
| "step": 14720 |
| }, |
| { |
| "epoch": 3.791017887015828, |
| "grad_norm": 0.9267153739929199, |
| "learning_rate": 1.20978120978121e-05, |
| "loss": 0.4058, |
| "step": 14730 |
| }, |
| { |
| "epoch": 3.7935915583579978, |
| "grad_norm": 1.0847712755203247, |
| "learning_rate": 1.2072072072072073e-05, |
| "loss": 0.428, |
| "step": 14740 |
| }, |
| { |
| "epoch": 3.7961652297001676, |
| "grad_norm": 1.368102788925171, |
| "learning_rate": 1.2046332046332048e-05, |
| "loss": 0.4002, |
| "step": 14750 |
| }, |
| { |
| "epoch": 3.798738901042337, |
| "grad_norm": 1.2423478364944458, |
| "learning_rate": 1.2020592020592021e-05, |
| "loss": 0.4696, |
| "step": 14760 |
| }, |
| { |
| "epoch": 3.8013125723845063, |
| "grad_norm": 1.3219197988510132, |
| "learning_rate": 1.1994851994851996e-05, |
| "loss": 0.4566, |
| "step": 14770 |
| }, |
| { |
| "epoch": 3.803886243726676, |
| "grad_norm": 0.9948756098747253, |
| "learning_rate": 1.196911196911197e-05, |
| "loss": 0.4422, |
| "step": 14780 |
| }, |
| { |
| "epoch": 3.806459915068846, |
| "grad_norm": 1.5322728157043457, |
| "learning_rate": 1.1943371943371944e-05, |
| "loss": 0.4079, |
| "step": 14790 |
| }, |
| { |
| "epoch": 3.809033586411015, |
| "grad_norm": 1.384191870689392, |
| "learning_rate": 1.1917631917631918e-05, |
| "loss": 0.4617, |
| "step": 14800 |
| }, |
| { |
| "epoch": 3.811607257753185, |
| "grad_norm": 1.3496747016906738, |
| "learning_rate": 1.1891891891891893e-05, |
| "loss": 0.4324, |
| "step": 14810 |
| }, |
| { |
| "epoch": 3.8141809290953548, |
| "grad_norm": 1.0405890941619873, |
| "learning_rate": 1.1866151866151868e-05, |
| "loss": 0.3989, |
| "step": 14820 |
| }, |
| { |
| "epoch": 3.816754600437524, |
| "grad_norm": 0.8725833296775818, |
| "learning_rate": 1.1840411840411841e-05, |
| "loss": 0.4727, |
| "step": 14830 |
| }, |
| { |
| "epoch": 3.8193282717796935, |
| "grad_norm": 1.375089406967163, |
| "learning_rate": 1.1814671814671816e-05, |
| "loss": 0.5229, |
| "step": 14840 |
| }, |
| { |
| "epoch": 3.8219019431218633, |
| "grad_norm": 1.0093263387680054, |
| "learning_rate": 1.178893178893179e-05, |
| "loss": 0.4561, |
| "step": 14850 |
| }, |
| { |
| "epoch": 3.824475614464033, |
| "grad_norm": 1.45624577999115, |
| "learning_rate": 1.1763191763191764e-05, |
| "loss": 0.4518, |
| "step": 14860 |
| }, |
| { |
| "epoch": 3.8270492858062024, |
| "grad_norm": 0.9073015451431274, |
| "learning_rate": 1.1737451737451738e-05, |
| "loss": 0.4404, |
| "step": 14870 |
| }, |
| { |
| "epoch": 3.829622957148372, |
| "grad_norm": 1.0322130918502808, |
| "learning_rate": 1.1711711711711713e-05, |
| "loss": 0.4328, |
| "step": 14880 |
| }, |
| { |
| "epoch": 3.832196628490542, |
| "grad_norm": 1.549871563911438, |
| "learning_rate": 1.1685971685971686e-05, |
| "loss": 0.4882, |
| "step": 14890 |
| }, |
| { |
| "epoch": 3.8347702998327113, |
| "grad_norm": 0.8814623355865479, |
| "learning_rate": 1.166023166023166e-05, |
| "loss": 0.4859, |
| "step": 14900 |
| }, |
| { |
| "epoch": 3.837343971174881, |
| "grad_norm": 1.2014186382293701, |
| "learning_rate": 1.1634491634491634e-05, |
| "loss": 0.4584, |
| "step": 14910 |
| }, |
| { |
| "epoch": 3.8399176425170505, |
| "grad_norm": 1.3947405815124512, |
| "learning_rate": 1.1608751608751609e-05, |
| "loss": 0.4389, |
| "step": 14920 |
| }, |
| { |
| "epoch": 3.8424913138592203, |
| "grad_norm": 1.3492847681045532, |
| "learning_rate": 1.1583011583011584e-05, |
| "loss": 0.4413, |
| "step": 14930 |
| }, |
| { |
| "epoch": 3.8450649852013896, |
| "grad_norm": 0.9952030777931213, |
| "learning_rate": 1.1557271557271557e-05, |
| "loss": 0.4203, |
| "step": 14940 |
| }, |
| { |
| "epoch": 3.8476386565435594, |
| "grad_norm": 1.1566764116287231, |
| "learning_rate": 1.1531531531531532e-05, |
| "loss": 0.4489, |
| "step": 14950 |
| }, |
| { |
| "epoch": 3.850212327885729, |
| "grad_norm": 1.0022921562194824, |
| "learning_rate": 1.1505791505791506e-05, |
| "loss": 0.4422, |
| "step": 14960 |
| }, |
| { |
| "epoch": 3.8527859992278986, |
| "grad_norm": 1.1566778421401978, |
| "learning_rate": 1.148005148005148e-05, |
| "loss": 0.4089, |
| "step": 14970 |
| }, |
| { |
| "epoch": 3.8553596705700683, |
| "grad_norm": 1.3538066148757935, |
| "learning_rate": 1.1454311454311454e-05, |
| "loss": 0.4617, |
| "step": 14980 |
| }, |
| { |
| "epoch": 3.8579333419122377, |
| "grad_norm": 1.1044282913208008, |
| "learning_rate": 1.1428571428571429e-05, |
| "loss": 0.4734, |
| "step": 14990 |
| }, |
| { |
| "epoch": 3.8605070132544075, |
| "grad_norm": 1.195326805114746, |
| "learning_rate": 1.1402831402831402e-05, |
| "loss": 0.4171, |
| "step": 15000 |
| }, |
| { |
| "epoch": 3.8605070132544075, |
| "eval_loss": 0.8098340630531311, |
| "eval_runtime": 395.6098, |
| "eval_samples_per_second": 49.104, |
| "eval_steps_per_second": 2.457, |
| "eval_token_accuracy": 0.0003602365642010644, |
| "step": 15000 |
| }, |
| { |
| "epoch": 3.863080684596577, |
| "grad_norm": 3.7595245838165283, |
| "learning_rate": 1.1377091377091377e-05, |
| "loss": 0.4476, |
| "step": 15010 |
| }, |
| { |
| "epoch": 3.8656543559387466, |
| "grad_norm": 1.0867187976837158, |
| "learning_rate": 1.1351351351351352e-05, |
| "loss": 0.4016, |
| "step": 15020 |
| }, |
| { |
| "epoch": 3.8682280272809164, |
| "grad_norm": 1.2275842428207397, |
| "learning_rate": 1.1325611325611327e-05, |
| "loss": 0.4635, |
| "step": 15030 |
| }, |
| { |
| "epoch": 3.8708016986230858, |
| "grad_norm": 2.140843629837036, |
| "learning_rate": 1.12998712998713e-05, |
| "loss": 0.4431, |
| "step": 15040 |
| }, |
| { |
| "epoch": 3.8733753699652556, |
| "grad_norm": 5.140436172485352, |
| "learning_rate": 1.1274131274131276e-05, |
| "loss": 0.5156, |
| "step": 15050 |
| }, |
| { |
| "epoch": 3.875949041307425, |
| "grad_norm": 1.347815990447998, |
| "learning_rate": 1.1248391248391249e-05, |
| "loss": 0.4836, |
| "step": 15060 |
| }, |
| { |
| "epoch": 3.8785227126495947, |
| "grad_norm": 0.9270965456962585, |
| "learning_rate": 1.1222651222651224e-05, |
| "loss": 0.4724, |
| "step": 15070 |
| }, |
| { |
| "epoch": 3.881096383991764, |
| "grad_norm": 3.1542866230010986, |
| "learning_rate": 1.1196911196911197e-05, |
| "loss": 0.4842, |
| "step": 15080 |
| }, |
| { |
| "epoch": 3.883670055333934, |
| "grad_norm": 3.4730048179626465, |
| "learning_rate": 1.1171171171171172e-05, |
| "loss": 0.4559, |
| "step": 15090 |
| }, |
| { |
| "epoch": 3.8862437266761036, |
| "grad_norm": 1.7973135709762573, |
| "learning_rate": 1.1145431145431146e-05, |
| "loss": 0.4053, |
| "step": 15100 |
| }, |
| { |
| "epoch": 3.888817398018273, |
| "grad_norm": 1.1773796081542969, |
| "learning_rate": 1.1119691119691119e-05, |
| "loss": 0.3808, |
| "step": 15110 |
| }, |
| { |
| "epoch": 3.8913910693604428, |
| "grad_norm": 2.933912754058838, |
| "learning_rate": 1.1093951093951096e-05, |
| "loss": 0.4514, |
| "step": 15120 |
| }, |
| { |
| "epoch": 3.893964740702612, |
| "grad_norm": 1.0887664556503296, |
| "learning_rate": 1.1068211068211069e-05, |
| "loss": 0.4728, |
| "step": 15130 |
| }, |
| { |
| "epoch": 3.896538412044782, |
| "grad_norm": 1.090585708618164, |
| "learning_rate": 1.1042471042471044e-05, |
| "loss": 0.45, |
| "step": 15140 |
| }, |
| { |
| "epoch": 3.8991120833869513, |
| "grad_norm": 2.140028238296509, |
| "learning_rate": 1.1016731016731017e-05, |
| "loss": 0.4672, |
| "step": 15150 |
| }, |
| { |
| "epoch": 3.901685754729121, |
| "grad_norm": 0.912236213684082, |
| "learning_rate": 1.0990990990990992e-05, |
| "loss": 0.4462, |
| "step": 15160 |
| }, |
| { |
| "epoch": 3.904259426071291, |
| "grad_norm": 2.3811190128326416, |
| "learning_rate": 1.0965250965250965e-05, |
| "loss": 0.446, |
| "step": 15170 |
| }, |
| { |
| "epoch": 3.90683309741346, |
| "grad_norm": 1.1356325149536133, |
| "learning_rate": 1.093951093951094e-05, |
| "loss": 0.4317, |
| "step": 15180 |
| }, |
| { |
| "epoch": 3.90940676875563, |
| "grad_norm": 1.289565920829773, |
| "learning_rate": 1.0913770913770914e-05, |
| "loss": 0.4078, |
| "step": 15190 |
| }, |
| { |
| "epoch": 3.9119804400977998, |
| "grad_norm": 0.9899163842201233, |
| "learning_rate": 1.0888030888030889e-05, |
| "loss": 0.4366, |
| "step": 15200 |
| }, |
| { |
| "epoch": 3.914554111439969, |
| "grad_norm": 1.4149761199951172, |
| "learning_rate": 1.0862290862290862e-05, |
| "loss": 0.5163, |
| "step": 15210 |
| }, |
| { |
| "epoch": 3.9171277827821385, |
| "grad_norm": 1.1276021003723145, |
| "learning_rate": 1.0836550836550837e-05, |
| "loss": 0.4242, |
| "step": 15220 |
| }, |
| { |
| "epoch": 3.9197014541243083, |
| "grad_norm": 1.067687749862671, |
| "learning_rate": 1.0810810810810812e-05, |
| "loss": 0.4305, |
| "step": 15230 |
| }, |
| { |
| "epoch": 3.922275125466478, |
| "grad_norm": 2.6914680004119873, |
| "learning_rate": 1.0785070785070785e-05, |
| "loss": 0.3787, |
| "step": 15240 |
| }, |
| { |
| "epoch": 3.9248487968086474, |
| "grad_norm": 1.443280816078186, |
| "learning_rate": 1.075933075933076e-05, |
| "loss": 0.4043, |
| "step": 15250 |
| }, |
| { |
| "epoch": 3.927422468150817, |
| "grad_norm": 3.6616697311401367, |
| "learning_rate": 1.0733590733590734e-05, |
| "loss": 0.4174, |
| "step": 15260 |
| }, |
| { |
| "epoch": 3.929996139492987, |
| "grad_norm": 1.0795118808746338, |
| "learning_rate": 1.0707850707850709e-05, |
| "loss": 0.443, |
| "step": 15270 |
| }, |
| { |
| "epoch": 3.9325698108351563, |
| "grad_norm": 1.2405767440795898, |
| "learning_rate": 1.0682110682110682e-05, |
| "loss": 0.401, |
| "step": 15280 |
| }, |
| { |
| "epoch": 3.9351434821773257, |
| "grad_norm": 1.446061611175537, |
| "learning_rate": 1.0656370656370657e-05, |
| "loss": 0.4593, |
| "step": 15290 |
| }, |
| { |
| "epoch": 3.9377171535194955, |
| "grad_norm": 4.733116626739502, |
| "learning_rate": 1.063063063063063e-05, |
| "loss": 0.4237, |
| "step": 15300 |
| }, |
| { |
| "epoch": 3.9402908248616653, |
| "grad_norm": 1.0354984998703003, |
| "learning_rate": 1.0604890604890605e-05, |
| "loss": 0.4924, |
| "step": 15310 |
| }, |
| { |
| "epoch": 3.9428644962038346, |
| "grad_norm": 4.033681392669678, |
| "learning_rate": 1.0579150579150579e-05, |
| "loss": 0.4451, |
| "step": 15320 |
| }, |
| { |
| "epoch": 3.9454381675460044, |
| "grad_norm": 1.3079330921173096, |
| "learning_rate": 1.0553410553410555e-05, |
| "loss": 0.4257, |
| "step": 15330 |
| }, |
| { |
| "epoch": 3.948011838888174, |
| "grad_norm": 5.547299385070801, |
| "learning_rate": 1.0527670527670529e-05, |
| "loss": 0.4284, |
| "step": 15340 |
| }, |
| { |
| "epoch": 3.9505855102303435, |
| "grad_norm": 1.283423900604248, |
| "learning_rate": 1.0501930501930504e-05, |
| "loss": 0.4863, |
| "step": 15350 |
| }, |
| { |
| "epoch": 3.9531591815725133, |
| "grad_norm": 1.2620773315429688, |
| "learning_rate": 1.0476190476190477e-05, |
| "loss": 0.4162, |
| "step": 15360 |
| }, |
| { |
| "epoch": 3.9557328529146827, |
| "grad_norm": 1.2353228330612183, |
| "learning_rate": 1.0450450450450452e-05, |
| "loss": 0.4916, |
| "step": 15370 |
| }, |
| { |
| "epoch": 3.9583065242568525, |
| "grad_norm": 1.3218683004379272, |
| "learning_rate": 1.0424710424710425e-05, |
| "loss": 0.3846, |
| "step": 15380 |
| }, |
| { |
| "epoch": 3.960880195599022, |
| "grad_norm": 0.9560413956642151, |
| "learning_rate": 1.0398970398970398e-05, |
| "loss": 0.4208, |
| "step": 15390 |
| }, |
| { |
| "epoch": 3.9634538669411916, |
| "grad_norm": 1.0743191242218018, |
| "learning_rate": 1.0373230373230373e-05, |
| "loss": 0.4151, |
| "step": 15400 |
| }, |
| { |
| "epoch": 3.9660275382833614, |
| "grad_norm": 2.8277199268341064, |
| "learning_rate": 1.0347490347490347e-05, |
| "loss": 0.4162, |
| "step": 15410 |
| }, |
| { |
| "epoch": 3.9686012096255308, |
| "grad_norm": 1.490350604057312, |
| "learning_rate": 1.0321750321750322e-05, |
| "loss": 0.486, |
| "step": 15420 |
| }, |
| { |
| "epoch": 3.9711748809677005, |
| "grad_norm": 1.0201480388641357, |
| "learning_rate": 1.0296010296010295e-05, |
| "loss": 0.4567, |
| "step": 15430 |
| }, |
| { |
| "epoch": 3.97374855230987, |
| "grad_norm": 0.8804197311401367, |
| "learning_rate": 1.0270270270270272e-05, |
| "loss": 0.4716, |
| "step": 15440 |
| }, |
| { |
| "epoch": 3.9763222236520397, |
| "grad_norm": 3.7693941593170166, |
| "learning_rate": 1.0244530244530245e-05, |
| "loss": 0.478, |
| "step": 15450 |
| }, |
| { |
| "epoch": 3.978895894994209, |
| "grad_norm": 5.499643325805664, |
| "learning_rate": 1.021879021879022e-05, |
| "loss": 0.4223, |
| "step": 15460 |
| }, |
| { |
| "epoch": 3.981469566336379, |
| "grad_norm": 0.8914560675621033, |
| "learning_rate": 1.0193050193050193e-05, |
| "loss": 0.4079, |
| "step": 15470 |
| }, |
| { |
| "epoch": 3.9840432376785486, |
| "grad_norm": 1.4804257154464722, |
| "learning_rate": 1.0167310167310168e-05, |
| "loss": 0.4416, |
| "step": 15480 |
| }, |
| { |
| "epoch": 3.986616909020718, |
| "grad_norm": 1.3806978464126587, |
| "learning_rate": 1.0141570141570142e-05, |
| "loss": 0.4259, |
| "step": 15490 |
| }, |
| { |
| "epoch": 3.9891905803628878, |
| "grad_norm": 5.6986212730407715, |
| "learning_rate": 1.0115830115830117e-05, |
| "loss": 0.458, |
| "step": 15500 |
| }, |
| { |
| "epoch": 3.991764251705057, |
| "grad_norm": 1.4306472539901733, |
| "learning_rate": 1.009009009009009e-05, |
| "loss": 0.4718, |
| "step": 15510 |
| }, |
| { |
| "epoch": 3.994337923047227, |
| "grad_norm": 0.9893137216567993, |
| "learning_rate": 1.0064350064350065e-05, |
| "loss": 0.3794, |
| "step": 15520 |
| }, |
| { |
| "epoch": 3.9969115943893962, |
| "grad_norm": 1.7962301969528198, |
| "learning_rate": 1.0038610038610038e-05, |
| "loss": 0.4687, |
| "step": 15530 |
| }, |
| { |
| "epoch": 3.999485265731566, |
| "grad_norm": 0.967959463596344, |
| "learning_rate": 1.0012870012870013e-05, |
| "loss": 0.4552, |
| "step": 15540 |
| }, |
| { |
| "epoch": 4.002058937073736, |
| "grad_norm": 0.9137493371963501, |
| "learning_rate": 9.987129987129988e-06, |
| "loss": 0.3108, |
| "step": 15550 |
| }, |
| { |
| "epoch": 4.004632608415905, |
| "grad_norm": 1.2810945510864258, |
| "learning_rate": 9.961389961389962e-06, |
| "loss": 0.3835, |
| "step": 15560 |
| }, |
| { |
| "epoch": 4.0072062797580745, |
| "grad_norm": 0.8937966227531433, |
| "learning_rate": 9.935649935649937e-06, |
| "loss": 0.3403, |
| "step": 15570 |
| }, |
| { |
| "epoch": 4.009779951100245, |
| "grad_norm": 1.049453616142273, |
| "learning_rate": 9.90990990990991e-06, |
| "loss": 0.3809, |
| "step": 15580 |
| }, |
| { |
| "epoch": 4.012353622442414, |
| "grad_norm": 1.1060616970062256, |
| "learning_rate": 9.884169884169885e-06, |
| "loss": 0.3363, |
| "step": 15590 |
| }, |
| { |
| "epoch": 4.0149272937845835, |
| "grad_norm": 0.9460583329200745, |
| "learning_rate": 9.858429858429858e-06, |
| "loss": 0.3511, |
| "step": 15600 |
| }, |
| { |
| "epoch": 4.017500965126754, |
| "grad_norm": 0.9251907467842102, |
| "learning_rate": 9.832689832689833e-06, |
| "loss": 0.3859, |
| "step": 15610 |
| }, |
| { |
| "epoch": 4.020074636468923, |
| "grad_norm": 1.1467247009277344, |
| "learning_rate": 9.806949806949807e-06, |
| "loss": 0.3202, |
| "step": 15620 |
| }, |
| { |
| "epoch": 4.022648307811092, |
| "grad_norm": 1.3818519115447998, |
| "learning_rate": 9.781209781209782e-06, |
| "loss": 0.3372, |
| "step": 15630 |
| }, |
| { |
| "epoch": 4.025221979153262, |
| "grad_norm": 0.9987813234329224, |
| "learning_rate": 9.755469755469757e-06, |
| "loss": 0.3872, |
| "step": 15640 |
| }, |
| { |
| "epoch": 4.027795650495432, |
| "grad_norm": 0.863815188407898, |
| "learning_rate": 9.729729729729732e-06, |
| "loss": 0.3444, |
| "step": 15650 |
| }, |
| { |
| "epoch": 4.030369321837601, |
| "grad_norm": 1.5916346311569214, |
| "learning_rate": 9.703989703989705e-06, |
| "loss": 0.3448, |
| "step": 15660 |
| }, |
| { |
| "epoch": 4.032942993179771, |
| "grad_norm": 0.9141498804092407, |
| "learning_rate": 9.678249678249678e-06, |
| "loss": 0.3513, |
| "step": 15670 |
| }, |
| { |
| "epoch": 4.035516664521941, |
| "grad_norm": 0.990973174571991, |
| "learning_rate": 9.652509652509653e-06, |
| "loss": 0.3584, |
| "step": 15680 |
| }, |
| { |
| "epoch": 4.03809033586411, |
| "grad_norm": 1.2190366983413696, |
| "learning_rate": 9.626769626769626e-06, |
| "loss": 0.3294, |
| "step": 15690 |
| }, |
| { |
| "epoch": 4.04066400720628, |
| "grad_norm": 1.7479588985443115, |
| "learning_rate": 9.601029601029601e-06, |
| "loss": 0.3963, |
| "step": 15700 |
| }, |
| { |
| "epoch": 4.043237678548449, |
| "grad_norm": 1.09687340259552, |
| "learning_rate": 9.575289575289575e-06, |
| "loss": 0.3238, |
| "step": 15710 |
| }, |
| { |
| "epoch": 4.045811349890619, |
| "grad_norm": 0.9358935356140137, |
| "learning_rate": 9.54954954954955e-06, |
| "loss": 0.345, |
| "step": 15720 |
| }, |
| { |
| "epoch": 4.0483850212327885, |
| "grad_norm": 3.78627872467041, |
| "learning_rate": 9.523809523809523e-06, |
| "loss": 0.3594, |
| "step": 15730 |
| }, |
| { |
| "epoch": 4.050958692574958, |
| "grad_norm": 3.2878758907318115, |
| "learning_rate": 9.498069498069498e-06, |
| "loss": 0.3389, |
| "step": 15740 |
| }, |
| { |
| "epoch": 4.053532363917128, |
| "grad_norm": 1.448227882385254, |
| "learning_rate": 9.472329472329473e-06, |
| "loss": 0.3344, |
| "step": 15750 |
| }, |
| { |
| "epoch": 4.0561060352592975, |
| "grad_norm": 1.4450225830078125, |
| "learning_rate": 9.446589446589448e-06, |
| "loss": 0.3224, |
| "step": 15760 |
| }, |
| { |
| "epoch": 4.058679706601467, |
| "grad_norm": 1.051365852355957, |
| "learning_rate": 9.420849420849421e-06, |
| "loss": 0.2675, |
| "step": 15770 |
| }, |
| { |
| "epoch": 4.061253377943636, |
| "grad_norm": 3.0001602172851562, |
| "learning_rate": 9.395109395109396e-06, |
| "loss": 0.353, |
| "step": 15780 |
| }, |
| { |
| "epoch": 4.063827049285806, |
| "grad_norm": 1.0692496299743652, |
| "learning_rate": 9.36936936936937e-06, |
| "loss": 0.3422, |
| "step": 15790 |
| }, |
| { |
| "epoch": 4.066400720627976, |
| "grad_norm": 1.7115063667297363, |
| "learning_rate": 9.343629343629345e-06, |
| "loss": 0.4045, |
| "step": 15800 |
| }, |
| { |
| "epoch": 4.068974391970145, |
| "grad_norm": 4.681840896606445, |
| "learning_rate": 9.317889317889318e-06, |
| "loss": 0.3603, |
| "step": 15810 |
| }, |
| { |
| "epoch": 4.071548063312315, |
| "grad_norm": 1.7194322347640991, |
| "learning_rate": 9.292149292149293e-06, |
| "loss": 0.3498, |
| "step": 15820 |
| }, |
| { |
| "epoch": 4.074121734654485, |
| "grad_norm": 0.9989597797393799, |
| "learning_rate": 9.266409266409266e-06, |
| "loss": 0.3276, |
| "step": 15830 |
| }, |
| { |
| "epoch": 4.076695405996654, |
| "grad_norm": 0.9685651659965515, |
| "learning_rate": 9.240669240669241e-06, |
| "loss": 0.316, |
| "step": 15840 |
| }, |
| { |
| "epoch": 4.079269077338824, |
| "grad_norm": 1.5660096406936646, |
| "learning_rate": 9.214929214929216e-06, |
| "loss": 0.3929, |
| "step": 15850 |
| }, |
| { |
| "epoch": 4.081842748680994, |
| "grad_norm": 1.4593199491500854, |
| "learning_rate": 9.18918918918919e-06, |
| "loss": 0.3564, |
| "step": 15860 |
| }, |
| { |
| "epoch": 4.084416420023163, |
| "grad_norm": 1.3041324615478516, |
| "learning_rate": 9.163449163449165e-06, |
| "loss": 0.2997, |
| "step": 15870 |
| }, |
| { |
| "epoch": 4.086990091365332, |
| "grad_norm": 1.8447177410125732, |
| "learning_rate": 9.137709137709138e-06, |
| "loss": 0.3274, |
| "step": 15880 |
| }, |
| { |
| "epoch": 4.0895637627075025, |
| "grad_norm": 3.268892765045166, |
| "learning_rate": 9.111969111969113e-06, |
| "loss": 0.3505, |
| "step": 15890 |
| }, |
| { |
| "epoch": 4.092137434049672, |
| "grad_norm": 1.6105279922485352, |
| "learning_rate": 9.086229086229086e-06, |
| "loss": 0.3177, |
| "step": 15900 |
| }, |
| { |
| "epoch": 4.094711105391841, |
| "grad_norm": 1.8889609575271606, |
| "learning_rate": 9.060489060489061e-06, |
| "loss": 0.3904, |
| "step": 15910 |
| }, |
| { |
| "epoch": 4.0972847767340115, |
| "grad_norm": 1.7368106842041016, |
| "learning_rate": 9.034749034749034e-06, |
| "loss": 0.339, |
| "step": 15920 |
| }, |
| { |
| "epoch": 4.099858448076181, |
| "grad_norm": 1.3504408597946167, |
| "learning_rate": 9.00900900900901e-06, |
| "loss": 0.3543, |
| "step": 15930 |
| }, |
| { |
| "epoch": 4.10243211941835, |
| "grad_norm": 1.0322836637496948, |
| "learning_rate": 8.983268983268983e-06, |
| "loss": 0.3498, |
| "step": 15940 |
| }, |
| { |
| "epoch": 4.1050057907605195, |
| "grad_norm": 1.1150479316711426, |
| "learning_rate": 8.957528957528958e-06, |
| "loss": 0.3504, |
| "step": 15950 |
| }, |
| { |
| "epoch": 4.10757946210269, |
| "grad_norm": 0.9721202254295349, |
| "learning_rate": 8.931788931788933e-06, |
| "loss": 0.3605, |
| "step": 15960 |
| }, |
| { |
| "epoch": 4.110153133444859, |
| "grad_norm": 1.3478281497955322, |
| "learning_rate": 8.906048906048906e-06, |
| "loss": 0.322, |
| "step": 15970 |
| }, |
| { |
| "epoch": 4.1127268047870285, |
| "grad_norm": 3.2121946811676025, |
| "learning_rate": 8.880308880308881e-06, |
| "loss": 0.3216, |
| "step": 15980 |
| }, |
| { |
| "epoch": 4.115300476129199, |
| "grad_norm": 1.280942440032959, |
| "learning_rate": 8.854568854568854e-06, |
| "loss": 0.3064, |
| "step": 15990 |
| }, |
| { |
| "epoch": 4.117874147471368, |
| "grad_norm": 1.1329562664031982, |
| "learning_rate": 8.82882882882883e-06, |
| "loss": 0.3252, |
| "step": 16000 |
| }, |
| { |
| "epoch": 4.117874147471368, |
| "eval_loss": 0.8655849695205688, |
| "eval_runtime": 395.6994, |
| "eval_samples_per_second": 49.093, |
| "eval_steps_per_second": 2.456, |
| "eval_token_accuracy": 0.0003534038369559441, |
| "step": 16000 |
| }, |
| { |
| "epoch": 4.120447818813537, |
| "grad_norm": 1.2837986946105957, |
| "learning_rate": 8.803088803088803e-06, |
| "loss": 0.2988, |
| "step": 16010 |
| }, |
| { |
| "epoch": 4.123021490155707, |
| "grad_norm": 1.140343189239502, |
| "learning_rate": 8.777348777348778e-06, |
| "loss": 0.3398, |
| "step": 16020 |
| }, |
| { |
| "epoch": 4.125595161497877, |
| "grad_norm": 1.1737910509109497, |
| "learning_rate": 8.751608751608751e-06, |
| "loss": 0.3676, |
| "step": 16030 |
| }, |
| { |
| "epoch": 4.128168832840046, |
| "grad_norm": 1.1342344284057617, |
| "learning_rate": 8.725868725868726e-06, |
| "loss": 0.3281, |
| "step": 16040 |
| }, |
| { |
| "epoch": 4.130742504182216, |
| "grad_norm": 1.363948941230774, |
| "learning_rate": 8.7001287001287e-06, |
| "loss": 0.3137, |
| "step": 16050 |
| }, |
| { |
| "epoch": 4.133316175524386, |
| "grad_norm": 0.9117937684059143, |
| "learning_rate": 8.674388674388676e-06, |
| "loss": 0.3463, |
| "step": 16060 |
| }, |
| { |
| "epoch": 4.135889846866555, |
| "grad_norm": 4.627786636352539, |
| "learning_rate": 8.64864864864865e-06, |
| "loss": 0.3507, |
| "step": 16070 |
| }, |
| { |
| "epoch": 4.138463518208725, |
| "grad_norm": 1.0469121932983398, |
| "learning_rate": 8.622908622908624e-06, |
| "loss": 0.3416, |
| "step": 16080 |
| }, |
| { |
| "epoch": 4.141037189550894, |
| "grad_norm": 4.487536907196045, |
| "learning_rate": 8.597168597168598e-06, |
| "loss": 0.3084, |
| "step": 16090 |
| }, |
| { |
| "epoch": 4.143610860893064, |
| "grad_norm": 3.589993953704834, |
| "learning_rate": 8.571428571428573e-06, |
| "loss": 0.3498, |
| "step": 16100 |
| }, |
| { |
| "epoch": 4.1461845322352335, |
| "grad_norm": 0.8773334622383118, |
| "learning_rate": 8.545688545688546e-06, |
| "loss": 0.3859, |
| "step": 16110 |
| }, |
| { |
| "epoch": 4.148758203577403, |
| "grad_norm": 1.286348819732666, |
| "learning_rate": 8.519948519948521e-06, |
| "loss": 0.2771, |
| "step": 16120 |
| }, |
| { |
| "epoch": 4.151331874919573, |
| "grad_norm": 1.3166707754135132, |
| "learning_rate": 8.494208494208494e-06, |
| "loss": 0.4015, |
| "step": 16130 |
| }, |
| { |
| "epoch": 4.1539055462617425, |
| "grad_norm": 4.864274978637695, |
| "learning_rate": 8.46846846846847e-06, |
| "loss": 0.3171, |
| "step": 16140 |
| }, |
| { |
| "epoch": 4.156479217603912, |
| "grad_norm": 1.0414588451385498, |
| "learning_rate": 8.442728442728443e-06, |
| "loss": 0.3396, |
| "step": 16150 |
| }, |
| { |
| "epoch": 4.159052888946081, |
| "grad_norm": 5.264183521270752, |
| "learning_rate": 8.416988416988418e-06, |
| "loss": 0.3569, |
| "step": 16160 |
| }, |
| { |
| "epoch": 4.161626560288251, |
| "grad_norm": 1.0647825002670288, |
| "learning_rate": 8.391248391248393e-06, |
| "loss": 0.3317, |
| "step": 16170 |
| }, |
| { |
| "epoch": 4.164200231630421, |
| "grad_norm": 0.8864617347717285, |
| "learning_rate": 8.365508365508366e-06, |
| "loss": 0.3493, |
| "step": 16180 |
| }, |
| { |
| "epoch": 4.16677390297259, |
| "grad_norm": 1.1304059028625488, |
| "learning_rate": 8.33976833976834e-06, |
| "loss": 0.3038, |
| "step": 16190 |
| }, |
| { |
| "epoch": 4.16934757431476, |
| "grad_norm": 1.0398166179656982, |
| "learning_rate": 8.314028314028314e-06, |
| "loss": 0.3168, |
| "step": 16200 |
| }, |
| { |
| "epoch": 4.17192124565693, |
| "grad_norm": 0.8544594645500183, |
| "learning_rate": 8.288288288288289e-06, |
| "loss": 0.3365, |
| "step": 16210 |
| }, |
| { |
| "epoch": 4.174494916999099, |
| "grad_norm": 0.9230920672416687, |
| "learning_rate": 8.262548262548262e-06, |
| "loss": 0.3184, |
| "step": 16220 |
| }, |
| { |
| "epoch": 4.177068588341269, |
| "grad_norm": 6.984750270843506, |
| "learning_rate": 8.236808236808237e-06, |
| "loss": 0.3553, |
| "step": 16230 |
| }, |
| { |
| "epoch": 4.179642259683439, |
| "grad_norm": 4.589465141296387, |
| "learning_rate": 8.21106821106821e-06, |
| "loss": 0.3263, |
| "step": 16240 |
| }, |
| { |
| "epoch": 4.182215931025608, |
| "grad_norm": 1.004537582397461, |
| "learning_rate": 8.185328185328186e-06, |
| "loss": 0.341, |
| "step": 16250 |
| }, |
| { |
| "epoch": 4.184789602367777, |
| "grad_norm": 2.243464231491089, |
| "learning_rate": 8.159588159588159e-06, |
| "loss": 0.3709, |
| "step": 16260 |
| }, |
| { |
| "epoch": 4.1873632737099475, |
| "grad_norm": 1.2939341068267822, |
| "learning_rate": 8.133848133848134e-06, |
| "loss": 0.2929, |
| "step": 16270 |
| }, |
| { |
| "epoch": 4.189936945052117, |
| "grad_norm": 1.1193703413009644, |
| "learning_rate": 8.108108108108109e-06, |
| "loss": 0.3015, |
| "step": 16280 |
| }, |
| { |
| "epoch": 4.192510616394286, |
| "grad_norm": 0.9649505019187927, |
| "learning_rate": 8.082368082368082e-06, |
| "loss": 0.3044, |
| "step": 16290 |
| }, |
| { |
| "epoch": 4.1950842877364565, |
| "grad_norm": 1.2722036838531494, |
| "learning_rate": 8.056628056628057e-06, |
| "loss": 0.3654, |
| "step": 16300 |
| }, |
| { |
| "epoch": 4.197657959078626, |
| "grad_norm": 1.9029432535171509, |
| "learning_rate": 8.03088803088803e-06, |
| "loss": 0.33, |
| "step": 16310 |
| }, |
| { |
| "epoch": 4.200231630420795, |
| "grad_norm": 0.8544790744781494, |
| "learning_rate": 8.005148005148006e-06, |
| "loss": 0.3368, |
| "step": 16320 |
| }, |
| { |
| "epoch": 4.2028053017629645, |
| "grad_norm": 0.991193950176239, |
| "learning_rate": 7.979407979407979e-06, |
| "loss": 0.3242, |
| "step": 16330 |
| }, |
| { |
| "epoch": 4.205378973105135, |
| "grad_norm": 1.923233151435852, |
| "learning_rate": 7.953667953667954e-06, |
| "loss": 0.2868, |
| "step": 16340 |
| }, |
| { |
| "epoch": 4.207952644447304, |
| "grad_norm": 1.3587863445281982, |
| "learning_rate": 7.927927927927927e-06, |
| "loss": 0.3379, |
| "step": 16350 |
| }, |
| { |
| "epoch": 4.2105263157894735, |
| "grad_norm": 0.9470755457878113, |
| "learning_rate": 7.902187902187902e-06, |
| "loss": 0.311, |
| "step": 16360 |
| }, |
| { |
| "epoch": 4.213099987131644, |
| "grad_norm": 1.3828065395355225, |
| "learning_rate": 7.876447876447877e-06, |
| "loss": 0.3017, |
| "step": 16370 |
| }, |
| { |
| "epoch": 4.215673658473813, |
| "grad_norm": 1.4327831268310547, |
| "learning_rate": 7.850707850707852e-06, |
| "loss": 0.3226, |
| "step": 16380 |
| }, |
| { |
| "epoch": 4.218247329815982, |
| "grad_norm": 1.1601290702819824, |
| "learning_rate": 7.824967824967826e-06, |
| "loss": 0.3352, |
| "step": 16390 |
| }, |
| { |
| "epoch": 4.220821001158152, |
| "grad_norm": 1.0657103061676025, |
| "learning_rate": 7.7992277992278e-06, |
| "loss": 0.332, |
| "step": 16400 |
| }, |
| { |
| "epoch": 4.223394672500322, |
| "grad_norm": 3.6614043712615967, |
| "learning_rate": 7.773487773487774e-06, |
| "loss": 0.342, |
| "step": 16410 |
| }, |
| { |
| "epoch": 4.225968343842491, |
| "grad_norm": 1.095182180404663, |
| "learning_rate": 7.747747747747749e-06, |
| "loss": 0.3488, |
| "step": 16420 |
| }, |
| { |
| "epoch": 4.228542015184661, |
| "grad_norm": 1.0206010341644287, |
| "learning_rate": 7.722007722007722e-06, |
| "loss": 0.3571, |
| "step": 16430 |
| }, |
| { |
| "epoch": 4.231115686526831, |
| "grad_norm": 1.0806519985198975, |
| "learning_rate": 7.696267696267697e-06, |
| "loss": 0.3845, |
| "step": 16440 |
| }, |
| { |
| "epoch": 4.233689357869, |
| "grad_norm": 1.2475837469100952, |
| "learning_rate": 7.67052767052767e-06, |
| "loss": 0.3936, |
| "step": 16450 |
| }, |
| { |
| "epoch": 4.23626302921117, |
| "grad_norm": 2.754807233810425, |
| "learning_rate": 7.644787644787644e-06, |
| "loss": 0.3074, |
| "step": 16460 |
| }, |
| { |
| "epoch": 4.238836700553339, |
| "grad_norm": 2.6300132274627686, |
| "learning_rate": 7.6190476190476205e-06, |
| "loss": 0.3422, |
| "step": 16470 |
| }, |
| { |
| "epoch": 4.241410371895509, |
| "grad_norm": 0.9944770932197571, |
| "learning_rate": 7.593307593307595e-06, |
| "loss": 0.3585, |
| "step": 16480 |
| }, |
| { |
| "epoch": 4.2439840432376785, |
| "grad_norm": 0.8388077616691589, |
| "learning_rate": 7.567567567567568e-06, |
| "loss": 0.3416, |
| "step": 16490 |
| }, |
| { |
| "epoch": 4.246557714579848, |
| "grad_norm": 1.02458918094635, |
| "learning_rate": 7.544401544401544e-06, |
| "loss": 0.3858, |
| "step": 16500 |
| }, |
| { |
| "epoch": 4.249131385922018, |
| "grad_norm": 1.3384264707565308, |
| "learning_rate": 7.518661518661518e-06, |
| "loss": 0.3755, |
| "step": 16510 |
| }, |
| { |
| "epoch": 4.2517050572641875, |
| "grad_norm": 0.9491118788719177, |
| "learning_rate": 7.492921492921494e-06, |
| "loss": 0.3789, |
| "step": 16520 |
| }, |
| { |
| "epoch": 4.254278728606357, |
| "grad_norm": 1.3997193574905396, |
| "learning_rate": 7.467181467181468e-06, |
| "loss": 0.364, |
| "step": 16530 |
| }, |
| { |
| "epoch": 4.256852399948526, |
| "grad_norm": 1.488587498664856, |
| "learning_rate": 7.441441441441442e-06, |
| "loss": 0.3153, |
| "step": 16540 |
| }, |
| { |
| "epoch": 4.259426071290696, |
| "grad_norm": 1.079736590385437, |
| "learning_rate": 7.415701415701416e-06, |
| "loss": 0.3447, |
| "step": 16550 |
| }, |
| { |
| "epoch": 4.261999742632866, |
| "grad_norm": 1.142016887664795, |
| "learning_rate": 7.38996138996139e-06, |
| "loss": 0.325, |
| "step": 16560 |
| }, |
| { |
| "epoch": 4.264573413975035, |
| "grad_norm": 0.9195720553398132, |
| "learning_rate": 7.364221364221364e-06, |
| "loss": 0.3222, |
| "step": 16570 |
| }, |
| { |
| "epoch": 4.267147085317205, |
| "grad_norm": 3.8581366539001465, |
| "learning_rate": 7.338481338481338e-06, |
| "loss": 0.3699, |
| "step": 16580 |
| }, |
| { |
| "epoch": 4.269720756659375, |
| "grad_norm": 0.8989285230636597, |
| "learning_rate": 7.3127413127413125e-06, |
| "loss": 0.3438, |
| "step": 16590 |
| }, |
| { |
| "epoch": 4.272294428001544, |
| "grad_norm": 1.3981261253356934, |
| "learning_rate": 7.287001287001287e-06, |
| "loss": 0.3827, |
| "step": 16600 |
| }, |
| { |
| "epoch": 4.274868099343713, |
| "grad_norm": 1.5550109148025513, |
| "learning_rate": 7.261261261261261e-06, |
| "loss": 0.3755, |
| "step": 16610 |
| }, |
| { |
| "epoch": 4.277441770685884, |
| "grad_norm": 1.4705978631973267, |
| "learning_rate": 7.235521235521237e-06, |
| "loss": 0.3258, |
| "step": 16620 |
| }, |
| { |
| "epoch": 4.280015442028053, |
| "grad_norm": 1.0708073377609253, |
| "learning_rate": 7.209781209781211e-06, |
| "loss": 0.3469, |
| "step": 16630 |
| }, |
| { |
| "epoch": 4.282589113370222, |
| "grad_norm": 4.414210796356201, |
| "learning_rate": 7.184041184041185e-06, |
| "loss": 0.3964, |
| "step": 16640 |
| }, |
| { |
| "epoch": 4.2851627847123925, |
| "grad_norm": 1.2815184593200684, |
| "learning_rate": 7.158301158301159e-06, |
| "loss": 0.3284, |
| "step": 16650 |
| }, |
| { |
| "epoch": 4.287736456054562, |
| "grad_norm": 1.2706879377365112, |
| "learning_rate": 7.132561132561133e-06, |
| "loss": 0.3881, |
| "step": 16660 |
| }, |
| { |
| "epoch": 4.290310127396731, |
| "grad_norm": 0.791149914264679, |
| "learning_rate": 7.106821106821107e-06, |
| "loss": 0.33, |
| "step": 16670 |
| }, |
| { |
| "epoch": 4.292883798738901, |
| "grad_norm": 1.350524663925171, |
| "learning_rate": 7.0810810810810815e-06, |
| "loss": 0.4255, |
| "step": 16680 |
| }, |
| { |
| "epoch": 4.295457470081071, |
| "grad_norm": 1.2001677751541138, |
| "learning_rate": 7.055341055341056e-06, |
| "loss": 0.3677, |
| "step": 16690 |
| }, |
| { |
| "epoch": 4.29803114142324, |
| "grad_norm": 3.5927603244781494, |
| "learning_rate": 7.02960102960103e-06, |
| "loss": 0.377, |
| "step": 16700 |
| }, |
| { |
| "epoch": 4.3006048127654095, |
| "grad_norm": 1.3985093832015991, |
| "learning_rate": 7.003861003861004e-06, |
| "loss": 0.3129, |
| "step": 16710 |
| }, |
| { |
| "epoch": 4.30317848410758, |
| "grad_norm": 2.9167826175689697, |
| "learning_rate": 6.978120978120979e-06, |
| "loss": 0.3275, |
| "step": 16720 |
| }, |
| { |
| "epoch": 4.305752155449749, |
| "grad_norm": 5.278055667877197, |
| "learning_rate": 6.952380952380953e-06, |
| "loss": 0.3816, |
| "step": 16730 |
| }, |
| { |
| "epoch": 4.3083258267919184, |
| "grad_norm": 1.1263563632965088, |
| "learning_rate": 6.926640926640927e-06, |
| "loss": 0.3447, |
| "step": 16740 |
| }, |
| { |
| "epoch": 4.310899498134089, |
| "grad_norm": 1.0541788339614868, |
| "learning_rate": 6.9009009009009014e-06, |
| "loss": 0.3533, |
| "step": 16750 |
| }, |
| { |
| "epoch": 4.313473169476258, |
| "grad_norm": 1.7180668115615845, |
| "learning_rate": 6.875160875160876e-06, |
| "loss": 0.3639, |
| "step": 16760 |
| }, |
| { |
| "epoch": 4.316046840818427, |
| "grad_norm": 1.4886486530303955, |
| "learning_rate": 6.84942084942085e-06, |
| "loss": 0.3239, |
| "step": 16770 |
| }, |
| { |
| "epoch": 4.318620512160597, |
| "grad_norm": 1.4927681684494019, |
| "learning_rate": 6.823680823680824e-06, |
| "loss": 0.327, |
| "step": 16780 |
| }, |
| { |
| "epoch": 4.321194183502767, |
| "grad_norm": 1.2550019025802612, |
| "learning_rate": 6.797940797940798e-06, |
| "loss": 0.3321, |
| "step": 16790 |
| }, |
| { |
| "epoch": 4.323767854844936, |
| "grad_norm": 1.218307614326477, |
| "learning_rate": 6.772200772200772e-06, |
| "loss": 0.3482, |
| "step": 16800 |
| }, |
| { |
| "epoch": 4.326341526187106, |
| "grad_norm": 1.1296604871749878, |
| "learning_rate": 6.746460746460746e-06, |
| "loss": 0.3373, |
| "step": 16810 |
| }, |
| { |
| "epoch": 4.328915197529276, |
| "grad_norm": 3.720290184020996, |
| "learning_rate": 6.7207207207207205e-06, |
| "loss": 0.3701, |
| "step": 16820 |
| }, |
| { |
| "epoch": 4.331488868871445, |
| "grad_norm": 1.146220326423645, |
| "learning_rate": 6.6949806949806955e-06, |
| "loss": 0.3682, |
| "step": 16830 |
| }, |
| { |
| "epoch": 4.334062540213615, |
| "grad_norm": 1.438725233078003, |
| "learning_rate": 6.66924066924067e-06, |
| "loss": 0.337, |
| "step": 16840 |
| }, |
| { |
| "epoch": 4.336636211555784, |
| "grad_norm": 1.264733910560608, |
| "learning_rate": 6.643500643500644e-06, |
| "loss": 0.3546, |
| "step": 16850 |
| }, |
| { |
| "epoch": 4.339209882897954, |
| "grad_norm": 1.1695109605789185, |
| "learning_rate": 6.617760617760618e-06, |
| "loss": 0.37, |
| "step": 16860 |
| }, |
| { |
| "epoch": 4.3417835542401235, |
| "grad_norm": 1.0620770454406738, |
| "learning_rate": 6.592020592020592e-06, |
| "loss": 0.3563, |
| "step": 16870 |
| }, |
| { |
| "epoch": 4.344357225582293, |
| "grad_norm": 4.714122772216797, |
| "learning_rate": 6.566280566280566e-06, |
| "loss": 0.3119, |
| "step": 16880 |
| }, |
| { |
| "epoch": 4.346930896924463, |
| "grad_norm": 1.527566909790039, |
| "learning_rate": 6.54054054054054e-06, |
| "loss": 0.3352, |
| "step": 16890 |
| }, |
| { |
| "epoch": 4.3495045682666325, |
| "grad_norm": 1.4223473072052002, |
| "learning_rate": 6.5148005148005146e-06, |
| "loss": 0.3913, |
| "step": 16900 |
| }, |
| { |
| "epoch": 4.352078239608802, |
| "grad_norm": 1.3415018320083618, |
| "learning_rate": 6.489060489060489e-06, |
| "loss": 0.3759, |
| "step": 16910 |
| }, |
| { |
| "epoch": 4.354651910950971, |
| "grad_norm": 1.1462664604187012, |
| "learning_rate": 6.463320463320463e-06, |
| "loss": 0.3822, |
| "step": 16920 |
| }, |
| { |
| "epoch": 4.357225582293141, |
| "grad_norm": 1.2110233306884766, |
| "learning_rate": 6.437580437580439e-06, |
| "loss": 0.3454, |
| "step": 16930 |
| }, |
| { |
| "epoch": 4.359799253635311, |
| "grad_norm": 4.063595294952393, |
| "learning_rate": 6.411840411840413e-06, |
| "loss": 0.3543, |
| "step": 16940 |
| }, |
| { |
| "epoch": 4.36237292497748, |
| "grad_norm": 0.9260187149047852, |
| "learning_rate": 6.386100386100387e-06, |
| "loss": 0.3595, |
| "step": 16950 |
| }, |
| { |
| "epoch": 4.36494659631965, |
| "grad_norm": 1.3427493572235107, |
| "learning_rate": 6.360360360360361e-06, |
| "loss": 0.366, |
| "step": 16960 |
| }, |
| { |
| "epoch": 4.36752026766182, |
| "grad_norm": 1.0595113039016724, |
| "learning_rate": 6.334620334620335e-06, |
| "loss": 0.2813, |
| "step": 16970 |
| }, |
| { |
| "epoch": 4.370093939003989, |
| "grad_norm": 1.5860631465911865, |
| "learning_rate": 6.3088803088803095e-06, |
| "loss": 0.3394, |
| "step": 16980 |
| }, |
| { |
| "epoch": 4.372667610346159, |
| "grad_norm": 1.2586485147476196, |
| "learning_rate": 6.283140283140284e-06, |
| "loss": 0.3595, |
| "step": 16990 |
| }, |
| { |
| "epoch": 4.375241281688329, |
| "grad_norm": 1.0204342603683472, |
| "learning_rate": 6.257400257400258e-06, |
| "loss": 0.3536, |
| "step": 17000 |
| }, |
| { |
| "epoch": 4.375241281688329, |
| "eval_loss": 0.8610925674438477, |
| "eval_runtime": 395.7656, |
| "eval_samples_per_second": 49.085, |
| "eval_steps_per_second": 2.456, |
| "eval_token_accuracy": 0.00034752009960597937, |
| "step": 17000 |
| }, |
| { |
| "epoch": 4.377814953030498, |
| "grad_norm": 0.714625895023346, |
| "learning_rate": 6.231660231660232e-06, |
| "loss": 0.3166, |
| "step": 17010 |
| }, |
| { |
| "epoch": 4.380388624372667, |
| "grad_norm": 1.5543569326400757, |
| "learning_rate": 6.205920205920206e-06, |
| "loss": 0.3606, |
| "step": 17020 |
| }, |
| { |
| "epoch": 4.3829622957148375, |
| "grad_norm": 1.7975174188613892, |
| "learning_rate": 6.18018018018018e-06, |
| "loss": 0.3124, |
| "step": 17030 |
| }, |
| { |
| "epoch": 4.385535967057007, |
| "grad_norm": 4.076536178588867, |
| "learning_rate": 6.154440154440154e-06, |
| "loss": 0.3513, |
| "step": 17040 |
| }, |
| { |
| "epoch": 4.388109638399176, |
| "grad_norm": 1.210050344467163, |
| "learning_rate": 6.1287001287001285e-06, |
| "loss": 0.3522, |
| "step": 17050 |
| }, |
| { |
| "epoch": 4.3906833097413465, |
| "grad_norm": 1.1933529376983643, |
| "learning_rate": 6.1029601029601035e-06, |
| "loss": 0.3474, |
| "step": 17060 |
| }, |
| { |
| "epoch": 4.393256981083516, |
| "grad_norm": 1.305173635482788, |
| "learning_rate": 6.077220077220078e-06, |
| "loss": 0.3271, |
| "step": 17070 |
| }, |
| { |
| "epoch": 4.395830652425685, |
| "grad_norm": 1.0537171363830566, |
| "learning_rate": 6.051480051480052e-06, |
| "loss": 0.3602, |
| "step": 17080 |
| }, |
| { |
| "epoch": 4.3984043237678545, |
| "grad_norm": 1.2690879106521606, |
| "learning_rate": 6.025740025740026e-06, |
| "loss": 0.356, |
| "step": 17090 |
| }, |
| { |
| "epoch": 4.400977995110025, |
| "grad_norm": 1.4055087566375732, |
| "learning_rate": 6e-06, |
| "loss": 0.3896, |
| "step": 17100 |
| }, |
| { |
| "epoch": 4.403551666452194, |
| "grad_norm": 2.5787110328674316, |
| "learning_rate": 5.974259974259975e-06, |
| "loss": 0.3773, |
| "step": 17110 |
| }, |
| { |
| "epoch": 4.406125337794363, |
| "grad_norm": 1.122134804725647, |
| "learning_rate": 5.948519948519949e-06, |
| "loss": 0.3427, |
| "step": 17120 |
| }, |
| { |
| "epoch": 4.408699009136534, |
| "grad_norm": 0.8836435675621033, |
| "learning_rate": 5.9227799227799235e-06, |
| "loss": 0.3142, |
| "step": 17130 |
| }, |
| { |
| "epoch": 4.411272680478703, |
| "grad_norm": 1.6835286617279053, |
| "learning_rate": 5.897039897039898e-06, |
| "loss": 0.3993, |
| "step": 17140 |
| }, |
| { |
| "epoch": 4.413846351820872, |
| "grad_norm": 3.622476577758789, |
| "learning_rate": 5.871299871299872e-06, |
| "loss": 0.331, |
| "step": 17150 |
| }, |
| { |
| "epoch": 4.416420023163042, |
| "grad_norm": 1.1817376613616943, |
| "learning_rate": 5.845559845559846e-06, |
| "loss": 0.356, |
| "step": 17160 |
| }, |
| { |
| "epoch": 4.418993694505212, |
| "grad_norm": 1.0736690759658813, |
| "learning_rate": 5.81981981981982e-06, |
| "loss": 0.4044, |
| "step": 17170 |
| }, |
| { |
| "epoch": 4.421567365847381, |
| "grad_norm": 1.3568885326385498, |
| "learning_rate": 5.794079794079794e-06, |
| "loss": 0.3515, |
| "step": 17180 |
| }, |
| { |
| "epoch": 4.424141037189551, |
| "grad_norm": 1.0532563924789429, |
| "learning_rate": 5.768339768339768e-06, |
| "loss": 0.3542, |
| "step": 17190 |
| }, |
| { |
| "epoch": 4.426714708531721, |
| "grad_norm": 1.5962337255477905, |
| "learning_rate": 5.7425997425997425e-06, |
| "loss": 0.3249, |
| "step": 17200 |
| }, |
| { |
| "epoch": 4.42928837987389, |
| "grad_norm": 1.8830443620681763, |
| "learning_rate": 5.716859716859717e-06, |
| "loss": 0.3256, |
| "step": 17210 |
| }, |
| { |
| "epoch": 4.43186205121606, |
| "grad_norm": 1.0833598375320435, |
| "learning_rate": 5.691119691119692e-06, |
| "loss": 0.3783, |
| "step": 17220 |
| }, |
| { |
| "epoch": 4.434435722558229, |
| "grad_norm": 1.3849422931671143, |
| "learning_rate": 5.665379665379666e-06, |
| "loss": 0.3494, |
| "step": 17230 |
| }, |
| { |
| "epoch": 4.437009393900399, |
| "grad_norm": 2.959022045135498, |
| "learning_rate": 5.63963963963964e-06, |
| "loss": 0.3622, |
| "step": 17240 |
| }, |
| { |
| "epoch": 4.4395830652425685, |
| "grad_norm": 0.9271379113197327, |
| "learning_rate": 5.613899613899614e-06, |
| "loss": 0.3807, |
| "step": 17250 |
| }, |
| { |
| "epoch": 4.442156736584738, |
| "grad_norm": 1.0124372243881226, |
| "learning_rate": 5.588159588159588e-06, |
| "loss": 0.3751, |
| "step": 17260 |
| }, |
| { |
| "epoch": 4.444730407926908, |
| "grad_norm": 3.5358119010925293, |
| "learning_rate": 5.562419562419563e-06, |
| "loss": 0.2874, |
| "step": 17270 |
| }, |
| { |
| "epoch": 4.4473040792690774, |
| "grad_norm": 0.9238685369491577, |
| "learning_rate": 5.5366795366795374e-06, |
| "loss": 0.3796, |
| "step": 17280 |
| }, |
| { |
| "epoch": 4.449877750611247, |
| "grad_norm": 1.2194762229919434, |
| "learning_rate": 5.510939510939512e-06, |
| "loss": 0.3643, |
| "step": 17290 |
| }, |
| { |
| "epoch": 4.452451421953416, |
| "grad_norm": 1.2170203924179077, |
| "learning_rate": 5.485199485199486e-06, |
| "loss": 0.3835, |
| "step": 17300 |
| }, |
| { |
| "epoch": 4.455025093295586, |
| "grad_norm": 1.5656147003173828, |
| "learning_rate": 5.45945945945946e-06, |
| "loss": 0.3385, |
| "step": 17310 |
| }, |
| { |
| "epoch": 4.457598764637756, |
| "grad_norm": 1.351452112197876, |
| "learning_rate": 5.433719433719434e-06, |
| "loss": 0.3646, |
| "step": 17320 |
| }, |
| { |
| "epoch": 4.460172435979925, |
| "grad_norm": 1.1364386081695557, |
| "learning_rate": 5.407979407979408e-06, |
| "loss": 0.3132, |
| "step": 17330 |
| }, |
| { |
| "epoch": 4.462746107322095, |
| "grad_norm": 1.2257845401763916, |
| "learning_rate": 5.382239382239382e-06, |
| "loss": 0.3266, |
| "step": 17340 |
| }, |
| { |
| "epoch": 4.465319778664265, |
| "grad_norm": 1.087573766708374, |
| "learning_rate": 5.3564993564993565e-06, |
| "loss": 0.3644, |
| "step": 17350 |
| }, |
| { |
| "epoch": 4.467893450006434, |
| "grad_norm": 1.2101587057113647, |
| "learning_rate": 5.330759330759331e-06, |
| "loss": 0.3456, |
| "step": 17360 |
| }, |
| { |
| "epoch": 4.470467121348603, |
| "grad_norm": 1.4573382139205933, |
| "learning_rate": 5.305019305019306e-06, |
| "loss": 0.3113, |
| "step": 17370 |
| }, |
| { |
| "epoch": 4.473040792690774, |
| "grad_norm": 0.8906600475311279, |
| "learning_rate": 5.27927927927928e-06, |
| "loss": 0.3345, |
| "step": 17380 |
| }, |
| { |
| "epoch": 4.475614464032943, |
| "grad_norm": 1.079119086265564, |
| "learning_rate": 5.253539253539254e-06, |
| "loss": 0.3668, |
| "step": 17390 |
| }, |
| { |
| "epoch": 4.478188135375112, |
| "grad_norm": 1.0174261331558228, |
| "learning_rate": 5.227799227799228e-06, |
| "loss": 0.2963, |
| "step": 17400 |
| }, |
| { |
| "epoch": 4.4807618067172825, |
| "grad_norm": 5.470022678375244, |
| "learning_rate": 5.202059202059202e-06, |
| "loss": 0.39, |
| "step": 17410 |
| }, |
| { |
| "epoch": 4.483335478059452, |
| "grad_norm": 3.2277028560638428, |
| "learning_rate": 5.176319176319176e-06, |
| "loss": 0.3342, |
| "step": 17420 |
| }, |
| { |
| "epoch": 4.485909149401621, |
| "grad_norm": 1.0484484434127808, |
| "learning_rate": 5.150579150579151e-06, |
| "loss": 0.3536, |
| "step": 17430 |
| }, |
| { |
| "epoch": 4.488482820743791, |
| "grad_norm": 3.3900933265686035, |
| "learning_rate": 5.1248391248391256e-06, |
| "loss": 0.3409, |
| "step": 17440 |
| }, |
| { |
| "epoch": 4.491056492085961, |
| "grad_norm": 1.2388783693313599, |
| "learning_rate": 5.0990990990991e-06, |
| "loss": 0.3389, |
| "step": 17450 |
| }, |
| { |
| "epoch": 4.49363016342813, |
| "grad_norm": 1.4199916124343872, |
| "learning_rate": 5.073359073359073e-06, |
| "loss": 0.3302, |
| "step": 17460 |
| }, |
| { |
| "epoch": 4.4962038347702995, |
| "grad_norm": 1.3507987260818481, |
| "learning_rate": 5.047619047619047e-06, |
| "loss": 0.367, |
| "step": 17470 |
| }, |
| { |
| "epoch": 4.49877750611247, |
| "grad_norm": 1.1383529901504517, |
| "learning_rate": 5.021879021879022e-06, |
| "loss": 0.3361, |
| "step": 17480 |
| }, |
| { |
| "epoch": 4.501351177454639, |
| "grad_norm": 1.3563300371170044, |
| "learning_rate": 4.996138996138996e-06, |
| "loss": 0.3596, |
| "step": 17490 |
| }, |
| { |
| "epoch": 4.503924848796808, |
| "grad_norm": 1.2805308103561401, |
| "learning_rate": 4.9703989703989705e-06, |
| "loss": 0.338, |
| "step": 17500 |
| }, |
| { |
| "epoch": 4.506498520138978, |
| "grad_norm": 1.186610460281372, |
| "learning_rate": 4.944658944658945e-06, |
| "loss": 0.4018, |
| "step": 17510 |
| }, |
| { |
| "epoch": 4.509072191481148, |
| "grad_norm": 1.3856215476989746, |
| "learning_rate": 4.918918918918919e-06, |
| "loss": 0.3387, |
| "step": 17520 |
| }, |
| { |
| "epoch": 4.511645862823317, |
| "grad_norm": 1.4942715167999268, |
| "learning_rate": 4.893178893178894e-06, |
| "loss": 0.3774, |
| "step": 17530 |
| }, |
| { |
| "epoch": 4.514219534165487, |
| "grad_norm": 1.3784438371658325, |
| "learning_rate": 4.867438867438868e-06, |
| "loss": 0.3103, |
| "step": 17540 |
| }, |
| { |
| "epoch": 4.516793205507657, |
| "grad_norm": 1.6347748041152954, |
| "learning_rate": 4.841698841698842e-06, |
| "loss": 0.3187, |
| "step": 17550 |
| }, |
| { |
| "epoch": 4.519366876849826, |
| "grad_norm": 3.5454697608947754, |
| "learning_rate": 4.815958815958816e-06, |
| "loss": 0.3355, |
| "step": 17560 |
| }, |
| { |
| "epoch": 4.521940548191996, |
| "grad_norm": 1.6713000535964966, |
| "learning_rate": 4.79021879021879e-06, |
| "loss": 0.3521, |
| "step": 17570 |
| }, |
| { |
| "epoch": 4.524514219534165, |
| "grad_norm": 1.2210255861282349, |
| "learning_rate": 4.764478764478765e-06, |
| "loss": 0.3532, |
| "step": 17580 |
| }, |
| { |
| "epoch": 4.527087890876335, |
| "grad_norm": 1.9054685831069946, |
| "learning_rate": 4.7387387387387395e-06, |
| "loss": 0.3597, |
| "step": 17590 |
| }, |
| { |
| "epoch": 4.529661562218505, |
| "grad_norm": 1.692408561706543, |
| "learning_rate": 4.712998712998713e-06, |
| "loss": 0.329, |
| "step": 17600 |
| }, |
| { |
| "epoch": 4.532235233560674, |
| "grad_norm": 1.5581741333007812, |
| "learning_rate": 4.687258687258687e-06, |
| "loss": 0.3693, |
| "step": 17610 |
| }, |
| { |
| "epoch": 4.534808904902844, |
| "grad_norm": 1.223633050918579, |
| "learning_rate": 4.661518661518661e-06, |
| "loss": 0.3536, |
| "step": 17620 |
| }, |
| { |
| "epoch": 4.5373825762450135, |
| "grad_norm": 1.719165325164795, |
| "learning_rate": 4.635778635778636e-06, |
| "loss": 0.3352, |
| "step": 17630 |
| }, |
| { |
| "epoch": 4.539956247587183, |
| "grad_norm": 1.2446945905685425, |
| "learning_rate": 4.61003861003861e-06, |
| "loss": 0.305, |
| "step": 17640 |
| }, |
| { |
| "epoch": 4.542529918929353, |
| "grad_norm": 1.1125750541687012, |
| "learning_rate": 4.5842985842985844e-06, |
| "loss": 0.3379, |
| "step": 17650 |
| }, |
| { |
| "epoch": 4.545103590271522, |
| "grad_norm": 1.183079719543457, |
| "learning_rate": 4.558558558558559e-06, |
| "loss": 0.4016, |
| "step": 17660 |
| }, |
| { |
| "epoch": 4.547677261613692, |
| "grad_norm": 1.196256399154663, |
| "learning_rate": 4.532818532818533e-06, |
| "loss": 0.3503, |
| "step": 17670 |
| }, |
| { |
| "epoch": 4.550250932955862, |
| "grad_norm": 2.58664608001709, |
| "learning_rate": 4.507078507078507e-06, |
| "loss": 0.3455, |
| "step": 17680 |
| }, |
| { |
| "epoch": 4.552824604298031, |
| "grad_norm": 1.1959803104400635, |
| "learning_rate": 4.481338481338482e-06, |
| "loss": 0.3916, |
| "step": 17690 |
| }, |
| { |
| "epoch": 4.555398275640201, |
| "grad_norm": 0.9203136563301086, |
| "learning_rate": 4.455598455598456e-06, |
| "loss": 0.3098, |
| "step": 17700 |
| }, |
| { |
| "epoch": 4.55797194698237, |
| "grad_norm": 0.8788485527038574, |
| "learning_rate": 4.42985842985843e-06, |
| "loss": 0.3033, |
| "step": 17710 |
| }, |
| { |
| "epoch": 4.56054561832454, |
| "grad_norm": 0.9711852669715881, |
| "learning_rate": 4.404118404118404e-06, |
| "loss": 0.3704, |
| "step": 17720 |
| }, |
| { |
| "epoch": 4.56311928966671, |
| "grad_norm": 0.947658360004425, |
| "learning_rate": 4.3783783783783785e-06, |
| "loss": 0.2983, |
| "step": 17730 |
| }, |
| { |
| "epoch": 4.565692961008879, |
| "grad_norm": 0.8453003168106079, |
| "learning_rate": 4.3526383526383535e-06, |
| "loss": 0.2935, |
| "step": 17740 |
| }, |
| { |
| "epoch": 4.568266632351049, |
| "grad_norm": 1.273728370666504, |
| "learning_rate": 4.326898326898327e-06, |
| "loss": 0.3509, |
| "step": 17750 |
| }, |
| { |
| "epoch": 4.570840303693219, |
| "grad_norm": 1.0643731355667114, |
| "learning_rate": 4.301158301158301e-06, |
| "loss": 0.3753, |
| "step": 17760 |
| }, |
| { |
| "epoch": 4.573413975035388, |
| "grad_norm": 1.152840256690979, |
| "learning_rate": 4.275418275418275e-06, |
| "loss": 0.3816, |
| "step": 17770 |
| }, |
| { |
| "epoch": 4.575987646377557, |
| "grad_norm": 1.0938299894332886, |
| "learning_rate": 4.249678249678249e-06, |
| "loss": 0.3666, |
| "step": 17780 |
| }, |
| { |
| "epoch": 4.5785613177197275, |
| "grad_norm": 1.659525752067566, |
| "learning_rate": 4.223938223938224e-06, |
| "loss": 0.3586, |
| "step": 17790 |
| }, |
| { |
| "epoch": 4.581134989061897, |
| "grad_norm": 1.3359277248382568, |
| "learning_rate": 4.1981981981981984e-06, |
| "loss": 0.3487, |
| "step": 17800 |
| }, |
| { |
| "epoch": 4.583708660404066, |
| "grad_norm": 1.0927700996398926, |
| "learning_rate": 4.1724581724581726e-06, |
| "loss": 0.3639, |
| "step": 17810 |
| }, |
| { |
| "epoch": 4.5862823317462365, |
| "grad_norm": 0.907636284828186, |
| "learning_rate": 4.146718146718147e-06, |
| "loss": 0.3088, |
| "step": 17820 |
| }, |
| { |
| "epoch": 4.588856003088406, |
| "grad_norm": 1.650233268737793, |
| "learning_rate": 4.120978120978121e-06, |
| "loss": 0.3611, |
| "step": 17830 |
| }, |
| { |
| "epoch": 4.591429674430575, |
| "grad_norm": 0.9985641837120056, |
| "learning_rate": 4.095238095238096e-06, |
| "loss": 0.3119, |
| "step": 17840 |
| }, |
| { |
| "epoch": 4.5940033457727445, |
| "grad_norm": 1.0793542861938477, |
| "learning_rate": 4.06949806949807e-06, |
| "loss": 0.3594, |
| "step": 17850 |
| }, |
| { |
| "epoch": 4.596577017114915, |
| "grad_norm": 1.0795053243637085, |
| "learning_rate": 4.043758043758044e-06, |
| "loss": 0.3464, |
| "step": 17860 |
| }, |
| { |
| "epoch": 4.599150688457084, |
| "grad_norm": 0.9072697758674622, |
| "learning_rate": 4.018018018018018e-06, |
| "loss": 0.3176, |
| "step": 17870 |
| }, |
| { |
| "epoch": 4.601724359799253, |
| "grad_norm": 2.6680212020874023, |
| "learning_rate": 3.9922779922779925e-06, |
| "loss": 0.337, |
| "step": 17880 |
| }, |
| { |
| "epoch": 4.604298031141424, |
| "grad_norm": 1.0229748487472534, |
| "learning_rate": 3.966537966537967e-06, |
| "loss": 0.3516, |
| "step": 17890 |
| }, |
| { |
| "epoch": 4.606871702483593, |
| "grad_norm": 3.6243858337402344, |
| "learning_rate": 3.940797940797941e-06, |
| "loss": 0.3331, |
| "step": 17900 |
| }, |
| { |
| "epoch": 4.609445373825762, |
| "grad_norm": 1.2276082038879395, |
| "learning_rate": 3.915057915057915e-06, |
| "loss": 0.3215, |
| "step": 17910 |
| }, |
| { |
| "epoch": 4.612019045167932, |
| "grad_norm": 1.2664270401000977, |
| "learning_rate": 3.889317889317889e-06, |
| "loss": 0.3782, |
| "step": 17920 |
| }, |
| { |
| "epoch": 4.614592716510102, |
| "grad_norm": 1.5127272605895996, |
| "learning_rate": 3.863577863577863e-06, |
| "loss": 0.3471, |
| "step": 17930 |
| }, |
| { |
| "epoch": 4.617166387852271, |
| "grad_norm": 1.5340853929519653, |
| "learning_rate": 3.837837837837837e-06, |
| "loss": 0.3593, |
| "step": 17940 |
| }, |
| { |
| "epoch": 4.619740059194441, |
| "grad_norm": 0.8411980867385864, |
| "learning_rate": 3.8120978120978124e-06, |
| "loss": 0.3846, |
| "step": 17950 |
| }, |
| { |
| "epoch": 4.622313730536611, |
| "grad_norm": 3.0642356872558594, |
| "learning_rate": 3.7863577863577866e-06, |
| "loss": 0.372, |
| "step": 17960 |
| }, |
| { |
| "epoch": 4.62488740187878, |
| "grad_norm": 2.399491786956787, |
| "learning_rate": 3.7606177606177607e-06, |
| "loss": 0.3127, |
| "step": 17970 |
| }, |
| { |
| "epoch": 4.62746107322095, |
| "grad_norm": 1.3899736404418945, |
| "learning_rate": 3.734877734877735e-06, |
| "loss": 0.3518, |
| "step": 17980 |
| }, |
| { |
| "epoch": 4.630034744563119, |
| "grad_norm": 2.5960450172424316, |
| "learning_rate": 3.709137709137709e-06, |
| "loss": 0.3246, |
| "step": 17990 |
| }, |
| { |
| "epoch": 4.632608415905289, |
| "grad_norm": 1.2150206565856934, |
| "learning_rate": 3.683397683397684e-06, |
| "loss": 0.3655, |
| "step": 18000 |
| }, |
| { |
| "epoch": 4.632608415905289, |
| "eval_loss": 0.8564029932022095, |
| "eval_runtime": 395.6976, |
| "eval_samples_per_second": 49.093, |
| "eval_steps_per_second": 2.456, |
| "eval_token_accuracy": 0.00035587121068334865, |
| "step": 18000 |
| }, |
| { |
| "epoch": 4.6351820872474585, |
| "grad_norm": 1.110227108001709, |
| "learning_rate": 3.6576576576576577e-06, |
| "loss": 0.3679, |
| "step": 18010 |
| }, |
| { |
| "epoch": 4.637755758589628, |
| "grad_norm": 1.2694969177246094, |
| "learning_rate": 3.6344916344916345e-06, |
| "loss": 0.4168, |
| "step": 18020 |
| }, |
| { |
| "epoch": 4.640329429931798, |
| "grad_norm": 1.5729743242263794, |
| "learning_rate": 3.6087516087516086e-06, |
| "loss": 0.3952, |
| "step": 18030 |
| }, |
| { |
| "epoch": 4.642903101273967, |
| "grad_norm": 1.1754688024520874, |
| "learning_rate": 3.5830115830115836e-06, |
| "loss": 0.3475, |
| "step": 18040 |
| }, |
| { |
| "epoch": 4.645476772616137, |
| "grad_norm": 1.3123219013214111, |
| "learning_rate": 3.5572715572715578e-06, |
| "loss": 0.3589, |
| "step": 18050 |
| }, |
| { |
| "epoch": 4.648050443958306, |
| "grad_norm": 3.629625082015991, |
| "learning_rate": 3.5315315315315315e-06, |
| "loss": 0.3261, |
| "step": 18060 |
| }, |
| { |
| "epoch": 4.650624115300476, |
| "grad_norm": 1.0868263244628906, |
| "learning_rate": 3.5057915057915057e-06, |
| "loss": 0.3427, |
| "step": 18070 |
| }, |
| { |
| "epoch": 4.653197786642646, |
| "grad_norm": 1.1146104335784912, |
| "learning_rate": 3.48005148005148e-06, |
| "loss": 0.3577, |
| "step": 18080 |
| }, |
| { |
| "epoch": 4.655771457984815, |
| "grad_norm": 1.4053925275802612, |
| "learning_rate": 3.454311454311455e-06, |
| "loss": 0.3473, |
| "step": 18090 |
| }, |
| { |
| "epoch": 4.658345129326985, |
| "grad_norm": 1.5852116346359253, |
| "learning_rate": 3.428571428571429e-06, |
| "loss": 0.3616, |
| "step": 18100 |
| }, |
| { |
| "epoch": 4.660918800669155, |
| "grad_norm": 1.1257933378219604, |
| "learning_rate": 3.402831402831403e-06, |
| "loss": 0.3035, |
| "step": 18110 |
| }, |
| { |
| "epoch": 4.663492472011324, |
| "grad_norm": 1.2144465446472168, |
| "learning_rate": 3.3770913770913773e-06, |
| "loss": 0.3089, |
| "step": 18120 |
| }, |
| { |
| "epoch": 4.666066143353493, |
| "grad_norm": 1.458746314048767, |
| "learning_rate": 3.3513513513513514e-06, |
| "loss": 0.3289, |
| "step": 18130 |
| }, |
| { |
| "epoch": 4.668639814695664, |
| "grad_norm": 1.8117784261703491, |
| "learning_rate": 3.325611325611326e-06, |
| "loss": 0.3345, |
| "step": 18140 |
| }, |
| { |
| "epoch": 4.671213486037833, |
| "grad_norm": 1.3290542364120483, |
| "learning_rate": 3.2998712998713e-06, |
| "loss": 0.2998, |
| "step": 18150 |
| }, |
| { |
| "epoch": 4.673787157380002, |
| "grad_norm": 1.1426622867584229, |
| "learning_rate": 3.2741312741312743e-06, |
| "loss": 0.364, |
| "step": 18160 |
| }, |
| { |
| "epoch": 4.6763608287221725, |
| "grad_norm": 1.2317893505096436, |
| "learning_rate": 3.2483912483912484e-06, |
| "loss": 0.2657, |
| "step": 18170 |
| }, |
| { |
| "epoch": 4.678934500064342, |
| "grad_norm": 1.211256980895996, |
| "learning_rate": 3.2226512226512226e-06, |
| "loss": 0.4204, |
| "step": 18180 |
| }, |
| { |
| "epoch": 4.681508171406511, |
| "grad_norm": 1.5022300481796265, |
| "learning_rate": 3.1969111969111968e-06, |
| "loss": 0.3907, |
| "step": 18190 |
| }, |
| { |
| "epoch": 4.684081842748681, |
| "grad_norm": 1.2132724523544312, |
| "learning_rate": 3.1711711711711718e-06, |
| "loss": 0.3386, |
| "step": 18200 |
| }, |
| { |
| "epoch": 4.686655514090851, |
| "grad_norm": 1.2516412734985352, |
| "learning_rate": 3.1454311454311455e-06, |
| "loss": 0.3616, |
| "step": 18210 |
| }, |
| { |
| "epoch": 4.68922918543302, |
| "grad_norm": 1.1552441120147705, |
| "learning_rate": 3.1196911196911196e-06, |
| "loss": 0.3508, |
| "step": 18220 |
| }, |
| { |
| "epoch": 4.6918028567751895, |
| "grad_norm": 3.639354705810547, |
| "learning_rate": 3.093951093951094e-06, |
| "loss": 0.3425, |
| "step": 18230 |
| }, |
| { |
| "epoch": 4.69437652811736, |
| "grad_norm": 3.8008406162261963, |
| "learning_rate": 3.0682110682110684e-06, |
| "loss": 0.3468, |
| "step": 18240 |
| }, |
| { |
| "epoch": 4.696950199459529, |
| "grad_norm": 1.2858436107635498, |
| "learning_rate": 3.0424710424710425e-06, |
| "loss": 0.3806, |
| "step": 18250 |
| }, |
| { |
| "epoch": 4.699523870801698, |
| "grad_norm": 2.4241316318511963, |
| "learning_rate": 3.016731016731017e-06, |
| "loss": 0.3471, |
| "step": 18260 |
| }, |
| { |
| "epoch": 4.702097542143868, |
| "grad_norm": 1.0600461959838867, |
| "learning_rate": 2.9909909909909912e-06, |
| "loss": 0.301, |
| "step": 18270 |
| }, |
| { |
| "epoch": 4.704671213486038, |
| "grad_norm": 1.1587661504745483, |
| "learning_rate": 2.9652509652509654e-06, |
| "loss": 0.3476, |
| "step": 18280 |
| }, |
| { |
| "epoch": 4.707244884828207, |
| "grad_norm": 2.059361696243286, |
| "learning_rate": 2.9395109395109395e-06, |
| "loss": 0.3625, |
| "step": 18290 |
| }, |
| { |
| "epoch": 4.709818556170377, |
| "grad_norm": 1.0103809833526611, |
| "learning_rate": 2.9137709137709137e-06, |
| "loss": 0.3285, |
| "step": 18300 |
| }, |
| { |
| "epoch": 4.712392227512547, |
| "grad_norm": 3.1036524772644043, |
| "learning_rate": 2.8880308880308883e-06, |
| "loss": 0.4137, |
| "step": 18310 |
| }, |
| { |
| "epoch": 4.714965898854716, |
| "grad_norm": 1.7626713514328003, |
| "learning_rate": 2.8622908622908624e-06, |
| "loss": 0.3824, |
| "step": 18320 |
| }, |
| { |
| "epoch": 4.717539570196886, |
| "grad_norm": 1.179608941078186, |
| "learning_rate": 2.8365508365508366e-06, |
| "loss": 0.3257, |
| "step": 18330 |
| }, |
| { |
| "epoch": 4.720113241539055, |
| "grad_norm": 1.0469659566879272, |
| "learning_rate": 2.810810810810811e-06, |
| "loss": 0.356, |
| "step": 18340 |
| }, |
| { |
| "epoch": 4.722686912881225, |
| "grad_norm": 1.0116140842437744, |
| "learning_rate": 2.7850707850707853e-06, |
| "loss": 0.3352, |
| "step": 18350 |
| }, |
| { |
| "epoch": 4.725260584223395, |
| "grad_norm": 1.530369520187378, |
| "learning_rate": 2.7593307593307595e-06, |
| "loss": 0.4061, |
| "step": 18360 |
| }, |
| { |
| "epoch": 4.727834255565564, |
| "grad_norm": 1.096043586730957, |
| "learning_rate": 2.7335907335907336e-06, |
| "loss": 0.3633, |
| "step": 18370 |
| }, |
| { |
| "epoch": 4.730407926907734, |
| "grad_norm": 2.9259002208709717, |
| "learning_rate": 2.7078507078507078e-06, |
| "loss": 0.3762, |
| "step": 18380 |
| }, |
| { |
| "epoch": 4.7329815982499035, |
| "grad_norm": 2.655984878540039, |
| "learning_rate": 2.6821106821106823e-06, |
| "loss": 0.331, |
| "step": 18390 |
| }, |
| { |
| "epoch": 4.735555269592073, |
| "grad_norm": 1.2596089839935303, |
| "learning_rate": 2.6563706563706565e-06, |
| "loss": 0.314, |
| "step": 18400 |
| }, |
| { |
| "epoch": 4.738128940934243, |
| "grad_norm": 1.149182915687561, |
| "learning_rate": 2.6306306306306306e-06, |
| "loss": 0.3377, |
| "step": 18410 |
| }, |
| { |
| "epoch": 4.740702612276412, |
| "grad_norm": 0.9628931879997253, |
| "learning_rate": 2.6048906048906052e-06, |
| "loss": 0.404, |
| "step": 18420 |
| }, |
| { |
| "epoch": 4.743276283618582, |
| "grad_norm": 1.304672360420227, |
| "learning_rate": 2.5791505791505794e-06, |
| "loss": 0.3934, |
| "step": 18430 |
| }, |
| { |
| "epoch": 4.745849954960751, |
| "grad_norm": 1.1351431608200073, |
| "learning_rate": 2.5534105534105535e-06, |
| "loss": 0.3549, |
| "step": 18440 |
| }, |
| { |
| "epoch": 4.748423626302921, |
| "grad_norm": 3.7342593669891357, |
| "learning_rate": 2.5276705276705277e-06, |
| "loss": 0.3391, |
| "step": 18450 |
| }, |
| { |
| "epoch": 4.750997297645091, |
| "grad_norm": 1.1187039613723755, |
| "learning_rate": 2.501930501930502e-06, |
| "loss": 0.341, |
| "step": 18460 |
| }, |
| { |
| "epoch": 4.75357096898726, |
| "grad_norm": 1.6478817462921143, |
| "learning_rate": 2.4761904761904764e-06, |
| "loss": 0.3171, |
| "step": 18470 |
| }, |
| { |
| "epoch": 4.75614464032943, |
| "grad_norm": 1.0323374271392822, |
| "learning_rate": 2.4504504504504506e-06, |
| "loss": 0.2878, |
| "step": 18480 |
| }, |
| { |
| "epoch": 4.7587183116716, |
| "grad_norm": 1.34542977809906, |
| "learning_rate": 2.424710424710425e-06, |
| "loss": 0.33, |
| "step": 18490 |
| }, |
| { |
| "epoch": 4.761291983013769, |
| "grad_norm": 1.1959524154663086, |
| "learning_rate": 2.3989703989703993e-06, |
| "loss": 0.3031, |
| "step": 18500 |
| }, |
| { |
| "epoch": 4.763865654355939, |
| "grad_norm": 1.4889193773269653, |
| "learning_rate": 2.3732303732303734e-06, |
| "loss": 0.386, |
| "step": 18510 |
| }, |
| { |
| "epoch": 4.766439325698109, |
| "grad_norm": 1.1365255117416382, |
| "learning_rate": 2.3474903474903476e-06, |
| "loss": 0.3173, |
| "step": 18520 |
| }, |
| { |
| "epoch": 4.769012997040278, |
| "grad_norm": 1.1125775575637817, |
| "learning_rate": 2.3217503217503217e-06, |
| "loss": 0.37, |
| "step": 18530 |
| }, |
| { |
| "epoch": 4.771586668382447, |
| "grad_norm": 1.230258584022522, |
| "learning_rate": 2.296010296010296e-06, |
| "loss": 0.3115, |
| "step": 18540 |
| }, |
| { |
| "epoch": 4.7741603397246175, |
| "grad_norm": 1.1539350748062134, |
| "learning_rate": 2.2702702702702705e-06, |
| "loss": 0.3639, |
| "step": 18550 |
| }, |
| { |
| "epoch": 4.776734011066787, |
| "grad_norm": 1.2653287649154663, |
| "learning_rate": 2.2445302445302446e-06, |
| "loss": 0.3541, |
| "step": 18560 |
| }, |
| { |
| "epoch": 4.779307682408956, |
| "grad_norm": 0.9209233522415161, |
| "learning_rate": 2.218790218790219e-06, |
| "loss": 0.3348, |
| "step": 18570 |
| }, |
| { |
| "epoch": 4.781881353751126, |
| "grad_norm": 0.895807683467865, |
| "learning_rate": 2.1930501930501933e-06, |
| "loss": 0.3351, |
| "step": 18580 |
| }, |
| { |
| "epoch": 4.784455025093296, |
| "grad_norm": 0.8402043581008911, |
| "learning_rate": 2.167310167310167e-06, |
| "loss": 0.3198, |
| "step": 18590 |
| }, |
| { |
| "epoch": 4.787028696435465, |
| "grad_norm": 2.4166226387023926, |
| "learning_rate": 2.1415701415701416e-06, |
| "loss": 0.3464, |
| "step": 18600 |
| }, |
| { |
| "epoch": 4.7896023677776345, |
| "grad_norm": 2.75046706199646, |
| "learning_rate": 2.115830115830116e-06, |
| "loss": 0.3046, |
| "step": 18610 |
| }, |
| { |
| "epoch": 4.792176039119805, |
| "grad_norm": 1.4120715856552124, |
| "learning_rate": 2.0900900900900904e-06, |
| "loss": 0.387, |
| "step": 18620 |
| }, |
| { |
| "epoch": 4.794749710461974, |
| "grad_norm": 1.039160132408142, |
| "learning_rate": 2.0643500643500645e-06, |
| "loss": 0.3805, |
| "step": 18630 |
| }, |
| { |
| "epoch": 4.797323381804143, |
| "grad_norm": 3.3216381072998047, |
| "learning_rate": 2.0386100386100387e-06, |
| "loss": 0.3687, |
| "step": 18640 |
| }, |
| { |
| "epoch": 4.799897053146314, |
| "grad_norm": 1.0428948402404785, |
| "learning_rate": 2.0128700128700133e-06, |
| "loss": 0.3146, |
| "step": 18650 |
| }, |
| { |
| "epoch": 4.802470724488483, |
| "grad_norm": 1.2070870399475098, |
| "learning_rate": 1.9871299871299874e-06, |
| "loss": 0.3448, |
| "step": 18660 |
| }, |
| { |
| "epoch": 4.805044395830652, |
| "grad_norm": 0.9999447464942932, |
| "learning_rate": 1.961389961389961e-06, |
| "loss": 0.3139, |
| "step": 18670 |
| }, |
| { |
| "epoch": 4.807618067172822, |
| "grad_norm": 0.7918708920478821, |
| "learning_rate": 1.9356499356499357e-06, |
| "loss": 0.355, |
| "step": 18680 |
| }, |
| { |
| "epoch": 4.810191738514992, |
| "grad_norm": 1.158919334411621, |
| "learning_rate": 1.90990990990991e-06, |
| "loss": 0.3108, |
| "step": 18690 |
| }, |
| { |
| "epoch": 4.812765409857161, |
| "grad_norm": 1.4351342916488647, |
| "learning_rate": 1.8841698841698844e-06, |
| "loss": 0.3367, |
| "step": 18700 |
| }, |
| { |
| "epoch": 4.815339081199331, |
| "grad_norm": 1.1386419534683228, |
| "learning_rate": 1.8584298584298586e-06, |
| "loss": 0.3189, |
| "step": 18710 |
| }, |
| { |
| "epoch": 4.817912752541501, |
| "grad_norm": 1.5254247188568115, |
| "learning_rate": 1.8326898326898327e-06, |
| "loss": 0.3336, |
| "step": 18720 |
| }, |
| { |
| "epoch": 4.82048642388367, |
| "grad_norm": 1.0063539743423462, |
| "learning_rate": 1.8069498069498071e-06, |
| "loss": 0.3185, |
| "step": 18730 |
| }, |
| { |
| "epoch": 4.82306009522584, |
| "grad_norm": 1.1583902835845947, |
| "learning_rate": 1.7812097812097813e-06, |
| "loss": 0.3918, |
| "step": 18740 |
| }, |
| { |
| "epoch": 4.825633766568009, |
| "grad_norm": 1.1044045686721802, |
| "learning_rate": 1.7554697554697556e-06, |
| "loss": 0.2989, |
| "step": 18750 |
| }, |
| { |
| "epoch": 4.828207437910179, |
| "grad_norm": 1.114070177078247, |
| "learning_rate": 1.7297297297297298e-06, |
| "loss": 0.3586, |
| "step": 18760 |
| }, |
| { |
| "epoch": 4.8307811092523485, |
| "grad_norm": 1.4722037315368652, |
| "learning_rate": 1.703989703989704e-06, |
| "loss": 0.364, |
| "step": 18770 |
| }, |
| { |
| "epoch": 4.833354780594518, |
| "grad_norm": 1.1275004148483276, |
| "learning_rate": 1.6782496782496785e-06, |
| "loss": 0.3595, |
| "step": 18780 |
| }, |
| { |
| "epoch": 4.835928451936688, |
| "grad_norm": 1.0008904933929443, |
| "learning_rate": 1.6525096525096527e-06, |
| "loss": 0.3347, |
| "step": 18790 |
| }, |
| { |
| "epoch": 4.838502123278857, |
| "grad_norm": 0.9728559851646423, |
| "learning_rate": 1.6267696267696266e-06, |
| "loss": 0.3359, |
| "step": 18800 |
| }, |
| { |
| "epoch": 4.841075794621027, |
| "grad_norm": 1.343985915184021, |
| "learning_rate": 1.6010296010296012e-06, |
| "loss": 0.3496, |
| "step": 18810 |
| }, |
| { |
| "epoch": 4.843649465963196, |
| "grad_norm": 0.8457748889923096, |
| "learning_rate": 1.5752895752895753e-06, |
| "loss": 0.3495, |
| "step": 18820 |
| }, |
| { |
| "epoch": 4.846223137305366, |
| "grad_norm": 4.120772838592529, |
| "learning_rate": 1.5495495495495495e-06, |
| "loss": 0.3289, |
| "step": 18830 |
| }, |
| { |
| "epoch": 4.848796808647536, |
| "grad_norm": 1.3416566848754883, |
| "learning_rate": 1.5238095238095238e-06, |
| "loss": 0.341, |
| "step": 18840 |
| }, |
| { |
| "epoch": 4.851370479989705, |
| "grad_norm": 2.817333936691284, |
| "learning_rate": 1.4980694980694982e-06, |
| "loss": 0.3213, |
| "step": 18850 |
| }, |
| { |
| "epoch": 4.853944151331875, |
| "grad_norm": 1.17255699634552, |
| "learning_rate": 1.4723294723294726e-06, |
| "loss": 0.3291, |
| "step": 18860 |
| }, |
| { |
| "epoch": 4.856517822674045, |
| "grad_norm": 2.559112787246704, |
| "learning_rate": 1.4465894465894465e-06, |
| "loss": 0.2838, |
| "step": 18870 |
| }, |
| { |
| "epoch": 4.859091494016214, |
| "grad_norm": 1.171962022781372, |
| "learning_rate": 1.4208494208494209e-06, |
| "loss": 0.3366, |
| "step": 18880 |
| }, |
| { |
| "epoch": 4.861665165358383, |
| "grad_norm": 1.125060796737671, |
| "learning_rate": 1.3951093951093952e-06, |
| "loss": 0.3411, |
| "step": 18890 |
| }, |
| { |
| "epoch": 4.864238836700554, |
| "grad_norm": 1.2087688446044922, |
| "learning_rate": 1.3693693693693696e-06, |
| "loss": 0.2909, |
| "step": 18900 |
| }, |
| { |
| "epoch": 4.866812508042723, |
| "grad_norm": 1.3647404909133911, |
| "learning_rate": 1.3436293436293435e-06, |
| "loss": 0.3347, |
| "step": 18910 |
| }, |
| { |
| "epoch": 4.869386179384892, |
| "grad_norm": 1.1150912046432495, |
| "learning_rate": 1.317889317889318e-06, |
| "loss": 0.3467, |
| "step": 18920 |
| }, |
| { |
| "epoch": 4.8719598507270625, |
| "grad_norm": 1.2216849327087402, |
| "learning_rate": 1.2921492921492923e-06, |
| "loss": 0.3459, |
| "step": 18930 |
| }, |
| { |
| "epoch": 4.874533522069232, |
| "grad_norm": 1.0959711074829102, |
| "learning_rate": 1.2664092664092664e-06, |
| "loss": 0.3167, |
| "step": 18940 |
| }, |
| { |
| "epoch": 4.877107193411401, |
| "grad_norm": 1.3392401933670044, |
| "learning_rate": 1.2406692406692406e-06, |
| "loss": 0.2993, |
| "step": 18950 |
| }, |
| { |
| "epoch": 4.8796808647535705, |
| "grad_norm": 1.1086993217468262, |
| "learning_rate": 1.214929214929215e-06, |
| "loss": 0.3165, |
| "step": 18960 |
| }, |
| { |
| "epoch": 4.882254536095741, |
| "grad_norm": 1.0389692783355713, |
| "learning_rate": 1.1891891891891893e-06, |
| "loss": 0.349, |
| "step": 18970 |
| }, |
| { |
| "epoch": 4.88482820743791, |
| "grad_norm": 1.053175449371338, |
| "learning_rate": 1.1634491634491635e-06, |
| "loss": 0.3713, |
| "step": 18980 |
| }, |
| { |
| "epoch": 4.8874018787800795, |
| "grad_norm": 1.4085311889648438, |
| "learning_rate": 1.1377091377091378e-06, |
| "loss": 0.3424, |
| "step": 18990 |
| }, |
| { |
| "epoch": 4.88997555012225, |
| "grad_norm": 1.0388787984848022, |
| "learning_rate": 1.111969111969112e-06, |
| "loss": 0.3223, |
| "step": 19000 |
| }, |
| { |
| "epoch": 4.88997555012225, |
| "eval_loss": 0.8570228219032288, |
| "eval_runtime": 395.7978, |
| "eval_samples_per_second": 49.081, |
| "eval_steps_per_second": 2.456, |
| "eval_token_accuracy": 0.00035359363493497523, |
| "step": 19000 |
| }, |
| { |
| "epoch": 4.892549221464419, |
| "grad_norm": 1.3370487689971924, |
| "learning_rate": 1.0862290862290863e-06, |
| "loss": 0.3524, |
| "step": 19010 |
| }, |
| { |
| "epoch": 4.895122892806588, |
| "grad_norm": 1.1266289949417114, |
| "learning_rate": 1.0604890604890605e-06, |
| "loss": 0.3193, |
| "step": 19020 |
| }, |
| { |
| "epoch": 4.897696564148758, |
| "grad_norm": 1.6796464920043945, |
| "learning_rate": 1.0347490347490348e-06, |
| "loss": 0.3117, |
| "step": 19030 |
| }, |
| { |
| "epoch": 4.900270235490928, |
| "grad_norm": 1.1467808485031128, |
| "learning_rate": 1.009009009009009e-06, |
| "loss": 0.366, |
| "step": 19040 |
| }, |
| { |
| "epoch": 4.902843906833097, |
| "grad_norm": 0.8519312739372253, |
| "learning_rate": 9.832689832689834e-07, |
| "loss": 0.2902, |
| "step": 19050 |
| }, |
| { |
| "epoch": 4.905417578175267, |
| "grad_norm": 1.30669105052948, |
| "learning_rate": 9.575289575289575e-07, |
| "loss": 0.3016, |
| "step": 19060 |
| }, |
| { |
| "epoch": 4.907991249517437, |
| "grad_norm": 1.3413194417953491, |
| "learning_rate": 9.317889317889319e-07, |
| "loss": 0.3301, |
| "step": 19070 |
| }, |
| { |
| "epoch": 4.910564920859606, |
| "grad_norm": 1.2429314851760864, |
| "learning_rate": 9.06048906048906e-07, |
| "loss": 0.328, |
| "step": 19080 |
| }, |
| { |
| "epoch": 4.913138592201776, |
| "grad_norm": 1.04901123046875, |
| "learning_rate": 8.803088803088803e-07, |
| "loss": 0.3483, |
| "step": 19090 |
| }, |
| { |
| "epoch": 4.915712263543945, |
| "grad_norm": 0.9599467515945435, |
| "learning_rate": 8.545688545688547e-07, |
| "loss": 0.3241, |
| "step": 19100 |
| }, |
| { |
| "epoch": 4.918285934886115, |
| "grad_norm": 3.971909284591675, |
| "learning_rate": 8.288288288288289e-07, |
| "loss": 0.3647, |
| "step": 19110 |
| }, |
| { |
| "epoch": 4.9208596062282846, |
| "grad_norm": 1.1139452457427979, |
| "learning_rate": 8.030888030888032e-07, |
| "loss": 0.3326, |
| "step": 19120 |
| }, |
| { |
| "epoch": 4.923433277570454, |
| "grad_norm": 1.015626311302185, |
| "learning_rate": 7.773487773487774e-07, |
| "loss": 0.3422, |
| "step": 19130 |
| }, |
| { |
| "epoch": 4.926006948912624, |
| "grad_norm": 1.0744688510894775, |
| "learning_rate": 7.516087516087516e-07, |
| "loss": 0.3471, |
| "step": 19140 |
| }, |
| { |
| "epoch": 4.9285806202547935, |
| "grad_norm": 1.1969647407531738, |
| "learning_rate": 7.258687258687259e-07, |
| "loss": 0.3139, |
| "step": 19150 |
| }, |
| { |
| "epoch": 4.931154291596963, |
| "grad_norm": 0.8164909482002258, |
| "learning_rate": 7.001287001287001e-07, |
| "loss": 0.3355, |
| "step": 19160 |
| }, |
| { |
| "epoch": 4.933727962939132, |
| "grad_norm": 1.189275860786438, |
| "learning_rate": 6.743886743886745e-07, |
| "loss": 0.3782, |
| "step": 19170 |
| }, |
| { |
| "epoch": 4.936301634281302, |
| "grad_norm": 1.551167368888855, |
| "learning_rate": 6.486486486486486e-07, |
| "loss": 0.3773, |
| "step": 19180 |
| }, |
| { |
| "epoch": 4.938875305623472, |
| "grad_norm": 0.9430525898933411, |
| "learning_rate": 6.22908622908623e-07, |
| "loss": 0.3332, |
| "step": 19190 |
| }, |
| { |
| "epoch": 4.941448976965641, |
| "grad_norm": 1.2511941194534302, |
| "learning_rate": 5.971685971685971e-07, |
| "loss": 0.3804, |
| "step": 19200 |
| }, |
| { |
| "epoch": 4.944022648307811, |
| "grad_norm": 1.0374319553375244, |
| "learning_rate": 5.714285714285715e-07, |
| "loss": 0.3166, |
| "step": 19210 |
| }, |
| { |
| "epoch": 4.946596319649981, |
| "grad_norm": 0.9965717792510986, |
| "learning_rate": 5.456885456885458e-07, |
| "loss": 0.3132, |
| "step": 19220 |
| }, |
| { |
| "epoch": 4.94916999099215, |
| "grad_norm": 1.1020451784133911, |
| "learning_rate": 5.1994851994852e-07, |
| "loss": 0.3717, |
| "step": 19230 |
| }, |
| { |
| "epoch": 4.95174366233432, |
| "grad_norm": 0.7918252944946289, |
| "learning_rate": 4.942084942084943e-07, |
| "loss": 0.2936, |
| "step": 19240 |
| }, |
| { |
| "epoch": 4.95431733367649, |
| "grad_norm": 1.4422928094863892, |
| "learning_rate": 4.6846846846846847e-07, |
| "loss": 0.3374, |
| "step": 19250 |
| }, |
| { |
| "epoch": 4.956891005018659, |
| "grad_norm": 0.949720025062561, |
| "learning_rate": 4.427284427284428e-07, |
| "loss": 0.3771, |
| "step": 19260 |
| }, |
| { |
| "epoch": 4.959464676360829, |
| "grad_norm": 1.0243984460830688, |
| "learning_rate": 4.16988416988417e-07, |
| "loss": 0.3283, |
| "step": 19270 |
| }, |
| { |
| "epoch": 4.962038347702999, |
| "grad_norm": 1.0813188552856445, |
| "learning_rate": 3.912483912483913e-07, |
| "loss": 0.3885, |
| "step": 19280 |
| }, |
| { |
| "epoch": 4.964612019045168, |
| "grad_norm": 4.7011799812316895, |
| "learning_rate": 3.6550836550836556e-07, |
| "loss": 0.3065, |
| "step": 19290 |
| }, |
| { |
| "epoch": 4.967185690387337, |
| "grad_norm": 5.996508598327637, |
| "learning_rate": 3.397683397683398e-07, |
| "loss": 0.3949, |
| "step": 19300 |
| }, |
| { |
| "epoch": 4.9697593617295075, |
| "grad_norm": 1.202574610710144, |
| "learning_rate": 3.1402831402831407e-07, |
| "loss": 0.3265, |
| "step": 19310 |
| }, |
| { |
| "epoch": 4.972333033071677, |
| "grad_norm": 0.9816371202468872, |
| "learning_rate": 2.8828828828828833e-07, |
| "loss": 0.3646, |
| "step": 19320 |
| }, |
| { |
| "epoch": 4.974906704413846, |
| "grad_norm": 1.1760458946228027, |
| "learning_rate": 2.6254826254826254e-07, |
| "loss": 0.3221, |
| "step": 19330 |
| }, |
| { |
| "epoch": 4.977480375756016, |
| "grad_norm": 1.4523731470108032, |
| "learning_rate": 2.3680823680823682e-07, |
| "loss": 0.3247, |
| "step": 19340 |
| }, |
| { |
| "epoch": 4.980054047098186, |
| "grad_norm": 1.1711961030960083, |
| "learning_rate": 2.1106821106821108e-07, |
| "loss": 0.3514, |
| "step": 19350 |
| }, |
| { |
| "epoch": 4.982627718440355, |
| "grad_norm": 0.8875769972801208, |
| "learning_rate": 1.8532818532818534e-07, |
| "loss": 0.3135, |
| "step": 19360 |
| }, |
| { |
| "epoch": 4.9852013897825245, |
| "grad_norm": 1.7097541093826294, |
| "learning_rate": 1.595881595881596e-07, |
| "loss": 0.3252, |
| "step": 19370 |
| }, |
| { |
| "epoch": 4.987775061124695, |
| "grad_norm": 3.204772472381592, |
| "learning_rate": 1.3384813384813385e-07, |
| "loss": 0.2987, |
| "step": 19380 |
| }, |
| { |
| "epoch": 4.990348732466864, |
| "grad_norm": 2.616095542907715, |
| "learning_rate": 1.0810810810810811e-07, |
| "loss": 0.3303, |
| "step": 19390 |
| }, |
| { |
| "epoch": 4.992922403809033, |
| "grad_norm": 1.0616631507873535, |
| "learning_rate": 8.236808236808237e-08, |
| "loss": 0.3565, |
| "step": 19400 |
| }, |
| { |
| "epoch": 4.995496075151204, |
| "grad_norm": 0.8763009905815125, |
| "learning_rate": 5.662805662805663e-08, |
| "loss": 0.3151, |
| "step": 19410 |
| }, |
| { |
| "epoch": 4.998069746493373, |
| "grad_norm": 1.2131026983261108, |
| "learning_rate": 3.088803088803089e-08, |
| "loss": 0.3425, |
| "step": 19420 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 19425, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 2000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.4362014238809784e+18, |
| "train_batch_size": 10, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|