{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 705, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00851063829787234, "grad_norm": 0.9370151162147522, "learning_rate": 8.333333333333333e-07, "loss": 4.115976333618164, "step": 2 }, { "epoch": 0.01702127659574468, "grad_norm": 0.41166678071022034, "learning_rate": 2.4999999999999998e-06, "loss": 1.878818154335022, "step": 4 }, { "epoch": 0.02553191489361702, "grad_norm": 0.4669341444969177, "learning_rate": 4.166666666666667e-06, "loss": 2.0509164333343506, "step": 6 }, { "epoch": 0.03404255319148936, "grad_norm": 0.6062728762626648, "learning_rate": 5.833333333333334e-06, "loss": 1.7545194625854492, "step": 8 }, { "epoch": 0.0425531914893617, "grad_norm": 0.37637069821357727, "learning_rate": 7.5e-06, "loss": 1.7690430879592896, "step": 10 }, { "epoch": 0.05106382978723404, "grad_norm": 2.065159797668457, "learning_rate": 9.166666666666668e-06, "loss": 2.355275869369507, "step": 12 }, { "epoch": 0.059574468085106386, "grad_norm": 1.091937780380249, "learning_rate": 1.0833333333333334e-05, "loss": 1.876306176185608, "step": 14 }, { "epoch": 0.06808510638297872, "grad_norm": 0.49564921855926514, "learning_rate": 1.25e-05, "loss": 1.785620927810669, "step": 16 }, { "epoch": 0.07659574468085106, "grad_norm": 1.3353960514068604, "learning_rate": 1.4166666666666666e-05, "loss": 1.523728370666504, "step": 18 }, { "epoch": 0.0851063829787234, "grad_norm": 0.30626556277275085, "learning_rate": 1.5833333333333333e-05, "loss": 1.510640025138855, "step": 20 }, { "epoch": 0.09361702127659574, "grad_norm": 0.1502075493335724, "learning_rate": 1.7500000000000002e-05, "loss": 1.4279940128326416, "step": 22 }, { "epoch": 0.10212765957446808, "grad_norm": 0.8587890863418579, "learning_rate": 1.9166666666666667e-05, "loss": 1.1794553995132446, "step": 24 }, { "epoch": 0.11063829787234042, "grad_norm": 1.0492326021194458, "learning_rate": 2.0833333333333333e-05, "loss": 1.2579714059829712, "step": 26 }, { "epoch": 0.11914893617021277, "grad_norm": 0.264726847410202, "learning_rate": 2.25e-05, "loss": 0.9510668516159058, "step": 28 }, { "epoch": 0.1276595744680851, "grad_norm": 0.2859492301940918, "learning_rate": 2.4166666666666667e-05, "loss": 1.656806230545044, "step": 30 }, { "epoch": 0.13617021276595745, "grad_norm": 1.1864591836929321, "learning_rate": 2.5833333333333336e-05, "loss": 0.683425784111023, "step": 32 }, { "epoch": 0.14468085106382977, "grad_norm": 0.22737205028533936, "learning_rate": 2.75e-05, "loss": 1.2381809949874878, "step": 34 }, { "epoch": 0.15319148936170213, "grad_norm": 0.32182246446609497, "learning_rate": 2.9166666666666666e-05, "loss": 1.422176480293274, "step": 36 }, { "epoch": 0.16170212765957448, "grad_norm": 0.36561885476112366, "learning_rate": 2.9998734788806287e-05, "loss": 1.2621259689331055, "step": 38 }, { "epoch": 0.1702127659574468, "grad_norm": 0.7920733690261841, "learning_rate": 2.9988614605803806e-05, "loss": 1.2012425661087036, "step": 40 }, { "epoch": 0.17872340425531916, "grad_norm": 0.19826872646808624, "learning_rate": 2.99683822733885e-05, "loss": 1.2804166078567505, "step": 42 }, { "epoch": 0.18723404255319148, "grad_norm": 0.5331127643585205, "learning_rate": 2.9938053852362484e-05, "loss": 1.1188151836395264, "step": 44 }, { "epoch": 0.19574468085106383, "grad_norm": 0.22702525556087494, "learning_rate": 2.989765341799095e-05, "loss": 1.338016152381897, "step": 46 }, { "epoch": 0.20425531914893616, "grad_norm": 1.6932013034820557, "learning_rate": 2.9847213040890793e-05, "loss": 0.8224107027053833, "step": 48 }, { "epoch": 0.2127659574468085, "grad_norm": 0.49766504764556885, "learning_rate": 2.9786772761572335e-05, "loss": 0.9487060308456421, "step": 50 }, { "epoch": 0.22127659574468084, "grad_norm": 0.20401130616664886, "learning_rate": 2.9716380558654445e-05, "loss": 1.3093581199645996, "step": 52 }, { "epoch": 0.2297872340425532, "grad_norm": 0.4012207090854645, "learning_rate": 2.9636092310778195e-05, "loss": 1.2255353927612305, "step": 54 }, { "epoch": 0.23829787234042554, "grad_norm": 0.8393444418907166, "learning_rate": 2.954597175224938e-05, "loss": 0.8253700137138367, "step": 56 }, { "epoch": 0.24680851063829787, "grad_norm": 0.27147796750068665, "learning_rate": 2.9446090422445016e-05, "loss": 1.0494465827941895, "step": 58 }, { "epoch": 0.2553191489361702, "grad_norm": 0.42597004771232605, "learning_rate": 2.9336527609024072e-05, "loss": 1.358299970626831, "step": 60 }, { "epoch": 0.26382978723404255, "grad_norm": 0.20699705183506012, "learning_rate": 2.9217370284987434e-05, "loss": 1.335315465927124, "step": 62 }, { "epoch": 0.2723404255319149, "grad_norm": 0.16389895975589752, "learning_rate": 2.9088713039637117e-05, "loss": 1.2662416696548462, "step": 64 }, { "epoch": 0.28085106382978725, "grad_norm": 0.25043565034866333, "learning_rate": 2.8950658003489534e-05, "loss": 1.2614431381225586, "step": 66 }, { "epoch": 0.28936170212765955, "grad_norm": 0.19853799045085907, "learning_rate": 2.880331476720238e-05, "loss": 1.2736550569534302, "step": 68 }, { "epoch": 0.2978723404255319, "grad_norm": 2.3273322582244873, "learning_rate": 2.8646800294579517e-05, "loss": 1.2623916864395142, "step": 70 }, { "epoch": 0.30638297872340425, "grad_norm": 0.25601401925086975, "learning_rate": 2.848123882972295e-05, "loss": 1.2526882886886597, "step": 72 }, { "epoch": 0.3148936170212766, "grad_norm": 0.342012494802475, "learning_rate": 2.8306761798405526e-05, "loss": 0.9114877581596375, "step": 74 }, { "epoch": 0.32340425531914896, "grad_norm": 0.5187206864356995, "learning_rate": 2.812350770374273e-05, "loss": 1.2244638204574585, "step": 76 }, { "epoch": 0.33191489361702126, "grad_norm": 0.2703041434288025, "learning_rate": 2.793162201624631e-05, "loss": 1.2725831270217896, "step": 78 }, { "epoch": 0.3404255319148936, "grad_norm": 0.1924225091934204, "learning_rate": 2.77312570583471e-05, "loss": 0.7258127927780151, "step": 80 }, { "epoch": 0.34893617021276596, "grad_norm": 0.25936806201934814, "learning_rate": 2.752257188347862e-05, "loss": 1.2518136501312256, "step": 82 }, { "epoch": 0.3574468085106383, "grad_norm": 0.1409609615802765, "learning_rate": 2.730573214981751e-05, "loss": 0.9943649172782898, "step": 84 }, { "epoch": 0.3659574468085106, "grad_norm": 0.18737079203128815, "learning_rate": 2.7080909988780982e-05, "loss": 1.001371145248413, "step": 86 }, { "epoch": 0.37446808510638296, "grad_norm": 0.512141227722168, "learning_rate": 2.684828386838569e-05, "loss": 1.2457740306854248, "step": 88 }, { "epoch": 0.3829787234042553, "grad_norm": 0.7956476807594299, "learning_rate": 2.6608038451576528e-05, "loss": 0.9049152135848999, "step": 90 }, { "epoch": 0.39148936170212767, "grad_norm": 0.17900533974170685, "learning_rate": 2.636036444963769e-05, "loss": 1.0495070219039917, "step": 92 }, { "epoch": 0.4, "grad_norm": 0.28203219175338745, "learning_rate": 2.6105458470802563e-05, "loss": 1.163374900817871, "step": 94 }, { "epoch": 0.4085106382978723, "grad_norm": 0.2998636066913605, "learning_rate": 2.5843522864182394e-05, "loss": 1.064250111579895, "step": 96 }, { "epoch": 0.41702127659574467, "grad_norm": 0.34955132007598877, "learning_rate": 2.557476555913785e-05, "loss": 1.0862053632736206, "step": 98 }, { "epoch": 0.425531914893617, "grad_norm": 0.5751879215240479, "learning_rate": 2.5299399900220803e-05, "loss": 0.9672101736068726, "step": 100 }, { "epoch": 0.4340425531914894, "grad_norm": 0.556722104549408, "learning_rate": 2.5017644477817424e-05, "loss": 1.38407564163208, "step": 102 }, { "epoch": 0.4425531914893617, "grad_norm": 0.24831920862197876, "learning_rate": 2.47297229546271e-05, "loss": 1.0887106657028198, "step": 104 }, { "epoch": 0.451063829787234, "grad_norm": 0.39581331610679626, "learning_rate": 2.4435863888114814e-05, "loss": 0.8722383975982666, "step": 106 }, { "epoch": 0.4595744680851064, "grad_norm": 0.16548699140548706, "learning_rate": 2.4136300549077976e-05, "loss": 1.2673208713531494, "step": 108 }, { "epoch": 0.46808510638297873, "grad_norm": 0.20999404788017273, "learning_rate": 2.3831270736471703e-05, "loss": 1.2440522909164429, "step": 110 }, { "epoch": 0.4765957446808511, "grad_norm": 0.18370361626148224, "learning_rate": 2.352101658863959e-05, "loss": 1.1419363021850586, "step": 112 }, { "epoch": 0.4851063829787234, "grad_norm": 0.24067164957523346, "learning_rate": 2.32057843910998e-05, "loss": 1.0633783340454102, "step": 114 }, { "epoch": 0.49361702127659574, "grad_norm": 0.47072628140449524, "learning_rate": 2.288582438103903e-05, "loss": 1.179775595664978, "step": 116 }, { "epoch": 0.502127659574468, "grad_norm": 0.2161823809146881, "learning_rate": 2.256139054866955e-05, "loss": 1.2002534866333008, "step": 118 }, { "epoch": 0.5106382978723404, "grad_norm": 1.2126752138137817, "learning_rate": 2.2232740435607067e-05, "loss": 0.9621443748474121, "step": 120 }, { "epoch": 0.5191489361702127, "grad_norm": 0.1435491442680359, "learning_rate": 2.19001349304294e-05, "loss": 1.253274917602539, "step": 122 }, { "epoch": 0.5276595744680851, "grad_norm": 0.19416141510009766, "learning_rate": 2.156383806157826e-05, "loss": 1.3183720111846924, "step": 124 }, { "epoch": 0.5361702127659574, "grad_norm": 0.14461560547351837, "learning_rate": 2.1224116787768552e-05, "loss": 1.238662838935852, "step": 126 }, { "epoch": 0.5446808510638298, "grad_norm": 0.155701145529747, "learning_rate": 2.0881240786071588e-05, "loss": 1.1723605394363403, "step": 128 }, { "epoch": 0.5531914893617021, "grad_norm": 0.17010867595672607, "learning_rate": 2.05354822378404e-05, "loss": 1.021695852279663, "step": 130 }, { "epoch": 0.5617021276595745, "grad_norm": 0.17941494286060333, "learning_rate": 2.018711561264714e-05, "loss": 1.2407881021499634, "step": 132 }, { "epoch": 0.5702127659574469, "grad_norm": 0.3246072828769684, "learning_rate": 1.9836417450403978e-05, "loss": 1.186881422996521, "step": 134 }, { "epoch": 0.5787234042553191, "grad_norm": 0.32305997610092163, "learning_rate": 1.9483666141840615e-05, "loss": 0.965923547744751, "step": 136 }, { "epoch": 0.5872340425531914, "grad_norm": 0.15171490609645844, "learning_rate": 1.9129141707512508e-05, "loss": 0.5913432836532593, "step": 138 }, { "epoch": 0.5957446808510638, "grad_norm": 0.37349367141723633, "learning_rate": 1.8773125575515364e-05, "loss": 1.1055043935775757, "step": 140 }, { "epoch": 0.6042553191489362, "grad_norm": 0.15140922367572784, "learning_rate": 1.8415900358082268e-05, "loss": 1.3598577976226807, "step": 142 }, { "epoch": 0.6127659574468085, "grad_norm": 0.5196365118026733, "learning_rate": 1.805774962724083e-05, "loss": 0.9218302369117737, "step": 144 }, { "epoch": 0.6212765957446809, "grad_norm": 0.15164780616760254, "learning_rate": 1.7698957689708426e-05, "loss": 1.1118155717849731, "step": 146 }, { "epoch": 0.6297872340425532, "grad_norm": 0.2984839677810669, "learning_rate": 1.7339809361204252e-05, "loss": 1.2171759605407715, "step": 148 }, { "epoch": 0.6382978723404256, "grad_norm": 0.28176164627075195, "learning_rate": 1.6980589740357294e-05, "loss": 1.1775646209716797, "step": 150 }, { "epoch": 0.6468085106382979, "grad_norm": 0.19051885604858398, "learning_rate": 1.6621583982389707e-05, "loss": 1.1114767789840698, "step": 152 }, { "epoch": 0.6553191489361702, "grad_norm": 0.28726738691329956, "learning_rate": 1.6263077072755326e-05, "loss": 0.7064566612243652, "step": 154 }, { "epoch": 0.6638297872340425, "grad_norm": 0.3889693021774292, "learning_rate": 1.5905353600912898e-05, "loss": 0.8802153468132019, "step": 156 }, { "epoch": 0.6723404255319149, "grad_norm": 0.28420913219451904, "learning_rate": 1.5548697534413646e-05, "loss": 0.9915045499801636, "step": 158 }, { "epoch": 0.6808510638297872, "grad_norm": 0.2218930721282959, "learning_rate": 1.5193391993482582e-05, "loss": 1.2188233137130737, "step": 160 }, { "epoch": 0.6893617021276596, "grad_norm": 0.1267762929201126, "learning_rate": 1.4839719026272377e-05, "loss": 1.0931644439697266, "step": 162 }, { "epoch": 0.6978723404255319, "grad_norm": 0.20958174765110016, "learning_rate": 1.4487959384968272e-05, "loss": 0.875884473323822, "step": 164 }, { "epoch": 0.7063829787234043, "grad_norm": 0.2075873166322708, "learning_rate": 1.4138392302921813e-05, "loss": 0.9931239485740662, "step": 166 }, { "epoch": 0.7148936170212766, "grad_norm": 0.45486778020858765, "learning_rate": 1.3791295272990175e-05, "loss": 1.1112878322601318, "step": 168 }, { "epoch": 0.723404255319149, "grad_norm": 0.12606321275234222, "learning_rate": 1.344694382725718e-05, "loss": 0.8394916653633118, "step": 170 }, { "epoch": 0.7319148936170212, "grad_norm": 1.405964732170105, "learning_rate": 1.3105611318310818e-05, "loss": 0.8859961032867432, "step": 172 }, { "epoch": 0.7404255319148936, "grad_norm": 0.24160844087600708, "learning_rate": 1.2767568702250844e-05, "loss": 1.1839947700500488, "step": 174 }, { "epoch": 0.7489361702127659, "grad_norm": 0.28286510705947876, "learning_rate": 1.2433084323598791e-05, "loss": 0.6160851120948792, "step": 176 }, { "epoch": 0.7574468085106383, "grad_norm": 0.36609017848968506, "learning_rate": 1.2102423702281116e-05, "loss": 1.0064830780029297, "step": 178 }, { "epoch": 0.7659574468085106, "grad_norm": 0.25650474429130554, "learning_rate": 1.1775849322854516e-05, "loss": 1.0609067678451538, "step": 180 }, { "epoch": 0.774468085106383, "grad_norm": 0.3248478174209595, "learning_rate": 1.1453620426140795e-05, "loss": 0.7533571720123291, "step": 182 }, { "epoch": 0.7829787234042553, "grad_norm": 0.22153323888778687, "learning_rate": 1.1135992803436695e-05, "loss": 0.9635902643203735, "step": 184 }, { "epoch": 0.7914893617021277, "grad_norm": 0.11977202445268631, "learning_rate": 1.0823218593461992e-05, "loss": 0.9765074849128723, "step": 186 }, { "epoch": 0.8, "grad_norm": 0.2166285663843155, "learning_rate": 1.0515546082207097e-05, "loss": 0.730889081954956, "step": 188 }, { "epoch": 0.8085106382978723, "grad_norm": 0.6851885914802551, "learning_rate": 1.0213219505838983e-05, "loss": 0.966530442237854, "step": 190 }, { "epoch": 0.8170212765957446, "grad_norm": 0.33713918924331665, "learning_rate": 9.91647885682201e-06, "loss": 1.3054033517837524, "step": 192 }, { "epoch": 0.825531914893617, "grad_norm": 0.196068674325943, "learning_rate": 9.625559693407413e-06, "loss": 1.2358617782592773, "step": 194 }, { "epoch": 0.8340425531914893, "grad_norm": 0.2631889283657074, "learning_rate": 9.340692952642789e-06, "loss": 0.789035975933075, "step": 196 }, { "epoch": 0.8425531914893617, "grad_norm": 0.165288507938385, "learning_rate": 9.062104767049956e-06, "loss": 1.2202699184417725, "step": 198 }, { "epoch": 0.851063829787234, "grad_norm": 0.4081096649169922, "learning_rate": 8.790016285116763e-06, "loss": 1.005751132965088, "step": 200 }, { "epoch": 0.8595744680851064, "grad_norm": 0.16539311408996582, "learning_rate": 8.524643495745306e-06, "loss": 0.9970273971557617, "step": 202 }, { "epoch": 0.8680851063829788, "grad_norm": 0.1629868745803833, "learning_rate": 8.26619705679589e-06, "loss": 1.1567330360412598, "step": 204 }, { "epoch": 0.8765957446808511, "grad_norm": 0.13489580154418945, "learning_rate": 8.014882127862923e-06, "loss": 1.1192888021469116, "step": 206 }, { "epoch": 0.8851063829787233, "grad_norm": 0.1189538910984993, "learning_rate": 7.770898207415416e-06, "loss": 0.9340039491653442, "step": 208 }, { "epoch": 0.8936170212765957, "grad_norm": 0.12542982399463654, "learning_rate": 7.534438974431351e-06, "loss": 1.2615220546722412, "step": 210 }, { "epoch": 0.902127659574468, "grad_norm": 0.2067514955997467, "learning_rate": 7.305692134651742e-06, "loss": 0.7793570160865784, "step": 212 }, { "epoch": 0.9106382978723404, "grad_norm": 0.12745940685272217, "learning_rate": 7.084839271576291e-06, "loss": 0.9480925798416138, "step": 214 }, { "epoch": 0.9191489361702128, "grad_norm": 0.14118105173110962, "learning_rate": 6.872055702319054e-06, "loss": 0.7321507334709167, "step": 216 }, { "epoch": 0.9276595744680851, "grad_norm": 0.19066962599754333, "learning_rate": 6.667510338438419e-06, "loss": 1.2155194282531738, "step": 218 }, { "epoch": 0.9361702127659575, "grad_norm": 0.2428826540708542, "learning_rate": 6.471365551852012e-06, "loss": 0.39672887325286865, "step": 220 }, { "epoch": 0.9446808510638298, "grad_norm": 0.16802483797073364, "learning_rate": 6.2837770459428e-06, "loss": 1.254830241203308, "step": 222 }, { "epoch": 0.9531914893617022, "grad_norm": 0.1394779235124588, "learning_rate": 6.1048937319588676e-06, "loss": 1.138830304145813, "step": 224 }, { "epoch": 0.9617021276595744, "grad_norm": 0.2871001362800598, "learning_rate": 5.9348576108049065e-06, "loss": 0.8572854399681091, "step": 226 }, { "epoch": 0.9702127659574468, "grad_norm": 0.1366124153137207, "learning_rate": 5.773803660319234e-06, "loss": 1.0662239789962769, "step": 228 }, { "epoch": 0.9787234042553191, "grad_norm": 0.12555794417858124, "learning_rate": 5.621859728125884e-06, "loss": 1.222013235092163, "step": 230 }, { "epoch": 0.9872340425531915, "grad_norm": 0.29969581961631775, "learning_rate": 5.479146430146781e-06, "loss": 1.0417675971984863, "step": 232 }, { "epoch": 0.9957446808510638, "grad_norm": 0.4836138188838959, "learning_rate": 5.345777054854579e-06, "loss": 0.9495846629142761, "step": 234 }, { "epoch": 1.004255319148936, "grad_norm": 0.28257113695144653, "learning_rate": 5.221857473342149e-06, "loss": 0.843216061592102, "step": 236 }, { "epoch": 1.0127659574468084, "grad_norm": 0.1423388570547104, "learning_rate": 5.10748605528015e-06, "loss": 0.9633752107620239, "step": 238 }, { "epoch": 1.0212765957446808, "grad_norm": 0.12806400656700134, "learning_rate": 5.002753590829349e-06, "loss": 0.6067866683006287, "step": 240 }, { "epoch": 1.0297872340425531, "grad_norm": 0.243302121758461, "learning_rate": 4.9077432185697e-06, "loss": 0.9341376423835754, "step": 242 }, { "epoch": 1.0382978723404255, "grad_norm": 0.22620677947998047, "learning_rate": 4.822530359503393e-06, "loss": 1.046706199645996, "step": 244 }, { "epoch": 1.0468085106382978, "grad_norm": 0.1606007069349289, "learning_rate": 4.747182657184251e-06, "loss": 0.6613105535507202, "step": 246 }, { "epoch": 1.0553191489361702, "grad_norm": 0.3230328857898712, "learning_rate": 4.681759924021033e-06, "loss": 0.7598414421081543, "step": 248 }, { "epoch": 1.0638297872340425, "grad_norm": 0.13726186752319336, "learning_rate": 4.626314093797213e-06, "loss": 0.7334257364273071, "step": 250 }, { "epoch": 1.0723404255319149, "grad_norm": 0.3747357726097107, "learning_rate": 4.580889180444988e-06, "loss": 0.5062970519065857, "step": 252 }, { "epoch": 1.0808510638297872, "grad_norm": 0.15663602948188782, "learning_rate": 4.545521243106197e-06, "loss": 1.0599923133850098, "step": 254 }, { "epoch": 1.0893617021276596, "grad_norm": 0.1843012273311615, "learning_rate": 4.520238357507899e-06, "loss": 0.940389096736908, "step": 256 }, { "epoch": 1.097872340425532, "grad_norm": 0.22409912943840027, "learning_rate": 4.505060593675342e-06, "loss": 1.0018244981765747, "step": 258 }, { "epoch": 1.1063829787234043, "grad_norm": 0.45610666275024414, "learning_rate": 4.5e-06, "loss": 0.8150784969329834, "step": 260 }, { "epoch": 1.1148936170212767, "grad_norm": 0.19215510785579681, "learning_rate": 4.505060593675342e-06, "loss": 0.7950176000595093, "step": 262 }, { "epoch": 1.123404255319149, "grad_norm": 0.23423391580581665, "learning_rate": 4.520238357507899e-06, "loss": 0.7621920704841614, "step": 264 }, { "epoch": 1.1319148936170214, "grad_norm": 0.15685321390628815, "learning_rate": 4.545521243106197e-06, "loss": 0.8206462264060974, "step": 266 }, { "epoch": 1.1404255319148937, "grad_norm": 0.4606432020664215, "learning_rate": 4.580889180444988e-06, "loss": 0.9801955819129944, "step": 268 }, { "epoch": 1.148936170212766, "grad_norm": 0.21424169838428497, "learning_rate": 4.6263140937972124e-06, "loss": 0.8219439387321472, "step": 270 }, { "epoch": 1.1574468085106382, "grad_norm": 0.2082989364862442, "learning_rate": 4.6817599240210315e-06, "loss": 0.6543676257133484, "step": 272 }, { "epoch": 1.1659574468085105, "grad_norm": 0.1622384637594223, "learning_rate": 4.747182657184251e-06, "loss": 0.9012349247932434, "step": 274 }, { "epoch": 1.174468085106383, "grad_norm": 0.13174618780612946, "learning_rate": 4.822530359503391e-06, "loss": 0.7129980325698853, "step": 276 }, { "epoch": 1.1829787234042553, "grad_norm": 0.2709389925003052, "learning_rate": 4.9077432185697e-06, "loss": 0.5482591986656189, "step": 278 }, { "epoch": 1.1914893617021276, "grad_norm": 0.2576906979084015, "learning_rate": 5.002753590829348e-06, "loss": 0.5497639775276184, "step": 280 }, { "epoch": 1.2, "grad_norm": 0.1526031494140625, "learning_rate": 5.1074860552801466e-06, "loss": 0.6242255568504333, "step": 282 }, { "epoch": 1.2085106382978723, "grad_norm": 0.22073198854923248, "learning_rate": 5.2218574733421455e-06, "loss": 0.693662703037262, "step": 284 }, { "epoch": 1.2170212765957447, "grad_norm": 0.22191239893436432, "learning_rate": 5.345777054854578e-06, "loss": 0.9797627329826355, "step": 286 }, { "epoch": 1.225531914893617, "grad_norm": 0.17303584516048431, "learning_rate": 5.479146430146783e-06, "loss": 0.9435557126998901, "step": 288 }, { "epoch": 1.2340425531914894, "grad_norm": 0.2218431681394577, "learning_rate": 5.6218597281258834e-06, "loss": 0.6460038423538208, "step": 290 }, { "epoch": 1.2425531914893617, "grad_norm": 0.22332793474197388, "learning_rate": 5.773803660319232e-06, "loss": 0.8468654155731201, "step": 292 }, { "epoch": 1.251063829787234, "grad_norm": 0.13609401881694794, "learning_rate": 5.934857610804904e-06, "loss": 0.8879528641700745, "step": 294 }, { "epoch": 1.2595744680851064, "grad_norm": 0.18097440898418427, "learning_rate": 6.1048937319588676e-06, "loss": 0.7949476838111877, "step": 296 }, { "epoch": 1.2680851063829788, "grad_norm": 0.3170534074306488, "learning_rate": 6.283777045942798e-06, "loss": 0.8318912386894226, "step": 298 }, { "epoch": 1.2765957446808511, "grad_norm": 0.19770559668540955, "learning_rate": 6.471365551852011e-06, "loss": 1.0059137344360352, "step": 300 }, { "epoch": 1.2851063829787235, "grad_norm": 0.14630961418151855, "learning_rate": 6.66751033843842e-06, "loss": 0.39432206749916077, "step": 302 }, { "epoch": 1.2936170212765958, "grad_norm": 0.3450065553188324, "learning_rate": 6.872055702319048e-06, "loss": 0.9369863271713257, "step": 304 }, { "epoch": 1.302127659574468, "grad_norm": 0.18490372598171234, "learning_rate": 7.084839271576289e-06, "loss": 0.9826689958572388, "step": 306 }, { "epoch": 1.3106382978723405, "grad_norm": 0.1866351068019867, "learning_rate": 7.30569213465174e-06, "loss": 0.5573344826698303, "step": 308 }, { "epoch": 1.3191489361702127, "grad_norm": 0.1976342350244522, "learning_rate": 7.534438974431351e-06, "loss": 0.5633378624916077, "step": 310 }, { "epoch": 1.327659574468085, "grad_norm": 0.3162063956260681, "learning_rate": 7.770898207415414e-06, "loss": 0.7375507354736328, "step": 312 }, { "epoch": 1.3361702127659574, "grad_norm": 0.2016829252243042, "learning_rate": 8.014882127862923e-06, "loss": 0.4668198823928833, "step": 314 }, { "epoch": 1.3446808510638297, "grad_norm": 1.459976315498352, "learning_rate": 8.266197056795886e-06, "loss": 0.7280799150466919, "step": 316 }, { "epoch": 1.353191489361702, "grad_norm": 0.3398215174674988, "learning_rate": 8.5246434957453e-06, "loss": 0.5291551351547241, "step": 318 }, { "epoch": 1.3617021276595744, "grad_norm": 0.19973713159561157, "learning_rate": 8.790016285116766e-06, "loss": 0.5350017547607422, "step": 320 }, { "epoch": 1.3702127659574468, "grad_norm": 0.14308671653270721, "learning_rate": 9.062104767049957e-06, "loss": 0.9615387916564941, "step": 322 }, { "epoch": 1.3787234042553191, "grad_norm": 0.1939127892255783, "learning_rate": 9.340692952642787e-06, "loss": 0.6444461345672607, "step": 324 }, { "epoch": 1.3872340425531915, "grad_norm": 0.18625043332576752, "learning_rate": 9.625559693407413e-06, "loss": 0.9194571375846863, "step": 326 }, { "epoch": 1.3957446808510638, "grad_norm": 0.2951974868774414, "learning_rate": 9.916478856822006e-06, "loss": 0.9841423034667969, "step": 328 }, { "epoch": 1.4042553191489362, "grad_norm": 0.16953378915786743, "learning_rate": 1.0213219505838985e-05, "loss": 0.8577695488929749, "step": 330 }, { "epoch": 1.4127659574468086, "grad_norm": 0.2879262864589691, "learning_rate": 1.0515546082207094e-05, "loss": 0.8521637916564941, "step": 332 }, { "epoch": 1.421276595744681, "grad_norm": 0.22187970578670502, "learning_rate": 1.082321859346199e-05, "loss": 0.6609964966773987, "step": 334 }, { "epoch": 1.4297872340425533, "grad_norm": 0.19994378089904785, "learning_rate": 1.1135992803436696e-05, "loss": 0.5523483753204346, "step": 336 }, { "epoch": 1.4382978723404256, "grad_norm": 0.19332869350910187, "learning_rate": 1.1453620426140791e-05, "loss": 0.7419446110725403, "step": 338 }, { "epoch": 1.4468085106382977, "grad_norm": 0.17565295100212097, "learning_rate": 1.1775849322854508e-05, "loss": 0.7033047676086426, "step": 340 }, { "epoch": 1.4553191489361703, "grad_norm": 0.3124133050441742, "learning_rate": 1.210242370228112e-05, "loss": 0.5836398005485535, "step": 342 }, { "epoch": 1.4638297872340424, "grad_norm": 0.21754996478557587, "learning_rate": 1.2433084323598791e-05, "loss": 0.9612702131271362, "step": 344 }, { "epoch": 1.472340425531915, "grad_norm": 0.21039743721485138, "learning_rate": 1.2767568702250838e-05, "loss": 0.9444953799247742, "step": 346 }, { "epoch": 1.4808510638297872, "grad_norm": 0.42962566018104553, "learning_rate": 1.3105611318310811e-05, "loss": 0.8060452938079834, "step": 348 }, { "epoch": 1.4893617021276595, "grad_norm": 0.19513703882694244, "learning_rate": 1.3446943827257184e-05, "loss": 0.8150795698165894, "step": 350 }, { "epoch": 1.4978723404255319, "grad_norm": 0.2640067934989929, "learning_rate": 1.3791295272990172e-05, "loss": 0.3241785168647766, "step": 352 }, { "epoch": 1.5063829787234042, "grad_norm": 0.25144851207733154, "learning_rate": 1.413839230292182e-05, "loss": 0.7416596412658691, "step": 354 }, { "epoch": 1.5148936170212766, "grad_norm": 0.16774724423885345, "learning_rate": 1.4487959384968276e-05, "loss": 0.518079400062561, "step": 356 }, { "epoch": 1.523404255319149, "grad_norm": 0.1518772393465042, "learning_rate": 1.4839719026272373e-05, "loss": 0.5670950412750244, "step": 358 }, { "epoch": 1.5319148936170213, "grad_norm": 0.23323562741279602, "learning_rate": 1.5193391993482579e-05, "loss": 1.1480307579040527, "step": 360 }, { "epoch": 1.5404255319148936, "grad_norm": 0.1399996131658554, "learning_rate": 1.5548697534413636e-05, "loss": 0.5208263397216797, "step": 362 }, { "epoch": 1.548936170212766, "grad_norm": 0.19347982108592987, "learning_rate": 1.5905353600912894e-05, "loss": 0.9330644607543945, "step": 364 }, { "epoch": 1.5574468085106383, "grad_norm": 0.15133146941661835, "learning_rate": 1.6263077072755323e-05, "loss": 0.8360081911087036, "step": 366 }, { "epoch": 1.5659574468085107, "grad_norm": 0.18284079432487488, "learning_rate": 1.6621583982389707e-05, "loss": 0.7362462878227234, "step": 368 }, { "epoch": 1.574468085106383, "grad_norm": 0.19372136890888214, "learning_rate": 1.6980589740357294e-05, "loss": 0.6744348406791687, "step": 370 }, { "epoch": 1.5829787234042554, "grad_norm": 0.12256433069705963, "learning_rate": 1.733980936120425e-05, "loss": 0.9976662397384644, "step": 372 }, { "epoch": 1.5914893617021275, "grad_norm": 0.20529648661613464, "learning_rate": 1.7698957689708416e-05, "loss": 1.0515512228012085, "step": 374 }, { "epoch": 1.6, "grad_norm": 0.16672298312187195, "learning_rate": 1.805774962724083e-05, "loss": 0.7891988754272461, "step": 376 }, { "epoch": 1.6085106382978722, "grad_norm": 0.1434396356344223, "learning_rate": 1.8415900358082265e-05, "loss": 0.6114462018013, "step": 378 }, { "epoch": 1.6170212765957448, "grad_norm": 0.16634748876094818, "learning_rate": 1.8773125575515357e-05, "loss": 0.5973749756813049, "step": 380 }, { "epoch": 1.625531914893617, "grad_norm": 0.13240687549114227, "learning_rate": 1.912914170751251e-05, "loss": 0.7943100929260254, "step": 382 }, { "epoch": 1.6340425531914895, "grad_norm": 0.12580837309360504, "learning_rate": 1.9483666141840612e-05, "loss": 0.9919679760932922, "step": 384 }, { "epoch": 1.6425531914893616, "grad_norm": 0.35519301891326904, "learning_rate": 1.9836417450403974e-05, "loss": 0.8425494432449341, "step": 386 }, { "epoch": 1.6510638297872342, "grad_norm": 0.2488897144794464, "learning_rate": 2.0187115612647133e-05, "loss": 0.7412829399108887, "step": 388 }, { "epoch": 1.6595744680851063, "grad_norm": 0.5074354410171509, "learning_rate": 2.05354822378404e-05, "loss": 0.7992886900901794, "step": 390 }, { "epoch": 1.6680851063829787, "grad_norm": 0.12573714554309845, "learning_rate": 2.0881240786071595e-05, "loss": 0.909046471118927, "step": 392 }, { "epoch": 1.676595744680851, "grad_norm": 0.33457598090171814, "learning_rate": 2.122411678776854e-05, "loss": 0.6243407726287842, "step": 394 }, { "epoch": 1.6851063829787234, "grad_norm": 0.15411563217639923, "learning_rate": 2.1563838061578258e-05, "loss": 0.8719924688339233, "step": 396 }, { "epoch": 1.6936170212765957, "grad_norm": 0.13289184868335724, "learning_rate": 2.1900134930429396e-05, "loss": 1.0623129606246948, "step": 398 }, { "epoch": 1.702127659574468, "grad_norm": 0.3402920663356781, "learning_rate": 2.223274043560706e-05, "loss": 0.9607300162315369, "step": 400 }, { "epoch": 1.7106382978723405, "grad_norm": 0.20696532726287842, "learning_rate": 2.256139054866954e-05, "loss": 0.4819542169570923, "step": 402 }, { "epoch": 1.7191489361702128, "grad_norm": 0.1925085484981537, "learning_rate": 2.2885824381039028e-05, "loss": 0.6816413402557373, "step": 404 }, { "epoch": 1.7276595744680852, "grad_norm": 0.5462839603424072, "learning_rate": 2.3205784391099808e-05, "loss": 0.5322574377059937, "step": 406 }, { "epoch": 1.7361702127659573, "grad_norm": 0.12952132523059845, "learning_rate": 2.352101658863958e-05, "loss": 0.8366644382476807, "step": 408 }, { "epoch": 1.7446808510638299, "grad_norm": 0.15898458659648895, "learning_rate": 2.3831270736471703e-05, "loss": 0.8219179511070251, "step": 410 }, { "epoch": 1.753191489361702, "grad_norm": 0.10134760290384293, "learning_rate": 2.4136300549077973e-05, "loss": 0.8943849802017212, "step": 412 }, { "epoch": 1.7617021276595746, "grad_norm": 0.21209417283535004, "learning_rate": 2.443586388811481e-05, "loss": 0.7499610781669617, "step": 414 }, { "epoch": 1.7702127659574467, "grad_norm": 0.47646665573120117, "learning_rate": 2.47297229546271e-05, "loss": 0.6121576428413391, "step": 416 }, { "epoch": 1.7787234042553193, "grad_norm": 0.26675277948379517, "learning_rate": 2.5017644477817418e-05, "loss": 0.7867974638938904, "step": 418 }, { "epoch": 1.7872340425531914, "grad_norm": 0.840756356716156, "learning_rate": 2.5299399900220807e-05, "loss": 0.6045423150062561, "step": 420 }, { "epoch": 1.795744680851064, "grad_norm": 0.400618314743042, "learning_rate": 2.5574765559137848e-05, "loss": 0.7998336553573608, "step": 422 }, { "epoch": 1.804255319148936, "grad_norm": 0.15446501970291138, "learning_rate": 2.584352286418239e-05, "loss": 0.7926132678985596, "step": 424 }, { "epoch": 1.8127659574468085, "grad_norm": 0.15063929557800293, "learning_rate": 2.6105458470802563e-05, "loss": 0.8235700130462646, "step": 426 }, { "epoch": 1.8212765957446808, "grad_norm": 0.11289381980895996, "learning_rate": 2.6360364449637683e-05, "loss": 0.8320169448852539, "step": 428 }, { "epoch": 1.8297872340425532, "grad_norm": 0.14190740883350372, "learning_rate": 2.6608038451576528e-05, "loss": 0.8250552415847778, "step": 430 }, { "epoch": 1.8382978723404255, "grad_norm": 0.3685653507709503, "learning_rate": 2.684828386838569e-05, "loss": 0.8021817207336426, "step": 432 }, { "epoch": 1.8468085106382979, "grad_norm": 0.12853652238845825, "learning_rate": 2.7080909988780982e-05, "loss": 0.7879712581634521, "step": 434 }, { "epoch": 1.8553191489361702, "grad_norm": 0.3018054962158203, "learning_rate": 2.7305732149817502e-05, "loss": 0.7762024998664856, "step": 436 }, { "epoch": 1.8638297872340426, "grad_norm": 0.14753182232379913, "learning_rate": 2.7522571883478617e-05, "loss": 1.2395200729370117, "step": 438 }, { "epoch": 1.872340425531915, "grad_norm": 0.1134040355682373, "learning_rate": 2.77312570583471e-05, "loss": 0.8664835691452026, "step": 440 }, { "epoch": 1.8808510638297873, "grad_norm": 0.19325752556324005, "learning_rate": 2.7931622016246304e-05, "loss": 1.0493484735488892, "step": 442 }, { "epoch": 1.8893617021276596, "grad_norm": 0.15366853773593903, "learning_rate": 2.8123507703742727e-05, "loss": 0.8971328139305115, "step": 444 }, { "epoch": 1.8978723404255318, "grad_norm": 0.1713661104440689, "learning_rate": 2.8306761798405522e-05, "loss": 0.9821380376815796, "step": 446 }, { "epoch": 1.9063829787234043, "grad_norm": 0.4014803469181061, "learning_rate": 2.848123882972295e-05, "loss": 1.0046873092651367, "step": 448 }, { "epoch": 1.9148936170212765, "grad_norm": 0.1893383413553238, "learning_rate": 2.8646800294579514e-05, "loss": 0.647858738899231, "step": 450 }, { "epoch": 1.923404255319149, "grad_norm": 0.5622705221176147, "learning_rate": 2.8803314767202376e-05, "loss": 0.5715224742889404, "step": 452 }, { "epoch": 1.9319148936170212, "grad_norm": 0.14517684280872345, "learning_rate": 2.8950658003489534e-05, "loss": 0.8555135726928711, "step": 454 }, { "epoch": 1.9404255319148938, "grad_norm": 0.16155880689620972, "learning_rate": 2.908871303963711e-05, "loss": 0.9127392768859863, "step": 456 }, { "epoch": 1.9489361702127659, "grad_norm": 0.12455891817808151, "learning_rate": 2.9217370284987434e-05, "loss": 0.9905153512954712, "step": 458 }, { "epoch": 1.9574468085106385, "grad_norm": 0.1255619376897812, "learning_rate": 2.9336527609024075e-05, "loss": 1.022163987159729, "step": 460 }, { "epoch": 1.9659574468085106, "grad_norm": 0.38456299901008606, "learning_rate": 2.9446090422445016e-05, "loss": 0.8251804709434509, "step": 462 }, { "epoch": 1.974468085106383, "grad_norm": 0.12075914442539215, "learning_rate": 2.9545971752249376e-05, "loss": 1.0136280059814453, "step": 464 }, { "epoch": 1.9829787234042553, "grad_norm": 0.20235367119312286, "learning_rate": 2.9636092310778195e-05, "loss": 0.9567205905914307, "step": 466 }, { "epoch": 1.9914893617021276, "grad_norm": 0.542694091796875, "learning_rate": 2.9716380558654445e-05, "loss": 0.6730895042419434, "step": 468 }, { "epoch": 2.0, "grad_norm": 0.11073193699121475, "learning_rate": 2.9786772761572338e-05, "loss": 0.6498197317123413, "step": 470 }, { "epoch": 2.008510638297872, "grad_norm": 0.11644601076841354, "learning_rate": 2.9847213040890793e-05, "loss": 0.4679168164730072, "step": 472 }, { "epoch": 2.0170212765957447, "grad_norm": 0.14129982888698578, "learning_rate": 2.989765341799095e-05, "loss": 0.7698261737823486, "step": 474 }, { "epoch": 2.025531914893617, "grad_norm": 0.13531845808029175, "learning_rate": 2.9938053852362484e-05, "loss": 0.47750866413116455, "step": 476 }, { "epoch": 2.0340425531914894, "grad_norm": 0.37192943692207336, "learning_rate": 2.99683822733885e-05, "loss": 0.6370930075645447, "step": 478 }, { "epoch": 2.0425531914893615, "grad_norm": 0.13162164390087128, "learning_rate": 2.9988614605803806e-05, "loss": 0.44695600867271423, "step": 480 }, { "epoch": 2.051063829787234, "grad_norm": 0.20202530920505524, "learning_rate": 2.9998734788806287e-05, "loss": 0.4687855839729309, "step": 482 }, { "epoch": 2.0595744680851062, "grad_norm": 0.09361886233091354, "learning_rate": 2.9998734788806287e-05, "loss": 0.28051936626434326, "step": 484 }, { "epoch": 2.068085106382979, "grad_norm": 0.1879085749387741, "learning_rate": 2.9988614605803806e-05, "loss": 0.6579060554504395, "step": 486 }, { "epoch": 2.076595744680851, "grad_norm": 0.24100597202777863, "learning_rate": 2.99683822733885e-05, "loss": 0.5791319608688354, "step": 488 }, { "epoch": 2.0851063829787235, "grad_norm": 0.15436111390590668, "learning_rate": 2.9938053852362484e-05, "loss": 0.5440673828125, "step": 490 }, { "epoch": 2.0936170212765957, "grad_norm": 0.1665167361497879, "learning_rate": 2.989765341799095e-05, "loss": 0.7786204218864441, "step": 492 }, { "epoch": 2.1021276595744682, "grad_norm": 0.12963709235191345, "learning_rate": 2.9847213040890793e-05, "loss": 0.6289861798286438, "step": 494 }, { "epoch": 2.1106382978723404, "grad_norm": 0.23455171287059784, "learning_rate": 2.9786772761572338e-05, "loss": 0.7049335241317749, "step": 496 }, { "epoch": 2.119148936170213, "grad_norm": 0.10971709340810776, "learning_rate": 2.9716380558654445e-05, "loss": 0.9018527269363403, "step": 498 }, { "epoch": 2.127659574468085, "grad_norm": 0.3122045695781708, "learning_rate": 2.9636092310778195e-05, "loss": 0.6065503358840942, "step": 500 }, { "epoch": 2.1361702127659576, "grad_norm": 0.15959368646144867, "learning_rate": 2.9545971752249383e-05, "loss": 0.5845508575439453, "step": 502 }, { "epoch": 2.1446808510638298, "grad_norm": 0.12507954239845276, "learning_rate": 2.9446090422445016e-05, "loss": 0.6952933669090271, "step": 504 }, { "epoch": 2.153191489361702, "grad_norm": 0.12080808728933334, "learning_rate": 2.933652760902408e-05, "loss": 0.7144219279289246, "step": 506 }, { "epoch": 2.1617021276595745, "grad_norm": 0.3896176815032959, "learning_rate": 2.9217370284987434e-05, "loss": 0.5674622654914856, "step": 508 }, { "epoch": 2.1702127659574466, "grad_norm": 0.14054062962532043, "learning_rate": 2.908871303963712e-05, "loss": 0.28315988183021545, "step": 510 }, { "epoch": 2.178723404255319, "grad_norm": 0.1403239667415619, "learning_rate": 2.8950658003489534e-05, "loss": 0.430463582277298, "step": 512 }, { "epoch": 2.1872340425531913, "grad_norm": 0.14137110114097595, "learning_rate": 2.8803314767202387e-05, "loss": 0.4114861488342285, "step": 514 }, { "epoch": 2.195744680851064, "grad_norm": 0.1280188113451004, "learning_rate": 2.864680029457952e-05, "loss": 0.5308094620704651, "step": 516 }, { "epoch": 2.204255319148936, "grad_norm": 0.25649434328079224, "learning_rate": 2.8481238829722946e-05, "loss": 0.6915931701660156, "step": 518 }, { "epoch": 2.2127659574468086, "grad_norm": 0.10003609210252762, "learning_rate": 2.8306761798405533e-05, "loss": 0.4150570333003998, "step": 520 }, { "epoch": 2.2212765957446807, "grad_norm": 0.15974430739879608, "learning_rate": 2.812350770374273e-05, "loss": 0.36256328225135803, "step": 522 }, { "epoch": 2.2297872340425533, "grad_norm": 0.38845136761665344, "learning_rate": 2.793162201624631e-05, "loss": 0.37502366304397583, "step": 524 }, { "epoch": 2.2382978723404254, "grad_norm": 0.1723078191280365, "learning_rate": 2.7731257058347098e-05, "loss": 0.7607256770133972, "step": 526 }, { "epoch": 2.246808510638298, "grad_norm": 0.255744993686676, "learning_rate": 2.752257188347862e-05, "loss": 0.6797502040863037, "step": 528 }, { "epoch": 2.25531914893617, "grad_norm": 0.3702438771724701, "learning_rate": 2.7305732149817512e-05, "loss": 0.7978914380073547, "step": 530 }, { "epoch": 2.2638297872340427, "grad_norm": 0.26327913999557495, "learning_rate": 2.708090998878098e-05, "loss": 0.5080232620239258, "step": 532 }, { "epoch": 2.272340425531915, "grad_norm": 0.14510440826416016, "learning_rate": 2.68482838683857e-05, "loss": 0.2904958724975586, "step": 534 }, { "epoch": 2.2808510638297874, "grad_norm": 0.1330595761537552, "learning_rate": 2.6608038451576528e-05, "loss": 0.5988547205924988, "step": 536 }, { "epoch": 2.2893617021276595, "grad_norm": 0.15609091520309448, "learning_rate": 2.6360364449637686e-05, "loss": 0.8217217922210693, "step": 538 }, { "epoch": 2.297872340425532, "grad_norm": 0.27816882729530334, "learning_rate": 2.6105458470802563e-05, "loss": 0.29454290866851807, "step": 540 }, { "epoch": 2.3063829787234043, "grad_norm": 0.16807164251804352, "learning_rate": 2.5843522864182397e-05, "loss": 0.2661294937133789, "step": 542 }, { "epoch": 2.3148936170212764, "grad_norm": 0.2340618073940277, "learning_rate": 2.557476555913786e-05, "loss": 0.4138932228088379, "step": 544 }, { "epoch": 2.323404255319149, "grad_norm": 0.12209529429674149, "learning_rate": 2.52993999002208e-05, "loss": 0.47122570872306824, "step": 546 }, { "epoch": 2.331914893617021, "grad_norm": 0.22724129259586334, "learning_rate": 2.501764447781743e-05, "loss": 0.5940241813659668, "step": 548 }, { "epoch": 2.3404255319148937, "grad_norm": 0.10988523811101913, "learning_rate": 2.4729722954627106e-05, "loss": 0.443272203207016, "step": 550 }, { "epoch": 2.348936170212766, "grad_norm": 0.3907153904438019, "learning_rate": 2.4435863888114814e-05, "loss": 0.612157940864563, "step": 552 }, { "epoch": 2.3574468085106384, "grad_norm": 0.15928088128566742, "learning_rate": 2.413630054907798e-05, "loss": 0.6884116530418396, "step": 554 }, { "epoch": 2.3659574468085105, "grad_norm": 0.12084610015153885, "learning_rate": 2.383127073647171e-05, "loss": 0.6459460854530334, "step": 556 }, { "epoch": 2.374468085106383, "grad_norm": 0.14340610802173615, "learning_rate": 2.3521016588639598e-05, "loss": 0.5485546588897705, "step": 558 }, { "epoch": 2.382978723404255, "grad_norm": 0.1330912858247757, "learning_rate": 2.32057843910998e-05, "loss": 0.6871886253356934, "step": 560 }, { "epoch": 2.391489361702128, "grad_norm": 0.1461075246334076, "learning_rate": 2.2885824381039024e-05, "loss": 0.5943357944488525, "step": 562 }, { "epoch": 2.4, "grad_norm": 0.4632057547569275, "learning_rate": 2.2561390548669552e-05, "loss": 0.44520270824432373, "step": 564 }, { "epoch": 2.4085106382978725, "grad_norm": 0.16760534048080444, "learning_rate": 2.2232740435607067e-05, "loss": 0.4427604377269745, "step": 566 }, { "epoch": 2.4170212765957446, "grad_norm": 0.17341409623622894, "learning_rate": 2.1900134930429403e-05, "loss": 0.7208576798439026, "step": 568 }, { "epoch": 2.425531914893617, "grad_norm": 0.12514737248420715, "learning_rate": 2.1563838061578264e-05, "loss": 0.6061448454856873, "step": 570 }, { "epoch": 2.4340425531914893, "grad_norm": 0.275291383266449, "learning_rate": 2.122411678776855e-05, "loss": 0.48493388295173645, "step": 572 }, { "epoch": 2.4425531914893615, "grad_norm": 0.35213977098464966, "learning_rate": 2.0881240786071588e-05, "loss": 0.6101267337799072, "step": 574 }, { "epoch": 2.451063829787234, "grad_norm": 0.3173239827156067, "learning_rate": 2.0535482237840398e-05, "loss": 0.6204816102981567, "step": 576 }, { "epoch": 2.4595744680851066, "grad_norm": 0.12788061797618866, "learning_rate": 2.018711561264715e-05, "loss": 0.5858104825019836, "step": 578 }, { "epoch": 2.4680851063829787, "grad_norm": 0.21373364329338074, "learning_rate": 1.9836417450403978e-05, "loss": 0.5342880487442017, "step": 580 }, { "epoch": 2.476595744680851, "grad_norm": 0.1782151162624359, "learning_rate": 1.948366614184062e-05, "loss": 0.6852924823760986, "step": 582 }, { "epoch": 2.4851063829787234, "grad_norm": 0.12118978798389435, "learning_rate": 1.9129141707512514e-05, "loss": 0.5905125737190247, "step": 584 }, { "epoch": 2.4936170212765956, "grad_norm": 0.6266365647315979, "learning_rate": 1.877312557551536e-05, "loss": 0.6729872822761536, "step": 586 }, { "epoch": 2.502127659574468, "grad_norm": 0.13392919301986694, "learning_rate": 1.841590035808227e-05, "loss": 0.7322168350219727, "step": 588 }, { "epoch": 2.5106382978723403, "grad_norm": 0.27286240458488464, "learning_rate": 1.8057749627240822e-05, "loss": 0.6285769939422607, "step": 590 }, { "epoch": 2.519148936170213, "grad_norm": 0.10404688119888306, "learning_rate": 1.7698957689708436e-05, "loss": 0.766042172908783, "step": 592 }, { "epoch": 2.527659574468085, "grad_norm": 0.11238724738359451, "learning_rate": 1.7339809361204255e-05, "loss": 0.43586933612823486, "step": 594 }, { "epoch": 2.5361702127659576, "grad_norm": 0.12042172253131866, "learning_rate": 1.6980589740357287e-05, "loss": 0.6923111081123352, "step": 596 }, { "epoch": 2.5446808510638297, "grad_norm": 0.2841528058052063, "learning_rate": 1.6621583982389714e-05, "loss": 0.47686851024627686, "step": 598 }, { "epoch": 2.5531914893617023, "grad_norm": 0.11969015002250671, "learning_rate": 1.626307707275533e-05, "loss": 0.5013985633850098, "step": 600 }, { "epoch": 2.5617021276595744, "grad_norm": 0.09876365959644318, "learning_rate": 1.59053536009129e-05, "loss": 0.6609531044960022, "step": 602 }, { "epoch": 2.570212765957447, "grad_norm": 0.4649478495121002, "learning_rate": 1.5548697534413653e-05, "loss": 0.6927770972251892, "step": 604 }, { "epoch": 2.578723404255319, "grad_norm": 0.1714295893907547, "learning_rate": 1.5193391993482573e-05, "loss": 0.8361954689025879, "step": 606 }, { "epoch": 2.5872340425531917, "grad_norm": 0.15755438804626465, "learning_rate": 1.483971902627239e-05, "loss": 0.5158759951591492, "step": 608 }, { "epoch": 2.595744680851064, "grad_norm": 0.11576741188764572, "learning_rate": 1.4487959384968272e-05, "loss": 0.6141366362571716, "step": 610 }, { "epoch": 2.604255319148936, "grad_norm": 0.2397107034921646, "learning_rate": 1.4138392302921813e-05, "loss": 0.669286847114563, "step": 612 }, { "epoch": 2.6127659574468085, "grad_norm": 0.17677916586399078, "learning_rate": 1.3791295272990178e-05, "loss": 0.8155961036682129, "step": 614 }, { "epoch": 2.621276595744681, "grad_norm": 0.10534735023975372, "learning_rate": 1.344694382725719e-05, "loss": 0.6486954689025879, "step": 616 }, { "epoch": 2.629787234042553, "grad_norm": 0.17969125509262085, "learning_rate": 1.3105611318310828e-05, "loss": 0.7624613046646118, "step": 618 }, { "epoch": 2.6382978723404253, "grad_norm": 0.10477960854768753, "learning_rate": 1.2767568702250834e-05, "loss": 0.44974949955940247, "step": 620 }, { "epoch": 2.646808510638298, "grad_norm": 0.41400146484375, "learning_rate": 1.2433084323598806e-05, "loss": 0.5666506886482239, "step": 622 }, { "epoch": 2.65531914893617, "grad_norm": 0.12478289753198624, "learning_rate": 1.2102423702281114e-05, "loss": 0.7800987362861633, "step": 624 }, { "epoch": 2.6638297872340426, "grad_norm": 0.16413263976573944, "learning_rate": 1.1775849322854515e-05, "loss": 0.5514877438545227, "step": 626 }, { "epoch": 2.6723404255319148, "grad_norm": 0.09944586455821991, "learning_rate": 1.1453620426140796e-05, "loss": 0.554164469242096, "step": 628 }, { "epoch": 2.6808510638297873, "grad_norm": 0.2592844069004059, "learning_rate": 1.1135992803436701e-05, "loss": 0.5563373565673828, "step": 630 }, { "epoch": 2.6893617021276595, "grad_norm": 0.2102140486240387, "learning_rate": 1.0823218593462006e-05, "loss": 0.5627699494361877, "step": 632 }, { "epoch": 2.697872340425532, "grad_norm": 0.1483510136604309, "learning_rate": 1.0515546082207089e-05, "loss": 0.698486328125, "step": 634 }, { "epoch": 2.706382978723404, "grad_norm": 0.11229369789361954, "learning_rate": 1.0213219505839e-05, "loss": 0.7639659643173218, "step": 636 }, { "epoch": 2.7148936170212767, "grad_norm": 0.10941766947507858, "learning_rate": 9.916478856822011e-06, "loss": 0.6158884763717651, "step": 638 }, { "epoch": 2.723404255319149, "grad_norm": 0.1750131994485855, "learning_rate": 9.625559693407416e-06, "loss": 0.9152920842170715, "step": 640 }, { "epoch": 2.731914893617021, "grad_norm": 0.18939398229122162, "learning_rate": 9.34069295264279e-06, "loss": 0.8749684691429138, "step": 642 }, { "epoch": 2.7404255319148936, "grad_norm": 0.0939684808254242, "learning_rate": 9.062104767049962e-06, "loss": 0.8344287872314453, "step": 644 }, { "epoch": 2.748936170212766, "grad_norm": 0.10787810385227203, "learning_rate": 8.790016285116773e-06, "loss": 0.45629745721817017, "step": 646 }, { "epoch": 2.7574468085106383, "grad_norm": 0.10236497223377228, "learning_rate": 8.5246434957453e-06, "loss": 0.3841492533683777, "step": 648 }, { "epoch": 2.7659574468085104, "grad_norm": 0.31643790006637573, "learning_rate": 8.266197056795886e-06, "loss": 0.4197019338607788, "step": 650 }, { "epoch": 2.774468085106383, "grad_norm": 0.15869994461536407, "learning_rate": 8.014882127862926e-06, "loss": 0.6872882843017578, "step": 652 }, { "epoch": 2.7829787234042556, "grad_norm": 0.15020446479320526, "learning_rate": 7.77089820741542e-06, "loss": 0.6434124112129211, "step": 654 }, { "epoch": 2.7914893617021277, "grad_norm": 0.13434921205043793, "learning_rate": 7.534438974431356e-06, "loss": 0.5578271150588989, "step": 656 }, { "epoch": 2.8, "grad_norm": 0.1459910273551941, "learning_rate": 7.305692134651748e-06, "loss": 0.6800917983055115, "step": 658 }, { "epoch": 2.8085106382978724, "grad_norm": 0.13331423699855804, "learning_rate": 7.0848392715763e-06, "loss": 0.4372290074825287, "step": 660 }, { "epoch": 2.8170212765957445, "grad_norm": 0.11432936787605286, "learning_rate": 6.872055702319048e-06, "loss": 0.3426898121833801, "step": 662 }, { "epoch": 2.825531914893617, "grad_norm": 0.47585824131965637, "learning_rate": 6.66751033843842e-06, "loss": 0.5213098526000977, "step": 664 }, { "epoch": 2.8340425531914892, "grad_norm": 0.23138658702373505, "learning_rate": 6.471365551852014e-06, "loss": 0.40890708565711975, "step": 666 }, { "epoch": 2.842553191489362, "grad_norm": 0.1170782670378685, "learning_rate": 6.283777045942801e-06, "loss": 0.6171141266822815, "step": 668 }, { "epoch": 2.851063829787234, "grad_norm": 0.2987634837627411, "learning_rate": 6.104893731958872e-06, "loss": 0.79531329870224, "step": 670 }, { "epoch": 2.8595744680851065, "grad_norm": 0.16172188520431519, "learning_rate": 5.934857610804912e-06, "loss": 0.6109257936477661, "step": 672 }, { "epoch": 2.8680851063829786, "grad_norm": 0.19967441260814667, "learning_rate": 5.773803660319231e-06, "loss": 0.7840122580528259, "step": 674 }, { "epoch": 2.876595744680851, "grad_norm": 0.09976833313703537, "learning_rate": 5.6218597281258834e-06, "loss": 0.4490019679069519, "step": 676 }, { "epoch": 2.8851063829787233, "grad_norm": 0.1243966668844223, "learning_rate": 5.479146430146783e-06, "loss": 0.5424228310585022, "step": 678 }, { "epoch": 2.8936170212765955, "grad_norm": 0.17045721411705017, "learning_rate": 5.3457770548545805e-06, "loss": 0.5151770114898682, "step": 680 }, { "epoch": 2.902127659574468, "grad_norm": 0.1732950657606125, "learning_rate": 5.22185747334215e-06, "loss": 0.7383885979652405, "step": 682 }, { "epoch": 2.9106382978723406, "grad_norm": 0.21424034237861633, "learning_rate": 5.107486055280145e-06, "loss": 0.4885507822036743, "step": 684 }, { "epoch": 2.9191489361702128, "grad_norm": 1.2460668087005615, "learning_rate": 5.002753590829352e-06, "loss": 0.29081299901008606, "step": 686 }, { "epoch": 2.927659574468085, "grad_norm": 0.11700358986854553, "learning_rate": 4.9077432185697e-06, "loss": 0.436556339263916, "step": 688 }, { "epoch": 2.9361702127659575, "grad_norm": 0.2099040299654007, "learning_rate": 4.822530359503391e-06, "loss": 0.759993314743042, "step": 690 }, { "epoch": 2.94468085106383, "grad_norm": 0.12797491252422333, "learning_rate": 4.747182657184251e-06, "loss": 0.6529436111450195, "step": 692 }, { "epoch": 2.953191489361702, "grad_norm": 0.122190460562706, "learning_rate": 4.681759924021033e-06, "loss": 0.6622499227523804, "step": 694 }, { "epoch": 2.9617021276595743, "grad_norm": 0.19574476778507233, "learning_rate": 4.626314093797213e-06, "loss": 0.5456458926200867, "step": 696 }, { "epoch": 2.970212765957447, "grad_norm": 0.15566527843475342, "learning_rate": 4.580889180444988e-06, "loss": 0.738301157951355, "step": 698 }, { "epoch": 2.978723404255319, "grad_norm": 0.13139484822750092, "learning_rate": 4.5455212431061985e-06, "loss": 0.4914497435092926, "step": 700 }, { "epoch": 2.9872340425531916, "grad_norm": 0.24261365830898285, "learning_rate": 4.520238357507899e-06, "loss": 0.5938633680343628, "step": 702 }, { "epoch": 2.9957446808510637, "grad_norm": 0.09112333506345749, "learning_rate": 4.505060593675342e-06, "loss": 0.4166713058948517, "step": 704 }, { "epoch": 3.0, "step": 705, "total_flos": 3.0241141054567875e+18, "train_loss": 0.8476084547685393, "train_runtime": 9225.3576, "train_samples_per_second": 4.891, "train_steps_per_second": 0.076 } ], "logging_steps": 2, "max_steps": 705, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 99999, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.0241141054567875e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }