{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 2486, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004023335345001006, "grad_norm": 27.82010551409744, "learning_rate": 0.0, "loss": 1.7886, "step": 1 }, { "epoch": 0.0008046670690002012, "grad_norm": 19.872529474831897, "learning_rate": 2.008032128514056e-08, "loss": 1.8136, "step": 2 }, { "epoch": 0.0012070006035003018, "grad_norm": 24.180123303383308, "learning_rate": 4.016064257028112e-08, "loss": 1.9299, "step": 3 }, { "epoch": 0.0016093341380004024, "grad_norm": 21.489391381684396, "learning_rate": 6.02409638554217e-08, "loss": 1.92, "step": 4 }, { "epoch": 0.0020116676725005027, "grad_norm": 17.558745295130834, "learning_rate": 8.032128514056224e-08, "loss": 1.8227, "step": 5 }, { "epoch": 0.0024140012070006035, "grad_norm": 21.46699607887353, "learning_rate": 1.0040160642570281e-07, "loss": 1.8824, "step": 6 }, { "epoch": 0.002816334741500704, "grad_norm": 21.298475273313784, "learning_rate": 1.204819277108434e-07, "loss": 1.8179, "step": 7 }, { "epoch": 0.0032186682760008047, "grad_norm": 21.372007205869878, "learning_rate": 1.4056224899598394e-07, "loss": 1.6859, "step": 8 }, { "epoch": 0.003621001810500905, "grad_norm": 23.292357598058903, "learning_rate": 1.6064257028112448e-07, "loss": 2.0856, "step": 9 }, { "epoch": 0.0040233353450010055, "grad_norm": 23.233045584039576, "learning_rate": 1.8072289156626505e-07, "loss": 2.1395, "step": 10 }, { "epoch": 0.004425668879501106, "grad_norm": 19.4337570491022, "learning_rate": 2.0080321285140563e-07, "loss": 1.8688, "step": 11 }, { "epoch": 0.004828002414001207, "grad_norm": 24.82725637763247, "learning_rate": 2.208835341365462e-07, "loss": 1.8054, "step": 12 }, { "epoch": 0.005230335948501308, "grad_norm": 27.369177322875704, "learning_rate": 2.409638554216868e-07, "loss": 1.8794, "step": 13 }, { "epoch": 0.005632669483001408, "grad_norm": 20.854134960964704, "learning_rate": 2.610441767068273e-07, "loss": 1.8873, "step": 14 }, { "epoch": 0.006035003017501509, "grad_norm": 25.685971689153487, "learning_rate": 2.811244979919679e-07, "loss": 2.0743, "step": 15 }, { "epoch": 0.0064373365520016095, "grad_norm": 21.501547984115398, "learning_rate": 3.0120481927710845e-07, "loss": 1.8415, "step": 16 }, { "epoch": 0.00683967008650171, "grad_norm": 22.75991867094887, "learning_rate": 3.2128514056224897e-07, "loss": 1.9089, "step": 17 }, { "epoch": 0.00724200362100181, "grad_norm": 22.23601729043514, "learning_rate": 3.413654618473896e-07, "loss": 1.9641, "step": 18 }, { "epoch": 0.007644337155501911, "grad_norm": 26.35207563502954, "learning_rate": 3.614457831325301e-07, "loss": 2.0458, "step": 19 }, { "epoch": 0.008046670690002011, "grad_norm": 20.57786492934256, "learning_rate": 3.8152610441767073e-07, "loss": 1.7877, "step": 20 }, { "epoch": 0.008449004224502113, "grad_norm": 25.563694935501676, "learning_rate": 4.0160642570281125e-07, "loss": 2.1414, "step": 21 }, { "epoch": 0.008851337759002213, "grad_norm": 18.98280746175437, "learning_rate": 4.216867469879518e-07, "loss": 1.7995, "step": 22 }, { "epoch": 0.009253671293502314, "grad_norm": 22.278384224532818, "learning_rate": 4.417670682730924e-07, "loss": 2.0412, "step": 23 }, { "epoch": 0.009656004828002414, "grad_norm": 18.256806226009903, "learning_rate": 4.6184738955823296e-07, "loss": 1.7538, "step": 24 }, { "epoch": 0.010058338362502514, "grad_norm": 19.087838303092653, "learning_rate": 4.819277108433736e-07, "loss": 1.708, "step": 25 }, { "epoch": 0.010460671897002616, "grad_norm": 18.961034656693933, "learning_rate": 5.020080321285141e-07, "loss": 1.6661, "step": 26 }, { "epoch": 0.010863005431502716, "grad_norm": 18.11814523029599, "learning_rate": 5.220883534136546e-07, "loss": 1.753, "step": 27 }, { "epoch": 0.011265338966002816, "grad_norm": 19.63691340845534, "learning_rate": 5.421686746987952e-07, "loss": 1.7005, "step": 28 }, { "epoch": 0.011667672500502917, "grad_norm": 16.293048438980087, "learning_rate": 5.622489959839358e-07, "loss": 1.7269, "step": 29 }, { "epoch": 0.012070006035003017, "grad_norm": 15.86827584036101, "learning_rate": 5.823293172690764e-07, "loss": 1.5401, "step": 30 }, { "epoch": 0.012472339569503117, "grad_norm": 17.38302283506762, "learning_rate": 6.024096385542169e-07, "loss": 1.5943, "step": 31 }, { "epoch": 0.012874673104003219, "grad_norm": 15.374566189246599, "learning_rate": 6.224899598393574e-07, "loss": 1.6221, "step": 32 }, { "epoch": 0.013277006638503319, "grad_norm": 14.904500940732934, "learning_rate": 6.425702811244979e-07, "loss": 1.5452, "step": 33 }, { "epoch": 0.01367934017300342, "grad_norm": 16.29526851000047, "learning_rate": 6.626506024096387e-07, "loss": 1.2262, "step": 34 }, { "epoch": 0.01408167370750352, "grad_norm": 12.819998198911712, "learning_rate": 6.827309236947792e-07, "loss": 1.3476, "step": 35 }, { "epoch": 0.01448400724200362, "grad_norm": 13.25182925491938, "learning_rate": 7.028112449799197e-07, "loss": 1.4371, "step": 36 }, { "epoch": 0.014886340776503722, "grad_norm": 14.477871111934059, "learning_rate": 7.228915662650602e-07, "loss": 1.3476, "step": 37 }, { "epoch": 0.015288674311003822, "grad_norm": 14.6196272151239, "learning_rate": 7.429718875502008e-07, "loss": 1.5309, "step": 38 }, { "epoch": 0.015691007845503924, "grad_norm": 11.90353169885529, "learning_rate": 7.630522088353415e-07, "loss": 1.3026, "step": 39 }, { "epoch": 0.016093341380004022, "grad_norm": 13.271008324770724, "learning_rate": 7.83132530120482e-07, "loss": 1.4655, "step": 40 }, { "epoch": 0.016495674914504124, "grad_norm": 11.2349199407351, "learning_rate": 8.032128514056225e-07, "loss": 1.2782, "step": 41 }, { "epoch": 0.016898008449004225, "grad_norm": 9.663111453936613, "learning_rate": 8.232931726907631e-07, "loss": 1.0628, "step": 42 }, { "epoch": 0.017300341983504323, "grad_norm": 10.290302552273262, "learning_rate": 8.433734939759036e-07, "loss": 1.2398, "step": 43 }, { "epoch": 0.017702675518004425, "grad_norm": 9.437964794637868, "learning_rate": 8.634538152610443e-07, "loss": 1.1087, "step": 44 }, { "epoch": 0.018105009052504527, "grad_norm": 9.928972956598479, "learning_rate": 8.835341365461848e-07, "loss": 1.1231, "step": 45 }, { "epoch": 0.01850734258700463, "grad_norm": 7.193942700746645, "learning_rate": 9.036144578313254e-07, "loss": 0.9053, "step": 46 }, { "epoch": 0.018909676121504727, "grad_norm": 8.545887733782381, "learning_rate": 9.236947791164659e-07, "loss": 1.0188, "step": 47 }, { "epoch": 0.01931200965600483, "grad_norm": 9.964310833592453, "learning_rate": 9.437751004016064e-07, "loss": 1.2336, "step": 48 }, { "epoch": 0.01971434319050493, "grad_norm": 8.465102281501581, "learning_rate": 9.638554216867472e-07, "loss": 0.978, "step": 49 }, { "epoch": 0.020116676725005028, "grad_norm": 8.387761909932218, "learning_rate": 9.839357429718876e-07, "loss": 1.0664, "step": 50 }, { "epoch": 0.02051901025950513, "grad_norm": 8.853387211734997, "learning_rate": 1.0040160642570282e-06, "loss": 1.1316, "step": 51 }, { "epoch": 0.02092134379400523, "grad_norm": 8.36811488486395, "learning_rate": 1.0240963855421688e-06, "loss": 0.9175, "step": 52 }, { "epoch": 0.02132367732850533, "grad_norm": 8.394630361238342, "learning_rate": 1.0441767068273092e-06, "loss": 0.9371, "step": 53 }, { "epoch": 0.02172601086300543, "grad_norm": 7.411500707300452, "learning_rate": 1.0642570281124499e-06, "loss": 0.9567, "step": 54 }, { "epoch": 0.022128344397505533, "grad_norm": 8.194667792436356, "learning_rate": 1.0843373493975905e-06, "loss": 0.9994, "step": 55 }, { "epoch": 0.02253067793200563, "grad_norm": 7.141458407534788, "learning_rate": 1.1044176706827311e-06, "loss": 0.8714, "step": 56 }, { "epoch": 0.022933011466505733, "grad_norm": 5.765708921674402, "learning_rate": 1.1244979919678715e-06, "loss": 0.9238, "step": 57 }, { "epoch": 0.023335345001005835, "grad_norm": 7.924253998059311, "learning_rate": 1.1445783132530121e-06, "loss": 0.9884, "step": 58 }, { "epoch": 0.023737678535505933, "grad_norm": 6.597672306780267, "learning_rate": 1.1646586345381528e-06, "loss": 0.9904, "step": 59 }, { "epoch": 0.024140012070006035, "grad_norm": 6.937688487718752, "learning_rate": 1.1847389558232934e-06, "loss": 0.971, "step": 60 }, { "epoch": 0.024542345604506136, "grad_norm": 6.148469030855221, "learning_rate": 1.2048192771084338e-06, "loss": 0.7868, "step": 61 }, { "epoch": 0.024944679139006234, "grad_norm": 6.388105517333996, "learning_rate": 1.2248995983935744e-06, "loss": 0.8567, "step": 62 }, { "epoch": 0.025347012673506336, "grad_norm": 6.105324559387419, "learning_rate": 1.2449799196787148e-06, "loss": 0.9989, "step": 63 }, { "epoch": 0.025749346208006438, "grad_norm": 6.565241933640869, "learning_rate": 1.2650602409638555e-06, "loss": 0.8641, "step": 64 }, { "epoch": 0.02615167974250654, "grad_norm": 5.869907162685047, "learning_rate": 1.2851405622489959e-06, "loss": 0.9526, "step": 65 }, { "epoch": 0.026554013277006638, "grad_norm": 5.851371649159624, "learning_rate": 1.3052208835341367e-06, "loss": 0.8387, "step": 66 }, { "epoch": 0.02695634681150674, "grad_norm": 6.3493989766408765, "learning_rate": 1.3253012048192773e-06, "loss": 1.0458, "step": 67 }, { "epoch": 0.02735868034600684, "grad_norm": 5.9283912382403114, "learning_rate": 1.345381526104418e-06, "loss": 1.0158, "step": 68 }, { "epoch": 0.02776101388050694, "grad_norm": 6.184126721660312, "learning_rate": 1.3654618473895584e-06, "loss": 0.8218, "step": 69 }, { "epoch": 0.02816334741500704, "grad_norm": 6.0169493477739975, "learning_rate": 1.385542168674699e-06, "loss": 0.7966, "step": 70 }, { "epoch": 0.028565680949507143, "grad_norm": 6.084096016864222, "learning_rate": 1.4056224899598394e-06, "loss": 0.8221, "step": 71 }, { "epoch": 0.02896801448400724, "grad_norm": 6.441697966234361, "learning_rate": 1.42570281124498e-06, "loss": 0.7189, "step": 72 }, { "epoch": 0.029370348018507342, "grad_norm": 5.684758471405106, "learning_rate": 1.4457831325301204e-06, "loss": 0.811, "step": 73 }, { "epoch": 0.029772681553007444, "grad_norm": 6.066554340342954, "learning_rate": 1.465863453815261e-06, "loss": 0.8128, "step": 74 }, { "epoch": 0.030175015087507542, "grad_norm": 5.398586406654046, "learning_rate": 1.4859437751004017e-06, "loss": 0.8521, "step": 75 }, { "epoch": 0.030577348622007644, "grad_norm": 5.72164636854203, "learning_rate": 1.5060240963855425e-06, "loss": 0.8733, "step": 76 }, { "epoch": 0.030979682156507746, "grad_norm": 5.317662321058184, "learning_rate": 1.526104417670683e-06, "loss": 0.6758, "step": 77 }, { "epoch": 0.03138201569100785, "grad_norm": 5.008046012529329, "learning_rate": 1.5461847389558236e-06, "loss": 0.77, "step": 78 }, { "epoch": 0.03178434922550795, "grad_norm": 5.806938383916589, "learning_rate": 1.566265060240964e-06, "loss": 0.9218, "step": 79 }, { "epoch": 0.032186682760008044, "grad_norm": 5.687427641764964, "learning_rate": 1.5863453815261046e-06, "loss": 0.873, "step": 80 }, { "epoch": 0.032589016294508145, "grad_norm": 5.877177229539529, "learning_rate": 1.606425702811245e-06, "loss": 0.8347, "step": 81 }, { "epoch": 0.03299134982900825, "grad_norm": 5.661277204998088, "learning_rate": 1.6265060240963856e-06, "loss": 0.7537, "step": 82 }, { "epoch": 0.03339368336350835, "grad_norm": 6.0004706518847675, "learning_rate": 1.6465863453815263e-06, "loss": 0.9341, "step": 83 }, { "epoch": 0.03379601689800845, "grad_norm": 5.383372715468371, "learning_rate": 1.6666666666666667e-06, "loss": 0.716, "step": 84 }, { "epoch": 0.03419835043250855, "grad_norm": 5.870492937673813, "learning_rate": 1.6867469879518073e-06, "loss": 0.8515, "step": 85 }, { "epoch": 0.03460068396700865, "grad_norm": 5.40282247472889, "learning_rate": 1.7068273092369481e-06, "loss": 0.7575, "step": 86 }, { "epoch": 0.03500301750150875, "grad_norm": 5.244298542730251, "learning_rate": 1.7269076305220885e-06, "loss": 0.7342, "step": 87 }, { "epoch": 0.03540535103600885, "grad_norm": 5.118324863903489, "learning_rate": 1.7469879518072292e-06, "loss": 0.7633, "step": 88 }, { "epoch": 0.03580768457050895, "grad_norm": 5.559590746485852, "learning_rate": 1.7670682730923696e-06, "loss": 0.915, "step": 89 }, { "epoch": 0.036210018105009054, "grad_norm": 5.093580932305843, "learning_rate": 1.7871485943775102e-06, "loss": 0.7234, "step": 90 }, { "epoch": 0.036612351639509155, "grad_norm": 5.949076037097858, "learning_rate": 1.8072289156626508e-06, "loss": 0.9392, "step": 91 }, { "epoch": 0.03701468517400926, "grad_norm": 6.126962098905458, "learning_rate": 1.8273092369477912e-06, "loss": 0.7973, "step": 92 }, { "epoch": 0.03741701870850935, "grad_norm": 5.418925049178253, "learning_rate": 1.8473895582329318e-06, "loss": 0.7386, "step": 93 }, { "epoch": 0.03781935224300945, "grad_norm": 5.30622581323819, "learning_rate": 1.8674698795180723e-06, "loss": 0.7712, "step": 94 }, { "epoch": 0.038221685777509555, "grad_norm": 5.203905538553713, "learning_rate": 1.8875502008032129e-06, "loss": 0.7403, "step": 95 }, { "epoch": 0.03862401931200966, "grad_norm": 5.497752744670633, "learning_rate": 1.9076305220883537e-06, "loss": 0.8343, "step": 96 }, { "epoch": 0.03902635284650976, "grad_norm": 5.792666990034754, "learning_rate": 1.9277108433734943e-06, "loss": 0.9855, "step": 97 }, { "epoch": 0.03942868638100986, "grad_norm": 4.95103734609948, "learning_rate": 1.947791164658635e-06, "loss": 0.7676, "step": 98 }, { "epoch": 0.039831019915509955, "grad_norm": 5.313239350915902, "learning_rate": 1.967871485943775e-06, "loss": 0.8112, "step": 99 }, { "epoch": 0.040233353450010056, "grad_norm": 5.262581677945773, "learning_rate": 1.987951807228916e-06, "loss": 0.7726, "step": 100 }, { "epoch": 0.04063568698451016, "grad_norm": 5.7284063119812965, "learning_rate": 2.0080321285140564e-06, "loss": 0.7143, "step": 101 }, { "epoch": 0.04103802051901026, "grad_norm": 5.61074467493987, "learning_rate": 2.028112449799197e-06, "loss": 0.7821, "step": 102 }, { "epoch": 0.04144035405351036, "grad_norm": 5.668818940572837, "learning_rate": 2.0481927710843377e-06, "loss": 0.6217, "step": 103 }, { "epoch": 0.04184268758801046, "grad_norm": 6.102827563222031, "learning_rate": 2.068273092369478e-06, "loss": 0.7094, "step": 104 }, { "epoch": 0.04224502112251056, "grad_norm": 5.106901986680839, "learning_rate": 2.0883534136546185e-06, "loss": 0.7452, "step": 105 }, { "epoch": 0.04264735465701066, "grad_norm": 5.320689458556449, "learning_rate": 2.1084337349397595e-06, "loss": 0.8121, "step": 106 }, { "epoch": 0.04304968819151076, "grad_norm": 4.851575241509713, "learning_rate": 2.1285140562248997e-06, "loss": 0.7695, "step": 107 }, { "epoch": 0.04345202172601086, "grad_norm": 5.7294050539399315, "learning_rate": 2.1485943775100404e-06, "loss": 0.8343, "step": 108 }, { "epoch": 0.043854355260510965, "grad_norm": 6.2102432153860265, "learning_rate": 2.168674698795181e-06, "loss": 0.8478, "step": 109 }, { "epoch": 0.044256688795011066, "grad_norm": 5.8055227032618415, "learning_rate": 2.1887550200803216e-06, "loss": 0.7647, "step": 110 }, { "epoch": 0.04465902232951117, "grad_norm": 5.59165747057277, "learning_rate": 2.2088353413654622e-06, "loss": 0.6905, "step": 111 }, { "epoch": 0.04506135586401126, "grad_norm": 4.682897930842293, "learning_rate": 2.2289156626506024e-06, "loss": 0.6875, "step": 112 }, { "epoch": 0.045463689398511364, "grad_norm": 4.320466219380915, "learning_rate": 2.248995983935743e-06, "loss": 0.696, "step": 113 }, { "epoch": 0.045866022933011466, "grad_norm": 5.050663115560423, "learning_rate": 2.2690763052208837e-06, "loss": 0.6224, "step": 114 }, { "epoch": 0.04626835646751157, "grad_norm": 4.681780822596801, "learning_rate": 2.2891566265060243e-06, "loss": 0.6462, "step": 115 }, { "epoch": 0.04667069000201167, "grad_norm": 5.954135299109462, "learning_rate": 2.309236947791165e-06, "loss": 0.8848, "step": 116 }, { "epoch": 0.04707302353651177, "grad_norm": 5.573998592787985, "learning_rate": 2.3293172690763055e-06, "loss": 0.8088, "step": 117 }, { "epoch": 0.047475357071011866, "grad_norm": 5.070007694615057, "learning_rate": 2.349397590361446e-06, "loss": 0.7173, "step": 118 }, { "epoch": 0.04787769060551197, "grad_norm": 5.671554086319649, "learning_rate": 2.3694779116465868e-06, "loss": 0.8136, "step": 119 }, { "epoch": 0.04828002414001207, "grad_norm": 5.769697685018851, "learning_rate": 2.389558232931727e-06, "loss": 0.7733, "step": 120 }, { "epoch": 0.04868235767451217, "grad_norm": 4.920298621449247, "learning_rate": 2.4096385542168676e-06, "loss": 0.5931, "step": 121 }, { "epoch": 0.04908469120901227, "grad_norm": 5.49690667652572, "learning_rate": 2.4297188755020082e-06, "loss": 0.8079, "step": 122 }, { "epoch": 0.049487024743512374, "grad_norm": 5.369187249008034, "learning_rate": 2.449799196787149e-06, "loss": 0.726, "step": 123 }, { "epoch": 0.04988935827801247, "grad_norm": 4.8515152911462796, "learning_rate": 2.469879518072289e-06, "loss": 0.7691, "step": 124 }, { "epoch": 0.05029169181251257, "grad_norm": 5.099204794104859, "learning_rate": 2.4899598393574297e-06, "loss": 0.8361, "step": 125 }, { "epoch": 0.05069402534701267, "grad_norm": 5.390609640239935, "learning_rate": 2.5100401606425707e-06, "loss": 0.6527, "step": 126 }, { "epoch": 0.051096358881512774, "grad_norm": 5.053936099287885, "learning_rate": 2.530120481927711e-06, "loss": 0.7326, "step": 127 }, { "epoch": 0.051498692416012876, "grad_norm": 5.876719841742455, "learning_rate": 2.5502008032128516e-06, "loss": 0.8571, "step": 128 }, { "epoch": 0.05190102595051298, "grad_norm": 5.291120710287588, "learning_rate": 2.5702811244979918e-06, "loss": 0.7641, "step": 129 }, { "epoch": 0.05230335948501308, "grad_norm": 5.2648612797895336, "learning_rate": 2.590361445783133e-06, "loss": 0.768, "step": 130 }, { "epoch": 0.052705693019513174, "grad_norm": 5.146667666976231, "learning_rate": 2.6104417670682734e-06, "loss": 0.729, "step": 131 }, { "epoch": 0.053108026554013275, "grad_norm": 5.607251668752912, "learning_rate": 2.6305220883534136e-06, "loss": 0.7812, "step": 132 }, { "epoch": 0.05351036008851338, "grad_norm": 4.774953294338218, "learning_rate": 2.6506024096385547e-06, "loss": 0.6996, "step": 133 }, { "epoch": 0.05391269362301348, "grad_norm": 5.187652607031923, "learning_rate": 2.670682730923695e-06, "loss": 0.7029, "step": 134 }, { "epoch": 0.05431502715751358, "grad_norm": 5.631076339839678, "learning_rate": 2.690763052208836e-06, "loss": 0.7618, "step": 135 }, { "epoch": 0.05471736069201368, "grad_norm": 5.1080370644891735, "learning_rate": 2.710843373493976e-06, "loss": 0.6746, "step": 136 }, { "epoch": 0.05511969422651378, "grad_norm": 5.177774314908037, "learning_rate": 2.7309236947791167e-06, "loss": 0.6576, "step": 137 }, { "epoch": 0.05552202776101388, "grad_norm": 5.4428731816372355, "learning_rate": 2.751004016064257e-06, "loss": 0.8666, "step": 138 }, { "epoch": 0.05592436129551398, "grad_norm": 4.737577892176393, "learning_rate": 2.771084337349398e-06, "loss": 0.6303, "step": 139 }, { "epoch": 0.05632669483001408, "grad_norm": 5.958074686629733, "learning_rate": 2.791164658634538e-06, "loss": 0.7375, "step": 140 }, { "epoch": 0.056729028364514184, "grad_norm": 5.347638262573924, "learning_rate": 2.811244979919679e-06, "loss": 0.7496, "step": 141 }, { "epoch": 0.057131361899014285, "grad_norm": 5.25633214854894, "learning_rate": 2.83132530120482e-06, "loss": 0.7703, "step": 142 }, { "epoch": 0.05753369543351439, "grad_norm": 6.3493191578507355, "learning_rate": 2.85140562248996e-06, "loss": 0.7297, "step": 143 }, { "epoch": 0.05793602896801448, "grad_norm": 5.291536986822124, "learning_rate": 2.8714859437751007e-06, "loss": 0.7848, "step": 144 }, { "epoch": 0.05833836250251458, "grad_norm": 4.327776932092466, "learning_rate": 2.891566265060241e-06, "loss": 0.5537, "step": 145 }, { "epoch": 0.058740696037014685, "grad_norm": 4.827406966450388, "learning_rate": 2.911646586345382e-06, "loss": 0.6614, "step": 146 }, { "epoch": 0.05914302957151479, "grad_norm": 5.4512595989034605, "learning_rate": 2.931726907630522e-06, "loss": 0.5105, "step": 147 }, { "epoch": 0.05954536310601489, "grad_norm": 5.349806231832712, "learning_rate": 2.9518072289156627e-06, "loss": 0.7398, "step": 148 }, { "epoch": 0.05994769664051499, "grad_norm": 5.376464953972205, "learning_rate": 2.9718875502008034e-06, "loss": 0.832, "step": 149 }, { "epoch": 0.060350030175015085, "grad_norm": 4.4973784287811, "learning_rate": 2.991967871485944e-06, "loss": 0.6631, "step": 150 }, { "epoch": 0.060752363709515186, "grad_norm": 4.1574904576313925, "learning_rate": 3.012048192771085e-06, "loss": 0.5736, "step": 151 }, { "epoch": 0.06115469724401529, "grad_norm": 4.932454920014607, "learning_rate": 3.0321285140562252e-06, "loss": 0.6843, "step": 152 }, { "epoch": 0.06155703077851539, "grad_norm": 4.920072821237323, "learning_rate": 3.052208835341366e-06, "loss": 0.6115, "step": 153 }, { "epoch": 0.06195936431301549, "grad_norm": 5.069151288790082, "learning_rate": 3.072289156626506e-06, "loss": 0.6592, "step": 154 }, { "epoch": 0.06236169784751559, "grad_norm": 4.940734045454529, "learning_rate": 3.092369477911647e-06, "loss": 0.6823, "step": 155 }, { "epoch": 0.0627640313820157, "grad_norm": 4.7329970403721635, "learning_rate": 3.1124497991967873e-06, "loss": 0.6183, "step": 156 }, { "epoch": 0.0631663649165158, "grad_norm": 5.584324231160161, "learning_rate": 3.132530120481928e-06, "loss": 0.8407, "step": 157 }, { "epoch": 0.0635686984510159, "grad_norm": 4.9494988025514095, "learning_rate": 3.152610441767068e-06, "loss": 0.7362, "step": 158 }, { "epoch": 0.06397103198551599, "grad_norm": 4.826336881024485, "learning_rate": 3.172690763052209e-06, "loss": 0.6756, "step": 159 }, { "epoch": 0.06437336552001609, "grad_norm": 4.663584106854257, "learning_rate": 3.1927710843373494e-06, "loss": 0.6618, "step": 160 }, { "epoch": 0.06477569905451619, "grad_norm": 5.358113823802763, "learning_rate": 3.21285140562249e-06, "loss": 0.7845, "step": 161 }, { "epoch": 0.06517803258901629, "grad_norm": 4.393966562734061, "learning_rate": 3.232931726907631e-06, "loss": 0.6643, "step": 162 }, { "epoch": 0.06558036612351639, "grad_norm": 5.269005574739929, "learning_rate": 3.2530120481927713e-06, "loss": 0.7892, "step": 163 }, { "epoch": 0.0659826996580165, "grad_norm": 5.191377211307637, "learning_rate": 3.2730923694779123e-06, "loss": 0.6834, "step": 164 }, { "epoch": 0.0663850331925166, "grad_norm": 5.9411717854682875, "learning_rate": 3.2931726907630525e-06, "loss": 0.682, "step": 165 }, { "epoch": 0.0667873667270167, "grad_norm": 4.657895689681778, "learning_rate": 3.313253012048193e-06, "loss": 0.6533, "step": 166 }, { "epoch": 0.0671897002615168, "grad_norm": 4.689805393308291, "learning_rate": 3.3333333333333333e-06, "loss": 0.5851, "step": 167 }, { "epoch": 0.0675920337960169, "grad_norm": 4.934046837424603, "learning_rate": 3.3534136546184744e-06, "loss": 0.7834, "step": 168 }, { "epoch": 0.067994367330517, "grad_norm": 5.295304511873193, "learning_rate": 3.3734939759036146e-06, "loss": 0.6978, "step": 169 }, { "epoch": 0.0683967008650171, "grad_norm": 5.8805644086916935, "learning_rate": 3.393574297188755e-06, "loss": 0.7022, "step": 170 }, { "epoch": 0.0687990343995172, "grad_norm": 4.986410631963467, "learning_rate": 3.4136546184738962e-06, "loss": 0.8589, "step": 171 }, { "epoch": 0.0692013679340173, "grad_norm": 4.587536055949948, "learning_rate": 3.4337349397590364e-06, "loss": 0.5962, "step": 172 }, { "epoch": 0.0696037014685174, "grad_norm": 4.542521557186459, "learning_rate": 3.453815261044177e-06, "loss": 0.5778, "step": 173 }, { "epoch": 0.0700060350030175, "grad_norm": 5.006237220238328, "learning_rate": 3.4738955823293173e-06, "loss": 0.7066, "step": 174 }, { "epoch": 0.0704083685375176, "grad_norm": 5.03127022104005, "learning_rate": 3.4939759036144583e-06, "loss": 0.7143, "step": 175 }, { "epoch": 0.0708107020720177, "grad_norm": 4.978976609084576, "learning_rate": 3.5140562248995985e-06, "loss": 0.5806, "step": 176 }, { "epoch": 0.0712130356065178, "grad_norm": 4.953415145134539, "learning_rate": 3.534136546184739e-06, "loss": 0.6169, "step": 177 }, { "epoch": 0.0716153691410179, "grad_norm": 5.224274445585009, "learning_rate": 3.5542168674698798e-06, "loss": 0.7055, "step": 178 }, { "epoch": 0.072017702675518, "grad_norm": 4.793231833957601, "learning_rate": 3.5742971887550204e-06, "loss": 0.6589, "step": 179 }, { "epoch": 0.07242003621001811, "grad_norm": 4.897332398174339, "learning_rate": 3.5943775100401606e-06, "loss": 0.5835, "step": 180 }, { "epoch": 0.07282236974451821, "grad_norm": 5.461164108231627, "learning_rate": 3.6144578313253016e-06, "loss": 0.6153, "step": 181 }, { "epoch": 0.07322470327901831, "grad_norm": 4.750713371821981, "learning_rate": 3.6345381526104423e-06, "loss": 0.5576, "step": 182 }, { "epoch": 0.07362703681351841, "grad_norm": 4.612815212986164, "learning_rate": 3.6546184738955825e-06, "loss": 0.6695, "step": 183 }, { "epoch": 0.07402937034801851, "grad_norm": 5.035137988160493, "learning_rate": 3.6746987951807235e-06, "loss": 0.6451, "step": 184 }, { "epoch": 0.0744317038825186, "grad_norm": 5.189253752749726, "learning_rate": 3.6947791164658637e-06, "loss": 0.6897, "step": 185 }, { "epoch": 0.0748340374170187, "grad_norm": 5.061038391462879, "learning_rate": 3.7148594377510043e-06, "loss": 0.7363, "step": 186 }, { "epoch": 0.0752363709515188, "grad_norm": 5.12457801565824, "learning_rate": 3.7349397590361445e-06, "loss": 0.6662, "step": 187 }, { "epoch": 0.0756387044860189, "grad_norm": 5.468566358022457, "learning_rate": 3.7550200803212856e-06, "loss": 0.7333, "step": 188 }, { "epoch": 0.07604103802051901, "grad_norm": 4.83323388980342, "learning_rate": 3.7751004016064258e-06, "loss": 0.6179, "step": 189 }, { "epoch": 0.07644337155501911, "grad_norm": 5.618759075616741, "learning_rate": 3.7951807228915664e-06, "loss": 0.7861, "step": 190 }, { "epoch": 0.07684570508951921, "grad_norm": 5.571721878536531, "learning_rate": 3.8152610441767074e-06, "loss": 0.6481, "step": 191 }, { "epoch": 0.07724803862401931, "grad_norm": 5.067103682358002, "learning_rate": 3.835341365461848e-06, "loss": 0.7295, "step": 192 }, { "epoch": 0.07765037215851942, "grad_norm": 4.945134722999282, "learning_rate": 3.855421686746989e-06, "loss": 0.6275, "step": 193 }, { "epoch": 0.07805270569301952, "grad_norm": 4.995120702157204, "learning_rate": 3.875502008032129e-06, "loss": 0.6156, "step": 194 }, { "epoch": 0.07845503922751962, "grad_norm": 4.936376571967402, "learning_rate": 3.89558232931727e-06, "loss": 0.6141, "step": 195 }, { "epoch": 0.07885737276201972, "grad_norm": 5.014965363677265, "learning_rate": 3.91566265060241e-06, "loss": 0.6706, "step": 196 }, { "epoch": 0.07925970629651982, "grad_norm": 4.379077775855287, "learning_rate": 3.93574297188755e-06, "loss": 0.625, "step": 197 }, { "epoch": 0.07966203983101991, "grad_norm": 5.213065143955326, "learning_rate": 3.9558232931726905e-06, "loss": 0.6259, "step": 198 }, { "epoch": 0.08006437336552001, "grad_norm": 4.943005393732397, "learning_rate": 3.975903614457832e-06, "loss": 0.6867, "step": 199 }, { "epoch": 0.08046670690002011, "grad_norm": 6.1486973779873795, "learning_rate": 3.995983935742972e-06, "loss": 0.8347, "step": 200 }, { "epoch": 0.08086904043452021, "grad_norm": 4.733552314219692, "learning_rate": 4.016064257028113e-06, "loss": 0.6611, "step": 201 }, { "epoch": 0.08127137396902032, "grad_norm": 4.8269187112270195, "learning_rate": 4.036144578313254e-06, "loss": 0.6939, "step": 202 }, { "epoch": 0.08167370750352042, "grad_norm": 5.3439700963806125, "learning_rate": 4.056224899598394e-06, "loss": 0.7133, "step": 203 }, { "epoch": 0.08207604103802052, "grad_norm": 5.29998702305394, "learning_rate": 4.076305220883534e-06, "loss": 0.7792, "step": 204 }, { "epoch": 0.08247837457252062, "grad_norm": 5.408216152846704, "learning_rate": 4.096385542168675e-06, "loss": 0.6411, "step": 205 }, { "epoch": 0.08288070810702072, "grad_norm": 4.419630999712613, "learning_rate": 4.1164658634538155e-06, "loss": 0.615, "step": 206 }, { "epoch": 0.08328304164152082, "grad_norm": 4.543564292966689, "learning_rate": 4.136546184738956e-06, "loss": 0.5462, "step": 207 }, { "epoch": 0.08368537517602093, "grad_norm": 4.814136944508732, "learning_rate": 4.156626506024097e-06, "loss": 0.6407, "step": 208 }, { "epoch": 0.08408770871052103, "grad_norm": 5.118958059055448, "learning_rate": 4.176706827309237e-06, "loss": 0.6922, "step": 209 }, { "epoch": 0.08449004224502112, "grad_norm": 5.761968321529592, "learning_rate": 4.196787148594378e-06, "loss": 0.7785, "step": 210 }, { "epoch": 0.08489237577952122, "grad_norm": 4.7973887668760185, "learning_rate": 4.216867469879519e-06, "loss": 0.5855, "step": 211 }, { "epoch": 0.08529470931402132, "grad_norm": 5.063119851943765, "learning_rate": 4.236947791164659e-06, "loss": 0.5061, "step": 212 }, { "epoch": 0.08569704284852142, "grad_norm": 5.185457141038217, "learning_rate": 4.2570281124497995e-06, "loss": 0.6547, "step": 213 }, { "epoch": 0.08609937638302152, "grad_norm": 4.908063923355144, "learning_rate": 4.27710843373494e-06, "loss": 0.6325, "step": 214 }, { "epoch": 0.08650170991752162, "grad_norm": 4.902236509892525, "learning_rate": 4.297188755020081e-06, "loss": 0.771, "step": 215 }, { "epoch": 0.08690404345202173, "grad_norm": 4.939421470208826, "learning_rate": 4.317269076305221e-06, "loss": 0.6009, "step": 216 }, { "epoch": 0.08730637698652183, "grad_norm": 5.585627020639388, "learning_rate": 4.337349397590362e-06, "loss": 0.75, "step": 217 }, { "epoch": 0.08770871052102193, "grad_norm": 5.768733340993033, "learning_rate": 4.357429718875502e-06, "loss": 0.8889, "step": 218 }, { "epoch": 0.08811104405552203, "grad_norm": 4.756073859984964, "learning_rate": 4.377510040160643e-06, "loss": 0.5845, "step": 219 }, { "epoch": 0.08851337759002213, "grad_norm": 5.602957070495176, "learning_rate": 4.397590361445783e-06, "loss": 0.5766, "step": 220 }, { "epoch": 0.08891571112452223, "grad_norm": 5.51373108651177, "learning_rate": 4.4176706827309244e-06, "loss": 0.667, "step": 221 }, { "epoch": 0.08931804465902234, "grad_norm": 5.2818421493270895, "learning_rate": 4.437751004016065e-06, "loss": 0.6132, "step": 222 }, { "epoch": 0.08972037819352242, "grad_norm": 4.687481374933495, "learning_rate": 4.457831325301205e-06, "loss": 0.69, "step": 223 }, { "epoch": 0.09012271172802253, "grad_norm": 5.045402036516882, "learning_rate": 4.477911646586346e-06, "loss": 0.5496, "step": 224 }, { "epoch": 0.09052504526252263, "grad_norm": 4.949655367217769, "learning_rate": 4.497991967871486e-06, "loss": 0.6777, "step": 225 }, { "epoch": 0.09092737879702273, "grad_norm": 4.725446380096972, "learning_rate": 4.518072289156627e-06, "loss": 0.5211, "step": 226 }, { "epoch": 0.09132971233152283, "grad_norm": 5.627747083860127, "learning_rate": 4.538152610441767e-06, "loss": 0.7335, "step": 227 }, { "epoch": 0.09173204586602293, "grad_norm": 4.658915831199445, "learning_rate": 4.558232931726908e-06, "loss": 0.618, "step": 228 }, { "epoch": 0.09213437940052303, "grad_norm": 5.169380307261355, "learning_rate": 4.578313253012049e-06, "loss": 0.8083, "step": 229 }, { "epoch": 0.09253671293502314, "grad_norm": 4.462280637097748, "learning_rate": 4.598393574297189e-06, "loss": 0.5873, "step": 230 }, { "epoch": 0.09293904646952324, "grad_norm": 5.088180514844276, "learning_rate": 4.61847389558233e-06, "loss": 0.6948, "step": 231 }, { "epoch": 0.09334138000402334, "grad_norm": 4.911645340461795, "learning_rate": 4.63855421686747e-06, "loss": 0.5014, "step": 232 }, { "epoch": 0.09374371353852344, "grad_norm": 5.183650867648958, "learning_rate": 4.658634538152611e-06, "loss": 0.6731, "step": 233 }, { "epoch": 0.09414604707302354, "grad_norm": 4.942647346081943, "learning_rate": 4.678714859437751e-06, "loss": 0.796, "step": 234 }, { "epoch": 0.09454838060752364, "grad_norm": 5.413754367435916, "learning_rate": 4.698795180722892e-06, "loss": 0.7029, "step": 235 }, { "epoch": 0.09495071414202373, "grad_norm": 5.394996509930947, "learning_rate": 4.7188755020080325e-06, "loss": 0.7448, "step": 236 }, { "epoch": 0.09535304767652383, "grad_norm": 4.807099705141704, "learning_rate": 4.7389558232931736e-06, "loss": 0.6608, "step": 237 }, { "epoch": 0.09575538121102393, "grad_norm": 5.204336077501797, "learning_rate": 4.759036144578314e-06, "loss": 0.585, "step": 238 }, { "epoch": 0.09615771474552404, "grad_norm": 4.845053158029385, "learning_rate": 4.779116465863454e-06, "loss": 0.6299, "step": 239 }, { "epoch": 0.09656004828002414, "grad_norm": 5.037832345926416, "learning_rate": 4.799196787148594e-06, "loss": 0.6598, "step": 240 }, { "epoch": 0.09696238181452424, "grad_norm": 4.930571962391263, "learning_rate": 4.819277108433735e-06, "loss": 0.5918, "step": 241 }, { "epoch": 0.09736471534902434, "grad_norm": 4.749158619882368, "learning_rate": 4.839357429718876e-06, "loss": 0.5785, "step": 242 }, { "epoch": 0.09776704888352444, "grad_norm": 5.774299138017609, "learning_rate": 4.8594377510040165e-06, "loss": 0.7465, "step": 243 }, { "epoch": 0.09816938241802455, "grad_norm": 4.582929546310413, "learning_rate": 4.8795180722891575e-06, "loss": 0.5464, "step": 244 }, { "epoch": 0.09857171595252465, "grad_norm": 4.604283671045787, "learning_rate": 4.899598393574298e-06, "loss": 0.6183, "step": 245 }, { "epoch": 0.09897404948702475, "grad_norm": 5.126092092509994, "learning_rate": 4.919678714859438e-06, "loss": 0.7267, "step": 246 }, { "epoch": 0.09937638302152485, "grad_norm": 4.6087281951772345, "learning_rate": 4.939759036144578e-06, "loss": 0.5911, "step": 247 }, { "epoch": 0.09977871655602494, "grad_norm": 4.4308954136706165, "learning_rate": 4.959839357429719e-06, "loss": 0.6525, "step": 248 }, { "epoch": 0.10018105009052504, "grad_norm": 4.704767612087696, "learning_rate": 4.979919678714859e-06, "loss": 0.7537, "step": 249 }, { "epoch": 0.10058338362502514, "grad_norm": 5.171569719861641, "learning_rate": 5e-06, "loss": 0.5941, "step": 250 }, { "epoch": 0.10098571715952524, "grad_norm": 5.7449160884679245, "learning_rate": 4.99999753465491e-06, "loss": 0.6686, "step": 251 }, { "epoch": 0.10138805069402534, "grad_norm": 4.648308764039487, "learning_rate": 4.999990138624503e-06, "loss": 0.6264, "step": 252 }, { "epoch": 0.10179038422852545, "grad_norm": 5.064339952308565, "learning_rate": 4.999977811923365e-06, "loss": 0.6921, "step": 253 }, { "epoch": 0.10219271776302555, "grad_norm": 4.810326779178174, "learning_rate": 4.9999605545758065e-06, "loss": 0.6119, "step": 254 }, { "epoch": 0.10259505129752565, "grad_norm": 4.805314994215823, "learning_rate": 4.999938366615867e-06, "loss": 0.6364, "step": 255 }, { "epoch": 0.10299738483202575, "grad_norm": 4.473992853756316, "learning_rate": 4.999911248087305e-06, "loss": 0.5508, "step": 256 }, { "epoch": 0.10339971836652585, "grad_norm": 5.452559581429947, "learning_rate": 4.999879199043606e-06, "loss": 0.7831, "step": 257 }, { "epoch": 0.10380205190102595, "grad_norm": 5.061524793864594, "learning_rate": 4.99984221954798e-06, "loss": 0.5644, "step": 258 }, { "epoch": 0.10420438543552606, "grad_norm": 4.673975167806026, "learning_rate": 4.999800309673361e-06, "loss": 0.649, "step": 259 }, { "epoch": 0.10460671897002616, "grad_norm": 4.8181768852216855, "learning_rate": 4.999753469502406e-06, "loss": 0.6528, "step": 260 }, { "epoch": 0.10500905250452625, "grad_norm": 5.3922298074237345, "learning_rate": 4.999701699127497e-06, "loss": 0.8216, "step": 261 }, { "epoch": 0.10541138603902635, "grad_norm": 4.463138014998754, "learning_rate": 4.99964499865074e-06, "loss": 0.6824, "step": 262 }, { "epoch": 0.10581371957352645, "grad_norm": 4.937246753810877, "learning_rate": 4.999583368183965e-06, "loss": 0.5256, "step": 263 }, { "epoch": 0.10621605310802655, "grad_norm": 5.02716535200141, "learning_rate": 4.999516807848721e-06, "loss": 0.6658, "step": 264 }, { "epoch": 0.10661838664252665, "grad_norm": 4.719903057858775, "learning_rate": 4.999445317776286e-06, "loss": 0.5943, "step": 265 }, { "epoch": 0.10702072017702675, "grad_norm": 4.327891289019763, "learning_rate": 4.9993688981076574e-06, "loss": 0.6267, "step": 266 }, { "epoch": 0.10742305371152686, "grad_norm": 5.04792717283726, "learning_rate": 4.999287548993555e-06, "loss": 0.7649, "step": 267 }, { "epoch": 0.10782538724602696, "grad_norm": 4.398384896034094, "learning_rate": 4.999201270594424e-06, "loss": 0.6237, "step": 268 }, { "epoch": 0.10822772078052706, "grad_norm": 4.513204332172601, "learning_rate": 4.999110063080428e-06, "loss": 0.685, "step": 269 }, { "epoch": 0.10863005431502716, "grad_norm": 4.910504259364484, "learning_rate": 4.999013926631453e-06, "loss": 0.6393, "step": 270 }, { "epoch": 0.10903238784952726, "grad_norm": 3.8051932823856927, "learning_rate": 4.998912861437106e-06, "loss": 0.557, "step": 271 }, { "epoch": 0.10943472138402736, "grad_norm": 4.261798228362777, "learning_rate": 4.9988068676967176e-06, "loss": 0.6685, "step": 272 }, { "epoch": 0.10983705491852747, "grad_norm": 5.199991754679721, "learning_rate": 4.9986959456193345e-06, "loss": 0.6926, "step": 273 }, { "epoch": 0.11023938845302755, "grad_norm": 4.800143443642004, "learning_rate": 4.998580095423728e-06, "loss": 0.6262, "step": 274 }, { "epoch": 0.11064172198752766, "grad_norm": 5.703676474195908, "learning_rate": 4.998459317338383e-06, "loss": 0.8102, "step": 275 }, { "epoch": 0.11104405552202776, "grad_norm": 5.554521472398849, "learning_rate": 4.998333611601511e-06, "loss": 0.8502, "step": 276 }, { "epoch": 0.11144638905652786, "grad_norm": 5.223628882218702, "learning_rate": 4.998202978461038e-06, "loss": 0.6968, "step": 277 }, { "epoch": 0.11184872259102796, "grad_norm": 5.028503778823592, "learning_rate": 4.9980674181746055e-06, "loss": 0.6237, "step": 278 }, { "epoch": 0.11225105612552806, "grad_norm": 4.810062581233454, "learning_rate": 4.997926931009579e-06, "loss": 0.6566, "step": 279 }, { "epoch": 0.11265338966002816, "grad_norm": 4.895795240240787, "learning_rate": 4.997781517243037e-06, "loss": 0.6358, "step": 280 }, { "epoch": 0.11305572319452827, "grad_norm": 4.678689813671165, "learning_rate": 4.997631177161775e-06, "loss": 0.7098, "step": 281 }, { "epoch": 0.11345805672902837, "grad_norm": 4.8363206866892465, "learning_rate": 4.9974759110623065e-06, "loss": 0.7234, "step": 282 }, { "epoch": 0.11386039026352847, "grad_norm": 4.872936830415585, "learning_rate": 4.997315719250857e-06, "loss": 0.5937, "step": 283 }, { "epoch": 0.11426272379802857, "grad_norm": 4.28046131366056, "learning_rate": 4.997150602043371e-06, "loss": 0.6211, "step": 284 }, { "epoch": 0.11466505733252867, "grad_norm": 4.837073628599811, "learning_rate": 4.996980559765504e-06, "loss": 0.5828, "step": 285 }, { "epoch": 0.11506739086702877, "grad_norm": 4.889286615585772, "learning_rate": 4.996805592752626e-06, "loss": 0.7301, "step": 286 }, { "epoch": 0.11546972440152886, "grad_norm": 3.885336153971243, "learning_rate": 4.996625701349822e-06, "loss": 0.5091, "step": 287 }, { "epoch": 0.11587205793602896, "grad_norm": 4.738737155243486, "learning_rate": 4.996440885911886e-06, "loss": 0.5519, "step": 288 }, { "epoch": 0.11627439147052906, "grad_norm": 4.600896924297193, "learning_rate": 4.996251146803326e-06, "loss": 0.5357, "step": 289 }, { "epoch": 0.11667672500502917, "grad_norm": 4.348185410245171, "learning_rate": 4.996056484398359e-06, "loss": 0.606, "step": 290 }, { "epoch": 0.11707905853952927, "grad_norm": 4.609344088685658, "learning_rate": 4.995856899080914e-06, "loss": 0.7079, "step": 291 }, { "epoch": 0.11748139207402937, "grad_norm": 4.982538761307109, "learning_rate": 4.9956523912446274e-06, "loss": 0.6626, "step": 292 }, { "epoch": 0.11788372560852947, "grad_norm": 4.958292330122082, "learning_rate": 4.9954429612928455e-06, "loss": 0.5589, "step": 293 }, { "epoch": 0.11828605914302957, "grad_norm": 5.3063992889234015, "learning_rate": 4.995228609638623e-06, "loss": 0.6761, "step": 294 }, { "epoch": 0.11868839267752967, "grad_norm": 4.787491132159558, "learning_rate": 4.995009336704718e-06, "loss": 0.6907, "step": 295 }, { "epoch": 0.11909072621202978, "grad_norm": 5.042532894966667, "learning_rate": 4.994785142923599e-06, "loss": 0.7259, "step": 296 }, { "epoch": 0.11949305974652988, "grad_norm": 5.24654264942669, "learning_rate": 4.994556028737439e-06, "loss": 0.7376, "step": 297 }, { "epoch": 0.11989539328102998, "grad_norm": 4.7381264053196475, "learning_rate": 4.9943219945981126e-06, "loss": 0.6419, "step": 298 }, { "epoch": 0.12029772681553007, "grad_norm": 4.174895491105691, "learning_rate": 4.9940830409672e-06, "loss": 0.5846, "step": 299 }, { "epoch": 0.12070006035003017, "grad_norm": 5.265350639824991, "learning_rate": 4.993839168315985e-06, "loss": 0.7269, "step": 300 }, { "epoch": 0.12110239388453027, "grad_norm": 5.293670654130999, "learning_rate": 4.99359037712545e-06, "loss": 0.7539, "step": 301 }, { "epoch": 0.12150472741903037, "grad_norm": 4.832726818218027, "learning_rate": 4.993336667886281e-06, "loss": 0.6617, "step": 302 }, { "epoch": 0.12190706095353047, "grad_norm": 4.982756198499471, "learning_rate": 4.993078041098862e-06, "loss": 0.6517, "step": 303 }, { "epoch": 0.12230939448803058, "grad_norm": 4.787146336142149, "learning_rate": 4.9928144972732785e-06, "loss": 0.5943, "step": 304 }, { "epoch": 0.12271172802253068, "grad_norm": 5.725699716607812, "learning_rate": 4.992546036929309e-06, "loss": 0.7385, "step": 305 }, { "epoch": 0.12311406155703078, "grad_norm": 4.388246603675133, "learning_rate": 4.992272660596432e-06, "loss": 0.5268, "step": 306 }, { "epoch": 0.12351639509153088, "grad_norm": 4.323516644235707, "learning_rate": 4.991994368813823e-06, "loss": 0.5069, "step": 307 }, { "epoch": 0.12391872862603098, "grad_norm": 4.968139714274168, "learning_rate": 4.991711162130347e-06, "loss": 0.6089, "step": 308 }, { "epoch": 0.12432106216053108, "grad_norm": 5.393689160165955, "learning_rate": 4.99142304110457e-06, "loss": 0.5749, "step": 309 }, { "epoch": 0.12472339569503119, "grad_norm": 5.010889684729688, "learning_rate": 4.991130006304742e-06, "loss": 0.6868, "step": 310 }, { "epoch": 0.12512572922953127, "grad_norm": 4.77584804740267, "learning_rate": 4.99083205830881e-06, "loss": 0.7153, "step": 311 }, { "epoch": 0.1255280627640314, "grad_norm": 3.8671259573453343, "learning_rate": 4.990529197704411e-06, "loss": 0.4938, "step": 312 }, { "epoch": 0.12593039629853148, "grad_norm": 4.878545701273808, "learning_rate": 4.990221425088868e-06, "loss": 0.5954, "step": 313 }, { "epoch": 0.1263327298330316, "grad_norm": 4.481594062259529, "learning_rate": 4.989908741069195e-06, "loss": 0.6397, "step": 314 }, { "epoch": 0.12673506336753168, "grad_norm": 4.317626894827104, "learning_rate": 4.98959114626209e-06, "loss": 0.609, "step": 315 }, { "epoch": 0.1271373969020318, "grad_norm": 4.415809282581258, "learning_rate": 4.989268641293939e-06, "loss": 0.5092, "step": 316 }, { "epoch": 0.12753973043653188, "grad_norm": 4.4424914565481375, "learning_rate": 4.9889412268008096e-06, "loss": 0.6316, "step": 317 }, { "epoch": 0.12794206397103197, "grad_norm": 4.431277408902586, "learning_rate": 4.988608903428454e-06, "loss": 0.6015, "step": 318 }, { "epoch": 0.1283443975055321, "grad_norm": 4.618861534462686, "learning_rate": 4.988271671832305e-06, "loss": 0.5562, "step": 319 }, { "epoch": 0.12874673104003218, "grad_norm": 5.078731601531793, "learning_rate": 4.987929532677478e-06, "loss": 0.7308, "step": 320 }, { "epoch": 0.1291490645745323, "grad_norm": 4.814095780713649, "learning_rate": 4.987582486638763e-06, "loss": 0.6856, "step": 321 }, { "epoch": 0.12955139810903238, "grad_norm": 5.171953766801144, "learning_rate": 4.987230534400634e-06, "loss": 0.7413, "step": 322 }, { "epoch": 0.1299537316435325, "grad_norm": 4.527962518284751, "learning_rate": 4.986873676657237e-06, "loss": 0.5227, "step": 323 }, { "epoch": 0.13035606517803258, "grad_norm": 4.9194193843235405, "learning_rate": 4.986511914112392e-06, "loss": 0.6038, "step": 324 }, { "epoch": 0.1307583987125327, "grad_norm": 4.403379209157847, "learning_rate": 4.986145247479597e-06, "loss": 0.6624, "step": 325 }, { "epoch": 0.13116073224703279, "grad_norm": 4.784759384702467, "learning_rate": 4.985773677482018e-06, "loss": 0.7109, "step": 326 }, { "epoch": 0.1315630657815329, "grad_norm": 4.928746115672455, "learning_rate": 4.985397204852496e-06, "loss": 0.6241, "step": 327 }, { "epoch": 0.131965399316033, "grad_norm": 5.6685753103245755, "learning_rate": 4.985015830333536e-06, "loss": 0.7121, "step": 328 }, { "epoch": 0.1323677328505331, "grad_norm": 4.723811520101061, "learning_rate": 4.984629554677316e-06, "loss": 0.5683, "step": 329 }, { "epoch": 0.1327700663850332, "grad_norm": 4.106210405752034, "learning_rate": 4.984238378645677e-06, "loss": 0.6128, "step": 330 }, { "epoch": 0.13317239991953328, "grad_norm": 5.439987884236508, "learning_rate": 4.983842303010127e-06, "loss": 0.7294, "step": 331 }, { "epoch": 0.1335747334540334, "grad_norm": 4.76281391948892, "learning_rate": 4.9834413285518365e-06, "loss": 0.6116, "step": 332 }, { "epoch": 0.13397706698853348, "grad_norm": 4.594103270680001, "learning_rate": 4.983035456061637e-06, "loss": 0.5509, "step": 333 }, { "epoch": 0.1343794005230336, "grad_norm": 4.8229418474364785, "learning_rate": 4.982624686340022e-06, "loss": 0.6565, "step": 334 }, { "epoch": 0.1347817340575337, "grad_norm": 4.572125613909981, "learning_rate": 4.9822090201971414e-06, "loss": 0.5739, "step": 335 }, { "epoch": 0.1351840675920338, "grad_norm": 4.657722588184461, "learning_rate": 4.981788458452806e-06, "loss": 0.716, "step": 336 }, { "epoch": 0.1355864011265339, "grad_norm": 4.485714819813409, "learning_rate": 4.9813630019364765e-06, "loss": 0.5809, "step": 337 }, { "epoch": 0.135988734661034, "grad_norm": 4.730520981461843, "learning_rate": 4.9809326514872735e-06, "loss": 0.6006, "step": 338 }, { "epoch": 0.1363910681955341, "grad_norm": 4.407751381335739, "learning_rate": 4.980497407953966e-06, "loss": 0.6979, "step": 339 }, { "epoch": 0.1367934017300342, "grad_norm": 5.036488553382197, "learning_rate": 4.980057272194973e-06, "loss": 0.6719, "step": 340 }, { "epoch": 0.1371957352645343, "grad_norm": 4.131020183485464, "learning_rate": 4.979612245078366e-06, "loss": 0.656, "step": 341 }, { "epoch": 0.1375980687990344, "grad_norm": 4.494385344326802, "learning_rate": 4.979162327481859e-06, "loss": 0.5824, "step": 342 }, { "epoch": 0.1380004023335345, "grad_norm": 4.887767115382313, "learning_rate": 4.9787075202928155e-06, "loss": 0.7065, "step": 343 }, { "epoch": 0.1384027358680346, "grad_norm": 4.929080555596624, "learning_rate": 4.97824782440824e-06, "loss": 0.6462, "step": 344 }, { "epoch": 0.1388050694025347, "grad_norm": 4.768200287000246, "learning_rate": 4.97778324073478e-06, "loss": 0.7052, "step": 345 }, { "epoch": 0.1392074029370348, "grad_norm": 4.824866676754677, "learning_rate": 4.977313770188723e-06, "loss": 0.6169, "step": 346 }, { "epoch": 0.1396097364715349, "grad_norm": 5.0730067342381515, "learning_rate": 4.976839413695994e-06, "loss": 0.6907, "step": 347 }, { "epoch": 0.140012070006035, "grad_norm": 4.400048281159567, "learning_rate": 4.976360172192156e-06, "loss": 0.6843, "step": 348 }, { "epoch": 0.1404144035405351, "grad_norm": 4.826940942258961, "learning_rate": 4.975876046622404e-06, "loss": 0.6425, "step": 349 }, { "epoch": 0.1408167370750352, "grad_norm": 4.1980516292894645, "learning_rate": 4.975387037941568e-06, "loss": 0.513, "step": 350 }, { "epoch": 0.1412190706095353, "grad_norm": 4.599176289887721, "learning_rate": 4.974893147114108e-06, "loss": 0.6138, "step": 351 }, { "epoch": 0.1416214041440354, "grad_norm": 5.224094683564219, "learning_rate": 4.974394375114113e-06, "loss": 0.7688, "step": 352 }, { "epoch": 0.14202373767853552, "grad_norm": 4.886795665659055, "learning_rate": 4.9738907229253e-06, "loss": 0.6447, "step": 353 }, { "epoch": 0.1424260712130356, "grad_norm": 4.686162997153853, "learning_rate": 4.973382191541008e-06, "loss": 0.6489, "step": 354 }, { "epoch": 0.14282840474753572, "grad_norm": 4.633889069385792, "learning_rate": 4.972868781964204e-06, "loss": 0.6841, "step": 355 }, { "epoch": 0.1432307382820358, "grad_norm": 4.714123893258144, "learning_rate": 4.972350495207472e-06, "loss": 0.7282, "step": 356 }, { "epoch": 0.1436330718165359, "grad_norm": 4.620909154154889, "learning_rate": 4.971827332293017e-06, "loss": 0.6267, "step": 357 }, { "epoch": 0.144035405351036, "grad_norm": 5.112805770554144, "learning_rate": 4.971299294252659e-06, "loss": 0.7187, "step": 358 }, { "epoch": 0.1444377388855361, "grad_norm": 4.491917871078975, "learning_rate": 4.970766382127838e-06, "loss": 0.6243, "step": 359 }, { "epoch": 0.14484007242003621, "grad_norm": 4.91569774829319, "learning_rate": 4.9702285969696006e-06, "loss": 0.6044, "step": 360 }, { "epoch": 0.1452424059545363, "grad_norm": 4.347791813311231, "learning_rate": 4.96968593983861e-06, "loss": 0.6496, "step": 361 }, { "epoch": 0.14564473948903642, "grad_norm": 4.6534811500518645, "learning_rate": 4.9691384118051346e-06, "loss": 0.7414, "step": 362 }, { "epoch": 0.1460470730235365, "grad_norm": 4.5918592667861455, "learning_rate": 4.968586013949051e-06, "loss": 0.5802, "step": 363 }, { "epoch": 0.14644940655803662, "grad_norm": 4.435482175102624, "learning_rate": 4.96802874735984e-06, "loss": 0.5162, "step": 364 }, { "epoch": 0.1468517400925367, "grad_norm": 4.407001515831722, "learning_rate": 4.967466613136586e-06, "loss": 0.5361, "step": 365 }, { "epoch": 0.14725407362703682, "grad_norm": 4.369142207445966, "learning_rate": 4.966899612387972e-06, "loss": 0.664, "step": 366 }, { "epoch": 0.1476564071615369, "grad_norm": 4.607767758241303, "learning_rate": 4.966327746232281e-06, "loss": 0.5882, "step": 367 }, { "epoch": 0.14805874069603703, "grad_norm": 4.938226519089643, "learning_rate": 4.96575101579739e-06, "loss": 0.5841, "step": 368 }, { "epoch": 0.14846107423053712, "grad_norm": 4.096479564143721, "learning_rate": 4.965169422220771e-06, "loss": 0.5691, "step": 369 }, { "epoch": 0.1488634077650372, "grad_norm": 4.775093512712085, "learning_rate": 4.964582966649488e-06, "loss": 0.6063, "step": 370 }, { "epoch": 0.14926574129953732, "grad_norm": 4.211658294897708, "learning_rate": 4.963991650240192e-06, "loss": 0.5215, "step": 371 }, { "epoch": 0.1496680748340374, "grad_norm": 4.364777540143213, "learning_rate": 4.963395474159122e-06, "loss": 0.4866, "step": 372 }, { "epoch": 0.15007040836853752, "grad_norm": 4.681087632312734, "learning_rate": 4.962794439582102e-06, "loss": 0.6122, "step": 373 }, { "epoch": 0.1504727419030376, "grad_norm": 5.077820479708522, "learning_rate": 4.9621885476945405e-06, "loss": 0.7555, "step": 374 }, { "epoch": 0.15087507543753773, "grad_norm": 5.978600203961224, "learning_rate": 4.961577799691421e-06, "loss": 0.7113, "step": 375 }, { "epoch": 0.1512774089720378, "grad_norm": 5.376960173895862, "learning_rate": 4.960962196777307e-06, "loss": 0.577, "step": 376 }, { "epoch": 0.15167974250653793, "grad_norm": 5.18055770132837, "learning_rate": 4.960341740166338e-06, "loss": 0.6625, "step": 377 }, { "epoch": 0.15208207604103802, "grad_norm": 4.460009104960063, "learning_rate": 4.959716431082227e-06, "loss": 0.5306, "step": 378 }, { "epoch": 0.15248440957553813, "grad_norm": 4.303191111673615, "learning_rate": 4.959086270758255e-06, "loss": 0.5604, "step": 379 }, { "epoch": 0.15288674311003822, "grad_norm": 4.8364861459301185, "learning_rate": 4.9584512604372704e-06, "loss": 0.6154, "step": 380 }, { "epoch": 0.15328907664453834, "grad_norm": 4.638041078223287, "learning_rate": 4.957811401371692e-06, "loss": 0.5852, "step": 381 }, { "epoch": 0.15369141017903842, "grad_norm": 4.8155087041398295, "learning_rate": 4.9571666948234975e-06, "loss": 0.591, "step": 382 }, { "epoch": 0.1540937437135385, "grad_norm": 4.365724907978, "learning_rate": 4.956517142064226e-06, "loss": 0.5645, "step": 383 }, { "epoch": 0.15449607724803863, "grad_norm": 4.800750875035706, "learning_rate": 4.955862744374974e-06, "loss": 0.6958, "step": 384 }, { "epoch": 0.15489841078253871, "grad_norm": 4.218913109898912, "learning_rate": 4.9552035030463955e-06, "loss": 0.5981, "step": 385 }, { "epoch": 0.15530074431703883, "grad_norm": 3.8608056472996775, "learning_rate": 4.954539419378695e-06, "loss": 0.5516, "step": 386 }, { "epoch": 0.15570307785153892, "grad_norm": 4.422539247772956, "learning_rate": 4.953870494681631e-06, "loss": 0.466, "step": 387 }, { "epoch": 0.15610541138603903, "grad_norm": 5.389694964679017, "learning_rate": 4.9531967302745065e-06, "loss": 0.7187, "step": 388 }, { "epoch": 0.15650774492053912, "grad_norm": 4.641539036731872, "learning_rate": 4.952518127486171e-06, "loss": 0.6118, "step": 389 }, { "epoch": 0.15691007845503924, "grad_norm": 4.38792439545412, "learning_rate": 4.951834687655016e-06, "loss": 0.6125, "step": 390 }, { "epoch": 0.15731241198953932, "grad_norm": 4.859276712096591, "learning_rate": 4.9511464121289745e-06, "loss": 0.789, "step": 391 }, { "epoch": 0.15771474552403944, "grad_norm": 4.123781252290954, "learning_rate": 4.950453302265516e-06, "loss": 0.5189, "step": 392 }, { "epoch": 0.15811707905853953, "grad_norm": 4.5665024165035915, "learning_rate": 4.949755359431643e-06, "loss": 0.5843, "step": 393 }, { "epoch": 0.15851941259303964, "grad_norm": 4.554644389766188, "learning_rate": 4.949052585003892e-06, "loss": 0.5165, "step": 394 }, { "epoch": 0.15892174612753973, "grad_norm": 4.652445776343059, "learning_rate": 4.9483449803683295e-06, "loss": 0.6496, "step": 395 }, { "epoch": 0.15932407966203982, "grad_norm": 4.155340040403773, "learning_rate": 4.947632546920545e-06, "loss": 0.5823, "step": 396 }, { "epoch": 0.15972641319653993, "grad_norm": 5.685373741993176, "learning_rate": 4.946915286065656e-06, "loss": 0.7257, "step": 397 }, { "epoch": 0.16012874673104002, "grad_norm": 3.794107938623536, "learning_rate": 4.946193199218298e-06, "loss": 0.5539, "step": 398 }, { "epoch": 0.16053108026554014, "grad_norm": 3.345825002647911, "learning_rate": 4.945466287802625e-06, "loss": 0.4522, "step": 399 }, { "epoch": 0.16093341380004023, "grad_norm": 4.499068362412442, "learning_rate": 4.944734553252308e-06, "loss": 0.7294, "step": 400 }, { "epoch": 0.16133574733454034, "grad_norm": 4.710035321209727, "learning_rate": 4.94399799701053e-06, "loss": 0.6495, "step": 401 }, { "epoch": 0.16173808086904043, "grad_norm": 4.665060195725978, "learning_rate": 4.943256620529983e-06, "loss": 0.5549, "step": 402 }, { "epoch": 0.16214041440354054, "grad_norm": 4.526568640183799, "learning_rate": 4.942510425272864e-06, "loss": 0.5342, "step": 403 }, { "epoch": 0.16254274793804063, "grad_norm": 4.500912340574114, "learning_rate": 4.941759412710878e-06, "loss": 0.5154, "step": 404 }, { "epoch": 0.16294508147254075, "grad_norm": 4.871395743184042, "learning_rate": 4.941003584325229e-06, "loss": 0.7273, "step": 405 }, { "epoch": 0.16334741500704084, "grad_norm": 5.226443625458406, "learning_rate": 4.940242941606619e-06, "loss": 0.5695, "step": 406 }, { "epoch": 0.16374974854154092, "grad_norm": 4.3424277680190215, "learning_rate": 4.939477486055245e-06, "loss": 0.6903, "step": 407 }, { "epoch": 0.16415208207604104, "grad_norm": 4.63055317361875, "learning_rate": 4.938707219180797e-06, "loss": 0.7724, "step": 408 }, { "epoch": 0.16455441561054113, "grad_norm": 4.632025607743007, "learning_rate": 4.9379321425024545e-06, "loss": 0.5579, "step": 409 }, { "epoch": 0.16495674914504124, "grad_norm": 4.949408477476733, "learning_rate": 4.9371522575488814e-06, "loss": 0.5585, "step": 410 }, { "epoch": 0.16535908267954133, "grad_norm": 4.82833151193623, "learning_rate": 4.936367565858228e-06, "loss": 0.6281, "step": 411 }, { "epoch": 0.16576141621404145, "grad_norm": 4.135170899195267, "learning_rate": 4.935578068978121e-06, "loss": 0.5236, "step": 412 }, { "epoch": 0.16616374974854153, "grad_norm": 4.848865820547502, "learning_rate": 4.9347837684656675e-06, "loss": 0.7572, "step": 413 }, { "epoch": 0.16656608328304165, "grad_norm": 4.717705787081838, "learning_rate": 4.933984665887447e-06, "loss": 0.7282, "step": 414 }, { "epoch": 0.16696841681754174, "grad_norm": 4.528215697412672, "learning_rate": 4.93318076281951e-06, "loss": 0.5974, "step": 415 }, { "epoch": 0.16737075035204185, "grad_norm": 5.165943035201723, "learning_rate": 4.9323720608473754e-06, "loss": 0.6622, "step": 416 }, { "epoch": 0.16777308388654194, "grad_norm": 4.244718678666535, "learning_rate": 4.9315585615660275e-06, "loss": 0.5827, "step": 417 }, { "epoch": 0.16817541742104206, "grad_norm": 5.152052375848677, "learning_rate": 4.930740266579911e-06, "loss": 0.6625, "step": 418 }, { "epoch": 0.16857775095554214, "grad_norm": 4.55196067766976, "learning_rate": 4.92991717750293e-06, "loss": 0.5734, "step": 419 }, { "epoch": 0.16898008449004223, "grad_norm": 5.130317188995626, "learning_rate": 4.929089295958442e-06, "loss": 0.6448, "step": 420 }, { "epoch": 0.16938241802454235, "grad_norm": 4.221455100939359, "learning_rate": 4.928256623579259e-06, "loss": 0.5736, "step": 421 }, { "epoch": 0.16978475155904243, "grad_norm": 5.042585074077865, "learning_rate": 4.9274191620076405e-06, "loss": 0.6948, "step": 422 }, { "epoch": 0.17018708509354255, "grad_norm": 4.475716088239217, "learning_rate": 4.926576912895292e-06, "loss": 0.5762, "step": 423 }, { "epoch": 0.17058941862804264, "grad_norm": 4.629681308629434, "learning_rate": 4.9257298779033615e-06, "loss": 0.7584, "step": 424 }, { "epoch": 0.17099175216254275, "grad_norm": 4.624861162695507, "learning_rate": 4.924878058702436e-06, "loss": 0.6725, "step": 425 }, { "epoch": 0.17139408569704284, "grad_norm": 4.729048209294549, "learning_rate": 4.9240214569725375e-06, "loss": 0.7722, "step": 426 }, { "epoch": 0.17179641923154296, "grad_norm": 4.21954665313362, "learning_rate": 4.923160074403122e-06, "loss": 0.6253, "step": 427 }, { "epoch": 0.17219875276604305, "grad_norm": 3.755869045635417, "learning_rate": 4.922293912693072e-06, "loss": 0.4626, "step": 428 }, { "epoch": 0.17260108630054316, "grad_norm": 4.263780577574945, "learning_rate": 4.921422973550699e-06, "loss": 0.695, "step": 429 }, { "epoch": 0.17300341983504325, "grad_norm": 5.108523767367576, "learning_rate": 4.920547258693735e-06, "loss": 0.7092, "step": 430 }, { "epoch": 0.17340575336954336, "grad_norm": 4.317384116136861, "learning_rate": 4.919666769849332e-06, "loss": 0.4564, "step": 431 }, { "epoch": 0.17380808690404345, "grad_norm": 5.268931042465572, "learning_rate": 4.918781508754057e-06, "loss": 0.7064, "step": 432 }, { "epoch": 0.17421042043854354, "grad_norm": 4.295948103968784, "learning_rate": 4.9178914771538875e-06, "loss": 0.6892, "step": 433 }, { "epoch": 0.17461275397304366, "grad_norm": 4.214776194153066, "learning_rate": 4.916996676804214e-06, "loss": 0.5877, "step": 434 }, { "epoch": 0.17501508750754374, "grad_norm": 4.187051473349894, "learning_rate": 4.916097109469829e-06, "loss": 0.5542, "step": 435 }, { "epoch": 0.17541742104204386, "grad_norm": 4.7370251117595235, "learning_rate": 4.915192776924926e-06, "loss": 0.6131, "step": 436 }, { "epoch": 0.17581975457654395, "grad_norm": 4.790572516699059, "learning_rate": 4.9142836809531e-06, "loss": 0.6847, "step": 437 }, { "epoch": 0.17622208811104406, "grad_norm": 4.024460804862858, "learning_rate": 4.91336982334734e-06, "loss": 0.4475, "step": 438 }, { "epoch": 0.17662442164554415, "grad_norm": 4.675530435693111, "learning_rate": 4.912451205910024e-06, "loss": 0.641, "step": 439 }, { "epoch": 0.17702675518004427, "grad_norm": 4.070435859439258, "learning_rate": 4.91152783045292e-06, "loss": 0.5715, "step": 440 }, { "epoch": 0.17742908871454435, "grad_norm": 5.014878172443035, "learning_rate": 4.910599698797179e-06, "loss": 0.6227, "step": 441 }, { "epoch": 0.17783142224904447, "grad_norm": 4.471540054012383, "learning_rate": 4.909666812773333e-06, "loss": 0.5145, "step": 442 }, { "epoch": 0.17823375578354456, "grad_norm": 4.442414191534445, "learning_rate": 4.908729174221289e-06, "loss": 0.6688, "step": 443 }, { "epoch": 0.17863608931804467, "grad_norm": 4.624231827213014, "learning_rate": 4.9077867849903325e-06, "loss": 0.6327, "step": 444 }, { "epoch": 0.17903842285254476, "grad_norm": 4.743489707184984, "learning_rate": 4.906839646939113e-06, "loss": 0.7075, "step": 445 }, { "epoch": 0.17944075638704485, "grad_norm": 4.200250704275894, "learning_rate": 4.905887761935649e-06, "loss": 0.5171, "step": 446 }, { "epoch": 0.17984308992154496, "grad_norm": 4.0039914177118545, "learning_rate": 4.90493113185732e-06, "loss": 0.5334, "step": 447 }, { "epoch": 0.18024542345604505, "grad_norm": 4.437438466205263, "learning_rate": 4.903969758590865e-06, "loss": 0.584, "step": 448 }, { "epoch": 0.18064775699054517, "grad_norm": 4.668639275979892, "learning_rate": 4.9030036440323766e-06, "loss": 0.669, "step": 449 }, { "epoch": 0.18105009052504525, "grad_norm": 4.262713251256841, "learning_rate": 4.902032790087301e-06, "loss": 0.518, "step": 450 }, { "epoch": 0.18145242405954537, "grad_norm": 4.58562990715973, "learning_rate": 4.9010571986704295e-06, "loss": 0.5627, "step": 451 }, { "epoch": 0.18185475759404546, "grad_norm": 4.2511574164879615, "learning_rate": 4.900076871705897e-06, "loss": 0.6189, "step": 452 }, { "epoch": 0.18225709112854557, "grad_norm": 3.879096119034437, "learning_rate": 4.89909181112718e-06, "loss": 0.4688, "step": 453 }, { "epoch": 0.18265942466304566, "grad_norm": 4.4698856213971645, "learning_rate": 4.898102018877088e-06, "loss": 0.6272, "step": 454 }, { "epoch": 0.18306175819754578, "grad_norm": 4.25349966912043, "learning_rate": 4.897107496907767e-06, "loss": 0.4844, "step": 455 }, { "epoch": 0.18346409173204586, "grad_norm": 5.180942020863308, "learning_rate": 4.896108247180688e-06, "loss": 0.7363, "step": 456 }, { "epoch": 0.18386642526654598, "grad_norm": 4.464543366415725, "learning_rate": 4.895104271666647e-06, "loss": 0.507, "step": 457 }, { "epoch": 0.18426875880104607, "grad_norm": 4.35656943725317, "learning_rate": 4.8940955723457604e-06, "loss": 0.5626, "step": 458 }, { "epoch": 0.18467109233554616, "grad_norm": 4.46990880672421, "learning_rate": 4.893082151207464e-06, "loss": 0.5986, "step": 459 }, { "epoch": 0.18507342587004627, "grad_norm": 4.315333956337568, "learning_rate": 4.8920640102505005e-06, "loss": 0.5399, "step": 460 }, { "epoch": 0.18547575940454636, "grad_norm": 4.414262914713829, "learning_rate": 4.891041151482928e-06, "loss": 0.6866, "step": 461 }, { "epoch": 0.18587809293904647, "grad_norm": 4.68767919716049, "learning_rate": 4.8900135769221045e-06, "loss": 0.6205, "step": 462 }, { "epoch": 0.18628042647354656, "grad_norm": 4.55495109723687, "learning_rate": 4.888981288594692e-06, "loss": 0.6232, "step": 463 }, { "epoch": 0.18668276000804668, "grad_norm": 5.072112114859536, "learning_rate": 4.887944288536648e-06, "loss": 0.6246, "step": 464 }, { "epoch": 0.18708509354254677, "grad_norm": 4.598620261784645, "learning_rate": 4.886902578793221e-06, "loss": 0.6688, "step": 465 }, { "epoch": 0.18748742707704688, "grad_norm": 4.156182800147685, "learning_rate": 4.885856161418953e-06, "loss": 0.6185, "step": 466 }, { "epoch": 0.18788976061154697, "grad_norm": 5.123876789008787, "learning_rate": 4.8848050384776665e-06, "loss": 0.6485, "step": 467 }, { "epoch": 0.18829209414604708, "grad_norm": 4.7877986741742555, "learning_rate": 4.883749212042466e-06, "loss": 0.6336, "step": 468 }, { "epoch": 0.18869442768054717, "grad_norm": 4.170455407715718, "learning_rate": 4.882688684195732e-06, "loss": 0.5708, "step": 469 }, { "epoch": 0.1890967612150473, "grad_norm": 4.394557070936156, "learning_rate": 4.881623457029121e-06, "loss": 0.6584, "step": 470 }, { "epoch": 0.18949909474954738, "grad_norm": 4.4866621521500205, "learning_rate": 4.880553532643552e-06, "loss": 0.5816, "step": 471 }, { "epoch": 0.18990142828404746, "grad_norm": 4.263089971509693, "learning_rate": 4.879478913149212e-06, "loss": 0.6563, "step": 472 }, { "epoch": 0.19030376181854758, "grad_norm": 3.897184614846541, "learning_rate": 4.8783996006655485e-06, "loss": 0.5816, "step": 473 }, { "epoch": 0.19070609535304767, "grad_norm": 4.759719975025391, "learning_rate": 4.877315597321263e-06, "loss": 0.7219, "step": 474 }, { "epoch": 0.19110842888754778, "grad_norm": 4.561889671156211, "learning_rate": 4.876226905254309e-06, "loss": 0.6307, "step": 475 }, { "epoch": 0.19151076242204787, "grad_norm": 4.622879218120913, "learning_rate": 4.875133526611888e-06, "loss": 0.5739, "step": 476 }, { "epoch": 0.19191309595654799, "grad_norm": 4.183836330008031, "learning_rate": 4.874035463550445e-06, "loss": 0.4165, "step": 477 }, { "epoch": 0.19231542949104807, "grad_norm": 4.431780943086593, "learning_rate": 4.872932718235663e-06, "loss": 0.5768, "step": 478 }, { "epoch": 0.1927177630255482, "grad_norm": 4.700772397419956, "learning_rate": 4.87182529284246e-06, "loss": 0.6635, "step": 479 }, { "epoch": 0.19312009656004828, "grad_norm": 4.0384190687435915, "learning_rate": 4.870713189554985e-06, "loss": 0.4994, "step": 480 }, { "epoch": 0.1935224300945484, "grad_norm": 4.16066989214431, "learning_rate": 4.869596410566614e-06, "loss": 0.523, "step": 481 }, { "epoch": 0.19392476362904848, "grad_norm": 3.7758221004395565, "learning_rate": 4.8684749580799405e-06, "loss": 0.5106, "step": 482 }, { "epoch": 0.1943270971635486, "grad_norm": 4.792256321316127, "learning_rate": 4.867348834306781e-06, "loss": 0.5955, "step": 483 }, { "epoch": 0.19472943069804868, "grad_norm": 4.408528934091589, "learning_rate": 4.866218041468161e-06, "loss": 0.6032, "step": 484 }, { "epoch": 0.19513176423254877, "grad_norm": 4.425894822386934, "learning_rate": 4.865082581794317e-06, "loss": 0.5298, "step": 485 }, { "epoch": 0.1955340977670489, "grad_norm": 4.356368057153947, "learning_rate": 4.863942457524689e-06, "loss": 0.5381, "step": 486 }, { "epoch": 0.19593643130154897, "grad_norm": 3.990737447917161, "learning_rate": 4.862797670907915e-06, "loss": 0.5049, "step": 487 }, { "epoch": 0.1963387648360491, "grad_norm": 4.229237889406895, "learning_rate": 4.8616482242018325e-06, "loss": 0.5467, "step": 488 }, { "epoch": 0.19674109837054918, "grad_norm": 4.126506367371427, "learning_rate": 4.860494119673466e-06, "loss": 0.5455, "step": 489 }, { "epoch": 0.1971434319050493, "grad_norm": 5.231977983505604, "learning_rate": 4.859335359599031e-06, "loss": 0.7528, "step": 490 }, { "epoch": 0.19754576543954938, "grad_norm": 5.1291616835145115, "learning_rate": 4.85817194626392e-06, "loss": 0.679, "step": 491 }, { "epoch": 0.1979480989740495, "grad_norm": 3.737681461647685, "learning_rate": 4.857003881962705e-06, "loss": 0.4926, "step": 492 }, { "epoch": 0.19835043250854958, "grad_norm": 4.703925829656442, "learning_rate": 4.855831168999132e-06, "loss": 0.6439, "step": 493 }, { "epoch": 0.1987527660430497, "grad_norm": 4.627856862526262, "learning_rate": 4.854653809686115e-06, "loss": 0.6698, "step": 494 }, { "epoch": 0.1991550995775498, "grad_norm": 4.364999176797817, "learning_rate": 4.853471806345732e-06, "loss": 0.6015, "step": 495 }, { "epoch": 0.19955743311204988, "grad_norm": 4.602684651491329, "learning_rate": 4.852285161309218e-06, "loss": 0.5809, "step": 496 }, { "epoch": 0.19995976664655, "grad_norm": 4.325677574896615, "learning_rate": 4.851093876916967e-06, "loss": 0.526, "step": 497 }, { "epoch": 0.20036210018105008, "grad_norm": 4.81776828160967, "learning_rate": 4.849897955518518e-06, "loss": 0.6654, "step": 498 }, { "epoch": 0.2007644337155502, "grad_norm": 5.49681992539213, "learning_rate": 4.848697399472561e-06, "loss": 0.8468, "step": 499 }, { "epoch": 0.20116676725005028, "grad_norm": 4.782602085058219, "learning_rate": 4.8474922111469225e-06, "loss": 0.6339, "step": 500 }, { "epoch": 0.2015691007845504, "grad_norm": 4.705465090280715, "learning_rate": 4.846282392918566e-06, "loss": 0.5111, "step": 501 }, { "epoch": 0.20197143431905049, "grad_norm": 4.044280178788859, "learning_rate": 4.845067947173589e-06, "loss": 0.4453, "step": 502 }, { "epoch": 0.2023737678535506, "grad_norm": 5.299453770363048, "learning_rate": 4.843848876307211e-06, "loss": 0.7287, "step": 503 }, { "epoch": 0.2027761013880507, "grad_norm": 3.9060416143674, "learning_rate": 4.842625182723779e-06, "loss": 0.4903, "step": 504 }, { "epoch": 0.2031784349225508, "grad_norm": 4.320966869452432, "learning_rate": 4.841396868836753e-06, "loss": 0.5312, "step": 505 }, { "epoch": 0.2035807684570509, "grad_norm": 4.059590568863906, "learning_rate": 4.840163937068707e-06, "loss": 0.5953, "step": 506 }, { "epoch": 0.203983101991551, "grad_norm": 4.544560777009811, "learning_rate": 4.838926389851324e-06, "loss": 0.5274, "step": 507 }, { "epoch": 0.2043854355260511, "grad_norm": 4.643239965625325, "learning_rate": 4.837684229625389e-06, "loss": 0.6284, "step": 508 }, { "epoch": 0.20478776906055118, "grad_norm": 3.848260216211789, "learning_rate": 4.836437458840783e-06, "loss": 0.5052, "step": 509 }, { "epoch": 0.2051901025950513, "grad_norm": 4.65542258301039, "learning_rate": 4.835186079956483e-06, "loss": 0.5641, "step": 510 }, { "epoch": 0.2055924361295514, "grad_norm": 3.7048624001601738, "learning_rate": 4.8339300954405545e-06, "loss": 0.5507, "step": 511 }, { "epoch": 0.2059947696640515, "grad_norm": 4.721724963544143, "learning_rate": 4.832669507770144e-06, "loss": 0.6945, "step": 512 }, { "epoch": 0.2063971031985516, "grad_norm": 3.9937965024761892, "learning_rate": 4.83140431943148e-06, "loss": 0.5444, "step": 513 }, { "epoch": 0.2067994367330517, "grad_norm": 4.150283320138307, "learning_rate": 4.830134532919863e-06, "loss": 0.5835, "step": 514 }, { "epoch": 0.2072017702675518, "grad_norm": 5.2365671694488904, "learning_rate": 4.828860150739662e-06, "loss": 0.697, "step": 515 }, { "epoch": 0.2076041038020519, "grad_norm": 3.98085660862504, "learning_rate": 4.827581175404311e-06, "loss": 0.4872, "step": 516 }, { "epoch": 0.208006437336552, "grad_norm": 4.574435055815776, "learning_rate": 4.8262976094363016e-06, "loss": 0.5925, "step": 517 }, { "epoch": 0.2084087708710521, "grad_norm": 4.459834553682444, "learning_rate": 4.825009455367181e-06, "loss": 0.6426, "step": 518 }, { "epoch": 0.2088111044055522, "grad_norm": 3.8557681901771037, "learning_rate": 4.823716715737544e-06, "loss": 0.5719, "step": 519 }, { "epoch": 0.20921343794005232, "grad_norm": 4.776729631302629, "learning_rate": 4.82241939309703e-06, "loss": 0.6118, "step": 520 }, { "epoch": 0.2096157714745524, "grad_norm": 4.701717061474551, "learning_rate": 4.821117490004319e-06, "loss": 0.4766, "step": 521 }, { "epoch": 0.2100181050090525, "grad_norm": 4.486472816066567, "learning_rate": 4.819811009027122e-06, "loss": 0.6134, "step": 522 }, { "epoch": 0.2104204385435526, "grad_norm": 4.784573013198364, "learning_rate": 4.818499952742179e-06, "loss": 0.5982, "step": 523 }, { "epoch": 0.2108227720780527, "grad_norm": 4.953599282635229, "learning_rate": 4.817184323735258e-06, "loss": 0.5732, "step": 524 }, { "epoch": 0.2112251056125528, "grad_norm": 4.808483380881421, "learning_rate": 4.815864124601139e-06, "loss": 0.6771, "step": 525 }, { "epoch": 0.2116274391470529, "grad_norm": 4.024625165836511, "learning_rate": 4.814539357943622e-06, "loss": 0.6355, "step": 526 }, { "epoch": 0.212029772681553, "grad_norm": 3.943315026550887, "learning_rate": 4.813210026375513e-06, "loss": 0.521, "step": 527 }, { "epoch": 0.2124321062160531, "grad_norm": 3.463894710975418, "learning_rate": 4.811876132518618e-06, "loss": 0.4542, "step": 528 }, { "epoch": 0.21283443975055322, "grad_norm": 4.513349087640872, "learning_rate": 4.810537679003746e-06, "loss": 0.6093, "step": 529 }, { "epoch": 0.2132367732850533, "grad_norm": 4.649205790356446, "learning_rate": 4.8091946684706956e-06, "loss": 0.6726, "step": 530 }, { "epoch": 0.21363910681955342, "grad_norm": 4.2037442337203315, "learning_rate": 4.807847103568255e-06, "loss": 0.6265, "step": 531 }, { "epoch": 0.2140414403540535, "grad_norm": 4.893988522032171, "learning_rate": 4.806494986954195e-06, "loss": 0.7373, "step": 532 }, { "epoch": 0.21444377388855362, "grad_norm": 4.455715486715626, "learning_rate": 4.805138321295262e-06, "loss": 0.5457, "step": 533 }, { "epoch": 0.2148461074230537, "grad_norm": 4.649691607815862, "learning_rate": 4.803777109267174e-06, "loss": 0.5872, "step": 534 }, { "epoch": 0.2152484409575538, "grad_norm": 4.606972921739361, "learning_rate": 4.802411353554619e-06, "loss": 0.6556, "step": 535 }, { "epoch": 0.21565077449205391, "grad_norm": 4.262679120822924, "learning_rate": 4.801041056851244e-06, "loss": 0.5814, "step": 536 }, { "epoch": 0.216053108026554, "grad_norm": 4.734408703763602, "learning_rate": 4.7996662218596505e-06, "loss": 0.5975, "step": 537 }, { "epoch": 0.21645544156105412, "grad_norm": 4.0966775674750515, "learning_rate": 4.798286851291395e-06, "loss": 0.5373, "step": 538 }, { "epoch": 0.2168577750955542, "grad_norm": 4.794723705583196, "learning_rate": 4.796902947866976e-06, "loss": 0.6411, "step": 539 }, { "epoch": 0.21726010863005432, "grad_norm": 5.177706885118282, "learning_rate": 4.795514514315833e-06, "loss": 0.7129, "step": 540 }, { "epoch": 0.2176624421645544, "grad_norm": 4.235334038988407, "learning_rate": 4.794121553376341e-06, "loss": 0.6368, "step": 541 }, { "epoch": 0.21806477569905452, "grad_norm": 4.250039615427904, "learning_rate": 4.792724067795802e-06, "loss": 0.5897, "step": 542 }, { "epoch": 0.2184671092335546, "grad_norm": 4.248095551159398, "learning_rate": 4.7913220603304455e-06, "loss": 0.6119, "step": 543 }, { "epoch": 0.21886944276805473, "grad_norm": 4.7784528939759126, "learning_rate": 4.789915533745415e-06, "loss": 0.6428, "step": 544 }, { "epoch": 0.21927177630255482, "grad_norm": 3.7433210513797857, "learning_rate": 4.788504490814771e-06, "loss": 0.4616, "step": 545 }, { "epoch": 0.21967410983705493, "grad_norm": 4.477510756273104, "learning_rate": 4.78708893432148e-06, "loss": 0.5564, "step": 546 }, { "epoch": 0.22007644337155502, "grad_norm": 4.159189423897508, "learning_rate": 4.785668867057408e-06, "loss": 0.5005, "step": 547 }, { "epoch": 0.2204787769060551, "grad_norm": 4.801923291900065, "learning_rate": 4.784244291823321e-06, "loss": 0.6703, "step": 548 }, { "epoch": 0.22088111044055522, "grad_norm": 4.9705885588356065, "learning_rate": 4.782815211428875e-06, "loss": 0.6892, "step": 549 }, { "epoch": 0.2212834439750553, "grad_norm": 4.689425225044057, "learning_rate": 4.7813816286926116e-06, "loss": 0.6438, "step": 550 }, { "epoch": 0.22168577750955543, "grad_norm": 3.9964370496914836, "learning_rate": 4.7799435464419496e-06, "loss": 0.5283, "step": 551 }, { "epoch": 0.22208811104405551, "grad_norm": 3.9063906383285523, "learning_rate": 4.778500967513186e-06, "loss": 0.5477, "step": 552 }, { "epoch": 0.22249044457855563, "grad_norm": 4.972674347366186, "learning_rate": 4.777053894751484e-06, "loss": 0.6762, "step": 553 }, { "epoch": 0.22289277811305572, "grad_norm": 4.605207302352324, "learning_rate": 4.77560233101087e-06, "loss": 0.5823, "step": 554 }, { "epoch": 0.22329511164755583, "grad_norm": 5.023738846112498, "learning_rate": 4.774146279154231e-06, "loss": 0.645, "step": 555 }, { "epoch": 0.22369744518205592, "grad_norm": 4.21478405630217, "learning_rate": 4.772685742053299e-06, "loss": 0.6361, "step": 556 }, { "epoch": 0.22409977871655604, "grad_norm": 4.443333059344436, "learning_rate": 4.7712207225886605e-06, "loss": 0.5808, "step": 557 }, { "epoch": 0.22450211225105612, "grad_norm": 4.136335273179504, "learning_rate": 4.7697512236497355e-06, "loss": 0.4543, "step": 558 }, { "epoch": 0.22490444578555624, "grad_norm": 5.228702785570883, "learning_rate": 4.7682772481347835e-06, "loss": 0.6677, "step": 559 }, { "epoch": 0.22530677932005633, "grad_norm": 4.607023738968648, "learning_rate": 4.76679879895089e-06, "loss": 0.6972, "step": 560 }, { "epoch": 0.22570911285455642, "grad_norm": 3.9623979182055646, "learning_rate": 4.7653158790139655e-06, "loss": 0.562, "step": 561 }, { "epoch": 0.22611144638905653, "grad_norm": 4.785540664618103, "learning_rate": 4.763828491248737e-06, "loss": 0.6667, "step": 562 }, { "epoch": 0.22651377992355662, "grad_norm": 4.626975611158776, "learning_rate": 4.762336638588745e-06, "loss": 0.5222, "step": 563 }, { "epoch": 0.22691611345805673, "grad_norm": 4.148716543735861, "learning_rate": 4.760840323976333e-06, "loss": 0.5792, "step": 564 }, { "epoch": 0.22731844699255682, "grad_norm": 3.7576481851339785, "learning_rate": 4.759339550362647e-06, "loss": 0.4171, "step": 565 }, { "epoch": 0.22772078052705694, "grad_norm": 4.056873981914918, "learning_rate": 4.757834320707629e-06, "loss": 0.5823, "step": 566 }, { "epoch": 0.22812311406155703, "grad_norm": 4.485956120607346, "learning_rate": 4.7563246379800036e-06, "loss": 0.6038, "step": 567 }, { "epoch": 0.22852544759605714, "grad_norm": 4.095512060377993, "learning_rate": 4.754810505157285e-06, "loss": 0.5807, "step": 568 }, { "epoch": 0.22892778113055723, "grad_norm": 4.316692633895088, "learning_rate": 4.753291925225761e-06, "loss": 0.5689, "step": 569 }, { "epoch": 0.22933011466505734, "grad_norm": 4.0964431049127095, "learning_rate": 4.751768901180488e-06, "loss": 0.4724, "step": 570 }, { "epoch": 0.22973244819955743, "grad_norm": 3.8505086953071683, "learning_rate": 4.750241436025292e-06, "loss": 0.5599, "step": 571 }, { "epoch": 0.23013478173405755, "grad_norm": 4.97127913813276, "learning_rate": 4.748709532772756e-06, "loss": 0.7189, "step": 572 }, { "epoch": 0.23053711526855764, "grad_norm": 4.376792824386442, "learning_rate": 4.7471731944442154e-06, "loss": 0.6149, "step": 573 }, { "epoch": 0.23093944880305772, "grad_norm": 4.633754764886852, "learning_rate": 4.745632424069755e-06, "loss": 0.6118, "step": 574 }, { "epoch": 0.23134178233755784, "grad_norm": 4.682951429282118, "learning_rate": 4.744087224688197e-06, "loss": 0.6089, "step": 575 }, { "epoch": 0.23174411587205793, "grad_norm": 4.504821455509545, "learning_rate": 4.742537599347101e-06, "loss": 0.5747, "step": 576 }, { "epoch": 0.23214644940655804, "grad_norm": 4.922801626757808, "learning_rate": 4.740983551102759e-06, "loss": 0.7267, "step": 577 }, { "epoch": 0.23254878294105813, "grad_norm": 4.128104128218885, "learning_rate": 4.73942508302018e-06, "loss": 0.5344, "step": 578 }, { "epoch": 0.23295111647555825, "grad_norm": 4.502862129743137, "learning_rate": 4.7378621981730966e-06, "loss": 0.4465, "step": 579 }, { "epoch": 0.23335345001005833, "grad_norm": 4.476573377559871, "learning_rate": 4.736294899643946e-06, "loss": 0.5327, "step": 580 }, { "epoch": 0.23375578354455845, "grad_norm": 4.525722183672701, "learning_rate": 4.734723190523875e-06, "loss": 0.6182, "step": 581 }, { "epoch": 0.23415811707905854, "grad_norm": 4.4174724143850455, "learning_rate": 4.7331470739127284e-06, "loss": 0.582, "step": 582 }, { "epoch": 0.23456045061355865, "grad_norm": 4.087588153474757, "learning_rate": 4.731566552919042e-06, "loss": 0.5626, "step": 583 }, { "epoch": 0.23496278414805874, "grad_norm": 4.793168855698109, "learning_rate": 4.7299816306600395e-06, "loss": 0.712, "step": 584 }, { "epoch": 0.23536511768255883, "grad_norm": 4.088862969423477, "learning_rate": 4.728392310261628e-06, "loss": 0.5867, "step": 585 }, { "epoch": 0.23576745121705894, "grad_norm": 4.31442875545057, "learning_rate": 4.726798594858382e-06, "loss": 0.5255, "step": 586 }, { "epoch": 0.23616978475155903, "grad_norm": 4.2694537336500495, "learning_rate": 4.7252004875935506e-06, "loss": 0.6822, "step": 587 }, { "epoch": 0.23657211828605915, "grad_norm": 4.179846752454285, "learning_rate": 4.723597991619043e-06, "loss": 0.5433, "step": 588 }, { "epoch": 0.23697445182055923, "grad_norm": 4.1764050880208, "learning_rate": 4.721991110095422e-06, "loss": 0.586, "step": 589 }, { "epoch": 0.23737678535505935, "grad_norm": 3.699033328030597, "learning_rate": 4.720379846191903e-06, "loss": 0.4942, "step": 590 }, { "epoch": 0.23777911888955944, "grad_norm": 4.798015125682991, "learning_rate": 4.718764203086342e-06, "loss": 0.7269, "step": 591 }, { "epoch": 0.23818145242405955, "grad_norm": 4.974938507339555, "learning_rate": 4.717144183965233e-06, "loss": 0.6753, "step": 592 }, { "epoch": 0.23858378595855964, "grad_norm": 5.08563513422606, "learning_rate": 4.715519792023703e-06, "loss": 0.6613, "step": 593 }, { "epoch": 0.23898611949305976, "grad_norm": 4.6691045707182655, "learning_rate": 4.7138910304655e-06, "loss": 0.5635, "step": 594 }, { "epoch": 0.23938845302755984, "grad_norm": 4.366016787794075, "learning_rate": 4.712257902502992e-06, "loss": 0.6603, "step": 595 }, { "epoch": 0.23979078656205996, "grad_norm": 4.98238528037948, "learning_rate": 4.710620411357157e-06, "loss": 0.6241, "step": 596 }, { "epoch": 0.24019312009656005, "grad_norm": 4.023463566077988, "learning_rate": 4.7089785602575805e-06, "loss": 0.5202, "step": 597 }, { "epoch": 0.24059545363106014, "grad_norm": 4.040098710840788, "learning_rate": 4.707332352442446e-06, "loss": 0.5843, "step": 598 }, { "epoch": 0.24099778716556025, "grad_norm": 4.151014624312927, "learning_rate": 4.705681791158531e-06, "loss": 0.6206, "step": 599 }, { "epoch": 0.24140012070006034, "grad_norm": 4.0875213661566665, "learning_rate": 4.704026879661196e-06, "loss": 0.5614, "step": 600 }, { "epoch": 0.24180245423456045, "grad_norm": 4.553877858262055, "learning_rate": 4.7023676212143834e-06, "loss": 0.6649, "step": 601 }, { "epoch": 0.24220478776906054, "grad_norm": 4.674650194014434, "learning_rate": 4.70070401909061e-06, "loss": 0.606, "step": 602 }, { "epoch": 0.24260712130356066, "grad_norm": 4.63353917939836, "learning_rate": 4.699036076570959e-06, "loss": 0.6668, "step": 603 }, { "epoch": 0.24300945483806075, "grad_norm": 4.723870743698028, "learning_rate": 4.697363796945072e-06, "loss": 0.6728, "step": 604 }, { "epoch": 0.24341178837256086, "grad_norm": 4.386624516452244, "learning_rate": 4.6956871835111475e-06, "loss": 0.5139, "step": 605 }, { "epoch": 0.24381412190706095, "grad_norm": 4.244611479893601, "learning_rate": 4.694006239575929e-06, "loss": 0.609, "step": 606 }, { "epoch": 0.24421645544156106, "grad_norm": 4.072589096296677, "learning_rate": 4.692320968454702e-06, "loss": 0.5657, "step": 607 }, { "epoch": 0.24461878897606115, "grad_norm": 4.88910796894103, "learning_rate": 4.690631373471287e-06, "loss": 0.701, "step": 608 }, { "epoch": 0.24502112251056127, "grad_norm": 4.686738823395271, "learning_rate": 4.6889374579580315e-06, "loss": 0.637, "step": 609 }, { "epoch": 0.24542345604506136, "grad_norm": 4.905698840133413, "learning_rate": 4.687239225255805e-06, "loss": 0.6003, "step": 610 }, { "epoch": 0.24582578957956144, "grad_norm": 4.338740241210239, "learning_rate": 4.68553667871399e-06, "loss": 0.5906, "step": 611 }, { "epoch": 0.24622812311406156, "grad_norm": 4.119688719772878, "learning_rate": 4.68382982169048e-06, "loss": 0.4954, "step": 612 }, { "epoch": 0.24663045664856165, "grad_norm": 3.96691281411716, "learning_rate": 4.6821186575516665e-06, "loss": 0.6126, "step": 613 }, { "epoch": 0.24703279018306176, "grad_norm": 4.760926273551105, "learning_rate": 4.680403189672439e-06, "loss": 0.6613, "step": 614 }, { "epoch": 0.24743512371756185, "grad_norm": 4.4384316824035075, "learning_rate": 4.678683421436173e-06, "loss": 0.5687, "step": 615 }, { "epoch": 0.24783745725206197, "grad_norm": 3.931664070509481, "learning_rate": 4.676959356234726e-06, "loss": 0.5857, "step": 616 }, { "epoch": 0.24823979078656205, "grad_norm": 5.119336252733733, "learning_rate": 4.6752309974684315e-06, "loss": 0.6312, "step": 617 }, { "epoch": 0.24864212432106217, "grad_norm": 4.229793646214086, "learning_rate": 4.6734983485460895e-06, "loss": 0.5093, "step": 618 }, { "epoch": 0.24904445785556226, "grad_norm": 4.9794376894703944, "learning_rate": 4.671761412884962e-06, "loss": 0.7092, "step": 619 }, { "epoch": 0.24944679139006237, "grad_norm": 4.36204998657697, "learning_rate": 4.670020193910766e-06, "loss": 0.559, "step": 620 }, { "epoch": 0.24984912492456246, "grad_norm": 4.297208174823204, "learning_rate": 4.668274695057666e-06, "loss": 0.4965, "step": 621 }, { "epoch": 0.25025145845906255, "grad_norm": 4.235510447453734, "learning_rate": 4.666524919768267e-06, "loss": 0.5434, "step": 622 }, { "epoch": 0.25065379199356264, "grad_norm": 4.297045599708538, "learning_rate": 4.66477087149361e-06, "loss": 0.5559, "step": 623 }, { "epoch": 0.2510561255280628, "grad_norm": 5.638228122643803, "learning_rate": 4.663012553693161e-06, "loss": 0.6393, "step": 624 }, { "epoch": 0.25145845906256287, "grad_norm": 4.307369262356312, "learning_rate": 4.661249969834809e-06, "loss": 0.5516, "step": 625 }, { "epoch": 0.25186079259706295, "grad_norm": 4.348008968073508, "learning_rate": 4.659483123394855e-06, "loss": 0.5858, "step": 626 }, { "epoch": 0.25226312613156304, "grad_norm": 4.245775793874246, "learning_rate": 4.657712017858011e-06, "loss": 0.5985, "step": 627 }, { "epoch": 0.2526654596660632, "grad_norm": 4.27241102613026, "learning_rate": 4.6559366567173824e-06, "loss": 0.5247, "step": 628 }, { "epoch": 0.2530677932005633, "grad_norm": 3.9143192748019384, "learning_rate": 4.6541570434744735e-06, "loss": 0.4936, "step": 629 }, { "epoch": 0.25347012673506336, "grad_norm": 4.550429412487192, "learning_rate": 4.6523731816391725e-06, "loss": 0.7499, "step": 630 }, { "epoch": 0.25387246026956345, "grad_norm": 4.202697728155226, "learning_rate": 4.650585074729747e-06, "loss": 0.5184, "step": 631 }, { "epoch": 0.2542747938040636, "grad_norm": 4.083559252407811, "learning_rate": 4.648792726272838e-06, "loss": 0.5242, "step": 632 }, { "epoch": 0.2546771273385637, "grad_norm": 3.9861255665329267, "learning_rate": 4.646996139803452e-06, "loss": 0.5269, "step": 633 }, { "epoch": 0.25507946087306377, "grad_norm": 4.559490860669659, "learning_rate": 4.645195318864951e-06, "loss": 0.6372, "step": 634 }, { "epoch": 0.25548179440756386, "grad_norm": 4.394457869705605, "learning_rate": 4.643390267009054e-06, "loss": 0.5105, "step": 635 }, { "epoch": 0.25588412794206394, "grad_norm": 5.018376167864149, "learning_rate": 4.641580987795821e-06, "loss": 0.6628, "step": 636 }, { "epoch": 0.2562864614765641, "grad_norm": 3.7893948824739465, "learning_rate": 4.639767484793648e-06, "loss": 0.508, "step": 637 }, { "epoch": 0.2566887950110642, "grad_norm": 4.596240270523842, "learning_rate": 4.637949761579266e-06, "loss": 0.6529, "step": 638 }, { "epoch": 0.25709112854556426, "grad_norm": 5.140583964722019, "learning_rate": 4.636127821737726e-06, "loss": 0.6489, "step": 639 }, { "epoch": 0.25749346208006435, "grad_norm": 3.8394758085160823, "learning_rate": 4.634301668862397e-06, "loss": 0.4736, "step": 640 }, { "epoch": 0.2578957956145645, "grad_norm": 4.38356346084736, "learning_rate": 4.632471306554955e-06, "loss": 0.5353, "step": 641 }, { "epoch": 0.2582981291490646, "grad_norm": 3.8889360472386185, "learning_rate": 4.630636738425381e-06, "loss": 0.5511, "step": 642 }, { "epoch": 0.25870046268356467, "grad_norm": 4.817925477512709, "learning_rate": 4.62879796809195e-06, "loss": 0.6963, "step": 643 }, { "epoch": 0.25910279621806476, "grad_norm": 4.742191877421797, "learning_rate": 4.626954999181224e-06, "loss": 0.7752, "step": 644 }, { "epoch": 0.2595051297525649, "grad_norm": 4.245951623762571, "learning_rate": 4.625107835328048e-06, "loss": 0.5754, "step": 645 }, { "epoch": 0.259907463287065, "grad_norm": 4.836575633325676, "learning_rate": 4.623256480175537e-06, "loss": 0.6639, "step": 646 }, { "epoch": 0.2603097968215651, "grad_norm": 3.806176284035789, "learning_rate": 4.621400937375075e-06, "loss": 0.5292, "step": 647 }, { "epoch": 0.26071213035606516, "grad_norm": 4.025115976057339, "learning_rate": 4.619541210586307e-06, "loss": 0.4815, "step": 648 }, { "epoch": 0.26111446389056525, "grad_norm": 4.1347392050793825, "learning_rate": 4.617677303477124e-06, "loss": 0.5681, "step": 649 }, { "epoch": 0.2615167974250654, "grad_norm": 4.371554358928095, "learning_rate": 4.615809219723667e-06, "loss": 0.592, "step": 650 }, { "epoch": 0.2619191309595655, "grad_norm": 4.976927834726334, "learning_rate": 4.613936963010314e-06, "loss": 0.7898, "step": 651 }, { "epoch": 0.26232146449406557, "grad_norm": 4.80195477309382, "learning_rate": 4.612060537029671e-06, "loss": 0.6565, "step": 652 }, { "epoch": 0.26272379802856566, "grad_norm": 4.2203256747490805, "learning_rate": 4.610179945482568e-06, "loss": 0.5597, "step": 653 }, { "epoch": 0.2631261315630658, "grad_norm": 4.3119037789418035, "learning_rate": 4.608295192078051e-06, "loss": 0.5773, "step": 654 }, { "epoch": 0.2635284650975659, "grad_norm": 3.938767904065124, "learning_rate": 4.606406280533373e-06, "loss": 0.4691, "step": 655 }, { "epoch": 0.263930798632066, "grad_norm": 3.873313542196617, "learning_rate": 4.6045132145739914e-06, "loss": 0.5446, "step": 656 }, { "epoch": 0.26433313216656606, "grad_norm": 4.720535373828625, "learning_rate": 4.602615997933552e-06, "loss": 0.6451, "step": 657 }, { "epoch": 0.2647354657010662, "grad_norm": 5.319338434241136, "learning_rate": 4.600714634353893e-06, "loss": 0.6786, "step": 658 }, { "epoch": 0.2651377992355663, "grad_norm": 3.986441538874652, "learning_rate": 4.598809127585026e-06, "loss": 0.5815, "step": 659 }, { "epoch": 0.2655401327700664, "grad_norm": 4.190229942704305, "learning_rate": 4.596899481385137e-06, "loss": 0.5933, "step": 660 }, { "epoch": 0.26594246630456647, "grad_norm": 4.211284711325467, "learning_rate": 4.5949856995205745e-06, "loss": 0.6062, "step": 661 }, { "epoch": 0.26634479983906656, "grad_norm": 5.022372015405014, "learning_rate": 4.593067785765846e-06, "loss": 0.6632, "step": 662 }, { "epoch": 0.2667471333735667, "grad_norm": 3.9365356585193148, "learning_rate": 4.5911457439036075e-06, "loss": 0.5324, "step": 663 }, { "epoch": 0.2671494669080668, "grad_norm": 4.891303145874928, "learning_rate": 4.589219577724654e-06, "loss": 0.7504, "step": 664 }, { "epoch": 0.2675518004425669, "grad_norm": 3.855940902006332, "learning_rate": 4.5872892910279185e-06, "loss": 0.5476, "step": 665 }, { "epoch": 0.26795413397706697, "grad_norm": 4.37750980680622, "learning_rate": 4.58535488762046e-06, "loss": 0.5902, "step": 666 }, { "epoch": 0.2683564675115671, "grad_norm": 4.450877429194335, "learning_rate": 4.583416371317454e-06, "loss": 0.6345, "step": 667 }, { "epoch": 0.2687588010460672, "grad_norm": 4.162519879691083, "learning_rate": 4.581473745942191e-06, "loss": 0.444, "step": 668 }, { "epoch": 0.2691611345805673, "grad_norm": 4.2236499578643025, "learning_rate": 4.579527015326065e-06, "loss": 0.492, "step": 669 }, { "epoch": 0.2695634681150674, "grad_norm": 4.482147273633291, "learning_rate": 4.5775761833085645e-06, "loss": 0.5648, "step": 670 }, { "epoch": 0.2699658016495675, "grad_norm": 4.760060597047131, "learning_rate": 4.57562125373727e-06, "loss": 0.6674, "step": 671 }, { "epoch": 0.2703681351840676, "grad_norm": 4.362389446472747, "learning_rate": 4.573662230467844e-06, "loss": 0.577, "step": 672 }, { "epoch": 0.2707704687185677, "grad_norm": 3.8159558241349987, "learning_rate": 4.5716991173640165e-06, "loss": 0.5185, "step": 673 }, { "epoch": 0.2711728022530678, "grad_norm": 4.005051687655319, "learning_rate": 4.5697319182975944e-06, "loss": 0.5922, "step": 674 }, { "epoch": 0.27157513578756787, "grad_norm": 4.667139467518304, "learning_rate": 4.567760637148432e-06, "loss": 0.6032, "step": 675 }, { "epoch": 0.271977469322068, "grad_norm": 3.9606776473436405, "learning_rate": 4.5657852778044435e-06, "loss": 0.5174, "step": 676 }, { "epoch": 0.2723798028565681, "grad_norm": 4.322780347285095, "learning_rate": 4.5638058441615815e-06, "loss": 0.6089, "step": 677 }, { "epoch": 0.2727821363910682, "grad_norm": 4.433317399280547, "learning_rate": 4.561822340123836e-06, "loss": 0.6047, "step": 678 }, { "epoch": 0.2731844699255683, "grad_norm": 4.552805703827102, "learning_rate": 4.559834769603224e-06, "loss": 0.6281, "step": 679 }, { "epoch": 0.2735868034600684, "grad_norm": 5.128182512733822, "learning_rate": 4.557843136519784e-06, "loss": 0.6697, "step": 680 }, { "epoch": 0.2739891369945685, "grad_norm": 4.063920951512369, "learning_rate": 4.555847444801565e-06, "loss": 0.5885, "step": 681 }, { "epoch": 0.2743914705290686, "grad_norm": 4.908841427141714, "learning_rate": 4.5538476983846245e-06, "loss": 0.6771, "step": 682 }, { "epoch": 0.2747938040635687, "grad_norm": 4.405724720337232, "learning_rate": 4.551843901213012e-06, "loss": 0.6479, "step": 683 }, { "epoch": 0.2751961375980688, "grad_norm": 3.945129587052982, "learning_rate": 4.549836057238769e-06, "loss": 0.6039, "step": 684 }, { "epoch": 0.2755984711325689, "grad_norm": 4.96376428039534, "learning_rate": 4.547824170421921e-06, "loss": 0.7191, "step": 685 }, { "epoch": 0.276000804667069, "grad_norm": 4.378595779287015, "learning_rate": 4.54580824473046e-06, "loss": 0.5549, "step": 686 }, { "epoch": 0.2764031382015691, "grad_norm": 4.262426880188478, "learning_rate": 4.54378828414035e-06, "loss": 0.5024, "step": 687 }, { "epoch": 0.2768054717360692, "grad_norm": 4.426695285396216, "learning_rate": 4.541764292635512e-06, "loss": 0.7052, "step": 688 }, { "epoch": 0.2772078052705693, "grad_norm": 4.473404410086587, "learning_rate": 4.5397362742078145e-06, "loss": 0.5557, "step": 689 }, { "epoch": 0.2776101388050694, "grad_norm": 4.052844689460026, "learning_rate": 4.537704232857069e-06, "loss": 0.4541, "step": 690 }, { "epoch": 0.2780124723395695, "grad_norm": 3.89050386632389, "learning_rate": 4.5356681725910255e-06, "loss": 0.549, "step": 691 }, { "epoch": 0.2784148058740696, "grad_norm": 4.349538143222671, "learning_rate": 4.5336280974253535e-06, "loss": 0.6076, "step": 692 }, { "epoch": 0.2788171394085697, "grad_norm": 4.361849160433631, "learning_rate": 4.5315840113836454e-06, "loss": 0.6524, "step": 693 }, { "epoch": 0.2792194729430698, "grad_norm": 4.617234754063093, "learning_rate": 4.529535918497403e-06, "loss": 0.5616, "step": 694 }, { "epoch": 0.2796218064775699, "grad_norm": 4.09023083083644, "learning_rate": 4.5274838228060326e-06, "loss": 0.5407, "step": 695 }, { "epoch": 0.28002414001207, "grad_norm": 3.7157295921866718, "learning_rate": 4.525427728356832e-06, "loss": 0.5361, "step": 696 }, { "epoch": 0.28042647354657013, "grad_norm": 4.457227461962157, "learning_rate": 4.523367639204985e-06, "loss": 0.5593, "step": 697 }, { "epoch": 0.2808288070810702, "grad_norm": 4.348610370546784, "learning_rate": 4.52130355941356e-06, "loss": 0.5121, "step": 698 }, { "epoch": 0.2812311406155703, "grad_norm": 3.9173059552199194, "learning_rate": 4.519235493053491e-06, "loss": 0.4927, "step": 699 }, { "epoch": 0.2816334741500704, "grad_norm": 4.208431911956715, "learning_rate": 4.517163444203575e-06, "loss": 0.5504, "step": 700 }, { "epoch": 0.2820358076845705, "grad_norm": 4.404532536673193, "learning_rate": 4.515087416950464e-06, "loss": 0.6057, "step": 701 }, { "epoch": 0.2824381412190706, "grad_norm": 4.163697909062566, "learning_rate": 4.513007415388659e-06, "loss": 0.4682, "step": 702 }, { "epoch": 0.2828404747535707, "grad_norm": 5.152927142247097, "learning_rate": 4.510923443620494e-06, "loss": 0.6601, "step": 703 }, { "epoch": 0.2832428082880708, "grad_norm": 4.544361331270338, "learning_rate": 4.50883550575614e-06, "loss": 0.5968, "step": 704 }, { "epoch": 0.2836451418225709, "grad_norm": 4.134394917706537, "learning_rate": 4.5067436059135836e-06, "loss": 0.6647, "step": 705 }, { "epoch": 0.28404747535707103, "grad_norm": 4.734085611831448, "learning_rate": 4.504647748218633e-06, "loss": 0.601, "step": 706 }, { "epoch": 0.2844498088915711, "grad_norm": 4.681267518422373, "learning_rate": 4.502547936804894e-06, "loss": 0.5539, "step": 707 }, { "epoch": 0.2848521424260712, "grad_norm": 4.325088947732379, "learning_rate": 4.500444175813776e-06, "loss": 0.5872, "step": 708 }, { "epoch": 0.2852544759605713, "grad_norm": 3.850534498836246, "learning_rate": 4.4983364693944775e-06, "loss": 0.5061, "step": 709 }, { "epoch": 0.28565680949507144, "grad_norm": 4.373584617561936, "learning_rate": 4.496224821703977e-06, "loss": 0.6693, "step": 710 }, { "epoch": 0.2860591430295715, "grad_norm": 4.377477636040013, "learning_rate": 4.494109236907026e-06, "loss": 0.5966, "step": 711 }, { "epoch": 0.2864614765640716, "grad_norm": 4.271381874115667, "learning_rate": 4.491989719176142e-06, "loss": 0.6221, "step": 712 }, { "epoch": 0.2868638100985717, "grad_norm": 3.7259627257148202, "learning_rate": 4.489866272691599e-06, "loss": 0.4313, "step": 713 }, { "epoch": 0.2872661436330718, "grad_norm": 4.0336320558123155, "learning_rate": 4.48773890164142e-06, "loss": 0.5305, "step": 714 }, { "epoch": 0.28766847716757193, "grad_norm": 4.585992205034134, "learning_rate": 4.485607610221367e-06, "loss": 0.7636, "step": 715 }, { "epoch": 0.288070810702072, "grad_norm": 4.67820441036977, "learning_rate": 4.4834724026349376e-06, "loss": 0.5866, "step": 716 }, { "epoch": 0.2884731442365721, "grad_norm": 5.012556667522281, "learning_rate": 4.4813332830933484e-06, "loss": 0.6498, "step": 717 }, { "epoch": 0.2888754777710722, "grad_norm": 4.19440277086636, "learning_rate": 4.4791902558155345e-06, "loss": 0.5434, "step": 718 }, { "epoch": 0.28927781130557234, "grad_norm": 4.330716534353544, "learning_rate": 4.477043325028137e-06, "loss": 0.5507, "step": 719 }, { "epoch": 0.28968014484007243, "grad_norm": 4.097976140286037, "learning_rate": 4.474892494965495e-06, "loss": 0.4591, "step": 720 }, { "epoch": 0.2900824783745725, "grad_norm": 4.132010732456553, "learning_rate": 4.472737769869642e-06, "loss": 0.5217, "step": 721 }, { "epoch": 0.2904848119090726, "grad_norm": 3.7815121295908343, "learning_rate": 4.470579153990288e-06, "loss": 0.4932, "step": 722 }, { "epoch": 0.29088714544357275, "grad_norm": 3.724905186650606, "learning_rate": 4.468416651584822e-06, "loss": 0.5448, "step": 723 }, { "epoch": 0.29128947897807284, "grad_norm": 3.82443773314746, "learning_rate": 4.4662502669182935e-06, "loss": 0.5298, "step": 724 }, { "epoch": 0.2916918125125729, "grad_norm": 4.343371987708737, "learning_rate": 4.464080004263411e-06, "loss": 0.606, "step": 725 }, { "epoch": 0.292094146047073, "grad_norm": 4.009570342579401, "learning_rate": 4.461905867900535e-06, "loss": 0.5916, "step": 726 }, { "epoch": 0.2924964795815731, "grad_norm": 4.148496022814315, "learning_rate": 4.459727862117658e-06, "loss": 0.5155, "step": 727 }, { "epoch": 0.29289881311607324, "grad_norm": 4.568599931304069, "learning_rate": 4.457545991210412e-06, "loss": 0.555, "step": 728 }, { "epoch": 0.29330114665057333, "grad_norm": 4.026645287125634, "learning_rate": 4.455360259482047e-06, "loss": 0.5265, "step": 729 }, { "epoch": 0.2937034801850734, "grad_norm": 4.4466778657917265, "learning_rate": 4.4531706712434305e-06, "loss": 0.6318, "step": 730 }, { "epoch": 0.2941058137195735, "grad_norm": 4.444852619715758, "learning_rate": 4.450977230813035e-06, "loss": 0.671, "step": 731 }, { "epoch": 0.29450814725407365, "grad_norm": 5.136550532204147, "learning_rate": 4.44877994251693e-06, "loss": 0.611, "step": 732 }, { "epoch": 0.29491048078857374, "grad_norm": 4.472272090985854, "learning_rate": 4.446578810688774e-06, "loss": 0.595, "step": 733 }, { "epoch": 0.2953128143230738, "grad_norm": 4.755288406312762, "learning_rate": 4.444373839669808e-06, "loss": 0.801, "step": 734 }, { "epoch": 0.2957151478575739, "grad_norm": 4.964978523542175, "learning_rate": 4.442165033808843e-06, "loss": 0.6712, "step": 735 }, { "epoch": 0.29611748139207406, "grad_norm": 4.590633696444609, "learning_rate": 4.439952397462254e-06, "loss": 0.6214, "step": 736 }, { "epoch": 0.29651981492657414, "grad_norm": 4.05753322012328, "learning_rate": 4.43773593499397e-06, "loss": 0.4477, "step": 737 }, { "epoch": 0.29692214846107423, "grad_norm": 4.354711360779599, "learning_rate": 4.435515650775468e-06, "loss": 0.5618, "step": 738 }, { "epoch": 0.2973244819955743, "grad_norm": 4.543128550216277, "learning_rate": 4.433291549185761e-06, "loss": 0.4807, "step": 739 }, { "epoch": 0.2977268155300744, "grad_norm": 3.9444140759236737, "learning_rate": 4.431063634611392e-06, "loss": 0.5393, "step": 740 }, { "epoch": 0.29812914906457455, "grad_norm": 4.516456335270593, "learning_rate": 4.428831911446422e-06, "loss": 0.5839, "step": 741 }, { "epoch": 0.29853148259907464, "grad_norm": 4.4607761036583815, "learning_rate": 4.426596384092426e-06, "loss": 0.6538, "step": 742 }, { "epoch": 0.2989338161335747, "grad_norm": 6.196021389574422, "learning_rate": 4.424357056958483e-06, "loss": 0.6484, "step": 743 }, { "epoch": 0.2993361496680748, "grad_norm": 4.251042887568661, "learning_rate": 4.422113934461161e-06, "loss": 0.5523, "step": 744 }, { "epoch": 0.29973848320257496, "grad_norm": 4.697763472346534, "learning_rate": 4.4198670210245186e-06, "loss": 0.6053, "step": 745 }, { "epoch": 0.30014081673707504, "grad_norm": 4.665389980976664, "learning_rate": 4.41761632108009e-06, "loss": 0.6618, "step": 746 }, { "epoch": 0.30054315027157513, "grad_norm": 4.2980125151698525, "learning_rate": 4.415361839066874e-06, "loss": 0.4863, "step": 747 }, { "epoch": 0.3009454838060752, "grad_norm": 4.25422933800907, "learning_rate": 4.413103579431335e-06, "loss": 0.6285, "step": 748 }, { "epoch": 0.30134781734057536, "grad_norm": 4.408647829220997, "learning_rate": 4.410841546627383e-06, "loss": 0.6214, "step": 749 }, { "epoch": 0.30175015087507545, "grad_norm": 3.5964308822983626, "learning_rate": 4.408575745116371e-06, "loss": 0.4709, "step": 750 }, { "epoch": 0.30215248440957554, "grad_norm": 4.177190728366978, "learning_rate": 4.406306179367086e-06, "loss": 0.6584, "step": 751 }, { "epoch": 0.3025548179440756, "grad_norm": 3.9436659867851, "learning_rate": 4.404032853855736e-06, "loss": 0.5804, "step": 752 }, { "epoch": 0.3029571514785757, "grad_norm": 4.5139405502647465, "learning_rate": 4.40175577306595e-06, "loss": 0.6131, "step": 753 }, { "epoch": 0.30335948501307586, "grad_norm": 4.656972033870911, "learning_rate": 4.399474941488756e-06, "loss": 0.5378, "step": 754 }, { "epoch": 0.30376181854757595, "grad_norm": 4.2754391925078785, "learning_rate": 4.397190363622588e-06, "loss": 0.5117, "step": 755 }, { "epoch": 0.30416415208207603, "grad_norm": 4.057091286795055, "learning_rate": 4.3949020439732594e-06, "loss": 0.4969, "step": 756 }, { "epoch": 0.3045664856165761, "grad_norm": 3.914038182134245, "learning_rate": 4.392609987053972e-06, "loss": 0.4484, "step": 757 }, { "epoch": 0.30496881915107626, "grad_norm": 4.364733148673897, "learning_rate": 4.390314197385292e-06, "loss": 0.662, "step": 758 }, { "epoch": 0.30537115268557635, "grad_norm": 4.866819672714999, "learning_rate": 4.388014679495154e-06, "loss": 0.6832, "step": 759 }, { "epoch": 0.30577348622007644, "grad_norm": 4.563239673265389, "learning_rate": 4.385711437918839e-06, "loss": 0.633, "step": 760 }, { "epoch": 0.30617581975457653, "grad_norm": 3.952293727810083, "learning_rate": 4.383404477198976e-06, "loss": 0.4319, "step": 761 }, { "epoch": 0.30657815328907667, "grad_norm": 3.8982970475676866, "learning_rate": 4.38109380188553e-06, "loss": 0.5412, "step": 762 }, { "epoch": 0.30698048682357676, "grad_norm": 4.767355411844817, "learning_rate": 4.3787794165357875e-06, "loss": 0.6688, "step": 763 }, { "epoch": 0.30738282035807685, "grad_norm": 5.490594878274815, "learning_rate": 4.3764613257143585e-06, "loss": 0.8315, "step": 764 }, { "epoch": 0.30778515389257693, "grad_norm": 4.4753569929038965, "learning_rate": 4.374139533993157e-06, "loss": 0.5577, "step": 765 }, { "epoch": 0.308187487427077, "grad_norm": 4.234373998218113, "learning_rate": 4.371814045951396e-06, "loss": 0.6195, "step": 766 }, { "epoch": 0.30858982096157717, "grad_norm": 3.8789979804624277, "learning_rate": 4.369484866175581e-06, "loss": 0.5812, "step": 767 }, { "epoch": 0.30899215449607725, "grad_norm": 4.3774657256617635, "learning_rate": 4.367151999259498e-06, "loss": 0.5334, "step": 768 }, { "epoch": 0.30939448803057734, "grad_norm": 4.413468965671534, "learning_rate": 4.364815449804204e-06, "loss": 0.6043, "step": 769 }, { "epoch": 0.30979682156507743, "grad_norm": 4.977248443532888, "learning_rate": 4.362475222418019e-06, "loss": 0.7427, "step": 770 }, { "epoch": 0.3101991550995776, "grad_norm": 4.449181903975735, "learning_rate": 4.360131321716518e-06, "loss": 0.5336, "step": 771 }, { "epoch": 0.31060148863407766, "grad_norm": 4.203059807958007, "learning_rate": 4.357783752322522e-06, "loss": 0.5362, "step": 772 }, { "epoch": 0.31100382216857775, "grad_norm": 3.8919997316821004, "learning_rate": 4.355432518866084e-06, "loss": 0.5202, "step": 773 }, { "epoch": 0.31140615570307784, "grad_norm": 4.125644261476737, "learning_rate": 4.353077625984484e-06, "loss": 0.5136, "step": 774 }, { "epoch": 0.311808489237578, "grad_norm": 3.7230845852814687, "learning_rate": 4.3507190783222245e-06, "loss": 0.4731, "step": 775 }, { "epoch": 0.31221082277207807, "grad_norm": 4.619811557245368, "learning_rate": 4.348356880531011e-06, "loss": 0.698, "step": 776 }, { "epoch": 0.31261315630657815, "grad_norm": 4.064918048601741, "learning_rate": 4.345991037269748e-06, "loss": 0.5594, "step": 777 }, { "epoch": 0.31301548984107824, "grad_norm": 4.28030910141664, "learning_rate": 4.343621553204533e-06, "loss": 0.6272, "step": 778 }, { "epoch": 0.31341782337557833, "grad_norm": 4.3485146730819, "learning_rate": 4.341248433008645e-06, "loss": 0.6078, "step": 779 }, { "epoch": 0.3138201569100785, "grad_norm": 4.211683329853448, "learning_rate": 4.338871681362528e-06, "loss": 0.5247, "step": 780 }, { "epoch": 0.31422249044457856, "grad_norm": 3.8752448583148453, "learning_rate": 4.3364913029537946e-06, "loss": 0.5274, "step": 781 }, { "epoch": 0.31462482397907865, "grad_norm": 3.999853155987491, "learning_rate": 4.334107302477208e-06, "loss": 0.5421, "step": 782 }, { "epoch": 0.31502715751357874, "grad_norm": 4.961501923709214, "learning_rate": 4.331719684634676e-06, "loss": 0.7151, "step": 783 }, { "epoch": 0.3154294910480789, "grad_norm": 4.148580969978077, "learning_rate": 4.3293284541352384e-06, "loss": 0.5084, "step": 784 }, { "epoch": 0.31583182458257897, "grad_norm": 3.623114769478068, "learning_rate": 4.326933615695064e-06, "loss": 0.4912, "step": 785 }, { "epoch": 0.31623415811707906, "grad_norm": 4.179951168808134, "learning_rate": 4.324535174037433e-06, "loss": 0.4943, "step": 786 }, { "epoch": 0.31663649165157914, "grad_norm": 4.206035486231768, "learning_rate": 4.322133133892737e-06, "loss": 0.6175, "step": 787 }, { "epoch": 0.3170388251860793, "grad_norm": 4.029670988995613, "learning_rate": 4.31972749999846e-06, "loss": 0.5329, "step": 788 }, { "epoch": 0.3174411587205794, "grad_norm": 4.024639871547482, "learning_rate": 4.317318277099178e-06, "loss": 0.5121, "step": 789 }, { "epoch": 0.31784349225507946, "grad_norm": 4.04730827392964, "learning_rate": 4.314905469946542e-06, "loss": 0.4979, "step": 790 }, { "epoch": 0.31824582578957955, "grad_norm": 4.440133410808506, "learning_rate": 4.312489083299275e-06, "loss": 0.6333, "step": 791 }, { "epoch": 0.31864815932407964, "grad_norm": 4.079663513764819, "learning_rate": 4.310069121923158e-06, "loss": 0.5987, "step": 792 }, { "epoch": 0.3190504928585798, "grad_norm": 4.17692958034715, "learning_rate": 4.307645590591024e-06, "loss": 0.4798, "step": 793 }, { "epoch": 0.31945282639307987, "grad_norm": 4.417314019145798, "learning_rate": 4.305218494082744e-06, "loss": 0.4832, "step": 794 }, { "epoch": 0.31985515992757996, "grad_norm": 4.920563992053252, "learning_rate": 4.302787837185224e-06, "loss": 0.6288, "step": 795 }, { "epoch": 0.32025749346208005, "grad_norm": 4.067969775601857, "learning_rate": 4.300353624692389e-06, "loss": 0.52, "step": 796 }, { "epoch": 0.3206598269965802, "grad_norm": 4.232889302389981, "learning_rate": 4.2979158614051805e-06, "loss": 0.6205, "step": 797 }, { "epoch": 0.3210621605310803, "grad_norm": 4.335177482388414, "learning_rate": 4.295474552131538e-06, "loss": 0.6153, "step": 798 }, { "epoch": 0.32146449406558036, "grad_norm": 4.698659170570413, "learning_rate": 4.2930297016863985e-06, "loss": 0.6404, "step": 799 }, { "epoch": 0.32186682760008045, "grad_norm": 3.931383036386479, "learning_rate": 4.2905813148916816e-06, "loss": 0.599, "step": 800 }, { "epoch": 0.32226916113458054, "grad_norm": 5.048444025824699, "learning_rate": 4.288129396576284e-06, "loss": 0.7259, "step": 801 }, { "epoch": 0.3226714946690807, "grad_norm": 4.851397090992265, "learning_rate": 4.285673951576062e-06, "loss": 0.642, "step": 802 }, { "epoch": 0.32307382820358077, "grad_norm": 4.28645665558046, "learning_rate": 4.283214984733833e-06, "loss": 0.5315, "step": 803 }, { "epoch": 0.32347616173808086, "grad_norm": 4.265366711136778, "learning_rate": 4.28075250089936e-06, "loss": 0.5529, "step": 804 }, { "epoch": 0.32387849527258095, "grad_norm": 3.848835311590686, "learning_rate": 4.278286504929338e-06, "loss": 0.5323, "step": 805 }, { "epoch": 0.3242808288070811, "grad_norm": 4.627843495385828, "learning_rate": 4.2758170016873934e-06, "loss": 0.5561, "step": 806 }, { "epoch": 0.3246831623415812, "grad_norm": 4.363493591898399, "learning_rate": 4.273343996044068e-06, "loss": 0.4805, "step": 807 }, { "epoch": 0.32508549587608127, "grad_norm": 3.9805290188441043, "learning_rate": 4.270867492876812e-06, "loss": 0.5204, "step": 808 }, { "epoch": 0.32548782941058135, "grad_norm": 3.9644340799634588, "learning_rate": 4.268387497069974e-06, "loss": 0.4642, "step": 809 }, { "epoch": 0.3258901629450815, "grad_norm": 5.269711583475643, "learning_rate": 4.265904013514788e-06, "loss": 0.6183, "step": 810 }, { "epoch": 0.3262924964795816, "grad_norm": 4.6945618708498165, "learning_rate": 4.263417047109371e-06, "loss": 0.6567, "step": 811 }, { "epoch": 0.32669483001408167, "grad_norm": 4.413878483629632, "learning_rate": 4.260926602758707e-06, "loss": 0.5518, "step": 812 }, { "epoch": 0.32709716354858176, "grad_norm": 4.355371114130468, "learning_rate": 4.258432685374641e-06, "loss": 0.478, "step": 813 }, { "epoch": 0.32749949708308185, "grad_norm": 4.625144876838143, "learning_rate": 4.255935299875864e-06, "loss": 0.5918, "step": 814 }, { "epoch": 0.327901830617582, "grad_norm": 3.8722185488332994, "learning_rate": 4.253434451187911e-06, "loss": 0.4698, "step": 815 }, { "epoch": 0.3283041641520821, "grad_norm": 3.8155166378719403, "learning_rate": 4.250930144243147e-06, "loss": 0.43, "step": 816 }, { "epoch": 0.32870649768658217, "grad_norm": 4.373317134651141, "learning_rate": 4.248422383980756e-06, "loss": 0.581, "step": 817 }, { "epoch": 0.32910883122108225, "grad_norm": 4.316760308023842, "learning_rate": 4.245911175346733e-06, "loss": 0.525, "step": 818 }, { "epoch": 0.3295111647555824, "grad_norm": 4.1806833930239895, "learning_rate": 4.243396523293875e-06, "loss": 0.6027, "step": 819 }, { "epoch": 0.3299134982900825, "grad_norm": 4.011770765526375, "learning_rate": 4.240878432781769e-06, "loss": 0.5642, "step": 820 }, { "epoch": 0.3303158318245826, "grad_norm": 4.106735076414742, "learning_rate": 4.238356908776786e-06, "loss": 0.5576, "step": 821 }, { "epoch": 0.33071816535908266, "grad_norm": 4.7608485825878955, "learning_rate": 4.235831956252069e-06, "loss": 0.7088, "step": 822 }, { "epoch": 0.3311204988935828, "grad_norm": 4.8786607685170305, "learning_rate": 4.2333035801875175e-06, "loss": 0.6513, "step": 823 }, { "epoch": 0.3315228324280829, "grad_norm": 4.288963783688971, "learning_rate": 4.230771785569791e-06, "loss": 0.5073, "step": 824 }, { "epoch": 0.331925165962583, "grad_norm": 4.014903400829248, "learning_rate": 4.228236577392285e-06, "loss": 0.5901, "step": 825 }, { "epoch": 0.33232749949708307, "grad_norm": 4.154812074980563, "learning_rate": 4.225697960655131e-06, "loss": 0.5568, "step": 826 }, { "epoch": 0.33272983303158316, "grad_norm": 3.827755329803438, "learning_rate": 4.223155940365181e-06, "loss": 0.4738, "step": 827 }, { "epoch": 0.3331321665660833, "grad_norm": 3.856077516315061, "learning_rate": 4.220610521536002e-06, "loss": 0.4963, "step": 828 }, { "epoch": 0.3335345001005834, "grad_norm": 4.615466783911023, "learning_rate": 4.218061709187862e-06, "loss": 0.5656, "step": 829 }, { "epoch": 0.3339368336350835, "grad_norm": 4.176966772191371, "learning_rate": 4.215509508347723e-06, "loss": 0.5398, "step": 830 }, { "epoch": 0.33433916716958356, "grad_norm": 4.49327690032652, "learning_rate": 4.212953924049229e-06, "loss": 0.6772, "step": 831 }, { "epoch": 0.3347415007040837, "grad_norm": 4.067266904294674, "learning_rate": 4.210394961332698e-06, "loss": 0.5311, "step": 832 }, { "epoch": 0.3351438342385838, "grad_norm": 3.7375221543142128, "learning_rate": 4.207832625245112e-06, "loss": 0.4573, "step": 833 }, { "epoch": 0.3355461677730839, "grad_norm": 4.056435688935082, "learning_rate": 4.205266920840103e-06, "loss": 0.543, "step": 834 }, { "epoch": 0.33594850130758397, "grad_norm": 3.9887060748088254, "learning_rate": 4.202697853177951e-06, "loss": 0.4613, "step": 835 }, { "epoch": 0.3363508348420841, "grad_norm": 3.5896338953258473, "learning_rate": 4.2001254273255646e-06, "loss": 0.4031, "step": 836 }, { "epoch": 0.3367531683765842, "grad_norm": 4.094188015061212, "learning_rate": 4.197549648356478e-06, "loss": 0.5933, "step": 837 }, { "epoch": 0.3371555019110843, "grad_norm": 3.6605198066957216, "learning_rate": 4.19497052135084e-06, "loss": 0.417, "step": 838 }, { "epoch": 0.3375578354455844, "grad_norm": 5.538220960178179, "learning_rate": 4.192388051395398e-06, "loss": 0.7984, "step": 839 }, { "epoch": 0.33796016898008446, "grad_norm": 4.543985654558515, "learning_rate": 4.1898022435835e-06, "loss": 0.5526, "step": 840 }, { "epoch": 0.3383625025145846, "grad_norm": 3.719487495093868, "learning_rate": 4.187213103015069e-06, "loss": 0.4399, "step": 841 }, { "epoch": 0.3387648360490847, "grad_norm": 3.958292027050775, "learning_rate": 4.184620634796608e-06, "loss": 0.4891, "step": 842 }, { "epoch": 0.3391671695835848, "grad_norm": 4.219443448956219, "learning_rate": 4.182024844041177e-06, "loss": 0.6417, "step": 843 }, { "epoch": 0.33956950311808487, "grad_norm": 4.637518704467857, "learning_rate": 4.179425735868395e-06, "loss": 0.5912, "step": 844 }, { "epoch": 0.339971836652585, "grad_norm": 4.725084492132898, "learning_rate": 4.176823315404419e-06, "loss": 0.705, "step": 845 }, { "epoch": 0.3403741701870851, "grad_norm": 4.081067015807228, "learning_rate": 4.17421758778194e-06, "loss": 0.5006, "step": 846 }, { "epoch": 0.3407765037215852, "grad_norm": 4.431343567792434, "learning_rate": 4.1716085581401746e-06, "loss": 0.5081, "step": 847 }, { "epoch": 0.3411788372560853, "grad_norm": 4.173696495740008, "learning_rate": 4.1689962316248475e-06, "loss": 0.497, "step": 848 }, { "epoch": 0.3415811707905854, "grad_norm": 4.125681881515539, "learning_rate": 4.166380613388189e-06, "loss": 0.5498, "step": 849 }, { "epoch": 0.3419835043250855, "grad_norm": 4.333217037647524, "learning_rate": 4.163761708588919e-06, "loss": 0.5597, "step": 850 }, { "epoch": 0.3423858378595856, "grad_norm": 4.191848923537687, "learning_rate": 4.161139522392243e-06, "loss": 0.7259, "step": 851 }, { "epoch": 0.3427881713940857, "grad_norm": 3.1772920188403164, "learning_rate": 4.158514059969834e-06, "loss": 0.4225, "step": 852 }, { "epoch": 0.34319050492858577, "grad_norm": 4.166291911511059, "learning_rate": 4.15588532649983e-06, "loss": 0.6061, "step": 853 }, { "epoch": 0.3435928384630859, "grad_norm": 3.5991395036430447, "learning_rate": 4.1532533271668175e-06, "loss": 0.461, "step": 854 }, { "epoch": 0.343995171997586, "grad_norm": 4.29116994161306, "learning_rate": 4.150618067161828e-06, "loss": 0.567, "step": 855 }, { "epoch": 0.3443975055320861, "grad_norm": 4.403897532608296, "learning_rate": 4.14797955168232e-06, "loss": 0.6118, "step": 856 }, { "epoch": 0.3447998390665862, "grad_norm": 4.145689255622535, "learning_rate": 4.145337785932174e-06, "loss": 0.5709, "step": 857 }, { "epoch": 0.3452021726010863, "grad_norm": 4.540600097537597, "learning_rate": 4.142692775121684e-06, "loss": 0.5506, "step": 858 }, { "epoch": 0.3456045061355864, "grad_norm": 4.730037786980354, "learning_rate": 4.14004452446754e-06, "loss": 0.6088, "step": 859 }, { "epoch": 0.3460068396700865, "grad_norm": 3.9409862740254487, "learning_rate": 4.137393039192822e-06, "loss": 0.5775, "step": 860 }, { "epoch": 0.3464091732045866, "grad_norm": 4.113320460512886, "learning_rate": 4.1347383245269935e-06, "loss": 0.5695, "step": 861 }, { "epoch": 0.34681150673908673, "grad_norm": 4.6445735426137364, "learning_rate": 4.1320803857058835e-06, "loss": 0.6158, "step": 862 }, { "epoch": 0.3472138402735868, "grad_norm": 4.634455536740089, "learning_rate": 4.129419227971681e-06, "loss": 0.6799, "step": 863 }, { "epoch": 0.3476161738080869, "grad_norm": 4.767706536948578, "learning_rate": 4.1267548565729235e-06, "loss": 0.5473, "step": 864 }, { "epoch": 0.348018507342587, "grad_norm": 4.4478746327105405, "learning_rate": 4.124087276764488e-06, "loss": 0.6564, "step": 865 }, { "epoch": 0.3484208408770871, "grad_norm": 5.1282545461378195, "learning_rate": 4.121416493807577e-06, "loss": 0.617, "step": 866 }, { "epoch": 0.3488231744115872, "grad_norm": 3.850370414902996, "learning_rate": 4.118742512969713e-06, "loss": 0.5775, "step": 867 }, { "epoch": 0.3492255079460873, "grad_norm": 4.66805214245803, "learning_rate": 4.116065339524724e-06, "loss": 0.6724, "step": 868 }, { "epoch": 0.3496278414805874, "grad_norm": 4.7651508619199765, "learning_rate": 4.113384978752734e-06, "loss": 0.6151, "step": 869 }, { "epoch": 0.3500301750150875, "grad_norm": 4.207593613220341, "learning_rate": 4.110701435940157e-06, "loss": 0.641, "step": 870 }, { "epoch": 0.35043250854958763, "grad_norm": 4.3034195107515965, "learning_rate": 4.108014716379679e-06, "loss": 0.6542, "step": 871 }, { "epoch": 0.3508348420840877, "grad_norm": 3.987719667248183, "learning_rate": 4.105324825370251e-06, "loss": 0.4996, "step": 872 }, { "epoch": 0.3512371756185878, "grad_norm": 3.9071320373239824, "learning_rate": 4.102631768217083e-06, "loss": 0.5168, "step": 873 }, { "epoch": 0.3516395091530879, "grad_norm": 4.2277706570506295, "learning_rate": 4.099935550231626e-06, "loss": 0.5379, "step": 874 }, { "epoch": 0.35204184268758804, "grad_norm": 3.9154962483786373, "learning_rate": 4.097236176731567e-06, "loss": 0.5523, "step": 875 }, { "epoch": 0.3524441762220881, "grad_norm": 3.465039425796171, "learning_rate": 4.094533653040814e-06, "loss": 0.4001, "step": 876 }, { "epoch": 0.3528465097565882, "grad_norm": 3.4706674601181167, "learning_rate": 4.091827984489493e-06, "loss": 0.4629, "step": 877 }, { "epoch": 0.3532488432910883, "grad_norm": 4.340590023749821, "learning_rate": 4.089119176413926e-06, "loss": 0.6815, "step": 878 }, { "epoch": 0.3536511768255884, "grad_norm": 4.471466388864396, "learning_rate": 4.086407234156633e-06, "loss": 0.6459, "step": 879 }, { "epoch": 0.35405351036008853, "grad_norm": 4.449019271914241, "learning_rate": 4.08369216306631e-06, "loss": 0.5592, "step": 880 }, { "epoch": 0.3544558438945886, "grad_norm": 4.125369265434845, "learning_rate": 4.080973968497829e-06, "loss": 0.5138, "step": 881 }, { "epoch": 0.3548581774290887, "grad_norm": 4.523305588736027, "learning_rate": 4.07825265581222e-06, "loss": 0.5929, "step": 882 }, { "epoch": 0.3552605109635888, "grad_norm": 4.077469504247697, "learning_rate": 4.075528230376662e-06, "loss": 0.5288, "step": 883 }, { "epoch": 0.35566284449808894, "grad_norm": 5.092003038039258, "learning_rate": 4.072800697564474e-06, "loss": 0.6624, "step": 884 }, { "epoch": 0.356065178032589, "grad_norm": 3.7756634610021935, "learning_rate": 4.070070062755105e-06, "loss": 0.5265, "step": 885 }, { "epoch": 0.3564675115670891, "grad_norm": 4.106660405022582, "learning_rate": 4.0673363313341195e-06, "loss": 0.4853, "step": 886 }, { "epoch": 0.3568698451015892, "grad_norm": 4.215215996153993, "learning_rate": 4.064599508693191e-06, "loss": 0.5721, "step": 887 }, { "epoch": 0.35727217863608934, "grad_norm": 4.171756065223583, "learning_rate": 4.06185960023009e-06, "loss": 0.4649, "step": 888 }, { "epoch": 0.35767451217058943, "grad_norm": 5.138546885500691, "learning_rate": 4.05911661134867e-06, "loss": 0.5972, "step": 889 }, { "epoch": 0.3580768457050895, "grad_norm": 4.136963664338033, "learning_rate": 4.0563705474588655e-06, "loss": 0.5027, "step": 890 }, { "epoch": 0.3584791792395896, "grad_norm": 3.940446094705443, "learning_rate": 4.05362141397667e-06, "loss": 0.5627, "step": 891 }, { "epoch": 0.3588815127740897, "grad_norm": 4.024692229149252, "learning_rate": 4.0508692163241356e-06, "loss": 0.5521, "step": 892 }, { "epoch": 0.35928384630858984, "grad_norm": 3.866043251441755, "learning_rate": 4.048113959929354e-06, "loss": 0.5125, "step": 893 }, { "epoch": 0.3596861798430899, "grad_norm": 4.786779622248551, "learning_rate": 4.0453556502264535e-06, "loss": 0.6475, "step": 894 }, { "epoch": 0.36008851337759, "grad_norm": 4.127247200504385, "learning_rate": 4.042594292655581e-06, "loss": 0.4877, "step": 895 }, { "epoch": 0.3604908469120901, "grad_norm": 3.3408217784719634, "learning_rate": 4.039829892662897e-06, "loss": 0.4337, "step": 896 }, { "epoch": 0.36089318044659024, "grad_norm": 4.010232394860985, "learning_rate": 4.037062455700559e-06, "loss": 0.493, "step": 897 }, { "epoch": 0.36129551398109033, "grad_norm": 4.3695516591433785, "learning_rate": 4.03429198722672e-06, "loss": 0.6051, "step": 898 }, { "epoch": 0.3616978475155904, "grad_norm": 4.766410636581824, "learning_rate": 4.0315184927055065e-06, "loss": 0.6602, "step": 899 }, { "epoch": 0.3621001810500905, "grad_norm": 4.1149436048907955, "learning_rate": 4.028741977607016e-06, "loss": 0.599, "step": 900 }, { "epoch": 0.36250251458459065, "grad_norm": 3.4871089312720045, "learning_rate": 4.0259624474073025e-06, "loss": 0.4779, "step": 901 }, { "epoch": 0.36290484811909074, "grad_norm": 3.6367752409912266, "learning_rate": 4.023179907588367e-06, "loss": 0.4541, "step": 902 }, { "epoch": 0.3633071816535908, "grad_norm": 3.3961433962608685, "learning_rate": 4.020394363638147e-06, "loss": 0.4265, "step": 903 }, { "epoch": 0.3637095151880909, "grad_norm": 3.8352646417492524, "learning_rate": 4.0176058210505045e-06, "loss": 0.5351, "step": 904 }, { "epoch": 0.364111848722591, "grad_norm": 4.242220553627046, "learning_rate": 4.0148142853252125e-06, "loss": 0.5792, "step": 905 }, { "epoch": 0.36451418225709115, "grad_norm": 4.096283695840529, "learning_rate": 4.0120197619679536e-06, "loss": 0.5412, "step": 906 }, { "epoch": 0.36491651579159123, "grad_norm": 4.729251903541468, "learning_rate": 4.009222256490297e-06, "loss": 0.6285, "step": 907 }, { "epoch": 0.3653188493260913, "grad_norm": 4.084180186714877, "learning_rate": 4.006421774409697e-06, "loss": 0.6304, "step": 908 }, { "epoch": 0.3657211828605914, "grad_norm": 4.280731987602122, "learning_rate": 4.003618321249476e-06, "loss": 0.5632, "step": 909 }, { "epoch": 0.36612351639509155, "grad_norm": 3.868764716611033, "learning_rate": 4.000811902538821e-06, "loss": 0.4715, "step": 910 }, { "epoch": 0.36652584992959164, "grad_norm": 3.6445778686982155, "learning_rate": 3.99800252381276e-06, "loss": 0.4084, "step": 911 }, { "epoch": 0.36692818346409173, "grad_norm": 4.600601862430362, "learning_rate": 3.995190190612165e-06, "loss": 0.7218, "step": 912 }, { "epoch": 0.3673305169985918, "grad_norm": 4.09418361850337, "learning_rate": 3.992374908483735e-06, "loss": 0.5913, "step": 913 }, { "epoch": 0.36773285053309196, "grad_norm": 4.495329343052196, "learning_rate": 3.9895566829799825e-06, "loss": 0.6735, "step": 914 }, { "epoch": 0.36813518406759205, "grad_norm": 4.408452152812246, "learning_rate": 3.986735519659226e-06, "loss": 0.6896, "step": 915 }, { "epoch": 0.36853751760209214, "grad_norm": 3.7084757585295276, "learning_rate": 3.983911424085578e-06, "loss": 0.552, "step": 916 }, { "epoch": 0.3689398511365922, "grad_norm": 3.8368900456075985, "learning_rate": 3.981084401828937e-06, "loss": 0.4944, "step": 917 }, { "epoch": 0.3693421846710923, "grad_norm": 5.696133375208026, "learning_rate": 3.978254458464969e-06, "loss": 0.6338, "step": 918 }, { "epoch": 0.36974451820559245, "grad_norm": 4.144365741395642, "learning_rate": 3.975421599575103e-06, "loss": 0.6463, "step": 919 }, { "epoch": 0.37014685174009254, "grad_norm": 3.7665583063095682, "learning_rate": 3.972585830746522e-06, "loss": 0.4442, "step": 920 }, { "epoch": 0.37054918527459263, "grad_norm": 4.341604001350467, "learning_rate": 3.969747157572142e-06, "loss": 0.7191, "step": 921 }, { "epoch": 0.3709515188090927, "grad_norm": 4.724527719836307, "learning_rate": 3.966905585650611e-06, "loss": 0.7508, "step": 922 }, { "epoch": 0.37135385234359286, "grad_norm": 6.112644943798556, "learning_rate": 3.964061120586294e-06, "loss": 0.6744, "step": 923 }, { "epoch": 0.37175618587809295, "grad_norm": 4.630742134072846, "learning_rate": 3.961213767989261e-06, "loss": 0.5703, "step": 924 }, { "epoch": 0.37215851941259304, "grad_norm": 3.5499854666112802, "learning_rate": 3.958363533475277e-06, "loss": 0.4563, "step": 925 }, { "epoch": 0.3725608529470931, "grad_norm": 3.926859330076321, "learning_rate": 3.955510422665791e-06, "loss": 0.5159, "step": 926 }, { "epoch": 0.37296318648159327, "grad_norm": 4.029967110899967, "learning_rate": 3.952654441187927e-06, "loss": 0.5267, "step": 927 }, { "epoch": 0.37336552001609336, "grad_norm": 3.4483709639926414, "learning_rate": 3.9497955946744675e-06, "loss": 0.4584, "step": 928 }, { "epoch": 0.37376785355059344, "grad_norm": 4.752285945278346, "learning_rate": 3.946933888763847e-06, "loss": 0.5743, "step": 929 }, { "epoch": 0.37417018708509353, "grad_norm": 3.985801543832948, "learning_rate": 3.9440693291001406e-06, "loss": 0.5319, "step": 930 }, { "epoch": 0.3745725206195936, "grad_norm": 3.5946794819227903, "learning_rate": 3.941201921333048e-06, "loss": 0.4217, "step": 931 }, { "epoch": 0.37497485415409376, "grad_norm": 4.035357970865375, "learning_rate": 3.9383316711178946e-06, "loss": 0.503, "step": 932 }, { "epoch": 0.37537718768859385, "grad_norm": 4.034839036975055, "learning_rate": 3.935458584115599e-06, "loss": 0.4698, "step": 933 }, { "epoch": 0.37577952122309394, "grad_norm": 4.286963186422666, "learning_rate": 3.932582665992688e-06, "loss": 0.5901, "step": 934 }, { "epoch": 0.376181854757594, "grad_norm": 4.021909925329209, "learning_rate": 3.929703922421263e-06, "loss": 0.5111, "step": 935 }, { "epoch": 0.37658418829209417, "grad_norm": 4.0942098752342035, "learning_rate": 3.926822359079001e-06, "loss": 0.5472, "step": 936 }, { "epoch": 0.37698652182659426, "grad_norm": 4.261504012145757, "learning_rate": 3.923937981649143e-06, "loss": 0.6453, "step": 937 }, { "epoch": 0.37738885536109434, "grad_norm": 4.014117574115026, "learning_rate": 3.921050795820474e-06, "loss": 0.5056, "step": 938 }, { "epoch": 0.37779118889559443, "grad_norm": 3.904726379027022, "learning_rate": 3.9181608072873244e-06, "loss": 0.5304, "step": 939 }, { "epoch": 0.3781935224300946, "grad_norm": 3.881019901522325, "learning_rate": 3.9152680217495475e-06, "loss": 0.5298, "step": 940 }, { "epoch": 0.37859585596459466, "grad_norm": 4.058485666284913, "learning_rate": 3.912372444912517e-06, "loss": 0.574, "step": 941 }, { "epoch": 0.37899818949909475, "grad_norm": 3.9630383581476205, "learning_rate": 3.909474082487107e-06, "loss": 0.5804, "step": 942 }, { "epoch": 0.37940052303359484, "grad_norm": 3.8676063595588683, "learning_rate": 3.906572940189691e-06, "loss": 0.4804, "step": 943 }, { "epoch": 0.3798028565680949, "grad_norm": 4.32834774261345, "learning_rate": 3.9036690237421215e-06, "loss": 0.6161, "step": 944 }, { "epoch": 0.38020519010259507, "grad_norm": 4.726002477400136, "learning_rate": 3.900762338871723e-06, "loss": 0.6898, "step": 945 }, { "epoch": 0.38060752363709516, "grad_norm": 4.577558255257823, "learning_rate": 3.897852891311282e-06, "loss": 0.5504, "step": 946 }, { "epoch": 0.38100985717159525, "grad_norm": 4.300878005170684, "learning_rate": 3.89494068679903e-06, "loss": 0.6734, "step": 947 }, { "epoch": 0.38141219070609533, "grad_norm": 4.519778162808581, "learning_rate": 3.892025731078641e-06, "loss": 0.5694, "step": 948 }, { "epoch": 0.3818145242405955, "grad_norm": 3.8176275411987683, "learning_rate": 3.88910802989921e-06, "loss": 0.4444, "step": 949 }, { "epoch": 0.38221685777509556, "grad_norm": 4.014909553832162, "learning_rate": 3.886187589015251e-06, "loss": 0.6054, "step": 950 }, { "epoch": 0.38261919130959565, "grad_norm": 4.267781377287406, "learning_rate": 3.883264414186677e-06, "loss": 0.6257, "step": 951 }, { "epoch": 0.38302152484409574, "grad_norm": 4.345044782459552, "learning_rate": 3.8803385111788e-06, "loss": 0.5543, "step": 952 }, { "epoch": 0.3834238583785959, "grad_norm": 4.221128544535545, "learning_rate": 3.877409885762305e-06, "loss": 0.5544, "step": 953 }, { "epoch": 0.38382619191309597, "grad_norm": 4.000060887768021, "learning_rate": 3.8744785437132506e-06, "loss": 0.5878, "step": 954 }, { "epoch": 0.38422852544759606, "grad_norm": 3.8372993108948488, "learning_rate": 3.871544490813054e-06, "loss": 0.5076, "step": 955 }, { "epoch": 0.38463085898209615, "grad_norm": 4.059706126497868, "learning_rate": 3.868607732848475e-06, "loss": 0.5353, "step": 956 }, { "epoch": 0.38503319251659623, "grad_norm": 4.408924474844376, "learning_rate": 3.865668275611614e-06, "loss": 0.519, "step": 957 }, { "epoch": 0.3854355260510964, "grad_norm": 4.010089079135815, "learning_rate": 3.862726124899889e-06, "loss": 0.5952, "step": 958 }, { "epoch": 0.38583785958559647, "grad_norm": 5.459071377748317, "learning_rate": 3.859781286516037e-06, "loss": 0.7175, "step": 959 }, { "epoch": 0.38624019312009655, "grad_norm": 4.542741280511702, "learning_rate": 3.85683376626809e-06, "loss": 0.5667, "step": 960 }, { "epoch": 0.38664252665459664, "grad_norm": 3.99184341470565, "learning_rate": 3.853883569969372e-06, "loss": 0.6148, "step": 961 }, { "epoch": 0.3870448601890968, "grad_norm": 4.078873510656604, "learning_rate": 3.850930703438484e-06, "loss": 0.5383, "step": 962 }, { "epoch": 0.38744719372359687, "grad_norm": 3.896065226196885, "learning_rate": 3.847975172499295e-06, "loss": 0.6785, "step": 963 }, { "epoch": 0.38784952725809696, "grad_norm": 4.328062559714918, "learning_rate": 3.845016982980929e-06, "loss": 0.5543, "step": 964 }, { "epoch": 0.38825186079259705, "grad_norm": 4.2981177827507935, "learning_rate": 3.842056140717749e-06, "loss": 0.6935, "step": 965 }, { "epoch": 0.3886541943270972, "grad_norm": 4.489409385825919, "learning_rate": 3.839092651549357e-06, "loss": 0.6412, "step": 966 }, { "epoch": 0.3890565278615973, "grad_norm": 4.162311148181462, "learning_rate": 3.836126521320569e-06, "loss": 0.5133, "step": 967 }, { "epoch": 0.38945886139609737, "grad_norm": 3.7564756341509464, "learning_rate": 3.833157755881414e-06, "loss": 0.5198, "step": 968 }, { "epoch": 0.38986119493059745, "grad_norm": 4.2353823981014385, "learning_rate": 3.830186361087117e-06, "loss": 0.4271, "step": 969 }, { "epoch": 0.39026352846509754, "grad_norm": 3.4858668216855437, "learning_rate": 3.827212342798089e-06, "loss": 0.4654, "step": 970 }, { "epoch": 0.3906658619995977, "grad_norm": 3.920259549785761, "learning_rate": 3.824235706879915e-06, "loss": 0.5822, "step": 971 }, { "epoch": 0.3910681955340978, "grad_norm": 4.239371514940848, "learning_rate": 3.821256459203343e-06, "loss": 0.5093, "step": 972 }, { "epoch": 0.39147052906859786, "grad_norm": 4.037277711642943, "learning_rate": 3.81827460564427e-06, "loss": 0.5222, "step": 973 }, { "epoch": 0.39187286260309795, "grad_norm": 4.456827976048942, "learning_rate": 3.815290152083737e-06, "loss": 0.6783, "step": 974 }, { "epoch": 0.3922751961375981, "grad_norm": 4.246220187755789, "learning_rate": 3.8123031044079084e-06, "loss": 0.5888, "step": 975 }, { "epoch": 0.3926775296720982, "grad_norm": 4.3203808668287635, "learning_rate": 3.809313468508068e-06, "loss": 0.5569, "step": 976 }, { "epoch": 0.39307986320659827, "grad_norm": 4.860143257226567, "learning_rate": 3.8063212502806035e-06, "loss": 0.5953, "step": 977 }, { "epoch": 0.39348219674109836, "grad_norm": 4.233378915892834, "learning_rate": 3.8033264556269954e-06, "loss": 0.6167, "step": 978 }, { "epoch": 0.39388453027559844, "grad_norm": 4.3534789683754305, "learning_rate": 3.8003290904538035e-06, "loss": 0.6133, "step": 979 }, { "epoch": 0.3942868638100986, "grad_norm": 3.9519882332598257, "learning_rate": 3.7973291606726614e-06, "loss": 0.4515, "step": 980 }, { "epoch": 0.3946891973445987, "grad_norm": 4.138408229253922, "learning_rate": 3.794326672200258e-06, "loss": 0.5607, "step": 981 }, { "epoch": 0.39509153087909876, "grad_norm": 4.3761985221008635, "learning_rate": 3.79132163095833e-06, "loss": 0.5507, "step": 982 }, { "epoch": 0.39549386441359885, "grad_norm": 4.023180334056979, "learning_rate": 3.7883140428736477e-06, "loss": 0.5314, "step": 983 }, { "epoch": 0.395896197948099, "grad_norm": 4.151232898341051, "learning_rate": 3.7853039138780057e-06, "loss": 0.5392, "step": 984 }, { "epoch": 0.3962985314825991, "grad_norm": 4.539638233125882, "learning_rate": 3.7822912499082087e-06, "loss": 0.5791, "step": 985 }, { "epoch": 0.39670086501709917, "grad_norm": 4.335330132423181, "learning_rate": 3.7792760569060626e-06, "loss": 0.5178, "step": 986 }, { "epoch": 0.39710319855159926, "grad_norm": 4.143033413053681, "learning_rate": 3.7762583408183594e-06, "loss": 0.5815, "step": 987 }, { "epoch": 0.3975055320860994, "grad_norm": 4.915299915525361, "learning_rate": 3.7732381075968694e-06, "loss": 0.6269, "step": 988 }, { "epoch": 0.3979078656205995, "grad_norm": 3.7212476195736506, "learning_rate": 3.770215363198325e-06, "loss": 0.4599, "step": 989 }, { "epoch": 0.3983101991550996, "grad_norm": 3.9977745943956884, "learning_rate": 3.7671901135844148e-06, "loss": 0.5973, "step": 990 }, { "epoch": 0.39871253268959966, "grad_norm": 4.3088468446821615, "learning_rate": 3.764162364721764e-06, "loss": 0.5963, "step": 991 }, { "epoch": 0.39911486622409975, "grad_norm": 4.2991674025596955, "learning_rate": 3.761132122581931e-06, "loss": 0.6923, "step": 992 }, { "epoch": 0.3995171997585999, "grad_norm": 3.952746214125807, "learning_rate": 3.758099393141388e-06, "loss": 0.5384, "step": 993 }, { "epoch": 0.3999195332931, "grad_norm": 4.327315800519712, "learning_rate": 3.755064182381516e-06, "loss": 0.5144, "step": 994 }, { "epoch": 0.40032186682760007, "grad_norm": 4.713412610694241, "learning_rate": 3.752026496288588e-06, "loss": 0.641, "step": 995 }, { "epoch": 0.40072420036210016, "grad_norm": 3.9912240405323507, "learning_rate": 3.74898634085376e-06, "loss": 0.5646, "step": 996 }, { "epoch": 0.4011265338966003, "grad_norm": 4.530973641905437, "learning_rate": 3.7459437220730583e-06, "loss": 0.5614, "step": 997 }, { "epoch": 0.4015288674311004, "grad_norm": 6.864543852090396, "learning_rate": 3.742898645947366e-06, "loss": 0.5336, "step": 998 }, { "epoch": 0.4019312009656005, "grad_norm": 4.429956976671279, "learning_rate": 3.7398511184824144e-06, "loss": 0.7004, "step": 999 }, { "epoch": 0.40233353450010056, "grad_norm": 3.164463326468864, "learning_rate": 3.7368011456887693e-06, "loss": 0.4163, "step": 1000 }, { "epoch": 0.4027358680346007, "grad_norm": 3.7984416296897456, "learning_rate": 3.7337487335818185e-06, "loss": 0.5342, "step": 1001 }, { "epoch": 0.4031382015691008, "grad_norm": 4.186189188641626, "learning_rate": 3.730693888181761e-06, "loss": 0.6125, "step": 1002 }, { "epoch": 0.4035405351036009, "grad_norm": 3.8782636909404506, "learning_rate": 3.7276366155135968e-06, "loss": 0.5669, "step": 1003 }, { "epoch": 0.40394286863810097, "grad_norm": 4.222464351010827, "learning_rate": 3.7245769216071104e-06, "loss": 0.5574, "step": 1004 }, { "epoch": 0.40434520217260106, "grad_norm": 4.0429197887930135, "learning_rate": 3.721514812496863e-06, "loss": 0.5031, "step": 1005 }, { "epoch": 0.4047475357071012, "grad_norm": 4.015964628315664, "learning_rate": 3.718450294222179e-06, "loss": 0.5994, "step": 1006 }, { "epoch": 0.4051498692416013, "grad_norm": 3.846521589103569, "learning_rate": 3.7153833728271356e-06, "loss": 0.4928, "step": 1007 }, { "epoch": 0.4055522027761014, "grad_norm": 4.138445980201307, "learning_rate": 3.712314054360547e-06, "loss": 0.5722, "step": 1008 }, { "epoch": 0.40595453631060147, "grad_norm": 4.476195845673172, "learning_rate": 3.7092423448759577e-06, "loss": 0.563, "step": 1009 }, { "epoch": 0.4063568698451016, "grad_norm": 3.786638148470317, "learning_rate": 3.7061682504316264e-06, "loss": 0.5181, "step": 1010 }, { "epoch": 0.4067592033796017, "grad_norm": 4.563364096910093, "learning_rate": 3.7030917770905155e-06, "loss": 0.5247, "step": 1011 }, { "epoch": 0.4071615369141018, "grad_norm": 4.276417493626447, "learning_rate": 3.7000129309202804e-06, "loss": 0.5815, "step": 1012 }, { "epoch": 0.4075638704486019, "grad_norm": 4.495474983317033, "learning_rate": 3.696931717993256e-06, "loss": 0.7269, "step": 1013 }, { "epoch": 0.407966203983102, "grad_norm": 3.6238415974427527, "learning_rate": 3.693848144386444e-06, "loss": 0.5462, "step": 1014 }, { "epoch": 0.4083685375176021, "grad_norm": 4.262996632479673, "learning_rate": 3.6907622161815027e-06, "loss": 0.64, "step": 1015 }, { "epoch": 0.4087708710521022, "grad_norm": 3.8744140465414745, "learning_rate": 3.687673939464736e-06, "loss": 0.5281, "step": 1016 }, { "epoch": 0.4091732045866023, "grad_norm": 4.021193749407921, "learning_rate": 3.6845833203270766e-06, "loss": 0.5238, "step": 1017 }, { "epoch": 0.40957553812110237, "grad_norm": 3.7942462225606084, "learning_rate": 3.6814903648640787e-06, "loss": 0.5274, "step": 1018 }, { "epoch": 0.4099778716556025, "grad_norm": 3.8584763272892193, "learning_rate": 3.6783950791759053e-06, "loss": 0.6652, "step": 1019 }, { "epoch": 0.4103802051901026, "grad_norm": 4.224914273954211, "learning_rate": 3.6752974693673144e-06, "loss": 0.6418, "step": 1020 }, { "epoch": 0.4107825387246027, "grad_norm": 3.611777243329413, "learning_rate": 3.6721975415476475e-06, "loss": 0.4478, "step": 1021 }, { "epoch": 0.4111848722591028, "grad_norm": 4.492308490256186, "learning_rate": 3.669095301830816e-06, "loss": 0.6383, "step": 1022 }, { "epoch": 0.4115872057936029, "grad_norm": 3.966328790708119, "learning_rate": 3.6659907563352964e-06, "loss": 0.5496, "step": 1023 }, { "epoch": 0.411989539328103, "grad_norm": 4.331090241917638, "learning_rate": 3.6628839111841067e-06, "loss": 0.6138, "step": 1024 }, { "epoch": 0.4123918728626031, "grad_norm": 4.039380276956274, "learning_rate": 3.6597747725048054e-06, "loss": 0.514, "step": 1025 }, { "epoch": 0.4127942063971032, "grad_norm": 4.2678553912438515, "learning_rate": 3.656663346429471e-06, "loss": 0.5304, "step": 1026 }, { "epoch": 0.4131965399316033, "grad_norm": 3.6311274537700444, "learning_rate": 3.6535496390946958e-06, "loss": 0.5324, "step": 1027 }, { "epoch": 0.4135988734661034, "grad_norm": 4.179608207864191, "learning_rate": 3.650433656641569e-06, "loss": 0.5913, "step": 1028 }, { "epoch": 0.4140012070006035, "grad_norm": 3.8042489733382308, "learning_rate": 3.647315405215668e-06, "loss": 0.5007, "step": 1029 }, { "epoch": 0.4144035405351036, "grad_norm": 3.4526666984193404, "learning_rate": 3.644194890967047e-06, "loss": 0.4141, "step": 1030 }, { "epoch": 0.4148058740696037, "grad_norm": 4.657500083163389, "learning_rate": 3.641072120050221e-06, "loss": 0.6175, "step": 1031 }, { "epoch": 0.4152082076041038, "grad_norm": 3.5394692246958934, "learning_rate": 3.637947098624156e-06, "loss": 0.493, "step": 1032 }, { "epoch": 0.4156105411386039, "grad_norm": 3.6382852897066003, "learning_rate": 3.6348198328522565e-06, "loss": 0.4733, "step": 1033 }, { "epoch": 0.416012874673104, "grad_norm": 4.701593231983809, "learning_rate": 3.631690328902355e-06, "loss": 0.6759, "step": 1034 }, { "epoch": 0.4164152082076041, "grad_norm": 4.126519542614592, "learning_rate": 3.6285585929466982e-06, "loss": 0.6065, "step": 1035 }, { "epoch": 0.4168175417421042, "grad_norm": 4.174303064525279, "learning_rate": 3.6254246311619317e-06, "loss": 0.6019, "step": 1036 }, { "epoch": 0.4172198752766043, "grad_norm": 4.369071505943041, "learning_rate": 3.6222884497290937e-06, "loss": 0.47, "step": 1037 }, { "epoch": 0.4176222088111044, "grad_norm": 4.547283701461554, "learning_rate": 3.619150054833601e-06, "loss": 0.5666, "step": 1038 }, { "epoch": 0.4180245423456045, "grad_norm": 6.593645962994248, "learning_rate": 3.616009452665234e-06, "loss": 0.4854, "step": 1039 }, { "epoch": 0.41842687588010463, "grad_norm": 4.3350300541777145, "learning_rate": 3.6128666494181274e-06, "loss": 0.6522, "step": 1040 }, { "epoch": 0.4188292094146047, "grad_norm": 3.4916298319626233, "learning_rate": 3.6097216512907574e-06, "loss": 0.5449, "step": 1041 }, { "epoch": 0.4192315429491048, "grad_norm": 3.721480691979447, "learning_rate": 3.6065744644859276e-06, "loss": 0.4152, "step": 1042 }, { "epoch": 0.4196338764836049, "grad_norm": 4.419075622535455, "learning_rate": 3.6034250952107598e-06, "loss": 0.5998, "step": 1043 }, { "epoch": 0.420036210018105, "grad_norm": 4.066379360835346, "learning_rate": 3.6002735496766787e-06, "loss": 0.4443, "step": 1044 }, { "epoch": 0.4204385435526051, "grad_norm": 5.277931637633872, "learning_rate": 3.5971198340994035e-06, "loss": 0.6578, "step": 1045 }, { "epoch": 0.4208408770871052, "grad_norm": 4.171317951485179, "learning_rate": 3.5939639546989315e-06, "loss": 0.5362, "step": 1046 }, { "epoch": 0.4212432106216053, "grad_norm": 3.6581061080480786, "learning_rate": 3.5908059176995274e-06, "loss": 0.4868, "step": 1047 }, { "epoch": 0.4216455441561054, "grad_norm": 4.089304060454962, "learning_rate": 3.587645729329713e-06, "loss": 0.4962, "step": 1048 }, { "epoch": 0.42204787769060553, "grad_norm": 4.193929975116655, "learning_rate": 3.584483395822252e-06, "loss": 0.585, "step": 1049 }, { "epoch": 0.4224502112251056, "grad_norm": 4.4833727752476324, "learning_rate": 3.5813189234141386e-06, "loss": 0.5295, "step": 1050 }, { "epoch": 0.4228525447596057, "grad_norm": 4.131740789977084, "learning_rate": 3.5781523183465868e-06, "loss": 0.5408, "step": 1051 }, { "epoch": 0.4232548782941058, "grad_norm": 3.972093255632944, "learning_rate": 3.574983586865015e-06, "loss": 0.5256, "step": 1052 }, { "epoch": 0.42365721182860594, "grad_norm": 4.619320510735413, "learning_rate": 3.571812735219037e-06, "loss": 0.6365, "step": 1053 }, { "epoch": 0.424059545363106, "grad_norm": 4.543758819054478, "learning_rate": 3.5686397696624486e-06, "loss": 0.5358, "step": 1054 }, { "epoch": 0.4244618788976061, "grad_norm": 3.9488894694458305, "learning_rate": 3.565464696453212e-06, "loss": 0.5101, "step": 1055 }, { "epoch": 0.4248642124321062, "grad_norm": 3.9008766737779217, "learning_rate": 3.5622875218534493e-06, "loss": 0.6157, "step": 1056 }, { "epoch": 0.4252665459666063, "grad_norm": 4.072195384389267, "learning_rate": 3.5591082521294264e-06, "loss": 0.572, "step": 1057 }, { "epoch": 0.42566887950110643, "grad_norm": 3.7407260188563622, "learning_rate": 3.555926893551539e-06, "loss": 0.4571, "step": 1058 }, { "epoch": 0.4260712130356065, "grad_norm": 4.030808733877317, "learning_rate": 3.552743452394306e-06, "loss": 0.5615, "step": 1059 }, { "epoch": 0.4264735465701066, "grad_norm": 3.833659387273303, "learning_rate": 3.5495579349363517e-06, "loss": 0.4978, "step": 1060 }, { "epoch": 0.4268758801046067, "grad_norm": 4.304988237070821, "learning_rate": 3.5463703474603964e-06, "loss": 0.6441, "step": 1061 }, { "epoch": 0.42727821363910684, "grad_norm": 4.68914366889499, "learning_rate": 3.5431806962532426e-06, "loss": 0.5009, "step": 1062 }, { "epoch": 0.42768054717360693, "grad_norm": 4.208417431491987, "learning_rate": 3.539988987605763e-06, "loss": 0.5556, "step": 1063 }, { "epoch": 0.428082880708107, "grad_norm": 3.913021062753838, "learning_rate": 3.5367952278128877e-06, "loss": 0.3997, "step": 1064 }, { "epoch": 0.4284852142426071, "grad_norm": 3.975035687293758, "learning_rate": 3.5335994231735927e-06, "loss": 0.5394, "step": 1065 }, { "epoch": 0.42888754777710725, "grad_norm": 3.969954958131227, "learning_rate": 3.5304015799908875e-06, "loss": 0.579, "step": 1066 }, { "epoch": 0.42928988131160734, "grad_norm": 4.942462188783348, "learning_rate": 3.5272017045718016e-06, "loss": 0.6723, "step": 1067 }, { "epoch": 0.4296922148461074, "grad_norm": 4.354492013681113, "learning_rate": 3.5239998032273727e-06, "loss": 0.5724, "step": 1068 }, { "epoch": 0.4300945483806075, "grad_norm": 4.275789917585218, "learning_rate": 3.520795882272634e-06, "loss": 0.5592, "step": 1069 }, { "epoch": 0.4304968819151076, "grad_norm": 4.0220453706013535, "learning_rate": 3.5175899480266023e-06, "loss": 0.6004, "step": 1070 }, { "epoch": 0.43089921544960774, "grad_norm": 3.956820759857449, "learning_rate": 3.5143820068122647e-06, "loss": 0.4667, "step": 1071 }, { "epoch": 0.43130154898410783, "grad_norm": 3.9359268522062996, "learning_rate": 3.5111720649565685e-06, "loss": 0.4766, "step": 1072 }, { "epoch": 0.4317038825186079, "grad_norm": 3.9665074015294555, "learning_rate": 3.507960128790402e-06, "loss": 0.5496, "step": 1073 }, { "epoch": 0.432106216053108, "grad_norm": 4.086635346156316, "learning_rate": 3.5047462046485934e-06, "loss": 0.4863, "step": 1074 }, { "epoch": 0.43250854958760815, "grad_norm": 4.234886651798155, "learning_rate": 3.501530298869886e-06, "loss": 0.5373, "step": 1075 }, { "epoch": 0.43291088312210824, "grad_norm": 4.578190826029045, "learning_rate": 3.4983124177969353e-06, "loss": 0.6952, "step": 1076 }, { "epoch": 0.4333132166566083, "grad_norm": 4.198352580671215, "learning_rate": 3.495092567776291e-06, "loss": 0.6753, "step": 1077 }, { "epoch": 0.4337155501911084, "grad_norm": 4.25314346947828, "learning_rate": 3.4918707551583853e-06, "loss": 0.6139, "step": 1078 }, { "epoch": 0.43411788372560856, "grad_norm": 3.7809774120139084, "learning_rate": 3.4886469862975224e-06, "loss": 0.5401, "step": 1079 }, { "epoch": 0.43452021726010864, "grad_norm": 3.861747118430867, "learning_rate": 3.485421267551865e-06, "loss": 0.4343, "step": 1080 }, { "epoch": 0.43492255079460873, "grad_norm": 4.1971689225974655, "learning_rate": 3.482193605283421e-06, "loss": 0.4651, "step": 1081 }, { "epoch": 0.4353248843291088, "grad_norm": 4.719733540576249, "learning_rate": 3.478964005858031e-06, "loss": 0.6265, "step": 1082 }, { "epoch": 0.4357272178636089, "grad_norm": 4.199121895188385, "learning_rate": 3.475732475645357e-06, "loss": 0.5915, "step": 1083 }, { "epoch": 0.43612955139810905, "grad_norm": 4.117901413415391, "learning_rate": 3.4724990210188693e-06, "loss": 0.5859, "step": 1084 }, { "epoch": 0.43653188493260914, "grad_norm": 4.063314519073365, "learning_rate": 3.4692636483558316e-06, "loss": 0.5799, "step": 1085 }, { "epoch": 0.4369342184671092, "grad_norm": 4.291259902841789, "learning_rate": 3.4660263640372926e-06, "loss": 0.5972, "step": 1086 }, { "epoch": 0.4373365520016093, "grad_norm": 4.317671430467203, "learning_rate": 3.462787174448071e-06, "loss": 0.5818, "step": 1087 }, { "epoch": 0.43773888553610946, "grad_norm": 4.041367856828536, "learning_rate": 3.459546085976743e-06, "loss": 0.5635, "step": 1088 }, { "epoch": 0.43814121907060954, "grad_norm": 5.295724094479603, "learning_rate": 3.45630310501563e-06, "loss": 0.6712, "step": 1089 }, { "epoch": 0.43854355260510963, "grad_norm": 4.415113705137869, "learning_rate": 3.453058237960785e-06, "loss": 0.578, "step": 1090 }, { "epoch": 0.4389458861396097, "grad_norm": 4.107191353105501, "learning_rate": 3.4498114912119817e-06, "loss": 0.5199, "step": 1091 }, { "epoch": 0.43934821967410986, "grad_norm": 4.1779462968909336, "learning_rate": 3.4465628711727022e-06, "loss": 0.577, "step": 1092 }, { "epoch": 0.43975055320860995, "grad_norm": 3.779499327289251, "learning_rate": 3.4433123842501205e-06, "loss": 0.4508, "step": 1093 }, { "epoch": 0.44015288674311004, "grad_norm": 3.9872787051249077, "learning_rate": 3.440060036855095e-06, "loss": 0.5583, "step": 1094 }, { "epoch": 0.4405552202776101, "grad_norm": 4.317372927534622, "learning_rate": 3.4368058354021526e-06, "loss": 0.5136, "step": 1095 }, { "epoch": 0.4409575538121102, "grad_norm": 3.7148972800031053, "learning_rate": 3.4335497863094773e-06, "loss": 0.514, "step": 1096 }, { "epoch": 0.44135988734661036, "grad_norm": 4.342233656787418, "learning_rate": 3.430291895998896e-06, "loss": 0.6742, "step": 1097 }, { "epoch": 0.44176222088111045, "grad_norm": 4.799719138290185, "learning_rate": 3.4270321708958686e-06, "loss": 0.6306, "step": 1098 }, { "epoch": 0.44216455441561053, "grad_norm": 3.9209860725209382, "learning_rate": 3.4237706174294726e-06, "loss": 0.5836, "step": 1099 }, { "epoch": 0.4425668879501106, "grad_norm": 4.127404724187905, "learning_rate": 3.420507242032392e-06, "loss": 0.5628, "step": 1100 }, { "epoch": 0.44296922148461076, "grad_norm": 4.956274683013535, "learning_rate": 3.4172420511409033e-06, "loss": 0.5366, "step": 1101 }, { "epoch": 0.44337155501911085, "grad_norm": 3.9527811079754285, "learning_rate": 3.413975051194865e-06, "loss": 0.4531, "step": 1102 }, { "epoch": 0.44377388855361094, "grad_norm": 4.364141896413706, "learning_rate": 3.4107062486377028e-06, "loss": 0.4914, "step": 1103 }, { "epoch": 0.44417622208811103, "grad_norm": 4.349108898092637, "learning_rate": 3.4074356499163976e-06, "loss": 0.7239, "step": 1104 }, { "epoch": 0.44457855562261117, "grad_norm": 4.038974541606429, "learning_rate": 3.4041632614814734e-06, "loss": 0.6375, "step": 1105 }, { "epoch": 0.44498088915711126, "grad_norm": 4.811612078810641, "learning_rate": 3.400889089786984e-06, "loss": 0.4963, "step": 1106 }, { "epoch": 0.44538322269161135, "grad_norm": 4.07473806896573, "learning_rate": 3.397613141290499e-06, "loss": 0.575, "step": 1107 }, { "epoch": 0.44578555622611143, "grad_norm": 4.160302353185681, "learning_rate": 3.3943354224530934e-06, "loss": 0.4766, "step": 1108 }, { "epoch": 0.4461878897606115, "grad_norm": 3.6900007830747934, "learning_rate": 3.3910559397393335e-06, "loss": 0.5544, "step": 1109 }, { "epoch": 0.44659022329511167, "grad_norm": 4.333978444807443, "learning_rate": 3.3877746996172645e-06, "loss": 0.5637, "step": 1110 }, { "epoch": 0.44699255682961175, "grad_norm": 3.768061403793603, "learning_rate": 3.384491708558399e-06, "loss": 0.4659, "step": 1111 }, { "epoch": 0.44739489036411184, "grad_norm": 3.9713683928936208, "learning_rate": 3.3812069730377007e-06, "loss": 0.539, "step": 1112 }, { "epoch": 0.44779722389861193, "grad_norm": 4.95129840549997, "learning_rate": 3.3779204995335747e-06, "loss": 0.6069, "step": 1113 }, { "epoch": 0.4481995574331121, "grad_norm": 4.316287155035261, "learning_rate": 3.3746322945278543e-06, "loss": 0.66, "step": 1114 }, { "epoch": 0.44860189096761216, "grad_norm": 4.697629623928124, "learning_rate": 3.3713423645057873e-06, "loss": 0.4885, "step": 1115 }, { "epoch": 0.44900422450211225, "grad_norm": 3.920242841924085, "learning_rate": 3.368050715956025e-06, "loss": 0.5213, "step": 1116 }, { "epoch": 0.44940655803661234, "grad_norm": 4.521566910606868, "learning_rate": 3.3647573553706052e-06, "loss": 0.5743, "step": 1117 }, { "epoch": 0.4498088915711125, "grad_norm": 4.082210637179168, "learning_rate": 3.3614622892449465e-06, "loss": 0.5361, "step": 1118 }, { "epoch": 0.45021122510561257, "grad_norm": 4.017855379880608, "learning_rate": 3.3581655240778277e-06, "loss": 0.5492, "step": 1119 }, { "epoch": 0.45061355864011265, "grad_norm": 3.3069156613524577, "learning_rate": 3.35486706637138e-06, "loss": 0.4142, "step": 1120 }, { "epoch": 0.45101589217461274, "grad_norm": 4.230766496789279, "learning_rate": 3.3515669226310716e-06, "loss": 0.5883, "step": 1121 }, { "epoch": 0.45141822570911283, "grad_norm": 3.932193721134766, "learning_rate": 3.3482650993656983e-06, "loss": 0.4732, "step": 1122 }, { "epoch": 0.451820559243613, "grad_norm": 3.9297564952360577, "learning_rate": 3.344961603087367e-06, "loss": 0.5852, "step": 1123 }, { "epoch": 0.45222289277811306, "grad_norm": 4.909401223605511, "learning_rate": 3.341656440311484e-06, "loss": 0.5709, "step": 1124 }, { "epoch": 0.45262522631261315, "grad_norm": 3.6889862040375525, "learning_rate": 3.3383496175567435e-06, "loss": 0.5966, "step": 1125 }, { "epoch": 0.45302755984711324, "grad_norm": 3.4699341807873694, "learning_rate": 3.3350411413451125e-06, "loss": 0.4692, "step": 1126 }, { "epoch": 0.4534298933816134, "grad_norm": 4.481843121210536, "learning_rate": 3.3317310182018186e-06, "loss": 0.6374, "step": 1127 }, { "epoch": 0.45383222691611347, "grad_norm": 4.674277827776204, "learning_rate": 3.328419254655339e-06, "loss": 0.7174, "step": 1128 }, { "epoch": 0.45423456045061356, "grad_norm": 3.8262305960434575, "learning_rate": 3.325105857237386e-06, "loss": 0.4618, "step": 1129 }, { "epoch": 0.45463689398511364, "grad_norm": 3.9899678016482305, "learning_rate": 3.3217908324828942e-06, "loss": 0.5839, "step": 1130 }, { "epoch": 0.4550392275196138, "grad_norm": 3.39198458807276, "learning_rate": 3.3184741869300073e-06, "loss": 0.5191, "step": 1131 }, { "epoch": 0.4554415610541139, "grad_norm": 4.032286065452944, "learning_rate": 3.3151559271200656e-06, "loss": 0.4835, "step": 1132 }, { "epoch": 0.45584389458861396, "grad_norm": 4.381591989574539, "learning_rate": 3.3118360595975947e-06, "loss": 0.5592, "step": 1133 }, { "epoch": 0.45624622812311405, "grad_norm": 4.022213508423109, "learning_rate": 3.3085145909102885e-06, "loss": 0.5304, "step": 1134 }, { "epoch": 0.45664856165761414, "grad_norm": 3.834367329162151, "learning_rate": 3.305191527609e-06, "loss": 0.4678, "step": 1135 }, { "epoch": 0.4570508951921143, "grad_norm": 4.448203020700801, "learning_rate": 3.3018668762477294e-06, "loss": 0.5172, "step": 1136 }, { "epoch": 0.45745322872661437, "grad_norm": 4.128965375687235, "learning_rate": 3.2985406433836055e-06, "loss": 0.4742, "step": 1137 }, { "epoch": 0.45785556226111446, "grad_norm": 4.129504251009841, "learning_rate": 3.295212835576878e-06, "loss": 0.5873, "step": 1138 }, { "epoch": 0.45825789579561454, "grad_norm": 4.422224105540513, "learning_rate": 3.2918834593909028e-06, "loss": 0.6049, "step": 1139 }, { "epoch": 0.4586602293301147, "grad_norm": 4.434937349809798, "learning_rate": 3.2885525213921286e-06, "loss": 0.5708, "step": 1140 }, { "epoch": 0.4590625628646148, "grad_norm": 3.668465429492197, "learning_rate": 3.2852200281500855e-06, "loss": 0.4559, "step": 1141 }, { "epoch": 0.45946489639911486, "grad_norm": 4.232532968956556, "learning_rate": 3.2818859862373696e-06, "loss": 0.5721, "step": 1142 }, { "epoch": 0.45986722993361495, "grad_norm": 4.7338258325238085, "learning_rate": 3.2785504022296316e-06, "loss": 0.6157, "step": 1143 }, { "epoch": 0.4602695634681151, "grad_norm": 3.7791426452155723, "learning_rate": 3.2752132827055646e-06, "loss": 0.5458, "step": 1144 }, { "epoch": 0.4606718970026152, "grad_norm": 4.1019637445172314, "learning_rate": 3.2718746342468894e-06, "loss": 0.5972, "step": 1145 }, { "epoch": 0.46107423053711527, "grad_norm": 5.277495798462077, "learning_rate": 3.268534463438342e-06, "loss": 0.8266, "step": 1146 }, { "epoch": 0.46147656407161536, "grad_norm": 4.184453273932426, "learning_rate": 3.2651927768676626e-06, "loss": 0.4562, "step": 1147 }, { "epoch": 0.46187889760611545, "grad_norm": 4.45907273350243, "learning_rate": 3.261849581125579e-06, "loss": 0.6532, "step": 1148 }, { "epoch": 0.4622812311406156, "grad_norm": 3.965076229787784, "learning_rate": 3.258504882805796e-06, "loss": 0.5247, "step": 1149 }, { "epoch": 0.4626835646751157, "grad_norm": 4.461055711184354, "learning_rate": 3.2551586885049814e-06, "loss": 0.4588, "step": 1150 }, { "epoch": 0.46308589820961576, "grad_norm": 4.3124706302870255, "learning_rate": 3.2518110048227553e-06, "loss": 0.7004, "step": 1151 }, { "epoch": 0.46348823174411585, "grad_norm": 4.326890541788323, "learning_rate": 3.2484618383616735e-06, "loss": 0.5989, "step": 1152 }, { "epoch": 0.463890565278616, "grad_norm": 4.323105374837098, "learning_rate": 3.2451111957272173e-06, "loss": 0.5451, "step": 1153 }, { "epoch": 0.4642928988131161, "grad_norm": 4.134824677878268, "learning_rate": 3.241759083527779e-06, "loss": 0.5589, "step": 1154 }, { "epoch": 0.46469523234761617, "grad_norm": 4.0981640141993365, "learning_rate": 3.238405508374649e-06, "loss": 0.5771, "step": 1155 }, { "epoch": 0.46509756588211626, "grad_norm": 3.7455178072585644, "learning_rate": 3.2350504768820036e-06, "loss": 0.5125, "step": 1156 }, { "epoch": 0.46549989941661635, "grad_norm": 3.72359460364635, "learning_rate": 3.23169399566689e-06, "loss": 0.4213, "step": 1157 }, { "epoch": 0.4659022329511165, "grad_norm": 3.509972924609966, "learning_rate": 3.228336071349218e-06, "loss": 0.4757, "step": 1158 }, { "epoch": 0.4663045664856166, "grad_norm": 4.028400275439829, "learning_rate": 3.2249767105517395e-06, "loss": 0.5022, "step": 1159 }, { "epoch": 0.46670690002011667, "grad_norm": 4.150972255373417, "learning_rate": 3.221615919900042e-06, "loss": 0.6257, "step": 1160 }, { "epoch": 0.46710923355461675, "grad_norm": 4.494189407683215, "learning_rate": 3.218253706022533e-06, "loss": 0.5914, "step": 1161 }, { "epoch": 0.4675115670891169, "grad_norm": 3.558180537366331, "learning_rate": 3.2148900755504253e-06, "loss": 0.5102, "step": 1162 }, { "epoch": 0.467913900623617, "grad_norm": 3.438406963531958, "learning_rate": 3.2115250351177274e-06, "loss": 0.515, "step": 1163 }, { "epoch": 0.4683162341581171, "grad_norm": 4.323925078750027, "learning_rate": 3.208158591361228e-06, "loss": 0.5109, "step": 1164 }, { "epoch": 0.46871856769261716, "grad_norm": 4.1534707963201285, "learning_rate": 3.204790750920484e-06, "loss": 0.5736, "step": 1165 }, { "epoch": 0.4691209012271173, "grad_norm": 3.6871720181954406, "learning_rate": 3.2014215204378064e-06, "loss": 0.5765, "step": 1166 }, { "epoch": 0.4695232347616174, "grad_norm": 3.9220478852176264, "learning_rate": 3.1980509065582476e-06, "loss": 0.5545, "step": 1167 }, { "epoch": 0.4699255682961175, "grad_norm": 4.603512689826395, "learning_rate": 3.194678915929589e-06, "loss": 0.5587, "step": 1168 }, { "epoch": 0.47032790183061757, "grad_norm": 4.0356910482849075, "learning_rate": 3.1913055552023263e-06, "loss": 0.4642, "step": 1169 }, { "epoch": 0.47073023536511766, "grad_norm": 4.251943286279792, "learning_rate": 3.1879308310296586e-06, "loss": 0.5496, "step": 1170 }, { "epoch": 0.4711325688996178, "grad_norm": 4.189042957707438, "learning_rate": 3.1845547500674737e-06, "loss": 0.5646, "step": 1171 }, { "epoch": 0.4715349024341179, "grad_norm": 3.3326084801089353, "learning_rate": 3.1811773189743355e-06, "loss": 0.4837, "step": 1172 }, { "epoch": 0.471937235968618, "grad_norm": 4.523273563587766, "learning_rate": 3.17779854441147e-06, "loss": 0.6807, "step": 1173 }, { "epoch": 0.47233956950311806, "grad_norm": 3.7232216855517315, "learning_rate": 3.1744184330427543e-06, "loss": 0.4492, "step": 1174 }, { "epoch": 0.4727419030376182, "grad_norm": 3.8162014027986437, "learning_rate": 3.1710369915347e-06, "loss": 0.5626, "step": 1175 }, { "epoch": 0.4731442365721183, "grad_norm": 4.575379469711398, "learning_rate": 3.1676542265564443e-06, "loss": 0.6358, "step": 1176 }, { "epoch": 0.4735465701066184, "grad_norm": 4.563047737373793, "learning_rate": 3.1642701447797324e-06, "loss": 0.6282, "step": 1177 }, { "epoch": 0.47394890364111847, "grad_norm": 4.254463762469985, "learning_rate": 3.1608847528789096e-06, "loss": 0.4974, "step": 1178 }, { "epoch": 0.4743512371756186, "grad_norm": 4.290561801295637, "learning_rate": 3.1574980575309015e-06, "loss": 0.6116, "step": 1179 }, { "epoch": 0.4747535707101187, "grad_norm": 4.276362615412023, "learning_rate": 3.154110065415208e-06, "loss": 0.5274, "step": 1180 }, { "epoch": 0.4751559042446188, "grad_norm": 3.8389354369931374, "learning_rate": 3.150720783213883e-06, "loss": 0.5338, "step": 1181 }, { "epoch": 0.4755582377791189, "grad_norm": 4.448041403696648, "learning_rate": 3.1473302176115283e-06, "loss": 0.5265, "step": 1182 }, { "epoch": 0.47596057131361896, "grad_norm": 4.46050378901124, "learning_rate": 3.1439383752952753e-06, "loss": 0.6299, "step": 1183 }, { "epoch": 0.4763629048481191, "grad_norm": 4.111077775261664, "learning_rate": 3.140545262954772e-06, "loss": 0.5723, "step": 1184 }, { "epoch": 0.4767652383826192, "grad_norm": 3.678492409328739, "learning_rate": 3.137150887282174e-06, "loss": 0.4178, "step": 1185 }, { "epoch": 0.4771675719171193, "grad_norm": 3.661689522078919, "learning_rate": 3.1337552549721262e-06, "loss": 0.4658, "step": 1186 }, { "epoch": 0.47756990545161937, "grad_norm": 4.250351645694971, "learning_rate": 3.130358372721754e-06, "loss": 0.5614, "step": 1187 }, { "epoch": 0.4779722389861195, "grad_norm": 3.599583256588181, "learning_rate": 3.126960247230646e-06, "loss": 0.4569, "step": 1188 }, { "epoch": 0.4783745725206196, "grad_norm": 3.821896657809243, "learning_rate": 3.1235608852008446e-06, "loss": 0.4834, "step": 1189 }, { "epoch": 0.4787769060551197, "grad_norm": 3.9335780159969906, "learning_rate": 3.1201602933368308e-06, "loss": 0.5152, "step": 1190 }, { "epoch": 0.4791792395896198, "grad_norm": 4.770540317196732, "learning_rate": 3.116758478345509e-06, "loss": 0.5948, "step": 1191 }, { "epoch": 0.4795815731241199, "grad_norm": 3.998481220511283, "learning_rate": 3.1133554469361976e-06, "loss": 0.4881, "step": 1192 }, { "epoch": 0.47998390665862, "grad_norm": 4.208648289802026, "learning_rate": 3.109951205820615e-06, "loss": 0.4933, "step": 1193 }, { "epoch": 0.4803862401931201, "grad_norm": 3.752607346638518, "learning_rate": 3.1065457617128635e-06, "loss": 0.547, "step": 1194 }, { "epoch": 0.4807885737276202, "grad_norm": 4.151251812002077, "learning_rate": 3.1031391213294203e-06, "loss": 0.5015, "step": 1195 }, { "epoch": 0.48119090726212027, "grad_norm": 4.7037289072786095, "learning_rate": 3.0997312913891196e-06, "loss": 0.6795, "step": 1196 }, { "epoch": 0.4815932407966204, "grad_norm": 4.364147422750159, "learning_rate": 3.0963222786131435e-06, "loss": 0.5422, "step": 1197 }, { "epoch": 0.4819955743311205, "grad_norm": 3.301076213262608, "learning_rate": 3.0929120897250066e-06, "loss": 0.4646, "step": 1198 }, { "epoch": 0.4823979078656206, "grad_norm": 3.803770968997955, "learning_rate": 3.0895007314505415e-06, "loss": 0.4579, "step": 1199 }, { "epoch": 0.4828002414001207, "grad_norm": 3.6319529462360816, "learning_rate": 3.0860882105178897e-06, "loss": 0.5152, "step": 1200 }, { "epoch": 0.4832025749346208, "grad_norm": 3.7950010562493453, "learning_rate": 3.082674533657484e-06, "loss": 0.588, "step": 1201 }, { "epoch": 0.4836049084691209, "grad_norm": 3.9284108228666796, "learning_rate": 3.079259707602038e-06, "loss": 0.5565, "step": 1202 }, { "epoch": 0.484007242003621, "grad_norm": 4.764319789283065, "learning_rate": 3.0758437390865314e-06, "loss": 0.6115, "step": 1203 }, { "epoch": 0.4844095755381211, "grad_norm": 3.781006793375389, "learning_rate": 3.0724266348481966e-06, "loss": 0.5239, "step": 1204 }, { "epoch": 0.4848119090726212, "grad_norm": 4.259150542468107, "learning_rate": 3.069008401626507e-06, "loss": 0.5505, "step": 1205 }, { "epoch": 0.4852142426071213, "grad_norm": 4.222901377467788, "learning_rate": 3.065589046163162e-06, "loss": 0.5425, "step": 1206 }, { "epoch": 0.4856165761416214, "grad_norm": 3.783266465549031, "learning_rate": 3.062168575202075e-06, "loss": 0.542, "step": 1207 }, { "epoch": 0.4860189096761215, "grad_norm": 4.428497795054238, "learning_rate": 3.0587469954893584e-06, "loss": 0.6567, "step": 1208 }, { "epoch": 0.4864212432106216, "grad_norm": 3.691837740063112, "learning_rate": 3.0553243137733125e-06, "loss": 0.4558, "step": 1209 }, { "epoch": 0.4868235767451217, "grad_norm": 4.073296743584824, "learning_rate": 3.051900536804411e-06, "loss": 0.5664, "step": 1210 }, { "epoch": 0.4872259102796218, "grad_norm": 3.7192683213034234, "learning_rate": 3.0484756713352858e-06, "loss": 0.494, "step": 1211 }, { "epoch": 0.4876282438141219, "grad_norm": 3.6377595381026557, "learning_rate": 3.045049724120718e-06, "loss": 0.4885, "step": 1212 }, { "epoch": 0.488030577348622, "grad_norm": 4.338683717827157, "learning_rate": 3.0416227019176208e-06, "loss": 0.6165, "step": 1213 }, { "epoch": 0.48843291088312213, "grad_norm": 4.374868800699678, "learning_rate": 3.0381946114850293e-06, "loss": 0.5755, "step": 1214 }, { "epoch": 0.4888352444176222, "grad_norm": 4.54077234147531, "learning_rate": 3.034765459584083e-06, "loss": 0.5954, "step": 1215 }, { "epoch": 0.4892375779521223, "grad_norm": 4.166268412885333, "learning_rate": 3.031335252978017e-06, "loss": 0.6657, "step": 1216 }, { "epoch": 0.4896399114866224, "grad_norm": 3.675310094808539, "learning_rate": 3.027903998432145e-06, "loss": 0.4676, "step": 1217 }, { "epoch": 0.49004224502112254, "grad_norm": 4.091571307318168, "learning_rate": 3.0244717027138494e-06, "loss": 0.579, "step": 1218 }, { "epoch": 0.4904445785556226, "grad_norm": 4.2620047470347595, "learning_rate": 3.021038372592563e-06, "loss": 0.577, "step": 1219 }, { "epoch": 0.4908469120901227, "grad_norm": 3.687909147915886, "learning_rate": 3.017604014839762e-06, "loss": 0.557, "step": 1220 }, { "epoch": 0.4912492456246228, "grad_norm": 4.003092692480048, "learning_rate": 3.014168636228948e-06, "loss": 0.5517, "step": 1221 }, { "epoch": 0.4916515791591229, "grad_norm": 4.349298145062102, "learning_rate": 3.010732243535636e-06, "loss": 0.5991, "step": 1222 }, { "epoch": 0.49205391269362303, "grad_norm": 3.9645619754575963, "learning_rate": 3.0072948435373408e-06, "loss": 0.5145, "step": 1223 }, { "epoch": 0.4924562462281231, "grad_norm": 4.216377733109406, "learning_rate": 3.003856443013564e-06, "loss": 0.6017, "step": 1224 }, { "epoch": 0.4928585797626232, "grad_norm": 3.9206543304610584, "learning_rate": 3.0004170487457813e-06, "loss": 0.5468, "step": 1225 }, { "epoch": 0.4932609132971233, "grad_norm": 4.196664240626691, "learning_rate": 2.996976667517427e-06, "loss": 0.5191, "step": 1226 }, { "epoch": 0.49366324683162344, "grad_norm": 3.7775234111981493, "learning_rate": 2.9935353061138833e-06, "loss": 0.5094, "step": 1227 }, { "epoch": 0.4940655803661235, "grad_norm": 4.412452469732304, "learning_rate": 2.990092971322464e-06, "loss": 0.566, "step": 1228 }, { "epoch": 0.4944679139006236, "grad_norm": 3.967665512410053, "learning_rate": 2.986649669932405e-06, "loss": 0.5005, "step": 1229 }, { "epoch": 0.4948702474351237, "grad_norm": 3.8061725506913353, "learning_rate": 2.983205408734846e-06, "loss": 0.5214, "step": 1230 }, { "epoch": 0.49527258096962384, "grad_norm": 3.615684737615807, "learning_rate": 2.9797601945228214e-06, "loss": 0.4647, "step": 1231 }, { "epoch": 0.49567491450412393, "grad_norm": 4.27866020604673, "learning_rate": 2.9763140340912463e-06, "loss": 0.6409, "step": 1232 }, { "epoch": 0.496077248038624, "grad_norm": 3.596076423790147, "learning_rate": 2.972866934236898e-06, "loss": 0.412, "step": 1233 }, { "epoch": 0.4964795815731241, "grad_norm": 4.0355620490752955, "learning_rate": 2.9694189017584097e-06, "loss": 0.5596, "step": 1234 }, { "epoch": 0.4968819151076242, "grad_norm": 4.148051125186017, "learning_rate": 2.965969943456254e-06, "loss": 0.5191, "step": 1235 }, { "epoch": 0.49728424864212434, "grad_norm": 4.2692044767988, "learning_rate": 2.962520066132728e-06, "loss": 0.5066, "step": 1236 }, { "epoch": 0.4976865821766244, "grad_norm": 4.28381423937242, "learning_rate": 2.9590692765919436e-06, "loss": 0.598, "step": 1237 }, { "epoch": 0.4980889157111245, "grad_norm": 4.175358078657294, "learning_rate": 2.955617581639809e-06, "loss": 0.5953, "step": 1238 }, { "epoch": 0.4984912492456246, "grad_norm": 4.5191335184254955, "learning_rate": 2.952164988084021e-06, "loss": 0.6236, "step": 1239 }, { "epoch": 0.49889358278012474, "grad_norm": 4.126996114146587, "learning_rate": 2.9487115027340458e-06, "loss": 0.5455, "step": 1240 }, { "epoch": 0.49929591631462483, "grad_norm": 3.978032878939077, "learning_rate": 2.9452571324011103e-06, "loss": 0.5658, "step": 1241 }, { "epoch": 0.4996982498491249, "grad_norm": 4.494778767380183, "learning_rate": 2.941801883898187e-06, "loss": 0.6639, "step": 1242 }, { "epoch": 0.5001005833836251, "grad_norm": 3.7291179386624376, "learning_rate": 2.9383457640399794e-06, "loss": 0.4724, "step": 1243 }, { "epoch": 0.5005029169181251, "grad_norm": 4.0158904670389175, "learning_rate": 2.9348887796429104e-06, "loss": 0.5611, "step": 1244 }, { "epoch": 0.5009052504526252, "grad_norm": 3.93870398764487, "learning_rate": 2.9314309375251065e-06, "loss": 0.5596, "step": 1245 }, { "epoch": 0.5013075839871253, "grad_norm": 4.111802544399985, "learning_rate": 2.9279722445063883e-06, "loss": 0.5872, "step": 1246 }, { "epoch": 0.5017099175216254, "grad_norm": 3.948101448350297, "learning_rate": 2.9245127074082535e-06, "loss": 0.5131, "step": 1247 }, { "epoch": 0.5021122510561256, "grad_norm": 4.345129844652909, "learning_rate": 2.9210523330538625e-06, "loss": 0.6389, "step": 1248 }, { "epoch": 0.5025145845906256, "grad_norm": 3.6367326824064112, "learning_rate": 2.9175911282680307e-06, "loss": 0.4518, "step": 1249 }, { "epoch": 0.5029169181251257, "grad_norm": 4.027476673290246, "learning_rate": 2.914129099877208e-06, "loss": 0.5234, "step": 1250 }, { "epoch": 0.5033192516596259, "grad_norm": 3.450546244138055, "learning_rate": 2.9106662547094724e-06, "loss": 0.3807, "step": 1251 }, { "epoch": 0.5037215851941259, "grad_norm": 3.696528336339893, "learning_rate": 2.907202599594508e-06, "loss": 0.3884, "step": 1252 }, { "epoch": 0.504123918728626, "grad_norm": 4.252344685184106, "learning_rate": 2.9037381413636e-06, "loss": 0.6935, "step": 1253 }, { "epoch": 0.5045262522631261, "grad_norm": 4.336362311318111, "learning_rate": 2.900272886849616e-06, "loss": 0.5331, "step": 1254 }, { "epoch": 0.5049285857976262, "grad_norm": 3.6842567382976688, "learning_rate": 2.8968068428869955e-06, "loss": 0.5545, "step": 1255 }, { "epoch": 0.5053309193321264, "grad_norm": 3.699409904804457, "learning_rate": 2.8933400163117337e-06, "loss": 0.3754, "step": 1256 }, { "epoch": 0.5057332528666264, "grad_norm": 4.205702275086939, "learning_rate": 2.889872413961369e-06, "loss": 0.6633, "step": 1257 }, { "epoch": 0.5061355864011265, "grad_norm": 3.6639853190462626, "learning_rate": 2.886404042674972e-06, "loss": 0.3892, "step": 1258 }, { "epoch": 0.5065379199356266, "grad_norm": 3.5640332071999765, "learning_rate": 2.882934909293127e-06, "loss": 0.4084, "step": 1259 }, { "epoch": 0.5069402534701267, "grad_norm": 4.0471975082058025, "learning_rate": 2.8794650206579234e-06, "loss": 0.5812, "step": 1260 }, { "epoch": 0.5073425870046269, "grad_norm": 4.0654643808596616, "learning_rate": 2.875994383612939e-06, "loss": 0.568, "step": 1261 }, { "epoch": 0.5077449205391269, "grad_norm": 4.330240433028303, "learning_rate": 2.872523005003229e-06, "loss": 0.5165, "step": 1262 }, { "epoch": 0.508147254073627, "grad_norm": 4.015050113056623, "learning_rate": 2.869050891675309e-06, "loss": 0.5002, "step": 1263 }, { "epoch": 0.5085495876081272, "grad_norm": 4.393983632365189, "learning_rate": 2.8655780504771476e-06, "loss": 0.573, "step": 1264 }, { "epoch": 0.5089519211426272, "grad_norm": 3.193554396038774, "learning_rate": 2.8621044882581433e-06, "loss": 0.3709, "step": 1265 }, { "epoch": 0.5093542546771274, "grad_norm": 3.3841882285013813, "learning_rate": 2.8586302118691223e-06, "loss": 0.4387, "step": 1266 }, { "epoch": 0.5097565882116274, "grad_norm": 3.9970258932238742, "learning_rate": 2.8551552281623157e-06, "loss": 0.5225, "step": 1267 }, { "epoch": 0.5101589217461275, "grad_norm": 4.513606814044003, "learning_rate": 2.85167954399135e-06, "loss": 0.5528, "step": 1268 }, { "epoch": 0.5105612552806277, "grad_norm": 4.684727392976221, "learning_rate": 2.848203166211235e-06, "loss": 0.6793, "step": 1269 }, { "epoch": 0.5109635888151277, "grad_norm": 4.222583621631339, "learning_rate": 2.8447261016783474e-06, "loss": 0.6053, "step": 1270 }, { "epoch": 0.5113659223496279, "grad_norm": 3.8724976223748744, "learning_rate": 2.841248357250418e-06, "loss": 0.4797, "step": 1271 }, { "epoch": 0.5117682558841279, "grad_norm": 3.5925982833624652, "learning_rate": 2.8377699397865184e-06, "loss": 0.6524, "step": 1272 }, { "epoch": 0.512170589418628, "grad_norm": 4.4527497576787285, "learning_rate": 2.8342908561470496e-06, "loss": 0.6426, "step": 1273 }, { "epoch": 0.5125729229531282, "grad_norm": 3.758694762638942, "learning_rate": 2.8308111131937237e-06, "loss": 0.4863, "step": 1274 }, { "epoch": 0.5129752564876282, "grad_norm": 3.432356742958401, "learning_rate": 2.827330717789555e-06, "loss": 0.5223, "step": 1275 }, { "epoch": 0.5133775900221283, "grad_norm": 4.33004996309828, "learning_rate": 2.8238496767988433e-06, "loss": 0.6004, "step": 1276 }, { "epoch": 0.5137799235566285, "grad_norm": 4.200608717420535, "learning_rate": 2.820367997087164e-06, "loss": 0.6228, "step": 1277 }, { "epoch": 0.5141822570911285, "grad_norm": 4.0453448815375035, "learning_rate": 2.8168856855213493e-06, "loss": 0.6242, "step": 1278 }, { "epoch": 0.5145845906256287, "grad_norm": 4.227997150733838, "learning_rate": 2.81340274896948e-06, "loss": 0.6175, "step": 1279 }, { "epoch": 0.5149869241601287, "grad_norm": 3.68997651052952, "learning_rate": 2.8099191943008673e-06, "loss": 0.3891, "step": 1280 }, { "epoch": 0.5153892576946288, "grad_norm": 3.983876769382806, "learning_rate": 2.806435028386044e-06, "loss": 0.4676, "step": 1281 }, { "epoch": 0.515791591229129, "grad_norm": 3.7821657559918362, "learning_rate": 2.8029502580967467e-06, "loss": 0.5712, "step": 1282 }, { "epoch": 0.516193924763629, "grad_norm": 4.613891160258615, "learning_rate": 2.7994648903059034e-06, "loss": 0.6892, "step": 1283 }, { "epoch": 0.5165962582981292, "grad_norm": 4.260048115855587, "learning_rate": 2.7959789318876235e-06, "loss": 0.587, "step": 1284 }, { "epoch": 0.5169985918326292, "grad_norm": 3.2651956443139905, "learning_rate": 2.792492389717178e-06, "loss": 0.4202, "step": 1285 }, { "epoch": 0.5174009253671293, "grad_norm": 3.8441567338972136, "learning_rate": 2.7890052706709913e-06, "loss": 0.5274, "step": 1286 }, { "epoch": 0.5178032589016295, "grad_norm": 4.244436141220556, "learning_rate": 2.785517581626624e-06, "loss": 0.5557, "step": 1287 }, { "epoch": 0.5182055924361295, "grad_norm": 3.777185766227343, "learning_rate": 2.782029329462763e-06, "loss": 0.5508, "step": 1288 }, { "epoch": 0.5186079259706297, "grad_norm": 4.024624316501224, "learning_rate": 2.778540521059204e-06, "loss": 0.4091, "step": 1289 }, { "epoch": 0.5190102595051298, "grad_norm": 3.429656579938711, "learning_rate": 2.775051163296839e-06, "loss": 0.373, "step": 1290 }, { "epoch": 0.5194125930396298, "grad_norm": 4.115516308278128, "learning_rate": 2.771561263057647e-06, "loss": 0.6296, "step": 1291 }, { "epoch": 0.51981492657413, "grad_norm": 4.5340536303593, "learning_rate": 2.7680708272246738e-06, "loss": 0.5478, "step": 1292 }, { "epoch": 0.52021726010863, "grad_norm": 4.1573055283514675, "learning_rate": 2.7645798626820226e-06, "loss": 0.5928, "step": 1293 }, { "epoch": 0.5206195936431302, "grad_norm": 3.5540381627057926, "learning_rate": 2.7610883763148382e-06, "loss": 0.4204, "step": 1294 }, { "epoch": 0.5210219271776303, "grad_norm": 3.7082383118644406, "learning_rate": 2.7575963750092976e-06, "loss": 0.5897, "step": 1295 }, { "epoch": 0.5214242607121303, "grad_norm": 3.670768517938044, "learning_rate": 2.7541038656525906e-06, "loss": 0.4524, "step": 1296 }, { "epoch": 0.5218265942466305, "grad_norm": 3.9599618494551336, "learning_rate": 2.7506108551329084e-06, "loss": 0.5334, "step": 1297 }, { "epoch": 0.5222289277811305, "grad_norm": 3.8332760152691425, "learning_rate": 2.747117350339434e-06, "loss": 0.5027, "step": 1298 }, { "epoch": 0.5226312613156306, "grad_norm": 3.5387447785248463, "learning_rate": 2.743623358162322e-06, "loss": 0.4453, "step": 1299 }, { "epoch": 0.5230335948501308, "grad_norm": 3.665534245478194, "learning_rate": 2.74012888549269e-06, "loss": 0.4541, "step": 1300 }, { "epoch": 0.5234359283846308, "grad_norm": 4.118363799618076, "learning_rate": 2.736633939222604e-06, "loss": 0.4863, "step": 1301 }, { "epoch": 0.523838261919131, "grad_norm": 4.07159865432201, "learning_rate": 2.733138526245061e-06, "loss": 0.5836, "step": 1302 }, { "epoch": 0.5242405954536311, "grad_norm": 4.9413620392110955, "learning_rate": 2.729642653453981e-06, "loss": 0.6331, "step": 1303 }, { "epoch": 0.5246429289881311, "grad_norm": 4.349366316003809, "learning_rate": 2.7261463277441913e-06, "loss": 0.6584, "step": 1304 }, { "epoch": 0.5250452625226313, "grad_norm": 4.013818084372327, "learning_rate": 2.7226495560114097e-06, "loss": 0.6444, "step": 1305 }, { "epoch": 0.5254475960571313, "grad_norm": 3.9731448823583158, "learning_rate": 2.719152345152237e-06, "loss": 0.4717, "step": 1306 }, { "epoch": 0.5258499295916315, "grad_norm": 3.836192951376196, "learning_rate": 2.7156547020641377e-06, "loss": 0.4813, "step": 1307 }, { "epoch": 0.5262522631261316, "grad_norm": 4.526603353793837, "learning_rate": 2.7121566336454295e-06, "loss": 0.6223, "step": 1308 }, { "epoch": 0.5266545966606316, "grad_norm": 4.542236066581728, "learning_rate": 2.70865814679527e-06, "loss": 0.541, "step": 1309 }, { "epoch": 0.5270569301951318, "grad_norm": 4.477320709723044, "learning_rate": 2.70515924841364e-06, "loss": 0.4898, "step": 1310 }, { "epoch": 0.5274592637296318, "grad_norm": 4.188969544607867, "learning_rate": 2.7016599454013324e-06, "loss": 0.478, "step": 1311 }, { "epoch": 0.527861597264132, "grad_norm": 3.57921084312417, "learning_rate": 2.6981602446599408e-06, "loss": 0.4797, "step": 1312 }, { "epoch": 0.5282639307986321, "grad_norm": 3.4828294543575784, "learning_rate": 2.69466015309184e-06, "loss": 0.4464, "step": 1313 }, { "epoch": 0.5286662643331321, "grad_norm": 3.6542821945407002, "learning_rate": 2.691159677600177e-06, "loss": 0.5241, "step": 1314 }, { "epoch": 0.5290685978676323, "grad_norm": 4.138582178147402, "learning_rate": 2.6876588250888556e-06, "loss": 0.4976, "step": 1315 }, { "epoch": 0.5294709314021324, "grad_norm": 4.156252344667509, "learning_rate": 2.6841576024625247e-06, "loss": 0.5346, "step": 1316 }, { "epoch": 0.5298732649366324, "grad_norm": 3.9783539180873118, "learning_rate": 2.680656016626561e-06, "loss": 0.5054, "step": 1317 }, { "epoch": 0.5302755984711326, "grad_norm": 4.167749289113631, "learning_rate": 2.677154074487057e-06, "loss": 0.5308, "step": 1318 }, { "epoch": 0.5306779320056326, "grad_norm": 3.6243864679157536, "learning_rate": 2.6736517829508124e-06, "loss": 0.4969, "step": 1319 }, { "epoch": 0.5310802655401328, "grad_norm": 4.080652153118834, "learning_rate": 2.6701491489253107e-06, "loss": 0.5166, "step": 1320 }, { "epoch": 0.5314825990746329, "grad_norm": 4.386897362383017, "learning_rate": 2.6666461793187143e-06, "loss": 0.4953, "step": 1321 }, { "epoch": 0.5318849326091329, "grad_norm": 4.286991486124671, "learning_rate": 2.6631428810398453e-06, "loss": 0.4804, "step": 1322 }, { "epoch": 0.5322872661436331, "grad_norm": 3.996180824655324, "learning_rate": 2.659639260998176e-06, "loss": 0.5182, "step": 1323 }, { "epoch": 0.5326895996781331, "grad_norm": 4.305112080018121, "learning_rate": 2.6561353261038127e-06, "loss": 0.6548, "step": 1324 }, { "epoch": 0.5330919332126333, "grad_norm": 3.6190226327616557, "learning_rate": 2.6526310832674807e-06, "loss": 0.5383, "step": 1325 }, { "epoch": 0.5334942667471334, "grad_norm": 4.316996697367821, "learning_rate": 2.6491265394005157e-06, "loss": 0.5516, "step": 1326 }, { "epoch": 0.5338966002816334, "grad_norm": 4.0648368660213325, "learning_rate": 2.645621701414845e-06, "loss": 0.6482, "step": 1327 }, { "epoch": 0.5342989338161336, "grad_norm": 3.7739876249793047, "learning_rate": 2.6421165762229777e-06, "loss": 0.4783, "step": 1328 }, { "epoch": 0.5347012673506337, "grad_norm": 3.789987275308021, "learning_rate": 2.6386111707379873e-06, "loss": 0.4525, "step": 1329 }, { "epoch": 0.5351036008851338, "grad_norm": 3.630575939542573, "learning_rate": 2.635105491873502e-06, "loss": 0.5329, "step": 1330 }, { "epoch": 0.5355059344196339, "grad_norm": 3.961557081061209, "learning_rate": 2.6315995465436878e-06, "loss": 0.5432, "step": 1331 }, { "epoch": 0.5359082679541339, "grad_norm": 3.8780055120857204, "learning_rate": 2.628093341663236e-06, "loss": 0.454, "step": 1332 }, { "epoch": 0.5363106014886341, "grad_norm": 4.031126634163913, "learning_rate": 2.6245868841473525e-06, "loss": 0.5452, "step": 1333 }, { "epoch": 0.5367129350231342, "grad_norm": 3.3788818500452997, "learning_rate": 2.6210801809117377e-06, "loss": 0.4364, "step": 1334 }, { "epoch": 0.5371152685576343, "grad_norm": 4.566291916332284, "learning_rate": 2.61757323887258e-06, "loss": 0.6427, "step": 1335 }, { "epoch": 0.5375176020921344, "grad_norm": 4.46318346856033, "learning_rate": 2.614066064946536e-06, "loss": 0.6142, "step": 1336 }, { "epoch": 0.5379199356266344, "grad_norm": 3.8724650951472377, "learning_rate": 2.6105586660507214e-06, "loss": 0.4865, "step": 1337 }, { "epoch": 0.5383222691611346, "grad_norm": 4.107266394665508, "learning_rate": 2.607051049102696e-06, "loss": 0.5618, "step": 1338 }, { "epoch": 0.5387246026956347, "grad_norm": 4.131578269950126, "learning_rate": 2.6035432210204475e-06, "loss": 0.6571, "step": 1339 }, { "epoch": 0.5391269362301347, "grad_norm": 4.135237490425349, "learning_rate": 2.6000351887223817e-06, "loss": 0.506, "step": 1340 }, { "epoch": 0.5395292697646349, "grad_norm": 4.2799950491642695, "learning_rate": 2.5965269591273075e-06, "loss": 0.5054, "step": 1341 }, { "epoch": 0.539931603299135, "grad_norm": 4.111277509443795, "learning_rate": 2.59301853915442e-06, "loss": 0.5896, "step": 1342 }, { "epoch": 0.5403339368336351, "grad_norm": 3.8507472387773416, "learning_rate": 2.5895099357232946e-06, "loss": 0.4784, "step": 1343 }, { "epoch": 0.5407362703681352, "grad_norm": 4.553167071691355, "learning_rate": 2.5860011557538644e-06, "loss": 0.5889, "step": 1344 }, { "epoch": 0.5411386039026352, "grad_norm": 4.8377815361815175, "learning_rate": 2.5824922061664125e-06, "loss": 0.7153, "step": 1345 }, { "epoch": 0.5415409374371354, "grad_norm": 4.026163246451428, "learning_rate": 2.578983093881557e-06, "loss": 0.5578, "step": 1346 }, { "epoch": 0.5419432709716355, "grad_norm": 4.396870764657521, "learning_rate": 2.5754738258202345e-06, "loss": 0.7799, "step": 1347 }, { "epoch": 0.5423456045061356, "grad_norm": 4.349518153263542, "learning_rate": 2.5719644089036916e-06, "loss": 0.5919, "step": 1348 }, { "epoch": 0.5427479380406357, "grad_norm": 4.355976508524607, "learning_rate": 2.568454850053467e-06, "loss": 0.5903, "step": 1349 }, { "epoch": 0.5431502715751357, "grad_norm": 3.951612074252993, "learning_rate": 2.5649451561913796e-06, "loss": 0.5038, "step": 1350 }, { "epoch": 0.5435526051096359, "grad_norm": 4.030293462511148, "learning_rate": 2.561435334239515e-06, "loss": 0.5308, "step": 1351 }, { "epoch": 0.543954938644136, "grad_norm": 4.530833691231785, "learning_rate": 2.55792539112021e-06, "loss": 0.5977, "step": 1352 }, { "epoch": 0.544357272178636, "grad_norm": 4.1989735383766185, "learning_rate": 2.5544153337560422e-06, "loss": 0.5686, "step": 1353 }, { "epoch": 0.5447596057131362, "grad_norm": 4.32105387076282, "learning_rate": 2.5509051690698137e-06, "loss": 0.6257, "step": 1354 }, { "epoch": 0.5451619392476363, "grad_norm": 3.34919768311592, "learning_rate": 2.5473949039845388e-06, "loss": 0.4131, "step": 1355 }, { "epoch": 0.5455642727821364, "grad_norm": 4.128988896294745, "learning_rate": 2.5438845454234284e-06, "loss": 0.6007, "step": 1356 }, { "epoch": 0.5459666063166365, "grad_norm": 4.122788652813818, "learning_rate": 2.5403741003098793e-06, "loss": 0.5223, "step": 1357 }, { "epoch": 0.5463689398511365, "grad_norm": 4.420118909526267, "learning_rate": 2.5368635755674586e-06, "loss": 0.5506, "step": 1358 }, { "epoch": 0.5467712733856367, "grad_norm": 4.2195230085645345, "learning_rate": 2.5333529781198895e-06, "loss": 0.5989, "step": 1359 }, { "epoch": 0.5471736069201368, "grad_norm": 4.20795164503954, "learning_rate": 2.5298423148910396e-06, "loss": 0.4462, "step": 1360 }, { "epoch": 0.5475759404546369, "grad_norm": 4.010223353178494, "learning_rate": 2.5263315928049063e-06, "loss": 0.5192, "step": 1361 }, { "epoch": 0.547978273989137, "grad_norm": 4.816701841105469, "learning_rate": 2.5228208187856024e-06, "loss": 0.6599, "step": 1362 }, { "epoch": 0.548380607523637, "grad_norm": 4.023413458632839, "learning_rate": 2.5193099997573437e-06, "loss": 0.6185, "step": 1363 }, { "epoch": 0.5487829410581372, "grad_norm": 3.6516400727809115, "learning_rate": 2.5157991426444343e-06, "loss": 0.5859, "step": 1364 }, { "epoch": 0.5491852745926373, "grad_norm": 4.119751084462459, "learning_rate": 2.5122882543712546e-06, "loss": 0.5255, "step": 1365 }, { "epoch": 0.5495876081271374, "grad_norm": 3.957835467170672, "learning_rate": 2.5087773418622447e-06, "loss": 0.4659, "step": 1366 }, { "epoch": 0.5499899416616375, "grad_norm": 3.500258182406765, "learning_rate": 2.505266412041893e-06, "loss": 0.4385, "step": 1367 }, { "epoch": 0.5503922751961376, "grad_norm": 4.041408070624076, "learning_rate": 2.5017554718347233e-06, "loss": 0.5367, "step": 1368 }, { "epoch": 0.5507946087306377, "grad_norm": 4.091765389792879, "learning_rate": 2.4982445281652775e-06, "loss": 0.5744, "step": 1369 }, { "epoch": 0.5511969422651378, "grad_norm": 4.469717920616285, "learning_rate": 2.4947335879581073e-06, "loss": 0.5818, "step": 1370 }, { "epoch": 0.5515992757996379, "grad_norm": 3.9779088164073833, "learning_rate": 2.4912226581377566e-06, "loss": 0.5893, "step": 1371 }, { "epoch": 0.552001609334138, "grad_norm": 3.627034776419984, "learning_rate": 2.4877117456287463e-06, "loss": 0.4902, "step": 1372 }, { "epoch": 0.5524039428686381, "grad_norm": 3.8562269767576405, "learning_rate": 2.484200857355566e-06, "loss": 0.6268, "step": 1373 }, { "epoch": 0.5528062764031382, "grad_norm": 4.192164362531014, "learning_rate": 2.4806900002426567e-06, "loss": 0.6121, "step": 1374 }, { "epoch": 0.5532086099376383, "grad_norm": 3.916801705190097, "learning_rate": 2.477179181214398e-06, "loss": 0.4926, "step": 1375 }, { "epoch": 0.5536109434721384, "grad_norm": 3.693356946438789, "learning_rate": 2.473668407195095e-06, "loss": 0.5189, "step": 1376 }, { "epoch": 0.5540132770066385, "grad_norm": 4.229576715855437, "learning_rate": 2.4701576851089613e-06, "loss": 0.6398, "step": 1377 }, { "epoch": 0.5544156105411386, "grad_norm": 3.6167470608574197, "learning_rate": 2.4666470218801113e-06, "loss": 0.5175, "step": 1378 }, { "epoch": 0.5548179440756387, "grad_norm": 5.070171032981664, "learning_rate": 2.463136424432542e-06, "loss": 0.4988, "step": 1379 }, { "epoch": 0.5552202776101388, "grad_norm": 4.058368062649386, "learning_rate": 2.4596258996901215e-06, "loss": 0.4313, "step": 1380 }, { "epoch": 0.555622611144639, "grad_norm": 3.9184186874819464, "learning_rate": 2.4561154545765724e-06, "loss": 0.5338, "step": 1381 }, { "epoch": 0.556024944679139, "grad_norm": 3.1244549306950784, "learning_rate": 2.4526050960154616e-06, "loss": 0.4289, "step": 1382 }, { "epoch": 0.5564272782136391, "grad_norm": 4.04400253855949, "learning_rate": 2.4490948309301867e-06, "loss": 0.5257, "step": 1383 }, { "epoch": 0.5568296117481392, "grad_norm": 3.9303715740181513, "learning_rate": 2.445584666243959e-06, "loss": 0.5149, "step": 1384 }, { "epoch": 0.5572319452826393, "grad_norm": 4.395655391522824, "learning_rate": 2.442074608879791e-06, "loss": 0.6249, "step": 1385 }, { "epoch": 0.5576342788171395, "grad_norm": 3.8268439749597576, "learning_rate": 2.438564665760486e-06, "loss": 0.5613, "step": 1386 }, { "epoch": 0.5580366123516395, "grad_norm": 4.255793020118025, "learning_rate": 2.4350548438086212e-06, "loss": 0.6155, "step": 1387 }, { "epoch": 0.5584389458861396, "grad_norm": 3.979739013550483, "learning_rate": 2.4315451499465336e-06, "loss": 0.5394, "step": 1388 }, { "epoch": 0.5588412794206397, "grad_norm": 3.781411432288367, "learning_rate": 2.4280355910963097e-06, "loss": 0.4507, "step": 1389 }, { "epoch": 0.5592436129551398, "grad_norm": 4.371781717934846, "learning_rate": 2.4245261741797663e-06, "loss": 0.5305, "step": 1390 }, { "epoch": 0.55964594648964, "grad_norm": 3.52895283713103, "learning_rate": 2.421016906118444e-06, "loss": 0.5194, "step": 1391 }, { "epoch": 0.56004828002414, "grad_norm": 3.872042926144624, "learning_rate": 2.417507793833587e-06, "loss": 0.4998, "step": 1392 }, { "epoch": 0.5604506135586401, "grad_norm": 3.391234256928363, "learning_rate": 2.413998844246136e-06, "loss": 0.4093, "step": 1393 }, { "epoch": 0.5608529470931403, "grad_norm": 3.9654243758806174, "learning_rate": 2.4104900642767066e-06, "loss": 0.4515, "step": 1394 }, { "epoch": 0.5612552806276403, "grad_norm": 3.5879075344069378, "learning_rate": 2.4069814608455804e-06, "loss": 0.5207, "step": 1395 }, { "epoch": 0.5616576141621404, "grad_norm": 4.104094820028554, "learning_rate": 2.4034730408726938e-06, "loss": 0.5625, "step": 1396 }, { "epoch": 0.5620599476966405, "grad_norm": 3.772298699595217, "learning_rate": 2.3999648112776183e-06, "loss": 0.5068, "step": 1397 }, { "epoch": 0.5624622812311406, "grad_norm": 3.688052093132043, "learning_rate": 2.3964567789795533e-06, "loss": 0.5042, "step": 1398 }, { "epoch": 0.5628646147656408, "grad_norm": 3.8931716258044498, "learning_rate": 2.3929489508973047e-06, "loss": 0.5485, "step": 1399 }, { "epoch": 0.5632669483001408, "grad_norm": 4.224728616738161, "learning_rate": 2.389441333949279e-06, "loss": 0.6582, "step": 1400 }, { "epoch": 0.5636692818346409, "grad_norm": 3.8494878180062897, "learning_rate": 2.3859339350534646e-06, "loss": 0.5379, "step": 1401 }, { "epoch": 0.564071615369141, "grad_norm": 4.200325419085908, "learning_rate": 2.382426761127421e-06, "loss": 0.5997, "step": 1402 }, { "epoch": 0.5644739489036411, "grad_norm": 4.143252972394019, "learning_rate": 2.378919819088263e-06, "loss": 0.6195, "step": 1403 }, { "epoch": 0.5648762824381413, "grad_norm": 3.7174533902571145, "learning_rate": 2.3754131158526483e-06, "loss": 0.5317, "step": 1404 }, { "epoch": 0.5652786159726413, "grad_norm": 4.0056931921159356, "learning_rate": 2.371906658336764e-06, "loss": 0.5232, "step": 1405 }, { "epoch": 0.5656809495071414, "grad_norm": 3.830247393182765, "learning_rate": 2.368400453456313e-06, "loss": 0.4898, "step": 1406 }, { "epoch": 0.5660832830416416, "grad_norm": 3.7820185066794787, "learning_rate": 2.364894508126499e-06, "loss": 0.5438, "step": 1407 }, { "epoch": 0.5664856165761416, "grad_norm": 4.4898443665627, "learning_rate": 2.3613888292620135e-06, "loss": 0.6322, "step": 1408 }, { "epoch": 0.5668879501106417, "grad_norm": 4.157236186991902, "learning_rate": 2.3578834237770227e-06, "loss": 0.5978, "step": 1409 }, { "epoch": 0.5672902836451418, "grad_norm": 4.576188931404494, "learning_rate": 2.354378298585155e-06, "loss": 0.5774, "step": 1410 }, { "epoch": 0.5676926171796419, "grad_norm": 3.714487248907507, "learning_rate": 2.3508734605994855e-06, "loss": 0.4305, "step": 1411 }, { "epoch": 0.5680949507141421, "grad_norm": 3.9270791753173313, "learning_rate": 2.3473689167325205e-06, "loss": 0.4912, "step": 1412 }, { "epoch": 0.5684972842486421, "grad_norm": 3.655781362606588, "learning_rate": 2.3438646738961886e-06, "loss": 0.5125, "step": 1413 }, { "epoch": 0.5688996177831422, "grad_norm": 4.2097153394323845, "learning_rate": 2.3403607390018246e-06, "loss": 0.4273, "step": 1414 }, { "epoch": 0.5693019513176423, "grad_norm": 3.7418997500983, "learning_rate": 2.3368571189601556e-06, "loss": 0.4534, "step": 1415 }, { "epoch": 0.5697042848521424, "grad_norm": 3.7385152001909154, "learning_rate": 2.3333538206812874e-06, "loss": 0.4461, "step": 1416 }, { "epoch": 0.5701066183866426, "grad_norm": 4.243485694858217, "learning_rate": 2.32985085107469e-06, "loss": 0.4943, "step": 1417 }, { "epoch": 0.5705089519211426, "grad_norm": 4.064181274521891, "learning_rate": 2.3263482170491884e-06, "loss": 0.6696, "step": 1418 }, { "epoch": 0.5709112854556427, "grad_norm": 4.168819688893492, "learning_rate": 2.3228459255129425e-06, "loss": 0.5914, "step": 1419 }, { "epoch": 0.5713136189901429, "grad_norm": 3.299064883781313, "learning_rate": 2.3193439833734403e-06, "loss": 0.3987, "step": 1420 }, { "epoch": 0.5717159525246429, "grad_norm": 4.887635853989916, "learning_rate": 2.3158423975374757e-06, "loss": 0.4842, "step": 1421 }, { "epoch": 0.572118286059143, "grad_norm": 3.4644648967520806, "learning_rate": 2.312341174911145e-06, "loss": 0.4643, "step": 1422 }, { "epoch": 0.5725206195936431, "grad_norm": 3.75941250297541, "learning_rate": 2.3088403223998235e-06, "loss": 0.4772, "step": 1423 }, { "epoch": 0.5729229531281432, "grad_norm": 3.4615485090255578, "learning_rate": 2.3053398469081613e-06, "loss": 0.5075, "step": 1424 }, { "epoch": 0.5733252866626434, "grad_norm": 4.2227564973481995, "learning_rate": 2.30183975534006e-06, "loss": 0.5125, "step": 1425 }, { "epoch": 0.5737276201971434, "grad_norm": 4.090888673139273, "learning_rate": 2.298340054598668e-06, "loss": 0.5183, "step": 1426 }, { "epoch": 0.5741299537316435, "grad_norm": 4.548845256935699, "learning_rate": 2.294840751586361e-06, "loss": 0.639, "step": 1427 }, { "epoch": 0.5745322872661436, "grad_norm": 3.901263423192965, "learning_rate": 2.291341853204731e-06, "loss": 0.623, "step": 1428 }, { "epoch": 0.5749346208006437, "grad_norm": 3.347357658304439, "learning_rate": 2.287843366354571e-06, "loss": 0.4009, "step": 1429 }, { "epoch": 0.5753369543351439, "grad_norm": 4.512229941763247, "learning_rate": 2.284345297935863e-06, "loss": 0.6336, "step": 1430 }, { "epoch": 0.5757392878696439, "grad_norm": 4.132485335550154, "learning_rate": 2.2808476548477636e-06, "loss": 0.5015, "step": 1431 }, { "epoch": 0.576141621404144, "grad_norm": 4.226000440762799, "learning_rate": 2.2773504439885903e-06, "loss": 0.5063, "step": 1432 }, { "epoch": 0.5765439549386442, "grad_norm": 4.010604789353783, "learning_rate": 2.27385367225581e-06, "loss": 0.5085, "step": 1433 }, { "epoch": 0.5769462884731442, "grad_norm": 3.71838877160847, "learning_rate": 2.2703573465460194e-06, "loss": 0.483, "step": 1434 }, { "epoch": 0.5773486220076444, "grad_norm": 3.715720870136391, "learning_rate": 2.2668614737549395e-06, "loss": 0.5292, "step": 1435 }, { "epoch": 0.5777509555421444, "grad_norm": 3.7875519229722756, "learning_rate": 2.263366060777397e-06, "loss": 0.4504, "step": 1436 }, { "epoch": 0.5781532890766445, "grad_norm": 4.235567100677654, "learning_rate": 2.25987111450731e-06, "loss": 0.6075, "step": 1437 }, { "epoch": 0.5785556226111447, "grad_norm": 4.23015897289991, "learning_rate": 2.256376641837679e-06, "loss": 0.6036, "step": 1438 }, { "epoch": 0.5789579561456447, "grad_norm": 4.397376552533781, "learning_rate": 2.2528826496605672e-06, "loss": 0.6584, "step": 1439 }, { "epoch": 0.5793602896801449, "grad_norm": 3.987430088200103, "learning_rate": 2.249389144867092e-06, "loss": 0.5312, "step": 1440 }, { "epoch": 0.5797626232146449, "grad_norm": 4.349402058381357, "learning_rate": 2.24589613434741e-06, "loss": 0.4806, "step": 1441 }, { "epoch": 0.580164956749145, "grad_norm": 4.389627692568862, "learning_rate": 2.2424036249907032e-06, "loss": 0.649, "step": 1442 }, { "epoch": 0.5805672902836452, "grad_norm": 3.889422512107434, "learning_rate": 2.238911623685162e-06, "loss": 0.5553, "step": 1443 }, { "epoch": 0.5809696238181452, "grad_norm": 4.1471977764513674, "learning_rate": 2.235420137317978e-06, "loss": 0.5527, "step": 1444 }, { "epoch": 0.5813719573526454, "grad_norm": 3.642246289302764, "learning_rate": 2.2319291727753266e-06, "loss": 0.4669, "step": 1445 }, { "epoch": 0.5817742908871455, "grad_norm": 3.7186046973608406, "learning_rate": 2.228438736942354e-06, "loss": 0.3984, "step": 1446 }, { "epoch": 0.5821766244216455, "grad_norm": 4.208948818822306, "learning_rate": 2.2249488367031612e-06, "loss": 0.574, "step": 1447 }, { "epoch": 0.5825789579561457, "grad_norm": 3.4944920810310633, "learning_rate": 2.221459478940797e-06, "loss": 0.4565, "step": 1448 }, { "epoch": 0.5829812914906457, "grad_norm": 3.8209272903130325, "learning_rate": 2.2179706705372377e-06, "loss": 0.4116, "step": 1449 }, { "epoch": 0.5833836250251458, "grad_norm": 4.562727204828236, "learning_rate": 2.214482418373376e-06, "loss": 0.6272, "step": 1450 }, { "epoch": 0.583785958559646, "grad_norm": 3.986021082183279, "learning_rate": 2.21099472932901e-06, "loss": 0.4946, "step": 1451 }, { "epoch": 0.584188292094146, "grad_norm": 3.619941584357637, "learning_rate": 2.207507610282823e-06, "loss": 0.5552, "step": 1452 }, { "epoch": 0.5845906256286462, "grad_norm": 4.265275522318885, "learning_rate": 2.2040210681123773e-06, "loss": 0.5387, "step": 1453 }, { "epoch": 0.5849929591631462, "grad_norm": 3.592671313761321, "learning_rate": 2.2005351096940966e-06, "loss": 0.4221, "step": 1454 }, { "epoch": 0.5853952926976463, "grad_norm": 4.0275300106962755, "learning_rate": 2.1970497419032546e-06, "loss": 0.5781, "step": 1455 }, { "epoch": 0.5857976262321465, "grad_norm": 3.8884576976533425, "learning_rate": 2.1935649716139567e-06, "loss": 0.5121, "step": 1456 }, { "epoch": 0.5861999597666465, "grad_norm": 4.481785017500937, "learning_rate": 2.1900808056991335e-06, "loss": 0.6321, "step": 1457 }, { "epoch": 0.5866022933011467, "grad_norm": 3.74252328127819, "learning_rate": 2.186597251030521e-06, "loss": 0.4935, "step": 1458 }, { "epoch": 0.5870046268356468, "grad_norm": 3.9659714356438442, "learning_rate": 2.183114314478651e-06, "loss": 0.5503, "step": 1459 }, { "epoch": 0.5874069603701468, "grad_norm": 3.435746642131932, "learning_rate": 2.179632002912837e-06, "loss": 0.3508, "step": 1460 }, { "epoch": 0.587809293904647, "grad_norm": 4.32896011694038, "learning_rate": 2.176150323201157e-06, "loss": 0.5728, "step": 1461 }, { "epoch": 0.588211627439147, "grad_norm": 4.1710482850436215, "learning_rate": 2.1726692822104455e-06, "loss": 0.5115, "step": 1462 }, { "epoch": 0.5886139609736472, "grad_norm": 4.1244090573644145, "learning_rate": 2.169188886806277e-06, "loss": 0.3078, "step": 1463 }, { "epoch": 0.5890162945081473, "grad_norm": 3.462366030557245, "learning_rate": 2.1657091438529517e-06, "loss": 0.472, "step": 1464 }, { "epoch": 0.5894186280426473, "grad_norm": 3.920420130653373, "learning_rate": 2.1622300602134824e-06, "loss": 0.5559, "step": 1465 }, { "epoch": 0.5898209615771475, "grad_norm": 4.574556709525141, "learning_rate": 2.158751642749583e-06, "loss": 0.5525, "step": 1466 }, { "epoch": 0.5902232951116475, "grad_norm": 3.8284931733353513, "learning_rate": 2.155273898321653e-06, "loss": 0.5681, "step": 1467 }, { "epoch": 0.5906256286461476, "grad_norm": 3.3784102201483677, "learning_rate": 2.151796833788766e-06, "loss": 0.4727, "step": 1468 }, { "epoch": 0.5910279621806478, "grad_norm": 4.047681657794816, "learning_rate": 2.1483204560086507e-06, "loss": 0.5665, "step": 1469 }, { "epoch": 0.5914302957151478, "grad_norm": 4.308357381109115, "learning_rate": 2.144844771837685e-06, "loss": 0.6112, "step": 1470 }, { "epoch": 0.591832629249648, "grad_norm": 3.733873078989285, "learning_rate": 2.1413697881308785e-06, "loss": 0.494, "step": 1471 }, { "epoch": 0.5922349627841481, "grad_norm": 4.130337043839704, "learning_rate": 2.1378955117418567e-06, "loss": 0.445, "step": 1472 }, { "epoch": 0.5926372963186481, "grad_norm": 4.189878829812914, "learning_rate": 2.1344219495228537e-06, "loss": 0.569, "step": 1473 }, { "epoch": 0.5930396298531483, "grad_norm": 3.926746164109752, "learning_rate": 2.1309491083246915e-06, "loss": 0.5696, "step": 1474 }, { "epoch": 0.5934419633876483, "grad_norm": 3.688932516880435, "learning_rate": 2.127476994996772e-06, "loss": 0.5001, "step": 1475 }, { "epoch": 0.5938442969221485, "grad_norm": 4.084824375606503, "learning_rate": 2.1240056163870614e-06, "loss": 0.613, "step": 1476 }, { "epoch": 0.5942466304566486, "grad_norm": 4.614880824241214, "learning_rate": 2.120534979342078e-06, "loss": 0.5252, "step": 1477 }, { "epoch": 0.5946489639911486, "grad_norm": 3.0888633468235014, "learning_rate": 2.1170650907068742e-06, "loss": 0.4531, "step": 1478 }, { "epoch": 0.5950512975256488, "grad_norm": 4.213120019047452, "learning_rate": 2.113595957325029e-06, "loss": 0.6458, "step": 1479 }, { "epoch": 0.5954536310601488, "grad_norm": 3.610258712936895, "learning_rate": 2.1101275860386313e-06, "loss": 0.5652, "step": 1480 }, { "epoch": 0.595855964594649, "grad_norm": 3.901115632452379, "learning_rate": 2.1066599836882667e-06, "loss": 0.4731, "step": 1481 }, { "epoch": 0.5962582981291491, "grad_norm": 4.576321666746087, "learning_rate": 2.1031931571130053e-06, "loss": 0.6799, "step": 1482 }, { "epoch": 0.5966606316636491, "grad_norm": 4.188708168775149, "learning_rate": 2.0997271131503847e-06, "loss": 0.5172, "step": 1483 }, { "epoch": 0.5970629651981493, "grad_norm": 3.993460741506799, "learning_rate": 2.0962618586364005e-06, "loss": 0.4828, "step": 1484 }, { "epoch": 0.5974652987326494, "grad_norm": 3.977543353419007, "learning_rate": 2.092797400405493e-06, "loss": 0.5368, "step": 1485 }, { "epoch": 0.5978676322671495, "grad_norm": 4.079669745022429, "learning_rate": 2.0893337452905293e-06, "loss": 0.4879, "step": 1486 }, { "epoch": 0.5982699658016496, "grad_norm": 4.000645952876263, "learning_rate": 2.0858709001227924e-06, "loss": 0.5338, "step": 1487 }, { "epoch": 0.5986722993361496, "grad_norm": 4.483257438703314, "learning_rate": 2.0824088717319697e-06, "loss": 0.5771, "step": 1488 }, { "epoch": 0.5990746328706498, "grad_norm": 3.493611740579638, "learning_rate": 2.0789476669461375e-06, "loss": 0.4758, "step": 1489 }, { "epoch": 0.5994769664051499, "grad_norm": 3.9869817394957385, "learning_rate": 2.0754872925917465e-06, "loss": 0.5482, "step": 1490 }, { "epoch": 0.59987929993965, "grad_norm": 3.655771235240259, "learning_rate": 2.072027755493612e-06, "loss": 0.4924, "step": 1491 }, { "epoch": 0.6002816334741501, "grad_norm": 3.959985484700295, "learning_rate": 2.0685690624748943e-06, "loss": 0.5653, "step": 1492 }, { "epoch": 0.6006839670086501, "grad_norm": 3.4092513778358047, "learning_rate": 2.0651112203570904e-06, "loss": 0.4375, "step": 1493 }, { "epoch": 0.6010863005431503, "grad_norm": 4.050380522574125, "learning_rate": 2.061654235960021e-06, "loss": 0.5627, "step": 1494 }, { "epoch": 0.6014886340776504, "grad_norm": 4.329223167800351, "learning_rate": 2.058198116101814e-06, "loss": 0.5642, "step": 1495 }, { "epoch": 0.6018909676121504, "grad_norm": 3.475966176009881, "learning_rate": 2.0547428675988906e-06, "loss": 0.4417, "step": 1496 }, { "epoch": 0.6022933011466506, "grad_norm": 3.7228401404060443, "learning_rate": 2.051288497265955e-06, "loss": 0.4509, "step": 1497 }, { "epoch": 0.6026956346811507, "grad_norm": 3.9298800993895284, "learning_rate": 2.04783501191598e-06, "loss": 0.4373, "step": 1498 }, { "epoch": 0.6030979682156508, "grad_norm": 4.3431668642603105, "learning_rate": 2.0443824183601917e-06, "loss": 0.6361, "step": 1499 }, { "epoch": 0.6035003017501509, "grad_norm": 3.759176356037184, "learning_rate": 2.0409307234080577e-06, "loss": 0.4214, "step": 1500 }, { "epoch": 0.6039026352846509, "grad_norm": 3.881662137011941, "learning_rate": 2.0374799338672723e-06, "loss": 0.4868, "step": 1501 }, { "epoch": 0.6043049688191511, "grad_norm": 4.786052455382573, "learning_rate": 2.034030056543747e-06, "loss": 0.6772, "step": 1502 }, { "epoch": 0.6047073023536512, "grad_norm": 4.137240403825967, "learning_rate": 2.0305810982415907e-06, "loss": 0.4981, "step": 1503 }, { "epoch": 0.6051096358881513, "grad_norm": 3.9245214475911947, "learning_rate": 2.0271330657631034e-06, "loss": 0.461, "step": 1504 }, { "epoch": 0.6055119694226514, "grad_norm": 4.043701387123113, "learning_rate": 2.023685965908755e-06, "loss": 0.4861, "step": 1505 }, { "epoch": 0.6059143029571514, "grad_norm": 4.578623188458431, "learning_rate": 2.020239805477179e-06, "loss": 0.5516, "step": 1506 }, { "epoch": 0.6063166364916516, "grad_norm": 3.939503557149013, "learning_rate": 2.0167945912651542e-06, "loss": 0.5348, "step": 1507 }, { "epoch": 0.6067189700261517, "grad_norm": 4.274904255142657, "learning_rate": 2.0133503300675963e-06, "loss": 0.5992, "step": 1508 }, { "epoch": 0.6071213035606517, "grad_norm": 3.732143003199723, "learning_rate": 2.0099070286775367e-06, "loss": 0.4891, "step": 1509 }, { "epoch": 0.6075236370951519, "grad_norm": 4.093400863247009, "learning_rate": 2.0064646938861175e-06, "loss": 0.4785, "step": 1510 }, { "epoch": 0.607925970629652, "grad_norm": 3.850637017187709, "learning_rate": 2.0030233324825734e-06, "loss": 0.5371, "step": 1511 }, { "epoch": 0.6083283041641521, "grad_norm": 3.726527595640385, "learning_rate": 1.999582951254219e-06, "loss": 0.4105, "step": 1512 }, { "epoch": 0.6087306376986522, "grad_norm": 4.0679286236770045, "learning_rate": 1.9961435569864366e-06, "loss": 0.524, "step": 1513 }, { "epoch": 0.6091329712331522, "grad_norm": 4.191017273712236, "learning_rate": 1.99270515646266e-06, "loss": 0.422, "step": 1514 }, { "epoch": 0.6095353047676524, "grad_norm": 4.549567455680612, "learning_rate": 1.9892677564643643e-06, "loss": 0.5753, "step": 1515 }, { "epoch": 0.6099376383021525, "grad_norm": 4.239380717200701, "learning_rate": 1.985831363771052e-06, "loss": 0.5251, "step": 1516 }, { "epoch": 0.6103399718366526, "grad_norm": 3.8382650524932433, "learning_rate": 1.9823959851602387e-06, "loss": 0.6749, "step": 1517 }, { "epoch": 0.6107423053711527, "grad_norm": 4.058628419689791, "learning_rate": 1.9789616274074375e-06, "loss": 0.5202, "step": 1518 }, { "epoch": 0.6111446389056527, "grad_norm": 4.1466696163681815, "learning_rate": 1.9755282972861515e-06, "loss": 0.4595, "step": 1519 }, { "epoch": 0.6115469724401529, "grad_norm": 4.193830162309673, "learning_rate": 1.9720960015678555e-06, "loss": 0.485, "step": 1520 }, { "epoch": 0.611949305974653, "grad_norm": 4.239153001501294, "learning_rate": 1.9686647470219835e-06, "loss": 0.6782, "step": 1521 }, { "epoch": 0.6123516395091531, "grad_norm": 4.138990574689833, "learning_rate": 1.965234540415918e-06, "loss": 0.6284, "step": 1522 }, { "epoch": 0.6127539730436532, "grad_norm": 4.079634170381014, "learning_rate": 1.9618053885149715e-06, "loss": 0.4656, "step": 1523 }, { "epoch": 0.6131563065781533, "grad_norm": 3.9398678794833204, "learning_rate": 1.9583772980823797e-06, "loss": 0.4345, "step": 1524 }, { "epoch": 0.6135586401126534, "grad_norm": 3.8999859116475237, "learning_rate": 1.9549502758792825e-06, "loss": 0.4463, "step": 1525 }, { "epoch": 0.6139609736471535, "grad_norm": 3.8236374179780652, "learning_rate": 1.9515243286647155e-06, "loss": 0.5032, "step": 1526 }, { "epoch": 0.6143633071816536, "grad_norm": 3.9105069856639787, "learning_rate": 1.9480994631955904e-06, "loss": 0.5249, "step": 1527 }, { "epoch": 0.6147656407161537, "grad_norm": 4.090697174474606, "learning_rate": 1.944675686226688e-06, "loss": 0.5986, "step": 1528 }, { "epoch": 0.6151679742506538, "grad_norm": 4.026970204591501, "learning_rate": 1.941253004510642e-06, "loss": 0.5695, "step": 1529 }, { "epoch": 0.6155703077851539, "grad_norm": 3.649099842703245, "learning_rate": 1.937831424797926e-06, "loss": 0.436, "step": 1530 }, { "epoch": 0.615972641319654, "grad_norm": 3.8376284628790773, "learning_rate": 1.9344109538368388e-06, "loss": 0.4986, "step": 1531 }, { "epoch": 0.616374974854154, "grad_norm": 4.048438236221887, "learning_rate": 1.9309915983734933e-06, "loss": 0.5608, "step": 1532 }, { "epoch": 0.6167773083886542, "grad_norm": 4.036284880877361, "learning_rate": 1.9275733651518038e-06, "loss": 0.552, "step": 1533 }, { "epoch": 0.6171796419231543, "grad_norm": 4.098386930913388, "learning_rate": 1.9241562609134694e-06, "loss": 0.5564, "step": 1534 }, { "epoch": 0.6175819754576544, "grad_norm": 4.4349116098990065, "learning_rate": 1.9207402923979633e-06, "loss": 0.5937, "step": 1535 }, { "epoch": 0.6179843089921545, "grad_norm": 3.9955554180753556, "learning_rate": 1.9173254663425165e-06, "loss": 0.4773, "step": 1536 }, { "epoch": 0.6183866425266547, "grad_norm": 3.7895980004495553, "learning_rate": 1.9139117894821107e-06, "loss": 0.5117, "step": 1537 }, { "epoch": 0.6187889760611547, "grad_norm": 4.503044474467485, "learning_rate": 1.910499268549459e-06, "loss": 0.5728, "step": 1538 }, { "epoch": 0.6191913095956548, "grad_norm": 4.5048282169496865, "learning_rate": 1.9070879102749944e-06, "loss": 0.4952, "step": 1539 }, { "epoch": 0.6195936431301549, "grad_norm": 3.6152226921367516, "learning_rate": 1.9036777213868567e-06, "loss": 0.4901, "step": 1540 }, { "epoch": 0.619995976664655, "grad_norm": 4.56665728701204, "learning_rate": 1.9002687086108808e-06, "loss": 0.6217, "step": 1541 }, { "epoch": 0.6203983101991551, "grad_norm": 3.8747261446432972, "learning_rate": 1.8968608786705803e-06, "loss": 0.4346, "step": 1542 }, { "epoch": 0.6208006437336552, "grad_norm": 3.553549231891504, "learning_rate": 1.8934542382871365e-06, "loss": 0.4086, "step": 1543 }, { "epoch": 0.6212029772681553, "grad_norm": 4.273007467466175, "learning_rate": 1.8900487941793862e-06, "loss": 0.5589, "step": 1544 }, { "epoch": 0.6216053108026554, "grad_norm": 4.033041498302765, "learning_rate": 1.8866445530638034e-06, "loss": 0.5258, "step": 1545 }, { "epoch": 0.6220076443371555, "grad_norm": 3.847764801262625, "learning_rate": 1.883241521654492e-06, "loss": 0.5286, "step": 1546 }, { "epoch": 0.6224099778716556, "grad_norm": 4.141005318996514, "learning_rate": 1.87983970666317e-06, "loss": 0.5428, "step": 1547 }, { "epoch": 0.6228123114061557, "grad_norm": 3.8545860722248286, "learning_rate": 1.8764391147991556e-06, "loss": 0.4504, "step": 1548 }, { "epoch": 0.6232146449406558, "grad_norm": 3.9064270159769543, "learning_rate": 1.8730397527693544e-06, "loss": 0.49, "step": 1549 }, { "epoch": 0.623616978475156, "grad_norm": 3.5305364890500903, "learning_rate": 1.8696416272782467e-06, "loss": 0.4747, "step": 1550 }, { "epoch": 0.624019312009656, "grad_norm": 4.367606237537799, "learning_rate": 1.866244745027874e-06, "loss": 0.6066, "step": 1551 }, { "epoch": 0.6244216455441561, "grad_norm": 3.778565429919078, "learning_rate": 1.8628491127178272e-06, "loss": 0.4401, "step": 1552 }, { "epoch": 0.6248239790786562, "grad_norm": 4.4484086477744675, "learning_rate": 1.8594547370452288e-06, "loss": 0.5702, "step": 1553 }, { "epoch": 0.6252263126131563, "grad_norm": 4.133384847499752, "learning_rate": 1.8560616247047253e-06, "loss": 0.4958, "step": 1554 }, { "epoch": 0.6256286461476565, "grad_norm": 4.481750156438252, "learning_rate": 1.8526697823884721e-06, "loss": 0.5982, "step": 1555 }, { "epoch": 0.6260309796821565, "grad_norm": 4.679844131200749, "learning_rate": 1.849279216786117e-06, "loss": 0.5699, "step": 1556 }, { "epoch": 0.6264333132166566, "grad_norm": 4.035894316525195, "learning_rate": 1.8458899345847936e-06, "loss": 0.6036, "step": 1557 }, { "epoch": 0.6268356467511567, "grad_norm": 3.3784184252716836, "learning_rate": 1.8425019424690993e-06, "loss": 0.3667, "step": 1558 }, { "epoch": 0.6272379802856568, "grad_norm": 3.7155353324243325, "learning_rate": 1.839115247121091e-06, "loss": 0.4744, "step": 1559 }, { "epoch": 0.627640313820157, "grad_norm": 4.062605549457995, "learning_rate": 1.8357298552202676e-06, "loss": 0.4846, "step": 1560 }, { "epoch": 0.628042647354657, "grad_norm": 3.968579655925073, "learning_rate": 1.832345773443557e-06, "loss": 0.4563, "step": 1561 }, { "epoch": 0.6284449808891571, "grad_norm": 3.859404877482336, "learning_rate": 1.8289630084653009e-06, "loss": 0.5339, "step": 1562 }, { "epoch": 0.6288473144236573, "grad_norm": 4.32493582991546, "learning_rate": 1.825581566957247e-06, "loss": 0.5206, "step": 1563 }, { "epoch": 0.6292496479581573, "grad_norm": 3.7620622738243474, "learning_rate": 1.8222014555885303e-06, "loss": 0.5187, "step": 1564 }, { "epoch": 0.6296519814926574, "grad_norm": 3.9245871423313616, "learning_rate": 1.8188226810256647e-06, "loss": 0.5302, "step": 1565 }, { "epoch": 0.6300543150271575, "grad_norm": 4.0730169054317305, "learning_rate": 1.8154452499325273e-06, "loss": 0.482, "step": 1566 }, { "epoch": 0.6304566485616576, "grad_norm": 4.0887356034823314, "learning_rate": 1.8120691689703423e-06, "loss": 0.6323, "step": 1567 }, { "epoch": 0.6308589820961578, "grad_norm": 4.134013360882183, "learning_rate": 1.8086944447976745e-06, "loss": 0.4215, "step": 1568 }, { "epoch": 0.6312613156306578, "grad_norm": 4.267968482401992, "learning_rate": 1.805321084070412e-06, "loss": 0.6421, "step": 1569 }, { "epoch": 0.6316636491651579, "grad_norm": 4.337432367486679, "learning_rate": 1.8019490934417533e-06, "loss": 0.614, "step": 1570 }, { "epoch": 0.632065982699658, "grad_norm": 4.325604571245914, "learning_rate": 1.7985784795621946e-06, "loss": 0.6082, "step": 1571 }, { "epoch": 0.6324683162341581, "grad_norm": 4.7255506567581005, "learning_rate": 1.7952092490795165e-06, "loss": 0.5848, "step": 1572 }, { "epoch": 0.6328706497686583, "grad_norm": 3.9917358158112513, "learning_rate": 1.7918414086387723e-06, "loss": 0.51, "step": 1573 }, { "epoch": 0.6332729833031583, "grad_norm": 3.9759190309543806, "learning_rate": 1.788474964882273e-06, "loss": 0.4348, "step": 1574 }, { "epoch": 0.6336753168376584, "grad_norm": 3.9060781229407673, "learning_rate": 1.7851099244495761e-06, "loss": 0.5328, "step": 1575 }, { "epoch": 0.6340776503721586, "grad_norm": 3.794629121755811, "learning_rate": 1.7817462939774683e-06, "loss": 0.5406, "step": 1576 }, { "epoch": 0.6344799839066586, "grad_norm": 4.434204491241485, "learning_rate": 1.778384080099959e-06, "loss": 0.6955, "step": 1577 }, { "epoch": 0.6348823174411587, "grad_norm": 4.243047298973574, "learning_rate": 1.775023289448261e-06, "loss": 0.6021, "step": 1578 }, { "epoch": 0.6352846509756588, "grad_norm": 3.906621293593377, "learning_rate": 1.7716639286507834e-06, "loss": 0.5017, "step": 1579 }, { "epoch": 0.6356869845101589, "grad_norm": 4.433075604901836, "learning_rate": 1.7683060043331102e-06, "loss": 0.6304, "step": 1580 }, { "epoch": 0.6360893180446591, "grad_norm": 4.0324680233039425, "learning_rate": 1.764949523117997e-06, "loss": 0.4593, "step": 1581 }, { "epoch": 0.6364916515791591, "grad_norm": 4.861200946299266, "learning_rate": 1.7615944916253511e-06, "loss": 0.5755, "step": 1582 }, { "epoch": 0.6368939851136592, "grad_norm": 3.4859779578870462, "learning_rate": 1.7582409164722219e-06, "loss": 0.4494, "step": 1583 }, { "epoch": 0.6372963186481593, "grad_norm": 4.394364956739334, "learning_rate": 1.7548888042727835e-06, "loss": 0.5756, "step": 1584 }, { "epoch": 0.6376986521826594, "grad_norm": 4.474860851761582, "learning_rate": 1.751538161638327e-06, "loss": 0.5217, "step": 1585 }, { "epoch": 0.6381009857171596, "grad_norm": 4.665357328027418, "learning_rate": 1.748188995177245e-06, "loss": 0.5599, "step": 1586 }, { "epoch": 0.6385033192516596, "grad_norm": 3.186769209343777, "learning_rate": 1.744841311495019e-06, "loss": 0.3409, "step": 1587 }, { "epoch": 0.6389056527861597, "grad_norm": 3.3506177658392744, "learning_rate": 1.7414951171942057e-06, "loss": 0.5176, "step": 1588 }, { "epoch": 0.6393079863206599, "grad_norm": 3.4136910306135935, "learning_rate": 1.7381504188744218e-06, "loss": 0.4464, "step": 1589 }, { "epoch": 0.6397103198551599, "grad_norm": 3.4917018490110467, "learning_rate": 1.7348072231323382e-06, "loss": 0.438, "step": 1590 }, { "epoch": 0.6401126533896601, "grad_norm": 3.421803275946972, "learning_rate": 1.7314655365616584e-06, "loss": 0.4189, "step": 1591 }, { "epoch": 0.6405149869241601, "grad_norm": 4.433987296079437, "learning_rate": 1.7281253657531123e-06, "loss": 0.606, "step": 1592 }, { "epoch": 0.6409173204586602, "grad_norm": 3.6737703307949983, "learning_rate": 1.7247867172944367e-06, "loss": 0.4349, "step": 1593 }, { "epoch": 0.6413196539931604, "grad_norm": 3.8221357123925914, "learning_rate": 1.721449597770369e-06, "loss": 0.484, "step": 1594 }, { "epoch": 0.6417219875276604, "grad_norm": 4.069693321778052, "learning_rate": 1.718114013762631e-06, "loss": 0.6507, "step": 1595 }, { "epoch": 0.6421243210621606, "grad_norm": 4.268934323506783, "learning_rate": 1.714779971849915e-06, "loss": 0.498, "step": 1596 }, { "epoch": 0.6425266545966606, "grad_norm": 4.426082622301671, "learning_rate": 1.7114474786078716e-06, "loss": 0.6755, "step": 1597 }, { "epoch": 0.6429289881311607, "grad_norm": 3.6655062613985288, "learning_rate": 1.708116540609098e-06, "loss": 0.5078, "step": 1598 }, { "epoch": 0.6433313216656609, "grad_norm": 4.096742196755438, "learning_rate": 1.7047871644231226e-06, "loss": 0.5085, "step": 1599 }, { "epoch": 0.6437336552001609, "grad_norm": 4.341802218643631, "learning_rate": 1.7014593566163949e-06, "loss": 0.6022, "step": 1600 }, { "epoch": 0.644135988734661, "grad_norm": 4.167303639683251, "learning_rate": 1.6981331237522716e-06, "loss": 0.595, "step": 1601 }, { "epoch": 0.6445383222691611, "grad_norm": 3.860753980307849, "learning_rate": 1.6948084723910002e-06, "loss": 0.4189, "step": 1602 }, { "epoch": 0.6449406558036612, "grad_norm": 3.801435117916446, "learning_rate": 1.6914854090897124e-06, "loss": 0.4691, "step": 1603 }, { "epoch": 0.6453429893381614, "grad_norm": 4.309844638484278, "learning_rate": 1.6881639404024063e-06, "loss": 0.4802, "step": 1604 }, { "epoch": 0.6457453228726614, "grad_norm": 3.8218645060402285, "learning_rate": 1.6848440728799346e-06, "loss": 0.6087, "step": 1605 }, { "epoch": 0.6461476564071615, "grad_norm": 4.318515117319728, "learning_rate": 1.6815258130699938e-06, "loss": 0.6644, "step": 1606 }, { "epoch": 0.6465499899416617, "grad_norm": 3.8706029461346114, "learning_rate": 1.6782091675171064e-06, "loss": 0.5314, "step": 1607 }, { "epoch": 0.6469523234761617, "grad_norm": 4.588786654868146, "learning_rate": 1.6748941427626142e-06, "loss": 0.5713, "step": 1608 }, { "epoch": 0.6473546570106619, "grad_norm": 4.206226833574013, "learning_rate": 1.671580745344661e-06, "loss": 0.5207, "step": 1609 }, { "epoch": 0.6477569905451619, "grad_norm": 4.1241193689590565, "learning_rate": 1.6682689817981827e-06, "loss": 0.5325, "step": 1610 }, { "epoch": 0.648159324079662, "grad_norm": 4.013331656155903, "learning_rate": 1.6649588586548888e-06, "loss": 0.5593, "step": 1611 }, { "epoch": 0.6485616576141622, "grad_norm": 4.701503684635364, "learning_rate": 1.6616503824432573e-06, "loss": 0.6878, "step": 1612 }, { "epoch": 0.6489639911486622, "grad_norm": 4.047494584818948, "learning_rate": 1.658343559688516e-06, "loss": 0.5131, "step": 1613 }, { "epoch": 0.6493663246831624, "grad_norm": 4.0317139401208575, "learning_rate": 1.6550383969126341e-06, "loss": 0.512, "step": 1614 }, { "epoch": 0.6497686582176624, "grad_norm": 3.9682317404566034, "learning_rate": 1.6517349006343025e-06, "loss": 0.5921, "step": 1615 }, { "epoch": 0.6501709917521625, "grad_norm": 3.824331386002604, "learning_rate": 1.6484330773689292e-06, "loss": 0.4712, "step": 1616 }, { "epoch": 0.6505733252866627, "grad_norm": 3.7019715035054204, "learning_rate": 1.645132933628621e-06, "loss": 0.4426, "step": 1617 }, { "epoch": 0.6509756588211627, "grad_norm": 4.0459494986124245, "learning_rate": 1.6418344759221734e-06, "loss": 0.6728, "step": 1618 }, { "epoch": 0.6513779923556628, "grad_norm": 4.154812270453969, "learning_rate": 1.6385377107550543e-06, "loss": 0.559, "step": 1619 }, { "epoch": 0.651780325890163, "grad_norm": 4.703387446268752, "learning_rate": 1.6352426446293952e-06, "loss": 0.5548, "step": 1620 }, { "epoch": 0.652182659424663, "grad_norm": 3.894555312354832, "learning_rate": 1.631949284043976e-06, "loss": 0.4987, "step": 1621 }, { "epoch": 0.6525849929591632, "grad_norm": 4.281016820029128, "learning_rate": 1.628657635494213e-06, "loss": 0.6007, "step": 1622 }, { "epoch": 0.6529873264936632, "grad_norm": 3.708021853167261, "learning_rate": 1.625367705472147e-06, "loss": 0.5041, "step": 1623 }, { "epoch": 0.6533896600281633, "grad_norm": 4.184667272519736, "learning_rate": 1.6220795004664264e-06, "loss": 0.563, "step": 1624 }, { "epoch": 0.6537919935626635, "grad_norm": 4.386544791097414, "learning_rate": 1.6187930269623001e-06, "loss": 0.6591, "step": 1625 }, { "epoch": 0.6541943270971635, "grad_norm": 4.095997806253382, "learning_rate": 1.6155082914416018e-06, "loss": 0.5122, "step": 1626 }, { "epoch": 0.6545966606316637, "grad_norm": 3.97582656513418, "learning_rate": 1.6122253003827353e-06, "loss": 0.5805, "step": 1627 }, { "epoch": 0.6549989941661637, "grad_norm": 3.990901166079429, "learning_rate": 1.6089440602606675e-06, "loss": 0.569, "step": 1628 }, { "epoch": 0.6554013277006638, "grad_norm": 3.8624368116988403, "learning_rate": 1.6056645775469077e-06, "loss": 0.4729, "step": 1629 }, { "epoch": 0.655803661235164, "grad_norm": 4.529362920268054, "learning_rate": 1.6023868587095016e-06, "loss": 0.5482, "step": 1630 }, { "epoch": 0.656205994769664, "grad_norm": 4.017565552354934, "learning_rate": 1.599110910213016e-06, "loss": 0.4916, "step": 1631 }, { "epoch": 0.6566083283041642, "grad_norm": 4.4182156313212575, "learning_rate": 1.595836738518527e-06, "loss": 0.5792, "step": 1632 }, { "epoch": 0.6570106618386643, "grad_norm": 3.5677962985553524, "learning_rate": 1.592564350083603e-06, "loss": 0.4697, "step": 1633 }, { "epoch": 0.6574129953731643, "grad_norm": 3.909711067921425, "learning_rate": 1.5892937513622977e-06, "loss": 0.4383, "step": 1634 }, { "epoch": 0.6578153289076645, "grad_norm": 4.098825788465001, "learning_rate": 1.5860249488051355e-06, "loss": 0.4745, "step": 1635 }, { "epoch": 0.6582176624421645, "grad_norm": 3.2004894026603616, "learning_rate": 1.5827579488590977e-06, "loss": 0.4775, "step": 1636 }, { "epoch": 0.6586199959766647, "grad_norm": 3.970341598297501, "learning_rate": 1.5794927579676089e-06, "loss": 0.5471, "step": 1637 }, { "epoch": 0.6590223295111648, "grad_norm": 4.241718162619275, "learning_rate": 1.5762293825705276e-06, "loss": 0.6544, "step": 1638 }, { "epoch": 0.6594246630456648, "grad_norm": 3.924819583311759, "learning_rate": 1.5729678291041318e-06, "loss": 0.5368, "step": 1639 }, { "epoch": 0.659826996580165, "grad_norm": 3.8382551527997397, "learning_rate": 1.5697081040011043e-06, "loss": 0.5121, "step": 1640 }, { "epoch": 0.660229330114665, "grad_norm": 3.6682266397110337, "learning_rate": 1.5664502136905236e-06, "loss": 0.4353, "step": 1641 }, { "epoch": 0.6606316636491651, "grad_norm": 3.9060845453518143, "learning_rate": 1.563194164597848e-06, "loss": 0.4814, "step": 1642 }, { "epoch": 0.6610339971836653, "grad_norm": 3.985541657912861, "learning_rate": 1.5599399631449052e-06, "loss": 0.4643, "step": 1643 }, { "epoch": 0.6614363307181653, "grad_norm": 4.246558106503365, "learning_rate": 1.5566876157498797e-06, "loss": 0.5457, "step": 1644 }, { "epoch": 0.6618386642526655, "grad_norm": 4.538356504223212, "learning_rate": 1.5534371288272992e-06, "loss": 0.5586, "step": 1645 }, { "epoch": 0.6622409977871656, "grad_norm": 3.512817123486096, "learning_rate": 1.5501885087880187e-06, "loss": 0.3777, "step": 1646 }, { "epoch": 0.6626433313216656, "grad_norm": 3.611315735282059, "learning_rate": 1.546941762039216e-06, "loss": 0.4559, "step": 1647 }, { "epoch": 0.6630456648561658, "grad_norm": 3.804900233889627, "learning_rate": 1.5436968949843705e-06, "loss": 0.5593, "step": 1648 }, { "epoch": 0.6634479983906658, "grad_norm": 4.290281020900778, "learning_rate": 1.5404539140232572e-06, "loss": 0.5433, "step": 1649 }, { "epoch": 0.663850331925166, "grad_norm": 4.261399800624298, "learning_rate": 1.5372128255519298e-06, "loss": 0.5669, "step": 1650 }, { "epoch": 0.6642526654596661, "grad_norm": 4.8206169191275094, "learning_rate": 1.533973635962708e-06, "loss": 0.535, "step": 1651 }, { "epoch": 0.6646549989941661, "grad_norm": 3.5705866846183882, "learning_rate": 1.5307363516441693e-06, "loss": 0.4436, "step": 1652 }, { "epoch": 0.6650573325286663, "grad_norm": 4.015284104793611, "learning_rate": 1.5275009789811317e-06, "loss": 0.4606, "step": 1653 }, { "epoch": 0.6654596660631663, "grad_norm": 3.746888831379427, "learning_rate": 1.5242675243546435e-06, "loss": 0.5344, "step": 1654 }, { "epoch": 0.6658619995976665, "grad_norm": 5.350374969998737, "learning_rate": 1.5210359941419698e-06, "loss": 0.7187, "step": 1655 }, { "epoch": 0.6662643331321666, "grad_norm": 4.992414873463243, "learning_rate": 1.5178063947165795e-06, "loss": 0.5597, "step": 1656 }, { "epoch": 0.6666666666666666, "grad_norm": 3.9536138723501413, "learning_rate": 1.514578732448135e-06, "loss": 0.5345, "step": 1657 }, { "epoch": 0.6670690002011668, "grad_norm": 4.023274067891729, "learning_rate": 1.5113530137024774e-06, "loss": 0.4893, "step": 1658 }, { "epoch": 0.6674713337356669, "grad_norm": 4.108018429855622, "learning_rate": 1.5081292448416158e-06, "loss": 0.645, "step": 1659 }, { "epoch": 0.667873667270167, "grad_norm": 3.817031046521698, "learning_rate": 1.5049074322237097e-06, "loss": 0.4838, "step": 1660 }, { "epoch": 0.6682760008046671, "grad_norm": 4.10734795999481, "learning_rate": 1.5016875822030651e-06, "loss": 0.5853, "step": 1661 }, { "epoch": 0.6686783343391671, "grad_norm": 4.104736058188134, "learning_rate": 1.4984697011301142e-06, "loss": 0.5032, "step": 1662 }, { "epoch": 0.6690806678736673, "grad_norm": 4.213736489738209, "learning_rate": 1.4952537953514078e-06, "loss": 0.4234, "step": 1663 }, { "epoch": 0.6694830014081674, "grad_norm": 4.592368631770386, "learning_rate": 1.4920398712095985e-06, "loss": 0.5323, "step": 1664 }, { "epoch": 0.6698853349426674, "grad_norm": 4.593839464020305, "learning_rate": 1.4888279350434325e-06, "loss": 0.5821, "step": 1665 }, { "epoch": 0.6702876684771676, "grad_norm": 3.881133737659202, "learning_rate": 1.485617993187735e-06, "loss": 0.5224, "step": 1666 }, { "epoch": 0.6706900020116676, "grad_norm": 4.088068946795046, "learning_rate": 1.4824100519733985e-06, "loss": 0.556, "step": 1667 }, { "epoch": 0.6710923355461678, "grad_norm": 4.328495324161391, "learning_rate": 1.4792041177273668e-06, "loss": 0.6097, "step": 1668 }, { "epoch": 0.6714946690806679, "grad_norm": 4.220316709489982, "learning_rate": 1.4760001967726283e-06, "loss": 0.6015, "step": 1669 }, { "epoch": 0.6718970026151679, "grad_norm": 3.788396547796328, "learning_rate": 1.472798295428199e-06, "loss": 0.5863, "step": 1670 }, { "epoch": 0.6722993361496681, "grad_norm": 3.8086553750558494, "learning_rate": 1.4695984200091129e-06, "loss": 0.5155, "step": 1671 }, { "epoch": 0.6727016696841682, "grad_norm": 3.792467737324882, "learning_rate": 1.4664005768264084e-06, "loss": 0.4517, "step": 1672 }, { "epoch": 0.6731040032186683, "grad_norm": 4.528886592192105, "learning_rate": 1.4632047721871134e-06, "loss": 0.5514, "step": 1673 }, { "epoch": 0.6735063367531684, "grad_norm": 3.687134247150264, "learning_rate": 1.4600110123942376e-06, "loss": 0.5765, "step": 1674 }, { "epoch": 0.6739086702876684, "grad_norm": 3.651752055706122, "learning_rate": 1.4568193037467574e-06, "loss": 0.4434, "step": 1675 }, { "epoch": 0.6743110038221686, "grad_norm": 4.196555677667497, "learning_rate": 1.453629652539604e-06, "loss": 0.5832, "step": 1676 }, { "epoch": 0.6747133373566687, "grad_norm": 3.3768494467370047, "learning_rate": 1.4504420650636485e-06, "loss": 0.4094, "step": 1677 }, { "epoch": 0.6751156708911688, "grad_norm": 3.9350013310511582, "learning_rate": 1.4472565476056952e-06, "loss": 0.4875, "step": 1678 }, { "epoch": 0.6755180044256689, "grad_norm": 3.9364895166754383, "learning_rate": 1.444073106448462e-06, "loss": 0.626, "step": 1679 }, { "epoch": 0.6759203379601689, "grad_norm": 4.218774138421791, "learning_rate": 1.4408917478705748e-06, "loss": 0.5318, "step": 1680 }, { "epoch": 0.6763226714946691, "grad_norm": 3.557852927029882, "learning_rate": 1.4377124781465517e-06, "loss": 0.4564, "step": 1681 }, { "epoch": 0.6767250050291692, "grad_norm": 4.131292301591179, "learning_rate": 1.4345353035467888e-06, "loss": 0.5458, "step": 1682 }, { "epoch": 0.6771273385636692, "grad_norm": 4.261120314757306, "learning_rate": 1.4313602303375523e-06, "loss": 0.6402, "step": 1683 }, { "epoch": 0.6775296720981694, "grad_norm": 4.027439558487913, "learning_rate": 1.4281872647809628e-06, "loss": 0.5762, "step": 1684 }, { "epoch": 0.6779320056326695, "grad_norm": 3.9491401859678605, "learning_rate": 1.4250164131349858e-06, "loss": 0.4807, "step": 1685 }, { "epoch": 0.6783343391671696, "grad_norm": 3.60940986670999, "learning_rate": 1.421847681653414e-06, "loss": 0.3659, "step": 1686 }, { "epoch": 0.6787366727016697, "grad_norm": 3.5822407819650515, "learning_rate": 1.4186810765858616e-06, "loss": 0.5165, "step": 1687 }, { "epoch": 0.6791390062361697, "grad_norm": 4.161182758555724, "learning_rate": 1.4155166041777483e-06, "loss": 0.495, "step": 1688 }, { "epoch": 0.6795413397706699, "grad_norm": 4.076431248542256, "learning_rate": 1.4123542706702868e-06, "loss": 0.4569, "step": 1689 }, { "epoch": 0.67994367330517, "grad_norm": 3.401944784975411, "learning_rate": 1.409194082300473e-06, "loss": 0.4661, "step": 1690 }, { "epoch": 0.6803460068396701, "grad_norm": 4.053896920545759, "learning_rate": 1.4060360453010691e-06, "loss": 0.6945, "step": 1691 }, { "epoch": 0.6807483403741702, "grad_norm": 3.6315876209619726, "learning_rate": 1.4028801659005975e-06, "loss": 0.5088, "step": 1692 }, { "epoch": 0.6811506739086702, "grad_norm": 3.9033263695876617, "learning_rate": 1.3997264503233222e-06, "loss": 0.5535, "step": 1693 }, { "epoch": 0.6815530074431704, "grad_norm": 3.6520508454132568, "learning_rate": 1.3965749047892421e-06, "loss": 0.4627, "step": 1694 }, { "epoch": 0.6819553409776705, "grad_norm": 4.253145100591827, "learning_rate": 1.3934255355140735e-06, "loss": 0.6059, "step": 1695 }, { "epoch": 0.6823576745121706, "grad_norm": 3.6442609114551168, "learning_rate": 1.3902783487092436e-06, "loss": 0.437, "step": 1696 }, { "epoch": 0.6827600080466707, "grad_norm": 3.6843000132081083, "learning_rate": 1.387133350581873e-06, "loss": 0.4741, "step": 1697 }, { "epoch": 0.6831623415811708, "grad_norm": 4.082924701084028, "learning_rate": 1.383990547334767e-06, "loss": 0.5523, "step": 1698 }, { "epoch": 0.6835646751156709, "grad_norm": 4.525193517476986, "learning_rate": 1.3808499451663998e-06, "loss": 0.5733, "step": 1699 }, { "epoch": 0.683967008650171, "grad_norm": 3.8434172838289142, "learning_rate": 1.377711550270907e-06, "loss": 0.4042, "step": 1700 }, { "epoch": 0.684369342184671, "grad_norm": 3.475844585621674, "learning_rate": 1.3745753688380692e-06, "loss": 0.4492, "step": 1701 }, { "epoch": 0.6847716757191712, "grad_norm": 4.106466165584271, "learning_rate": 1.3714414070533022e-06, "loss": 0.4795, "step": 1702 }, { "epoch": 0.6851740092536713, "grad_norm": 3.9408796988863477, "learning_rate": 1.368309671097645e-06, "loss": 0.4586, "step": 1703 }, { "epoch": 0.6855763427881714, "grad_norm": 3.892878635048642, "learning_rate": 1.3651801671477435e-06, "loss": 0.5256, "step": 1704 }, { "epoch": 0.6859786763226715, "grad_norm": 3.602589673252341, "learning_rate": 1.3620529013758444e-06, "loss": 0.4474, "step": 1705 }, { "epoch": 0.6863810098571715, "grad_norm": 4.29050518808401, "learning_rate": 1.3589278799497802e-06, "loss": 0.4799, "step": 1706 }, { "epoch": 0.6867833433916717, "grad_norm": 4.530188596691838, "learning_rate": 1.3558051090329536e-06, "loss": 0.6044, "step": 1707 }, { "epoch": 0.6871856769261718, "grad_norm": 3.8842771157381932, "learning_rate": 1.352684594784333e-06, "loss": 0.4759, "step": 1708 }, { "epoch": 0.6875880104606719, "grad_norm": 4.206785571596244, "learning_rate": 1.3495663433584327e-06, "loss": 0.58, "step": 1709 }, { "epoch": 0.687990343995172, "grad_norm": 4.648975138578508, "learning_rate": 1.3464503609053053e-06, "loss": 0.5508, "step": 1710 }, { "epoch": 0.6883926775296721, "grad_norm": 4.513492173355037, "learning_rate": 1.3433366535705292e-06, "loss": 0.6606, "step": 1711 }, { "epoch": 0.6887950110641722, "grad_norm": 4.149085872177257, "learning_rate": 1.3402252274951954e-06, "loss": 0.5438, "step": 1712 }, { "epoch": 0.6891973445986723, "grad_norm": 4.3080582119504935, "learning_rate": 1.3371160888158935e-06, "loss": 0.7277, "step": 1713 }, { "epoch": 0.6895996781331724, "grad_norm": 4.195944255893043, "learning_rate": 1.3340092436647045e-06, "loss": 0.5682, "step": 1714 }, { "epoch": 0.6900020116676725, "grad_norm": 4.22717459587227, "learning_rate": 1.330904698169184e-06, "loss": 0.5394, "step": 1715 }, { "epoch": 0.6904043452021726, "grad_norm": 3.8328998433470334, "learning_rate": 1.3278024584523542e-06, "loss": 0.519, "step": 1716 }, { "epoch": 0.6908066787366727, "grad_norm": 3.809809815071977, "learning_rate": 1.324702530632686e-06, "loss": 0.4609, "step": 1717 }, { "epoch": 0.6912090122711728, "grad_norm": 4.408644772842878, "learning_rate": 1.3216049208240945e-06, "loss": 0.5259, "step": 1718 }, { "epoch": 0.6916113458056729, "grad_norm": 4.351040955206057, "learning_rate": 1.3185096351359211e-06, "loss": 0.531, "step": 1719 }, { "epoch": 0.692013679340173, "grad_norm": 3.8862458710646135, "learning_rate": 1.3154166796729245e-06, "loss": 0.5521, "step": 1720 }, { "epoch": 0.6924160128746731, "grad_norm": 3.8732946833809283, "learning_rate": 1.3123260605352647e-06, "loss": 0.4054, "step": 1721 }, { "epoch": 0.6928183464091732, "grad_norm": 4.014263742357025, "learning_rate": 1.3092377838184984e-06, "loss": 0.5096, "step": 1722 }, { "epoch": 0.6932206799436733, "grad_norm": 3.8457014699738306, "learning_rate": 1.3061518556135572e-06, "loss": 0.5361, "step": 1723 }, { "epoch": 0.6936230134781735, "grad_norm": 4.700764158732767, "learning_rate": 1.3030682820067447e-06, "loss": 0.6363, "step": 1724 }, { "epoch": 0.6940253470126735, "grad_norm": 4.056861341126563, "learning_rate": 1.2999870690797207e-06, "loss": 0.5499, "step": 1725 }, { "epoch": 0.6944276805471736, "grad_norm": 3.824223111444158, "learning_rate": 1.2969082229094853e-06, "loss": 0.5049, "step": 1726 }, { "epoch": 0.6948300140816737, "grad_norm": 3.8066989110426332, "learning_rate": 1.2938317495683744e-06, "loss": 0.487, "step": 1727 }, { "epoch": 0.6952323476161738, "grad_norm": 4.382851529690001, "learning_rate": 1.2907576551240425e-06, "loss": 0.5016, "step": 1728 }, { "epoch": 0.695634681150674, "grad_norm": 4.2709658357684415, "learning_rate": 1.2876859456394537e-06, "loss": 0.5287, "step": 1729 }, { "epoch": 0.696037014685174, "grad_norm": 4.10428141385855, "learning_rate": 1.284616627172865e-06, "loss": 0.5048, "step": 1730 }, { "epoch": 0.6964393482196741, "grad_norm": 3.9240197132726133, "learning_rate": 1.281549705777821e-06, "loss": 0.5834, "step": 1731 }, { "epoch": 0.6968416817541742, "grad_norm": 3.917832710676123, "learning_rate": 1.278485187503137e-06, "loss": 0.4978, "step": 1732 }, { "epoch": 0.6972440152886743, "grad_norm": 3.9886241574028123, "learning_rate": 1.2754230783928907e-06, "loss": 0.5046, "step": 1733 }, { "epoch": 0.6976463488231744, "grad_norm": 3.8252585006207744, "learning_rate": 1.2723633844864036e-06, "loss": 0.4337, "step": 1734 }, { "epoch": 0.6980486823576745, "grad_norm": 4.327909530843214, "learning_rate": 1.2693061118182399e-06, "loss": 0.5663, "step": 1735 }, { "epoch": 0.6984510158921746, "grad_norm": 3.9729308104472225, "learning_rate": 1.2662512664181826e-06, "loss": 0.4753, "step": 1736 }, { "epoch": 0.6988533494266748, "grad_norm": 3.4193536839822434, "learning_rate": 1.2631988543112314e-06, "loss": 0.3984, "step": 1737 }, { "epoch": 0.6992556829611748, "grad_norm": 3.199663843957577, "learning_rate": 1.2601488815175867e-06, "loss": 0.3615, "step": 1738 }, { "epoch": 0.6996580164956749, "grad_norm": 4.275095481044542, "learning_rate": 1.257101354052635e-06, "loss": 0.635, "step": 1739 }, { "epoch": 0.700060350030175, "grad_norm": 4.154894699295693, "learning_rate": 1.2540562779269421e-06, "loss": 0.468, "step": 1740 }, { "epoch": 0.7004626835646751, "grad_norm": 4.21271106717443, "learning_rate": 1.2510136591462399e-06, "loss": 0.557, "step": 1741 }, { "epoch": 0.7008650170991753, "grad_norm": 3.9488032178250276, "learning_rate": 1.2479735037114118e-06, "loss": 0.4743, "step": 1742 }, { "epoch": 0.7012673506336753, "grad_norm": 3.6437407466020453, "learning_rate": 1.2449358176184848e-06, "loss": 0.4958, "step": 1743 }, { "epoch": 0.7016696841681754, "grad_norm": 4.011291060246906, "learning_rate": 1.2419006068586125e-06, "loss": 0.5109, "step": 1744 }, { "epoch": 0.7020720177026755, "grad_norm": 3.7545390991024026, "learning_rate": 1.2388678774180698e-06, "loss": 0.5269, "step": 1745 }, { "epoch": 0.7024743512371756, "grad_norm": 3.821504931037996, "learning_rate": 1.2358376352782358e-06, "loss": 0.4905, "step": 1746 }, { "epoch": 0.7028766847716758, "grad_norm": 4.3580981207011655, "learning_rate": 1.232809886415586e-06, "loss": 0.4904, "step": 1747 }, { "epoch": 0.7032790183061758, "grad_norm": 3.539055820090707, "learning_rate": 1.2297846368016748e-06, "loss": 0.4124, "step": 1748 }, { "epoch": 0.7036813518406759, "grad_norm": 3.9844890147713254, "learning_rate": 1.2267618924031317e-06, "loss": 0.5321, "step": 1749 }, { "epoch": 0.7040836853751761, "grad_norm": 4.178112023828336, "learning_rate": 1.2237416591816412e-06, "loss": 0.6362, "step": 1750 }, { "epoch": 0.7044860189096761, "grad_norm": 4.0472275399599695, "learning_rate": 1.220723943093939e-06, "loss": 0.5101, "step": 1751 }, { "epoch": 0.7048883524441762, "grad_norm": 4.312523214514084, "learning_rate": 1.2177087500917925e-06, "loss": 0.5277, "step": 1752 }, { "epoch": 0.7052906859786763, "grad_norm": 4.133768938615848, "learning_rate": 1.2146960861219954e-06, "loss": 0.5271, "step": 1753 }, { "epoch": 0.7056930195131764, "grad_norm": 4.430552507921185, "learning_rate": 1.211685957126353e-06, "loss": 0.711, "step": 1754 }, { "epoch": 0.7060953530476766, "grad_norm": 4.119451270646705, "learning_rate": 1.2086783690416704e-06, "loss": 0.4358, "step": 1755 }, { "epoch": 0.7064976865821766, "grad_norm": 4.295693476571627, "learning_rate": 1.205673327799743e-06, "loss": 0.5466, "step": 1756 }, { "epoch": 0.7069000201166767, "grad_norm": 4.312446284251359, "learning_rate": 1.2026708393273392e-06, "loss": 0.5861, "step": 1757 }, { "epoch": 0.7073023536511768, "grad_norm": 3.9109516304734733, "learning_rate": 1.1996709095461967e-06, "loss": 0.4782, "step": 1758 }, { "epoch": 0.7077046871856769, "grad_norm": 4.425022693894618, "learning_rate": 1.1966735443730052e-06, "loss": 0.5627, "step": 1759 }, { "epoch": 0.7081070207201771, "grad_norm": 4.259706435608531, "learning_rate": 1.193678749719397e-06, "loss": 0.5274, "step": 1760 }, { "epoch": 0.7085093542546771, "grad_norm": 4.219902699982978, "learning_rate": 1.190686531491932e-06, "loss": 0.5397, "step": 1761 }, { "epoch": 0.7089116877891772, "grad_norm": 4.105043115699033, "learning_rate": 1.1876968955920916e-06, "loss": 0.4845, "step": 1762 }, { "epoch": 0.7093140213236774, "grad_norm": 3.905684008250733, "learning_rate": 1.1847098479162644e-06, "loss": 0.5769, "step": 1763 }, { "epoch": 0.7097163548581774, "grad_norm": 4.0434341189775616, "learning_rate": 1.1817253943557308e-06, "loss": 0.4338, "step": 1764 }, { "epoch": 0.7101186883926776, "grad_norm": 4.331648953621162, "learning_rate": 1.1787435407966593e-06, "loss": 0.6231, "step": 1765 }, { "epoch": 0.7105210219271776, "grad_norm": 4.225800580764582, "learning_rate": 1.1757642931200864e-06, "loss": 0.4946, "step": 1766 }, { "epoch": 0.7109233554616777, "grad_norm": 3.832934508222693, "learning_rate": 1.1727876572019117e-06, "loss": 0.5319, "step": 1767 }, { "epoch": 0.7113256889961779, "grad_norm": 3.994089631167334, "learning_rate": 1.1698136389128833e-06, "loss": 0.5005, "step": 1768 }, { "epoch": 0.7117280225306779, "grad_norm": 3.8399445750919425, "learning_rate": 1.166842244118587e-06, "loss": 0.4726, "step": 1769 }, { "epoch": 0.712130356065178, "grad_norm": 4.043691249667768, "learning_rate": 1.163873478679432e-06, "loss": 0.4885, "step": 1770 }, { "epoch": 0.7125326895996781, "grad_norm": 4.26586065507147, "learning_rate": 1.1609073484506441e-06, "loss": 0.5955, "step": 1771 }, { "epoch": 0.7129350231341782, "grad_norm": 3.2723853825243006, "learning_rate": 1.157943859282251e-06, "loss": 0.393, "step": 1772 }, { "epoch": 0.7133373566686784, "grad_norm": 4.228399556429077, "learning_rate": 1.1549830170190714e-06, "loss": 0.6138, "step": 1773 }, { "epoch": 0.7137396902031784, "grad_norm": 3.703790105757895, "learning_rate": 1.152024827500705e-06, "loss": 0.4952, "step": 1774 }, { "epoch": 0.7141420237376785, "grad_norm": 4.037603384099717, "learning_rate": 1.149069296561516e-06, "loss": 0.5791, "step": 1775 }, { "epoch": 0.7145443572721787, "grad_norm": 3.958021889837016, "learning_rate": 1.1461164300306292e-06, "loss": 0.495, "step": 1776 }, { "epoch": 0.7149466908066787, "grad_norm": 4.142430978673458, "learning_rate": 1.1431662337319107e-06, "loss": 0.6534, "step": 1777 }, { "epoch": 0.7153490243411789, "grad_norm": 4.484413621549814, "learning_rate": 1.1402187134839643e-06, "loss": 0.4885, "step": 1778 }, { "epoch": 0.7157513578756789, "grad_norm": 3.6695545389300395, "learning_rate": 1.1372738751001111e-06, "loss": 0.4446, "step": 1779 }, { "epoch": 0.716153691410179, "grad_norm": 3.492829736247777, "learning_rate": 1.134331724388387e-06, "loss": 0.5245, "step": 1780 }, { "epoch": 0.7165560249446792, "grad_norm": 3.8848835038152107, "learning_rate": 1.1313922671515252e-06, "loss": 0.5042, "step": 1781 }, { "epoch": 0.7169583584791792, "grad_norm": 4.4510516838349705, "learning_rate": 1.128455509186948e-06, "loss": 0.6898, "step": 1782 }, { "epoch": 0.7173606920136794, "grad_norm": 4.381017376532981, "learning_rate": 1.1255214562867503e-06, "loss": 0.5289, "step": 1783 }, { "epoch": 0.7177630255481794, "grad_norm": 3.5079718266262416, "learning_rate": 1.122590114237696e-06, "loss": 0.4724, "step": 1784 }, { "epoch": 0.7181653590826795, "grad_norm": 4.292226353007704, "learning_rate": 1.1196614888212007e-06, "loss": 0.5796, "step": 1785 }, { "epoch": 0.7185676926171797, "grad_norm": 4.052087753788445, "learning_rate": 1.1167355858133226e-06, "loss": 0.4938, "step": 1786 }, { "epoch": 0.7189700261516797, "grad_norm": 5.157980356295375, "learning_rate": 1.1138124109847503e-06, "loss": 0.5949, "step": 1787 }, { "epoch": 0.7193723596861799, "grad_norm": 3.285940613285488, "learning_rate": 1.1108919701007907e-06, "loss": 0.3908, "step": 1788 }, { "epoch": 0.71977469322068, "grad_norm": 3.5335717637010937, "learning_rate": 1.1079742689213593e-06, "loss": 0.5013, "step": 1789 }, { "epoch": 0.72017702675518, "grad_norm": 4.450180240936036, "learning_rate": 1.1050593132009703e-06, "loss": 0.6329, "step": 1790 }, { "epoch": 0.7205793602896802, "grad_norm": 4.404576812133769, "learning_rate": 1.1021471086887187e-06, "loss": 0.5781, "step": 1791 }, { "epoch": 0.7209816938241802, "grad_norm": 3.3602741770117177, "learning_rate": 1.0992376611282776e-06, "loss": 0.3621, "step": 1792 }, { "epoch": 0.7213840273586803, "grad_norm": 3.381272077924684, "learning_rate": 1.0963309762578795e-06, "loss": 0.384, "step": 1793 }, { "epoch": 0.7217863608931805, "grad_norm": 3.5780173251972025, "learning_rate": 1.0934270598103094e-06, "loss": 0.5372, "step": 1794 }, { "epoch": 0.7221886944276805, "grad_norm": 3.7738716419210814, "learning_rate": 1.0905259175128931e-06, "loss": 0.4336, "step": 1795 }, { "epoch": 0.7225910279621807, "grad_norm": 3.820342940812247, "learning_rate": 1.0876275550874846e-06, "loss": 0.5335, "step": 1796 }, { "epoch": 0.7229933614966807, "grad_norm": 4.176570559423123, "learning_rate": 1.0847319782504532e-06, "loss": 0.5038, "step": 1797 }, { "epoch": 0.7233956950311808, "grad_norm": 3.67077010562448, "learning_rate": 1.0818391927126764e-06, "loss": 0.4342, "step": 1798 }, { "epoch": 0.723798028565681, "grad_norm": 3.9880069700930423, "learning_rate": 1.078949204179526e-06, "loss": 0.4953, "step": 1799 }, { "epoch": 0.724200362100181, "grad_norm": 4.161334696726195, "learning_rate": 1.0760620183508582e-06, "loss": 0.5799, "step": 1800 }, { "epoch": 0.7246026956346812, "grad_norm": 4.126464363948578, "learning_rate": 1.073177640920999e-06, "loss": 0.5138, "step": 1801 }, { "epoch": 0.7250050291691813, "grad_norm": 4.1895057334300505, "learning_rate": 1.0702960775787374e-06, "loss": 0.5121, "step": 1802 }, { "epoch": 0.7254073627036813, "grad_norm": 3.884515999748438, "learning_rate": 1.0674173340073121e-06, "loss": 0.4891, "step": 1803 }, { "epoch": 0.7258096962381815, "grad_norm": 4.286867137892461, "learning_rate": 1.0645414158844012e-06, "loss": 0.6118, "step": 1804 }, { "epoch": 0.7262120297726815, "grad_norm": 3.8959234040057806, "learning_rate": 1.0616683288821067e-06, "loss": 0.5234, "step": 1805 }, { "epoch": 0.7266143633071817, "grad_norm": 4.239978882514847, "learning_rate": 1.058798078666952e-06, "loss": 0.5183, "step": 1806 }, { "epoch": 0.7270166968416818, "grad_norm": 3.7122065670739235, "learning_rate": 1.0559306708998607e-06, "loss": 0.501, "step": 1807 }, { "epoch": 0.7274190303761818, "grad_norm": 4.556831231079228, "learning_rate": 1.0530661112361533e-06, "loss": 0.5257, "step": 1808 }, { "epoch": 0.727821363910682, "grad_norm": 4.308685532929913, "learning_rate": 1.050204405325534e-06, "loss": 0.5534, "step": 1809 }, { "epoch": 0.728223697445182, "grad_norm": 3.9297951909048368, "learning_rate": 1.0473455588120738e-06, "loss": 0.4681, "step": 1810 }, { "epoch": 0.7286260309796821, "grad_norm": 3.98667448491781, "learning_rate": 1.0444895773342092e-06, "loss": 0.4785, "step": 1811 }, { "epoch": 0.7290283645141823, "grad_norm": 4.2934600586771845, "learning_rate": 1.0416364665247236e-06, "loss": 0.5408, "step": 1812 }, { "epoch": 0.7294306980486823, "grad_norm": 3.9347493779534, "learning_rate": 1.0387862320107403e-06, "loss": 0.5895, "step": 1813 }, { "epoch": 0.7298330315831825, "grad_norm": 4.338941092246621, "learning_rate": 1.0359388794137068e-06, "loss": 0.5648, "step": 1814 }, { "epoch": 0.7302353651176826, "grad_norm": 3.792647800273631, "learning_rate": 1.0330944143493892e-06, "loss": 0.5449, "step": 1815 }, { "epoch": 0.7306376986521826, "grad_norm": 3.821766624740813, "learning_rate": 1.0302528424278585e-06, "loss": 0.4544, "step": 1816 }, { "epoch": 0.7310400321866828, "grad_norm": 4.176558757203708, "learning_rate": 1.0274141692534784e-06, "loss": 0.4802, "step": 1817 }, { "epoch": 0.7314423657211828, "grad_norm": 3.5626608894002803, "learning_rate": 1.0245784004248968e-06, "loss": 0.4273, "step": 1818 }, { "epoch": 0.731844699255683, "grad_norm": 3.8719939712343936, "learning_rate": 1.0217455415350316e-06, "loss": 0.5409, "step": 1819 }, { "epoch": 0.7322470327901831, "grad_norm": 4.537877117506395, "learning_rate": 1.0189155981710642e-06, "loss": 0.5705, "step": 1820 }, { "epoch": 0.7326493663246831, "grad_norm": 3.932184586794533, "learning_rate": 1.0160885759144217e-06, "loss": 0.4904, "step": 1821 }, { "epoch": 0.7330516998591833, "grad_norm": 4.266661778185627, "learning_rate": 1.0132644803407753e-06, "loss": 0.4769, "step": 1822 }, { "epoch": 0.7334540333936833, "grad_norm": 3.7589693532265827, "learning_rate": 1.0104433170200188e-06, "loss": 0.4891, "step": 1823 }, { "epoch": 0.7338563669281835, "grad_norm": 3.8562553041075263, "learning_rate": 1.007625091516266e-06, "loss": 0.417, "step": 1824 }, { "epoch": 0.7342587004626836, "grad_norm": 4.130360177598847, "learning_rate": 1.0048098093878352e-06, "loss": 0.5207, "step": 1825 }, { "epoch": 0.7346610339971836, "grad_norm": 4.16513812007776, "learning_rate": 1.001997476187241e-06, "loss": 0.5213, "step": 1826 }, { "epoch": 0.7350633675316838, "grad_norm": 4.4208377164885055, "learning_rate": 9.991880974611809e-07, "loss": 0.6269, "step": 1827 }, { "epoch": 0.7354657010661839, "grad_norm": 4.867032713419264, "learning_rate": 9.96381678750524e-07, "loss": 0.4808, "step": 1828 }, { "epoch": 0.735868034600684, "grad_norm": 3.982545780740484, "learning_rate": 9.935782255903034e-07, "loss": 0.5066, "step": 1829 }, { "epoch": 0.7362703681351841, "grad_norm": 4.286351965336647, "learning_rate": 9.907777435097029e-07, "loss": 0.5151, "step": 1830 }, { "epoch": 0.7366727016696841, "grad_norm": 4.362229906545009, "learning_rate": 9.879802380320473e-07, "loss": 0.5936, "step": 1831 }, { "epoch": 0.7370750352041843, "grad_norm": 4.660993080328005, "learning_rate": 9.851857146747873e-07, "loss": 0.6829, "step": 1832 }, { "epoch": 0.7374773687386844, "grad_norm": 4.222997821431192, "learning_rate": 9.82394178949497e-07, "loss": 0.514, "step": 1833 }, { "epoch": 0.7378797022731844, "grad_norm": 4.369792251301722, "learning_rate": 9.796056363618533e-07, "loss": 0.5742, "step": 1834 }, { "epoch": 0.7382820358076846, "grad_norm": 4.422672878513299, "learning_rate": 9.768200924116338e-07, "loss": 0.568, "step": 1835 }, { "epoch": 0.7386843693421846, "grad_norm": 4.507594407110887, "learning_rate": 9.740375525926988e-07, "loss": 0.6017, "step": 1836 }, { "epoch": 0.7390867028766848, "grad_norm": 3.8125967279749835, "learning_rate": 9.71258022392985e-07, "loss": 0.5959, "step": 1837 }, { "epoch": 0.7394890364111849, "grad_norm": 3.7273585439154586, "learning_rate": 9.684815072944946e-07, "loss": 0.3825, "step": 1838 }, { "epoch": 0.7398913699456849, "grad_norm": 3.738805120372397, "learning_rate": 9.657080127732807e-07, "loss": 0.4931, "step": 1839 }, { "epoch": 0.7402937034801851, "grad_norm": 3.9828190911127526, "learning_rate": 9.629375442994418e-07, "loss": 0.5182, "step": 1840 }, { "epoch": 0.7406960370146852, "grad_norm": 3.9798516908695882, "learning_rate": 9.601701073371045e-07, "loss": 0.548, "step": 1841 }, { "epoch": 0.7410983705491853, "grad_norm": 3.7199861649210715, "learning_rate": 9.574057073444195e-07, "loss": 0.468, "step": 1842 }, { "epoch": 0.7415007040836854, "grad_norm": 4.0680484984054335, "learning_rate": 9.546443497735467e-07, "loss": 0.5782, "step": 1843 }, { "epoch": 0.7419030376181854, "grad_norm": 3.4109362486846115, "learning_rate": 9.518860400706465e-07, "loss": 0.4245, "step": 1844 }, { "epoch": 0.7423053711526856, "grad_norm": 3.931799967484987, "learning_rate": 9.491307836758651e-07, "loss": 0.4863, "step": 1845 }, { "epoch": 0.7427077046871857, "grad_norm": 4.211928044347116, "learning_rate": 9.463785860233301e-07, "loss": 0.5459, "step": 1846 }, { "epoch": 0.7431100382216858, "grad_norm": 4.0902800868050715, "learning_rate": 9.436294525411357e-07, "loss": 0.6375, "step": 1847 }, { "epoch": 0.7435123717561859, "grad_norm": 3.524293359481542, "learning_rate": 9.408833886513303e-07, "loss": 0.444, "step": 1848 }, { "epoch": 0.7439147052906859, "grad_norm": 3.6061388537955827, "learning_rate": 9.381403997699117e-07, "loss": 0.5217, "step": 1849 }, { "epoch": 0.7443170388251861, "grad_norm": 3.6547603411472838, "learning_rate": 9.354004913068096e-07, "loss": 0.5229, "step": 1850 }, { "epoch": 0.7447193723596862, "grad_norm": 4.340731245011316, "learning_rate": 9.326636686658808e-07, "loss": 0.5924, "step": 1851 }, { "epoch": 0.7451217058941862, "grad_norm": 3.4273270658083166, "learning_rate": 9.299299372448953e-07, "loss": 0.4941, "step": 1852 }, { "epoch": 0.7455240394286864, "grad_norm": 3.9330296000490668, "learning_rate": 9.271993024355263e-07, "loss": 0.5988, "step": 1853 }, { "epoch": 0.7459263729631865, "grad_norm": 4.79047981984662, "learning_rate": 9.244717696233388e-07, "loss": 0.6193, "step": 1854 }, { "epoch": 0.7463287064976866, "grad_norm": 3.6277035412847924, "learning_rate": 9.217473441877805e-07, "loss": 0.4905, "step": 1855 }, { "epoch": 0.7467310400321867, "grad_norm": 4.652255454626382, "learning_rate": 9.19026031502171e-07, "loss": 0.6231, "step": 1856 }, { "epoch": 0.7471333735666867, "grad_norm": 4.139619345656057, "learning_rate": 9.1630783693369e-07, "loss": 0.4981, "step": 1857 }, { "epoch": 0.7475357071011869, "grad_norm": 5.059383612132635, "learning_rate": 9.135927658433685e-07, "loss": 0.7292, "step": 1858 }, { "epoch": 0.747938040635687, "grad_norm": 4.664959181536374, "learning_rate": 9.108808235860744e-07, "loss": 0.6118, "step": 1859 }, { "epoch": 0.7483403741701871, "grad_norm": 4.243108848129753, "learning_rate": 9.081720155105076e-07, "loss": 0.5616, "step": 1860 }, { "epoch": 0.7487427077046872, "grad_norm": 3.4693576707426548, "learning_rate": 9.054663469591862e-07, "loss": 0.4451, "step": 1861 }, { "epoch": 0.7491450412391872, "grad_norm": 4.037761942838199, "learning_rate": 9.027638232684339e-07, "loss": 0.5354, "step": 1862 }, { "epoch": 0.7495473747736874, "grad_norm": 4.037140323875402, "learning_rate": 9.00064449768375e-07, "loss": 0.3946, "step": 1863 }, { "epoch": 0.7499497083081875, "grad_norm": 4.420498117292749, "learning_rate": 8.97368231782918e-07, "loss": 0.6091, "step": 1864 }, { "epoch": 0.7503520418426876, "grad_norm": 4.098542987137017, "learning_rate": 8.946751746297494e-07, "loss": 0.6347, "step": 1865 }, { "epoch": 0.7507543753771877, "grad_norm": 3.718783314907544, "learning_rate": 8.919852836203224e-07, "loss": 0.5523, "step": 1866 }, { "epoch": 0.7511567089116878, "grad_norm": 4.238567081312867, "learning_rate": 8.892985640598434e-07, "loss": 0.5716, "step": 1867 }, { "epoch": 0.7515590424461879, "grad_norm": 3.878960051531824, "learning_rate": 8.866150212472657e-07, "loss": 0.5552, "step": 1868 }, { "epoch": 0.751961375980688, "grad_norm": 4.0405823701369155, "learning_rate": 8.839346604752763e-07, "loss": 0.5286, "step": 1869 }, { "epoch": 0.752363709515188, "grad_norm": 4.251404967808422, "learning_rate": 8.81257487030287e-07, "loss": 0.4872, "step": 1870 }, { "epoch": 0.7527660430496882, "grad_norm": 3.6336762556932163, "learning_rate": 8.785835061924234e-07, "loss": 0.4301, "step": 1871 }, { "epoch": 0.7531683765841883, "grad_norm": 4.324873818573403, "learning_rate": 8.759127232355128e-07, "loss": 0.5405, "step": 1872 }, { "epoch": 0.7535707101186884, "grad_norm": 4.42447965555661, "learning_rate": 8.732451434270767e-07, "loss": 0.5209, "step": 1873 }, { "epoch": 0.7539730436531885, "grad_norm": 4.251290292687751, "learning_rate": 8.705807720283199e-07, "loss": 0.5951, "step": 1874 }, { "epoch": 0.7543753771876885, "grad_norm": 4.171584555966765, "learning_rate": 8.679196142941173e-07, "loss": 0.6086, "step": 1875 }, { "epoch": 0.7547777107221887, "grad_norm": 4.0958958819445614, "learning_rate": 8.652616754730075e-07, "loss": 0.5693, "step": 1876 }, { "epoch": 0.7551800442566888, "grad_norm": 3.683062028222347, "learning_rate": 8.626069608071785e-07, "loss": 0.4584, "step": 1877 }, { "epoch": 0.7555823777911889, "grad_norm": 3.9012710111822924, "learning_rate": 8.599554755324613e-07, "loss": 0.4611, "step": 1878 }, { "epoch": 0.755984711325689, "grad_norm": 3.5489182194802704, "learning_rate": 8.573072248783165e-07, "loss": 0.5083, "step": 1879 }, { "epoch": 0.7563870448601892, "grad_norm": 3.9884722155489225, "learning_rate": 8.546622140678265e-07, "loss": 0.5094, "step": 1880 }, { "epoch": 0.7567893783946892, "grad_norm": 4.328049459143733, "learning_rate": 8.520204483176814e-07, "loss": 0.3656, "step": 1881 }, { "epoch": 0.7571917119291893, "grad_norm": 3.9339309591347087, "learning_rate": 8.493819328381731e-07, "loss": 0.5317, "step": 1882 }, { "epoch": 0.7575940454636894, "grad_norm": 4.2270154261756385, "learning_rate": 8.467466728331828e-07, "loss": 0.4588, "step": 1883 }, { "epoch": 0.7579963789981895, "grad_norm": 3.303945186981914, "learning_rate": 8.441146735001715e-07, "loss": 0.4649, "step": 1884 }, { "epoch": 0.7583987125326896, "grad_norm": 4.054188026080457, "learning_rate": 8.414859400301667e-07, "loss": 0.5645, "step": 1885 }, { "epoch": 0.7588010460671897, "grad_norm": 4.17992148808867, "learning_rate": 8.388604776077575e-07, "loss": 0.5204, "step": 1886 }, { "epoch": 0.7592033796016898, "grad_norm": 4.428173356861454, "learning_rate": 8.362382914110804e-07, "loss": 0.6007, "step": 1887 }, { "epoch": 0.7596057131361899, "grad_norm": 4.255468940795954, "learning_rate": 8.336193866118117e-07, "loss": 0.5153, "step": 1888 }, { "epoch": 0.76000804667069, "grad_norm": 4.025791541260584, "learning_rate": 8.310037683751527e-07, "loss": 0.4859, "step": 1889 }, { "epoch": 0.7604103802051901, "grad_norm": 3.8573782279099533, "learning_rate": 8.283914418598263e-07, "loss": 0.5058, "step": 1890 }, { "epoch": 0.7608127137396902, "grad_norm": 3.332235385366874, "learning_rate": 8.257824122180602e-07, "loss": 0.3737, "step": 1891 }, { "epoch": 0.7612150472741903, "grad_norm": 3.394561752123056, "learning_rate": 8.231766845955818e-07, "loss": 0.4197, "step": 1892 }, { "epoch": 0.7616173808086905, "grad_norm": 3.3171785759316443, "learning_rate": 8.205742641316064e-07, "loss": 0.417, "step": 1893 }, { "epoch": 0.7620197143431905, "grad_norm": 4.487532016958629, "learning_rate": 8.179751559588234e-07, "loss": 0.5663, "step": 1894 }, { "epoch": 0.7624220478776906, "grad_norm": 4.546271967177341, "learning_rate": 8.153793652033931e-07, "loss": 0.5597, "step": 1895 }, { "epoch": 0.7628243814121907, "grad_norm": 3.914164846084329, "learning_rate": 8.12786896984931e-07, "loss": 0.4677, "step": 1896 }, { "epoch": 0.7632267149466908, "grad_norm": 3.852819616304247, "learning_rate": 8.101977564165011e-07, "loss": 0.503, "step": 1897 }, { "epoch": 0.763629048481191, "grad_norm": 4.067929177534588, "learning_rate": 8.076119486046019e-07, "loss": 0.5396, "step": 1898 }, { "epoch": 0.764031382015691, "grad_norm": 4.010571344025329, "learning_rate": 8.050294786491611e-07, "loss": 0.4829, "step": 1899 }, { "epoch": 0.7644337155501911, "grad_norm": 4.199207618578668, "learning_rate": 8.024503516435222e-07, "loss": 0.6123, "step": 1900 }, { "epoch": 0.7648360490846912, "grad_norm": 4.313823267969129, "learning_rate": 7.998745726744358e-07, "loss": 0.5597, "step": 1901 }, { "epoch": 0.7652383826191913, "grad_norm": 3.7661528466618956, "learning_rate": 7.973021468220502e-07, "loss": 0.5036, "step": 1902 }, { "epoch": 0.7656407161536914, "grad_norm": 4.017310363498917, "learning_rate": 7.947330791598973e-07, "loss": 0.5435, "step": 1903 }, { "epoch": 0.7660430496881915, "grad_norm": 4.447099011053462, "learning_rate": 7.921673747548895e-07, "loss": 0.4892, "step": 1904 }, { "epoch": 0.7664453832226916, "grad_norm": 4.1491367299305555, "learning_rate": 7.896050386673026e-07, "loss": 0.5234, "step": 1905 }, { "epoch": 0.7668477167571918, "grad_norm": 4.640959967611087, "learning_rate": 7.870460759507722e-07, "loss": 0.5619, "step": 1906 }, { "epoch": 0.7672500502916918, "grad_norm": 4.207681163367058, "learning_rate": 7.844904916522784e-07, "loss": 0.6313, "step": 1907 }, { "epoch": 0.7676523838261919, "grad_norm": 4.624725241802207, "learning_rate": 7.819382908121386e-07, "loss": 0.6169, "step": 1908 }, { "epoch": 0.768054717360692, "grad_norm": 3.891844454436063, "learning_rate": 7.793894784639986e-07, "loss": 0.4932, "step": 1909 }, { "epoch": 0.7684570508951921, "grad_norm": 3.6005349724480573, "learning_rate": 7.768440596348192e-07, "loss": 0.5835, "step": 1910 }, { "epoch": 0.7688593844296923, "grad_norm": 4.073804566700674, "learning_rate": 7.743020393448702e-07, "loss": 0.4882, "step": 1911 }, { "epoch": 0.7692617179641923, "grad_norm": 3.959969867245135, "learning_rate": 7.717634226077156e-07, "loss": 0.4945, "step": 1912 }, { "epoch": 0.7696640514986924, "grad_norm": 3.599166244946097, "learning_rate": 7.692282144302094e-07, "loss": 0.3889, "step": 1913 }, { "epoch": 0.7700663850331925, "grad_norm": 3.539375774753211, "learning_rate": 7.666964198124819e-07, "loss": 0.523, "step": 1914 }, { "epoch": 0.7704687185676926, "grad_norm": 3.912187073942566, "learning_rate": 7.64168043747932e-07, "loss": 0.5567, "step": 1915 }, { "epoch": 0.7708710521021928, "grad_norm": 4.034008046236846, "learning_rate": 7.616430912232137e-07, "loss": 0.5219, "step": 1916 }, { "epoch": 0.7712733856366928, "grad_norm": 4.074005253663992, "learning_rate": 7.59121567218232e-07, "loss": 0.5302, "step": 1917 }, { "epoch": 0.7716757191711929, "grad_norm": 3.4776790939949347, "learning_rate": 7.566034767061265e-07, "loss": 0.2992, "step": 1918 }, { "epoch": 0.7720780527056931, "grad_norm": 3.755741404325393, "learning_rate": 7.540888246532688e-07, "loss": 0.4337, "step": 1919 }, { "epoch": 0.7724803862401931, "grad_norm": 3.854464881913325, "learning_rate": 7.515776160192454e-07, "loss": 0.5286, "step": 1920 }, { "epoch": 0.7728827197746932, "grad_norm": 3.74441109291879, "learning_rate": 7.490698557568535e-07, "loss": 0.5449, "step": 1921 }, { "epoch": 0.7732850533091933, "grad_norm": 4.570844263947877, "learning_rate": 7.465655488120891e-07, "loss": 0.6153, "step": 1922 }, { "epoch": 0.7736873868436934, "grad_norm": 3.7650085288496613, "learning_rate": 7.440647001241363e-07, "loss": 0.5166, "step": 1923 }, { "epoch": 0.7740897203781936, "grad_norm": 3.6719483984817236, "learning_rate": 7.415673146253604e-07, "loss": 0.4842, "step": 1924 }, { "epoch": 0.7744920539126936, "grad_norm": 4.057471839423865, "learning_rate": 7.390733972412933e-07, "loss": 0.6075, "step": 1925 }, { "epoch": 0.7748943874471937, "grad_norm": 4.4626847789976924, "learning_rate": 7.365829528906293e-07, "loss": 0.5292, "step": 1926 }, { "epoch": 0.7752967209816938, "grad_norm": 3.76677011259477, "learning_rate": 7.340959864852124e-07, "loss": 0.405, "step": 1927 }, { "epoch": 0.7756990545161939, "grad_norm": 3.8469846124186082, "learning_rate": 7.316125029300275e-07, "loss": 0.4827, "step": 1928 }, { "epoch": 0.7761013880506941, "grad_norm": 3.8371445392089827, "learning_rate": 7.291325071231883e-07, "loss": 0.5134, "step": 1929 }, { "epoch": 0.7765037215851941, "grad_norm": 3.8383451450589887, "learning_rate": 7.26656003955932e-07, "loss": 0.4061, "step": 1930 }, { "epoch": 0.7769060551196942, "grad_norm": 4.037666657124557, "learning_rate": 7.241829983126075e-07, "loss": 0.4976, "step": 1931 }, { "epoch": 0.7773083886541944, "grad_norm": 3.980302342869356, "learning_rate": 7.217134950706625e-07, "loss": 0.5149, "step": 1932 }, { "epoch": 0.7777107221886944, "grad_norm": 3.7463100803001277, "learning_rate": 7.192474991006416e-07, "loss": 0.459, "step": 1933 }, { "epoch": 0.7781130557231946, "grad_norm": 3.9126255555142597, "learning_rate": 7.167850152661673e-07, "loss": 0.5537, "step": 1934 }, { "epoch": 0.7785153892576946, "grad_norm": 4.306777769767276, "learning_rate": 7.143260484239386e-07, "loss": 0.4997, "step": 1935 }, { "epoch": 0.7789177227921947, "grad_norm": 4.472035583740821, "learning_rate": 7.118706034237172e-07, "loss": 0.6373, "step": 1936 }, { "epoch": 0.7793200563266949, "grad_norm": 3.8082808763272356, "learning_rate": 7.094186851083187e-07, "loss": 0.5713, "step": 1937 }, { "epoch": 0.7797223898611949, "grad_norm": 3.907355998411772, "learning_rate": 7.069702983136023e-07, "loss": 0.4916, "step": 1938 }, { "epoch": 0.780124723395695, "grad_norm": 3.4404115391335623, "learning_rate": 7.045254478684626e-07, "loss": 0.4552, "step": 1939 }, { "epoch": 0.7805270569301951, "grad_norm": 4.072479874224982, "learning_rate": 7.020841385948199e-07, "loss": 0.5061, "step": 1940 }, { "epoch": 0.7809293904646952, "grad_norm": 5.014496338999608, "learning_rate": 6.996463753076102e-07, "loss": 0.5401, "step": 1941 }, { "epoch": 0.7813317239991954, "grad_norm": 4.5982221422710134, "learning_rate": 6.972121628147766e-07, "loss": 0.6385, "step": 1942 }, { "epoch": 0.7817340575336954, "grad_norm": 3.866309336941733, "learning_rate": 6.94781505917256e-07, "loss": 0.5007, "step": 1943 }, { "epoch": 0.7821363910681955, "grad_norm": 4.000641087223532, "learning_rate": 6.923544094089762e-07, "loss": 0.4097, "step": 1944 }, { "epoch": 0.7825387246026957, "grad_norm": 4.183627371809607, "learning_rate": 6.899308780768423e-07, "loss": 0.4986, "step": 1945 }, { "epoch": 0.7829410581371957, "grad_norm": 3.834784873841407, "learning_rate": 6.875109167007255e-07, "loss": 0.4719, "step": 1946 }, { "epoch": 0.7833433916716959, "grad_norm": 4.624045837237235, "learning_rate": 6.85094530053459e-07, "loss": 0.5742, "step": 1947 }, { "epoch": 0.7837457252061959, "grad_norm": 4.968557042198568, "learning_rate": 6.826817229008232e-07, "loss": 0.6943, "step": 1948 }, { "epoch": 0.784148058740696, "grad_norm": 3.798467897259186, "learning_rate": 6.802725000015406e-07, "loss": 0.5297, "step": 1949 }, { "epoch": 0.7845503922751962, "grad_norm": 3.974644659610978, "learning_rate": 6.778668661072643e-07, "loss": 0.4803, "step": 1950 }, { "epoch": 0.7849527258096962, "grad_norm": 4.048333500653427, "learning_rate": 6.754648259625673e-07, "loss": 0.5941, "step": 1951 }, { "epoch": 0.7853550593441964, "grad_norm": 4.3035958811807875, "learning_rate": 6.730663843049367e-07, "loss": 0.5955, "step": 1952 }, { "epoch": 0.7857573928786964, "grad_norm": 3.6544635394961764, "learning_rate": 6.706715458647615e-07, "loss": 0.4851, "step": 1953 }, { "epoch": 0.7861597264131965, "grad_norm": 4.218917868283341, "learning_rate": 6.682803153653241e-07, "loss": 0.6081, "step": 1954 }, { "epoch": 0.7865620599476967, "grad_norm": 3.617134711107577, "learning_rate": 6.658926975227923e-07, "loss": 0.4414, "step": 1955 }, { "epoch": 0.7869643934821967, "grad_norm": 4.041925903109421, "learning_rate": 6.635086970462057e-07, "loss": 0.4589, "step": 1956 }, { "epoch": 0.7873667270166969, "grad_norm": 3.9272206421591944, "learning_rate": 6.611283186374723e-07, "loss": 0.4697, "step": 1957 }, { "epoch": 0.7877690605511969, "grad_norm": 4.242979124153208, "learning_rate": 6.587515669913558e-07, "loss": 0.4185, "step": 1958 }, { "epoch": 0.788171394085697, "grad_norm": 3.38359591341345, "learning_rate": 6.563784467954668e-07, "loss": 0.3828, "step": 1959 }, { "epoch": 0.7885737276201972, "grad_norm": 4.348954859955942, "learning_rate": 6.540089627302523e-07, "loss": 0.5046, "step": 1960 }, { "epoch": 0.7889760611546972, "grad_norm": 4.120739882383607, "learning_rate": 6.516431194689907e-07, "loss": 0.5227, "step": 1961 }, { "epoch": 0.7893783946891973, "grad_norm": 3.6740741354898305, "learning_rate": 6.492809216777762e-07, "loss": 0.4134, "step": 1962 }, { "epoch": 0.7897807282236975, "grad_norm": 3.690318810926642, "learning_rate": 6.469223740155159e-07, "loss": 0.4901, "step": 1963 }, { "epoch": 0.7901830617581975, "grad_norm": 4.2381781760809485, "learning_rate": 6.445674811339175e-07, "loss": 0.4225, "step": 1964 }, { "epoch": 0.7905853952926977, "grad_norm": 4.106368765680081, "learning_rate": 6.422162476774788e-07, "loss": 0.5645, "step": 1965 }, { "epoch": 0.7909877288271977, "grad_norm": 3.706517698008605, "learning_rate": 6.398686782834815e-07, "loss": 0.4995, "step": 1966 }, { "epoch": 0.7913900623616978, "grad_norm": 4.432403407342872, "learning_rate": 6.375247775819809e-07, "loss": 0.5127, "step": 1967 }, { "epoch": 0.791792395896198, "grad_norm": 4.022769337353194, "learning_rate": 6.351845501957971e-07, "loss": 0.4525, "step": 1968 }, { "epoch": 0.792194729430698, "grad_norm": 4.43806071523494, "learning_rate": 6.328480007405027e-07, "loss": 0.5067, "step": 1969 }, { "epoch": 0.7925970629651982, "grad_norm": 3.7265320273813027, "learning_rate": 6.305151338244192e-07, "loss": 0.4277, "step": 1970 }, { "epoch": 0.7929993964996982, "grad_norm": 4.06221248413846, "learning_rate": 6.281859540486043e-07, "loss": 0.6297, "step": 1971 }, { "epoch": 0.7934017300341983, "grad_norm": 4.391919257148749, "learning_rate": 6.258604660068443e-07, "loss": 0.4994, "step": 1972 }, { "epoch": 0.7938040635686985, "grad_norm": 4.252838693305146, "learning_rate": 6.23538674285642e-07, "loss": 0.5949, "step": 1973 }, { "epoch": 0.7942063971031985, "grad_norm": 4.325947652020019, "learning_rate": 6.212205834642132e-07, "loss": 0.5428, "step": 1974 }, { "epoch": 0.7946087306376987, "grad_norm": 4.185176650642697, "learning_rate": 6.189061981144715e-07, "loss": 0.5732, "step": 1975 }, { "epoch": 0.7950110641721988, "grad_norm": 3.2994931408989627, "learning_rate": 6.165955228010242e-07, "loss": 0.37, "step": 1976 }, { "epoch": 0.7954133977066988, "grad_norm": 3.9581400842405277, "learning_rate": 6.14288562081162e-07, "loss": 0.6003, "step": 1977 }, { "epoch": 0.795815731241199, "grad_norm": 3.855839195313672, "learning_rate": 6.11985320504847e-07, "loss": 0.5755, "step": 1978 }, { "epoch": 0.796218064775699, "grad_norm": 4.28483670710571, "learning_rate": 6.096858026147079e-07, "loss": 0.5674, "step": 1979 }, { "epoch": 0.7966203983101992, "grad_norm": 3.946092310322622, "learning_rate": 6.073900129460286e-07, "loss": 0.4429, "step": 1980 }, { "epoch": 0.7970227318446993, "grad_norm": 3.9442626812628725, "learning_rate": 6.050979560267414e-07, "loss": 0.4953, "step": 1981 }, { "epoch": 0.7974250653791993, "grad_norm": 4.023203847345498, "learning_rate": 6.028096363774136e-07, "loss": 0.5437, "step": 1982 }, { "epoch": 0.7978273989136995, "grad_norm": 4.5536613551126335, "learning_rate": 6.005250585112437e-07, "loss": 0.6177, "step": 1983 }, { "epoch": 0.7982297324481995, "grad_norm": 3.8312030174786313, "learning_rate": 5.982442269340505e-07, "loss": 0.4801, "step": 1984 }, { "epoch": 0.7986320659826996, "grad_norm": 3.8614656531277896, "learning_rate": 5.959671461442634e-07, "loss": 0.4395, "step": 1985 }, { "epoch": 0.7990343995171998, "grad_norm": 4.191209006767766, "learning_rate": 5.936938206329148e-07, "loss": 0.4913, "step": 1986 }, { "epoch": 0.7994367330516998, "grad_norm": 4.201209209604027, "learning_rate": 5.914242548836291e-07, "loss": 0.567, "step": 1987 }, { "epoch": 0.7998390665862, "grad_norm": 4.711191490796992, "learning_rate": 5.891584533726177e-07, "loss": 0.6445, "step": 1988 }, { "epoch": 0.8002414001207001, "grad_norm": 4.050282830317415, "learning_rate": 5.868964205686653e-07, "loss": 0.4854, "step": 1989 }, { "epoch": 0.8006437336552001, "grad_norm": 3.4477078683288838, "learning_rate": 5.84638160933127e-07, "loss": 0.4578, "step": 1990 }, { "epoch": 0.8010460671897003, "grad_norm": 3.977635083967608, "learning_rate": 5.82383678919912e-07, "loss": 0.4966, "step": 1991 }, { "epoch": 0.8014484007242003, "grad_norm": 4.254984216161458, "learning_rate": 5.801329789754826e-07, "loss": 0.5101, "step": 1992 }, { "epoch": 0.8018507342587005, "grad_norm": 4.1472621686996005, "learning_rate": 5.778860655388399e-07, "loss": 0.6366, "step": 1993 }, { "epoch": 0.8022530677932006, "grad_norm": 4.0712688803226715, "learning_rate": 5.75642943041518e-07, "loss": 0.4908, "step": 1994 }, { "epoch": 0.8026554013277006, "grad_norm": 4.69998065888643, "learning_rate": 5.734036159075743e-07, "loss": 0.6769, "step": 1995 }, { "epoch": 0.8030577348622008, "grad_norm": 3.5455182945935144, "learning_rate": 5.711680885535784e-07, "loss": 0.4177, "step": 1996 }, { "epoch": 0.8034600683967008, "grad_norm": 3.853566513480482, "learning_rate": 5.689363653886085e-07, "loss": 0.4434, "step": 1997 }, { "epoch": 0.803862401931201, "grad_norm": 4.050333441948282, "learning_rate": 5.667084508142386e-07, "loss": 0.4822, "step": 1998 }, { "epoch": 0.8042647354657011, "grad_norm": 4.442980318720399, "learning_rate": 5.644843492245322e-07, "loss": 0.4872, "step": 1999 }, { "epoch": 0.8046670690002011, "grad_norm": 4.457653531733048, "learning_rate": 5.622640650060301e-07, "loss": 0.6444, "step": 2000 }, { "epoch": 0.8050694025347013, "grad_norm": 3.921778521793536, "learning_rate": 5.600476025377463e-07, "loss": 0.5391, "step": 2001 }, { "epoch": 0.8054717360692014, "grad_norm": 3.8081912666035755, "learning_rate": 5.578349661911578e-07, "loss": 0.4802, "step": 2002 }, { "epoch": 0.8058740696037014, "grad_norm": 4.001868301091059, "learning_rate": 5.556261603301924e-07, "loss": 0.5459, "step": 2003 }, { "epoch": 0.8062764031382016, "grad_norm": 3.847308434546458, "learning_rate": 5.534211893112268e-07, "loss": 0.4606, "step": 2004 }, { "epoch": 0.8066787366727016, "grad_norm": 4.15027221954597, "learning_rate": 5.512200574830712e-07, "loss": 0.5705, "step": 2005 }, { "epoch": 0.8070810702072018, "grad_norm": 3.8098780638264027, "learning_rate": 5.490227691869656e-07, "loss": 0.491, "step": 2006 }, { "epoch": 0.8074834037417019, "grad_norm": 3.9362205164209616, "learning_rate": 5.468293287565696e-07, "loss": 0.4342, "step": 2007 }, { "epoch": 0.8078857372762019, "grad_norm": 3.621695424225381, "learning_rate": 5.446397405179534e-07, "loss": 0.4943, "step": 2008 }, { "epoch": 0.8082880708107021, "grad_norm": 4.894139218776488, "learning_rate": 5.424540087895888e-07, "loss": 0.7108, "step": 2009 }, { "epoch": 0.8086904043452021, "grad_norm": 4.43306730596569, "learning_rate": 5.402721378823422e-07, "loss": 0.5759, "step": 2010 }, { "epoch": 0.8090927378797023, "grad_norm": 4.303012456014276, "learning_rate": 5.380941320994659e-07, "loss": 0.575, "step": 2011 }, { "epoch": 0.8094950714142024, "grad_norm": 4.313613214512251, "learning_rate": 5.35919995736589e-07, "loss": 0.4761, "step": 2012 }, { "epoch": 0.8098974049487024, "grad_norm": 3.8521996380430394, "learning_rate": 5.337497330817074e-07, "loss": 0.5352, "step": 2013 }, { "epoch": 0.8102997384832026, "grad_norm": 4.528703112399277, "learning_rate": 5.315833484151786e-07, "loss": 0.5959, "step": 2014 }, { "epoch": 0.8107020720177027, "grad_norm": 3.878303004429545, "learning_rate": 5.294208460097125e-07, "loss": 0.4815, "step": 2015 }, { "epoch": 0.8111044055522028, "grad_norm": 4.490283907824032, "learning_rate": 5.272622301303587e-07, "loss": 0.6368, "step": 2016 }, { "epoch": 0.8115067390867029, "grad_norm": 4.279104111581364, "learning_rate": 5.251075050345056e-07, "loss": 0.5216, "step": 2017 }, { "epoch": 0.8119090726212029, "grad_norm": 3.6608651815923086, "learning_rate": 5.229566749718645e-07, "loss": 0.4809, "step": 2018 }, { "epoch": 0.8123114061557031, "grad_norm": 4.666782295826111, "learning_rate": 5.208097441844668e-07, "loss": 0.504, "step": 2019 }, { "epoch": 0.8127137396902032, "grad_norm": 3.84925143665575, "learning_rate": 5.186667169066523e-07, "loss": 0.5423, "step": 2020 }, { "epoch": 0.8131160732247033, "grad_norm": 3.578156563304788, "learning_rate": 5.165275973650633e-07, "loss": 0.3936, "step": 2021 }, { "epoch": 0.8135184067592034, "grad_norm": 4.074315611637044, "learning_rate": 5.143923897786329e-07, "loss": 0.6302, "step": 2022 }, { "epoch": 0.8139207402937034, "grad_norm": 4.085845386699454, "learning_rate": 5.122610983585808e-07, "loss": 0.5789, "step": 2023 }, { "epoch": 0.8143230738282036, "grad_norm": 4.000394354287652, "learning_rate": 5.101337273084017e-07, "loss": 0.4229, "step": 2024 }, { "epoch": 0.8147254073627037, "grad_norm": 3.835079915243062, "learning_rate": 5.080102808238585e-07, "loss": 0.5005, "step": 2025 }, { "epoch": 0.8151277408972037, "grad_norm": 3.590556700281588, "learning_rate": 5.058907630929749e-07, "loss": 0.4452, "step": 2026 }, { "epoch": 0.8155300744317039, "grad_norm": 4.254640813185372, "learning_rate": 5.037751782960234e-07, "loss": 0.6328, "step": 2027 }, { "epoch": 0.815932407966204, "grad_norm": 4.290085466017056, "learning_rate": 5.016635306055223e-07, "loss": 0.5785, "step": 2028 }, { "epoch": 0.8163347415007041, "grad_norm": 4.159416054122212, "learning_rate": 4.995558241862242e-07, "loss": 0.5129, "step": 2029 }, { "epoch": 0.8167370750352042, "grad_norm": 3.7061504996754153, "learning_rate": 4.974520631951069e-07, "loss": 0.5305, "step": 2030 }, { "epoch": 0.8171394085697042, "grad_norm": 4.178397491082031, "learning_rate": 4.953522517813689e-07, "loss": 0.5855, "step": 2031 }, { "epoch": 0.8175417421042044, "grad_norm": 4.154495627484645, "learning_rate": 4.932563940864168e-07, "loss": 0.4164, "step": 2032 }, { "epoch": 0.8179440756387045, "grad_norm": 4.283562955009001, "learning_rate": 4.911644942438614e-07, "loss": 0.4942, "step": 2033 }, { "epoch": 0.8183464091732046, "grad_norm": 3.307871102947408, "learning_rate": 4.890765563795072e-07, "loss": 0.4552, "step": 2034 }, { "epoch": 0.8187487427077047, "grad_norm": 4.15735416688211, "learning_rate": 4.86992584611343e-07, "loss": 0.5797, "step": 2035 }, { "epoch": 0.8191510762422047, "grad_norm": 4.111111195260256, "learning_rate": 4.849125830495366e-07, "loss": 0.5776, "step": 2036 }, { "epoch": 0.8195534097767049, "grad_norm": 3.534887324822181, "learning_rate": 4.828365557964257e-07, "loss": 0.4913, "step": 2037 }, { "epoch": 0.819955743311205, "grad_norm": 3.8702017938683606, "learning_rate": 4.807645069465095e-07, "loss": 0.4638, "step": 2038 }, { "epoch": 0.820358076845705, "grad_norm": 4.105810885704208, "learning_rate": 4.786964405864403e-07, "loss": 0.51, "step": 2039 }, { "epoch": 0.8207604103802052, "grad_norm": 3.7721413153025183, "learning_rate": 4.766323607950149e-07, "loss": 0.5327, "step": 2040 }, { "epoch": 0.8211627439147053, "grad_norm": 4.604383899796593, "learning_rate": 4.7457227164316914e-07, "loss": 0.6425, "step": 2041 }, { "epoch": 0.8215650774492054, "grad_norm": 4.7452724983070595, "learning_rate": 4.725161771939679e-07, "loss": 0.5968, "step": 2042 }, { "epoch": 0.8219674109837055, "grad_norm": 4.13939509952924, "learning_rate": 4.704640815025971e-07, "loss": 0.5454, "step": 2043 }, { "epoch": 0.8223697445182055, "grad_norm": 4.07811911040908, "learning_rate": 4.684159886163553e-07, "loss": 0.4846, "step": 2044 }, { "epoch": 0.8227720780527057, "grad_norm": 4.223857527886817, "learning_rate": 4.663719025746477e-07, "loss": 0.548, "step": 2045 }, { "epoch": 0.8231744115872058, "grad_norm": 4.055398221449843, "learning_rate": 4.643318274089756e-07, "loss": 0.4762, "step": 2046 }, { "epoch": 0.8235767451217059, "grad_norm": 3.93292252326897, "learning_rate": 4.6229576714293067e-07, "loss": 0.5506, "step": 2047 }, { "epoch": 0.823979078656206, "grad_norm": 3.5755501548250357, "learning_rate": 4.6026372579218686e-07, "loss": 0.5277, "step": 2048 }, { "epoch": 0.824381412190706, "grad_norm": 3.900846576287911, "learning_rate": 4.5823570736448913e-07, "loss": 0.4289, "step": 2049 }, { "epoch": 0.8247837457252062, "grad_norm": 4.422345897561269, "learning_rate": 4.5621171585965046e-07, "loss": 0.5439, "step": 2050 }, { "epoch": 0.8251860792597063, "grad_norm": 3.86188238287213, "learning_rate": 4.5419175526954074e-07, "loss": 0.4971, "step": 2051 }, { "epoch": 0.8255884127942064, "grad_norm": 4.149840183197456, "learning_rate": 4.5217582957808064e-07, "loss": 0.4913, "step": 2052 }, { "epoch": 0.8259907463287065, "grad_norm": 4.14442670188834, "learning_rate": 4.501639427612309e-07, "loss": 0.593, "step": 2053 }, { "epoch": 0.8263930798632066, "grad_norm": 4.160435006991155, "learning_rate": 4.4815609878698847e-07, "loss": 0.5702, "step": 2054 }, { "epoch": 0.8267954133977067, "grad_norm": 4.699994325668033, "learning_rate": 4.46152301615376e-07, "loss": 0.6052, "step": 2055 }, { "epoch": 0.8271977469322068, "grad_norm": 4.158216057441291, "learning_rate": 4.441525551984349e-07, "loss": 0.6375, "step": 2056 }, { "epoch": 0.8276000804667069, "grad_norm": 5.052807884367178, "learning_rate": 4.421568634802165e-07, "loss": 0.687, "step": 2057 }, { "epoch": 0.828002414001207, "grad_norm": 3.9353061661541133, "learning_rate": 4.40165230396776e-07, "loss": 0.4774, "step": 2058 }, { "epoch": 0.8284047475357071, "grad_norm": 4.01193456780117, "learning_rate": 4.3817765987616474e-07, "loss": 0.4932, "step": 2059 }, { "epoch": 0.8288070810702072, "grad_norm": 4.976729164334829, "learning_rate": 4.361941558384189e-07, "loss": 0.6807, "step": 2060 }, { "epoch": 0.8292094146047073, "grad_norm": 4.339682582742787, "learning_rate": 4.3421472219555725e-07, "loss": 0.4727, "step": 2061 }, { "epoch": 0.8296117481392074, "grad_norm": 4.503522087170257, "learning_rate": 4.3223936285156836e-07, "loss": 0.6022, "step": 2062 }, { "epoch": 0.8300140816737075, "grad_norm": 4.009166626990359, "learning_rate": 4.3026808170240686e-07, "loss": 0.6318, "step": 2063 }, { "epoch": 0.8304164152082076, "grad_norm": 4.433472277211533, "learning_rate": 4.283008826359833e-07, "loss": 0.6059, "step": 2064 }, { "epoch": 0.8308187487427077, "grad_norm": 4.239654566860423, "learning_rate": 4.263377695321577e-07, "loss": 0.5319, "step": 2065 }, { "epoch": 0.8312210822772078, "grad_norm": 4.33872236149242, "learning_rate": 4.2437874626273024e-07, "loss": 0.5192, "step": 2066 }, { "epoch": 0.831623415811708, "grad_norm": 4.396517305178651, "learning_rate": 4.2242381669143603e-07, "loss": 0.6266, "step": 2067 }, { "epoch": 0.832025749346208, "grad_norm": 4.213477790744039, "learning_rate": 4.204729846739358e-07, "loss": 0.6738, "step": 2068 }, { "epoch": 0.8324280828807081, "grad_norm": 3.312219624247552, "learning_rate": 4.1852625405780914e-07, "loss": 0.4833, "step": 2069 }, { "epoch": 0.8328304164152082, "grad_norm": 3.9392383731854257, "learning_rate": 4.165836286825464e-07, "loss": 0.4513, "step": 2070 }, { "epoch": 0.8332327499497083, "grad_norm": 3.8224909723307734, "learning_rate": 4.1464511237954056e-07, "loss": 0.5628, "step": 2071 }, { "epoch": 0.8336350834842085, "grad_norm": 4.5465506777870885, "learning_rate": 4.1271070897208166e-07, "loss": 0.5761, "step": 2072 }, { "epoch": 0.8340374170187085, "grad_norm": 3.87942801566727, "learning_rate": 4.107804222753464e-07, "loss": 0.4364, "step": 2073 }, { "epoch": 0.8344397505532086, "grad_norm": 3.901824805945848, "learning_rate": 4.088542560963937e-07, "loss": 0.4897, "step": 2074 }, { "epoch": 0.8348420840877087, "grad_norm": 4.2369387169530155, "learning_rate": 4.0693221423415447e-07, "loss": 0.6186, "step": 2075 }, { "epoch": 0.8352444176222088, "grad_norm": 4.1329987533853005, "learning_rate": 4.0501430047942603e-07, "loss": 0.5358, "step": 2076 }, { "epoch": 0.835646751156709, "grad_norm": 4.533598883351018, "learning_rate": 4.03100518614864e-07, "loss": 0.5522, "step": 2077 }, { "epoch": 0.836049084691209, "grad_norm": 3.998009974659786, "learning_rate": 4.011908724149746e-07, "loss": 0.3743, "step": 2078 }, { "epoch": 0.8364514182257091, "grad_norm": 3.417865078970055, "learning_rate": 3.9928536564610767e-07, "loss": 0.4035, "step": 2079 }, { "epoch": 0.8368537517602093, "grad_norm": 3.9749987307006345, "learning_rate": 3.9738400206644766e-07, "loss": 0.6477, "step": 2080 }, { "epoch": 0.8372560852947093, "grad_norm": 4.037697255647064, "learning_rate": 3.9548678542600886e-07, "loss": 0.5661, "step": 2081 }, { "epoch": 0.8376584188292094, "grad_norm": 4.431095797135675, "learning_rate": 3.935937194666267e-07, "loss": 0.4818, "step": 2082 }, { "epoch": 0.8380607523637095, "grad_norm": 4.260045700942229, "learning_rate": 3.9170480792194985e-07, "loss": 0.4263, "step": 2083 }, { "epoch": 0.8384630858982096, "grad_norm": 4.051159338313962, "learning_rate": 3.8982005451743243e-07, "loss": 0.487, "step": 2084 }, { "epoch": 0.8388654194327098, "grad_norm": 3.505632578812268, "learning_rate": 3.8793946297032926e-07, "loss": 0.5173, "step": 2085 }, { "epoch": 0.8392677529672098, "grad_norm": 3.4621570841739917, "learning_rate": 3.8606303698968643e-07, "loss": 0.3609, "step": 2086 }, { "epoch": 0.8396700865017099, "grad_norm": 3.708654324006204, "learning_rate": 3.841907802763331e-07, "loss": 0.4437, "step": 2087 }, { "epoch": 0.84007242003621, "grad_norm": 3.3369961693311447, "learning_rate": 3.8232269652287713e-07, "loss": 0.4295, "step": 2088 }, { "epoch": 0.8404747535707101, "grad_norm": 3.8475631558462884, "learning_rate": 3.8045878941369434e-07, "loss": 0.5628, "step": 2089 }, { "epoch": 0.8408770871052103, "grad_norm": 3.827602289260064, "learning_rate": 3.7859906262492503e-07, "loss": 0.5158, "step": 2090 }, { "epoch": 0.8412794206397103, "grad_norm": 4.093693428637088, "learning_rate": 3.7674351982446347e-07, "loss": 0.4742, "step": 2091 }, { "epoch": 0.8416817541742104, "grad_norm": 4.310046930338328, "learning_rate": 3.7489216467195304e-07, "loss": 0.6097, "step": 2092 }, { "epoch": 0.8420840877087106, "grad_norm": 4.017527345482499, "learning_rate": 3.730450008187761e-07, "loss": 0.5075, "step": 2093 }, { "epoch": 0.8424864212432106, "grad_norm": 3.946920016020563, "learning_rate": 3.7120203190805034e-07, "loss": 0.4275, "step": 2094 }, { "epoch": 0.8428887547777107, "grad_norm": 4.145255268119006, "learning_rate": 3.6936326157461925e-07, "loss": 0.5914, "step": 2095 }, { "epoch": 0.8432910883122108, "grad_norm": 4.2460232246744525, "learning_rate": 3.6752869344504586e-07, "loss": 0.5152, "step": 2096 }, { "epoch": 0.8436934218467109, "grad_norm": 3.95033426326261, "learning_rate": 3.656983311376042e-07, "loss": 0.4448, "step": 2097 }, { "epoch": 0.8440957553812111, "grad_norm": 3.8931938435117286, "learning_rate": 3.638721782622745e-07, "loss": 0.499, "step": 2098 }, { "epoch": 0.8444980889157111, "grad_norm": 4.3118732957405985, "learning_rate": 3.620502384207342e-07, "loss": 0.49, "step": 2099 }, { "epoch": 0.8449004224502112, "grad_norm": 4.005112629756995, "learning_rate": 3.602325152063524e-07, "loss": 0.4125, "step": 2100 }, { "epoch": 0.8453027559847113, "grad_norm": 3.9812849673226114, "learning_rate": 3.5841901220418e-07, "loss": 0.5451, "step": 2101 }, { "epoch": 0.8457050895192114, "grad_norm": 3.948178072792975, "learning_rate": 3.566097329909468e-07, "loss": 0.6427, "step": 2102 }, { "epoch": 0.8461074230537116, "grad_norm": 3.7188722401214016, "learning_rate": 3.5480468113504946e-07, "loss": 0.405, "step": 2103 }, { "epoch": 0.8465097565882116, "grad_norm": 3.913596983806647, "learning_rate": 3.530038601965491e-07, "loss": 0.4635, "step": 2104 }, { "epoch": 0.8469120901227117, "grad_norm": 4.411578338116122, "learning_rate": 3.5120727372716295e-07, "loss": 0.5511, "step": 2105 }, { "epoch": 0.8473144236572119, "grad_norm": 3.282083239679109, "learning_rate": 3.494149252702536e-07, "loss": 0.4441, "step": 2106 }, { "epoch": 0.8477167571917119, "grad_norm": 4.011319326831284, "learning_rate": 3.476268183608281e-07, "loss": 0.5018, "step": 2107 }, { "epoch": 0.848119090726212, "grad_norm": 4.284870919327636, "learning_rate": 3.4584295652552667e-07, "loss": 0.4846, "step": 2108 }, { "epoch": 0.8485214242607121, "grad_norm": 4.276416896965254, "learning_rate": 3.4406334328261765e-07, "loss": 0.5376, "step": 2109 }, { "epoch": 0.8489237577952122, "grad_norm": 3.834125175179826, "learning_rate": 3.4228798214198974e-07, "loss": 0.5651, "step": 2110 }, { "epoch": 0.8493260913297124, "grad_norm": 3.9930841906813517, "learning_rate": 3.405168766051445e-07, "loss": 0.3851, "step": 2111 }, { "epoch": 0.8497284248642124, "grad_norm": 4.2112351410317626, "learning_rate": 3.3875003016519137e-07, "loss": 0.5431, "step": 2112 }, { "epoch": 0.8501307583987125, "grad_norm": 4.52788654812861, "learning_rate": 3.369874463068398e-07, "loss": 0.5849, "step": 2113 }, { "epoch": 0.8505330919332126, "grad_norm": 3.8509825855331163, "learning_rate": 3.3522912850639106e-07, "loss": 0.5748, "step": 2114 }, { "epoch": 0.8509354254677127, "grad_norm": 3.510509250668585, "learning_rate": 3.334750802317338e-07, "loss": 0.4397, "step": 2115 }, { "epoch": 0.8513377590022129, "grad_norm": 3.8349389092945114, "learning_rate": 3.3172530494233467e-07, "loss": 0.4769, "step": 2116 }, { "epoch": 0.8517400925367129, "grad_norm": 3.8288751826393197, "learning_rate": 3.2997980608923426e-07, "loss": 0.5434, "step": 2117 }, { "epoch": 0.852142426071213, "grad_norm": 3.9001403159636374, "learning_rate": 3.282385871150387e-07, "loss": 0.4523, "step": 2118 }, { "epoch": 0.8525447596057132, "grad_norm": 3.6933669568597978, "learning_rate": 3.26501651453911e-07, "loss": 0.3905, "step": 2119 }, { "epoch": 0.8529470931402132, "grad_norm": 3.811615026103731, "learning_rate": 3.247690025315689e-07, "loss": 0.4305, "step": 2120 }, { "epoch": 0.8533494266747134, "grad_norm": 3.6210800698038925, "learning_rate": 3.2304064376527417e-07, "loss": 0.5332, "step": 2121 }, { "epoch": 0.8537517602092134, "grad_norm": 4.231885438858511, "learning_rate": 3.2131657856382753e-07, "loss": 0.5911, "step": 2122 }, { "epoch": 0.8541540937437135, "grad_norm": 4.464935656077254, "learning_rate": 3.195968103275618e-07, "loss": 0.5702, "step": 2123 }, { "epoch": 0.8545564272782137, "grad_norm": 4.3078684853827465, "learning_rate": 3.17881342448334e-07, "loss": 0.5691, "step": 2124 }, { "epoch": 0.8549587608127137, "grad_norm": 3.8241264163503255, "learning_rate": 3.1617017830952084e-07, "loss": 0.5375, "step": 2125 }, { "epoch": 0.8553610943472139, "grad_norm": 3.519525197941784, "learning_rate": 3.1446332128601015e-07, "loss": 0.4416, "step": 2126 }, { "epoch": 0.8557634278817139, "grad_norm": 3.4697912989841524, "learning_rate": 3.12760774744196e-07, "loss": 0.4459, "step": 2127 }, { "epoch": 0.856165761416214, "grad_norm": 3.7851796176468455, "learning_rate": 3.1106254204196883e-07, "loss": 0.4682, "step": 2128 }, { "epoch": 0.8565680949507142, "grad_norm": 3.936606771811409, "learning_rate": 3.093686265287138e-07, "loss": 0.5729, "step": 2129 }, { "epoch": 0.8569704284852142, "grad_norm": 4.190455136191216, "learning_rate": 3.076790315452988e-07, "loss": 0.5277, "step": 2130 }, { "epoch": 0.8573727620197144, "grad_norm": 4.108892268475147, "learning_rate": 3.05993760424072e-07, "loss": 0.4329, "step": 2131 }, { "epoch": 0.8577750955542145, "grad_norm": 4.390564938079403, "learning_rate": 3.043128164888537e-07, "loss": 0.5011, "step": 2132 }, { "epoch": 0.8581774290887145, "grad_norm": 4.367475352759402, "learning_rate": 3.0263620305492877e-07, "loss": 0.4797, "step": 2133 }, { "epoch": 0.8585797626232147, "grad_norm": 3.8444843612579924, "learning_rate": 3.009639234290418e-07, "loss": 0.5327, "step": 2134 }, { "epoch": 0.8589820961577147, "grad_norm": 3.6090606638950318, "learning_rate": 2.992959809093901e-07, "loss": 0.4413, "step": 2135 }, { "epoch": 0.8593844296922148, "grad_norm": 4.388230282282369, "learning_rate": 2.976323787856175e-07, "loss": 0.5717, "step": 2136 }, { "epoch": 0.859786763226715, "grad_norm": 4.333473286328121, "learning_rate": 2.9597312033880557e-07, "loss": 0.6314, "step": 2137 }, { "epoch": 0.860189096761215, "grad_norm": 3.890510795148687, "learning_rate": 2.943182088414701e-07, "loss": 0.433, "step": 2138 }, { "epoch": 0.8605914302957152, "grad_norm": 4.234653691147445, "learning_rate": 2.9266764755755406e-07, "loss": 0.54, "step": 2139 }, { "epoch": 0.8609937638302152, "grad_norm": 4.183758834087082, "learning_rate": 2.910214397424202e-07, "loss": 0.4845, "step": 2140 }, { "epoch": 0.8613960973647153, "grad_norm": 3.799668423286151, "learning_rate": 2.8937958864284353e-07, "loss": 0.4333, "step": 2141 }, { "epoch": 0.8617984308992155, "grad_norm": 4.133997388279442, "learning_rate": 2.877420974970088e-07, "loss": 0.4396, "step": 2142 }, { "epoch": 0.8622007644337155, "grad_norm": 4.0525547445688614, "learning_rate": 2.861089695345007e-07, "loss": 0.4857, "step": 2143 }, { "epoch": 0.8626030979682157, "grad_norm": 3.7880628956269904, "learning_rate": 2.8448020797629735e-07, "loss": 0.5046, "step": 2144 }, { "epoch": 0.8630054315027158, "grad_norm": 4.143267859487349, "learning_rate": 2.8285581603476747e-07, "loss": 0.5291, "step": 2145 }, { "epoch": 0.8634077650372158, "grad_norm": 3.8047765913147993, "learning_rate": 2.8123579691365923e-07, "loss": 0.5513, "step": 2146 }, { "epoch": 0.863810098571716, "grad_norm": 4.041711864076377, "learning_rate": 2.796201538080981e-07, "loss": 0.448, "step": 2147 }, { "epoch": 0.864212432106216, "grad_norm": 4.193683351365294, "learning_rate": 2.7800888990457855e-07, "loss": 0.5598, "step": 2148 }, { "epoch": 0.8646147656407162, "grad_norm": 3.6613935569874583, "learning_rate": 2.7640200838095793e-07, "loss": 0.4067, "step": 2149 }, { "epoch": 0.8650170991752163, "grad_norm": 4.104620267298875, "learning_rate": 2.7479951240644953e-07, "loss": 0.5648, "step": 2150 }, { "epoch": 0.8654194327097163, "grad_norm": 3.7600215255952256, "learning_rate": 2.7320140514161846e-07, "loss": 0.3946, "step": 2151 }, { "epoch": 0.8658217662442165, "grad_norm": 3.9525358112645965, "learning_rate": 2.7160768973837293e-07, "loss": 0.53, "step": 2152 }, { "epoch": 0.8662240997787165, "grad_norm": 3.5450533427675226, "learning_rate": 2.7001836933995995e-07, "loss": 0.5298, "step": 2153 }, { "epoch": 0.8666264333132166, "grad_norm": 3.6668666878628815, "learning_rate": 2.684334470809588e-07, "loss": 0.5056, "step": 2154 }, { "epoch": 0.8670287668477168, "grad_norm": 4.224237505147509, "learning_rate": 2.6685292608727236e-07, "loss": 0.5218, "step": 2155 }, { "epoch": 0.8674311003822168, "grad_norm": 3.69557439406065, "learning_rate": 2.652768094761257e-07, "loss": 0.3756, "step": 2156 }, { "epoch": 0.867833433916717, "grad_norm": 4.281090167368473, "learning_rate": 2.637051003560548e-07, "loss": 0.6402, "step": 2157 }, { "epoch": 0.8682357674512171, "grad_norm": 3.9633212522861054, "learning_rate": 2.621378018269047e-07, "loss": 0.5075, "step": 2158 }, { "epoch": 0.8686381009857171, "grad_norm": 3.8583945548771825, "learning_rate": 2.6057491697982e-07, "loss": 0.6033, "step": 2159 }, { "epoch": 0.8690404345202173, "grad_norm": 3.3558856738663487, "learning_rate": 2.5901644889724186e-07, "loss": 0.4015, "step": 2160 }, { "epoch": 0.8694427680547173, "grad_norm": 3.5749358804434483, "learning_rate": 2.574624006528992e-07, "loss": 0.4944, "step": 2161 }, { "epoch": 0.8698451015892175, "grad_norm": 3.86587826976141, "learning_rate": 2.55912775311804e-07, "loss": 0.4891, "step": 2162 }, { "epoch": 0.8702474351237176, "grad_norm": 3.9917581972004785, "learning_rate": 2.543675759302464e-07, "loss": 0.5079, "step": 2163 }, { "epoch": 0.8706497686582176, "grad_norm": 4.393976999836958, "learning_rate": 2.5282680555578477e-07, "loss": 0.6145, "step": 2164 }, { "epoch": 0.8710521021927178, "grad_norm": 4.04228191167184, "learning_rate": 2.51290467227244e-07, "loss": 0.5429, "step": 2165 }, { "epoch": 0.8714544357272178, "grad_norm": 4.4282484060475324, "learning_rate": 2.497585639747077e-07, "loss": 0.6325, "step": 2166 }, { "epoch": 0.871856769261718, "grad_norm": 3.816275160612967, "learning_rate": 2.482310988195125e-07, "loss": 0.5543, "step": 2167 }, { "epoch": 0.8722591027962181, "grad_norm": 4.1145602142866675, "learning_rate": 2.4670807477424007e-07, "loss": 0.5527, "step": 2168 }, { "epoch": 0.8726614363307181, "grad_norm": 4.231321602541556, "learning_rate": 2.4518949484271525e-07, "loss": 0.6496, "step": 2169 }, { "epoch": 0.8730637698652183, "grad_norm": 4.220523454555819, "learning_rate": 2.4367536201999695e-07, "loss": 0.5652, "step": 2170 }, { "epoch": 0.8734661033997184, "grad_norm": 3.938889256449136, "learning_rate": 2.421656792923724e-07, "loss": 0.456, "step": 2171 }, { "epoch": 0.8738684369342185, "grad_norm": 4.5217908297671485, "learning_rate": 2.406604496373535e-07, "loss": 0.6334, "step": 2172 }, { "epoch": 0.8742707704687186, "grad_norm": 3.7526929266944773, "learning_rate": 2.391596760236681e-07, "loss": 0.4743, "step": 2173 }, { "epoch": 0.8746731040032186, "grad_norm": 3.710705344410006, "learning_rate": 2.376633614112561e-07, "loss": 0.5661, "step": 2174 }, { "epoch": 0.8750754375377188, "grad_norm": 3.2609967694053243, "learning_rate": 2.361715087512631e-07, "loss": 0.3589, "step": 2175 }, { "epoch": 0.8754777710722189, "grad_norm": 4.108101136746485, "learning_rate": 2.346841209860351e-07, "loss": 0.4368, "step": 2176 }, { "epoch": 0.875880104606719, "grad_norm": 3.763267262463342, "learning_rate": 2.3320120104911053e-07, "loss": 0.4953, "step": 2177 }, { "epoch": 0.8762824381412191, "grad_norm": 5.025294007587999, "learning_rate": 2.317227518652168e-07, "loss": 0.6524, "step": 2178 }, { "epoch": 0.8766847716757191, "grad_norm": 3.9367012271057584, "learning_rate": 2.302487763502645e-07, "loss": 0.4149, "step": 2179 }, { "epoch": 0.8770871052102193, "grad_norm": 4.236893564539156, "learning_rate": 2.2877927741134036e-07, "loss": 0.5532, "step": 2180 }, { "epoch": 0.8774894387447194, "grad_norm": 3.5424218920413373, "learning_rate": 2.2731425794670135e-07, "loss": 0.4691, "step": 2181 }, { "epoch": 0.8778917722792194, "grad_norm": 4.04407474745495, "learning_rate": 2.2585372084577033e-07, "loss": 0.5125, "step": 2182 }, { "epoch": 0.8782941058137196, "grad_norm": 3.8612028242647147, "learning_rate": 2.2439766898913e-07, "loss": 0.5256, "step": 2183 }, { "epoch": 0.8786964393482197, "grad_norm": 4.271299959657279, "learning_rate": 2.2294610524851685e-07, "loss": 0.4944, "step": 2184 }, { "epoch": 0.8790987728827198, "grad_norm": 4.465135814281888, "learning_rate": 2.2149903248681464e-07, "loss": 0.6834, "step": 2185 }, { "epoch": 0.8795011064172199, "grad_norm": 3.9117028793633173, "learning_rate": 2.2005645355805127e-07, "loss": 0.4605, "step": 2186 }, { "epoch": 0.8799034399517199, "grad_norm": 3.8402220541360292, "learning_rate": 2.1861837130738945e-07, "loss": 0.4467, "step": 2187 }, { "epoch": 0.8803057734862201, "grad_norm": 4.223182944719613, "learning_rate": 2.1718478857112506e-07, "loss": 0.5347, "step": 2188 }, { "epoch": 0.8807081070207202, "grad_norm": 3.8811613151194213, "learning_rate": 2.1575570817667952e-07, "loss": 0.5265, "step": 2189 }, { "epoch": 0.8811104405552203, "grad_norm": 3.8705546986459956, "learning_rate": 2.143311329425926e-07, "loss": 0.4427, "step": 2190 }, { "epoch": 0.8815127740897204, "grad_norm": 4.5414937373063955, "learning_rate": 2.1291106567852105e-07, "loss": 0.6442, "step": 2191 }, { "epoch": 0.8819151076242204, "grad_norm": 3.588299387450043, "learning_rate": 2.1149550918522887e-07, "loss": 0.4173, "step": 2192 }, { "epoch": 0.8823174411587206, "grad_norm": 4.165225372551161, "learning_rate": 2.100844662545848e-07, "loss": 0.5496, "step": 2193 }, { "epoch": 0.8827197746932207, "grad_norm": 4.217297273907552, "learning_rate": 2.086779396695554e-07, "loss": 0.5329, "step": 2194 }, { "epoch": 0.8831221082277207, "grad_norm": 3.980212414160413, "learning_rate": 2.0727593220419812e-07, "loss": 0.4682, "step": 2195 }, { "epoch": 0.8835244417622209, "grad_norm": 4.015069610064507, "learning_rate": 2.0587844662365963e-07, "loss": 0.4975, "step": 2196 }, { "epoch": 0.883926775296721, "grad_norm": 3.675033120744641, "learning_rate": 2.0448548568416715e-07, "loss": 0.4932, "step": 2197 }, { "epoch": 0.8843291088312211, "grad_norm": 3.7917456823069178, "learning_rate": 2.030970521330247e-07, "loss": 0.4523, "step": 2198 }, { "epoch": 0.8847314423657212, "grad_norm": 4.2084887531102195, "learning_rate": 2.0171314870860548e-07, "loss": 0.5797, "step": 2199 }, { "epoch": 0.8851337759002212, "grad_norm": 4.1920678247066245, "learning_rate": 2.0033377814034999e-07, "loss": 0.5458, "step": 2200 }, { "epoch": 0.8855361094347214, "grad_norm": 4.341033030777012, "learning_rate": 1.9895894314875707e-07, "loss": 0.5069, "step": 2201 }, { "epoch": 0.8859384429692215, "grad_norm": 4.251489033634488, "learning_rate": 1.9758864644538188e-07, "loss": 0.5658, "step": 2202 }, { "epoch": 0.8863407765037216, "grad_norm": 4.118913583140747, "learning_rate": 1.9622289073282675e-07, "loss": 0.5186, "step": 2203 }, { "epoch": 0.8867431100382217, "grad_norm": 4.17198274381597, "learning_rate": 1.9486167870473915e-07, "loss": 0.5394, "step": 2204 }, { "epoch": 0.8871454435727217, "grad_norm": 4.1140466186765865, "learning_rate": 1.9350501304580577e-07, "loss": 0.6302, "step": 2205 }, { "epoch": 0.8875477771072219, "grad_norm": 4.252952260852059, "learning_rate": 1.9215289643174485e-07, "loss": 0.5715, "step": 2206 }, { "epoch": 0.887950110641722, "grad_norm": 3.9053915680472673, "learning_rate": 1.9080533152930486e-07, "loss": 0.502, "step": 2207 }, { "epoch": 0.8883524441762221, "grad_norm": 3.702225135993101, "learning_rate": 1.8946232099625482e-07, "loss": 0.3917, "step": 2208 }, { "epoch": 0.8887547777107222, "grad_norm": 3.5875962058050863, "learning_rate": 1.8812386748138234e-07, "loss": 0.3413, "step": 2209 }, { "epoch": 0.8891571112452223, "grad_norm": 3.6894845944247723, "learning_rate": 1.8678997362448743e-07, "loss": 0.4334, "step": 2210 }, { "epoch": 0.8895594447797224, "grad_norm": 3.977873097022255, "learning_rate": 1.8546064205637758e-07, "loss": 0.4746, "step": 2211 }, { "epoch": 0.8899617783142225, "grad_norm": 3.4865669801340147, "learning_rate": 1.8413587539886064e-07, "loss": 0.4173, "step": 2212 }, { "epoch": 0.8903641118487226, "grad_norm": 4.35862373407535, "learning_rate": 1.828156762647429e-07, "loss": 0.5653, "step": 2213 }, { "epoch": 0.8907664453832227, "grad_norm": 4.2652495550914775, "learning_rate": 1.81500047257821e-07, "loss": 0.5071, "step": 2214 }, { "epoch": 0.8911687789177228, "grad_norm": 3.285190042692873, "learning_rate": 1.8018899097287906e-07, "loss": 0.3731, "step": 2215 }, { "epoch": 0.8915711124522229, "grad_norm": 3.782205241606253, "learning_rate": 1.78882509995682e-07, "loss": 0.4907, "step": 2216 }, { "epoch": 0.891973445986723, "grad_norm": 4.391188417961942, "learning_rate": 1.7758060690297035e-07, "loss": 0.3594, "step": 2217 }, { "epoch": 0.892375779521223, "grad_norm": 3.172977998916911, "learning_rate": 1.7628328426245677e-07, "loss": 0.3866, "step": 2218 }, { "epoch": 0.8927781130557232, "grad_norm": 4.109857114119739, "learning_rate": 1.7499054463281985e-07, "loss": 0.5489, "step": 2219 }, { "epoch": 0.8931804465902233, "grad_norm": 3.6387926247562214, "learning_rate": 1.7370239056369946e-07, "loss": 0.3892, "step": 2220 }, { "epoch": 0.8935827801247234, "grad_norm": 3.897940749417438, "learning_rate": 1.7241882459569008e-07, "loss": 0.6287, "step": 2221 }, { "epoch": 0.8939851136592235, "grad_norm": 3.968476012728493, "learning_rate": 1.7113984926033832e-07, "loss": 0.459, "step": 2222 }, { "epoch": 0.8943874471937237, "grad_norm": 4.209938898766111, "learning_rate": 1.6986546708013722e-07, "loss": 0.5426, "step": 2223 }, { "epoch": 0.8947897807282237, "grad_norm": 3.791580194250541, "learning_rate": 1.6859568056852016e-07, "loss": 0.498, "step": 2224 }, { "epoch": 0.8951921142627238, "grad_norm": 3.8639160637014665, "learning_rate": 1.673304922298563e-07, "loss": 0.4049, "step": 2225 }, { "epoch": 0.8955944477972239, "grad_norm": 4.147228062957371, "learning_rate": 1.6606990455944634e-07, "loss": 0.5258, "step": 2226 }, { "epoch": 0.895996781331724, "grad_norm": 4.247308433616599, "learning_rate": 1.648139200435178e-07, "loss": 0.5252, "step": 2227 }, { "epoch": 0.8963991148662241, "grad_norm": 4.057558052673904, "learning_rate": 1.635625411592179e-07, "loss": 0.4851, "step": 2228 }, { "epoch": 0.8968014484007242, "grad_norm": 4.529371570037193, "learning_rate": 1.6231577037461228e-07, "loss": 0.6091, "step": 2229 }, { "epoch": 0.8972037819352243, "grad_norm": 4.092889891063284, "learning_rate": 1.6107361014867622e-07, "loss": 0.5051, "step": 2230 }, { "epoch": 0.8976061154697244, "grad_norm": 4.057510211081129, "learning_rate": 1.598360629312931e-07, "loss": 0.5379, "step": 2231 }, { "epoch": 0.8980084490042245, "grad_norm": 4.08275289824886, "learning_rate": 1.586031311632477e-07, "loss": 0.5542, "step": 2232 }, { "epoch": 0.8984107825387246, "grad_norm": 5.7116466876850875, "learning_rate": 1.5737481727622212e-07, "loss": 0.5021, "step": 2233 }, { "epoch": 0.8988131160732247, "grad_norm": 3.507507523132951, "learning_rate": 1.5615112369278957e-07, "loss": 0.391, "step": 2234 }, { "epoch": 0.8992154496077248, "grad_norm": 3.7113726990024194, "learning_rate": 1.5493205282641228e-07, "loss": 0.4515, "step": 2235 }, { "epoch": 0.899617783142225, "grad_norm": 3.512898207115907, "learning_rate": 1.5371760708143447e-07, "loss": 0.4935, "step": 2236 }, { "epoch": 0.900020116676725, "grad_norm": 4.213491522925021, "learning_rate": 1.525077888530782e-07, "loss": 0.538, "step": 2237 }, { "epoch": 0.9004224502112251, "grad_norm": 4.482285849278304, "learning_rate": 1.5130260052743955e-07, "loss": 0.6545, "step": 2238 }, { "epoch": 0.9008247837457252, "grad_norm": 3.7492718480617384, "learning_rate": 1.5010204448148214e-07, "loss": 0.4442, "step": 2239 }, { "epoch": 0.9012271172802253, "grad_norm": 3.9838020107249195, "learning_rate": 1.489061230830338e-07, "loss": 0.4953, "step": 2240 }, { "epoch": 0.9016294508147255, "grad_norm": 3.9003308620272796, "learning_rate": 1.4771483869078228e-07, "loss": 0.5208, "step": 2241 }, { "epoch": 0.9020317843492255, "grad_norm": 3.788631062888711, "learning_rate": 1.4652819365426867e-07, "loss": 0.5613, "step": 2242 }, { "epoch": 0.9024341178837256, "grad_norm": 4.676276663949487, "learning_rate": 1.453461903138853e-07, "loss": 0.5352, "step": 2243 }, { "epoch": 0.9028364514182257, "grad_norm": 4.525346271987768, "learning_rate": 1.441688310008682e-07, "loss": 0.5099, "step": 2244 }, { "epoch": 0.9032387849527258, "grad_norm": 3.625199458865597, "learning_rate": 1.4299611803729547e-07, "loss": 0.4766, "step": 2245 }, { "epoch": 0.903641118487226, "grad_norm": 3.8753089109297787, "learning_rate": 1.4182805373608055e-07, "loss": 0.5326, "step": 2246 }, { "epoch": 0.904043452021726, "grad_norm": 4.021228834417677, "learning_rate": 1.4066464040096956e-07, "loss": 0.5691, "step": 2247 }, { "epoch": 0.9044457855562261, "grad_norm": 4.6752863808538105, "learning_rate": 1.395058803265334e-07, "loss": 0.7598, "step": 2248 }, { "epoch": 0.9048481190907263, "grad_norm": 3.7179471187703514, "learning_rate": 1.3835177579816783e-07, "loss": 0.3955, "step": 2249 }, { "epoch": 0.9052504526252263, "grad_norm": 5.00701237752423, "learning_rate": 1.3720232909208541e-07, "loss": 0.6319, "step": 2250 }, { "epoch": 0.9056527861597264, "grad_norm": 3.7834417036951895, "learning_rate": 1.3605754247531245e-07, "loss": 0.5174, "step": 2251 }, { "epoch": 0.9060551196942265, "grad_norm": 3.680557056193272, "learning_rate": 1.3491741820568366e-07, "loss": 0.4133, "step": 2252 }, { "epoch": 0.9064574532287266, "grad_norm": 4.046295696144512, "learning_rate": 1.337819585318395e-07, "loss": 0.6443, "step": 2253 }, { "epoch": 0.9068597867632268, "grad_norm": 4.363328152667618, "learning_rate": 1.3265116569322e-07, "loss": 0.5437, "step": 2254 }, { "epoch": 0.9072621202977268, "grad_norm": 4.437499912355556, "learning_rate": 1.3152504192006005e-07, "loss": 0.5439, "step": 2255 }, { "epoch": 0.9076644538322269, "grad_norm": 3.904533797833089, "learning_rate": 1.3040358943338742e-07, "loss": 0.4058, "step": 2256 }, { "epoch": 0.908066787366727, "grad_norm": 3.990363526582056, "learning_rate": 1.2928681044501513e-07, "loss": 0.4204, "step": 2257 }, { "epoch": 0.9084691209012271, "grad_norm": 4.0760801231394606, "learning_rate": 1.281747071575404e-07, "loss": 0.6104, "step": 2258 }, { "epoch": 0.9088714544357273, "grad_norm": 4.256580027345346, "learning_rate": 1.270672817643376e-07, "loss": 0.6544, "step": 2259 }, { "epoch": 0.9092737879702273, "grad_norm": 4.71579332313011, "learning_rate": 1.2596453644955597e-07, "loss": 0.6273, "step": 2260 }, { "epoch": 0.9096761215047274, "grad_norm": 3.483958223270458, "learning_rate": 1.2486647338811248e-07, "loss": 0.4546, "step": 2261 }, { "epoch": 0.9100784550392276, "grad_norm": 3.9093619147070946, "learning_rate": 1.2377309474569165e-07, "loss": 0.4985, "step": 2262 }, { "epoch": 0.9104807885737276, "grad_norm": 4.057320495855987, "learning_rate": 1.226844026787377e-07, "loss": 0.5448, "step": 2263 }, { "epoch": 0.9108831221082277, "grad_norm": 3.5360903799800503, "learning_rate": 1.2160039933445222e-07, "loss": 0.4164, "step": 2264 }, { "epoch": 0.9112854556427278, "grad_norm": 4.087420290794404, "learning_rate": 1.2052108685078834e-07, "loss": 0.594, "step": 2265 }, { "epoch": 0.9116877891772279, "grad_norm": 3.9832710853164763, "learning_rate": 1.1944646735644882e-07, "loss": 0.452, "step": 2266 }, { "epoch": 0.9120901227117281, "grad_norm": 4.282566514995135, "learning_rate": 1.1837654297087997e-07, "loss": 0.5473, "step": 2267 }, { "epoch": 0.9124924562462281, "grad_norm": 3.9435273915235656, "learning_rate": 1.1731131580426796e-07, "loss": 0.6224, "step": 2268 }, { "epoch": 0.9128947897807282, "grad_norm": 4.3103292605724155, "learning_rate": 1.1625078795753476e-07, "loss": 0.5467, "step": 2269 }, { "epoch": 0.9132971233152283, "grad_norm": 4.22573396631058, "learning_rate": 1.1519496152233439e-07, "loss": 0.5037, "step": 2270 }, { "epoch": 0.9136994568497284, "grad_norm": 3.7886525775152338, "learning_rate": 1.1414383858104756e-07, "loss": 0.4498, "step": 2271 }, { "epoch": 0.9141017903842286, "grad_norm": 3.2224197154073617, "learning_rate": 1.1309742120677897e-07, "loss": 0.4794, "step": 2272 }, { "epoch": 0.9145041239187286, "grad_norm": 4.705013349470272, "learning_rate": 1.1205571146335303e-07, "loss": 0.6041, "step": 2273 }, { "epoch": 0.9149064574532287, "grad_norm": 4.6403427413148535, "learning_rate": 1.1101871140530851e-07, "loss": 0.6308, "step": 2274 }, { "epoch": 0.9153087909877289, "grad_norm": 3.819381857552772, "learning_rate": 1.0998642307789576e-07, "loss": 0.5461, "step": 2275 }, { "epoch": 0.9157111245222289, "grad_norm": 3.997200472540085, "learning_rate": 1.0895884851707255e-07, "loss": 0.4926, "step": 2276 }, { "epoch": 0.9161134580567291, "grad_norm": 4.162425768576248, "learning_rate": 1.079359897494997e-07, "loss": 0.489, "step": 2277 }, { "epoch": 0.9165157915912291, "grad_norm": 3.8422971328140454, "learning_rate": 1.0691784879253736e-07, "loss": 0.5291, "step": 2278 }, { "epoch": 0.9169181251257292, "grad_norm": 4.112928250644688, "learning_rate": 1.0590442765423953e-07, "loss": 0.4926, "step": 2279 }, { "epoch": 0.9173204586602294, "grad_norm": 3.9378155981778664, "learning_rate": 1.0489572833335349e-07, "loss": 0.4722, "step": 2280 }, { "epoch": 0.9177227921947294, "grad_norm": 4.631927118939465, "learning_rate": 1.038917528193123e-07, "loss": 0.5274, "step": 2281 }, { "epoch": 0.9181251257292296, "grad_norm": 4.035377760065258, "learning_rate": 1.0289250309223314e-07, "loss": 0.5788, "step": 2282 }, { "epoch": 0.9185274592637296, "grad_norm": 3.868825647605808, "learning_rate": 1.0189798112291177e-07, "loss": 0.5046, "step": 2283 }, { "epoch": 0.9189297927982297, "grad_norm": 4.194215283322997, "learning_rate": 1.0090818887282112e-07, "loss": 0.6029, "step": 2284 }, { "epoch": 0.9193321263327299, "grad_norm": 4.215786203028234, "learning_rate": 9.99231282941035e-08, "loss": 0.4987, "step": 2285 }, { "epoch": 0.9197344598672299, "grad_norm": 4.911243421308879, "learning_rate": 9.894280132957124e-08, "loss": 0.6475, "step": 2286 }, { "epoch": 0.92013679340173, "grad_norm": 4.551350447477857, "learning_rate": 9.79672099126991e-08, "loss": 0.6439, "step": 2287 }, { "epoch": 0.9205391269362302, "grad_norm": 4.0194710398308615, "learning_rate": 9.69963559676232e-08, "loss": 0.5376, "step": 2288 }, { "epoch": 0.9209414604707302, "grad_norm": 3.83021269408099, "learning_rate": 9.60302414091352e-08, "loss": 0.5093, "step": 2289 }, { "epoch": 0.9213437940052304, "grad_norm": 4.1940419822947534, "learning_rate": 9.506886814268007e-08, "loss": 0.5702, "step": 2290 }, { "epoch": 0.9217461275397304, "grad_norm": 4.341103613511488, "learning_rate": 9.41122380643511e-08, "loss": 0.5271, "step": 2291 }, { "epoch": 0.9221484610742305, "grad_norm": 4.421018842084841, "learning_rate": 9.316035306088684e-08, "loss": 0.5923, "step": 2292 }, { "epoch": 0.9225507946087307, "grad_norm": 4.192150302264493, "learning_rate": 9.221321500966746e-08, "loss": 0.5475, "step": 2293 }, { "epoch": 0.9229531281432307, "grad_norm": 4.56023548088152, "learning_rate": 9.127082577871038e-08, "loss": 0.6055, "step": 2294 }, { "epoch": 0.9233554616777309, "grad_norm": 3.727375694339219, "learning_rate": 9.033318722666801e-08, "loss": 0.3706, "step": 2295 }, { "epoch": 0.9237577952122309, "grad_norm": 3.7651015398148897, "learning_rate": 8.940030120282162e-08, "loss": 0.5118, "step": 2296 }, { "epoch": 0.924160128746731, "grad_norm": 4.305229814421829, "learning_rate": 8.847216954708055e-08, "loss": 0.5453, "step": 2297 }, { "epoch": 0.9245624622812312, "grad_norm": 4.1747564223004066, "learning_rate": 8.754879408997636e-08, "loss": 0.5156, "step": 2298 }, { "epoch": 0.9249647958157312, "grad_norm": 3.649399704340268, "learning_rate": 8.663017665266032e-08, "loss": 0.3701, "step": 2299 }, { "epoch": 0.9253671293502314, "grad_norm": 3.7596577501273734, "learning_rate": 8.571631904690014e-08, "loss": 0.4399, "step": 2300 }, { "epoch": 0.9257694628847315, "grad_norm": 4.230082435346815, "learning_rate": 8.480722307507461e-08, "loss": 0.4874, "step": 2301 }, { "epoch": 0.9261717964192315, "grad_norm": 3.9612627542935197, "learning_rate": 8.390289053017226e-08, "loss": 0.5485, "step": 2302 }, { "epoch": 0.9265741299537317, "grad_norm": 3.5438909670467003, "learning_rate": 8.300332319578664e-08, "loss": 0.3951, "step": 2303 }, { "epoch": 0.9269764634882317, "grad_norm": 3.2833308636054217, "learning_rate": 8.210852284611299e-08, "loss": 0.407, "step": 2304 }, { "epoch": 0.9273787970227318, "grad_norm": 3.4444311785022865, "learning_rate": 8.121849124594405e-08, "loss": 0.3925, "step": 2305 }, { "epoch": 0.927781130557232, "grad_norm": 4.548198093339736, "learning_rate": 8.033323015066818e-08, "loss": 0.719, "step": 2306 }, { "epoch": 0.928183464091732, "grad_norm": 3.6123982321182284, "learning_rate": 7.945274130626484e-08, "loss": 0.3828, "step": 2307 }, { "epoch": 0.9285857976262322, "grad_norm": 3.177723392355348, "learning_rate": 7.857702644930133e-08, "loss": 0.353, "step": 2308 }, { "epoch": 0.9289881311607322, "grad_norm": 4.133100828896777, "learning_rate": 7.770608730692852e-08, "loss": 0.4977, "step": 2309 }, { "epoch": 0.9293904646952323, "grad_norm": 3.562342202197354, "learning_rate": 7.683992559687881e-08, "loss": 0.3892, "step": 2310 }, { "epoch": 0.9297927982297325, "grad_norm": 3.898366854418511, "learning_rate": 7.597854302746288e-08, "loss": 0.4583, "step": 2311 }, { "epoch": 0.9301951317642325, "grad_norm": 4.612414420760638, "learning_rate": 7.512194129756428e-08, "loss": 0.5581, "step": 2312 }, { "epoch": 0.9305974652987327, "grad_norm": 3.988029769560431, "learning_rate": 7.42701220966388e-08, "loss": 0.4744, "step": 2313 }, { "epoch": 0.9309997988332327, "grad_norm": 4.481476524243926, "learning_rate": 7.342308710470814e-08, "loss": 0.59, "step": 2314 }, { "epoch": 0.9314021323677328, "grad_norm": 4.562506876878105, "learning_rate": 7.258083799235982e-08, "loss": 0.6742, "step": 2315 }, { "epoch": 0.931804465902233, "grad_norm": 4.628598132831451, "learning_rate": 7.174337642074147e-08, "loss": 0.6008, "step": 2316 }, { "epoch": 0.932206799436733, "grad_norm": 4.283379928862786, "learning_rate": 7.091070404155854e-08, "loss": 0.5053, "step": 2317 }, { "epoch": 0.9326091329712332, "grad_norm": 3.9776126177559155, "learning_rate": 7.008282249707065e-08, "loss": 0.4839, "step": 2318 }, { "epoch": 0.9330114665057333, "grad_norm": 4.069433607383545, "learning_rate": 6.925973342008897e-08, "loss": 0.453, "step": 2319 }, { "epoch": 0.9334138000402333, "grad_norm": 4.037972743715134, "learning_rate": 6.844143843397238e-08, "loss": 0.5427, "step": 2320 }, { "epoch": 0.9338161335747335, "grad_norm": 3.547048708842676, "learning_rate": 6.76279391526241e-08, "loss": 0.4561, "step": 2321 }, { "epoch": 0.9342184671092335, "grad_norm": 3.7831575419487256, "learning_rate": 6.681923718049016e-08, "loss": 0.5184, "step": 2322 }, { "epoch": 0.9346208006437337, "grad_norm": 3.7960605831677428, "learning_rate": 6.601533411255328e-08, "loss": 0.428, "step": 2323 }, { "epoch": 0.9350231341782338, "grad_norm": 3.699019611927255, "learning_rate": 6.52162315343327e-08, "loss": 0.4711, "step": 2324 }, { "epoch": 0.9354254677127338, "grad_norm": 3.7499929742301465, "learning_rate": 6.44219310218791e-08, "loss": 0.494, "step": 2325 }, { "epoch": 0.935827801247234, "grad_norm": 3.982088440613016, "learning_rate": 6.363243414177239e-08, "loss": 0.5013, "step": 2326 }, { "epoch": 0.936230134781734, "grad_norm": 3.5405567172339034, "learning_rate": 6.284774245111868e-08, "loss": 0.4151, "step": 2327 }, { "epoch": 0.9366324683162341, "grad_norm": 3.9069783277866894, "learning_rate": 6.206785749754613e-08, "loss": 0.5257, "step": 2328 }, { "epoch": 0.9370348018507343, "grad_norm": 4.560908800971963, "learning_rate": 6.12927808192032e-08, "loss": 0.5827, "step": 2329 }, { "epoch": 0.9374371353852343, "grad_norm": 3.6067893534884967, "learning_rate": 6.052251394475544e-08, "loss": 0.4533, "step": 2330 }, { "epoch": 0.9378394689197345, "grad_norm": 4.228506402107548, "learning_rate": 5.975705839338208e-08, "loss": 0.5367, "step": 2331 }, { "epoch": 0.9382418024542346, "grad_norm": 4.377928096595712, "learning_rate": 5.89964156747716e-08, "loss": 0.487, "step": 2332 }, { "epoch": 0.9386441359887346, "grad_norm": 4.001917632135577, "learning_rate": 5.8240587289122576e-08, "loss": 0.5991, "step": 2333 }, { "epoch": 0.9390464695232348, "grad_norm": 3.89102054624204, "learning_rate": 5.7489574727136734e-08, "loss": 0.5289, "step": 2334 }, { "epoch": 0.9394488030577348, "grad_norm": 3.9529925607216314, "learning_rate": 5.67433794700184e-08, "loss": 0.5516, "step": 2335 }, { "epoch": 0.939851136592235, "grad_norm": 3.9965015035833114, "learning_rate": 5.600200298947034e-08, "loss": 0.5758, "step": 2336 }, { "epoch": 0.9402534701267351, "grad_norm": 4.354462501990094, "learning_rate": 5.52654467476918e-08, "loss": 0.5793, "step": 2337 }, { "epoch": 0.9406558036612351, "grad_norm": 4.199766571714081, "learning_rate": 5.453371219737491e-08, "loss": 0.5688, "step": 2338 }, { "epoch": 0.9410581371957353, "grad_norm": 3.3551684650390454, "learning_rate": 5.380680078170275e-08, "loss": 0.3524, "step": 2339 }, { "epoch": 0.9414604707302353, "grad_norm": 4.408645627917106, "learning_rate": 5.3084713934344613e-08, "loss": 0.5442, "step": 2340 }, { "epoch": 0.9418628042647355, "grad_norm": 4.4356672674608895, "learning_rate": 5.236745307945545e-08, "loss": 0.5028, "step": 2341 }, { "epoch": 0.9422651377992356, "grad_norm": 3.46236072019421, "learning_rate": 5.165501963167174e-08, "loss": 0.3921, "step": 2342 }, { "epoch": 0.9426674713337356, "grad_norm": 4.4269078668736075, "learning_rate": 5.094741499610839e-08, "loss": 0.5983, "step": 2343 }, { "epoch": 0.9430698048682358, "grad_norm": 4.503787561410546, "learning_rate": 5.02446405683582e-08, "loss": 0.585, "step": 2344 }, { "epoch": 0.9434721384027359, "grad_norm": 3.658547311389223, "learning_rate": 4.9546697734485224e-08, "loss": 0.4762, "step": 2345 }, { "epoch": 0.943874471937236, "grad_norm": 4.353601396348682, "learning_rate": 4.885358787102584e-08, "loss": 0.6102, "step": 2346 }, { "epoch": 0.9442768054717361, "grad_norm": 4.143732790296102, "learning_rate": 4.816531234498406e-08, "loss": 0.6124, "step": 2347 }, { "epoch": 0.9446791390062361, "grad_norm": 4.3975559021632575, "learning_rate": 4.7481872513829575e-08, "loss": 0.5633, "step": 2348 }, { "epoch": 0.9450814725407363, "grad_norm": 4.066972538920918, "learning_rate": 4.680326972549387e-08, "loss": 0.5118, "step": 2349 }, { "epoch": 0.9454838060752364, "grad_norm": 4.03878765773838, "learning_rate": 4.6129505318369127e-08, "loss": 0.4778, "step": 2350 }, { "epoch": 0.9458861396097364, "grad_norm": 4.201321170733065, "learning_rate": 4.546058062130487e-08, "loss": 0.5128, "step": 2351 }, { "epoch": 0.9462884731442366, "grad_norm": 3.8094183727880395, "learning_rate": 4.4796496953605493e-08, "loss": 0.5441, "step": 2352 }, { "epoch": 0.9466908066787366, "grad_norm": 3.5299699606250834, "learning_rate": 4.413725562502691e-08, "loss": 0.4678, "step": 2353 }, { "epoch": 0.9470931402132368, "grad_norm": 3.9087226991955717, "learning_rate": 4.348285793577517e-08, "loss": 0.5329, "step": 2354 }, { "epoch": 0.9474954737477369, "grad_norm": 4.0121109992844755, "learning_rate": 4.2833305176503136e-08, "loss": 0.5837, "step": 2355 }, { "epoch": 0.9478978072822369, "grad_norm": 4.544145628732853, "learning_rate": 4.218859862830826e-08, "loss": 0.3829, "step": 2356 }, { "epoch": 0.9483001408167371, "grad_norm": 3.89076415048523, "learning_rate": 4.154873956272981e-08, "loss": 0.4691, "step": 2357 }, { "epoch": 0.9487024743512372, "grad_norm": 4.1725175451999, "learning_rate": 4.091372924174636e-08, "loss": 0.4885, "step": 2358 }, { "epoch": 0.9491048078857373, "grad_norm": 4.3450167781713285, "learning_rate": 4.028356891777385e-08, "loss": 0.5124, "step": 2359 }, { "epoch": 0.9495071414202374, "grad_norm": 4.4322372412295605, "learning_rate": 3.9658259833662284e-08, "loss": 0.5234, "step": 2360 }, { "epoch": 0.9499094749547374, "grad_norm": 3.707524532117256, "learning_rate": 3.903780322269374e-08, "loss": 0.4149, "step": 2361 }, { "epoch": 0.9503118084892376, "grad_norm": 5.065679790929374, "learning_rate": 3.842220030858018e-08, "loss": 0.688, "step": 2362 }, { "epoch": 0.9507141420237377, "grad_norm": 3.786912616101777, "learning_rate": 3.781145230546013e-08, "loss": 0.43, "step": 2363 }, { "epoch": 0.9511164755582378, "grad_norm": 3.6227665097152224, "learning_rate": 3.720556041789752e-08, "loss": 0.4382, "step": 2364 }, { "epoch": 0.9515188090927379, "grad_norm": 3.999125768684484, "learning_rate": 3.660452584087815e-08, "loss": 0.5716, "step": 2365 }, { "epoch": 0.9519211426272379, "grad_norm": 3.409457081881056, "learning_rate": 3.600834975980877e-08, "loss": 0.5407, "step": 2366 }, { "epoch": 0.9523234761617381, "grad_norm": 3.892658686650397, "learning_rate": 3.5417033350512444e-08, "loss": 0.4589, "step": 2367 }, { "epoch": 0.9527258096962382, "grad_norm": 4.000437062384583, "learning_rate": 3.483057777922905e-08, "loss": 0.4875, "step": 2368 }, { "epoch": 0.9531281432307382, "grad_norm": 4.046666707315621, "learning_rate": 3.42489842026103e-08, "loss": 0.5162, "step": 2369 }, { "epoch": 0.9535304767652384, "grad_norm": 4.290796216817974, "learning_rate": 3.3672253767719485e-08, "loss": 0.6233, "step": 2370 }, { "epoch": 0.9539328102997385, "grad_norm": 4.023014608954607, "learning_rate": 3.310038761202839e-08, "loss": 0.5062, "step": 2371 }, { "epoch": 0.9543351438342386, "grad_norm": 4.43210170452369, "learning_rate": 3.2533386863414784e-08, "loss": 0.5594, "step": 2372 }, { "epoch": 0.9547374773687387, "grad_norm": 3.8979482975499624, "learning_rate": 3.1971252640160545e-08, "loss": 0.5128, "step": 2373 }, { "epoch": 0.9551398109032387, "grad_norm": 4.162386221402383, "learning_rate": 3.141398605094992e-08, "loss": 0.5848, "step": 2374 }, { "epoch": 0.9555421444377389, "grad_norm": 4.022603602017269, "learning_rate": 3.086158819486651e-08, "loss": 0.5771, "step": 2375 }, { "epoch": 0.955944477972239, "grad_norm": 3.8920410241916863, "learning_rate": 3.031406016139077e-08, "loss": 0.5498, "step": 2376 }, { "epoch": 0.9563468115067391, "grad_norm": 4.084788242660989, "learning_rate": 2.9771403030399713e-08, "loss": 0.598, "step": 2377 }, { "epoch": 0.9567491450412392, "grad_norm": 4.048190861072185, "learning_rate": 2.923361787216278e-08, "loss": 0.4405, "step": 2378 }, { "epoch": 0.9571514785757392, "grad_norm": 4.502700834332798, "learning_rate": 2.8700705747340963e-08, "loss": 0.5169, "step": 2379 }, { "epoch": 0.9575538121102394, "grad_norm": 4.054509620797073, "learning_rate": 2.8172667706983792e-08, "loss": 0.53, "step": 2380 }, { "epoch": 0.9579561456447395, "grad_norm": 3.939484097293469, "learning_rate": 2.7649504792528204e-08, "loss": 0.5566, "step": 2381 }, { "epoch": 0.9583584791792396, "grad_norm": 4.299563624727568, "learning_rate": 2.7131218035796324e-08, "loss": 0.5603, "step": 2382 }, { "epoch": 0.9587608127137397, "grad_norm": 3.723520114625019, "learning_rate": 2.6617808458991867e-08, "loss": 0.4346, "step": 2383 }, { "epoch": 0.9591631462482398, "grad_norm": 3.667012045051573, "learning_rate": 2.6109277074700955e-08, "loss": 0.3638, "step": 2384 }, { "epoch": 0.9595654797827399, "grad_norm": 3.725825844080702, "learning_rate": 2.5605624885887414e-08, "loss": 0.4541, "step": 2385 }, { "epoch": 0.95996781331724, "grad_norm": 3.7094680890029776, "learning_rate": 2.5106852885892764e-08, "loss": 0.4693, "step": 2386 }, { "epoch": 0.96037014685174, "grad_norm": 3.7229547246513937, "learning_rate": 2.461296205843261e-08, "loss": 0.5144, "step": 2387 }, { "epoch": 0.9607724803862402, "grad_norm": 4.159612295768075, "learning_rate": 2.412395337759693e-08, "loss": 0.522, "step": 2388 }, { "epoch": 0.9611748139207403, "grad_norm": 3.8986140359372286, "learning_rate": 2.3639827807844796e-08, "loss": 0.5237, "step": 2389 }, { "epoch": 0.9615771474552404, "grad_norm": 4.121244015419092, "learning_rate": 2.3160586304006026e-08, "loss": 0.538, "step": 2390 }, { "epoch": 0.9619794809897405, "grad_norm": 4.166559231330322, "learning_rate": 2.268622981127705e-08, "loss": 0.5822, "step": 2391 }, { "epoch": 0.9623818145242405, "grad_norm": 3.9211071919172813, "learning_rate": 2.2216759265220044e-08, "loss": 0.4897, "step": 2392 }, { "epoch": 0.9627841480587407, "grad_norm": 4.349097825226008, "learning_rate": 2.1752175591760184e-08, "loss": 0.5211, "step": 2393 }, { "epoch": 0.9631864815932408, "grad_norm": 3.9282047251137677, "learning_rate": 2.1292479707184787e-08, "loss": 0.5741, "step": 2394 }, { "epoch": 0.9635888151277409, "grad_norm": 4.159989263744697, "learning_rate": 2.0837672518141395e-08, "loss": 0.5238, "step": 2395 }, { "epoch": 0.963991148662241, "grad_norm": 3.8198728228381293, "learning_rate": 2.0387754921634972e-08, "loss": 0.5144, "step": 2396 }, { "epoch": 0.9643934821967411, "grad_norm": 3.346255200741844, "learning_rate": 1.9942727805027373e-08, "loss": 0.4384, "step": 2397 }, { "epoch": 0.9647958157312412, "grad_norm": 3.8469019196516845, "learning_rate": 1.9502592046035107e-08, "loss": 0.486, "step": 2398 }, { "epoch": 0.9651981492657413, "grad_norm": 3.5776773638982893, "learning_rate": 1.9067348512726847e-08, "loss": 0.4414, "step": 2399 }, { "epoch": 0.9656004828002414, "grad_norm": 4.031486537261451, "learning_rate": 1.8636998063523705e-08, "loss": 0.5328, "step": 2400 }, { "epoch": 0.9660028163347415, "grad_norm": 4.260664576753229, "learning_rate": 1.8211541547195067e-08, "loss": 0.5136, "step": 2401 }, { "epoch": 0.9664051498692416, "grad_norm": 4.152186034058617, "learning_rate": 1.7790979802858876e-08, "loss": 0.6316, "step": 2402 }, { "epoch": 0.9668074834037417, "grad_norm": 3.995357070807909, "learning_rate": 1.7375313659978855e-08, "loss": 0.5377, "step": 2403 }, { "epoch": 0.9672098169382418, "grad_norm": 3.4651763373212474, "learning_rate": 1.6964543938363388e-08, "loss": 0.4157, "step": 2404 }, { "epoch": 0.9676121504727418, "grad_norm": 4.360198459376172, "learning_rate": 1.6558671448163866e-08, "loss": 0.6273, "step": 2405 }, { "epoch": 0.968014484007242, "grad_norm": 4.006498747343065, "learning_rate": 1.6157696989873017e-08, "loss": 0.4914, "step": 2406 }, { "epoch": 0.9684168175417421, "grad_norm": 3.7398733702546822, "learning_rate": 1.5761621354322965e-08, "loss": 0.4022, "step": 2407 }, { "epoch": 0.9688191510762422, "grad_norm": 4.1110217996882294, "learning_rate": 1.5370445322684392e-08, "loss": 0.4698, "step": 2408 }, { "epoch": 0.9692214846107423, "grad_norm": 4.362254470529691, "learning_rate": 1.498416966646432e-08, "loss": 0.5598, "step": 2409 }, { "epoch": 0.9696238181452425, "grad_norm": 3.8880330136972328, "learning_rate": 1.4602795147504734e-08, "loss": 0.473, "step": 2410 }, { "epoch": 0.9700261516797425, "grad_norm": 4.110874787245324, "learning_rate": 1.4226322517981728e-08, "loss": 0.4901, "step": 2411 }, { "epoch": 0.9704284852142426, "grad_norm": 4.138814644855194, "learning_rate": 1.3854752520403303e-08, "loss": 0.5618, "step": 2412 }, { "epoch": 0.9708308187487427, "grad_norm": 4.481371755440245, "learning_rate": 1.3488085887607972e-08, "loss": 0.6152, "step": 2413 }, { "epoch": 0.9712331522832428, "grad_norm": 3.6050373310270176, "learning_rate": 1.3126323342763648e-08, "loss": 0.5525, "step": 2414 }, { "epoch": 0.971635485817743, "grad_norm": 4.061084112826385, "learning_rate": 1.2769465599365982e-08, "loss": 0.4619, "step": 2415 }, { "epoch": 0.972037819352243, "grad_norm": 3.8920028399978097, "learning_rate": 1.24175133612367e-08, "loss": 0.4833, "step": 2416 }, { "epoch": 0.9724401528867431, "grad_norm": 3.909035344776143, "learning_rate": 1.2070467322522762e-08, "loss": 0.5188, "step": 2417 }, { "epoch": 0.9728424864212432, "grad_norm": 4.225663392112197, "learning_rate": 1.1728328167695258e-08, "loss": 0.6696, "step": 2418 }, { "epoch": 0.9732448199557433, "grad_norm": 4.6741890030042725, "learning_rate": 1.1391096571546633e-08, "loss": 0.5815, "step": 2419 }, { "epoch": 0.9736471534902434, "grad_norm": 4.51608876386972, "learning_rate": 1.1058773199190965e-08, "loss": 0.5173, "step": 2420 }, { "epoch": 0.9740494870247435, "grad_norm": 4.0654982873308665, "learning_rate": 1.0731358706061456e-08, "loss": 0.5749, "step": 2421 }, { "epoch": 0.9744518205592436, "grad_norm": 4.2304161552160835, "learning_rate": 1.0408853737909896e-08, "loss": 0.5813, "step": 2422 }, { "epoch": 0.9748541540937438, "grad_norm": 3.4742799346124493, "learning_rate": 1.0091258930805259e-08, "loss": 0.4705, "step": 2423 }, { "epoch": 0.9752564876282438, "grad_norm": 4.08672880848899, "learning_rate": 9.778574911132044e-09, "loss": 0.516, "step": 2424 }, { "epoch": 0.9756588211627439, "grad_norm": 3.669930620914309, "learning_rate": 9.470802295589442e-09, "loss": 0.4138, "step": 2425 }, { "epoch": 0.976061154697244, "grad_norm": 3.5860479619434673, "learning_rate": 9.16794169119023e-09, "loss": 0.4224, "step": 2426 }, { "epoch": 0.9764634882317441, "grad_norm": 4.642054584241187, "learning_rate": 8.869993695258816e-09, "loss": 0.611, "step": 2427 }, { "epoch": 0.9768658217662443, "grad_norm": 3.977155407637389, "learning_rate": 8.576958895431531e-09, "loss": 0.5681, "step": 2428 }, { "epoch": 0.9772681553007443, "grad_norm": 3.6884328747774857, "learning_rate": 8.28883786965301e-09, "loss": 0.4945, "step": 2429 }, { "epoch": 0.9776704888352444, "grad_norm": 4.023502331465119, "learning_rate": 8.005631186177864e-09, "loss": 0.4201, "step": 2430 }, { "epoch": 0.9780728223697445, "grad_norm": 3.901800973146699, "learning_rate": 7.727339403568179e-09, "loss": 0.4977, "step": 2431 }, { "epoch": 0.9784751559042446, "grad_norm": 3.9967038572140856, "learning_rate": 7.45396307069185e-09, "loss": 0.6418, "step": 2432 }, { "epoch": 0.9788774894387448, "grad_norm": 4.291073567899429, "learning_rate": 7.185502726722305e-09, "loss": 0.6601, "step": 2433 }, { "epoch": 0.9792798229732448, "grad_norm": 3.974019553234897, "learning_rate": 6.921958901137948e-09, "loss": 0.4781, "step": 2434 }, { "epoch": 0.9796821565077449, "grad_norm": 3.833029352112665, "learning_rate": 6.66333211371939e-09, "loss": 0.4373, "step": 2435 }, { "epoch": 0.9800844900422451, "grad_norm": 4.398810707509668, "learning_rate": 6.409622874550547e-09, "loss": 0.5795, "step": 2436 }, { "epoch": 0.9804868235767451, "grad_norm": 3.5697882170733557, "learning_rate": 6.160831684015878e-09, "loss": 0.398, "step": 2437 }, { "epoch": 0.9808891571112452, "grad_norm": 3.904433602202901, "learning_rate": 5.9169590328003735e-09, "loss": 0.5029, "step": 2438 }, { "epoch": 0.9812914906457453, "grad_norm": 3.7353800882065715, "learning_rate": 5.678005401888176e-09, "loss": 0.4911, "step": 2439 }, { "epoch": 0.9816938241802454, "grad_norm": 3.795949925181833, "learning_rate": 5.443971262561465e-09, "loss": 0.4405, "step": 2440 }, { "epoch": 0.9820961577147456, "grad_norm": 3.617609885039579, "learning_rate": 5.214857076400736e-09, "loss": 0.4329, "step": 2441 }, { "epoch": 0.9824984912492456, "grad_norm": 3.5249358490195637, "learning_rate": 4.990663295282305e-09, "loss": 0.4285, "step": 2442 }, { "epoch": 0.9829008247837457, "grad_norm": 4.142775577116706, "learning_rate": 4.771390361377748e-09, "loss": 0.4777, "step": 2443 }, { "epoch": 0.9833031583182458, "grad_norm": 4.312344250176257, "learning_rate": 4.5570387071544595e-09, "loss": 0.6627, "step": 2444 }, { "epoch": 0.9837054918527459, "grad_norm": 3.67386973137855, "learning_rate": 4.347608755372601e-09, "loss": 0.4431, "step": 2445 }, { "epoch": 0.9841078253872461, "grad_norm": 4.191733748165821, "learning_rate": 4.143100919086207e-09, "loss": 0.4884, "step": 2446 }, { "epoch": 0.9845101589217461, "grad_norm": 4.38016174030762, "learning_rate": 3.943515601640968e-09, "loss": 0.5822, "step": 2447 }, { "epoch": 0.9849124924562462, "grad_norm": 4.227081529941771, "learning_rate": 3.748853196673952e-09, "loss": 0.49, "step": 2448 }, { "epoch": 0.9853148259907464, "grad_norm": 3.6169704926991124, "learning_rate": 3.5591140881138797e-09, "loss": 0.5642, "step": 2449 }, { "epoch": 0.9857171595252464, "grad_norm": 3.228283450712886, "learning_rate": 3.374298650178076e-09, "loss": 0.3926, "step": 2450 }, { "epoch": 0.9861194930597466, "grad_norm": 3.996805720599477, "learning_rate": 3.194407247373854e-09, "loss": 0.5175, "step": 2451 }, { "epoch": 0.9865218265942466, "grad_norm": 3.786750548322366, "learning_rate": 3.0194402344965733e-09, "loss": 0.4496, "step": 2452 }, { "epoch": 0.9869241601287467, "grad_norm": 3.730366038034417, "learning_rate": 2.8493979566296403e-09, "loss": 0.4701, "step": 2453 }, { "epoch": 0.9873264936632469, "grad_norm": 4.616677722561612, "learning_rate": 2.6842807491433974e-09, "loss": 0.6207, "step": 2454 }, { "epoch": 0.9877288271977469, "grad_norm": 3.9983804243694645, "learning_rate": 2.524088937694291e-09, "loss": 0.4734, "step": 2455 }, { "epoch": 0.988131160732247, "grad_norm": 4.364062652259154, "learning_rate": 2.3688228382251487e-09, "loss": 0.6254, "step": 2456 }, { "epoch": 0.9885334942667471, "grad_norm": 3.97492888774164, "learning_rate": 2.218482756963236e-09, "loss": 0.5645, "step": 2457 }, { "epoch": 0.9889358278012472, "grad_norm": 3.9671569540217493, "learning_rate": 2.07306899042109e-09, "loss": 0.4821, "step": 2458 }, { "epoch": 0.9893381613357474, "grad_norm": 3.808293451273474, "learning_rate": 1.9325818253945748e-09, "loss": 0.4502, "step": 2459 }, { "epoch": 0.9897404948702474, "grad_norm": 4.467378355634954, "learning_rate": 1.7970215389628842e-09, "loss": 0.5963, "step": 2460 }, { "epoch": 0.9901428284047475, "grad_norm": 4.625199727103312, "learning_rate": 1.6663883984888163e-09, "loss": 0.6419, "step": 2461 }, { "epoch": 0.9905451619392477, "grad_norm": 3.9326948253065184, "learning_rate": 1.5406826616168324e-09, "loss": 0.5575, "step": 2462 }, { "epoch": 0.9909474954737477, "grad_norm": 4.4568921518738485, "learning_rate": 1.4199045762730569e-09, "loss": 0.6483, "step": 2463 }, { "epoch": 0.9913498290082479, "grad_norm": 4.351427492104435, "learning_rate": 1.3040543806658313e-09, "loss": 0.5948, "step": 2464 }, { "epoch": 0.9917521625427479, "grad_norm": 3.4818450271582857, "learning_rate": 1.1931323032832177e-09, "loss": 0.3535, "step": 2465 }, { "epoch": 0.992154496077248, "grad_norm": 4.11227816177031, "learning_rate": 1.0871385628941077e-09, "loss": 0.5317, "step": 2466 }, { "epoch": 0.9925568296117482, "grad_norm": 4.418916055074932, "learning_rate": 9.860733685479463e-10, "loss": 0.6433, "step": 2467 }, { "epoch": 0.9929591631462482, "grad_norm": 3.6128250845259724, "learning_rate": 8.899369195727869e-10, "loss": 0.5447, "step": 2468 }, { "epoch": 0.9933614966807484, "grad_norm": 4.364090118207612, "learning_rate": 7.987294055761264e-10, "loss": 0.6099, "step": 2469 }, { "epoch": 0.9937638302152484, "grad_norm": 3.739315819157321, "learning_rate": 7.124510064446255e-10, "loss": 0.4998, "step": 2470 }, { "epoch": 0.9941661637497485, "grad_norm": 4.035011038094296, "learning_rate": 6.311018923432776e-10, "loss": 0.4558, "step": 2471 }, { "epoch": 0.9945684972842487, "grad_norm": 3.577550936704116, "learning_rate": 5.546822237145755e-10, "loss": 0.4408, "step": 2472 }, { "epoch": 0.9949708308187487, "grad_norm": 3.7748701163123584, "learning_rate": 4.831921512796211e-10, "loss": 0.4403, "step": 2473 }, { "epoch": 0.9953731643532489, "grad_norm": 3.7403701320267686, "learning_rate": 4.166318160361837e-10, "loss": 0.4989, "step": 2474 }, { "epoch": 0.995775497887749, "grad_norm": 3.8787870872673857, "learning_rate": 3.5500134925980925e-10, "loss": 0.5066, "step": 2475 }, { "epoch": 0.996177831422249, "grad_norm": 4.321954937291901, "learning_rate": 2.983008725029879e-10, "loss": 0.498, "step": 2476 }, { "epoch": 0.9965801649567492, "grad_norm": 3.574339539948393, "learning_rate": 2.4653049759432167e-10, "loss": 0.4164, "step": 2477 }, { "epoch": 0.9969824984912492, "grad_norm": 3.7965194640316478, "learning_rate": 1.9969032663935683e-10, "loss": 0.4769, "step": 2478 }, { "epoch": 0.9973848320257493, "grad_norm": 4.23022008583803, "learning_rate": 1.5778045202002878e-10, "loss": 0.4481, "step": 2479 }, { "epoch": 0.9977871655602495, "grad_norm": 4.071535851310751, "learning_rate": 1.2080095639410704e-10, "loss": 0.535, "step": 2480 }, { "epoch": 0.9981894990947495, "grad_norm": 3.9007108087469646, "learning_rate": 8.875191269547279e-11, "loss": 0.542, "step": 2481 }, { "epoch": 0.9985918326292497, "grad_norm": 4.05416067842318, "learning_rate": 6.163338413328612e-11, "loss": 0.5738, "step": 2482 }, { "epoch": 0.9989941661637497, "grad_norm": 3.9020981788420386, "learning_rate": 3.944542419337394e-11, "loss": 0.4897, "step": 2483 }, { "epoch": 0.9993964996982498, "grad_norm": 3.6120331011996085, "learning_rate": 2.2188076636009415e-11, "loss": 0.5195, "step": 2484 }, { "epoch": 0.99979883323275, "grad_norm": 3.4028120769775385, "learning_rate": 9.861375497577375e-12, "loss": 0.3763, "step": 2485 }, { "epoch": 1.0, "grad_norm": 3.4028120769775385, "learning_rate": 2.4653450900191666e-12, "loss": 0.4064, "step": 2486 }, { "epoch": 1.0, "step": 2486, "total_flos": 882515131957248.0, "train_loss": 0.5852410430066162, "train_runtime": 194010.5784, "train_samples_per_second": 1.64, "train_steps_per_second": 0.013 } ], "logging_steps": 1.0, "max_steps": 2486, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 882515131957248.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }