diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,17445 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 2486, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0004023335345001006, + "grad_norm": 27.82010551409744, + "learning_rate": 0.0, + "loss": 1.7886, + "step": 1 + }, + { + "epoch": 0.0008046670690002012, + "grad_norm": 19.872529474831897, + "learning_rate": 2.008032128514056e-08, + "loss": 1.8136, + "step": 2 + }, + { + "epoch": 0.0012070006035003018, + "grad_norm": 24.180123303383308, + "learning_rate": 4.016064257028112e-08, + "loss": 1.9299, + "step": 3 + }, + { + "epoch": 0.0016093341380004024, + "grad_norm": 21.489391381684396, + "learning_rate": 6.02409638554217e-08, + "loss": 1.92, + "step": 4 + }, + { + "epoch": 0.0020116676725005027, + "grad_norm": 17.558745295130834, + "learning_rate": 8.032128514056224e-08, + "loss": 1.8227, + "step": 5 + }, + { + "epoch": 0.0024140012070006035, + "grad_norm": 21.46699607887353, + "learning_rate": 1.0040160642570281e-07, + "loss": 1.8824, + "step": 6 + }, + { + "epoch": 0.002816334741500704, + "grad_norm": 21.298475273313784, + "learning_rate": 1.204819277108434e-07, + "loss": 1.8179, + "step": 7 + }, + { + "epoch": 0.0032186682760008047, + "grad_norm": 21.372007205869878, + "learning_rate": 1.4056224899598394e-07, + "loss": 1.6859, + "step": 8 + }, + { + "epoch": 0.003621001810500905, + "grad_norm": 23.292357598058903, + "learning_rate": 1.6064257028112448e-07, + "loss": 2.0856, + "step": 9 + }, + { + "epoch": 0.0040233353450010055, + "grad_norm": 23.233045584039576, + "learning_rate": 1.8072289156626505e-07, + "loss": 2.1395, + "step": 10 + }, + { + "epoch": 0.004425668879501106, + "grad_norm": 19.4337570491022, + "learning_rate": 2.0080321285140563e-07, + "loss": 1.8688, + "step": 11 + }, + { + "epoch": 0.004828002414001207, + "grad_norm": 24.82725637763247, + "learning_rate": 2.208835341365462e-07, + "loss": 1.8054, + "step": 12 + }, + { + "epoch": 0.005230335948501308, + "grad_norm": 27.369177322875704, + "learning_rate": 2.409638554216868e-07, + "loss": 1.8794, + "step": 13 + }, + { + "epoch": 0.005632669483001408, + "grad_norm": 20.854134960964704, + "learning_rate": 2.610441767068273e-07, + "loss": 1.8873, + "step": 14 + }, + { + "epoch": 0.006035003017501509, + "grad_norm": 25.685971689153487, + "learning_rate": 2.811244979919679e-07, + "loss": 2.0743, + "step": 15 + }, + { + "epoch": 0.0064373365520016095, + "grad_norm": 21.501547984115398, + "learning_rate": 3.0120481927710845e-07, + "loss": 1.8415, + "step": 16 + }, + { + "epoch": 0.00683967008650171, + "grad_norm": 22.75991867094887, + "learning_rate": 3.2128514056224897e-07, + "loss": 1.9089, + "step": 17 + }, + { + "epoch": 0.00724200362100181, + "grad_norm": 22.23601729043514, + "learning_rate": 3.413654618473896e-07, + "loss": 1.9641, + "step": 18 + }, + { + "epoch": 0.007644337155501911, + "grad_norm": 26.35207563502954, + "learning_rate": 3.614457831325301e-07, + "loss": 2.0458, + "step": 19 + }, + { + "epoch": 0.008046670690002011, + "grad_norm": 20.57786492934256, + "learning_rate": 3.8152610441767073e-07, + "loss": 1.7877, + "step": 20 + }, + { + "epoch": 0.008449004224502113, + "grad_norm": 25.563694935501676, + "learning_rate": 4.0160642570281125e-07, + "loss": 2.1414, + "step": 21 + }, + { + "epoch": 0.008851337759002213, + "grad_norm": 18.98280746175437, + "learning_rate": 4.216867469879518e-07, + "loss": 1.7995, + "step": 22 + }, + { + "epoch": 0.009253671293502314, + "grad_norm": 22.278384224532818, + "learning_rate": 4.417670682730924e-07, + "loss": 2.0412, + "step": 23 + }, + { + "epoch": 0.009656004828002414, + "grad_norm": 18.256806226009903, + "learning_rate": 4.6184738955823296e-07, + "loss": 1.7538, + "step": 24 + }, + { + "epoch": 0.010058338362502514, + "grad_norm": 19.087838303092653, + "learning_rate": 4.819277108433736e-07, + "loss": 1.708, + "step": 25 + }, + { + "epoch": 0.010460671897002616, + "grad_norm": 18.961034656693933, + "learning_rate": 5.020080321285141e-07, + "loss": 1.6661, + "step": 26 + }, + { + "epoch": 0.010863005431502716, + "grad_norm": 18.11814523029599, + "learning_rate": 5.220883534136546e-07, + "loss": 1.753, + "step": 27 + }, + { + "epoch": 0.011265338966002816, + "grad_norm": 19.63691340845534, + "learning_rate": 5.421686746987952e-07, + "loss": 1.7005, + "step": 28 + }, + { + "epoch": 0.011667672500502917, + "grad_norm": 16.293048438980087, + "learning_rate": 5.622489959839358e-07, + "loss": 1.7269, + "step": 29 + }, + { + "epoch": 0.012070006035003017, + "grad_norm": 15.86827584036101, + "learning_rate": 5.823293172690764e-07, + "loss": 1.5401, + "step": 30 + }, + { + "epoch": 0.012472339569503117, + "grad_norm": 17.38302283506762, + "learning_rate": 6.024096385542169e-07, + "loss": 1.5943, + "step": 31 + }, + { + "epoch": 0.012874673104003219, + "grad_norm": 15.374566189246599, + "learning_rate": 6.224899598393574e-07, + "loss": 1.6221, + "step": 32 + }, + { + "epoch": 0.013277006638503319, + "grad_norm": 14.904500940732934, + "learning_rate": 6.425702811244979e-07, + "loss": 1.5452, + "step": 33 + }, + { + "epoch": 0.01367934017300342, + "grad_norm": 16.29526851000047, + "learning_rate": 6.626506024096387e-07, + "loss": 1.2262, + "step": 34 + }, + { + "epoch": 0.01408167370750352, + "grad_norm": 12.819998198911712, + "learning_rate": 6.827309236947792e-07, + "loss": 1.3476, + "step": 35 + }, + { + "epoch": 0.01448400724200362, + "grad_norm": 13.25182925491938, + "learning_rate": 7.028112449799197e-07, + "loss": 1.4371, + "step": 36 + }, + { + "epoch": 0.014886340776503722, + "grad_norm": 14.477871111934059, + "learning_rate": 7.228915662650602e-07, + "loss": 1.3476, + "step": 37 + }, + { + "epoch": 0.015288674311003822, + "grad_norm": 14.6196272151239, + "learning_rate": 7.429718875502008e-07, + "loss": 1.5309, + "step": 38 + }, + { + "epoch": 0.015691007845503924, + "grad_norm": 11.90353169885529, + "learning_rate": 7.630522088353415e-07, + "loss": 1.3026, + "step": 39 + }, + { + "epoch": 0.016093341380004022, + "grad_norm": 13.271008324770724, + "learning_rate": 7.83132530120482e-07, + "loss": 1.4655, + "step": 40 + }, + { + "epoch": 0.016495674914504124, + "grad_norm": 11.2349199407351, + "learning_rate": 8.032128514056225e-07, + "loss": 1.2782, + "step": 41 + }, + { + "epoch": 0.016898008449004225, + "grad_norm": 9.663111453936613, + "learning_rate": 8.232931726907631e-07, + "loss": 1.0628, + "step": 42 + }, + { + "epoch": 0.017300341983504323, + "grad_norm": 10.290302552273262, + "learning_rate": 8.433734939759036e-07, + "loss": 1.2398, + "step": 43 + }, + { + "epoch": 0.017702675518004425, + "grad_norm": 9.437964794637868, + "learning_rate": 8.634538152610443e-07, + "loss": 1.1087, + "step": 44 + }, + { + "epoch": 0.018105009052504527, + "grad_norm": 9.928972956598479, + "learning_rate": 8.835341365461848e-07, + "loss": 1.1231, + "step": 45 + }, + { + "epoch": 0.01850734258700463, + "grad_norm": 7.193942700746645, + "learning_rate": 9.036144578313254e-07, + "loss": 0.9053, + "step": 46 + }, + { + "epoch": 0.018909676121504727, + "grad_norm": 8.545887733782381, + "learning_rate": 9.236947791164659e-07, + "loss": 1.0188, + "step": 47 + }, + { + "epoch": 0.01931200965600483, + "grad_norm": 9.964310833592453, + "learning_rate": 9.437751004016064e-07, + "loss": 1.2336, + "step": 48 + }, + { + "epoch": 0.01971434319050493, + "grad_norm": 8.465102281501581, + "learning_rate": 9.638554216867472e-07, + "loss": 0.978, + "step": 49 + }, + { + "epoch": 0.020116676725005028, + "grad_norm": 8.387761909932218, + "learning_rate": 9.839357429718876e-07, + "loss": 1.0664, + "step": 50 + }, + { + "epoch": 0.02051901025950513, + "grad_norm": 8.853387211734997, + "learning_rate": 1.0040160642570282e-06, + "loss": 1.1316, + "step": 51 + }, + { + "epoch": 0.02092134379400523, + "grad_norm": 8.36811488486395, + "learning_rate": 1.0240963855421688e-06, + "loss": 0.9175, + "step": 52 + }, + { + "epoch": 0.02132367732850533, + "grad_norm": 8.394630361238342, + "learning_rate": 1.0441767068273092e-06, + "loss": 0.9371, + "step": 53 + }, + { + "epoch": 0.02172601086300543, + "grad_norm": 7.411500707300452, + "learning_rate": 1.0642570281124499e-06, + "loss": 0.9567, + "step": 54 + }, + { + "epoch": 0.022128344397505533, + "grad_norm": 8.194667792436356, + "learning_rate": 1.0843373493975905e-06, + "loss": 0.9994, + "step": 55 + }, + { + "epoch": 0.02253067793200563, + "grad_norm": 7.141458407534788, + "learning_rate": 1.1044176706827311e-06, + "loss": 0.8714, + "step": 56 + }, + { + "epoch": 0.022933011466505733, + "grad_norm": 5.765708921674402, + "learning_rate": 1.1244979919678715e-06, + "loss": 0.9238, + "step": 57 + }, + { + "epoch": 0.023335345001005835, + "grad_norm": 7.924253998059311, + "learning_rate": 1.1445783132530121e-06, + "loss": 0.9884, + "step": 58 + }, + { + "epoch": 0.023737678535505933, + "grad_norm": 6.597672306780267, + "learning_rate": 1.1646586345381528e-06, + "loss": 0.9904, + "step": 59 + }, + { + "epoch": 0.024140012070006035, + "grad_norm": 6.937688487718752, + "learning_rate": 1.1847389558232934e-06, + "loss": 0.971, + "step": 60 + }, + { + "epoch": 0.024542345604506136, + "grad_norm": 6.148469030855221, + "learning_rate": 1.2048192771084338e-06, + "loss": 0.7868, + "step": 61 + }, + { + "epoch": 0.024944679139006234, + "grad_norm": 6.388105517333996, + "learning_rate": 1.2248995983935744e-06, + "loss": 0.8567, + "step": 62 + }, + { + "epoch": 0.025347012673506336, + "grad_norm": 6.105324559387419, + "learning_rate": 1.2449799196787148e-06, + "loss": 0.9989, + "step": 63 + }, + { + "epoch": 0.025749346208006438, + "grad_norm": 6.565241933640869, + "learning_rate": 1.2650602409638555e-06, + "loss": 0.8641, + "step": 64 + }, + { + "epoch": 0.02615167974250654, + "grad_norm": 5.869907162685047, + "learning_rate": 1.2851405622489959e-06, + "loss": 0.9526, + "step": 65 + }, + { + "epoch": 0.026554013277006638, + "grad_norm": 5.851371649159624, + "learning_rate": 1.3052208835341367e-06, + "loss": 0.8387, + "step": 66 + }, + { + "epoch": 0.02695634681150674, + "grad_norm": 6.3493989766408765, + "learning_rate": 1.3253012048192773e-06, + "loss": 1.0458, + "step": 67 + }, + { + "epoch": 0.02735868034600684, + "grad_norm": 5.9283912382403114, + "learning_rate": 1.345381526104418e-06, + "loss": 1.0158, + "step": 68 + }, + { + "epoch": 0.02776101388050694, + "grad_norm": 6.184126721660312, + "learning_rate": 1.3654618473895584e-06, + "loss": 0.8218, + "step": 69 + }, + { + "epoch": 0.02816334741500704, + "grad_norm": 6.0169493477739975, + "learning_rate": 1.385542168674699e-06, + "loss": 0.7966, + "step": 70 + }, + { + "epoch": 0.028565680949507143, + "grad_norm": 6.084096016864222, + "learning_rate": 1.4056224899598394e-06, + "loss": 0.8221, + "step": 71 + }, + { + "epoch": 0.02896801448400724, + "grad_norm": 6.441697966234361, + "learning_rate": 1.42570281124498e-06, + "loss": 0.7189, + "step": 72 + }, + { + "epoch": 0.029370348018507342, + "grad_norm": 5.684758471405106, + "learning_rate": 1.4457831325301204e-06, + "loss": 0.811, + "step": 73 + }, + { + "epoch": 0.029772681553007444, + "grad_norm": 6.066554340342954, + "learning_rate": 1.465863453815261e-06, + "loss": 0.8128, + "step": 74 + }, + { + "epoch": 0.030175015087507542, + "grad_norm": 5.398586406654046, + "learning_rate": 1.4859437751004017e-06, + "loss": 0.8521, + "step": 75 + }, + { + "epoch": 0.030577348622007644, + "grad_norm": 5.72164636854203, + "learning_rate": 1.5060240963855425e-06, + "loss": 0.8733, + "step": 76 + }, + { + "epoch": 0.030979682156507746, + "grad_norm": 5.317662321058184, + "learning_rate": 1.526104417670683e-06, + "loss": 0.6758, + "step": 77 + }, + { + "epoch": 0.03138201569100785, + "grad_norm": 5.008046012529329, + "learning_rate": 1.5461847389558236e-06, + "loss": 0.77, + "step": 78 + }, + { + "epoch": 0.03178434922550795, + "grad_norm": 5.806938383916589, + "learning_rate": 1.566265060240964e-06, + "loss": 0.9218, + "step": 79 + }, + { + "epoch": 0.032186682760008044, + "grad_norm": 5.687427641764964, + "learning_rate": 1.5863453815261046e-06, + "loss": 0.873, + "step": 80 + }, + { + "epoch": 0.032589016294508145, + "grad_norm": 5.877177229539529, + "learning_rate": 1.606425702811245e-06, + "loss": 0.8347, + "step": 81 + }, + { + "epoch": 0.03299134982900825, + "grad_norm": 5.661277204998088, + "learning_rate": 1.6265060240963856e-06, + "loss": 0.7537, + "step": 82 + }, + { + "epoch": 0.03339368336350835, + "grad_norm": 6.0004706518847675, + "learning_rate": 1.6465863453815263e-06, + "loss": 0.9341, + "step": 83 + }, + { + "epoch": 0.03379601689800845, + "grad_norm": 5.383372715468371, + "learning_rate": 1.6666666666666667e-06, + "loss": 0.716, + "step": 84 + }, + { + "epoch": 0.03419835043250855, + "grad_norm": 5.870492937673813, + "learning_rate": 1.6867469879518073e-06, + "loss": 0.8515, + "step": 85 + }, + { + "epoch": 0.03460068396700865, + "grad_norm": 5.40282247472889, + "learning_rate": 1.7068273092369481e-06, + "loss": 0.7575, + "step": 86 + }, + { + "epoch": 0.03500301750150875, + "grad_norm": 5.244298542730251, + "learning_rate": 1.7269076305220885e-06, + "loss": 0.7342, + "step": 87 + }, + { + "epoch": 0.03540535103600885, + "grad_norm": 5.118324863903489, + "learning_rate": 1.7469879518072292e-06, + "loss": 0.7633, + "step": 88 + }, + { + "epoch": 0.03580768457050895, + "grad_norm": 5.559590746485852, + "learning_rate": 1.7670682730923696e-06, + "loss": 0.915, + "step": 89 + }, + { + "epoch": 0.036210018105009054, + "grad_norm": 5.093580932305843, + "learning_rate": 1.7871485943775102e-06, + "loss": 0.7234, + "step": 90 + }, + { + "epoch": 0.036612351639509155, + "grad_norm": 5.949076037097858, + "learning_rate": 1.8072289156626508e-06, + "loss": 0.9392, + "step": 91 + }, + { + "epoch": 0.03701468517400926, + "grad_norm": 6.126962098905458, + "learning_rate": 1.8273092369477912e-06, + "loss": 0.7973, + "step": 92 + }, + { + "epoch": 0.03741701870850935, + "grad_norm": 5.418925049178253, + "learning_rate": 1.8473895582329318e-06, + "loss": 0.7386, + "step": 93 + }, + { + "epoch": 0.03781935224300945, + "grad_norm": 5.30622581323819, + "learning_rate": 1.8674698795180723e-06, + "loss": 0.7712, + "step": 94 + }, + { + "epoch": 0.038221685777509555, + "grad_norm": 5.203905538553713, + "learning_rate": 1.8875502008032129e-06, + "loss": 0.7403, + "step": 95 + }, + { + "epoch": 0.03862401931200966, + "grad_norm": 5.497752744670633, + "learning_rate": 1.9076305220883537e-06, + "loss": 0.8343, + "step": 96 + }, + { + "epoch": 0.03902635284650976, + "grad_norm": 5.792666990034754, + "learning_rate": 1.9277108433734943e-06, + "loss": 0.9855, + "step": 97 + }, + { + "epoch": 0.03942868638100986, + "grad_norm": 4.95103734609948, + "learning_rate": 1.947791164658635e-06, + "loss": 0.7676, + "step": 98 + }, + { + "epoch": 0.039831019915509955, + "grad_norm": 5.313239350915902, + "learning_rate": 1.967871485943775e-06, + "loss": 0.8112, + "step": 99 + }, + { + "epoch": 0.040233353450010056, + "grad_norm": 5.262581677945773, + "learning_rate": 1.987951807228916e-06, + "loss": 0.7726, + "step": 100 + }, + { + "epoch": 0.04063568698451016, + "grad_norm": 5.7284063119812965, + "learning_rate": 2.0080321285140564e-06, + "loss": 0.7143, + "step": 101 + }, + { + "epoch": 0.04103802051901026, + "grad_norm": 5.61074467493987, + "learning_rate": 2.028112449799197e-06, + "loss": 0.7821, + "step": 102 + }, + { + "epoch": 0.04144035405351036, + "grad_norm": 5.668818940572837, + "learning_rate": 2.0481927710843377e-06, + "loss": 0.6217, + "step": 103 + }, + { + "epoch": 0.04184268758801046, + "grad_norm": 6.102827563222031, + "learning_rate": 2.068273092369478e-06, + "loss": 0.7094, + "step": 104 + }, + { + "epoch": 0.04224502112251056, + "grad_norm": 5.106901986680839, + "learning_rate": 2.0883534136546185e-06, + "loss": 0.7452, + "step": 105 + }, + { + "epoch": 0.04264735465701066, + "grad_norm": 5.320689458556449, + "learning_rate": 2.1084337349397595e-06, + "loss": 0.8121, + "step": 106 + }, + { + "epoch": 0.04304968819151076, + "grad_norm": 4.851575241509713, + "learning_rate": 2.1285140562248997e-06, + "loss": 0.7695, + "step": 107 + }, + { + "epoch": 0.04345202172601086, + "grad_norm": 5.7294050539399315, + "learning_rate": 2.1485943775100404e-06, + "loss": 0.8343, + "step": 108 + }, + { + "epoch": 0.043854355260510965, + "grad_norm": 6.2102432153860265, + "learning_rate": 2.168674698795181e-06, + "loss": 0.8478, + "step": 109 + }, + { + "epoch": 0.044256688795011066, + "grad_norm": 5.8055227032618415, + "learning_rate": 2.1887550200803216e-06, + "loss": 0.7647, + "step": 110 + }, + { + "epoch": 0.04465902232951117, + "grad_norm": 5.59165747057277, + "learning_rate": 2.2088353413654622e-06, + "loss": 0.6905, + "step": 111 + }, + { + "epoch": 0.04506135586401126, + "grad_norm": 4.682897930842293, + "learning_rate": 2.2289156626506024e-06, + "loss": 0.6875, + "step": 112 + }, + { + "epoch": 0.045463689398511364, + "grad_norm": 4.320466219380915, + "learning_rate": 2.248995983935743e-06, + "loss": 0.696, + "step": 113 + }, + { + "epoch": 0.045866022933011466, + "grad_norm": 5.050663115560423, + "learning_rate": 2.2690763052208837e-06, + "loss": 0.6224, + "step": 114 + }, + { + "epoch": 0.04626835646751157, + "grad_norm": 4.681780822596801, + "learning_rate": 2.2891566265060243e-06, + "loss": 0.6462, + "step": 115 + }, + { + "epoch": 0.04667069000201167, + "grad_norm": 5.954135299109462, + "learning_rate": 2.309236947791165e-06, + "loss": 0.8848, + "step": 116 + }, + { + "epoch": 0.04707302353651177, + "grad_norm": 5.573998592787985, + "learning_rate": 2.3293172690763055e-06, + "loss": 0.8088, + "step": 117 + }, + { + "epoch": 0.047475357071011866, + "grad_norm": 5.070007694615057, + "learning_rate": 2.349397590361446e-06, + "loss": 0.7173, + "step": 118 + }, + { + "epoch": 0.04787769060551197, + "grad_norm": 5.671554086319649, + "learning_rate": 2.3694779116465868e-06, + "loss": 0.8136, + "step": 119 + }, + { + "epoch": 0.04828002414001207, + "grad_norm": 5.769697685018851, + "learning_rate": 2.389558232931727e-06, + "loss": 0.7733, + "step": 120 + }, + { + "epoch": 0.04868235767451217, + "grad_norm": 4.920298621449247, + "learning_rate": 2.4096385542168676e-06, + "loss": 0.5931, + "step": 121 + }, + { + "epoch": 0.04908469120901227, + "grad_norm": 5.49690667652572, + "learning_rate": 2.4297188755020082e-06, + "loss": 0.8079, + "step": 122 + }, + { + "epoch": 0.049487024743512374, + "grad_norm": 5.369187249008034, + "learning_rate": 2.449799196787149e-06, + "loss": 0.726, + "step": 123 + }, + { + "epoch": 0.04988935827801247, + "grad_norm": 4.8515152911462796, + "learning_rate": 2.469879518072289e-06, + "loss": 0.7691, + "step": 124 + }, + { + "epoch": 0.05029169181251257, + "grad_norm": 5.099204794104859, + "learning_rate": 2.4899598393574297e-06, + "loss": 0.8361, + "step": 125 + }, + { + "epoch": 0.05069402534701267, + "grad_norm": 5.390609640239935, + "learning_rate": 2.5100401606425707e-06, + "loss": 0.6527, + "step": 126 + }, + { + "epoch": 0.051096358881512774, + "grad_norm": 5.053936099287885, + "learning_rate": 2.530120481927711e-06, + "loss": 0.7326, + "step": 127 + }, + { + "epoch": 0.051498692416012876, + "grad_norm": 5.876719841742455, + "learning_rate": 2.5502008032128516e-06, + "loss": 0.8571, + "step": 128 + }, + { + "epoch": 0.05190102595051298, + "grad_norm": 5.291120710287588, + "learning_rate": 2.5702811244979918e-06, + "loss": 0.7641, + "step": 129 + }, + { + "epoch": 0.05230335948501308, + "grad_norm": 5.2648612797895336, + "learning_rate": 2.590361445783133e-06, + "loss": 0.768, + "step": 130 + }, + { + "epoch": 0.052705693019513174, + "grad_norm": 5.146667666976231, + "learning_rate": 2.6104417670682734e-06, + "loss": 0.729, + "step": 131 + }, + { + "epoch": 0.053108026554013275, + "grad_norm": 5.607251668752912, + "learning_rate": 2.6305220883534136e-06, + "loss": 0.7812, + "step": 132 + }, + { + "epoch": 0.05351036008851338, + "grad_norm": 4.774953294338218, + "learning_rate": 2.6506024096385547e-06, + "loss": 0.6996, + "step": 133 + }, + { + "epoch": 0.05391269362301348, + "grad_norm": 5.187652607031923, + "learning_rate": 2.670682730923695e-06, + "loss": 0.7029, + "step": 134 + }, + { + "epoch": 0.05431502715751358, + "grad_norm": 5.631076339839678, + "learning_rate": 2.690763052208836e-06, + "loss": 0.7618, + "step": 135 + }, + { + "epoch": 0.05471736069201368, + "grad_norm": 5.1080370644891735, + "learning_rate": 2.710843373493976e-06, + "loss": 0.6746, + "step": 136 + }, + { + "epoch": 0.05511969422651378, + "grad_norm": 5.177774314908037, + "learning_rate": 2.7309236947791167e-06, + "loss": 0.6576, + "step": 137 + }, + { + "epoch": 0.05552202776101388, + "grad_norm": 5.4428731816372355, + "learning_rate": 2.751004016064257e-06, + "loss": 0.8666, + "step": 138 + }, + { + "epoch": 0.05592436129551398, + "grad_norm": 4.737577892176393, + "learning_rate": 2.771084337349398e-06, + "loss": 0.6303, + "step": 139 + }, + { + "epoch": 0.05632669483001408, + "grad_norm": 5.958074686629733, + "learning_rate": 2.791164658634538e-06, + "loss": 0.7375, + "step": 140 + }, + { + "epoch": 0.056729028364514184, + "grad_norm": 5.347638262573924, + "learning_rate": 2.811244979919679e-06, + "loss": 0.7496, + "step": 141 + }, + { + "epoch": 0.057131361899014285, + "grad_norm": 5.25633214854894, + "learning_rate": 2.83132530120482e-06, + "loss": 0.7703, + "step": 142 + }, + { + "epoch": 0.05753369543351439, + "grad_norm": 6.3493191578507355, + "learning_rate": 2.85140562248996e-06, + "loss": 0.7297, + "step": 143 + }, + { + "epoch": 0.05793602896801448, + "grad_norm": 5.291536986822124, + "learning_rate": 2.8714859437751007e-06, + "loss": 0.7848, + "step": 144 + }, + { + "epoch": 0.05833836250251458, + "grad_norm": 4.327776932092466, + "learning_rate": 2.891566265060241e-06, + "loss": 0.5537, + "step": 145 + }, + { + "epoch": 0.058740696037014685, + "grad_norm": 4.827406966450388, + "learning_rate": 2.911646586345382e-06, + "loss": 0.6614, + "step": 146 + }, + { + "epoch": 0.05914302957151479, + "grad_norm": 5.4512595989034605, + "learning_rate": 2.931726907630522e-06, + "loss": 0.5105, + "step": 147 + }, + { + "epoch": 0.05954536310601489, + "grad_norm": 5.349806231832712, + "learning_rate": 2.9518072289156627e-06, + "loss": 0.7398, + "step": 148 + }, + { + "epoch": 0.05994769664051499, + "grad_norm": 5.376464953972205, + "learning_rate": 2.9718875502008034e-06, + "loss": 0.832, + "step": 149 + }, + { + "epoch": 0.060350030175015085, + "grad_norm": 4.4973784287811, + "learning_rate": 2.991967871485944e-06, + "loss": 0.6631, + "step": 150 + }, + { + "epoch": 0.060752363709515186, + "grad_norm": 4.1574904576313925, + "learning_rate": 3.012048192771085e-06, + "loss": 0.5736, + "step": 151 + }, + { + "epoch": 0.06115469724401529, + "grad_norm": 4.932454920014607, + "learning_rate": 3.0321285140562252e-06, + "loss": 0.6843, + "step": 152 + }, + { + "epoch": 0.06155703077851539, + "grad_norm": 4.920072821237323, + "learning_rate": 3.052208835341366e-06, + "loss": 0.6115, + "step": 153 + }, + { + "epoch": 0.06195936431301549, + "grad_norm": 5.069151288790082, + "learning_rate": 3.072289156626506e-06, + "loss": 0.6592, + "step": 154 + }, + { + "epoch": 0.06236169784751559, + "grad_norm": 4.940734045454529, + "learning_rate": 3.092369477911647e-06, + "loss": 0.6823, + "step": 155 + }, + { + "epoch": 0.0627640313820157, + "grad_norm": 4.7329970403721635, + "learning_rate": 3.1124497991967873e-06, + "loss": 0.6183, + "step": 156 + }, + { + "epoch": 0.0631663649165158, + "grad_norm": 5.584324231160161, + "learning_rate": 3.132530120481928e-06, + "loss": 0.8407, + "step": 157 + }, + { + "epoch": 0.0635686984510159, + "grad_norm": 4.9494988025514095, + "learning_rate": 3.152610441767068e-06, + "loss": 0.7362, + "step": 158 + }, + { + "epoch": 0.06397103198551599, + "grad_norm": 4.826336881024485, + "learning_rate": 3.172690763052209e-06, + "loss": 0.6756, + "step": 159 + }, + { + "epoch": 0.06437336552001609, + "grad_norm": 4.663584106854257, + "learning_rate": 3.1927710843373494e-06, + "loss": 0.6618, + "step": 160 + }, + { + "epoch": 0.06477569905451619, + "grad_norm": 5.358113823802763, + "learning_rate": 3.21285140562249e-06, + "loss": 0.7845, + "step": 161 + }, + { + "epoch": 0.06517803258901629, + "grad_norm": 4.393966562734061, + "learning_rate": 3.232931726907631e-06, + "loss": 0.6643, + "step": 162 + }, + { + "epoch": 0.06558036612351639, + "grad_norm": 5.269005574739929, + "learning_rate": 3.2530120481927713e-06, + "loss": 0.7892, + "step": 163 + }, + { + "epoch": 0.0659826996580165, + "grad_norm": 5.191377211307637, + "learning_rate": 3.2730923694779123e-06, + "loss": 0.6834, + "step": 164 + }, + { + "epoch": 0.0663850331925166, + "grad_norm": 5.9411717854682875, + "learning_rate": 3.2931726907630525e-06, + "loss": 0.682, + "step": 165 + }, + { + "epoch": 0.0667873667270167, + "grad_norm": 4.657895689681778, + "learning_rate": 3.313253012048193e-06, + "loss": 0.6533, + "step": 166 + }, + { + "epoch": 0.0671897002615168, + "grad_norm": 4.689805393308291, + "learning_rate": 3.3333333333333333e-06, + "loss": 0.5851, + "step": 167 + }, + { + "epoch": 0.0675920337960169, + "grad_norm": 4.934046837424603, + "learning_rate": 3.3534136546184744e-06, + "loss": 0.7834, + "step": 168 + }, + { + "epoch": 0.067994367330517, + "grad_norm": 5.295304511873193, + "learning_rate": 3.3734939759036146e-06, + "loss": 0.6978, + "step": 169 + }, + { + "epoch": 0.0683967008650171, + "grad_norm": 5.8805644086916935, + "learning_rate": 3.393574297188755e-06, + "loss": 0.7022, + "step": 170 + }, + { + "epoch": 0.0687990343995172, + "grad_norm": 4.986410631963467, + "learning_rate": 3.4136546184738962e-06, + "loss": 0.8589, + "step": 171 + }, + { + "epoch": 0.0692013679340173, + "grad_norm": 4.587536055949948, + "learning_rate": 3.4337349397590364e-06, + "loss": 0.5962, + "step": 172 + }, + { + "epoch": 0.0696037014685174, + "grad_norm": 4.542521557186459, + "learning_rate": 3.453815261044177e-06, + "loss": 0.5778, + "step": 173 + }, + { + "epoch": 0.0700060350030175, + "grad_norm": 5.006237220238328, + "learning_rate": 3.4738955823293173e-06, + "loss": 0.7066, + "step": 174 + }, + { + "epoch": 0.0704083685375176, + "grad_norm": 5.03127022104005, + "learning_rate": 3.4939759036144583e-06, + "loss": 0.7143, + "step": 175 + }, + { + "epoch": 0.0708107020720177, + "grad_norm": 4.978976609084576, + "learning_rate": 3.5140562248995985e-06, + "loss": 0.5806, + "step": 176 + }, + { + "epoch": 0.0712130356065178, + "grad_norm": 4.953415145134539, + "learning_rate": 3.534136546184739e-06, + "loss": 0.6169, + "step": 177 + }, + { + "epoch": 0.0716153691410179, + "grad_norm": 5.224274445585009, + "learning_rate": 3.5542168674698798e-06, + "loss": 0.7055, + "step": 178 + }, + { + "epoch": 0.072017702675518, + "grad_norm": 4.793231833957601, + "learning_rate": 3.5742971887550204e-06, + "loss": 0.6589, + "step": 179 + }, + { + "epoch": 0.07242003621001811, + "grad_norm": 4.897332398174339, + "learning_rate": 3.5943775100401606e-06, + "loss": 0.5835, + "step": 180 + }, + { + "epoch": 0.07282236974451821, + "grad_norm": 5.461164108231627, + "learning_rate": 3.6144578313253016e-06, + "loss": 0.6153, + "step": 181 + }, + { + "epoch": 0.07322470327901831, + "grad_norm": 4.750713371821981, + "learning_rate": 3.6345381526104423e-06, + "loss": 0.5576, + "step": 182 + }, + { + "epoch": 0.07362703681351841, + "grad_norm": 4.612815212986164, + "learning_rate": 3.6546184738955825e-06, + "loss": 0.6695, + "step": 183 + }, + { + "epoch": 0.07402937034801851, + "grad_norm": 5.035137988160493, + "learning_rate": 3.6746987951807235e-06, + "loss": 0.6451, + "step": 184 + }, + { + "epoch": 0.0744317038825186, + "grad_norm": 5.189253752749726, + "learning_rate": 3.6947791164658637e-06, + "loss": 0.6897, + "step": 185 + }, + { + "epoch": 0.0748340374170187, + "grad_norm": 5.061038391462879, + "learning_rate": 3.7148594377510043e-06, + "loss": 0.7363, + "step": 186 + }, + { + "epoch": 0.0752363709515188, + "grad_norm": 5.12457801565824, + "learning_rate": 3.7349397590361445e-06, + "loss": 0.6662, + "step": 187 + }, + { + "epoch": 0.0756387044860189, + "grad_norm": 5.468566358022457, + "learning_rate": 3.7550200803212856e-06, + "loss": 0.7333, + "step": 188 + }, + { + "epoch": 0.07604103802051901, + "grad_norm": 4.83323388980342, + "learning_rate": 3.7751004016064258e-06, + "loss": 0.6179, + "step": 189 + }, + { + "epoch": 0.07644337155501911, + "grad_norm": 5.618759075616741, + "learning_rate": 3.7951807228915664e-06, + "loss": 0.7861, + "step": 190 + }, + { + "epoch": 0.07684570508951921, + "grad_norm": 5.571721878536531, + "learning_rate": 3.8152610441767074e-06, + "loss": 0.6481, + "step": 191 + }, + { + "epoch": 0.07724803862401931, + "grad_norm": 5.067103682358002, + "learning_rate": 3.835341365461848e-06, + "loss": 0.7295, + "step": 192 + }, + { + "epoch": 0.07765037215851942, + "grad_norm": 4.945134722999282, + "learning_rate": 3.855421686746989e-06, + "loss": 0.6275, + "step": 193 + }, + { + "epoch": 0.07805270569301952, + "grad_norm": 4.995120702157204, + "learning_rate": 3.875502008032129e-06, + "loss": 0.6156, + "step": 194 + }, + { + "epoch": 0.07845503922751962, + "grad_norm": 4.936376571967402, + "learning_rate": 3.89558232931727e-06, + "loss": 0.6141, + "step": 195 + }, + { + "epoch": 0.07885737276201972, + "grad_norm": 5.014965363677265, + "learning_rate": 3.91566265060241e-06, + "loss": 0.6706, + "step": 196 + }, + { + "epoch": 0.07925970629651982, + "grad_norm": 4.379077775855287, + "learning_rate": 3.93574297188755e-06, + "loss": 0.625, + "step": 197 + }, + { + "epoch": 0.07966203983101991, + "grad_norm": 5.213065143955326, + "learning_rate": 3.9558232931726905e-06, + "loss": 0.6259, + "step": 198 + }, + { + "epoch": 0.08006437336552001, + "grad_norm": 4.943005393732397, + "learning_rate": 3.975903614457832e-06, + "loss": 0.6867, + "step": 199 + }, + { + "epoch": 0.08046670690002011, + "grad_norm": 6.1486973779873795, + "learning_rate": 3.995983935742972e-06, + "loss": 0.8347, + "step": 200 + }, + { + "epoch": 0.08086904043452021, + "grad_norm": 4.733552314219692, + "learning_rate": 4.016064257028113e-06, + "loss": 0.6611, + "step": 201 + }, + { + "epoch": 0.08127137396902032, + "grad_norm": 4.8269187112270195, + "learning_rate": 4.036144578313254e-06, + "loss": 0.6939, + "step": 202 + }, + { + "epoch": 0.08167370750352042, + "grad_norm": 5.3439700963806125, + "learning_rate": 4.056224899598394e-06, + "loss": 0.7133, + "step": 203 + }, + { + "epoch": 0.08207604103802052, + "grad_norm": 5.29998702305394, + "learning_rate": 4.076305220883534e-06, + "loss": 0.7792, + "step": 204 + }, + { + "epoch": 0.08247837457252062, + "grad_norm": 5.408216152846704, + "learning_rate": 4.096385542168675e-06, + "loss": 0.6411, + "step": 205 + }, + { + "epoch": 0.08288070810702072, + "grad_norm": 4.419630999712613, + "learning_rate": 4.1164658634538155e-06, + "loss": 0.615, + "step": 206 + }, + { + "epoch": 0.08328304164152082, + "grad_norm": 4.543564292966689, + "learning_rate": 4.136546184738956e-06, + "loss": 0.5462, + "step": 207 + }, + { + "epoch": 0.08368537517602093, + "grad_norm": 4.814136944508732, + "learning_rate": 4.156626506024097e-06, + "loss": 0.6407, + "step": 208 + }, + { + "epoch": 0.08408770871052103, + "grad_norm": 5.118958059055448, + "learning_rate": 4.176706827309237e-06, + "loss": 0.6922, + "step": 209 + }, + { + "epoch": 0.08449004224502112, + "grad_norm": 5.761968321529592, + "learning_rate": 4.196787148594378e-06, + "loss": 0.7785, + "step": 210 + }, + { + "epoch": 0.08489237577952122, + "grad_norm": 4.7973887668760185, + "learning_rate": 4.216867469879519e-06, + "loss": 0.5855, + "step": 211 + }, + { + "epoch": 0.08529470931402132, + "grad_norm": 5.063119851943765, + "learning_rate": 4.236947791164659e-06, + "loss": 0.5061, + "step": 212 + }, + { + "epoch": 0.08569704284852142, + "grad_norm": 5.185457141038217, + "learning_rate": 4.2570281124497995e-06, + "loss": 0.6547, + "step": 213 + }, + { + "epoch": 0.08609937638302152, + "grad_norm": 4.908063923355144, + "learning_rate": 4.27710843373494e-06, + "loss": 0.6325, + "step": 214 + }, + { + "epoch": 0.08650170991752162, + "grad_norm": 4.902236509892525, + "learning_rate": 4.297188755020081e-06, + "loss": 0.771, + "step": 215 + }, + { + "epoch": 0.08690404345202173, + "grad_norm": 4.939421470208826, + "learning_rate": 4.317269076305221e-06, + "loss": 0.6009, + "step": 216 + }, + { + "epoch": 0.08730637698652183, + "grad_norm": 5.585627020639388, + "learning_rate": 4.337349397590362e-06, + "loss": 0.75, + "step": 217 + }, + { + "epoch": 0.08770871052102193, + "grad_norm": 5.768733340993033, + "learning_rate": 4.357429718875502e-06, + "loss": 0.8889, + "step": 218 + }, + { + "epoch": 0.08811104405552203, + "grad_norm": 4.756073859984964, + "learning_rate": 4.377510040160643e-06, + "loss": 0.5845, + "step": 219 + }, + { + "epoch": 0.08851337759002213, + "grad_norm": 5.602957070495176, + "learning_rate": 4.397590361445783e-06, + "loss": 0.5766, + "step": 220 + }, + { + "epoch": 0.08891571112452223, + "grad_norm": 5.51373108651177, + "learning_rate": 4.4176706827309244e-06, + "loss": 0.667, + "step": 221 + }, + { + "epoch": 0.08931804465902234, + "grad_norm": 5.2818421493270895, + "learning_rate": 4.437751004016065e-06, + "loss": 0.6132, + "step": 222 + }, + { + "epoch": 0.08972037819352242, + "grad_norm": 4.687481374933495, + "learning_rate": 4.457831325301205e-06, + "loss": 0.69, + "step": 223 + }, + { + "epoch": 0.09012271172802253, + "grad_norm": 5.045402036516882, + "learning_rate": 4.477911646586346e-06, + "loss": 0.5496, + "step": 224 + }, + { + "epoch": 0.09052504526252263, + "grad_norm": 4.949655367217769, + "learning_rate": 4.497991967871486e-06, + "loss": 0.6777, + "step": 225 + }, + { + "epoch": 0.09092737879702273, + "grad_norm": 4.725446380096972, + "learning_rate": 4.518072289156627e-06, + "loss": 0.5211, + "step": 226 + }, + { + "epoch": 0.09132971233152283, + "grad_norm": 5.627747083860127, + "learning_rate": 4.538152610441767e-06, + "loss": 0.7335, + "step": 227 + }, + { + "epoch": 0.09173204586602293, + "grad_norm": 4.658915831199445, + "learning_rate": 4.558232931726908e-06, + "loss": 0.618, + "step": 228 + }, + { + "epoch": 0.09213437940052303, + "grad_norm": 5.169380307261355, + "learning_rate": 4.578313253012049e-06, + "loss": 0.8083, + "step": 229 + }, + { + "epoch": 0.09253671293502314, + "grad_norm": 4.462280637097748, + "learning_rate": 4.598393574297189e-06, + "loss": 0.5873, + "step": 230 + }, + { + "epoch": 0.09293904646952324, + "grad_norm": 5.088180514844276, + "learning_rate": 4.61847389558233e-06, + "loss": 0.6948, + "step": 231 + }, + { + "epoch": 0.09334138000402334, + "grad_norm": 4.911645340461795, + "learning_rate": 4.63855421686747e-06, + "loss": 0.5014, + "step": 232 + }, + { + "epoch": 0.09374371353852344, + "grad_norm": 5.183650867648958, + "learning_rate": 4.658634538152611e-06, + "loss": 0.6731, + "step": 233 + }, + { + "epoch": 0.09414604707302354, + "grad_norm": 4.942647346081943, + "learning_rate": 4.678714859437751e-06, + "loss": 0.796, + "step": 234 + }, + { + "epoch": 0.09454838060752364, + "grad_norm": 5.413754367435916, + "learning_rate": 4.698795180722892e-06, + "loss": 0.7029, + "step": 235 + }, + { + "epoch": 0.09495071414202373, + "grad_norm": 5.394996509930947, + "learning_rate": 4.7188755020080325e-06, + "loss": 0.7448, + "step": 236 + }, + { + "epoch": 0.09535304767652383, + "grad_norm": 4.807099705141704, + "learning_rate": 4.7389558232931736e-06, + "loss": 0.6608, + "step": 237 + }, + { + "epoch": 0.09575538121102393, + "grad_norm": 5.204336077501797, + "learning_rate": 4.759036144578314e-06, + "loss": 0.585, + "step": 238 + }, + { + "epoch": 0.09615771474552404, + "grad_norm": 4.845053158029385, + "learning_rate": 4.779116465863454e-06, + "loss": 0.6299, + "step": 239 + }, + { + "epoch": 0.09656004828002414, + "grad_norm": 5.037832345926416, + "learning_rate": 4.799196787148594e-06, + "loss": 0.6598, + "step": 240 + }, + { + "epoch": 0.09696238181452424, + "grad_norm": 4.930571962391263, + "learning_rate": 4.819277108433735e-06, + "loss": 0.5918, + "step": 241 + }, + { + "epoch": 0.09736471534902434, + "grad_norm": 4.749158619882368, + "learning_rate": 4.839357429718876e-06, + "loss": 0.5785, + "step": 242 + }, + { + "epoch": 0.09776704888352444, + "grad_norm": 5.774299138017609, + "learning_rate": 4.8594377510040165e-06, + "loss": 0.7465, + "step": 243 + }, + { + "epoch": 0.09816938241802455, + "grad_norm": 4.582929546310413, + "learning_rate": 4.8795180722891575e-06, + "loss": 0.5464, + "step": 244 + }, + { + "epoch": 0.09857171595252465, + "grad_norm": 4.604283671045787, + "learning_rate": 4.899598393574298e-06, + "loss": 0.6183, + "step": 245 + }, + { + "epoch": 0.09897404948702475, + "grad_norm": 5.126092092509994, + "learning_rate": 4.919678714859438e-06, + "loss": 0.7267, + "step": 246 + }, + { + "epoch": 0.09937638302152485, + "grad_norm": 4.6087281951772345, + "learning_rate": 4.939759036144578e-06, + "loss": 0.5911, + "step": 247 + }, + { + "epoch": 0.09977871655602494, + "grad_norm": 4.4308954136706165, + "learning_rate": 4.959839357429719e-06, + "loss": 0.6525, + "step": 248 + }, + { + "epoch": 0.10018105009052504, + "grad_norm": 4.704767612087696, + "learning_rate": 4.979919678714859e-06, + "loss": 0.7537, + "step": 249 + }, + { + "epoch": 0.10058338362502514, + "grad_norm": 5.171569719861641, + "learning_rate": 5e-06, + "loss": 0.5941, + "step": 250 + }, + { + "epoch": 0.10098571715952524, + "grad_norm": 5.7449160884679245, + "learning_rate": 4.99999753465491e-06, + "loss": 0.6686, + "step": 251 + }, + { + "epoch": 0.10138805069402534, + "grad_norm": 4.648308764039487, + "learning_rate": 4.999990138624503e-06, + "loss": 0.6264, + "step": 252 + }, + { + "epoch": 0.10179038422852545, + "grad_norm": 5.064339952308565, + "learning_rate": 4.999977811923365e-06, + "loss": 0.6921, + "step": 253 + }, + { + "epoch": 0.10219271776302555, + "grad_norm": 4.810326779178174, + "learning_rate": 4.9999605545758065e-06, + "loss": 0.6119, + "step": 254 + }, + { + "epoch": 0.10259505129752565, + "grad_norm": 4.805314994215823, + "learning_rate": 4.999938366615867e-06, + "loss": 0.6364, + "step": 255 + }, + { + "epoch": 0.10299738483202575, + "grad_norm": 4.473992853756316, + "learning_rate": 4.999911248087305e-06, + "loss": 0.5508, + "step": 256 + }, + { + "epoch": 0.10339971836652585, + "grad_norm": 5.452559581429947, + "learning_rate": 4.999879199043606e-06, + "loss": 0.7831, + "step": 257 + }, + { + "epoch": 0.10380205190102595, + "grad_norm": 5.061524793864594, + "learning_rate": 4.99984221954798e-06, + "loss": 0.5644, + "step": 258 + }, + { + "epoch": 0.10420438543552606, + "grad_norm": 4.673975167806026, + "learning_rate": 4.999800309673361e-06, + "loss": 0.649, + "step": 259 + }, + { + "epoch": 0.10460671897002616, + "grad_norm": 4.8181768852216855, + "learning_rate": 4.999753469502406e-06, + "loss": 0.6528, + "step": 260 + }, + { + "epoch": 0.10500905250452625, + "grad_norm": 5.3922298074237345, + "learning_rate": 4.999701699127497e-06, + "loss": 0.8216, + "step": 261 + }, + { + "epoch": 0.10541138603902635, + "grad_norm": 4.463138014998754, + "learning_rate": 4.99964499865074e-06, + "loss": 0.6824, + "step": 262 + }, + { + "epoch": 0.10581371957352645, + "grad_norm": 4.937246753810877, + "learning_rate": 4.999583368183965e-06, + "loss": 0.5256, + "step": 263 + }, + { + "epoch": 0.10621605310802655, + "grad_norm": 5.02716535200141, + "learning_rate": 4.999516807848721e-06, + "loss": 0.6658, + "step": 264 + }, + { + "epoch": 0.10661838664252665, + "grad_norm": 4.719903057858775, + "learning_rate": 4.999445317776286e-06, + "loss": 0.5943, + "step": 265 + }, + { + "epoch": 0.10702072017702675, + "grad_norm": 4.327891289019763, + "learning_rate": 4.9993688981076574e-06, + "loss": 0.6267, + "step": 266 + }, + { + "epoch": 0.10742305371152686, + "grad_norm": 5.04792717283726, + "learning_rate": 4.999287548993555e-06, + "loss": 0.7649, + "step": 267 + }, + { + "epoch": 0.10782538724602696, + "grad_norm": 4.398384896034094, + "learning_rate": 4.999201270594424e-06, + "loss": 0.6237, + "step": 268 + }, + { + "epoch": 0.10822772078052706, + "grad_norm": 4.513204332172601, + "learning_rate": 4.999110063080428e-06, + "loss": 0.685, + "step": 269 + }, + { + "epoch": 0.10863005431502716, + "grad_norm": 4.910504259364484, + "learning_rate": 4.999013926631453e-06, + "loss": 0.6393, + "step": 270 + }, + { + "epoch": 0.10903238784952726, + "grad_norm": 3.8051932823856927, + "learning_rate": 4.998912861437106e-06, + "loss": 0.557, + "step": 271 + }, + { + "epoch": 0.10943472138402736, + "grad_norm": 4.261798228362777, + "learning_rate": 4.9988068676967176e-06, + "loss": 0.6685, + "step": 272 + }, + { + "epoch": 0.10983705491852747, + "grad_norm": 5.199991754679721, + "learning_rate": 4.9986959456193345e-06, + "loss": 0.6926, + "step": 273 + }, + { + "epoch": 0.11023938845302755, + "grad_norm": 4.800143443642004, + "learning_rate": 4.998580095423728e-06, + "loss": 0.6262, + "step": 274 + }, + { + "epoch": 0.11064172198752766, + "grad_norm": 5.703676474195908, + "learning_rate": 4.998459317338383e-06, + "loss": 0.8102, + "step": 275 + }, + { + "epoch": 0.11104405552202776, + "grad_norm": 5.554521472398849, + "learning_rate": 4.998333611601511e-06, + "loss": 0.8502, + "step": 276 + }, + { + "epoch": 0.11144638905652786, + "grad_norm": 5.223628882218702, + "learning_rate": 4.998202978461038e-06, + "loss": 0.6968, + "step": 277 + }, + { + "epoch": 0.11184872259102796, + "grad_norm": 5.028503778823592, + "learning_rate": 4.9980674181746055e-06, + "loss": 0.6237, + "step": 278 + }, + { + "epoch": 0.11225105612552806, + "grad_norm": 4.810062581233454, + "learning_rate": 4.997926931009579e-06, + "loss": 0.6566, + "step": 279 + }, + { + "epoch": 0.11265338966002816, + "grad_norm": 4.895795240240787, + "learning_rate": 4.997781517243037e-06, + "loss": 0.6358, + "step": 280 + }, + { + "epoch": 0.11305572319452827, + "grad_norm": 4.678689813671165, + "learning_rate": 4.997631177161775e-06, + "loss": 0.7098, + "step": 281 + }, + { + "epoch": 0.11345805672902837, + "grad_norm": 4.8363206866892465, + "learning_rate": 4.9974759110623065e-06, + "loss": 0.7234, + "step": 282 + }, + { + "epoch": 0.11386039026352847, + "grad_norm": 4.872936830415585, + "learning_rate": 4.997315719250857e-06, + "loss": 0.5937, + "step": 283 + }, + { + "epoch": 0.11426272379802857, + "grad_norm": 4.28046131366056, + "learning_rate": 4.997150602043371e-06, + "loss": 0.6211, + "step": 284 + }, + { + "epoch": 0.11466505733252867, + "grad_norm": 4.837073628599811, + "learning_rate": 4.996980559765504e-06, + "loss": 0.5828, + "step": 285 + }, + { + "epoch": 0.11506739086702877, + "grad_norm": 4.889286615585772, + "learning_rate": 4.996805592752626e-06, + "loss": 0.7301, + "step": 286 + }, + { + "epoch": 0.11546972440152886, + "grad_norm": 3.885336153971243, + "learning_rate": 4.996625701349822e-06, + "loss": 0.5091, + "step": 287 + }, + { + "epoch": 0.11587205793602896, + "grad_norm": 4.738737155243486, + "learning_rate": 4.996440885911886e-06, + "loss": 0.5519, + "step": 288 + }, + { + "epoch": 0.11627439147052906, + "grad_norm": 4.600896924297193, + "learning_rate": 4.996251146803326e-06, + "loss": 0.5357, + "step": 289 + }, + { + "epoch": 0.11667672500502917, + "grad_norm": 4.348185410245171, + "learning_rate": 4.996056484398359e-06, + "loss": 0.606, + "step": 290 + }, + { + "epoch": 0.11707905853952927, + "grad_norm": 4.609344088685658, + "learning_rate": 4.995856899080914e-06, + "loss": 0.7079, + "step": 291 + }, + { + "epoch": 0.11748139207402937, + "grad_norm": 4.982538761307109, + "learning_rate": 4.9956523912446274e-06, + "loss": 0.6626, + "step": 292 + }, + { + "epoch": 0.11788372560852947, + "grad_norm": 4.958292330122082, + "learning_rate": 4.9954429612928455e-06, + "loss": 0.5589, + "step": 293 + }, + { + "epoch": 0.11828605914302957, + "grad_norm": 5.3063992889234015, + "learning_rate": 4.995228609638623e-06, + "loss": 0.6761, + "step": 294 + }, + { + "epoch": 0.11868839267752967, + "grad_norm": 4.787491132159558, + "learning_rate": 4.995009336704718e-06, + "loss": 0.6907, + "step": 295 + }, + { + "epoch": 0.11909072621202978, + "grad_norm": 5.042532894966667, + "learning_rate": 4.994785142923599e-06, + "loss": 0.7259, + "step": 296 + }, + { + "epoch": 0.11949305974652988, + "grad_norm": 5.24654264942669, + "learning_rate": 4.994556028737439e-06, + "loss": 0.7376, + "step": 297 + }, + { + "epoch": 0.11989539328102998, + "grad_norm": 4.7381264053196475, + "learning_rate": 4.9943219945981126e-06, + "loss": 0.6419, + "step": 298 + }, + { + "epoch": 0.12029772681553007, + "grad_norm": 4.174895491105691, + "learning_rate": 4.9940830409672e-06, + "loss": 0.5846, + "step": 299 + }, + { + "epoch": 0.12070006035003017, + "grad_norm": 5.265350639824991, + "learning_rate": 4.993839168315985e-06, + "loss": 0.7269, + "step": 300 + }, + { + "epoch": 0.12110239388453027, + "grad_norm": 5.293670654130999, + "learning_rate": 4.99359037712545e-06, + "loss": 0.7539, + "step": 301 + }, + { + "epoch": 0.12150472741903037, + "grad_norm": 4.832726818218027, + "learning_rate": 4.993336667886281e-06, + "loss": 0.6617, + "step": 302 + }, + { + "epoch": 0.12190706095353047, + "grad_norm": 4.982756198499471, + "learning_rate": 4.993078041098862e-06, + "loss": 0.6517, + "step": 303 + }, + { + "epoch": 0.12230939448803058, + "grad_norm": 4.787146336142149, + "learning_rate": 4.9928144972732785e-06, + "loss": 0.5943, + "step": 304 + }, + { + "epoch": 0.12271172802253068, + "grad_norm": 5.725699716607812, + "learning_rate": 4.992546036929309e-06, + "loss": 0.7385, + "step": 305 + }, + { + "epoch": 0.12311406155703078, + "grad_norm": 4.388246603675133, + "learning_rate": 4.992272660596432e-06, + "loss": 0.5268, + "step": 306 + }, + { + "epoch": 0.12351639509153088, + "grad_norm": 4.323516644235707, + "learning_rate": 4.991994368813823e-06, + "loss": 0.5069, + "step": 307 + }, + { + "epoch": 0.12391872862603098, + "grad_norm": 4.968139714274168, + "learning_rate": 4.991711162130347e-06, + "loss": 0.6089, + "step": 308 + }, + { + "epoch": 0.12432106216053108, + "grad_norm": 5.393689160165955, + "learning_rate": 4.99142304110457e-06, + "loss": 0.5749, + "step": 309 + }, + { + "epoch": 0.12472339569503119, + "grad_norm": 5.010889684729688, + "learning_rate": 4.991130006304742e-06, + "loss": 0.6868, + "step": 310 + }, + { + "epoch": 0.12512572922953127, + "grad_norm": 4.77584804740267, + "learning_rate": 4.99083205830881e-06, + "loss": 0.7153, + "step": 311 + }, + { + "epoch": 0.1255280627640314, + "grad_norm": 3.8671259573453343, + "learning_rate": 4.990529197704411e-06, + "loss": 0.4938, + "step": 312 + }, + { + "epoch": 0.12593039629853148, + "grad_norm": 4.878545701273808, + "learning_rate": 4.990221425088868e-06, + "loss": 0.5954, + "step": 313 + }, + { + "epoch": 0.1263327298330316, + "grad_norm": 4.481594062259529, + "learning_rate": 4.989908741069195e-06, + "loss": 0.6397, + "step": 314 + }, + { + "epoch": 0.12673506336753168, + "grad_norm": 4.317626894827104, + "learning_rate": 4.98959114626209e-06, + "loss": 0.609, + "step": 315 + }, + { + "epoch": 0.1271373969020318, + "grad_norm": 4.415809282581258, + "learning_rate": 4.989268641293939e-06, + "loss": 0.5092, + "step": 316 + }, + { + "epoch": 0.12753973043653188, + "grad_norm": 4.4424914565481375, + "learning_rate": 4.9889412268008096e-06, + "loss": 0.6316, + "step": 317 + }, + { + "epoch": 0.12794206397103197, + "grad_norm": 4.431277408902586, + "learning_rate": 4.988608903428454e-06, + "loss": 0.6015, + "step": 318 + }, + { + "epoch": 0.1283443975055321, + "grad_norm": 4.618861534462686, + "learning_rate": 4.988271671832305e-06, + "loss": 0.5562, + "step": 319 + }, + { + "epoch": 0.12874673104003218, + "grad_norm": 5.078731601531793, + "learning_rate": 4.987929532677478e-06, + "loss": 0.7308, + "step": 320 + }, + { + "epoch": 0.1291490645745323, + "grad_norm": 4.814095780713649, + "learning_rate": 4.987582486638763e-06, + "loss": 0.6856, + "step": 321 + }, + { + "epoch": 0.12955139810903238, + "grad_norm": 5.171953766801144, + "learning_rate": 4.987230534400634e-06, + "loss": 0.7413, + "step": 322 + }, + { + "epoch": 0.1299537316435325, + "grad_norm": 4.527962518284751, + "learning_rate": 4.986873676657237e-06, + "loss": 0.5227, + "step": 323 + }, + { + "epoch": 0.13035606517803258, + "grad_norm": 4.9194193843235405, + "learning_rate": 4.986511914112392e-06, + "loss": 0.6038, + "step": 324 + }, + { + "epoch": 0.1307583987125327, + "grad_norm": 4.403379209157847, + "learning_rate": 4.986145247479597e-06, + "loss": 0.6624, + "step": 325 + }, + { + "epoch": 0.13116073224703279, + "grad_norm": 4.784759384702467, + "learning_rate": 4.985773677482018e-06, + "loss": 0.7109, + "step": 326 + }, + { + "epoch": 0.1315630657815329, + "grad_norm": 4.928746115672455, + "learning_rate": 4.985397204852496e-06, + "loss": 0.6241, + "step": 327 + }, + { + "epoch": 0.131965399316033, + "grad_norm": 5.6685753103245755, + "learning_rate": 4.985015830333536e-06, + "loss": 0.7121, + "step": 328 + }, + { + "epoch": 0.1323677328505331, + "grad_norm": 4.723811520101061, + "learning_rate": 4.984629554677316e-06, + "loss": 0.5683, + "step": 329 + }, + { + "epoch": 0.1327700663850332, + "grad_norm": 4.106210405752034, + "learning_rate": 4.984238378645677e-06, + "loss": 0.6128, + "step": 330 + }, + { + "epoch": 0.13317239991953328, + "grad_norm": 5.439987884236508, + "learning_rate": 4.983842303010127e-06, + "loss": 0.7294, + "step": 331 + }, + { + "epoch": 0.1335747334540334, + "grad_norm": 4.76281391948892, + "learning_rate": 4.9834413285518365e-06, + "loss": 0.6116, + "step": 332 + }, + { + "epoch": 0.13397706698853348, + "grad_norm": 4.594103270680001, + "learning_rate": 4.983035456061637e-06, + "loss": 0.5509, + "step": 333 + }, + { + "epoch": 0.1343794005230336, + "grad_norm": 4.8229418474364785, + "learning_rate": 4.982624686340022e-06, + "loss": 0.6565, + "step": 334 + }, + { + "epoch": 0.1347817340575337, + "grad_norm": 4.572125613909981, + "learning_rate": 4.9822090201971414e-06, + "loss": 0.5739, + "step": 335 + }, + { + "epoch": 0.1351840675920338, + "grad_norm": 4.657722588184461, + "learning_rate": 4.981788458452806e-06, + "loss": 0.716, + "step": 336 + }, + { + "epoch": 0.1355864011265339, + "grad_norm": 4.485714819813409, + "learning_rate": 4.9813630019364765e-06, + "loss": 0.5809, + "step": 337 + }, + { + "epoch": 0.135988734661034, + "grad_norm": 4.730520981461843, + "learning_rate": 4.9809326514872735e-06, + "loss": 0.6006, + "step": 338 + }, + { + "epoch": 0.1363910681955341, + "grad_norm": 4.407751381335739, + "learning_rate": 4.980497407953966e-06, + "loss": 0.6979, + "step": 339 + }, + { + "epoch": 0.1367934017300342, + "grad_norm": 5.036488553382197, + "learning_rate": 4.980057272194973e-06, + "loss": 0.6719, + "step": 340 + }, + { + "epoch": 0.1371957352645343, + "grad_norm": 4.131020183485464, + "learning_rate": 4.979612245078366e-06, + "loss": 0.656, + "step": 341 + }, + { + "epoch": 0.1375980687990344, + "grad_norm": 4.494385344326802, + "learning_rate": 4.979162327481859e-06, + "loss": 0.5824, + "step": 342 + }, + { + "epoch": 0.1380004023335345, + "grad_norm": 4.887767115382313, + "learning_rate": 4.9787075202928155e-06, + "loss": 0.7065, + "step": 343 + }, + { + "epoch": 0.1384027358680346, + "grad_norm": 4.929080555596624, + "learning_rate": 4.97824782440824e-06, + "loss": 0.6462, + "step": 344 + }, + { + "epoch": 0.1388050694025347, + "grad_norm": 4.768200287000246, + "learning_rate": 4.97778324073478e-06, + "loss": 0.7052, + "step": 345 + }, + { + "epoch": 0.1392074029370348, + "grad_norm": 4.824866676754677, + "learning_rate": 4.977313770188723e-06, + "loss": 0.6169, + "step": 346 + }, + { + "epoch": 0.1396097364715349, + "grad_norm": 5.0730067342381515, + "learning_rate": 4.976839413695994e-06, + "loss": 0.6907, + "step": 347 + }, + { + "epoch": 0.140012070006035, + "grad_norm": 4.400048281159567, + "learning_rate": 4.976360172192156e-06, + "loss": 0.6843, + "step": 348 + }, + { + "epoch": 0.1404144035405351, + "grad_norm": 4.826940942258961, + "learning_rate": 4.975876046622404e-06, + "loss": 0.6425, + "step": 349 + }, + { + "epoch": 0.1408167370750352, + "grad_norm": 4.1980516292894645, + "learning_rate": 4.975387037941568e-06, + "loss": 0.513, + "step": 350 + }, + { + "epoch": 0.1412190706095353, + "grad_norm": 4.599176289887721, + "learning_rate": 4.974893147114108e-06, + "loss": 0.6138, + "step": 351 + }, + { + "epoch": 0.1416214041440354, + "grad_norm": 5.224094683564219, + "learning_rate": 4.974394375114113e-06, + "loss": 0.7688, + "step": 352 + }, + { + "epoch": 0.14202373767853552, + "grad_norm": 4.886795665659055, + "learning_rate": 4.9738907229253e-06, + "loss": 0.6447, + "step": 353 + }, + { + "epoch": 0.1424260712130356, + "grad_norm": 4.686162997153853, + "learning_rate": 4.973382191541008e-06, + "loss": 0.6489, + "step": 354 + }, + { + "epoch": 0.14282840474753572, + "grad_norm": 4.633889069385792, + "learning_rate": 4.972868781964204e-06, + "loss": 0.6841, + "step": 355 + }, + { + "epoch": 0.1432307382820358, + "grad_norm": 4.714123893258144, + "learning_rate": 4.972350495207472e-06, + "loss": 0.7282, + "step": 356 + }, + { + "epoch": 0.1436330718165359, + "grad_norm": 4.620909154154889, + "learning_rate": 4.971827332293017e-06, + "loss": 0.6267, + "step": 357 + }, + { + "epoch": 0.144035405351036, + "grad_norm": 5.112805770554144, + "learning_rate": 4.971299294252659e-06, + "loss": 0.7187, + "step": 358 + }, + { + "epoch": 0.1444377388855361, + "grad_norm": 4.491917871078975, + "learning_rate": 4.970766382127838e-06, + "loss": 0.6243, + "step": 359 + }, + { + "epoch": 0.14484007242003621, + "grad_norm": 4.91569774829319, + "learning_rate": 4.9702285969696006e-06, + "loss": 0.6044, + "step": 360 + }, + { + "epoch": 0.1452424059545363, + "grad_norm": 4.347791813311231, + "learning_rate": 4.96968593983861e-06, + "loss": 0.6496, + "step": 361 + }, + { + "epoch": 0.14564473948903642, + "grad_norm": 4.6534811500518645, + "learning_rate": 4.9691384118051346e-06, + "loss": 0.7414, + "step": 362 + }, + { + "epoch": 0.1460470730235365, + "grad_norm": 4.5918592667861455, + "learning_rate": 4.968586013949051e-06, + "loss": 0.5802, + "step": 363 + }, + { + "epoch": 0.14644940655803662, + "grad_norm": 4.435482175102624, + "learning_rate": 4.96802874735984e-06, + "loss": 0.5162, + "step": 364 + }, + { + "epoch": 0.1468517400925367, + "grad_norm": 4.407001515831722, + "learning_rate": 4.967466613136586e-06, + "loss": 0.5361, + "step": 365 + }, + { + "epoch": 0.14725407362703682, + "grad_norm": 4.369142207445966, + "learning_rate": 4.966899612387972e-06, + "loss": 0.664, + "step": 366 + }, + { + "epoch": 0.1476564071615369, + "grad_norm": 4.607767758241303, + "learning_rate": 4.966327746232281e-06, + "loss": 0.5882, + "step": 367 + }, + { + "epoch": 0.14805874069603703, + "grad_norm": 4.938226519089643, + "learning_rate": 4.96575101579739e-06, + "loss": 0.5841, + "step": 368 + }, + { + "epoch": 0.14846107423053712, + "grad_norm": 4.096479564143721, + "learning_rate": 4.965169422220771e-06, + "loss": 0.5691, + "step": 369 + }, + { + "epoch": 0.1488634077650372, + "grad_norm": 4.775093512712085, + "learning_rate": 4.964582966649488e-06, + "loss": 0.6063, + "step": 370 + }, + { + "epoch": 0.14926574129953732, + "grad_norm": 4.211658294897708, + "learning_rate": 4.963991650240192e-06, + "loss": 0.5215, + "step": 371 + }, + { + "epoch": 0.1496680748340374, + "grad_norm": 4.364777540143213, + "learning_rate": 4.963395474159122e-06, + "loss": 0.4866, + "step": 372 + }, + { + "epoch": 0.15007040836853752, + "grad_norm": 4.681087632312734, + "learning_rate": 4.962794439582102e-06, + "loss": 0.6122, + "step": 373 + }, + { + "epoch": 0.1504727419030376, + "grad_norm": 5.077820479708522, + "learning_rate": 4.9621885476945405e-06, + "loss": 0.7555, + "step": 374 + }, + { + "epoch": 0.15087507543753773, + "grad_norm": 5.978600203961224, + "learning_rate": 4.961577799691421e-06, + "loss": 0.7113, + "step": 375 + }, + { + "epoch": 0.1512774089720378, + "grad_norm": 5.376960173895862, + "learning_rate": 4.960962196777307e-06, + "loss": 0.577, + "step": 376 + }, + { + "epoch": 0.15167974250653793, + "grad_norm": 5.18055770132837, + "learning_rate": 4.960341740166338e-06, + "loss": 0.6625, + "step": 377 + }, + { + "epoch": 0.15208207604103802, + "grad_norm": 4.460009104960063, + "learning_rate": 4.959716431082227e-06, + "loss": 0.5306, + "step": 378 + }, + { + "epoch": 0.15248440957553813, + "grad_norm": 4.303191111673615, + "learning_rate": 4.959086270758255e-06, + "loss": 0.5604, + "step": 379 + }, + { + "epoch": 0.15288674311003822, + "grad_norm": 4.8364861459301185, + "learning_rate": 4.9584512604372704e-06, + "loss": 0.6154, + "step": 380 + }, + { + "epoch": 0.15328907664453834, + "grad_norm": 4.638041078223287, + "learning_rate": 4.957811401371692e-06, + "loss": 0.5852, + "step": 381 + }, + { + "epoch": 0.15369141017903842, + "grad_norm": 4.8155087041398295, + "learning_rate": 4.9571666948234975e-06, + "loss": 0.591, + "step": 382 + }, + { + "epoch": 0.1540937437135385, + "grad_norm": 4.365724907978, + "learning_rate": 4.956517142064226e-06, + "loss": 0.5645, + "step": 383 + }, + { + "epoch": 0.15449607724803863, + "grad_norm": 4.800750875035706, + "learning_rate": 4.955862744374974e-06, + "loss": 0.6958, + "step": 384 + }, + { + "epoch": 0.15489841078253871, + "grad_norm": 4.218913109898912, + "learning_rate": 4.9552035030463955e-06, + "loss": 0.5981, + "step": 385 + }, + { + "epoch": 0.15530074431703883, + "grad_norm": 3.8608056472996775, + "learning_rate": 4.954539419378695e-06, + "loss": 0.5516, + "step": 386 + }, + { + "epoch": 0.15570307785153892, + "grad_norm": 4.422539247772956, + "learning_rate": 4.953870494681631e-06, + "loss": 0.466, + "step": 387 + }, + { + "epoch": 0.15610541138603903, + "grad_norm": 5.389694964679017, + "learning_rate": 4.9531967302745065e-06, + "loss": 0.7187, + "step": 388 + }, + { + "epoch": 0.15650774492053912, + "grad_norm": 4.641539036731872, + "learning_rate": 4.952518127486171e-06, + "loss": 0.6118, + "step": 389 + }, + { + "epoch": 0.15691007845503924, + "grad_norm": 4.38792439545412, + "learning_rate": 4.951834687655016e-06, + "loss": 0.6125, + "step": 390 + }, + { + "epoch": 0.15731241198953932, + "grad_norm": 4.859276712096591, + "learning_rate": 4.9511464121289745e-06, + "loss": 0.789, + "step": 391 + }, + { + "epoch": 0.15771474552403944, + "grad_norm": 4.123781252290954, + "learning_rate": 4.950453302265516e-06, + "loss": 0.5189, + "step": 392 + }, + { + "epoch": 0.15811707905853953, + "grad_norm": 4.5665024165035915, + "learning_rate": 4.949755359431643e-06, + "loss": 0.5843, + "step": 393 + }, + { + "epoch": 0.15851941259303964, + "grad_norm": 4.554644389766188, + "learning_rate": 4.949052585003892e-06, + "loss": 0.5165, + "step": 394 + }, + { + "epoch": 0.15892174612753973, + "grad_norm": 4.652445776343059, + "learning_rate": 4.9483449803683295e-06, + "loss": 0.6496, + "step": 395 + }, + { + "epoch": 0.15932407966203982, + "grad_norm": 4.155340040403773, + "learning_rate": 4.947632546920545e-06, + "loss": 0.5823, + "step": 396 + }, + { + "epoch": 0.15972641319653993, + "grad_norm": 5.685373741993176, + "learning_rate": 4.946915286065656e-06, + "loss": 0.7257, + "step": 397 + }, + { + "epoch": 0.16012874673104002, + "grad_norm": 3.794107938623536, + "learning_rate": 4.946193199218298e-06, + "loss": 0.5539, + "step": 398 + }, + { + "epoch": 0.16053108026554014, + "grad_norm": 3.345825002647911, + "learning_rate": 4.945466287802625e-06, + "loss": 0.4522, + "step": 399 + }, + { + "epoch": 0.16093341380004023, + "grad_norm": 4.499068362412442, + "learning_rate": 4.944734553252308e-06, + "loss": 0.7294, + "step": 400 + }, + { + "epoch": 0.16133574733454034, + "grad_norm": 4.710035321209727, + "learning_rate": 4.94399799701053e-06, + "loss": 0.6495, + "step": 401 + }, + { + "epoch": 0.16173808086904043, + "grad_norm": 4.665060195725978, + "learning_rate": 4.943256620529983e-06, + "loss": 0.5549, + "step": 402 + }, + { + "epoch": 0.16214041440354054, + "grad_norm": 4.526568640183799, + "learning_rate": 4.942510425272864e-06, + "loss": 0.5342, + "step": 403 + }, + { + "epoch": 0.16254274793804063, + "grad_norm": 4.500912340574114, + "learning_rate": 4.941759412710878e-06, + "loss": 0.5154, + "step": 404 + }, + { + "epoch": 0.16294508147254075, + "grad_norm": 4.871395743184042, + "learning_rate": 4.941003584325229e-06, + "loss": 0.7273, + "step": 405 + }, + { + "epoch": 0.16334741500704084, + "grad_norm": 5.226443625458406, + "learning_rate": 4.940242941606619e-06, + "loss": 0.5695, + "step": 406 + }, + { + "epoch": 0.16374974854154092, + "grad_norm": 4.3424277680190215, + "learning_rate": 4.939477486055245e-06, + "loss": 0.6903, + "step": 407 + }, + { + "epoch": 0.16415208207604104, + "grad_norm": 4.63055317361875, + "learning_rate": 4.938707219180797e-06, + "loss": 0.7724, + "step": 408 + }, + { + "epoch": 0.16455441561054113, + "grad_norm": 4.632025607743007, + "learning_rate": 4.9379321425024545e-06, + "loss": 0.5579, + "step": 409 + }, + { + "epoch": 0.16495674914504124, + "grad_norm": 4.949408477476733, + "learning_rate": 4.9371522575488814e-06, + "loss": 0.5585, + "step": 410 + }, + { + "epoch": 0.16535908267954133, + "grad_norm": 4.82833151193623, + "learning_rate": 4.936367565858228e-06, + "loss": 0.6281, + "step": 411 + }, + { + "epoch": 0.16576141621404145, + "grad_norm": 4.135170899195267, + "learning_rate": 4.935578068978121e-06, + "loss": 0.5236, + "step": 412 + }, + { + "epoch": 0.16616374974854153, + "grad_norm": 4.848865820547502, + "learning_rate": 4.9347837684656675e-06, + "loss": 0.7572, + "step": 413 + }, + { + "epoch": 0.16656608328304165, + "grad_norm": 4.717705787081838, + "learning_rate": 4.933984665887447e-06, + "loss": 0.7282, + "step": 414 + }, + { + "epoch": 0.16696841681754174, + "grad_norm": 4.528215697412672, + "learning_rate": 4.93318076281951e-06, + "loss": 0.5974, + "step": 415 + }, + { + "epoch": 0.16737075035204185, + "grad_norm": 5.165943035201723, + "learning_rate": 4.9323720608473754e-06, + "loss": 0.6622, + "step": 416 + }, + { + "epoch": 0.16777308388654194, + "grad_norm": 4.244718678666535, + "learning_rate": 4.9315585615660275e-06, + "loss": 0.5827, + "step": 417 + }, + { + "epoch": 0.16817541742104206, + "grad_norm": 5.152052375848677, + "learning_rate": 4.930740266579911e-06, + "loss": 0.6625, + "step": 418 + }, + { + "epoch": 0.16857775095554214, + "grad_norm": 4.55196067766976, + "learning_rate": 4.92991717750293e-06, + "loss": 0.5734, + "step": 419 + }, + { + "epoch": 0.16898008449004223, + "grad_norm": 5.130317188995626, + "learning_rate": 4.929089295958442e-06, + "loss": 0.6448, + "step": 420 + }, + { + "epoch": 0.16938241802454235, + "grad_norm": 4.221455100939359, + "learning_rate": 4.928256623579259e-06, + "loss": 0.5736, + "step": 421 + }, + { + "epoch": 0.16978475155904243, + "grad_norm": 5.042585074077865, + "learning_rate": 4.9274191620076405e-06, + "loss": 0.6948, + "step": 422 + }, + { + "epoch": 0.17018708509354255, + "grad_norm": 4.475716088239217, + "learning_rate": 4.926576912895292e-06, + "loss": 0.5762, + "step": 423 + }, + { + "epoch": 0.17058941862804264, + "grad_norm": 4.629681308629434, + "learning_rate": 4.9257298779033615e-06, + "loss": 0.7584, + "step": 424 + }, + { + "epoch": 0.17099175216254275, + "grad_norm": 4.624861162695507, + "learning_rate": 4.924878058702436e-06, + "loss": 0.6725, + "step": 425 + }, + { + "epoch": 0.17139408569704284, + "grad_norm": 4.729048209294549, + "learning_rate": 4.9240214569725375e-06, + "loss": 0.7722, + "step": 426 + }, + { + "epoch": 0.17179641923154296, + "grad_norm": 4.21954665313362, + "learning_rate": 4.923160074403122e-06, + "loss": 0.6253, + "step": 427 + }, + { + "epoch": 0.17219875276604305, + "grad_norm": 3.755869045635417, + "learning_rate": 4.922293912693072e-06, + "loss": 0.4626, + "step": 428 + }, + { + "epoch": 0.17260108630054316, + "grad_norm": 4.263780577574945, + "learning_rate": 4.921422973550699e-06, + "loss": 0.695, + "step": 429 + }, + { + "epoch": 0.17300341983504325, + "grad_norm": 5.108523767367576, + "learning_rate": 4.920547258693735e-06, + "loss": 0.7092, + "step": 430 + }, + { + "epoch": 0.17340575336954336, + "grad_norm": 4.317384116136861, + "learning_rate": 4.919666769849332e-06, + "loss": 0.4564, + "step": 431 + }, + { + "epoch": 0.17380808690404345, + "grad_norm": 5.268931042465572, + "learning_rate": 4.918781508754057e-06, + "loss": 0.7064, + "step": 432 + }, + { + "epoch": 0.17421042043854354, + "grad_norm": 4.295948103968784, + "learning_rate": 4.9178914771538875e-06, + "loss": 0.6892, + "step": 433 + }, + { + "epoch": 0.17461275397304366, + "grad_norm": 4.214776194153066, + "learning_rate": 4.916996676804214e-06, + "loss": 0.5877, + "step": 434 + }, + { + "epoch": 0.17501508750754374, + "grad_norm": 4.187051473349894, + "learning_rate": 4.916097109469829e-06, + "loss": 0.5542, + "step": 435 + }, + { + "epoch": 0.17541742104204386, + "grad_norm": 4.7370251117595235, + "learning_rate": 4.915192776924926e-06, + "loss": 0.6131, + "step": 436 + }, + { + "epoch": 0.17581975457654395, + "grad_norm": 4.790572516699059, + "learning_rate": 4.9142836809531e-06, + "loss": 0.6847, + "step": 437 + }, + { + "epoch": 0.17622208811104406, + "grad_norm": 4.024460804862858, + "learning_rate": 4.91336982334734e-06, + "loss": 0.4475, + "step": 438 + }, + { + "epoch": 0.17662442164554415, + "grad_norm": 4.675530435693111, + "learning_rate": 4.912451205910024e-06, + "loss": 0.641, + "step": 439 + }, + { + "epoch": 0.17702675518004427, + "grad_norm": 4.070435859439258, + "learning_rate": 4.91152783045292e-06, + "loss": 0.5715, + "step": 440 + }, + { + "epoch": 0.17742908871454435, + "grad_norm": 5.014878172443035, + "learning_rate": 4.910599698797179e-06, + "loss": 0.6227, + "step": 441 + }, + { + "epoch": 0.17783142224904447, + "grad_norm": 4.471540054012383, + "learning_rate": 4.909666812773333e-06, + "loss": 0.5145, + "step": 442 + }, + { + "epoch": 0.17823375578354456, + "grad_norm": 4.442414191534445, + "learning_rate": 4.908729174221289e-06, + "loss": 0.6688, + "step": 443 + }, + { + "epoch": 0.17863608931804467, + "grad_norm": 4.624231827213014, + "learning_rate": 4.9077867849903325e-06, + "loss": 0.6327, + "step": 444 + }, + { + "epoch": 0.17903842285254476, + "grad_norm": 4.743489707184984, + "learning_rate": 4.906839646939113e-06, + "loss": 0.7075, + "step": 445 + }, + { + "epoch": 0.17944075638704485, + "grad_norm": 4.200250704275894, + "learning_rate": 4.905887761935649e-06, + "loss": 0.5171, + "step": 446 + }, + { + "epoch": 0.17984308992154496, + "grad_norm": 4.0039914177118545, + "learning_rate": 4.90493113185732e-06, + "loss": 0.5334, + "step": 447 + }, + { + "epoch": 0.18024542345604505, + "grad_norm": 4.437438466205263, + "learning_rate": 4.903969758590865e-06, + "loss": 0.584, + "step": 448 + }, + { + "epoch": 0.18064775699054517, + "grad_norm": 4.668639275979892, + "learning_rate": 4.9030036440323766e-06, + "loss": 0.669, + "step": 449 + }, + { + "epoch": 0.18105009052504525, + "grad_norm": 4.262713251256841, + "learning_rate": 4.902032790087301e-06, + "loss": 0.518, + "step": 450 + }, + { + "epoch": 0.18145242405954537, + "grad_norm": 4.58562990715973, + "learning_rate": 4.9010571986704295e-06, + "loss": 0.5627, + "step": 451 + }, + { + "epoch": 0.18185475759404546, + "grad_norm": 4.2511574164879615, + "learning_rate": 4.900076871705897e-06, + "loss": 0.6189, + "step": 452 + }, + { + "epoch": 0.18225709112854557, + "grad_norm": 3.879096119034437, + "learning_rate": 4.89909181112718e-06, + "loss": 0.4688, + "step": 453 + }, + { + "epoch": 0.18265942466304566, + "grad_norm": 4.4698856213971645, + "learning_rate": 4.898102018877088e-06, + "loss": 0.6272, + "step": 454 + }, + { + "epoch": 0.18306175819754578, + "grad_norm": 4.25349966912043, + "learning_rate": 4.897107496907767e-06, + "loss": 0.4844, + "step": 455 + }, + { + "epoch": 0.18346409173204586, + "grad_norm": 5.180942020863308, + "learning_rate": 4.896108247180688e-06, + "loss": 0.7363, + "step": 456 + }, + { + "epoch": 0.18386642526654598, + "grad_norm": 4.464543366415725, + "learning_rate": 4.895104271666647e-06, + "loss": 0.507, + "step": 457 + }, + { + "epoch": 0.18426875880104607, + "grad_norm": 4.35656943725317, + "learning_rate": 4.8940955723457604e-06, + "loss": 0.5626, + "step": 458 + }, + { + "epoch": 0.18467109233554616, + "grad_norm": 4.46990880672421, + "learning_rate": 4.893082151207464e-06, + "loss": 0.5986, + "step": 459 + }, + { + "epoch": 0.18507342587004627, + "grad_norm": 4.315333956337568, + "learning_rate": 4.8920640102505005e-06, + "loss": 0.5399, + "step": 460 + }, + { + "epoch": 0.18547575940454636, + "grad_norm": 4.414262914713829, + "learning_rate": 4.891041151482928e-06, + "loss": 0.6866, + "step": 461 + }, + { + "epoch": 0.18587809293904647, + "grad_norm": 4.68767919716049, + "learning_rate": 4.8900135769221045e-06, + "loss": 0.6205, + "step": 462 + }, + { + "epoch": 0.18628042647354656, + "grad_norm": 4.55495109723687, + "learning_rate": 4.888981288594692e-06, + "loss": 0.6232, + "step": 463 + }, + { + "epoch": 0.18668276000804668, + "grad_norm": 5.072112114859536, + "learning_rate": 4.887944288536648e-06, + "loss": 0.6246, + "step": 464 + }, + { + "epoch": 0.18708509354254677, + "grad_norm": 4.598620261784645, + "learning_rate": 4.886902578793221e-06, + "loss": 0.6688, + "step": 465 + }, + { + "epoch": 0.18748742707704688, + "grad_norm": 4.156182800147685, + "learning_rate": 4.885856161418953e-06, + "loss": 0.6185, + "step": 466 + }, + { + "epoch": 0.18788976061154697, + "grad_norm": 5.123876789008787, + "learning_rate": 4.8848050384776665e-06, + "loss": 0.6485, + "step": 467 + }, + { + "epoch": 0.18829209414604708, + "grad_norm": 4.7877986741742555, + "learning_rate": 4.883749212042466e-06, + "loss": 0.6336, + "step": 468 + }, + { + "epoch": 0.18869442768054717, + "grad_norm": 4.170455407715718, + "learning_rate": 4.882688684195732e-06, + "loss": 0.5708, + "step": 469 + }, + { + "epoch": 0.1890967612150473, + "grad_norm": 4.394557070936156, + "learning_rate": 4.881623457029121e-06, + "loss": 0.6584, + "step": 470 + }, + { + "epoch": 0.18949909474954738, + "grad_norm": 4.4866621521500205, + "learning_rate": 4.880553532643552e-06, + "loss": 0.5816, + "step": 471 + }, + { + "epoch": 0.18990142828404746, + "grad_norm": 4.263089971509693, + "learning_rate": 4.879478913149212e-06, + "loss": 0.6563, + "step": 472 + }, + { + "epoch": 0.19030376181854758, + "grad_norm": 3.897184614846541, + "learning_rate": 4.8783996006655485e-06, + "loss": 0.5816, + "step": 473 + }, + { + "epoch": 0.19070609535304767, + "grad_norm": 4.759719975025391, + "learning_rate": 4.877315597321263e-06, + "loss": 0.7219, + "step": 474 + }, + { + "epoch": 0.19110842888754778, + "grad_norm": 4.561889671156211, + "learning_rate": 4.876226905254309e-06, + "loss": 0.6307, + "step": 475 + }, + { + "epoch": 0.19151076242204787, + "grad_norm": 4.622879218120913, + "learning_rate": 4.875133526611888e-06, + "loss": 0.5739, + "step": 476 + }, + { + "epoch": 0.19191309595654799, + "grad_norm": 4.183836330008031, + "learning_rate": 4.874035463550445e-06, + "loss": 0.4165, + "step": 477 + }, + { + "epoch": 0.19231542949104807, + "grad_norm": 4.431780943086593, + "learning_rate": 4.872932718235663e-06, + "loss": 0.5768, + "step": 478 + }, + { + "epoch": 0.1927177630255482, + "grad_norm": 4.700772397419956, + "learning_rate": 4.87182529284246e-06, + "loss": 0.6635, + "step": 479 + }, + { + "epoch": 0.19312009656004828, + "grad_norm": 4.0384190687435915, + "learning_rate": 4.870713189554985e-06, + "loss": 0.4994, + "step": 480 + }, + { + "epoch": 0.1935224300945484, + "grad_norm": 4.16066989214431, + "learning_rate": 4.869596410566614e-06, + "loss": 0.523, + "step": 481 + }, + { + "epoch": 0.19392476362904848, + "grad_norm": 3.7758221004395565, + "learning_rate": 4.8684749580799405e-06, + "loss": 0.5106, + "step": 482 + }, + { + "epoch": 0.1943270971635486, + "grad_norm": 4.792256321316127, + "learning_rate": 4.867348834306781e-06, + "loss": 0.5955, + "step": 483 + }, + { + "epoch": 0.19472943069804868, + "grad_norm": 4.408528934091589, + "learning_rate": 4.866218041468161e-06, + "loss": 0.6032, + "step": 484 + }, + { + "epoch": 0.19513176423254877, + "grad_norm": 4.425894822386934, + "learning_rate": 4.865082581794317e-06, + "loss": 0.5298, + "step": 485 + }, + { + "epoch": 0.1955340977670489, + "grad_norm": 4.356368057153947, + "learning_rate": 4.863942457524689e-06, + "loss": 0.5381, + "step": 486 + }, + { + "epoch": 0.19593643130154897, + "grad_norm": 3.990737447917161, + "learning_rate": 4.862797670907915e-06, + "loss": 0.5049, + "step": 487 + }, + { + "epoch": 0.1963387648360491, + "grad_norm": 4.229237889406895, + "learning_rate": 4.8616482242018325e-06, + "loss": 0.5467, + "step": 488 + }, + { + "epoch": 0.19674109837054918, + "grad_norm": 4.126506367371427, + "learning_rate": 4.860494119673466e-06, + "loss": 0.5455, + "step": 489 + }, + { + "epoch": 0.1971434319050493, + "grad_norm": 5.231977983505604, + "learning_rate": 4.859335359599031e-06, + "loss": 0.7528, + "step": 490 + }, + { + "epoch": 0.19754576543954938, + "grad_norm": 5.1291616835145115, + "learning_rate": 4.85817194626392e-06, + "loss": 0.679, + "step": 491 + }, + { + "epoch": 0.1979480989740495, + "grad_norm": 3.737681461647685, + "learning_rate": 4.857003881962705e-06, + "loss": 0.4926, + "step": 492 + }, + { + "epoch": 0.19835043250854958, + "grad_norm": 4.703925829656442, + "learning_rate": 4.855831168999132e-06, + "loss": 0.6439, + "step": 493 + }, + { + "epoch": 0.1987527660430497, + "grad_norm": 4.627856862526262, + "learning_rate": 4.854653809686115e-06, + "loss": 0.6698, + "step": 494 + }, + { + "epoch": 0.1991550995775498, + "grad_norm": 4.364999176797817, + "learning_rate": 4.853471806345732e-06, + "loss": 0.6015, + "step": 495 + }, + { + "epoch": 0.19955743311204988, + "grad_norm": 4.602684651491329, + "learning_rate": 4.852285161309218e-06, + "loss": 0.5809, + "step": 496 + }, + { + "epoch": 0.19995976664655, + "grad_norm": 4.325677574896615, + "learning_rate": 4.851093876916967e-06, + "loss": 0.526, + "step": 497 + }, + { + "epoch": 0.20036210018105008, + "grad_norm": 4.81776828160967, + "learning_rate": 4.849897955518518e-06, + "loss": 0.6654, + "step": 498 + }, + { + "epoch": 0.2007644337155502, + "grad_norm": 5.49681992539213, + "learning_rate": 4.848697399472561e-06, + "loss": 0.8468, + "step": 499 + }, + { + "epoch": 0.20116676725005028, + "grad_norm": 4.782602085058219, + "learning_rate": 4.8474922111469225e-06, + "loss": 0.6339, + "step": 500 + }, + { + "epoch": 0.2015691007845504, + "grad_norm": 4.705465090280715, + "learning_rate": 4.846282392918566e-06, + "loss": 0.5111, + "step": 501 + }, + { + "epoch": 0.20197143431905049, + "grad_norm": 4.044280178788859, + "learning_rate": 4.845067947173589e-06, + "loss": 0.4453, + "step": 502 + }, + { + "epoch": 0.2023737678535506, + "grad_norm": 5.299453770363048, + "learning_rate": 4.843848876307211e-06, + "loss": 0.7287, + "step": 503 + }, + { + "epoch": 0.2027761013880507, + "grad_norm": 3.9060416143674, + "learning_rate": 4.842625182723779e-06, + "loss": 0.4903, + "step": 504 + }, + { + "epoch": 0.2031784349225508, + "grad_norm": 4.320966869452432, + "learning_rate": 4.841396868836753e-06, + "loss": 0.5312, + "step": 505 + }, + { + "epoch": 0.2035807684570509, + "grad_norm": 4.059590568863906, + "learning_rate": 4.840163937068707e-06, + "loss": 0.5953, + "step": 506 + }, + { + "epoch": 0.203983101991551, + "grad_norm": 4.544560777009811, + "learning_rate": 4.838926389851324e-06, + "loss": 0.5274, + "step": 507 + }, + { + "epoch": 0.2043854355260511, + "grad_norm": 4.643239965625325, + "learning_rate": 4.837684229625389e-06, + "loss": 0.6284, + "step": 508 + }, + { + "epoch": 0.20478776906055118, + "grad_norm": 3.848260216211789, + "learning_rate": 4.836437458840783e-06, + "loss": 0.5052, + "step": 509 + }, + { + "epoch": 0.2051901025950513, + "grad_norm": 4.65542258301039, + "learning_rate": 4.835186079956483e-06, + "loss": 0.5641, + "step": 510 + }, + { + "epoch": 0.2055924361295514, + "grad_norm": 3.7048624001601738, + "learning_rate": 4.8339300954405545e-06, + "loss": 0.5507, + "step": 511 + }, + { + "epoch": 0.2059947696640515, + "grad_norm": 4.721724963544143, + "learning_rate": 4.832669507770144e-06, + "loss": 0.6945, + "step": 512 + }, + { + "epoch": 0.2063971031985516, + "grad_norm": 3.9937965024761892, + "learning_rate": 4.83140431943148e-06, + "loss": 0.5444, + "step": 513 + }, + { + "epoch": 0.2067994367330517, + "grad_norm": 4.150283320138307, + "learning_rate": 4.830134532919863e-06, + "loss": 0.5835, + "step": 514 + }, + { + "epoch": 0.2072017702675518, + "grad_norm": 5.2365671694488904, + "learning_rate": 4.828860150739662e-06, + "loss": 0.697, + "step": 515 + }, + { + "epoch": 0.2076041038020519, + "grad_norm": 3.98085660862504, + "learning_rate": 4.827581175404311e-06, + "loss": 0.4872, + "step": 516 + }, + { + "epoch": 0.208006437336552, + "grad_norm": 4.574435055815776, + "learning_rate": 4.8262976094363016e-06, + "loss": 0.5925, + "step": 517 + }, + { + "epoch": 0.2084087708710521, + "grad_norm": 4.459834553682444, + "learning_rate": 4.825009455367181e-06, + "loss": 0.6426, + "step": 518 + }, + { + "epoch": 0.2088111044055522, + "grad_norm": 3.8557681901771037, + "learning_rate": 4.823716715737544e-06, + "loss": 0.5719, + "step": 519 + }, + { + "epoch": 0.20921343794005232, + "grad_norm": 4.776729631302629, + "learning_rate": 4.82241939309703e-06, + "loss": 0.6118, + "step": 520 + }, + { + "epoch": 0.2096157714745524, + "grad_norm": 4.701717061474551, + "learning_rate": 4.821117490004319e-06, + "loss": 0.4766, + "step": 521 + }, + { + "epoch": 0.2100181050090525, + "grad_norm": 4.486472816066567, + "learning_rate": 4.819811009027122e-06, + "loss": 0.6134, + "step": 522 + }, + { + "epoch": 0.2104204385435526, + "grad_norm": 4.784573013198364, + "learning_rate": 4.818499952742179e-06, + "loss": 0.5982, + "step": 523 + }, + { + "epoch": 0.2108227720780527, + "grad_norm": 4.953599282635229, + "learning_rate": 4.817184323735258e-06, + "loss": 0.5732, + "step": 524 + }, + { + "epoch": 0.2112251056125528, + "grad_norm": 4.808483380881421, + "learning_rate": 4.815864124601139e-06, + "loss": 0.6771, + "step": 525 + }, + { + "epoch": 0.2116274391470529, + "grad_norm": 4.024625165836511, + "learning_rate": 4.814539357943622e-06, + "loss": 0.6355, + "step": 526 + }, + { + "epoch": 0.212029772681553, + "grad_norm": 3.943315026550887, + "learning_rate": 4.813210026375513e-06, + "loss": 0.521, + "step": 527 + }, + { + "epoch": 0.2124321062160531, + "grad_norm": 3.463894710975418, + "learning_rate": 4.811876132518618e-06, + "loss": 0.4542, + "step": 528 + }, + { + "epoch": 0.21283443975055322, + "grad_norm": 4.513349087640872, + "learning_rate": 4.810537679003746e-06, + "loss": 0.6093, + "step": 529 + }, + { + "epoch": 0.2132367732850533, + "grad_norm": 4.649205790356446, + "learning_rate": 4.8091946684706956e-06, + "loss": 0.6726, + "step": 530 + }, + { + "epoch": 0.21363910681955342, + "grad_norm": 4.2037442337203315, + "learning_rate": 4.807847103568255e-06, + "loss": 0.6265, + "step": 531 + }, + { + "epoch": 0.2140414403540535, + "grad_norm": 4.893988522032171, + "learning_rate": 4.806494986954195e-06, + "loss": 0.7373, + "step": 532 + }, + { + "epoch": 0.21444377388855362, + "grad_norm": 4.455715486715626, + "learning_rate": 4.805138321295262e-06, + "loss": 0.5457, + "step": 533 + }, + { + "epoch": 0.2148461074230537, + "grad_norm": 4.649691607815862, + "learning_rate": 4.803777109267174e-06, + "loss": 0.5872, + "step": 534 + }, + { + "epoch": 0.2152484409575538, + "grad_norm": 4.606972921739361, + "learning_rate": 4.802411353554619e-06, + "loss": 0.6556, + "step": 535 + }, + { + "epoch": 0.21565077449205391, + "grad_norm": 4.262679120822924, + "learning_rate": 4.801041056851244e-06, + "loss": 0.5814, + "step": 536 + }, + { + "epoch": 0.216053108026554, + "grad_norm": 4.734408703763602, + "learning_rate": 4.7996662218596505e-06, + "loss": 0.5975, + "step": 537 + }, + { + "epoch": 0.21645544156105412, + "grad_norm": 4.0966775674750515, + "learning_rate": 4.798286851291395e-06, + "loss": 0.5373, + "step": 538 + }, + { + "epoch": 0.2168577750955542, + "grad_norm": 4.794723705583196, + "learning_rate": 4.796902947866976e-06, + "loss": 0.6411, + "step": 539 + }, + { + "epoch": 0.21726010863005432, + "grad_norm": 5.177706885118282, + "learning_rate": 4.795514514315833e-06, + "loss": 0.7129, + "step": 540 + }, + { + "epoch": 0.2176624421645544, + "grad_norm": 4.235334038988407, + "learning_rate": 4.794121553376341e-06, + "loss": 0.6368, + "step": 541 + }, + { + "epoch": 0.21806477569905452, + "grad_norm": 4.250039615427904, + "learning_rate": 4.792724067795802e-06, + "loss": 0.5897, + "step": 542 + }, + { + "epoch": 0.2184671092335546, + "grad_norm": 4.248095551159398, + "learning_rate": 4.7913220603304455e-06, + "loss": 0.6119, + "step": 543 + }, + { + "epoch": 0.21886944276805473, + "grad_norm": 4.7784528939759126, + "learning_rate": 4.789915533745415e-06, + "loss": 0.6428, + "step": 544 + }, + { + "epoch": 0.21927177630255482, + "grad_norm": 3.7433210513797857, + "learning_rate": 4.788504490814771e-06, + "loss": 0.4616, + "step": 545 + }, + { + "epoch": 0.21967410983705493, + "grad_norm": 4.477510756273104, + "learning_rate": 4.78708893432148e-06, + "loss": 0.5564, + "step": 546 + }, + { + "epoch": 0.22007644337155502, + "grad_norm": 4.159189423897508, + "learning_rate": 4.785668867057408e-06, + "loss": 0.5005, + "step": 547 + }, + { + "epoch": 0.2204787769060551, + "grad_norm": 4.801923291900065, + "learning_rate": 4.784244291823321e-06, + "loss": 0.6703, + "step": 548 + }, + { + "epoch": 0.22088111044055522, + "grad_norm": 4.9705885588356065, + "learning_rate": 4.782815211428875e-06, + "loss": 0.6892, + "step": 549 + }, + { + "epoch": 0.2212834439750553, + "grad_norm": 4.689425225044057, + "learning_rate": 4.7813816286926116e-06, + "loss": 0.6438, + "step": 550 + }, + { + "epoch": 0.22168577750955543, + "grad_norm": 3.9964370496914836, + "learning_rate": 4.7799435464419496e-06, + "loss": 0.5283, + "step": 551 + }, + { + "epoch": 0.22208811104405551, + "grad_norm": 3.9063906383285523, + "learning_rate": 4.778500967513186e-06, + "loss": 0.5477, + "step": 552 + }, + { + "epoch": 0.22249044457855563, + "grad_norm": 4.972674347366186, + "learning_rate": 4.777053894751484e-06, + "loss": 0.6762, + "step": 553 + }, + { + "epoch": 0.22289277811305572, + "grad_norm": 4.605207302352324, + "learning_rate": 4.77560233101087e-06, + "loss": 0.5823, + "step": 554 + }, + { + "epoch": 0.22329511164755583, + "grad_norm": 5.023738846112498, + "learning_rate": 4.774146279154231e-06, + "loss": 0.645, + "step": 555 + }, + { + "epoch": 0.22369744518205592, + "grad_norm": 4.21478405630217, + "learning_rate": 4.772685742053299e-06, + "loss": 0.6361, + "step": 556 + }, + { + "epoch": 0.22409977871655604, + "grad_norm": 4.443333059344436, + "learning_rate": 4.7712207225886605e-06, + "loss": 0.5808, + "step": 557 + }, + { + "epoch": 0.22450211225105612, + "grad_norm": 4.136335273179504, + "learning_rate": 4.7697512236497355e-06, + "loss": 0.4543, + "step": 558 + }, + { + "epoch": 0.22490444578555624, + "grad_norm": 5.228702785570883, + "learning_rate": 4.7682772481347835e-06, + "loss": 0.6677, + "step": 559 + }, + { + "epoch": 0.22530677932005633, + "grad_norm": 4.607023738968648, + "learning_rate": 4.76679879895089e-06, + "loss": 0.6972, + "step": 560 + }, + { + "epoch": 0.22570911285455642, + "grad_norm": 3.9623979182055646, + "learning_rate": 4.7653158790139655e-06, + "loss": 0.562, + "step": 561 + }, + { + "epoch": 0.22611144638905653, + "grad_norm": 4.785540664618103, + "learning_rate": 4.763828491248737e-06, + "loss": 0.6667, + "step": 562 + }, + { + "epoch": 0.22651377992355662, + "grad_norm": 4.626975611158776, + "learning_rate": 4.762336638588745e-06, + "loss": 0.5222, + "step": 563 + }, + { + "epoch": 0.22691611345805673, + "grad_norm": 4.148716543735861, + "learning_rate": 4.760840323976333e-06, + "loss": 0.5792, + "step": 564 + }, + { + "epoch": 0.22731844699255682, + "grad_norm": 3.7576481851339785, + "learning_rate": 4.759339550362647e-06, + "loss": 0.4171, + "step": 565 + }, + { + "epoch": 0.22772078052705694, + "grad_norm": 4.056873981914918, + "learning_rate": 4.757834320707629e-06, + "loss": 0.5823, + "step": 566 + }, + { + "epoch": 0.22812311406155703, + "grad_norm": 4.485956120607346, + "learning_rate": 4.7563246379800036e-06, + "loss": 0.6038, + "step": 567 + }, + { + "epoch": 0.22852544759605714, + "grad_norm": 4.095512060377993, + "learning_rate": 4.754810505157285e-06, + "loss": 0.5807, + "step": 568 + }, + { + "epoch": 0.22892778113055723, + "grad_norm": 4.316692633895088, + "learning_rate": 4.753291925225761e-06, + "loss": 0.5689, + "step": 569 + }, + { + "epoch": 0.22933011466505734, + "grad_norm": 4.0964431049127095, + "learning_rate": 4.751768901180488e-06, + "loss": 0.4724, + "step": 570 + }, + { + "epoch": 0.22973244819955743, + "grad_norm": 3.8505086953071683, + "learning_rate": 4.750241436025292e-06, + "loss": 0.5599, + "step": 571 + }, + { + "epoch": 0.23013478173405755, + "grad_norm": 4.97127913813276, + "learning_rate": 4.748709532772756e-06, + "loss": 0.7189, + "step": 572 + }, + { + "epoch": 0.23053711526855764, + "grad_norm": 4.376792824386442, + "learning_rate": 4.7471731944442154e-06, + "loss": 0.6149, + "step": 573 + }, + { + "epoch": 0.23093944880305772, + "grad_norm": 4.633754764886852, + "learning_rate": 4.745632424069755e-06, + "loss": 0.6118, + "step": 574 + }, + { + "epoch": 0.23134178233755784, + "grad_norm": 4.682951429282118, + "learning_rate": 4.744087224688197e-06, + "loss": 0.6089, + "step": 575 + }, + { + "epoch": 0.23174411587205793, + "grad_norm": 4.504821455509545, + "learning_rate": 4.742537599347101e-06, + "loss": 0.5747, + "step": 576 + }, + { + "epoch": 0.23214644940655804, + "grad_norm": 4.922801626757808, + "learning_rate": 4.740983551102759e-06, + "loss": 0.7267, + "step": 577 + }, + { + "epoch": 0.23254878294105813, + "grad_norm": 4.128104128218885, + "learning_rate": 4.73942508302018e-06, + "loss": 0.5344, + "step": 578 + }, + { + "epoch": 0.23295111647555825, + "grad_norm": 4.502862129743137, + "learning_rate": 4.7378621981730966e-06, + "loss": 0.4465, + "step": 579 + }, + { + "epoch": 0.23335345001005833, + "grad_norm": 4.476573377559871, + "learning_rate": 4.736294899643946e-06, + "loss": 0.5327, + "step": 580 + }, + { + "epoch": 0.23375578354455845, + "grad_norm": 4.525722183672701, + "learning_rate": 4.734723190523875e-06, + "loss": 0.6182, + "step": 581 + }, + { + "epoch": 0.23415811707905854, + "grad_norm": 4.4174724143850455, + "learning_rate": 4.7331470739127284e-06, + "loss": 0.582, + "step": 582 + }, + { + "epoch": 0.23456045061355865, + "grad_norm": 4.087588153474757, + "learning_rate": 4.731566552919042e-06, + "loss": 0.5626, + "step": 583 + }, + { + "epoch": 0.23496278414805874, + "grad_norm": 4.793168855698109, + "learning_rate": 4.7299816306600395e-06, + "loss": 0.712, + "step": 584 + }, + { + "epoch": 0.23536511768255883, + "grad_norm": 4.088862969423477, + "learning_rate": 4.728392310261628e-06, + "loss": 0.5867, + "step": 585 + }, + { + "epoch": 0.23576745121705894, + "grad_norm": 4.31442875545057, + "learning_rate": 4.726798594858382e-06, + "loss": 0.5255, + "step": 586 + }, + { + "epoch": 0.23616978475155903, + "grad_norm": 4.2694537336500495, + "learning_rate": 4.7252004875935506e-06, + "loss": 0.6822, + "step": 587 + }, + { + "epoch": 0.23657211828605915, + "grad_norm": 4.179846752454285, + "learning_rate": 4.723597991619043e-06, + "loss": 0.5433, + "step": 588 + }, + { + "epoch": 0.23697445182055923, + "grad_norm": 4.1764050880208, + "learning_rate": 4.721991110095422e-06, + "loss": 0.586, + "step": 589 + }, + { + "epoch": 0.23737678535505935, + "grad_norm": 3.699033328030597, + "learning_rate": 4.720379846191903e-06, + "loss": 0.4942, + "step": 590 + }, + { + "epoch": 0.23777911888955944, + "grad_norm": 4.798015125682991, + "learning_rate": 4.718764203086342e-06, + "loss": 0.7269, + "step": 591 + }, + { + "epoch": 0.23818145242405955, + "grad_norm": 4.974938507339555, + "learning_rate": 4.717144183965233e-06, + "loss": 0.6753, + "step": 592 + }, + { + "epoch": 0.23858378595855964, + "grad_norm": 5.08563513422606, + "learning_rate": 4.715519792023703e-06, + "loss": 0.6613, + "step": 593 + }, + { + "epoch": 0.23898611949305976, + "grad_norm": 4.6691045707182655, + "learning_rate": 4.7138910304655e-06, + "loss": 0.5635, + "step": 594 + }, + { + "epoch": 0.23938845302755984, + "grad_norm": 4.366016787794075, + "learning_rate": 4.712257902502992e-06, + "loss": 0.6603, + "step": 595 + }, + { + "epoch": 0.23979078656205996, + "grad_norm": 4.98238528037948, + "learning_rate": 4.710620411357157e-06, + "loss": 0.6241, + "step": 596 + }, + { + "epoch": 0.24019312009656005, + "grad_norm": 4.023463566077988, + "learning_rate": 4.7089785602575805e-06, + "loss": 0.5202, + "step": 597 + }, + { + "epoch": 0.24059545363106014, + "grad_norm": 4.040098710840788, + "learning_rate": 4.707332352442446e-06, + "loss": 0.5843, + "step": 598 + }, + { + "epoch": 0.24099778716556025, + "grad_norm": 4.151014624312927, + "learning_rate": 4.705681791158531e-06, + "loss": 0.6206, + "step": 599 + }, + { + "epoch": 0.24140012070006034, + "grad_norm": 4.0875213661566665, + "learning_rate": 4.704026879661196e-06, + "loss": 0.5614, + "step": 600 + }, + { + "epoch": 0.24180245423456045, + "grad_norm": 4.553877858262055, + "learning_rate": 4.7023676212143834e-06, + "loss": 0.6649, + "step": 601 + }, + { + "epoch": 0.24220478776906054, + "grad_norm": 4.674650194014434, + "learning_rate": 4.70070401909061e-06, + "loss": 0.606, + "step": 602 + }, + { + "epoch": 0.24260712130356066, + "grad_norm": 4.63353917939836, + "learning_rate": 4.699036076570959e-06, + "loss": 0.6668, + "step": 603 + }, + { + "epoch": 0.24300945483806075, + "grad_norm": 4.723870743698028, + "learning_rate": 4.697363796945072e-06, + "loss": 0.6728, + "step": 604 + }, + { + "epoch": 0.24341178837256086, + "grad_norm": 4.386624516452244, + "learning_rate": 4.6956871835111475e-06, + "loss": 0.5139, + "step": 605 + }, + { + "epoch": 0.24381412190706095, + "grad_norm": 4.244611479893601, + "learning_rate": 4.694006239575929e-06, + "loss": 0.609, + "step": 606 + }, + { + "epoch": 0.24421645544156106, + "grad_norm": 4.072589096296677, + "learning_rate": 4.692320968454702e-06, + "loss": 0.5657, + "step": 607 + }, + { + "epoch": 0.24461878897606115, + "grad_norm": 4.88910796894103, + "learning_rate": 4.690631373471287e-06, + "loss": 0.701, + "step": 608 + }, + { + "epoch": 0.24502112251056127, + "grad_norm": 4.686738823395271, + "learning_rate": 4.6889374579580315e-06, + "loss": 0.637, + "step": 609 + }, + { + "epoch": 0.24542345604506136, + "grad_norm": 4.905698840133413, + "learning_rate": 4.687239225255805e-06, + "loss": 0.6003, + "step": 610 + }, + { + "epoch": 0.24582578957956144, + "grad_norm": 4.338740241210239, + "learning_rate": 4.68553667871399e-06, + "loss": 0.5906, + "step": 611 + }, + { + "epoch": 0.24622812311406156, + "grad_norm": 4.119688719772878, + "learning_rate": 4.68382982169048e-06, + "loss": 0.4954, + "step": 612 + }, + { + "epoch": 0.24663045664856165, + "grad_norm": 3.96691281411716, + "learning_rate": 4.6821186575516665e-06, + "loss": 0.6126, + "step": 613 + }, + { + "epoch": 0.24703279018306176, + "grad_norm": 4.760926273551105, + "learning_rate": 4.680403189672439e-06, + "loss": 0.6613, + "step": 614 + }, + { + "epoch": 0.24743512371756185, + "grad_norm": 4.4384316824035075, + "learning_rate": 4.678683421436173e-06, + "loss": 0.5687, + "step": 615 + }, + { + "epoch": 0.24783745725206197, + "grad_norm": 3.931664070509481, + "learning_rate": 4.676959356234726e-06, + "loss": 0.5857, + "step": 616 + }, + { + "epoch": 0.24823979078656205, + "grad_norm": 5.119336252733733, + "learning_rate": 4.6752309974684315e-06, + "loss": 0.6312, + "step": 617 + }, + { + "epoch": 0.24864212432106217, + "grad_norm": 4.229793646214086, + "learning_rate": 4.6734983485460895e-06, + "loss": 0.5093, + "step": 618 + }, + { + "epoch": 0.24904445785556226, + "grad_norm": 4.9794376894703944, + "learning_rate": 4.671761412884962e-06, + "loss": 0.7092, + "step": 619 + }, + { + "epoch": 0.24944679139006237, + "grad_norm": 4.36204998657697, + "learning_rate": 4.670020193910766e-06, + "loss": 0.559, + "step": 620 + }, + { + "epoch": 0.24984912492456246, + "grad_norm": 4.297208174823204, + "learning_rate": 4.668274695057666e-06, + "loss": 0.4965, + "step": 621 + }, + { + "epoch": 0.25025145845906255, + "grad_norm": 4.235510447453734, + "learning_rate": 4.666524919768267e-06, + "loss": 0.5434, + "step": 622 + }, + { + "epoch": 0.25065379199356264, + "grad_norm": 4.297045599708538, + "learning_rate": 4.66477087149361e-06, + "loss": 0.5559, + "step": 623 + }, + { + "epoch": 0.2510561255280628, + "grad_norm": 5.638228122643803, + "learning_rate": 4.663012553693161e-06, + "loss": 0.6393, + "step": 624 + }, + { + "epoch": 0.25145845906256287, + "grad_norm": 4.307369262356312, + "learning_rate": 4.661249969834809e-06, + "loss": 0.5516, + "step": 625 + }, + { + "epoch": 0.25186079259706295, + "grad_norm": 4.348008968073508, + "learning_rate": 4.659483123394855e-06, + "loss": 0.5858, + "step": 626 + }, + { + "epoch": 0.25226312613156304, + "grad_norm": 4.245775793874246, + "learning_rate": 4.657712017858011e-06, + "loss": 0.5985, + "step": 627 + }, + { + "epoch": 0.2526654596660632, + "grad_norm": 4.27241102613026, + "learning_rate": 4.6559366567173824e-06, + "loss": 0.5247, + "step": 628 + }, + { + "epoch": 0.2530677932005633, + "grad_norm": 3.9143192748019384, + "learning_rate": 4.6541570434744735e-06, + "loss": 0.4936, + "step": 629 + }, + { + "epoch": 0.25347012673506336, + "grad_norm": 4.550429412487192, + "learning_rate": 4.6523731816391725e-06, + "loss": 0.7499, + "step": 630 + }, + { + "epoch": 0.25387246026956345, + "grad_norm": 4.202697728155226, + "learning_rate": 4.650585074729747e-06, + "loss": 0.5184, + "step": 631 + }, + { + "epoch": 0.2542747938040636, + "grad_norm": 4.083559252407811, + "learning_rate": 4.648792726272838e-06, + "loss": 0.5242, + "step": 632 + }, + { + "epoch": 0.2546771273385637, + "grad_norm": 3.9861255665329267, + "learning_rate": 4.646996139803452e-06, + "loss": 0.5269, + "step": 633 + }, + { + "epoch": 0.25507946087306377, + "grad_norm": 4.559490860669659, + "learning_rate": 4.645195318864951e-06, + "loss": 0.6372, + "step": 634 + }, + { + "epoch": 0.25548179440756386, + "grad_norm": 4.394457869705605, + "learning_rate": 4.643390267009054e-06, + "loss": 0.5105, + "step": 635 + }, + { + "epoch": 0.25588412794206394, + "grad_norm": 5.018376167864149, + "learning_rate": 4.641580987795821e-06, + "loss": 0.6628, + "step": 636 + }, + { + "epoch": 0.2562864614765641, + "grad_norm": 3.7893948824739465, + "learning_rate": 4.639767484793648e-06, + "loss": 0.508, + "step": 637 + }, + { + "epoch": 0.2566887950110642, + "grad_norm": 4.596240270523842, + "learning_rate": 4.637949761579266e-06, + "loss": 0.6529, + "step": 638 + }, + { + "epoch": 0.25709112854556426, + "grad_norm": 5.140583964722019, + "learning_rate": 4.636127821737726e-06, + "loss": 0.6489, + "step": 639 + }, + { + "epoch": 0.25749346208006435, + "grad_norm": 3.8394758085160823, + "learning_rate": 4.634301668862397e-06, + "loss": 0.4736, + "step": 640 + }, + { + "epoch": 0.2578957956145645, + "grad_norm": 4.38356346084736, + "learning_rate": 4.632471306554955e-06, + "loss": 0.5353, + "step": 641 + }, + { + "epoch": 0.2582981291490646, + "grad_norm": 3.8889360472386185, + "learning_rate": 4.630636738425381e-06, + "loss": 0.5511, + "step": 642 + }, + { + "epoch": 0.25870046268356467, + "grad_norm": 4.817925477512709, + "learning_rate": 4.62879796809195e-06, + "loss": 0.6963, + "step": 643 + }, + { + "epoch": 0.25910279621806476, + "grad_norm": 4.742191877421797, + "learning_rate": 4.626954999181224e-06, + "loss": 0.7752, + "step": 644 + }, + { + "epoch": 0.2595051297525649, + "grad_norm": 4.245951623762571, + "learning_rate": 4.625107835328048e-06, + "loss": 0.5754, + "step": 645 + }, + { + "epoch": 0.259907463287065, + "grad_norm": 4.836575633325676, + "learning_rate": 4.623256480175537e-06, + "loss": 0.6639, + "step": 646 + }, + { + "epoch": 0.2603097968215651, + "grad_norm": 3.806176284035789, + "learning_rate": 4.621400937375075e-06, + "loss": 0.5292, + "step": 647 + }, + { + "epoch": 0.26071213035606516, + "grad_norm": 4.025115976057339, + "learning_rate": 4.619541210586307e-06, + "loss": 0.4815, + "step": 648 + }, + { + "epoch": 0.26111446389056525, + "grad_norm": 4.1347392050793825, + "learning_rate": 4.617677303477124e-06, + "loss": 0.5681, + "step": 649 + }, + { + "epoch": 0.2615167974250654, + "grad_norm": 4.371554358928095, + "learning_rate": 4.615809219723667e-06, + "loss": 0.592, + "step": 650 + }, + { + "epoch": 0.2619191309595655, + "grad_norm": 4.976927834726334, + "learning_rate": 4.613936963010314e-06, + "loss": 0.7898, + "step": 651 + }, + { + "epoch": 0.26232146449406557, + "grad_norm": 4.80195477309382, + "learning_rate": 4.612060537029671e-06, + "loss": 0.6565, + "step": 652 + }, + { + "epoch": 0.26272379802856566, + "grad_norm": 4.2203256747490805, + "learning_rate": 4.610179945482568e-06, + "loss": 0.5597, + "step": 653 + }, + { + "epoch": 0.2631261315630658, + "grad_norm": 4.3119037789418035, + "learning_rate": 4.608295192078051e-06, + "loss": 0.5773, + "step": 654 + }, + { + "epoch": 0.2635284650975659, + "grad_norm": 3.938767904065124, + "learning_rate": 4.606406280533373e-06, + "loss": 0.4691, + "step": 655 + }, + { + "epoch": 0.263930798632066, + "grad_norm": 3.873313542196617, + "learning_rate": 4.6045132145739914e-06, + "loss": 0.5446, + "step": 656 + }, + { + "epoch": 0.26433313216656606, + "grad_norm": 4.720535373828625, + "learning_rate": 4.602615997933552e-06, + "loss": 0.6451, + "step": 657 + }, + { + "epoch": 0.2647354657010662, + "grad_norm": 5.319338434241136, + "learning_rate": 4.600714634353893e-06, + "loss": 0.6786, + "step": 658 + }, + { + "epoch": 0.2651377992355663, + "grad_norm": 3.986441538874652, + "learning_rate": 4.598809127585026e-06, + "loss": 0.5815, + "step": 659 + }, + { + "epoch": 0.2655401327700664, + "grad_norm": 4.190229942704305, + "learning_rate": 4.596899481385137e-06, + "loss": 0.5933, + "step": 660 + }, + { + "epoch": 0.26594246630456647, + "grad_norm": 4.211284711325467, + "learning_rate": 4.5949856995205745e-06, + "loss": 0.6062, + "step": 661 + }, + { + "epoch": 0.26634479983906656, + "grad_norm": 5.022372015405014, + "learning_rate": 4.593067785765846e-06, + "loss": 0.6632, + "step": 662 + }, + { + "epoch": 0.2667471333735667, + "grad_norm": 3.9365356585193148, + "learning_rate": 4.5911457439036075e-06, + "loss": 0.5324, + "step": 663 + }, + { + "epoch": 0.2671494669080668, + "grad_norm": 4.891303145874928, + "learning_rate": 4.589219577724654e-06, + "loss": 0.7504, + "step": 664 + }, + { + "epoch": 0.2675518004425669, + "grad_norm": 3.855940902006332, + "learning_rate": 4.5872892910279185e-06, + "loss": 0.5476, + "step": 665 + }, + { + "epoch": 0.26795413397706697, + "grad_norm": 4.37750980680622, + "learning_rate": 4.58535488762046e-06, + "loss": 0.5902, + "step": 666 + }, + { + "epoch": 0.2683564675115671, + "grad_norm": 4.450877429194335, + "learning_rate": 4.583416371317454e-06, + "loss": 0.6345, + "step": 667 + }, + { + "epoch": 0.2687588010460672, + "grad_norm": 4.162519879691083, + "learning_rate": 4.581473745942191e-06, + "loss": 0.444, + "step": 668 + }, + { + "epoch": 0.2691611345805673, + "grad_norm": 4.2236499578643025, + "learning_rate": 4.579527015326065e-06, + "loss": 0.492, + "step": 669 + }, + { + "epoch": 0.2695634681150674, + "grad_norm": 4.482147273633291, + "learning_rate": 4.5775761833085645e-06, + "loss": 0.5648, + "step": 670 + }, + { + "epoch": 0.2699658016495675, + "grad_norm": 4.760060597047131, + "learning_rate": 4.57562125373727e-06, + "loss": 0.6674, + "step": 671 + }, + { + "epoch": 0.2703681351840676, + "grad_norm": 4.362389446472747, + "learning_rate": 4.573662230467844e-06, + "loss": 0.577, + "step": 672 + }, + { + "epoch": 0.2707704687185677, + "grad_norm": 3.8159558241349987, + "learning_rate": 4.5716991173640165e-06, + "loss": 0.5185, + "step": 673 + }, + { + "epoch": 0.2711728022530678, + "grad_norm": 4.005051687655319, + "learning_rate": 4.5697319182975944e-06, + "loss": 0.5922, + "step": 674 + }, + { + "epoch": 0.27157513578756787, + "grad_norm": 4.667139467518304, + "learning_rate": 4.567760637148432e-06, + "loss": 0.6032, + "step": 675 + }, + { + "epoch": 0.271977469322068, + "grad_norm": 3.9606776473436405, + "learning_rate": 4.5657852778044435e-06, + "loss": 0.5174, + "step": 676 + }, + { + "epoch": 0.2723798028565681, + "grad_norm": 4.322780347285095, + "learning_rate": 4.5638058441615815e-06, + "loss": 0.6089, + "step": 677 + }, + { + "epoch": 0.2727821363910682, + "grad_norm": 4.433317399280547, + "learning_rate": 4.561822340123836e-06, + "loss": 0.6047, + "step": 678 + }, + { + "epoch": 0.2731844699255683, + "grad_norm": 4.552805703827102, + "learning_rate": 4.559834769603224e-06, + "loss": 0.6281, + "step": 679 + }, + { + "epoch": 0.2735868034600684, + "grad_norm": 5.128182512733822, + "learning_rate": 4.557843136519784e-06, + "loss": 0.6697, + "step": 680 + }, + { + "epoch": 0.2739891369945685, + "grad_norm": 4.063920951512369, + "learning_rate": 4.555847444801565e-06, + "loss": 0.5885, + "step": 681 + }, + { + "epoch": 0.2743914705290686, + "grad_norm": 4.908841427141714, + "learning_rate": 4.5538476983846245e-06, + "loss": 0.6771, + "step": 682 + }, + { + "epoch": 0.2747938040635687, + "grad_norm": 4.405724720337232, + "learning_rate": 4.551843901213012e-06, + "loss": 0.6479, + "step": 683 + }, + { + "epoch": 0.2751961375980688, + "grad_norm": 3.945129587052982, + "learning_rate": 4.549836057238769e-06, + "loss": 0.6039, + "step": 684 + }, + { + "epoch": 0.2755984711325689, + "grad_norm": 4.96376428039534, + "learning_rate": 4.547824170421921e-06, + "loss": 0.7191, + "step": 685 + }, + { + "epoch": 0.276000804667069, + "grad_norm": 4.378595779287015, + "learning_rate": 4.54580824473046e-06, + "loss": 0.5549, + "step": 686 + }, + { + "epoch": 0.2764031382015691, + "grad_norm": 4.262426880188478, + "learning_rate": 4.54378828414035e-06, + "loss": 0.5024, + "step": 687 + }, + { + "epoch": 0.2768054717360692, + "grad_norm": 4.426695285396216, + "learning_rate": 4.541764292635512e-06, + "loss": 0.7052, + "step": 688 + }, + { + "epoch": 0.2772078052705693, + "grad_norm": 4.473404410086587, + "learning_rate": 4.5397362742078145e-06, + "loss": 0.5557, + "step": 689 + }, + { + "epoch": 0.2776101388050694, + "grad_norm": 4.052844689460026, + "learning_rate": 4.537704232857069e-06, + "loss": 0.4541, + "step": 690 + }, + { + "epoch": 0.2780124723395695, + "grad_norm": 3.89050386632389, + "learning_rate": 4.5356681725910255e-06, + "loss": 0.549, + "step": 691 + }, + { + "epoch": 0.2784148058740696, + "grad_norm": 4.349538143222671, + "learning_rate": 4.5336280974253535e-06, + "loss": 0.6076, + "step": 692 + }, + { + "epoch": 0.2788171394085697, + "grad_norm": 4.361849160433631, + "learning_rate": 4.5315840113836454e-06, + "loss": 0.6524, + "step": 693 + }, + { + "epoch": 0.2792194729430698, + "grad_norm": 4.617234754063093, + "learning_rate": 4.529535918497403e-06, + "loss": 0.5616, + "step": 694 + }, + { + "epoch": 0.2796218064775699, + "grad_norm": 4.09023083083644, + "learning_rate": 4.5274838228060326e-06, + "loss": 0.5407, + "step": 695 + }, + { + "epoch": 0.28002414001207, + "grad_norm": 3.7157295921866718, + "learning_rate": 4.525427728356832e-06, + "loss": 0.5361, + "step": 696 + }, + { + "epoch": 0.28042647354657013, + "grad_norm": 4.457227461962157, + "learning_rate": 4.523367639204985e-06, + "loss": 0.5593, + "step": 697 + }, + { + "epoch": 0.2808288070810702, + "grad_norm": 4.348610370546784, + "learning_rate": 4.52130355941356e-06, + "loss": 0.5121, + "step": 698 + }, + { + "epoch": 0.2812311406155703, + "grad_norm": 3.9173059552199194, + "learning_rate": 4.519235493053491e-06, + "loss": 0.4927, + "step": 699 + }, + { + "epoch": 0.2816334741500704, + "grad_norm": 4.208431911956715, + "learning_rate": 4.517163444203575e-06, + "loss": 0.5504, + "step": 700 + }, + { + "epoch": 0.2820358076845705, + "grad_norm": 4.404532536673193, + "learning_rate": 4.515087416950464e-06, + "loss": 0.6057, + "step": 701 + }, + { + "epoch": 0.2824381412190706, + "grad_norm": 4.163697909062566, + "learning_rate": 4.513007415388659e-06, + "loss": 0.4682, + "step": 702 + }, + { + "epoch": 0.2828404747535707, + "grad_norm": 5.152927142247097, + "learning_rate": 4.510923443620494e-06, + "loss": 0.6601, + "step": 703 + }, + { + "epoch": 0.2832428082880708, + "grad_norm": 4.544361331270338, + "learning_rate": 4.50883550575614e-06, + "loss": 0.5968, + "step": 704 + }, + { + "epoch": 0.2836451418225709, + "grad_norm": 4.134394917706537, + "learning_rate": 4.5067436059135836e-06, + "loss": 0.6647, + "step": 705 + }, + { + "epoch": 0.28404747535707103, + "grad_norm": 4.734085611831448, + "learning_rate": 4.504647748218633e-06, + "loss": 0.601, + "step": 706 + }, + { + "epoch": 0.2844498088915711, + "grad_norm": 4.681267518422373, + "learning_rate": 4.502547936804894e-06, + "loss": 0.5539, + "step": 707 + }, + { + "epoch": 0.2848521424260712, + "grad_norm": 4.325088947732379, + "learning_rate": 4.500444175813776e-06, + "loss": 0.5872, + "step": 708 + }, + { + "epoch": 0.2852544759605713, + "grad_norm": 3.850534498836246, + "learning_rate": 4.4983364693944775e-06, + "loss": 0.5061, + "step": 709 + }, + { + "epoch": 0.28565680949507144, + "grad_norm": 4.373584617561936, + "learning_rate": 4.496224821703977e-06, + "loss": 0.6693, + "step": 710 + }, + { + "epoch": 0.2860591430295715, + "grad_norm": 4.377477636040013, + "learning_rate": 4.494109236907026e-06, + "loss": 0.5966, + "step": 711 + }, + { + "epoch": 0.2864614765640716, + "grad_norm": 4.271381874115667, + "learning_rate": 4.491989719176142e-06, + "loss": 0.6221, + "step": 712 + }, + { + "epoch": 0.2868638100985717, + "grad_norm": 3.7259627257148202, + "learning_rate": 4.489866272691599e-06, + "loss": 0.4313, + "step": 713 + }, + { + "epoch": 0.2872661436330718, + "grad_norm": 4.0336320558123155, + "learning_rate": 4.48773890164142e-06, + "loss": 0.5305, + "step": 714 + }, + { + "epoch": 0.28766847716757193, + "grad_norm": 4.585992205034134, + "learning_rate": 4.485607610221367e-06, + "loss": 0.7636, + "step": 715 + }, + { + "epoch": 0.288070810702072, + "grad_norm": 4.67820441036977, + "learning_rate": 4.4834724026349376e-06, + "loss": 0.5866, + "step": 716 + }, + { + "epoch": 0.2884731442365721, + "grad_norm": 5.012556667522281, + "learning_rate": 4.4813332830933484e-06, + "loss": 0.6498, + "step": 717 + }, + { + "epoch": 0.2888754777710722, + "grad_norm": 4.19440277086636, + "learning_rate": 4.4791902558155345e-06, + "loss": 0.5434, + "step": 718 + }, + { + "epoch": 0.28927781130557234, + "grad_norm": 4.330716534353544, + "learning_rate": 4.477043325028137e-06, + "loss": 0.5507, + "step": 719 + }, + { + "epoch": 0.28968014484007243, + "grad_norm": 4.097976140286037, + "learning_rate": 4.474892494965495e-06, + "loss": 0.4591, + "step": 720 + }, + { + "epoch": 0.2900824783745725, + "grad_norm": 4.132010732456553, + "learning_rate": 4.472737769869642e-06, + "loss": 0.5217, + "step": 721 + }, + { + "epoch": 0.2904848119090726, + "grad_norm": 3.7815121295908343, + "learning_rate": 4.470579153990288e-06, + "loss": 0.4932, + "step": 722 + }, + { + "epoch": 0.29088714544357275, + "grad_norm": 3.724905186650606, + "learning_rate": 4.468416651584822e-06, + "loss": 0.5448, + "step": 723 + }, + { + "epoch": 0.29128947897807284, + "grad_norm": 3.82443773314746, + "learning_rate": 4.4662502669182935e-06, + "loss": 0.5298, + "step": 724 + }, + { + "epoch": 0.2916918125125729, + "grad_norm": 4.343371987708737, + "learning_rate": 4.464080004263411e-06, + "loss": 0.606, + "step": 725 + }, + { + "epoch": 0.292094146047073, + "grad_norm": 4.009570342579401, + "learning_rate": 4.461905867900535e-06, + "loss": 0.5916, + "step": 726 + }, + { + "epoch": 0.2924964795815731, + "grad_norm": 4.148496022814315, + "learning_rate": 4.459727862117658e-06, + "loss": 0.5155, + "step": 727 + }, + { + "epoch": 0.29289881311607324, + "grad_norm": 4.568599931304069, + "learning_rate": 4.457545991210412e-06, + "loss": 0.555, + "step": 728 + }, + { + "epoch": 0.29330114665057333, + "grad_norm": 4.026645287125634, + "learning_rate": 4.455360259482047e-06, + "loss": 0.5265, + "step": 729 + }, + { + "epoch": 0.2937034801850734, + "grad_norm": 4.4466778657917265, + "learning_rate": 4.4531706712434305e-06, + "loss": 0.6318, + "step": 730 + }, + { + "epoch": 0.2941058137195735, + "grad_norm": 4.444852619715758, + "learning_rate": 4.450977230813035e-06, + "loss": 0.671, + "step": 731 + }, + { + "epoch": 0.29450814725407365, + "grad_norm": 5.136550532204147, + "learning_rate": 4.44877994251693e-06, + "loss": 0.611, + "step": 732 + }, + { + "epoch": 0.29491048078857374, + "grad_norm": 4.472272090985854, + "learning_rate": 4.446578810688774e-06, + "loss": 0.595, + "step": 733 + }, + { + "epoch": 0.2953128143230738, + "grad_norm": 4.755288406312762, + "learning_rate": 4.444373839669808e-06, + "loss": 0.801, + "step": 734 + }, + { + "epoch": 0.2957151478575739, + "grad_norm": 4.964978523542175, + "learning_rate": 4.442165033808843e-06, + "loss": 0.6712, + "step": 735 + }, + { + "epoch": 0.29611748139207406, + "grad_norm": 4.590633696444609, + "learning_rate": 4.439952397462254e-06, + "loss": 0.6214, + "step": 736 + }, + { + "epoch": 0.29651981492657414, + "grad_norm": 4.05753322012328, + "learning_rate": 4.43773593499397e-06, + "loss": 0.4477, + "step": 737 + }, + { + "epoch": 0.29692214846107423, + "grad_norm": 4.354711360779599, + "learning_rate": 4.435515650775468e-06, + "loss": 0.5618, + "step": 738 + }, + { + "epoch": 0.2973244819955743, + "grad_norm": 4.543128550216277, + "learning_rate": 4.433291549185761e-06, + "loss": 0.4807, + "step": 739 + }, + { + "epoch": 0.2977268155300744, + "grad_norm": 3.9444140759236737, + "learning_rate": 4.431063634611392e-06, + "loss": 0.5393, + "step": 740 + }, + { + "epoch": 0.29812914906457455, + "grad_norm": 4.516456335270593, + "learning_rate": 4.428831911446422e-06, + "loss": 0.5839, + "step": 741 + }, + { + "epoch": 0.29853148259907464, + "grad_norm": 4.4607761036583815, + "learning_rate": 4.426596384092426e-06, + "loss": 0.6538, + "step": 742 + }, + { + "epoch": 0.2989338161335747, + "grad_norm": 6.196021389574422, + "learning_rate": 4.424357056958483e-06, + "loss": 0.6484, + "step": 743 + }, + { + "epoch": 0.2993361496680748, + "grad_norm": 4.251042887568661, + "learning_rate": 4.422113934461161e-06, + "loss": 0.5523, + "step": 744 + }, + { + "epoch": 0.29973848320257496, + "grad_norm": 4.697763472346534, + "learning_rate": 4.4198670210245186e-06, + "loss": 0.6053, + "step": 745 + }, + { + "epoch": 0.30014081673707504, + "grad_norm": 4.665389980976664, + "learning_rate": 4.41761632108009e-06, + "loss": 0.6618, + "step": 746 + }, + { + "epoch": 0.30054315027157513, + "grad_norm": 4.2980125151698525, + "learning_rate": 4.415361839066874e-06, + "loss": 0.4863, + "step": 747 + }, + { + "epoch": 0.3009454838060752, + "grad_norm": 4.25422933800907, + "learning_rate": 4.413103579431335e-06, + "loss": 0.6285, + "step": 748 + }, + { + "epoch": 0.30134781734057536, + "grad_norm": 4.408647829220997, + "learning_rate": 4.410841546627383e-06, + "loss": 0.6214, + "step": 749 + }, + { + "epoch": 0.30175015087507545, + "grad_norm": 3.5964308822983626, + "learning_rate": 4.408575745116371e-06, + "loss": 0.4709, + "step": 750 + }, + { + "epoch": 0.30215248440957554, + "grad_norm": 4.177190728366978, + "learning_rate": 4.406306179367086e-06, + "loss": 0.6584, + "step": 751 + }, + { + "epoch": 0.3025548179440756, + "grad_norm": 3.9436659867851, + "learning_rate": 4.404032853855736e-06, + "loss": 0.5804, + "step": 752 + }, + { + "epoch": 0.3029571514785757, + "grad_norm": 4.5139405502647465, + "learning_rate": 4.40175577306595e-06, + "loss": 0.6131, + "step": 753 + }, + { + "epoch": 0.30335948501307586, + "grad_norm": 4.656972033870911, + "learning_rate": 4.399474941488756e-06, + "loss": 0.5378, + "step": 754 + }, + { + "epoch": 0.30376181854757595, + "grad_norm": 4.2754391925078785, + "learning_rate": 4.397190363622588e-06, + "loss": 0.5117, + "step": 755 + }, + { + "epoch": 0.30416415208207603, + "grad_norm": 4.057091286795055, + "learning_rate": 4.3949020439732594e-06, + "loss": 0.4969, + "step": 756 + }, + { + "epoch": 0.3045664856165761, + "grad_norm": 3.914038182134245, + "learning_rate": 4.392609987053972e-06, + "loss": 0.4484, + "step": 757 + }, + { + "epoch": 0.30496881915107626, + "grad_norm": 4.364733148673897, + "learning_rate": 4.390314197385292e-06, + "loss": 0.662, + "step": 758 + }, + { + "epoch": 0.30537115268557635, + "grad_norm": 4.866819672714999, + "learning_rate": 4.388014679495154e-06, + "loss": 0.6832, + "step": 759 + }, + { + "epoch": 0.30577348622007644, + "grad_norm": 4.563239673265389, + "learning_rate": 4.385711437918839e-06, + "loss": 0.633, + "step": 760 + }, + { + "epoch": 0.30617581975457653, + "grad_norm": 3.952293727810083, + "learning_rate": 4.383404477198976e-06, + "loss": 0.4319, + "step": 761 + }, + { + "epoch": 0.30657815328907667, + "grad_norm": 3.8982970475676866, + "learning_rate": 4.38109380188553e-06, + "loss": 0.5412, + "step": 762 + }, + { + "epoch": 0.30698048682357676, + "grad_norm": 4.767355411844817, + "learning_rate": 4.3787794165357875e-06, + "loss": 0.6688, + "step": 763 + }, + { + "epoch": 0.30738282035807685, + "grad_norm": 5.490594878274815, + "learning_rate": 4.3764613257143585e-06, + "loss": 0.8315, + "step": 764 + }, + { + "epoch": 0.30778515389257693, + "grad_norm": 4.4753569929038965, + "learning_rate": 4.374139533993157e-06, + "loss": 0.5577, + "step": 765 + }, + { + "epoch": 0.308187487427077, + "grad_norm": 4.234373998218113, + "learning_rate": 4.371814045951396e-06, + "loss": 0.6195, + "step": 766 + }, + { + "epoch": 0.30858982096157717, + "grad_norm": 3.8789979804624277, + "learning_rate": 4.369484866175581e-06, + "loss": 0.5812, + "step": 767 + }, + { + "epoch": 0.30899215449607725, + "grad_norm": 4.3774657256617635, + "learning_rate": 4.367151999259498e-06, + "loss": 0.5334, + "step": 768 + }, + { + "epoch": 0.30939448803057734, + "grad_norm": 4.413468965671534, + "learning_rate": 4.364815449804204e-06, + "loss": 0.6043, + "step": 769 + }, + { + "epoch": 0.30979682156507743, + "grad_norm": 4.977248443532888, + "learning_rate": 4.362475222418019e-06, + "loss": 0.7427, + "step": 770 + }, + { + "epoch": 0.3101991550995776, + "grad_norm": 4.449181903975735, + "learning_rate": 4.360131321716518e-06, + "loss": 0.5336, + "step": 771 + }, + { + "epoch": 0.31060148863407766, + "grad_norm": 4.203059807958007, + "learning_rate": 4.357783752322522e-06, + "loss": 0.5362, + "step": 772 + }, + { + "epoch": 0.31100382216857775, + "grad_norm": 3.8919997316821004, + "learning_rate": 4.355432518866084e-06, + "loss": 0.5202, + "step": 773 + }, + { + "epoch": 0.31140615570307784, + "grad_norm": 4.125644261476737, + "learning_rate": 4.353077625984484e-06, + "loss": 0.5136, + "step": 774 + }, + { + "epoch": 0.311808489237578, + "grad_norm": 3.7230845852814687, + "learning_rate": 4.3507190783222245e-06, + "loss": 0.4731, + "step": 775 + }, + { + "epoch": 0.31221082277207807, + "grad_norm": 4.619811557245368, + "learning_rate": 4.348356880531011e-06, + "loss": 0.698, + "step": 776 + }, + { + "epoch": 0.31261315630657815, + "grad_norm": 4.064918048601741, + "learning_rate": 4.345991037269748e-06, + "loss": 0.5594, + "step": 777 + }, + { + "epoch": 0.31301548984107824, + "grad_norm": 4.28030910141664, + "learning_rate": 4.343621553204533e-06, + "loss": 0.6272, + "step": 778 + }, + { + "epoch": 0.31341782337557833, + "grad_norm": 4.3485146730819, + "learning_rate": 4.341248433008645e-06, + "loss": 0.6078, + "step": 779 + }, + { + "epoch": 0.3138201569100785, + "grad_norm": 4.211683329853448, + "learning_rate": 4.338871681362528e-06, + "loss": 0.5247, + "step": 780 + }, + { + "epoch": 0.31422249044457856, + "grad_norm": 3.8752448583148453, + "learning_rate": 4.3364913029537946e-06, + "loss": 0.5274, + "step": 781 + }, + { + "epoch": 0.31462482397907865, + "grad_norm": 3.999853155987491, + "learning_rate": 4.334107302477208e-06, + "loss": 0.5421, + "step": 782 + }, + { + "epoch": 0.31502715751357874, + "grad_norm": 4.961501923709214, + "learning_rate": 4.331719684634676e-06, + "loss": 0.7151, + "step": 783 + }, + { + "epoch": 0.3154294910480789, + "grad_norm": 4.148580969978077, + "learning_rate": 4.3293284541352384e-06, + "loss": 0.5084, + "step": 784 + }, + { + "epoch": 0.31583182458257897, + "grad_norm": 3.623114769478068, + "learning_rate": 4.326933615695064e-06, + "loss": 0.4912, + "step": 785 + }, + { + "epoch": 0.31623415811707906, + "grad_norm": 4.179951168808134, + "learning_rate": 4.324535174037433e-06, + "loss": 0.4943, + "step": 786 + }, + { + "epoch": 0.31663649165157914, + "grad_norm": 4.206035486231768, + "learning_rate": 4.322133133892737e-06, + "loss": 0.6175, + "step": 787 + }, + { + "epoch": 0.3170388251860793, + "grad_norm": 4.029670988995613, + "learning_rate": 4.31972749999846e-06, + "loss": 0.5329, + "step": 788 + }, + { + "epoch": 0.3174411587205794, + "grad_norm": 4.024639871547482, + "learning_rate": 4.317318277099178e-06, + "loss": 0.5121, + "step": 789 + }, + { + "epoch": 0.31784349225507946, + "grad_norm": 4.04730827392964, + "learning_rate": 4.314905469946542e-06, + "loss": 0.4979, + "step": 790 + }, + { + "epoch": 0.31824582578957955, + "grad_norm": 4.440133410808506, + "learning_rate": 4.312489083299275e-06, + "loss": 0.6333, + "step": 791 + }, + { + "epoch": 0.31864815932407964, + "grad_norm": 4.079663513764819, + "learning_rate": 4.310069121923158e-06, + "loss": 0.5987, + "step": 792 + }, + { + "epoch": 0.3190504928585798, + "grad_norm": 4.17692958034715, + "learning_rate": 4.307645590591024e-06, + "loss": 0.4798, + "step": 793 + }, + { + "epoch": 0.31945282639307987, + "grad_norm": 4.417314019145798, + "learning_rate": 4.305218494082744e-06, + "loss": 0.4832, + "step": 794 + }, + { + "epoch": 0.31985515992757996, + "grad_norm": 4.920563992053252, + "learning_rate": 4.302787837185224e-06, + "loss": 0.6288, + "step": 795 + }, + { + "epoch": 0.32025749346208005, + "grad_norm": 4.067969775601857, + "learning_rate": 4.300353624692389e-06, + "loss": 0.52, + "step": 796 + }, + { + "epoch": 0.3206598269965802, + "grad_norm": 4.232889302389981, + "learning_rate": 4.2979158614051805e-06, + "loss": 0.6205, + "step": 797 + }, + { + "epoch": 0.3210621605310803, + "grad_norm": 4.335177482388414, + "learning_rate": 4.295474552131538e-06, + "loss": 0.6153, + "step": 798 + }, + { + "epoch": 0.32146449406558036, + "grad_norm": 4.698659170570413, + "learning_rate": 4.2930297016863985e-06, + "loss": 0.6404, + "step": 799 + }, + { + "epoch": 0.32186682760008045, + "grad_norm": 3.931383036386479, + "learning_rate": 4.2905813148916816e-06, + "loss": 0.599, + "step": 800 + }, + { + "epoch": 0.32226916113458054, + "grad_norm": 5.048444025824699, + "learning_rate": 4.288129396576284e-06, + "loss": 0.7259, + "step": 801 + }, + { + "epoch": 0.3226714946690807, + "grad_norm": 4.851397090992265, + "learning_rate": 4.285673951576062e-06, + "loss": 0.642, + "step": 802 + }, + { + "epoch": 0.32307382820358077, + "grad_norm": 4.28645665558046, + "learning_rate": 4.283214984733833e-06, + "loss": 0.5315, + "step": 803 + }, + { + "epoch": 0.32347616173808086, + "grad_norm": 4.265366711136778, + "learning_rate": 4.28075250089936e-06, + "loss": 0.5529, + "step": 804 + }, + { + "epoch": 0.32387849527258095, + "grad_norm": 3.848835311590686, + "learning_rate": 4.278286504929338e-06, + "loss": 0.5323, + "step": 805 + }, + { + "epoch": 0.3242808288070811, + "grad_norm": 4.627843495385828, + "learning_rate": 4.2758170016873934e-06, + "loss": 0.5561, + "step": 806 + }, + { + "epoch": 0.3246831623415812, + "grad_norm": 4.363493591898399, + "learning_rate": 4.273343996044068e-06, + "loss": 0.4805, + "step": 807 + }, + { + "epoch": 0.32508549587608127, + "grad_norm": 3.9805290188441043, + "learning_rate": 4.270867492876812e-06, + "loss": 0.5204, + "step": 808 + }, + { + "epoch": 0.32548782941058135, + "grad_norm": 3.9644340799634588, + "learning_rate": 4.268387497069974e-06, + "loss": 0.4642, + "step": 809 + }, + { + "epoch": 0.3258901629450815, + "grad_norm": 5.269711583475643, + "learning_rate": 4.265904013514788e-06, + "loss": 0.6183, + "step": 810 + }, + { + "epoch": 0.3262924964795816, + "grad_norm": 4.6945618708498165, + "learning_rate": 4.263417047109371e-06, + "loss": 0.6567, + "step": 811 + }, + { + "epoch": 0.32669483001408167, + "grad_norm": 4.413878483629632, + "learning_rate": 4.260926602758707e-06, + "loss": 0.5518, + "step": 812 + }, + { + "epoch": 0.32709716354858176, + "grad_norm": 4.355371114130468, + "learning_rate": 4.258432685374641e-06, + "loss": 0.478, + "step": 813 + }, + { + "epoch": 0.32749949708308185, + "grad_norm": 4.625144876838143, + "learning_rate": 4.255935299875864e-06, + "loss": 0.5918, + "step": 814 + }, + { + "epoch": 0.327901830617582, + "grad_norm": 3.8722185488332994, + "learning_rate": 4.253434451187911e-06, + "loss": 0.4698, + "step": 815 + }, + { + "epoch": 0.3283041641520821, + "grad_norm": 3.8155166378719403, + "learning_rate": 4.250930144243147e-06, + "loss": 0.43, + "step": 816 + }, + { + "epoch": 0.32870649768658217, + "grad_norm": 4.373317134651141, + "learning_rate": 4.248422383980756e-06, + "loss": 0.581, + "step": 817 + }, + { + "epoch": 0.32910883122108225, + "grad_norm": 4.316760308023842, + "learning_rate": 4.245911175346733e-06, + "loss": 0.525, + "step": 818 + }, + { + "epoch": 0.3295111647555824, + "grad_norm": 4.1806833930239895, + "learning_rate": 4.243396523293875e-06, + "loss": 0.6027, + "step": 819 + }, + { + "epoch": 0.3299134982900825, + "grad_norm": 4.011770765526375, + "learning_rate": 4.240878432781769e-06, + "loss": 0.5642, + "step": 820 + }, + { + "epoch": 0.3303158318245826, + "grad_norm": 4.106735076414742, + "learning_rate": 4.238356908776786e-06, + "loss": 0.5576, + "step": 821 + }, + { + "epoch": 0.33071816535908266, + "grad_norm": 4.7608485825878955, + "learning_rate": 4.235831956252069e-06, + "loss": 0.7088, + "step": 822 + }, + { + "epoch": 0.3311204988935828, + "grad_norm": 4.8786607685170305, + "learning_rate": 4.2333035801875175e-06, + "loss": 0.6513, + "step": 823 + }, + { + "epoch": 0.3315228324280829, + "grad_norm": 4.288963783688971, + "learning_rate": 4.230771785569791e-06, + "loss": 0.5073, + "step": 824 + }, + { + "epoch": 0.331925165962583, + "grad_norm": 4.014903400829248, + "learning_rate": 4.228236577392285e-06, + "loss": 0.5901, + "step": 825 + }, + { + "epoch": 0.33232749949708307, + "grad_norm": 4.154812074980563, + "learning_rate": 4.225697960655131e-06, + "loss": 0.5568, + "step": 826 + }, + { + "epoch": 0.33272983303158316, + "grad_norm": 3.827755329803438, + "learning_rate": 4.223155940365181e-06, + "loss": 0.4738, + "step": 827 + }, + { + "epoch": 0.3331321665660833, + "grad_norm": 3.856077516315061, + "learning_rate": 4.220610521536002e-06, + "loss": 0.4963, + "step": 828 + }, + { + "epoch": 0.3335345001005834, + "grad_norm": 4.615466783911023, + "learning_rate": 4.218061709187862e-06, + "loss": 0.5656, + "step": 829 + }, + { + "epoch": 0.3339368336350835, + "grad_norm": 4.176966772191371, + "learning_rate": 4.215509508347723e-06, + "loss": 0.5398, + "step": 830 + }, + { + "epoch": 0.33433916716958356, + "grad_norm": 4.49327690032652, + "learning_rate": 4.212953924049229e-06, + "loss": 0.6772, + "step": 831 + }, + { + "epoch": 0.3347415007040837, + "grad_norm": 4.067266904294674, + "learning_rate": 4.210394961332698e-06, + "loss": 0.5311, + "step": 832 + }, + { + "epoch": 0.3351438342385838, + "grad_norm": 3.7375221543142128, + "learning_rate": 4.207832625245112e-06, + "loss": 0.4573, + "step": 833 + }, + { + "epoch": 0.3355461677730839, + "grad_norm": 4.056435688935082, + "learning_rate": 4.205266920840103e-06, + "loss": 0.543, + "step": 834 + }, + { + "epoch": 0.33594850130758397, + "grad_norm": 3.9887060748088254, + "learning_rate": 4.202697853177951e-06, + "loss": 0.4613, + "step": 835 + }, + { + "epoch": 0.3363508348420841, + "grad_norm": 3.5896338953258473, + "learning_rate": 4.2001254273255646e-06, + "loss": 0.4031, + "step": 836 + }, + { + "epoch": 0.3367531683765842, + "grad_norm": 4.094188015061212, + "learning_rate": 4.197549648356478e-06, + "loss": 0.5933, + "step": 837 + }, + { + "epoch": 0.3371555019110843, + "grad_norm": 3.6605198066957216, + "learning_rate": 4.19497052135084e-06, + "loss": 0.417, + "step": 838 + }, + { + "epoch": 0.3375578354455844, + "grad_norm": 5.538220960178179, + "learning_rate": 4.192388051395398e-06, + "loss": 0.7984, + "step": 839 + }, + { + "epoch": 0.33796016898008446, + "grad_norm": 4.543985654558515, + "learning_rate": 4.1898022435835e-06, + "loss": 0.5526, + "step": 840 + }, + { + "epoch": 0.3383625025145846, + "grad_norm": 3.719487495093868, + "learning_rate": 4.187213103015069e-06, + "loss": 0.4399, + "step": 841 + }, + { + "epoch": 0.3387648360490847, + "grad_norm": 3.958292027050775, + "learning_rate": 4.184620634796608e-06, + "loss": 0.4891, + "step": 842 + }, + { + "epoch": 0.3391671695835848, + "grad_norm": 4.219443448956219, + "learning_rate": 4.182024844041177e-06, + "loss": 0.6417, + "step": 843 + }, + { + "epoch": 0.33956950311808487, + "grad_norm": 4.637518704467857, + "learning_rate": 4.179425735868395e-06, + "loss": 0.5912, + "step": 844 + }, + { + "epoch": 0.339971836652585, + "grad_norm": 4.725084492132898, + "learning_rate": 4.176823315404419e-06, + "loss": 0.705, + "step": 845 + }, + { + "epoch": 0.3403741701870851, + "grad_norm": 4.081067015807228, + "learning_rate": 4.17421758778194e-06, + "loss": 0.5006, + "step": 846 + }, + { + "epoch": 0.3407765037215852, + "grad_norm": 4.431343567792434, + "learning_rate": 4.1716085581401746e-06, + "loss": 0.5081, + "step": 847 + }, + { + "epoch": 0.3411788372560853, + "grad_norm": 4.173696495740008, + "learning_rate": 4.1689962316248475e-06, + "loss": 0.497, + "step": 848 + }, + { + "epoch": 0.3415811707905854, + "grad_norm": 4.125681881515539, + "learning_rate": 4.166380613388189e-06, + "loss": 0.5498, + "step": 849 + }, + { + "epoch": 0.3419835043250855, + "grad_norm": 4.333217037647524, + "learning_rate": 4.163761708588919e-06, + "loss": 0.5597, + "step": 850 + }, + { + "epoch": 0.3423858378595856, + "grad_norm": 4.191848923537687, + "learning_rate": 4.161139522392243e-06, + "loss": 0.7259, + "step": 851 + }, + { + "epoch": 0.3427881713940857, + "grad_norm": 3.1772920188403164, + "learning_rate": 4.158514059969834e-06, + "loss": 0.4225, + "step": 852 + }, + { + "epoch": 0.34319050492858577, + "grad_norm": 4.166291911511059, + "learning_rate": 4.15588532649983e-06, + "loss": 0.6061, + "step": 853 + }, + { + "epoch": 0.3435928384630859, + "grad_norm": 3.5991395036430447, + "learning_rate": 4.1532533271668175e-06, + "loss": 0.461, + "step": 854 + }, + { + "epoch": 0.343995171997586, + "grad_norm": 4.29116994161306, + "learning_rate": 4.150618067161828e-06, + "loss": 0.567, + "step": 855 + }, + { + "epoch": 0.3443975055320861, + "grad_norm": 4.403897532608296, + "learning_rate": 4.14797955168232e-06, + "loss": 0.6118, + "step": 856 + }, + { + "epoch": 0.3447998390665862, + "grad_norm": 4.145689255622535, + "learning_rate": 4.145337785932174e-06, + "loss": 0.5709, + "step": 857 + }, + { + "epoch": 0.3452021726010863, + "grad_norm": 4.540600097537597, + "learning_rate": 4.142692775121684e-06, + "loss": 0.5506, + "step": 858 + }, + { + "epoch": 0.3456045061355864, + "grad_norm": 4.730037786980354, + "learning_rate": 4.14004452446754e-06, + "loss": 0.6088, + "step": 859 + }, + { + "epoch": 0.3460068396700865, + "grad_norm": 3.9409862740254487, + "learning_rate": 4.137393039192822e-06, + "loss": 0.5775, + "step": 860 + }, + { + "epoch": 0.3464091732045866, + "grad_norm": 4.113320460512886, + "learning_rate": 4.1347383245269935e-06, + "loss": 0.5695, + "step": 861 + }, + { + "epoch": 0.34681150673908673, + "grad_norm": 4.6445735426137364, + "learning_rate": 4.1320803857058835e-06, + "loss": 0.6158, + "step": 862 + }, + { + "epoch": 0.3472138402735868, + "grad_norm": 4.634455536740089, + "learning_rate": 4.129419227971681e-06, + "loss": 0.6799, + "step": 863 + }, + { + "epoch": 0.3476161738080869, + "grad_norm": 4.767706536948578, + "learning_rate": 4.1267548565729235e-06, + "loss": 0.5473, + "step": 864 + }, + { + "epoch": 0.348018507342587, + "grad_norm": 4.4478746327105405, + "learning_rate": 4.124087276764488e-06, + "loss": 0.6564, + "step": 865 + }, + { + "epoch": 0.3484208408770871, + "grad_norm": 5.1282545461378195, + "learning_rate": 4.121416493807577e-06, + "loss": 0.617, + "step": 866 + }, + { + "epoch": 0.3488231744115872, + "grad_norm": 3.850370414902996, + "learning_rate": 4.118742512969713e-06, + "loss": 0.5775, + "step": 867 + }, + { + "epoch": 0.3492255079460873, + "grad_norm": 4.66805214245803, + "learning_rate": 4.116065339524724e-06, + "loss": 0.6724, + "step": 868 + }, + { + "epoch": 0.3496278414805874, + "grad_norm": 4.7651508619199765, + "learning_rate": 4.113384978752734e-06, + "loss": 0.6151, + "step": 869 + }, + { + "epoch": 0.3500301750150875, + "grad_norm": 4.207593613220341, + "learning_rate": 4.110701435940157e-06, + "loss": 0.641, + "step": 870 + }, + { + "epoch": 0.35043250854958763, + "grad_norm": 4.3034195107515965, + "learning_rate": 4.108014716379679e-06, + "loss": 0.6542, + "step": 871 + }, + { + "epoch": 0.3508348420840877, + "grad_norm": 3.987719667248183, + "learning_rate": 4.105324825370251e-06, + "loss": 0.4996, + "step": 872 + }, + { + "epoch": 0.3512371756185878, + "grad_norm": 3.9071320373239824, + "learning_rate": 4.102631768217083e-06, + "loss": 0.5168, + "step": 873 + }, + { + "epoch": 0.3516395091530879, + "grad_norm": 4.2277706570506295, + "learning_rate": 4.099935550231626e-06, + "loss": 0.5379, + "step": 874 + }, + { + "epoch": 0.35204184268758804, + "grad_norm": 3.9154962483786373, + "learning_rate": 4.097236176731567e-06, + "loss": 0.5523, + "step": 875 + }, + { + "epoch": 0.3524441762220881, + "grad_norm": 3.465039425796171, + "learning_rate": 4.094533653040814e-06, + "loss": 0.4001, + "step": 876 + }, + { + "epoch": 0.3528465097565882, + "grad_norm": 3.4706674601181167, + "learning_rate": 4.091827984489493e-06, + "loss": 0.4629, + "step": 877 + }, + { + "epoch": 0.3532488432910883, + "grad_norm": 4.340590023749821, + "learning_rate": 4.089119176413926e-06, + "loss": 0.6815, + "step": 878 + }, + { + "epoch": 0.3536511768255884, + "grad_norm": 4.471466388864396, + "learning_rate": 4.086407234156633e-06, + "loss": 0.6459, + "step": 879 + }, + { + "epoch": 0.35405351036008853, + "grad_norm": 4.449019271914241, + "learning_rate": 4.08369216306631e-06, + "loss": 0.5592, + "step": 880 + }, + { + "epoch": 0.3544558438945886, + "grad_norm": 4.125369265434845, + "learning_rate": 4.080973968497829e-06, + "loss": 0.5138, + "step": 881 + }, + { + "epoch": 0.3548581774290887, + "grad_norm": 4.523305588736027, + "learning_rate": 4.07825265581222e-06, + "loss": 0.5929, + "step": 882 + }, + { + "epoch": 0.3552605109635888, + "grad_norm": 4.077469504247697, + "learning_rate": 4.075528230376662e-06, + "loss": 0.5288, + "step": 883 + }, + { + "epoch": 0.35566284449808894, + "grad_norm": 5.092003038039258, + "learning_rate": 4.072800697564474e-06, + "loss": 0.6624, + "step": 884 + }, + { + "epoch": 0.356065178032589, + "grad_norm": 3.7756634610021935, + "learning_rate": 4.070070062755105e-06, + "loss": 0.5265, + "step": 885 + }, + { + "epoch": 0.3564675115670891, + "grad_norm": 4.106660405022582, + "learning_rate": 4.0673363313341195e-06, + "loss": 0.4853, + "step": 886 + }, + { + "epoch": 0.3568698451015892, + "grad_norm": 4.215215996153993, + "learning_rate": 4.064599508693191e-06, + "loss": 0.5721, + "step": 887 + }, + { + "epoch": 0.35727217863608934, + "grad_norm": 4.171756065223583, + "learning_rate": 4.06185960023009e-06, + "loss": 0.4649, + "step": 888 + }, + { + "epoch": 0.35767451217058943, + "grad_norm": 5.138546885500691, + "learning_rate": 4.05911661134867e-06, + "loss": 0.5972, + "step": 889 + }, + { + "epoch": 0.3580768457050895, + "grad_norm": 4.136963664338033, + "learning_rate": 4.0563705474588655e-06, + "loss": 0.5027, + "step": 890 + }, + { + "epoch": 0.3584791792395896, + "grad_norm": 3.940446094705443, + "learning_rate": 4.05362141397667e-06, + "loss": 0.5627, + "step": 891 + }, + { + "epoch": 0.3588815127740897, + "grad_norm": 4.024692229149252, + "learning_rate": 4.0508692163241356e-06, + "loss": 0.5521, + "step": 892 + }, + { + "epoch": 0.35928384630858984, + "grad_norm": 3.866043251441755, + "learning_rate": 4.048113959929354e-06, + "loss": 0.5125, + "step": 893 + }, + { + "epoch": 0.3596861798430899, + "grad_norm": 4.786779622248551, + "learning_rate": 4.0453556502264535e-06, + "loss": 0.6475, + "step": 894 + }, + { + "epoch": 0.36008851337759, + "grad_norm": 4.127247200504385, + "learning_rate": 4.042594292655581e-06, + "loss": 0.4877, + "step": 895 + }, + { + "epoch": 0.3604908469120901, + "grad_norm": 3.3408217784719634, + "learning_rate": 4.039829892662897e-06, + "loss": 0.4337, + "step": 896 + }, + { + "epoch": 0.36089318044659024, + "grad_norm": 4.010232394860985, + "learning_rate": 4.037062455700559e-06, + "loss": 0.493, + "step": 897 + }, + { + "epoch": 0.36129551398109033, + "grad_norm": 4.3695516591433785, + "learning_rate": 4.03429198722672e-06, + "loss": 0.6051, + "step": 898 + }, + { + "epoch": 0.3616978475155904, + "grad_norm": 4.766410636581824, + "learning_rate": 4.0315184927055065e-06, + "loss": 0.6602, + "step": 899 + }, + { + "epoch": 0.3621001810500905, + "grad_norm": 4.1149436048907955, + "learning_rate": 4.028741977607016e-06, + "loss": 0.599, + "step": 900 + }, + { + "epoch": 0.36250251458459065, + "grad_norm": 3.4871089312720045, + "learning_rate": 4.0259624474073025e-06, + "loss": 0.4779, + "step": 901 + }, + { + "epoch": 0.36290484811909074, + "grad_norm": 3.6367752409912266, + "learning_rate": 4.023179907588367e-06, + "loss": 0.4541, + "step": 902 + }, + { + "epoch": 0.3633071816535908, + "grad_norm": 3.3961433962608685, + "learning_rate": 4.020394363638147e-06, + "loss": 0.4265, + "step": 903 + }, + { + "epoch": 0.3637095151880909, + "grad_norm": 3.8352646417492524, + "learning_rate": 4.0176058210505045e-06, + "loss": 0.5351, + "step": 904 + }, + { + "epoch": 0.364111848722591, + "grad_norm": 4.242220553627046, + "learning_rate": 4.0148142853252125e-06, + "loss": 0.5792, + "step": 905 + }, + { + "epoch": 0.36451418225709115, + "grad_norm": 4.096283695840529, + "learning_rate": 4.0120197619679536e-06, + "loss": 0.5412, + "step": 906 + }, + { + "epoch": 0.36491651579159123, + "grad_norm": 4.729251903541468, + "learning_rate": 4.009222256490297e-06, + "loss": 0.6285, + "step": 907 + }, + { + "epoch": 0.3653188493260913, + "grad_norm": 4.084180186714877, + "learning_rate": 4.006421774409697e-06, + "loss": 0.6304, + "step": 908 + }, + { + "epoch": 0.3657211828605914, + "grad_norm": 4.280731987602122, + "learning_rate": 4.003618321249476e-06, + "loss": 0.5632, + "step": 909 + }, + { + "epoch": 0.36612351639509155, + "grad_norm": 3.868764716611033, + "learning_rate": 4.000811902538821e-06, + "loss": 0.4715, + "step": 910 + }, + { + "epoch": 0.36652584992959164, + "grad_norm": 3.6445778686982155, + "learning_rate": 3.99800252381276e-06, + "loss": 0.4084, + "step": 911 + }, + { + "epoch": 0.36692818346409173, + "grad_norm": 4.600601862430362, + "learning_rate": 3.995190190612165e-06, + "loss": 0.7218, + "step": 912 + }, + { + "epoch": 0.3673305169985918, + "grad_norm": 4.09418361850337, + "learning_rate": 3.992374908483735e-06, + "loss": 0.5913, + "step": 913 + }, + { + "epoch": 0.36773285053309196, + "grad_norm": 4.495329343052196, + "learning_rate": 3.9895566829799825e-06, + "loss": 0.6735, + "step": 914 + }, + { + "epoch": 0.36813518406759205, + "grad_norm": 4.408452152812246, + "learning_rate": 3.986735519659226e-06, + "loss": 0.6896, + "step": 915 + }, + { + "epoch": 0.36853751760209214, + "grad_norm": 3.7084757585295276, + "learning_rate": 3.983911424085578e-06, + "loss": 0.552, + "step": 916 + }, + { + "epoch": 0.3689398511365922, + "grad_norm": 3.8368900456075985, + "learning_rate": 3.981084401828937e-06, + "loss": 0.4944, + "step": 917 + }, + { + "epoch": 0.3693421846710923, + "grad_norm": 5.696133375208026, + "learning_rate": 3.978254458464969e-06, + "loss": 0.6338, + "step": 918 + }, + { + "epoch": 0.36974451820559245, + "grad_norm": 4.144365741395642, + "learning_rate": 3.975421599575103e-06, + "loss": 0.6463, + "step": 919 + }, + { + "epoch": 0.37014685174009254, + "grad_norm": 3.7665583063095682, + "learning_rate": 3.972585830746522e-06, + "loss": 0.4442, + "step": 920 + }, + { + "epoch": 0.37054918527459263, + "grad_norm": 4.341604001350467, + "learning_rate": 3.969747157572142e-06, + "loss": 0.7191, + "step": 921 + }, + { + "epoch": 0.3709515188090927, + "grad_norm": 4.724527719836307, + "learning_rate": 3.966905585650611e-06, + "loss": 0.7508, + "step": 922 + }, + { + "epoch": 0.37135385234359286, + "grad_norm": 6.112644943798556, + "learning_rate": 3.964061120586294e-06, + "loss": 0.6744, + "step": 923 + }, + { + "epoch": 0.37175618587809295, + "grad_norm": 4.630742134072846, + "learning_rate": 3.961213767989261e-06, + "loss": 0.5703, + "step": 924 + }, + { + "epoch": 0.37215851941259304, + "grad_norm": 3.5499854666112802, + "learning_rate": 3.958363533475277e-06, + "loss": 0.4563, + "step": 925 + }, + { + "epoch": 0.3725608529470931, + "grad_norm": 3.926859330076321, + "learning_rate": 3.955510422665791e-06, + "loss": 0.5159, + "step": 926 + }, + { + "epoch": 0.37296318648159327, + "grad_norm": 4.029967110899967, + "learning_rate": 3.952654441187927e-06, + "loss": 0.5267, + "step": 927 + }, + { + "epoch": 0.37336552001609336, + "grad_norm": 3.4483709639926414, + "learning_rate": 3.9497955946744675e-06, + "loss": 0.4584, + "step": 928 + }, + { + "epoch": 0.37376785355059344, + "grad_norm": 4.752285945278346, + "learning_rate": 3.946933888763847e-06, + "loss": 0.5743, + "step": 929 + }, + { + "epoch": 0.37417018708509353, + "grad_norm": 3.985801543832948, + "learning_rate": 3.9440693291001406e-06, + "loss": 0.5319, + "step": 930 + }, + { + "epoch": 0.3745725206195936, + "grad_norm": 3.5946794819227903, + "learning_rate": 3.941201921333048e-06, + "loss": 0.4217, + "step": 931 + }, + { + "epoch": 0.37497485415409376, + "grad_norm": 4.035357970865375, + "learning_rate": 3.9383316711178946e-06, + "loss": 0.503, + "step": 932 + }, + { + "epoch": 0.37537718768859385, + "grad_norm": 4.034839036975055, + "learning_rate": 3.935458584115599e-06, + "loss": 0.4698, + "step": 933 + }, + { + "epoch": 0.37577952122309394, + "grad_norm": 4.286963186422666, + "learning_rate": 3.932582665992688e-06, + "loss": 0.5901, + "step": 934 + }, + { + "epoch": 0.376181854757594, + "grad_norm": 4.021909925329209, + "learning_rate": 3.929703922421263e-06, + "loss": 0.5111, + "step": 935 + }, + { + "epoch": 0.37658418829209417, + "grad_norm": 4.0942098752342035, + "learning_rate": 3.926822359079001e-06, + "loss": 0.5472, + "step": 936 + }, + { + "epoch": 0.37698652182659426, + "grad_norm": 4.261504012145757, + "learning_rate": 3.923937981649143e-06, + "loss": 0.6453, + "step": 937 + }, + { + "epoch": 0.37738885536109434, + "grad_norm": 4.014117574115026, + "learning_rate": 3.921050795820474e-06, + "loss": 0.5056, + "step": 938 + }, + { + "epoch": 0.37779118889559443, + "grad_norm": 3.904726379027022, + "learning_rate": 3.9181608072873244e-06, + "loss": 0.5304, + "step": 939 + }, + { + "epoch": 0.3781935224300946, + "grad_norm": 3.881019901522325, + "learning_rate": 3.9152680217495475e-06, + "loss": 0.5298, + "step": 940 + }, + { + "epoch": 0.37859585596459466, + "grad_norm": 4.058485666284913, + "learning_rate": 3.912372444912517e-06, + "loss": 0.574, + "step": 941 + }, + { + "epoch": 0.37899818949909475, + "grad_norm": 3.9630383581476205, + "learning_rate": 3.909474082487107e-06, + "loss": 0.5804, + "step": 942 + }, + { + "epoch": 0.37940052303359484, + "grad_norm": 3.8676063595588683, + "learning_rate": 3.906572940189691e-06, + "loss": 0.4804, + "step": 943 + }, + { + "epoch": 0.3798028565680949, + "grad_norm": 4.32834774261345, + "learning_rate": 3.9036690237421215e-06, + "loss": 0.6161, + "step": 944 + }, + { + "epoch": 0.38020519010259507, + "grad_norm": 4.726002477400136, + "learning_rate": 3.900762338871723e-06, + "loss": 0.6898, + "step": 945 + }, + { + "epoch": 0.38060752363709516, + "grad_norm": 4.577558255257823, + "learning_rate": 3.897852891311282e-06, + "loss": 0.5504, + "step": 946 + }, + { + "epoch": 0.38100985717159525, + "grad_norm": 4.300878005170684, + "learning_rate": 3.89494068679903e-06, + "loss": 0.6734, + "step": 947 + }, + { + "epoch": 0.38141219070609533, + "grad_norm": 4.519778162808581, + "learning_rate": 3.892025731078641e-06, + "loss": 0.5694, + "step": 948 + }, + { + "epoch": 0.3818145242405955, + "grad_norm": 3.8176275411987683, + "learning_rate": 3.88910802989921e-06, + "loss": 0.4444, + "step": 949 + }, + { + "epoch": 0.38221685777509556, + "grad_norm": 4.014909553832162, + "learning_rate": 3.886187589015251e-06, + "loss": 0.6054, + "step": 950 + }, + { + "epoch": 0.38261919130959565, + "grad_norm": 4.267781377287406, + "learning_rate": 3.883264414186677e-06, + "loss": 0.6257, + "step": 951 + }, + { + "epoch": 0.38302152484409574, + "grad_norm": 4.345044782459552, + "learning_rate": 3.8803385111788e-06, + "loss": 0.5543, + "step": 952 + }, + { + "epoch": 0.3834238583785959, + "grad_norm": 4.221128544535545, + "learning_rate": 3.877409885762305e-06, + "loss": 0.5544, + "step": 953 + }, + { + "epoch": 0.38382619191309597, + "grad_norm": 4.000060887768021, + "learning_rate": 3.8744785437132506e-06, + "loss": 0.5878, + "step": 954 + }, + { + "epoch": 0.38422852544759606, + "grad_norm": 3.8372993108948488, + "learning_rate": 3.871544490813054e-06, + "loss": 0.5076, + "step": 955 + }, + { + "epoch": 0.38463085898209615, + "grad_norm": 4.059706126497868, + "learning_rate": 3.868607732848475e-06, + "loss": 0.5353, + "step": 956 + }, + { + "epoch": 0.38503319251659623, + "grad_norm": 4.408924474844376, + "learning_rate": 3.865668275611614e-06, + "loss": 0.519, + "step": 957 + }, + { + "epoch": 0.3854355260510964, + "grad_norm": 4.010089079135815, + "learning_rate": 3.862726124899889e-06, + "loss": 0.5952, + "step": 958 + }, + { + "epoch": 0.38583785958559647, + "grad_norm": 5.459071377748317, + "learning_rate": 3.859781286516037e-06, + "loss": 0.7175, + "step": 959 + }, + { + "epoch": 0.38624019312009655, + "grad_norm": 4.542741280511702, + "learning_rate": 3.85683376626809e-06, + "loss": 0.5667, + "step": 960 + }, + { + "epoch": 0.38664252665459664, + "grad_norm": 3.99184341470565, + "learning_rate": 3.853883569969372e-06, + "loss": 0.6148, + "step": 961 + }, + { + "epoch": 0.3870448601890968, + "grad_norm": 4.078873510656604, + "learning_rate": 3.850930703438484e-06, + "loss": 0.5383, + "step": 962 + }, + { + "epoch": 0.38744719372359687, + "grad_norm": 3.896065226196885, + "learning_rate": 3.847975172499295e-06, + "loss": 0.6785, + "step": 963 + }, + { + "epoch": 0.38784952725809696, + "grad_norm": 4.328062559714918, + "learning_rate": 3.845016982980929e-06, + "loss": 0.5543, + "step": 964 + }, + { + "epoch": 0.38825186079259705, + "grad_norm": 4.2981177827507935, + "learning_rate": 3.842056140717749e-06, + "loss": 0.6935, + "step": 965 + }, + { + "epoch": 0.3886541943270972, + "grad_norm": 4.489409385825919, + "learning_rate": 3.839092651549357e-06, + "loss": 0.6412, + "step": 966 + }, + { + "epoch": 0.3890565278615973, + "grad_norm": 4.162311148181462, + "learning_rate": 3.836126521320569e-06, + "loss": 0.5133, + "step": 967 + }, + { + "epoch": 0.38945886139609737, + "grad_norm": 3.7564756341509464, + "learning_rate": 3.833157755881414e-06, + "loss": 0.5198, + "step": 968 + }, + { + "epoch": 0.38986119493059745, + "grad_norm": 4.2353823981014385, + "learning_rate": 3.830186361087117e-06, + "loss": 0.4271, + "step": 969 + }, + { + "epoch": 0.39026352846509754, + "grad_norm": 3.4858668216855437, + "learning_rate": 3.827212342798089e-06, + "loss": 0.4654, + "step": 970 + }, + { + "epoch": 0.3906658619995977, + "grad_norm": 3.920259549785761, + "learning_rate": 3.824235706879915e-06, + "loss": 0.5822, + "step": 971 + }, + { + "epoch": 0.3910681955340978, + "grad_norm": 4.239371514940848, + "learning_rate": 3.821256459203343e-06, + "loss": 0.5093, + "step": 972 + }, + { + "epoch": 0.39147052906859786, + "grad_norm": 4.037277711642943, + "learning_rate": 3.81827460564427e-06, + "loss": 0.5222, + "step": 973 + }, + { + "epoch": 0.39187286260309795, + "grad_norm": 4.456827976048942, + "learning_rate": 3.815290152083737e-06, + "loss": 0.6783, + "step": 974 + }, + { + "epoch": 0.3922751961375981, + "grad_norm": 4.246220187755789, + "learning_rate": 3.8123031044079084e-06, + "loss": 0.5888, + "step": 975 + }, + { + "epoch": 0.3926775296720982, + "grad_norm": 4.3203808668287635, + "learning_rate": 3.809313468508068e-06, + "loss": 0.5569, + "step": 976 + }, + { + "epoch": 0.39307986320659827, + "grad_norm": 4.860143257226567, + "learning_rate": 3.8063212502806035e-06, + "loss": 0.5953, + "step": 977 + }, + { + "epoch": 0.39348219674109836, + "grad_norm": 4.233378915892834, + "learning_rate": 3.8033264556269954e-06, + "loss": 0.6167, + "step": 978 + }, + { + "epoch": 0.39388453027559844, + "grad_norm": 4.3534789683754305, + "learning_rate": 3.8003290904538035e-06, + "loss": 0.6133, + "step": 979 + }, + { + "epoch": 0.3942868638100986, + "grad_norm": 3.9519882332598257, + "learning_rate": 3.7973291606726614e-06, + "loss": 0.4515, + "step": 980 + }, + { + "epoch": 0.3946891973445987, + "grad_norm": 4.138408229253922, + "learning_rate": 3.794326672200258e-06, + "loss": 0.5607, + "step": 981 + }, + { + "epoch": 0.39509153087909876, + "grad_norm": 4.3761985221008635, + "learning_rate": 3.79132163095833e-06, + "loss": 0.5507, + "step": 982 + }, + { + "epoch": 0.39549386441359885, + "grad_norm": 4.023180334056979, + "learning_rate": 3.7883140428736477e-06, + "loss": 0.5314, + "step": 983 + }, + { + "epoch": 0.395896197948099, + "grad_norm": 4.151232898341051, + "learning_rate": 3.7853039138780057e-06, + "loss": 0.5392, + "step": 984 + }, + { + "epoch": 0.3962985314825991, + "grad_norm": 4.539638233125882, + "learning_rate": 3.7822912499082087e-06, + "loss": 0.5791, + "step": 985 + }, + { + "epoch": 0.39670086501709917, + "grad_norm": 4.335330132423181, + "learning_rate": 3.7792760569060626e-06, + "loss": 0.5178, + "step": 986 + }, + { + "epoch": 0.39710319855159926, + "grad_norm": 4.143033413053681, + "learning_rate": 3.7762583408183594e-06, + "loss": 0.5815, + "step": 987 + }, + { + "epoch": 0.3975055320860994, + "grad_norm": 4.915299915525361, + "learning_rate": 3.7732381075968694e-06, + "loss": 0.6269, + "step": 988 + }, + { + "epoch": 0.3979078656205995, + "grad_norm": 3.7212476195736506, + "learning_rate": 3.770215363198325e-06, + "loss": 0.4599, + "step": 989 + }, + { + "epoch": 0.3983101991550996, + "grad_norm": 3.9977745943956884, + "learning_rate": 3.7671901135844148e-06, + "loss": 0.5973, + "step": 990 + }, + { + "epoch": 0.39871253268959966, + "grad_norm": 4.3088468446821615, + "learning_rate": 3.764162364721764e-06, + "loss": 0.5963, + "step": 991 + }, + { + "epoch": 0.39911486622409975, + "grad_norm": 4.2991674025596955, + "learning_rate": 3.761132122581931e-06, + "loss": 0.6923, + "step": 992 + }, + { + "epoch": 0.3995171997585999, + "grad_norm": 3.952746214125807, + "learning_rate": 3.758099393141388e-06, + "loss": 0.5384, + "step": 993 + }, + { + "epoch": 0.3999195332931, + "grad_norm": 4.327315800519712, + "learning_rate": 3.755064182381516e-06, + "loss": 0.5144, + "step": 994 + }, + { + "epoch": 0.40032186682760007, + "grad_norm": 4.713412610694241, + "learning_rate": 3.752026496288588e-06, + "loss": 0.641, + "step": 995 + }, + { + "epoch": 0.40072420036210016, + "grad_norm": 3.9912240405323507, + "learning_rate": 3.74898634085376e-06, + "loss": 0.5646, + "step": 996 + }, + { + "epoch": 0.4011265338966003, + "grad_norm": 4.530973641905437, + "learning_rate": 3.7459437220730583e-06, + "loss": 0.5614, + "step": 997 + }, + { + "epoch": 0.4015288674311004, + "grad_norm": 6.864543852090396, + "learning_rate": 3.742898645947366e-06, + "loss": 0.5336, + "step": 998 + }, + { + "epoch": 0.4019312009656005, + "grad_norm": 4.429956976671279, + "learning_rate": 3.7398511184824144e-06, + "loss": 0.7004, + "step": 999 + }, + { + "epoch": 0.40233353450010056, + "grad_norm": 3.164463326468864, + "learning_rate": 3.7368011456887693e-06, + "loss": 0.4163, + "step": 1000 + }, + { + "epoch": 0.4027358680346007, + "grad_norm": 3.7984416296897456, + "learning_rate": 3.7337487335818185e-06, + "loss": 0.5342, + "step": 1001 + }, + { + "epoch": 0.4031382015691008, + "grad_norm": 4.186189188641626, + "learning_rate": 3.730693888181761e-06, + "loss": 0.6125, + "step": 1002 + }, + { + "epoch": 0.4035405351036009, + "grad_norm": 3.8782636909404506, + "learning_rate": 3.7276366155135968e-06, + "loss": 0.5669, + "step": 1003 + }, + { + "epoch": 0.40394286863810097, + "grad_norm": 4.222464351010827, + "learning_rate": 3.7245769216071104e-06, + "loss": 0.5574, + "step": 1004 + }, + { + "epoch": 0.40434520217260106, + "grad_norm": 4.0429197887930135, + "learning_rate": 3.721514812496863e-06, + "loss": 0.5031, + "step": 1005 + }, + { + "epoch": 0.4047475357071012, + "grad_norm": 4.015964628315664, + "learning_rate": 3.718450294222179e-06, + "loss": 0.5994, + "step": 1006 + }, + { + "epoch": 0.4051498692416013, + "grad_norm": 3.846521589103569, + "learning_rate": 3.7153833728271356e-06, + "loss": 0.4928, + "step": 1007 + }, + { + "epoch": 0.4055522027761014, + "grad_norm": 4.138445980201307, + "learning_rate": 3.712314054360547e-06, + "loss": 0.5722, + "step": 1008 + }, + { + "epoch": 0.40595453631060147, + "grad_norm": 4.476195845673172, + "learning_rate": 3.7092423448759577e-06, + "loss": 0.563, + "step": 1009 + }, + { + "epoch": 0.4063568698451016, + "grad_norm": 3.786638148470317, + "learning_rate": 3.7061682504316264e-06, + "loss": 0.5181, + "step": 1010 + }, + { + "epoch": 0.4067592033796017, + "grad_norm": 4.563364096910093, + "learning_rate": 3.7030917770905155e-06, + "loss": 0.5247, + "step": 1011 + }, + { + "epoch": 0.4071615369141018, + "grad_norm": 4.276417493626447, + "learning_rate": 3.7000129309202804e-06, + "loss": 0.5815, + "step": 1012 + }, + { + "epoch": 0.4075638704486019, + "grad_norm": 4.495474983317033, + "learning_rate": 3.696931717993256e-06, + "loss": 0.7269, + "step": 1013 + }, + { + "epoch": 0.407966203983102, + "grad_norm": 3.6238415974427527, + "learning_rate": 3.693848144386444e-06, + "loss": 0.5462, + "step": 1014 + }, + { + "epoch": 0.4083685375176021, + "grad_norm": 4.262996632479673, + "learning_rate": 3.6907622161815027e-06, + "loss": 0.64, + "step": 1015 + }, + { + "epoch": 0.4087708710521022, + "grad_norm": 3.8744140465414745, + "learning_rate": 3.687673939464736e-06, + "loss": 0.5281, + "step": 1016 + }, + { + "epoch": 0.4091732045866023, + "grad_norm": 4.021193749407921, + "learning_rate": 3.6845833203270766e-06, + "loss": 0.5238, + "step": 1017 + }, + { + "epoch": 0.40957553812110237, + "grad_norm": 3.7942462225606084, + "learning_rate": 3.6814903648640787e-06, + "loss": 0.5274, + "step": 1018 + }, + { + "epoch": 0.4099778716556025, + "grad_norm": 3.8584763272892193, + "learning_rate": 3.6783950791759053e-06, + "loss": 0.6652, + "step": 1019 + }, + { + "epoch": 0.4103802051901026, + "grad_norm": 4.224914273954211, + "learning_rate": 3.6752974693673144e-06, + "loss": 0.6418, + "step": 1020 + }, + { + "epoch": 0.4107825387246027, + "grad_norm": 3.611777243329413, + "learning_rate": 3.6721975415476475e-06, + "loss": 0.4478, + "step": 1021 + }, + { + "epoch": 0.4111848722591028, + "grad_norm": 4.492308490256186, + "learning_rate": 3.669095301830816e-06, + "loss": 0.6383, + "step": 1022 + }, + { + "epoch": 0.4115872057936029, + "grad_norm": 3.966328790708119, + "learning_rate": 3.6659907563352964e-06, + "loss": 0.5496, + "step": 1023 + }, + { + "epoch": 0.411989539328103, + "grad_norm": 4.331090241917638, + "learning_rate": 3.6628839111841067e-06, + "loss": 0.6138, + "step": 1024 + }, + { + "epoch": 0.4123918728626031, + "grad_norm": 4.039380276956274, + "learning_rate": 3.6597747725048054e-06, + "loss": 0.514, + "step": 1025 + }, + { + "epoch": 0.4127942063971032, + "grad_norm": 4.2678553912438515, + "learning_rate": 3.656663346429471e-06, + "loss": 0.5304, + "step": 1026 + }, + { + "epoch": 0.4131965399316033, + "grad_norm": 3.6311274537700444, + "learning_rate": 3.6535496390946958e-06, + "loss": 0.5324, + "step": 1027 + }, + { + "epoch": 0.4135988734661034, + "grad_norm": 4.179608207864191, + "learning_rate": 3.650433656641569e-06, + "loss": 0.5913, + "step": 1028 + }, + { + "epoch": 0.4140012070006035, + "grad_norm": 3.8042489733382308, + "learning_rate": 3.647315405215668e-06, + "loss": 0.5007, + "step": 1029 + }, + { + "epoch": 0.4144035405351036, + "grad_norm": 3.4526666984193404, + "learning_rate": 3.644194890967047e-06, + "loss": 0.4141, + "step": 1030 + }, + { + "epoch": 0.4148058740696037, + "grad_norm": 4.657500083163389, + "learning_rate": 3.641072120050221e-06, + "loss": 0.6175, + "step": 1031 + }, + { + "epoch": 0.4152082076041038, + "grad_norm": 3.5394692246958934, + "learning_rate": 3.637947098624156e-06, + "loss": 0.493, + "step": 1032 + }, + { + "epoch": 0.4156105411386039, + "grad_norm": 3.6382852897066003, + "learning_rate": 3.6348198328522565e-06, + "loss": 0.4733, + "step": 1033 + }, + { + "epoch": 0.416012874673104, + "grad_norm": 4.701593231983809, + "learning_rate": 3.631690328902355e-06, + "loss": 0.6759, + "step": 1034 + }, + { + "epoch": 0.4164152082076041, + "grad_norm": 4.126519542614592, + "learning_rate": 3.6285585929466982e-06, + "loss": 0.6065, + "step": 1035 + }, + { + "epoch": 0.4168175417421042, + "grad_norm": 4.174303064525279, + "learning_rate": 3.6254246311619317e-06, + "loss": 0.6019, + "step": 1036 + }, + { + "epoch": 0.4172198752766043, + "grad_norm": 4.369071505943041, + "learning_rate": 3.6222884497290937e-06, + "loss": 0.47, + "step": 1037 + }, + { + "epoch": 0.4176222088111044, + "grad_norm": 4.547283701461554, + "learning_rate": 3.619150054833601e-06, + "loss": 0.5666, + "step": 1038 + }, + { + "epoch": 0.4180245423456045, + "grad_norm": 6.593645962994248, + "learning_rate": 3.616009452665234e-06, + "loss": 0.4854, + "step": 1039 + }, + { + "epoch": 0.41842687588010463, + "grad_norm": 4.3350300541777145, + "learning_rate": 3.6128666494181274e-06, + "loss": 0.6522, + "step": 1040 + }, + { + "epoch": 0.4188292094146047, + "grad_norm": 3.4916298319626233, + "learning_rate": 3.6097216512907574e-06, + "loss": 0.5449, + "step": 1041 + }, + { + "epoch": 0.4192315429491048, + "grad_norm": 3.721480691979447, + "learning_rate": 3.6065744644859276e-06, + "loss": 0.4152, + "step": 1042 + }, + { + "epoch": 0.4196338764836049, + "grad_norm": 4.419075622535455, + "learning_rate": 3.6034250952107598e-06, + "loss": 0.5998, + "step": 1043 + }, + { + "epoch": 0.420036210018105, + "grad_norm": 4.066379360835346, + "learning_rate": 3.6002735496766787e-06, + "loss": 0.4443, + "step": 1044 + }, + { + "epoch": 0.4204385435526051, + "grad_norm": 5.277931637633872, + "learning_rate": 3.5971198340994035e-06, + "loss": 0.6578, + "step": 1045 + }, + { + "epoch": 0.4208408770871052, + "grad_norm": 4.171317951485179, + "learning_rate": 3.5939639546989315e-06, + "loss": 0.5362, + "step": 1046 + }, + { + "epoch": 0.4212432106216053, + "grad_norm": 3.6581061080480786, + "learning_rate": 3.5908059176995274e-06, + "loss": 0.4868, + "step": 1047 + }, + { + "epoch": 0.4216455441561054, + "grad_norm": 4.089304060454962, + "learning_rate": 3.587645729329713e-06, + "loss": 0.4962, + "step": 1048 + }, + { + "epoch": 0.42204787769060553, + "grad_norm": 4.193929975116655, + "learning_rate": 3.584483395822252e-06, + "loss": 0.585, + "step": 1049 + }, + { + "epoch": 0.4224502112251056, + "grad_norm": 4.4833727752476324, + "learning_rate": 3.5813189234141386e-06, + "loss": 0.5295, + "step": 1050 + }, + { + "epoch": 0.4228525447596057, + "grad_norm": 4.131740789977084, + "learning_rate": 3.5781523183465868e-06, + "loss": 0.5408, + "step": 1051 + }, + { + "epoch": 0.4232548782941058, + "grad_norm": 3.972093255632944, + "learning_rate": 3.574983586865015e-06, + "loss": 0.5256, + "step": 1052 + }, + { + "epoch": 0.42365721182860594, + "grad_norm": 4.619320510735413, + "learning_rate": 3.571812735219037e-06, + "loss": 0.6365, + "step": 1053 + }, + { + "epoch": 0.424059545363106, + "grad_norm": 4.543758819054478, + "learning_rate": 3.5686397696624486e-06, + "loss": 0.5358, + "step": 1054 + }, + { + "epoch": 0.4244618788976061, + "grad_norm": 3.9488894694458305, + "learning_rate": 3.565464696453212e-06, + "loss": 0.5101, + "step": 1055 + }, + { + "epoch": 0.4248642124321062, + "grad_norm": 3.9008766737779217, + "learning_rate": 3.5622875218534493e-06, + "loss": 0.6157, + "step": 1056 + }, + { + "epoch": 0.4252665459666063, + "grad_norm": 4.072195384389267, + "learning_rate": 3.5591082521294264e-06, + "loss": 0.572, + "step": 1057 + }, + { + "epoch": 0.42566887950110643, + "grad_norm": 3.7407260188563622, + "learning_rate": 3.555926893551539e-06, + "loss": 0.4571, + "step": 1058 + }, + { + "epoch": 0.4260712130356065, + "grad_norm": 4.030808733877317, + "learning_rate": 3.552743452394306e-06, + "loss": 0.5615, + "step": 1059 + }, + { + "epoch": 0.4264735465701066, + "grad_norm": 3.833659387273303, + "learning_rate": 3.5495579349363517e-06, + "loss": 0.4978, + "step": 1060 + }, + { + "epoch": 0.4268758801046067, + "grad_norm": 4.304988237070821, + "learning_rate": 3.5463703474603964e-06, + "loss": 0.6441, + "step": 1061 + }, + { + "epoch": 0.42727821363910684, + "grad_norm": 4.68914366889499, + "learning_rate": 3.5431806962532426e-06, + "loss": 0.5009, + "step": 1062 + }, + { + "epoch": 0.42768054717360693, + "grad_norm": 4.208417431491987, + "learning_rate": 3.539988987605763e-06, + "loss": 0.5556, + "step": 1063 + }, + { + "epoch": 0.428082880708107, + "grad_norm": 3.913021062753838, + "learning_rate": 3.5367952278128877e-06, + "loss": 0.3997, + "step": 1064 + }, + { + "epoch": 0.4284852142426071, + "grad_norm": 3.975035687293758, + "learning_rate": 3.5335994231735927e-06, + "loss": 0.5394, + "step": 1065 + }, + { + "epoch": 0.42888754777710725, + "grad_norm": 3.969954958131227, + "learning_rate": 3.5304015799908875e-06, + "loss": 0.579, + "step": 1066 + }, + { + "epoch": 0.42928988131160734, + "grad_norm": 4.942462188783348, + "learning_rate": 3.5272017045718016e-06, + "loss": 0.6723, + "step": 1067 + }, + { + "epoch": 0.4296922148461074, + "grad_norm": 4.354492013681113, + "learning_rate": 3.5239998032273727e-06, + "loss": 0.5724, + "step": 1068 + }, + { + "epoch": 0.4300945483806075, + "grad_norm": 4.275789917585218, + "learning_rate": 3.520795882272634e-06, + "loss": 0.5592, + "step": 1069 + }, + { + "epoch": 0.4304968819151076, + "grad_norm": 4.0220453706013535, + "learning_rate": 3.5175899480266023e-06, + "loss": 0.6004, + "step": 1070 + }, + { + "epoch": 0.43089921544960774, + "grad_norm": 3.956820759857449, + "learning_rate": 3.5143820068122647e-06, + "loss": 0.4667, + "step": 1071 + }, + { + "epoch": 0.43130154898410783, + "grad_norm": 3.9359268522062996, + "learning_rate": 3.5111720649565685e-06, + "loss": 0.4766, + "step": 1072 + }, + { + "epoch": 0.4317038825186079, + "grad_norm": 3.9665074015294555, + "learning_rate": 3.507960128790402e-06, + "loss": 0.5496, + "step": 1073 + }, + { + "epoch": 0.432106216053108, + "grad_norm": 4.086635346156316, + "learning_rate": 3.5047462046485934e-06, + "loss": 0.4863, + "step": 1074 + }, + { + "epoch": 0.43250854958760815, + "grad_norm": 4.234886651798155, + "learning_rate": 3.501530298869886e-06, + "loss": 0.5373, + "step": 1075 + }, + { + "epoch": 0.43291088312210824, + "grad_norm": 4.578190826029045, + "learning_rate": 3.4983124177969353e-06, + "loss": 0.6952, + "step": 1076 + }, + { + "epoch": 0.4333132166566083, + "grad_norm": 4.198352580671215, + "learning_rate": 3.495092567776291e-06, + "loss": 0.6753, + "step": 1077 + }, + { + "epoch": 0.4337155501911084, + "grad_norm": 4.25314346947828, + "learning_rate": 3.4918707551583853e-06, + "loss": 0.6139, + "step": 1078 + }, + { + "epoch": 0.43411788372560856, + "grad_norm": 3.7809774120139084, + "learning_rate": 3.4886469862975224e-06, + "loss": 0.5401, + "step": 1079 + }, + { + "epoch": 0.43452021726010864, + "grad_norm": 3.861747118430867, + "learning_rate": 3.485421267551865e-06, + "loss": 0.4343, + "step": 1080 + }, + { + "epoch": 0.43492255079460873, + "grad_norm": 4.1971689225974655, + "learning_rate": 3.482193605283421e-06, + "loss": 0.4651, + "step": 1081 + }, + { + "epoch": 0.4353248843291088, + "grad_norm": 4.719733540576249, + "learning_rate": 3.478964005858031e-06, + "loss": 0.6265, + "step": 1082 + }, + { + "epoch": 0.4357272178636089, + "grad_norm": 4.199121895188385, + "learning_rate": 3.475732475645357e-06, + "loss": 0.5915, + "step": 1083 + }, + { + "epoch": 0.43612955139810905, + "grad_norm": 4.117901413415391, + "learning_rate": 3.4724990210188693e-06, + "loss": 0.5859, + "step": 1084 + }, + { + "epoch": 0.43653188493260914, + "grad_norm": 4.063314519073365, + "learning_rate": 3.4692636483558316e-06, + "loss": 0.5799, + "step": 1085 + }, + { + "epoch": 0.4369342184671092, + "grad_norm": 4.291259902841789, + "learning_rate": 3.4660263640372926e-06, + "loss": 0.5972, + "step": 1086 + }, + { + "epoch": 0.4373365520016093, + "grad_norm": 4.317671430467203, + "learning_rate": 3.462787174448071e-06, + "loss": 0.5818, + "step": 1087 + }, + { + "epoch": 0.43773888553610946, + "grad_norm": 4.041367856828536, + "learning_rate": 3.459546085976743e-06, + "loss": 0.5635, + "step": 1088 + }, + { + "epoch": 0.43814121907060954, + "grad_norm": 5.295724094479603, + "learning_rate": 3.45630310501563e-06, + "loss": 0.6712, + "step": 1089 + }, + { + "epoch": 0.43854355260510963, + "grad_norm": 4.415113705137869, + "learning_rate": 3.453058237960785e-06, + "loss": 0.578, + "step": 1090 + }, + { + "epoch": 0.4389458861396097, + "grad_norm": 4.107191353105501, + "learning_rate": 3.4498114912119817e-06, + "loss": 0.5199, + "step": 1091 + }, + { + "epoch": 0.43934821967410986, + "grad_norm": 4.1779462968909336, + "learning_rate": 3.4465628711727022e-06, + "loss": 0.577, + "step": 1092 + }, + { + "epoch": 0.43975055320860995, + "grad_norm": 3.779499327289251, + "learning_rate": 3.4433123842501205e-06, + "loss": 0.4508, + "step": 1093 + }, + { + "epoch": 0.44015288674311004, + "grad_norm": 3.9872787051249077, + "learning_rate": 3.440060036855095e-06, + "loss": 0.5583, + "step": 1094 + }, + { + "epoch": 0.4405552202776101, + "grad_norm": 4.317372927534622, + "learning_rate": 3.4368058354021526e-06, + "loss": 0.5136, + "step": 1095 + }, + { + "epoch": 0.4409575538121102, + "grad_norm": 3.7148972800031053, + "learning_rate": 3.4335497863094773e-06, + "loss": 0.514, + "step": 1096 + }, + { + "epoch": 0.44135988734661036, + "grad_norm": 4.342233656787418, + "learning_rate": 3.430291895998896e-06, + "loss": 0.6742, + "step": 1097 + }, + { + "epoch": 0.44176222088111045, + "grad_norm": 4.799719138290185, + "learning_rate": 3.4270321708958686e-06, + "loss": 0.6306, + "step": 1098 + }, + { + "epoch": 0.44216455441561053, + "grad_norm": 3.9209860725209382, + "learning_rate": 3.4237706174294726e-06, + "loss": 0.5836, + "step": 1099 + }, + { + "epoch": 0.4425668879501106, + "grad_norm": 4.127404724187905, + "learning_rate": 3.420507242032392e-06, + "loss": 0.5628, + "step": 1100 + }, + { + "epoch": 0.44296922148461076, + "grad_norm": 4.956274683013535, + "learning_rate": 3.4172420511409033e-06, + "loss": 0.5366, + "step": 1101 + }, + { + "epoch": 0.44337155501911085, + "grad_norm": 3.9527811079754285, + "learning_rate": 3.413975051194865e-06, + "loss": 0.4531, + "step": 1102 + }, + { + "epoch": 0.44377388855361094, + "grad_norm": 4.364141896413706, + "learning_rate": 3.4107062486377028e-06, + "loss": 0.4914, + "step": 1103 + }, + { + "epoch": 0.44417622208811103, + "grad_norm": 4.349108898092637, + "learning_rate": 3.4074356499163976e-06, + "loss": 0.7239, + "step": 1104 + }, + { + "epoch": 0.44457855562261117, + "grad_norm": 4.038974541606429, + "learning_rate": 3.4041632614814734e-06, + "loss": 0.6375, + "step": 1105 + }, + { + "epoch": 0.44498088915711126, + "grad_norm": 4.811612078810641, + "learning_rate": 3.400889089786984e-06, + "loss": 0.4963, + "step": 1106 + }, + { + "epoch": 0.44538322269161135, + "grad_norm": 4.07473806896573, + "learning_rate": 3.397613141290499e-06, + "loss": 0.575, + "step": 1107 + }, + { + "epoch": 0.44578555622611143, + "grad_norm": 4.160302353185681, + "learning_rate": 3.3943354224530934e-06, + "loss": 0.4766, + "step": 1108 + }, + { + "epoch": 0.4461878897606115, + "grad_norm": 3.6900007830747934, + "learning_rate": 3.3910559397393335e-06, + "loss": 0.5544, + "step": 1109 + }, + { + "epoch": 0.44659022329511167, + "grad_norm": 4.333978444807443, + "learning_rate": 3.3877746996172645e-06, + "loss": 0.5637, + "step": 1110 + }, + { + "epoch": 0.44699255682961175, + "grad_norm": 3.768061403793603, + "learning_rate": 3.384491708558399e-06, + "loss": 0.4659, + "step": 1111 + }, + { + "epoch": 0.44739489036411184, + "grad_norm": 3.9713683928936208, + "learning_rate": 3.3812069730377007e-06, + "loss": 0.539, + "step": 1112 + }, + { + "epoch": 0.44779722389861193, + "grad_norm": 4.95129840549997, + "learning_rate": 3.3779204995335747e-06, + "loss": 0.6069, + "step": 1113 + }, + { + "epoch": 0.4481995574331121, + "grad_norm": 4.316287155035261, + "learning_rate": 3.3746322945278543e-06, + "loss": 0.66, + "step": 1114 + }, + { + "epoch": 0.44860189096761216, + "grad_norm": 4.697629623928124, + "learning_rate": 3.3713423645057873e-06, + "loss": 0.4885, + "step": 1115 + }, + { + "epoch": 0.44900422450211225, + "grad_norm": 3.920242841924085, + "learning_rate": 3.368050715956025e-06, + "loss": 0.5213, + "step": 1116 + }, + { + "epoch": 0.44940655803661234, + "grad_norm": 4.521566910606868, + "learning_rate": 3.3647573553706052e-06, + "loss": 0.5743, + "step": 1117 + }, + { + "epoch": 0.4498088915711125, + "grad_norm": 4.082210637179168, + "learning_rate": 3.3614622892449465e-06, + "loss": 0.5361, + "step": 1118 + }, + { + "epoch": 0.45021122510561257, + "grad_norm": 4.017855379880608, + "learning_rate": 3.3581655240778277e-06, + "loss": 0.5492, + "step": 1119 + }, + { + "epoch": 0.45061355864011265, + "grad_norm": 3.3069156613524577, + "learning_rate": 3.35486706637138e-06, + "loss": 0.4142, + "step": 1120 + }, + { + "epoch": 0.45101589217461274, + "grad_norm": 4.230766496789279, + "learning_rate": 3.3515669226310716e-06, + "loss": 0.5883, + "step": 1121 + }, + { + "epoch": 0.45141822570911283, + "grad_norm": 3.932193721134766, + "learning_rate": 3.3482650993656983e-06, + "loss": 0.4732, + "step": 1122 + }, + { + "epoch": 0.451820559243613, + "grad_norm": 3.9297564952360577, + "learning_rate": 3.344961603087367e-06, + "loss": 0.5852, + "step": 1123 + }, + { + "epoch": 0.45222289277811306, + "grad_norm": 4.909401223605511, + "learning_rate": 3.341656440311484e-06, + "loss": 0.5709, + "step": 1124 + }, + { + "epoch": 0.45262522631261315, + "grad_norm": 3.6889862040375525, + "learning_rate": 3.3383496175567435e-06, + "loss": 0.5966, + "step": 1125 + }, + { + "epoch": 0.45302755984711324, + "grad_norm": 3.4699341807873694, + "learning_rate": 3.3350411413451125e-06, + "loss": 0.4692, + "step": 1126 + }, + { + "epoch": 0.4534298933816134, + "grad_norm": 4.481843121210536, + "learning_rate": 3.3317310182018186e-06, + "loss": 0.6374, + "step": 1127 + }, + { + "epoch": 0.45383222691611347, + "grad_norm": 4.674277827776204, + "learning_rate": 3.328419254655339e-06, + "loss": 0.7174, + "step": 1128 + }, + { + "epoch": 0.45423456045061356, + "grad_norm": 3.8262305960434575, + "learning_rate": 3.325105857237386e-06, + "loss": 0.4618, + "step": 1129 + }, + { + "epoch": 0.45463689398511364, + "grad_norm": 3.9899678016482305, + "learning_rate": 3.3217908324828942e-06, + "loss": 0.5839, + "step": 1130 + }, + { + "epoch": 0.4550392275196138, + "grad_norm": 3.39198458807276, + "learning_rate": 3.3184741869300073e-06, + "loss": 0.5191, + "step": 1131 + }, + { + "epoch": 0.4554415610541139, + "grad_norm": 4.032286065452944, + "learning_rate": 3.3151559271200656e-06, + "loss": 0.4835, + "step": 1132 + }, + { + "epoch": 0.45584389458861396, + "grad_norm": 4.381591989574539, + "learning_rate": 3.3118360595975947e-06, + "loss": 0.5592, + "step": 1133 + }, + { + "epoch": 0.45624622812311405, + "grad_norm": 4.022213508423109, + "learning_rate": 3.3085145909102885e-06, + "loss": 0.5304, + "step": 1134 + }, + { + "epoch": 0.45664856165761414, + "grad_norm": 3.834367329162151, + "learning_rate": 3.305191527609e-06, + "loss": 0.4678, + "step": 1135 + }, + { + "epoch": 0.4570508951921143, + "grad_norm": 4.448203020700801, + "learning_rate": 3.3018668762477294e-06, + "loss": 0.5172, + "step": 1136 + }, + { + "epoch": 0.45745322872661437, + "grad_norm": 4.128965375687235, + "learning_rate": 3.2985406433836055e-06, + "loss": 0.4742, + "step": 1137 + }, + { + "epoch": 0.45785556226111446, + "grad_norm": 4.129504251009841, + "learning_rate": 3.295212835576878e-06, + "loss": 0.5873, + "step": 1138 + }, + { + "epoch": 0.45825789579561454, + "grad_norm": 4.422224105540513, + "learning_rate": 3.2918834593909028e-06, + "loss": 0.6049, + "step": 1139 + }, + { + "epoch": 0.4586602293301147, + "grad_norm": 4.434937349809798, + "learning_rate": 3.2885525213921286e-06, + "loss": 0.5708, + "step": 1140 + }, + { + "epoch": 0.4590625628646148, + "grad_norm": 3.668465429492197, + "learning_rate": 3.2852200281500855e-06, + "loss": 0.4559, + "step": 1141 + }, + { + "epoch": 0.45946489639911486, + "grad_norm": 4.232532968956556, + "learning_rate": 3.2818859862373696e-06, + "loss": 0.5721, + "step": 1142 + }, + { + "epoch": 0.45986722993361495, + "grad_norm": 4.7338258325238085, + "learning_rate": 3.2785504022296316e-06, + "loss": 0.6157, + "step": 1143 + }, + { + "epoch": 0.4602695634681151, + "grad_norm": 3.7791426452155723, + "learning_rate": 3.2752132827055646e-06, + "loss": 0.5458, + "step": 1144 + }, + { + "epoch": 0.4606718970026152, + "grad_norm": 4.1019637445172314, + "learning_rate": 3.2718746342468894e-06, + "loss": 0.5972, + "step": 1145 + }, + { + "epoch": 0.46107423053711527, + "grad_norm": 5.277495798462077, + "learning_rate": 3.268534463438342e-06, + "loss": 0.8266, + "step": 1146 + }, + { + "epoch": 0.46147656407161536, + "grad_norm": 4.184453273932426, + "learning_rate": 3.2651927768676626e-06, + "loss": 0.4562, + "step": 1147 + }, + { + "epoch": 0.46187889760611545, + "grad_norm": 4.45907273350243, + "learning_rate": 3.261849581125579e-06, + "loss": 0.6532, + "step": 1148 + }, + { + "epoch": 0.4622812311406156, + "grad_norm": 3.965076229787784, + "learning_rate": 3.258504882805796e-06, + "loss": 0.5247, + "step": 1149 + }, + { + "epoch": 0.4626835646751157, + "grad_norm": 4.461055711184354, + "learning_rate": 3.2551586885049814e-06, + "loss": 0.4588, + "step": 1150 + }, + { + "epoch": 0.46308589820961576, + "grad_norm": 4.3124706302870255, + "learning_rate": 3.2518110048227553e-06, + "loss": 0.7004, + "step": 1151 + }, + { + "epoch": 0.46348823174411585, + "grad_norm": 4.326890541788323, + "learning_rate": 3.2484618383616735e-06, + "loss": 0.5989, + "step": 1152 + }, + { + "epoch": 0.463890565278616, + "grad_norm": 4.323105374837098, + "learning_rate": 3.2451111957272173e-06, + "loss": 0.5451, + "step": 1153 + }, + { + "epoch": 0.4642928988131161, + "grad_norm": 4.134824677878268, + "learning_rate": 3.241759083527779e-06, + "loss": 0.5589, + "step": 1154 + }, + { + "epoch": 0.46469523234761617, + "grad_norm": 4.0981640141993365, + "learning_rate": 3.238405508374649e-06, + "loss": 0.5771, + "step": 1155 + }, + { + "epoch": 0.46509756588211626, + "grad_norm": 3.7455178072585644, + "learning_rate": 3.2350504768820036e-06, + "loss": 0.5125, + "step": 1156 + }, + { + "epoch": 0.46549989941661635, + "grad_norm": 3.72359460364635, + "learning_rate": 3.23169399566689e-06, + "loss": 0.4213, + "step": 1157 + }, + { + "epoch": 0.4659022329511165, + "grad_norm": 3.509972924609966, + "learning_rate": 3.228336071349218e-06, + "loss": 0.4757, + "step": 1158 + }, + { + "epoch": 0.4663045664856166, + "grad_norm": 4.028400275439829, + "learning_rate": 3.2249767105517395e-06, + "loss": 0.5022, + "step": 1159 + }, + { + "epoch": 0.46670690002011667, + "grad_norm": 4.150972255373417, + "learning_rate": 3.221615919900042e-06, + "loss": 0.6257, + "step": 1160 + }, + { + "epoch": 0.46710923355461675, + "grad_norm": 4.494189407683215, + "learning_rate": 3.218253706022533e-06, + "loss": 0.5914, + "step": 1161 + }, + { + "epoch": 0.4675115670891169, + "grad_norm": 3.558180537366331, + "learning_rate": 3.2148900755504253e-06, + "loss": 0.5102, + "step": 1162 + }, + { + "epoch": 0.467913900623617, + "grad_norm": 3.438406963531958, + "learning_rate": 3.2115250351177274e-06, + "loss": 0.515, + "step": 1163 + }, + { + "epoch": 0.4683162341581171, + "grad_norm": 4.323925078750027, + "learning_rate": 3.208158591361228e-06, + "loss": 0.5109, + "step": 1164 + }, + { + "epoch": 0.46871856769261716, + "grad_norm": 4.1534707963201285, + "learning_rate": 3.204790750920484e-06, + "loss": 0.5736, + "step": 1165 + }, + { + "epoch": 0.4691209012271173, + "grad_norm": 3.6871720181954406, + "learning_rate": 3.2014215204378064e-06, + "loss": 0.5765, + "step": 1166 + }, + { + "epoch": 0.4695232347616174, + "grad_norm": 3.9220478852176264, + "learning_rate": 3.1980509065582476e-06, + "loss": 0.5545, + "step": 1167 + }, + { + "epoch": 0.4699255682961175, + "grad_norm": 4.603512689826395, + "learning_rate": 3.194678915929589e-06, + "loss": 0.5587, + "step": 1168 + }, + { + "epoch": 0.47032790183061757, + "grad_norm": 4.0356910482849075, + "learning_rate": 3.1913055552023263e-06, + "loss": 0.4642, + "step": 1169 + }, + { + "epoch": 0.47073023536511766, + "grad_norm": 4.251943286279792, + "learning_rate": 3.1879308310296586e-06, + "loss": 0.5496, + "step": 1170 + }, + { + "epoch": 0.4711325688996178, + "grad_norm": 4.189042957707438, + "learning_rate": 3.1845547500674737e-06, + "loss": 0.5646, + "step": 1171 + }, + { + "epoch": 0.4715349024341179, + "grad_norm": 3.3326084801089353, + "learning_rate": 3.1811773189743355e-06, + "loss": 0.4837, + "step": 1172 + }, + { + "epoch": 0.471937235968618, + "grad_norm": 4.523273563587766, + "learning_rate": 3.17779854441147e-06, + "loss": 0.6807, + "step": 1173 + }, + { + "epoch": 0.47233956950311806, + "grad_norm": 3.7232216855517315, + "learning_rate": 3.1744184330427543e-06, + "loss": 0.4492, + "step": 1174 + }, + { + "epoch": 0.4727419030376182, + "grad_norm": 3.8162014027986437, + "learning_rate": 3.1710369915347e-06, + "loss": 0.5626, + "step": 1175 + }, + { + "epoch": 0.4731442365721183, + "grad_norm": 4.575379469711398, + "learning_rate": 3.1676542265564443e-06, + "loss": 0.6358, + "step": 1176 + }, + { + "epoch": 0.4735465701066184, + "grad_norm": 4.563047737373793, + "learning_rate": 3.1642701447797324e-06, + "loss": 0.6282, + "step": 1177 + }, + { + "epoch": 0.47394890364111847, + "grad_norm": 4.254463762469985, + "learning_rate": 3.1608847528789096e-06, + "loss": 0.4974, + "step": 1178 + }, + { + "epoch": 0.4743512371756186, + "grad_norm": 4.290561801295637, + "learning_rate": 3.1574980575309015e-06, + "loss": 0.6116, + "step": 1179 + }, + { + "epoch": 0.4747535707101187, + "grad_norm": 4.276362615412023, + "learning_rate": 3.154110065415208e-06, + "loss": 0.5274, + "step": 1180 + }, + { + "epoch": 0.4751559042446188, + "grad_norm": 3.8389354369931374, + "learning_rate": 3.150720783213883e-06, + "loss": 0.5338, + "step": 1181 + }, + { + "epoch": 0.4755582377791189, + "grad_norm": 4.448041403696648, + "learning_rate": 3.1473302176115283e-06, + "loss": 0.5265, + "step": 1182 + }, + { + "epoch": 0.47596057131361896, + "grad_norm": 4.46050378901124, + "learning_rate": 3.1439383752952753e-06, + "loss": 0.6299, + "step": 1183 + }, + { + "epoch": 0.4763629048481191, + "grad_norm": 4.111077775261664, + "learning_rate": 3.140545262954772e-06, + "loss": 0.5723, + "step": 1184 + }, + { + "epoch": 0.4767652383826192, + "grad_norm": 3.678492409328739, + "learning_rate": 3.137150887282174e-06, + "loss": 0.4178, + "step": 1185 + }, + { + "epoch": 0.4771675719171193, + "grad_norm": 3.661689522078919, + "learning_rate": 3.1337552549721262e-06, + "loss": 0.4658, + "step": 1186 + }, + { + "epoch": 0.47756990545161937, + "grad_norm": 4.250351645694971, + "learning_rate": 3.130358372721754e-06, + "loss": 0.5614, + "step": 1187 + }, + { + "epoch": 0.4779722389861195, + "grad_norm": 3.599583256588181, + "learning_rate": 3.126960247230646e-06, + "loss": 0.4569, + "step": 1188 + }, + { + "epoch": 0.4783745725206196, + "grad_norm": 3.821896657809243, + "learning_rate": 3.1235608852008446e-06, + "loss": 0.4834, + "step": 1189 + }, + { + "epoch": 0.4787769060551197, + "grad_norm": 3.9335780159969906, + "learning_rate": 3.1201602933368308e-06, + "loss": 0.5152, + "step": 1190 + }, + { + "epoch": 0.4791792395896198, + "grad_norm": 4.770540317196732, + "learning_rate": 3.116758478345509e-06, + "loss": 0.5948, + "step": 1191 + }, + { + "epoch": 0.4795815731241199, + "grad_norm": 3.998481220511283, + "learning_rate": 3.1133554469361976e-06, + "loss": 0.4881, + "step": 1192 + }, + { + "epoch": 0.47998390665862, + "grad_norm": 4.208648289802026, + "learning_rate": 3.109951205820615e-06, + "loss": 0.4933, + "step": 1193 + }, + { + "epoch": 0.4803862401931201, + "grad_norm": 3.752607346638518, + "learning_rate": 3.1065457617128635e-06, + "loss": 0.547, + "step": 1194 + }, + { + "epoch": 0.4807885737276202, + "grad_norm": 4.151251812002077, + "learning_rate": 3.1031391213294203e-06, + "loss": 0.5015, + "step": 1195 + }, + { + "epoch": 0.48119090726212027, + "grad_norm": 4.7037289072786095, + "learning_rate": 3.0997312913891196e-06, + "loss": 0.6795, + "step": 1196 + }, + { + "epoch": 0.4815932407966204, + "grad_norm": 4.364147422750159, + "learning_rate": 3.0963222786131435e-06, + "loss": 0.5422, + "step": 1197 + }, + { + "epoch": 0.4819955743311205, + "grad_norm": 3.301076213262608, + "learning_rate": 3.0929120897250066e-06, + "loss": 0.4646, + "step": 1198 + }, + { + "epoch": 0.4823979078656206, + "grad_norm": 3.803770968997955, + "learning_rate": 3.0895007314505415e-06, + "loss": 0.4579, + "step": 1199 + }, + { + "epoch": 0.4828002414001207, + "grad_norm": 3.6319529462360816, + "learning_rate": 3.0860882105178897e-06, + "loss": 0.5152, + "step": 1200 + }, + { + "epoch": 0.4832025749346208, + "grad_norm": 3.7950010562493453, + "learning_rate": 3.082674533657484e-06, + "loss": 0.588, + "step": 1201 + }, + { + "epoch": 0.4836049084691209, + "grad_norm": 3.9284108228666796, + "learning_rate": 3.079259707602038e-06, + "loss": 0.5565, + "step": 1202 + }, + { + "epoch": 0.484007242003621, + "grad_norm": 4.764319789283065, + "learning_rate": 3.0758437390865314e-06, + "loss": 0.6115, + "step": 1203 + }, + { + "epoch": 0.4844095755381211, + "grad_norm": 3.781006793375389, + "learning_rate": 3.0724266348481966e-06, + "loss": 0.5239, + "step": 1204 + }, + { + "epoch": 0.4848119090726212, + "grad_norm": 4.259150542468107, + "learning_rate": 3.069008401626507e-06, + "loss": 0.5505, + "step": 1205 + }, + { + "epoch": 0.4852142426071213, + "grad_norm": 4.222901377467788, + "learning_rate": 3.065589046163162e-06, + "loss": 0.5425, + "step": 1206 + }, + { + "epoch": 0.4856165761416214, + "grad_norm": 3.783266465549031, + "learning_rate": 3.062168575202075e-06, + "loss": 0.542, + "step": 1207 + }, + { + "epoch": 0.4860189096761215, + "grad_norm": 4.428497795054238, + "learning_rate": 3.0587469954893584e-06, + "loss": 0.6567, + "step": 1208 + }, + { + "epoch": 0.4864212432106216, + "grad_norm": 3.691837740063112, + "learning_rate": 3.0553243137733125e-06, + "loss": 0.4558, + "step": 1209 + }, + { + "epoch": 0.4868235767451217, + "grad_norm": 4.073296743584824, + "learning_rate": 3.051900536804411e-06, + "loss": 0.5664, + "step": 1210 + }, + { + "epoch": 0.4872259102796218, + "grad_norm": 3.7192683213034234, + "learning_rate": 3.0484756713352858e-06, + "loss": 0.494, + "step": 1211 + }, + { + "epoch": 0.4876282438141219, + "grad_norm": 3.6377595381026557, + "learning_rate": 3.045049724120718e-06, + "loss": 0.4885, + "step": 1212 + }, + { + "epoch": 0.488030577348622, + "grad_norm": 4.338683717827157, + "learning_rate": 3.0416227019176208e-06, + "loss": 0.6165, + "step": 1213 + }, + { + "epoch": 0.48843291088312213, + "grad_norm": 4.374868800699678, + "learning_rate": 3.0381946114850293e-06, + "loss": 0.5755, + "step": 1214 + }, + { + "epoch": 0.4888352444176222, + "grad_norm": 4.54077234147531, + "learning_rate": 3.034765459584083e-06, + "loss": 0.5954, + "step": 1215 + }, + { + "epoch": 0.4892375779521223, + "grad_norm": 4.166268412885333, + "learning_rate": 3.031335252978017e-06, + "loss": 0.6657, + "step": 1216 + }, + { + "epoch": 0.4896399114866224, + "grad_norm": 3.675310094808539, + "learning_rate": 3.027903998432145e-06, + "loss": 0.4676, + "step": 1217 + }, + { + "epoch": 0.49004224502112254, + "grad_norm": 4.091571307318168, + "learning_rate": 3.0244717027138494e-06, + "loss": 0.579, + "step": 1218 + }, + { + "epoch": 0.4904445785556226, + "grad_norm": 4.2620047470347595, + "learning_rate": 3.021038372592563e-06, + "loss": 0.577, + "step": 1219 + }, + { + "epoch": 0.4908469120901227, + "grad_norm": 3.687909147915886, + "learning_rate": 3.017604014839762e-06, + "loss": 0.557, + "step": 1220 + }, + { + "epoch": 0.4912492456246228, + "grad_norm": 4.003092692480048, + "learning_rate": 3.014168636228948e-06, + "loss": 0.5517, + "step": 1221 + }, + { + "epoch": 0.4916515791591229, + "grad_norm": 4.349298145062102, + "learning_rate": 3.010732243535636e-06, + "loss": 0.5991, + "step": 1222 + }, + { + "epoch": 0.49205391269362303, + "grad_norm": 3.9645619754575963, + "learning_rate": 3.0072948435373408e-06, + "loss": 0.5145, + "step": 1223 + }, + { + "epoch": 0.4924562462281231, + "grad_norm": 4.216377733109406, + "learning_rate": 3.003856443013564e-06, + "loss": 0.6017, + "step": 1224 + }, + { + "epoch": 0.4928585797626232, + "grad_norm": 3.9206543304610584, + "learning_rate": 3.0004170487457813e-06, + "loss": 0.5468, + "step": 1225 + }, + { + "epoch": 0.4932609132971233, + "grad_norm": 4.196664240626691, + "learning_rate": 2.996976667517427e-06, + "loss": 0.5191, + "step": 1226 + }, + { + "epoch": 0.49366324683162344, + "grad_norm": 3.7775234111981493, + "learning_rate": 2.9935353061138833e-06, + "loss": 0.5094, + "step": 1227 + }, + { + "epoch": 0.4940655803661235, + "grad_norm": 4.412452469732304, + "learning_rate": 2.990092971322464e-06, + "loss": 0.566, + "step": 1228 + }, + { + "epoch": 0.4944679139006236, + "grad_norm": 3.967665512410053, + "learning_rate": 2.986649669932405e-06, + "loss": 0.5005, + "step": 1229 + }, + { + "epoch": 0.4948702474351237, + "grad_norm": 3.8061725506913353, + "learning_rate": 2.983205408734846e-06, + "loss": 0.5214, + "step": 1230 + }, + { + "epoch": 0.49527258096962384, + "grad_norm": 3.615684737615807, + "learning_rate": 2.9797601945228214e-06, + "loss": 0.4647, + "step": 1231 + }, + { + "epoch": 0.49567491450412393, + "grad_norm": 4.27866020604673, + "learning_rate": 2.9763140340912463e-06, + "loss": 0.6409, + "step": 1232 + }, + { + "epoch": 0.496077248038624, + "grad_norm": 3.596076423790147, + "learning_rate": 2.972866934236898e-06, + "loss": 0.412, + "step": 1233 + }, + { + "epoch": 0.4964795815731241, + "grad_norm": 4.0355620490752955, + "learning_rate": 2.9694189017584097e-06, + "loss": 0.5596, + "step": 1234 + }, + { + "epoch": 0.4968819151076242, + "grad_norm": 4.148051125186017, + "learning_rate": 2.965969943456254e-06, + "loss": 0.5191, + "step": 1235 + }, + { + "epoch": 0.49728424864212434, + "grad_norm": 4.2692044767988, + "learning_rate": 2.962520066132728e-06, + "loss": 0.5066, + "step": 1236 + }, + { + "epoch": 0.4976865821766244, + "grad_norm": 4.28381423937242, + "learning_rate": 2.9590692765919436e-06, + "loss": 0.598, + "step": 1237 + }, + { + "epoch": 0.4980889157111245, + "grad_norm": 4.175358078657294, + "learning_rate": 2.955617581639809e-06, + "loss": 0.5953, + "step": 1238 + }, + { + "epoch": 0.4984912492456246, + "grad_norm": 4.5191335184254955, + "learning_rate": 2.952164988084021e-06, + "loss": 0.6236, + "step": 1239 + }, + { + "epoch": 0.49889358278012474, + "grad_norm": 4.126996114146587, + "learning_rate": 2.9487115027340458e-06, + "loss": 0.5455, + "step": 1240 + }, + { + "epoch": 0.49929591631462483, + "grad_norm": 3.978032878939077, + "learning_rate": 2.9452571324011103e-06, + "loss": 0.5658, + "step": 1241 + }, + { + "epoch": 0.4996982498491249, + "grad_norm": 4.494778767380183, + "learning_rate": 2.941801883898187e-06, + "loss": 0.6639, + "step": 1242 + }, + { + "epoch": 0.5001005833836251, + "grad_norm": 3.7291179386624376, + "learning_rate": 2.9383457640399794e-06, + "loss": 0.4724, + "step": 1243 + }, + { + "epoch": 0.5005029169181251, + "grad_norm": 4.0158904670389175, + "learning_rate": 2.9348887796429104e-06, + "loss": 0.5611, + "step": 1244 + }, + { + "epoch": 0.5009052504526252, + "grad_norm": 3.93870398764487, + "learning_rate": 2.9314309375251065e-06, + "loss": 0.5596, + "step": 1245 + }, + { + "epoch": 0.5013075839871253, + "grad_norm": 4.111802544399985, + "learning_rate": 2.9279722445063883e-06, + "loss": 0.5872, + "step": 1246 + }, + { + "epoch": 0.5017099175216254, + "grad_norm": 3.948101448350297, + "learning_rate": 2.9245127074082535e-06, + "loss": 0.5131, + "step": 1247 + }, + { + "epoch": 0.5021122510561256, + "grad_norm": 4.345129844652909, + "learning_rate": 2.9210523330538625e-06, + "loss": 0.6389, + "step": 1248 + }, + { + "epoch": 0.5025145845906256, + "grad_norm": 3.6367326824064112, + "learning_rate": 2.9175911282680307e-06, + "loss": 0.4518, + "step": 1249 + }, + { + "epoch": 0.5029169181251257, + "grad_norm": 4.027476673290246, + "learning_rate": 2.914129099877208e-06, + "loss": 0.5234, + "step": 1250 + }, + { + "epoch": 0.5033192516596259, + "grad_norm": 3.450546244138055, + "learning_rate": 2.9106662547094724e-06, + "loss": 0.3807, + "step": 1251 + }, + { + "epoch": 0.5037215851941259, + "grad_norm": 3.696528336339893, + "learning_rate": 2.907202599594508e-06, + "loss": 0.3884, + "step": 1252 + }, + { + "epoch": 0.504123918728626, + "grad_norm": 4.252344685184106, + "learning_rate": 2.9037381413636e-06, + "loss": 0.6935, + "step": 1253 + }, + { + "epoch": 0.5045262522631261, + "grad_norm": 4.336362311318111, + "learning_rate": 2.900272886849616e-06, + "loss": 0.5331, + "step": 1254 + }, + { + "epoch": 0.5049285857976262, + "grad_norm": 3.6842567382976688, + "learning_rate": 2.8968068428869955e-06, + "loss": 0.5545, + "step": 1255 + }, + { + "epoch": 0.5053309193321264, + "grad_norm": 3.699409904804457, + "learning_rate": 2.8933400163117337e-06, + "loss": 0.3754, + "step": 1256 + }, + { + "epoch": 0.5057332528666264, + "grad_norm": 4.205702275086939, + "learning_rate": 2.889872413961369e-06, + "loss": 0.6633, + "step": 1257 + }, + { + "epoch": 0.5061355864011265, + "grad_norm": 3.6639853190462626, + "learning_rate": 2.886404042674972e-06, + "loss": 0.3892, + "step": 1258 + }, + { + "epoch": 0.5065379199356266, + "grad_norm": 3.5640332071999765, + "learning_rate": 2.882934909293127e-06, + "loss": 0.4084, + "step": 1259 + }, + { + "epoch": 0.5069402534701267, + "grad_norm": 4.0471975082058025, + "learning_rate": 2.8794650206579234e-06, + "loss": 0.5812, + "step": 1260 + }, + { + "epoch": 0.5073425870046269, + "grad_norm": 4.0654643808596616, + "learning_rate": 2.875994383612939e-06, + "loss": 0.568, + "step": 1261 + }, + { + "epoch": 0.5077449205391269, + "grad_norm": 4.330240433028303, + "learning_rate": 2.872523005003229e-06, + "loss": 0.5165, + "step": 1262 + }, + { + "epoch": 0.508147254073627, + "grad_norm": 4.015050113056623, + "learning_rate": 2.869050891675309e-06, + "loss": 0.5002, + "step": 1263 + }, + { + "epoch": 0.5085495876081272, + "grad_norm": 4.393983632365189, + "learning_rate": 2.8655780504771476e-06, + "loss": 0.573, + "step": 1264 + }, + { + "epoch": 0.5089519211426272, + "grad_norm": 3.193554396038774, + "learning_rate": 2.8621044882581433e-06, + "loss": 0.3709, + "step": 1265 + }, + { + "epoch": 0.5093542546771274, + "grad_norm": 3.3841882285013813, + "learning_rate": 2.8586302118691223e-06, + "loss": 0.4387, + "step": 1266 + }, + { + "epoch": 0.5097565882116274, + "grad_norm": 3.9970258932238742, + "learning_rate": 2.8551552281623157e-06, + "loss": 0.5225, + "step": 1267 + }, + { + "epoch": 0.5101589217461275, + "grad_norm": 4.513606814044003, + "learning_rate": 2.85167954399135e-06, + "loss": 0.5528, + "step": 1268 + }, + { + "epoch": 0.5105612552806277, + "grad_norm": 4.684727392976221, + "learning_rate": 2.848203166211235e-06, + "loss": 0.6793, + "step": 1269 + }, + { + "epoch": 0.5109635888151277, + "grad_norm": 4.222583621631339, + "learning_rate": 2.8447261016783474e-06, + "loss": 0.6053, + "step": 1270 + }, + { + "epoch": 0.5113659223496279, + "grad_norm": 3.8724976223748744, + "learning_rate": 2.841248357250418e-06, + "loss": 0.4797, + "step": 1271 + }, + { + "epoch": 0.5117682558841279, + "grad_norm": 3.5925982833624652, + "learning_rate": 2.8377699397865184e-06, + "loss": 0.6524, + "step": 1272 + }, + { + "epoch": 0.512170589418628, + "grad_norm": 4.4527497576787285, + "learning_rate": 2.8342908561470496e-06, + "loss": 0.6426, + "step": 1273 + }, + { + "epoch": 0.5125729229531282, + "grad_norm": 3.758694762638942, + "learning_rate": 2.8308111131937237e-06, + "loss": 0.4863, + "step": 1274 + }, + { + "epoch": 0.5129752564876282, + "grad_norm": 3.432356742958401, + "learning_rate": 2.827330717789555e-06, + "loss": 0.5223, + "step": 1275 + }, + { + "epoch": 0.5133775900221283, + "grad_norm": 4.33004996309828, + "learning_rate": 2.8238496767988433e-06, + "loss": 0.6004, + "step": 1276 + }, + { + "epoch": 0.5137799235566285, + "grad_norm": 4.200608717420535, + "learning_rate": 2.820367997087164e-06, + "loss": 0.6228, + "step": 1277 + }, + { + "epoch": 0.5141822570911285, + "grad_norm": 4.0453448815375035, + "learning_rate": 2.8168856855213493e-06, + "loss": 0.6242, + "step": 1278 + }, + { + "epoch": 0.5145845906256287, + "grad_norm": 4.227997150733838, + "learning_rate": 2.81340274896948e-06, + "loss": 0.6175, + "step": 1279 + }, + { + "epoch": 0.5149869241601287, + "grad_norm": 3.68997651052952, + "learning_rate": 2.8099191943008673e-06, + "loss": 0.3891, + "step": 1280 + }, + { + "epoch": 0.5153892576946288, + "grad_norm": 3.983876769382806, + "learning_rate": 2.806435028386044e-06, + "loss": 0.4676, + "step": 1281 + }, + { + "epoch": 0.515791591229129, + "grad_norm": 3.7821657559918362, + "learning_rate": 2.8029502580967467e-06, + "loss": 0.5712, + "step": 1282 + }, + { + "epoch": 0.516193924763629, + "grad_norm": 4.613891160258615, + "learning_rate": 2.7994648903059034e-06, + "loss": 0.6892, + "step": 1283 + }, + { + "epoch": 0.5165962582981292, + "grad_norm": 4.260048115855587, + "learning_rate": 2.7959789318876235e-06, + "loss": 0.587, + "step": 1284 + }, + { + "epoch": 0.5169985918326292, + "grad_norm": 3.2651956443139905, + "learning_rate": 2.792492389717178e-06, + "loss": 0.4202, + "step": 1285 + }, + { + "epoch": 0.5174009253671293, + "grad_norm": 3.8441567338972136, + "learning_rate": 2.7890052706709913e-06, + "loss": 0.5274, + "step": 1286 + }, + { + "epoch": 0.5178032589016295, + "grad_norm": 4.244436141220556, + "learning_rate": 2.785517581626624e-06, + "loss": 0.5557, + "step": 1287 + }, + { + "epoch": 0.5182055924361295, + "grad_norm": 3.777185766227343, + "learning_rate": 2.782029329462763e-06, + "loss": 0.5508, + "step": 1288 + }, + { + "epoch": 0.5186079259706297, + "grad_norm": 4.024624316501224, + "learning_rate": 2.778540521059204e-06, + "loss": 0.4091, + "step": 1289 + }, + { + "epoch": 0.5190102595051298, + "grad_norm": 3.429656579938711, + "learning_rate": 2.775051163296839e-06, + "loss": 0.373, + "step": 1290 + }, + { + "epoch": 0.5194125930396298, + "grad_norm": 4.115516308278128, + "learning_rate": 2.771561263057647e-06, + "loss": 0.6296, + "step": 1291 + }, + { + "epoch": 0.51981492657413, + "grad_norm": 4.5340536303593, + "learning_rate": 2.7680708272246738e-06, + "loss": 0.5478, + "step": 1292 + }, + { + "epoch": 0.52021726010863, + "grad_norm": 4.1573055283514675, + "learning_rate": 2.7645798626820226e-06, + "loss": 0.5928, + "step": 1293 + }, + { + "epoch": 0.5206195936431302, + "grad_norm": 3.5540381627057926, + "learning_rate": 2.7610883763148382e-06, + "loss": 0.4204, + "step": 1294 + }, + { + "epoch": 0.5210219271776303, + "grad_norm": 3.7082383118644406, + "learning_rate": 2.7575963750092976e-06, + "loss": 0.5897, + "step": 1295 + }, + { + "epoch": 0.5214242607121303, + "grad_norm": 3.670768517938044, + "learning_rate": 2.7541038656525906e-06, + "loss": 0.4524, + "step": 1296 + }, + { + "epoch": 0.5218265942466305, + "grad_norm": 3.9599618494551336, + "learning_rate": 2.7506108551329084e-06, + "loss": 0.5334, + "step": 1297 + }, + { + "epoch": 0.5222289277811305, + "grad_norm": 3.8332760152691425, + "learning_rate": 2.747117350339434e-06, + "loss": 0.5027, + "step": 1298 + }, + { + "epoch": 0.5226312613156306, + "grad_norm": 3.5387447785248463, + "learning_rate": 2.743623358162322e-06, + "loss": 0.4453, + "step": 1299 + }, + { + "epoch": 0.5230335948501308, + "grad_norm": 3.665534245478194, + "learning_rate": 2.74012888549269e-06, + "loss": 0.4541, + "step": 1300 + }, + { + "epoch": 0.5234359283846308, + "grad_norm": 4.118363799618076, + "learning_rate": 2.736633939222604e-06, + "loss": 0.4863, + "step": 1301 + }, + { + "epoch": 0.523838261919131, + "grad_norm": 4.07159865432201, + "learning_rate": 2.733138526245061e-06, + "loss": 0.5836, + "step": 1302 + }, + { + "epoch": 0.5242405954536311, + "grad_norm": 4.9413620392110955, + "learning_rate": 2.729642653453981e-06, + "loss": 0.6331, + "step": 1303 + }, + { + "epoch": 0.5246429289881311, + "grad_norm": 4.349366316003809, + "learning_rate": 2.7261463277441913e-06, + "loss": 0.6584, + "step": 1304 + }, + { + "epoch": 0.5250452625226313, + "grad_norm": 4.013818084372327, + "learning_rate": 2.7226495560114097e-06, + "loss": 0.6444, + "step": 1305 + }, + { + "epoch": 0.5254475960571313, + "grad_norm": 3.9731448823583158, + "learning_rate": 2.719152345152237e-06, + "loss": 0.4717, + "step": 1306 + }, + { + "epoch": 0.5258499295916315, + "grad_norm": 3.836192951376196, + "learning_rate": 2.7156547020641377e-06, + "loss": 0.4813, + "step": 1307 + }, + { + "epoch": 0.5262522631261316, + "grad_norm": 4.526603353793837, + "learning_rate": 2.7121566336454295e-06, + "loss": 0.6223, + "step": 1308 + }, + { + "epoch": 0.5266545966606316, + "grad_norm": 4.542236066581728, + "learning_rate": 2.70865814679527e-06, + "loss": 0.541, + "step": 1309 + }, + { + "epoch": 0.5270569301951318, + "grad_norm": 4.477320709723044, + "learning_rate": 2.70515924841364e-06, + "loss": 0.4898, + "step": 1310 + }, + { + "epoch": 0.5274592637296318, + "grad_norm": 4.188969544607867, + "learning_rate": 2.7016599454013324e-06, + "loss": 0.478, + "step": 1311 + }, + { + "epoch": 0.527861597264132, + "grad_norm": 3.57921084312417, + "learning_rate": 2.6981602446599408e-06, + "loss": 0.4797, + "step": 1312 + }, + { + "epoch": 0.5282639307986321, + "grad_norm": 3.4828294543575784, + "learning_rate": 2.69466015309184e-06, + "loss": 0.4464, + "step": 1313 + }, + { + "epoch": 0.5286662643331321, + "grad_norm": 3.6542821945407002, + "learning_rate": 2.691159677600177e-06, + "loss": 0.5241, + "step": 1314 + }, + { + "epoch": 0.5290685978676323, + "grad_norm": 4.138582178147402, + "learning_rate": 2.6876588250888556e-06, + "loss": 0.4976, + "step": 1315 + }, + { + "epoch": 0.5294709314021324, + "grad_norm": 4.156252344667509, + "learning_rate": 2.6841576024625247e-06, + "loss": 0.5346, + "step": 1316 + }, + { + "epoch": 0.5298732649366324, + "grad_norm": 3.9783539180873118, + "learning_rate": 2.680656016626561e-06, + "loss": 0.5054, + "step": 1317 + }, + { + "epoch": 0.5302755984711326, + "grad_norm": 4.167749289113631, + "learning_rate": 2.677154074487057e-06, + "loss": 0.5308, + "step": 1318 + }, + { + "epoch": 0.5306779320056326, + "grad_norm": 3.6243864679157536, + "learning_rate": 2.6736517829508124e-06, + "loss": 0.4969, + "step": 1319 + }, + { + "epoch": 0.5310802655401328, + "grad_norm": 4.080652153118834, + "learning_rate": 2.6701491489253107e-06, + "loss": 0.5166, + "step": 1320 + }, + { + "epoch": 0.5314825990746329, + "grad_norm": 4.386897362383017, + "learning_rate": 2.6666461793187143e-06, + "loss": 0.4953, + "step": 1321 + }, + { + "epoch": 0.5318849326091329, + "grad_norm": 4.286991486124671, + "learning_rate": 2.6631428810398453e-06, + "loss": 0.4804, + "step": 1322 + }, + { + "epoch": 0.5322872661436331, + "grad_norm": 3.996180824655324, + "learning_rate": 2.659639260998176e-06, + "loss": 0.5182, + "step": 1323 + }, + { + "epoch": 0.5326895996781331, + "grad_norm": 4.305112080018121, + "learning_rate": 2.6561353261038127e-06, + "loss": 0.6548, + "step": 1324 + }, + { + "epoch": 0.5330919332126333, + "grad_norm": 3.6190226327616557, + "learning_rate": 2.6526310832674807e-06, + "loss": 0.5383, + "step": 1325 + }, + { + "epoch": 0.5334942667471334, + "grad_norm": 4.316996697367821, + "learning_rate": 2.6491265394005157e-06, + "loss": 0.5516, + "step": 1326 + }, + { + "epoch": 0.5338966002816334, + "grad_norm": 4.0648368660213325, + "learning_rate": 2.645621701414845e-06, + "loss": 0.6482, + "step": 1327 + }, + { + "epoch": 0.5342989338161336, + "grad_norm": 3.7739876249793047, + "learning_rate": 2.6421165762229777e-06, + "loss": 0.4783, + "step": 1328 + }, + { + "epoch": 0.5347012673506337, + "grad_norm": 3.789987275308021, + "learning_rate": 2.6386111707379873e-06, + "loss": 0.4525, + "step": 1329 + }, + { + "epoch": 0.5351036008851338, + "grad_norm": 3.630575939542573, + "learning_rate": 2.635105491873502e-06, + "loss": 0.5329, + "step": 1330 + }, + { + "epoch": 0.5355059344196339, + "grad_norm": 3.961557081061209, + "learning_rate": 2.6315995465436878e-06, + "loss": 0.5432, + "step": 1331 + }, + { + "epoch": 0.5359082679541339, + "grad_norm": 3.8780055120857204, + "learning_rate": 2.628093341663236e-06, + "loss": 0.454, + "step": 1332 + }, + { + "epoch": 0.5363106014886341, + "grad_norm": 4.031126634163913, + "learning_rate": 2.6245868841473525e-06, + "loss": 0.5452, + "step": 1333 + }, + { + "epoch": 0.5367129350231342, + "grad_norm": 3.3788818500452997, + "learning_rate": 2.6210801809117377e-06, + "loss": 0.4364, + "step": 1334 + }, + { + "epoch": 0.5371152685576343, + "grad_norm": 4.566291916332284, + "learning_rate": 2.61757323887258e-06, + "loss": 0.6427, + "step": 1335 + }, + { + "epoch": 0.5375176020921344, + "grad_norm": 4.46318346856033, + "learning_rate": 2.614066064946536e-06, + "loss": 0.6142, + "step": 1336 + }, + { + "epoch": 0.5379199356266344, + "grad_norm": 3.8724650951472377, + "learning_rate": 2.6105586660507214e-06, + "loss": 0.4865, + "step": 1337 + }, + { + "epoch": 0.5383222691611346, + "grad_norm": 4.107266394665508, + "learning_rate": 2.607051049102696e-06, + "loss": 0.5618, + "step": 1338 + }, + { + "epoch": 0.5387246026956347, + "grad_norm": 4.131578269950126, + "learning_rate": 2.6035432210204475e-06, + "loss": 0.6571, + "step": 1339 + }, + { + "epoch": 0.5391269362301347, + "grad_norm": 4.135237490425349, + "learning_rate": 2.6000351887223817e-06, + "loss": 0.506, + "step": 1340 + }, + { + "epoch": 0.5395292697646349, + "grad_norm": 4.2799950491642695, + "learning_rate": 2.5965269591273075e-06, + "loss": 0.5054, + "step": 1341 + }, + { + "epoch": 0.539931603299135, + "grad_norm": 4.111277509443795, + "learning_rate": 2.59301853915442e-06, + "loss": 0.5896, + "step": 1342 + }, + { + "epoch": 0.5403339368336351, + "grad_norm": 3.8507472387773416, + "learning_rate": 2.5895099357232946e-06, + "loss": 0.4784, + "step": 1343 + }, + { + "epoch": 0.5407362703681352, + "grad_norm": 4.553167071691355, + "learning_rate": 2.5860011557538644e-06, + "loss": 0.5889, + "step": 1344 + }, + { + "epoch": 0.5411386039026352, + "grad_norm": 4.8377815361815175, + "learning_rate": 2.5824922061664125e-06, + "loss": 0.7153, + "step": 1345 + }, + { + "epoch": 0.5415409374371354, + "grad_norm": 4.026163246451428, + "learning_rate": 2.578983093881557e-06, + "loss": 0.5578, + "step": 1346 + }, + { + "epoch": 0.5419432709716355, + "grad_norm": 4.396870764657521, + "learning_rate": 2.5754738258202345e-06, + "loss": 0.7799, + "step": 1347 + }, + { + "epoch": 0.5423456045061356, + "grad_norm": 4.349518153263542, + "learning_rate": 2.5719644089036916e-06, + "loss": 0.5919, + "step": 1348 + }, + { + "epoch": 0.5427479380406357, + "grad_norm": 4.355976508524607, + "learning_rate": 2.568454850053467e-06, + "loss": 0.5903, + "step": 1349 + }, + { + "epoch": 0.5431502715751357, + "grad_norm": 3.951612074252993, + "learning_rate": 2.5649451561913796e-06, + "loss": 0.5038, + "step": 1350 + }, + { + "epoch": 0.5435526051096359, + "grad_norm": 4.030293462511148, + "learning_rate": 2.561435334239515e-06, + "loss": 0.5308, + "step": 1351 + }, + { + "epoch": 0.543954938644136, + "grad_norm": 4.530833691231785, + "learning_rate": 2.55792539112021e-06, + "loss": 0.5977, + "step": 1352 + }, + { + "epoch": 0.544357272178636, + "grad_norm": 4.1989735383766185, + "learning_rate": 2.5544153337560422e-06, + "loss": 0.5686, + "step": 1353 + }, + { + "epoch": 0.5447596057131362, + "grad_norm": 4.32105387076282, + "learning_rate": 2.5509051690698137e-06, + "loss": 0.6257, + "step": 1354 + }, + { + "epoch": 0.5451619392476363, + "grad_norm": 3.34919768311592, + "learning_rate": 2.5473949039845388e-06, + "loss": 0.4131, + "step": 1355 + }, + { + "epoch": 0.5455642727821364, + "grad_norm": 4.128988896294745, + "learning_rate": 2.5438845454234284e-06, + "loss": 0.6007, + "step": 1356 + }, + { + "epoch": 0.5459666063166365, + "grad_norm": 4.122788652813818, + "learning_rate": 2.5403741003098793e-06, + "loss": 0.5223, + "step": 1357 + }, + { + "epoch": 0.5463689398511365, + "grad_norm": 4.420118909526267, + "learning_rate": 2.5368635755674586e-06, + "loss": 0.5506, + "step": 1358 + }, + { + "epoch": 0.5467712733856367, + "grad_norm": 4.2195230085645345, + "learning_rate": 2.5333529781198895e-06, + "loss": 0.5989, + "step": 1359 + }, + { + "epoch": 0.5471736069201368, + "grad_norm": 4.20795164503954, + "learning_rate": 2.5298423148910396e-06, + "loss": 0.4462, + "step": 1360 + }, + { + "epoch": 0.5475759404546369, + "grad_norm": 4.010223353178494, + "learning_rate": 2.5263315928049063e-06, + "loss": 0.5192, + "step": 1361 + }, + { + "epoch": 0.547978273989137, + "grad_norm": 4.816701841105469, + "learning_rate": 2.5228208187856024e-06, + "loss": 0.6599, + "step": 1362 + }, + { + "epoch": 0.548380607523637, + "grad_norm": 4.023413458632839, + "learning_rate": 2.5193099997573437e-06, + "loss": 0.6185, + "step": 1363 + }, + { + "epoch": 0.5487829410581372, + "grad_norm": 3.6516400727809115, + "learning_rate": 2.5157991426444343e-06, + "loss": 0.5859, + "step": 1364 + }, + { + "epoch": 0.5491852745926373, + "grad_norm": 4.119751084462459, + "learning_rate": 2.5122882543712546e-06, + "loss": 0.5255, + "step": 1365 + }, + { + "epoch": 0.5495876081271374, + "grad_norm": 3.957835467170672, + "learning_rate": 2.5087773418622447e-06, + "loss": 0.4659, + "step": 1366 + }, + { + "epoch": 0.5499899416616375, + "grad_norm": 3.500258182406765, + "learning_rate": 2.505266412041893e-06, + "loss": 0.4385, + "step": 1367 + }, + { + "epoch": 0.5503922751961376, + "grad_norm": 4.041408070624076, + "learning_rate": 2.5017554718347233e-06, + "loss": 0.5367, + "step": 1368 + }, + { + "epoch": 0.5507946087306377, + "grad_norm": 4.091765389792879, + "learning_rate": 2.4982445281652775e-06, + "loss": 0.5744, + "step": 1369 + }, + { + "epoch": 0.5511969422651378, + "grad_norm": 4.469717920616285, + "learning_rate": 2.4947335879581073e-06, + "loss": 0.5818, + "step": 1370 + }, + { + "epoch": 0.5515992757996379, + "grad_norm": 3.9779088164073833, + "learning_rate": 2.4912226581377566e-06, + "loss": 0.5893, + "step": 1371 + }, + { + "epoch": 0.552001609334138, + "grad_norm": 3.627034776419984, + "learning_rate": 2.4877117456287463e-06, + "loss": 0.4902, + "step": 1372 + }, + { + "epoch": 0.5524039428686381, + "grad_norm": 3.8562269767576405, + "learning_rate": 2.484200857355566e-06, + "loss": 0.6268, + "step": 1373 + }, + { + "epoch": 0.5528062764031382, + "grad_norm": 4.192164362531014, + "learning_rate": 2.4806900002426567e-06, + "loss": 0.6121, + "step": 1374 + }, + { + "epoch": 0.5532086099376383, + "grad_norm": 3.916801705190097, + "learning_rate": 2.477179181214398e-06, + "loss": 0.4926, + "step": 1375 + }, + { + "epoch": 0.5536109434721384, + "grad_norm": 3.693356946438789, + "learning_rate": 2.473668407195095e-06, + "loss": 0.5189, + "step": 1376 + }, + { + "epoch": 0.5540132770066385, + "grad_norm": 4.229576715855437, + "learning_rate": 2.4701576851089613e-06, + "loss": 0.6398, + "step": 1377 + }, + { + "epoch": 0.5544156105411386, + "grad_norm": 3.6167470608574197, + "learning_rate": 2.4666470218801113e-06, + "loss": 0.5175, + "step": 1378 + }, + { + "epoch": 0.5548179440756387, + "grad_norm": 5.070171032981664, + "learning_rate": 2.463136424432542e-06, + "loss": 0.4988, + "step": 1379 + }, + { + "epoch": 0.5552202776101388, + "grad_norm": 4.058368062649386, + "learning_rate": 2.4596258996901215e-06, + "loss": 0.4313, + "step": 1380 + }, + { + "epoch": 0.555622611144639, + "grad_norm": 3.9184186874819464, + "learning_rate": 2.4561154545765724e-06, + "loss": 0.5338, + "step": 1381 + }, + { + "epoch": 0.556024944679139, + "grad_norm": 3.1244549306950784, + "learning_rate": 2.4526050960154616e-06, + "loss": 0.4289, + "step": 1382 + }, + { + "epoch": 0.5564272782136391, + "grad_norm": 4.04400253855949, + "learning_rate": 2.4490948309301867e-06, + "loss": 0.5257, + "step": 1383 + }, + { + "epoch": 0.5568296117481392, + "grad_norm": 3.9303715740181513, + "learning_rate": 2.445584666243959e-06, + "loss": 0.5149, + "step": 1384 + }, + { + "epoch": 0.5572319452826393, + "grad_norm": 4.395655391522824, + "learning_rate": 2.442074608879791e-06, + "loss": 0.6249, + "step": 1385 + }, + { + "epoch": 0.5576342788171395, + "grad_norm": 3.8268439749597576, + "learning_rate": 2.438564665760486e-06, + "loss": 0.5613, + "step": 1386 + }, + { + "epoch": 0.5580366123516395, + "grad_norm": 4.255793020118025, + "learning_rate": 2.4350548438086212e-06, + "loss": 0.6155, + "step": 1387 + }, + { + "epoch": 0.5584389458861396, + "grad_norm": 3.979739013550483, + "learning_rate": 2.4315451499465336e-06, + "loss": 0.5394, + "step": 1388 + }, + { + "epoch": 0.5588412794206397, + "grad_norm": 3.781411432288367, + "learning_rate": 2.4280355910963097e-06, + "loss": 0.4507, + "step": 1389 + }, + { + "epoch": 0.5592436129551398, + "grad_norm": 4.371781717934846, + "learning_rate": 2.4245261741797663e-06, + "loss": 0.5305, + "step": 1390 + }, + { + "epoch": 0.55964594648964, + "grad_norm": 3.52895283713103, + "learning_rate": 2.421016906118444e-06, + "loss": 0.5194, + "step": 1391 + }, + { + "epoch": 0.56004828002414, + "grad_norm": 3.872042926144624, + "learning_rate": 2.417507793833587e-06, + "loss": 0.4998, + "step": 1392 + }, + { + "epoch": 0.5604506135586401, + "grad_norm": 3.391234256928363, + "learning_rate": 2.413998844246136e-06, + "loss": 0.4093, + "step": 1393 + }, + { + "epoch": 0.5608529470931403, + "grad_norm": 3.9654243758806174, + "learning_rate": 2.4104900642767066e-06, + "loss": 0.4515, + "step": 1394 + }, + { + "epoch": 0.5612552806276403, + "grad_norm": 3.5879075344069378, + "learning_rate": 2.4069814608455804e-06, + "loss": 0.5207, + "step": 1395 + }, + { + "epoch": 0.5616576141621404, + "grad_norm": 4.104094820028554, + "learning_rate": 2.4034730408726938e-06, + "loss": 0.5625, + "step": 1396 + }, + { + "epoch": 0.5620599476966405, + "grad_norm": 3.772298699595217, + "learning_rate": 2.3999648112776183e-06, + "loss": 0.5068, + "step": 1397 + }, + { + "epoch": 0.5624622812311406, + "grad_norm": 3.688052093132043, + "learning_rate": 2.3964567789795533e-06, + "loss": 0.5042, + "step": 1398 + }, + { + "epoch": 0.5628646147656408, + "grad_norm": 3.8931716258044498, + "learning_rate": 2.3929489508973047e-06, + "loss": 0.5485, + "step": 1399 + }, + { + "epoch": 0.5632669483001408, + "grad_norm": 4.224728616738161, + "learning_rate": 2.389441333949279e-06, + "loss": 0.6582, + "step": 1400 + }, + { + "epoch": 0.5636692818346409, + "grad_norm": 3.8494878180062897, + "learning_rate": 2.3859339350534646e-06, + "loss": 0.5379, + "step": 1401 + }, + { + "epoch": 0.564071615369141, + "grad_norm": 4.200325419085908, + "learning_rate": 2.382426761127421e-06, + "loss": 0.5997, + "step": 1402 + }, + { + "epoch": 0.5644739489036411, + "grad_norm": 4.143252972394019, + "learning_rate": 2.378919819088263e-06, + "loss": 0.6195, + "step": 1403 + }, + { + "epoch": 0.5648762824381413, + "grad_norm": 3.7174533902571145, + "learning_rate": 2.3754131158526483e-06, + "loss": 0.5317, + "step": 1404 + }, + { + "epoch": 0.5652786159726413, + "grad_norm": 4.0056931921159356, + "learning_rate": 2.371906658336764e-06, + "loss": 0.5232, + "step": 1405 + }, + { + "epoch": 0.5656809495071414, + "grad_norm": 3.830247393182765, + "learning_rate": 2.368400453456313e-06, + "loss": 0.4898, + "step": 1406 + }, + { + "epoch": 0.5660832830416416, + "grad_norm": 3.7820185066794787, + "learning_rate": 2.364894508126499e-06, + "loss": 0.5438, + "step": 1407 + }, + { + "epoch": 0.5664856165761416, + "grad_norm": 4.4898443665627, + "learning_rate": 2.3613888292620135e-06, + "loss": 0.6322, + "step": 1408 + }, + { + "epoch": 0.5668879501106417, + "grad_norm": 4.157236186991902, + "learning_rate": 2.3578834237770227e-06, + "loss": 0.5978, + "step": 1409 + }, + { + "epoch": 0.5672902836451418, + "grad_norm": 4.576188931404494, + "learning_rate": 2.354378298585155e-06, + "loss": 0.5774, + "step": 1410 + }, + { + "epoch": 0.5676926171796419, + "grad_norm": 3.714487248907507, + "learning_rate": 2.3508734605994855e-06, + "loss": 0.4305, + "step": 1411 + }, + { + "epoch": 0.5680949507141421, + "grad_norm": 3.9270791753173313, + "learning_rate": 2.3473689167325205e-06, + "loss": 0.4912, + "step": 1412 + }, + { + "epoch": 0.5684972842486421, + "grad_norm": 3.655781362606588, + "learning_rate": 2.3438646738961886e-06, + "loss": 0.5125, + "step": 1413 + }, + { + "epoch": 0.5688996177831422, + "grad_norm": 4.2097153394323845, + "learning_rate": 2.3403607390018246e-06, + "loss": 0.4273, + "step": 1414 + }, + { + "epoch": 0.5693019513176423, + "grad_norm": 3.7418997500983, + "learning_rate": 2.3368571189601556e-06, + "loss": 0.4534, + "step": 1415 + }, + { + "epoch": 0.5697042848521424, + "grad_norm": 3.7385152001909154, + "learning_rate": 2.3333538206812874e-06, + "loss": 0.4461, + "step": 1416 + }, + { + "epoch": 0.5701066183866426, + "grad_norm": 4.243485694858217, + "learning_rate": 2.32985085107469e-06, + "loss": 0.4943, + "step": 1417 + }, + { + "epoch": 0.5705089519211426, + "grad_norm": 4.064181274521891, + "learning_rate": 2.3263482170491884e-06, + "loss": 0.6696, + "step": 1418 + }, + { + "epoch": 0.5709112854556427, + "grad_norm": 4.168819688893492, + "learning_rate": 2.3228459255129425e-06, + "loss": 0.5914, + "step": 1419 + }, + { + "epoch": 0.5713136189901429, + "grad_norm": 3.299064883781313, + "learning_rate": 2.3193439833734403e-06, + "loss": 0.3987, + "step": 1420 + }, + { + "epoch": 0.5717159525246429, + "grad_norm": 4.887635853989916, + "learning_rate": 2.3158423975374757e-06, + "loss": 0.4842, + "step": 1421 + }, + { + "epoch": 0.572118286059143, + "grad_norm": 3.4644648967520806, + "learning_rate": 2.312341174911145e-06, + "loss": 0.4643, + "step": 1422 + }, + { + "epoch": 0.5725206195936431, + "grad_norm": 3.75941250297541, + "learning_rate": 2.3088403223998235e-06, + "loss": 0.4772, + "step": 1423 + }, + { + "epoch": 0.5729229531281432, + "grad_norm": 3.4615485090255578, + "learning_rate": 2.3053398469081613e-06, + "loss": 0.5075, + "step": 1424 + }, + { + "epoch": 0.5733252866626434, + "grad_norm": 4.2227564973481995, + "learning_rate": 2.30183975534006e-06, + "loss": 0.5125, + "step": 1425 + }, + { + "epoch": 0.5737276201971434, + "grad_norm": 4.090888673139273, + "learning_rate": 2.298340054598668e-06, + "loss": 0.5183, + "step": 1426 + }, + { + "epoch": 0.5741299537316435, + "grad_norm": 4.548845256935699, + "learning_rate": 2.294840751586361e-06, + "loss": 0.639, + "step": 1427 + }, + { + "epoch": 0.5745322872661436, + "grad_norm": 3.901263423192965, + "learning_rate": 2.291341853204731e-06, + "loss": 0.623, + "step": 1428 + }, + { + "epoch": 0.5749346208006437, + "grad_norm": 3.347357658304439, + "learning_rate": 2.287843366354571e-06, + "loss": 0.4009, + "step": 1429 + }, + { + "epoch": 0.5753369543351439, + "grad_norm": 4.512229941763247, + "learning_rate": 2.284345297935863e-06, + "loss": 0.6336, + "step": 1430 + }, + { + "epoch": 0.5757392878696439, + "grad_norm": 4.132485335550154, + "learning_rate": 2.2808476548477636e-06, + "loss": 0.5015, + "step": 1431 + }, + { + "epoch": 0.576141621404144, + "grad_norm": 4.226000440762799, + "learning_rate": 2.2773504439885903e-06, + "loss": 0.5063, + "step": 1432 + }, + { + "epoch": 0.5765439549386442, + "grad_norm": 4.010604789353783, + "learning_rate": 2.27385367225581e-06, + "loss": 0.5085, + "step": 1433 + }, + { + "epoch": 0.5769462884731442, + "grad_norm": 3.71838877160847, + "learning_rate": 2.2703573465460194e-06, + "loss": 0.483, + "step": 1434 + }, + { + "epoch": 0.5773486220076444, + "grad_norm": 3.715720870136391, + "learning_rate": 2.2668614737549395e-06, + "loss": 0.5292, + "step": 1435 + }, + { + "epoch": 0.5777509555421444, + "grad_norm": 3.7875519229722756, + "learning_rate": 2.263366060777397e-06, + "loss": 0.4504, + "step": 1436 + }, + { + "epoch": 0.5781532890766445, + "grad_norm": 4.235567100677654, + "learning_rate": 2.25987111450731e-06, + "loss": 0.6075, + "step": 1437 + }, + { + "epoch": 0.5785556226111447, + "grad_norm": 4.23015897289991, + "learning_rate": 2.256376641837679e-06, + "loss": 0.6036, + "step": 1438 + }, + { + "epoch": 0.5789579561456447, + "grad_norm": 4.397376552533781, + "learning_rate": 2.2528826496605672e-06, + "loss": 0.6584, + "step": 1439 + }, + { + "epoch": 0.5793602896801449, + "grad_norm": 3.987430088200103, + "learning_rate": 2.249389144867092e-06, + "loss": 0.5312, + "step": 1440 + }, + { + "epoch": 0.5797626232146449, + "grad_norm": 4.349402058381357, + "learning_rate": 2.24589613434741e-06, + "loss": 0.4806, + "step": 1441 + }, + { + "epoch": 0.580164956749145, + "grad_norm": 4.389627692568862, + "learning_rate": 2.2424036249907032e-06, + "loss": 0.649, + "step": 1442 + }, + { + "epoch": 0.5805672902836452, + "grad_norm": 3.889422512107434, + "learning_rate": 2.238911623685162e-06, + "loss": 0.5553, + "step": 1443 + }, + { + "epoch": 0.5809696238181452, + "grad_norm": 4.1471977764513674, + "learning_rate": 2.235420137317978e-06, + "loss": 0.5527, + "step": 1444 + }, + { + "epoch": 0.5813719573526454, + "grad_norm": 3.642246289302764, + "learning_rate": 2.2319291727753266e-06, + "loss": 0.4669, + "step": 1445 + }, + { + "epoch": 0.5817742908871455, + "grad_norm": 3.7186046973608406, + "learning_rate": 2.228438736942354e-06, + "loss": 0.3984, + "step": 1446 + }, + { + "epoch": 0.5821766244216455, + "grad_norm": 4.208948818822306, + "learning_rate": 2.2249488367031612e-06, + "loss": 0.574, + "step": 1447 + }, + { + "epoch": 0.5825789579561457, + "grad_norm": 3.4944920810310633, + "learning_rate": 2.221459478940797e-06, + "loss": 0.4565, + "step": 1448 + }, + { + "epoch": 0.5829812914906457, + "grad_norm": 3.8209272903130325, + "learning_rate": 2.2179706705372377e-06, + "loss": 0.4116, + "step": 1449 + }, + { + "epoch": 0.5833836250251458, + "grad_norm": 4.562727204828236, + "learning_rate": 2.214482418373376e-06, + "loss": 0.6272, + "step": 1450 + }, + { + "epoch": 0.583785958559646, + "grad_norm": 3.986021082183279, + "learning_rate": 2.21099472932901e-06, + "loss": 0.4946, + "step": 1451 + }, + { + "epoch": 0.584188292094146, + "grad_norm": 3.619941584357637, + "learning_rate": 2.207507610282823e-06, + "loss": 0.5552, + "step": 1452 + }, + { + "epoch": 0.5845906256286462, + "grad_norm": 4.265275522318885, + "learning_rate": 2.2040210681123773e-06, + "loss": 0.5387, + "step": 1453 + }, + { + "epoch": 0.5849929591631462, + "grad_norm": 3.592671313761321, + "learning_rate": 2.2005351096940966e-06, + "loss": 0.4221, + "step": 1454 + }, + { + "epoch": 0.5853952926976463, + "grad_norm": 4.0275300106962755, + "learning_rate": 2.1970497419032546e-06, + "loss": 0.5781, + "step": 1455 + }, + { + "epoch": 0.5857976262321465, + "grad_norm": 3.8884576976533425, + "learning_rate": 2.1935649716139567e-06, + "loss": 0.5121, + "step": 1456 + }, + { + "epoch": 0.5861999597666465, + "grad_norm": 4.481785017500937, + "learning_rate": 2.1900808056991335e-06, + "loss": 0.6321, + "step": 1457 + }, + { + "epoch": 0.5866022933011467, + "grad_norm": 3.74252328127819, + "learning_rate": 2.186597251030521e-06, + "loss": 0.4935, + "step": 1458 + }, + { + "epoch": 0.5870046268356468, + "grad_norm": 3.9659714356438442, + "learning_rate": 2.183114314478651e-06, + "loss": 0.5503, + "step": 1459 + }, + { + "epoch": 0.5874069603701468, + "grad_norm": 3.435746642131932, + "learning_rate": 2.179632002912837e-06, + "loss": 0.3508, + "step": 1460 + }, + { + "epoch": 0.587809293904647, + "grad_norm": 4.32896011694038, + "learning_rate": 2.176150323201157e-06, + "loss": 0.5728, + "step": 1461 + }, + { + "epoch": 0.588211627439147, + "grad_norm": 4.1710482850436215, + "learning_rate": 2.1726692822104455e-06, + "loss": 0.5115, + "step": 1462 + }, + { + "epoch": 0.5886139609736472, + "grad_norm": 4.1244090573644145, + "learning_rate": 2.169188886806277e-06, + "loss": 0.3078, + "step": 1463 + }, + { + "epoch": 0.5890162945081473, + "grad_norm": 3.462366030557245, + "learning_rate": 2.1657091438529517e-06, + "loss": 0.472, + "step": 1464 + }, + { + "epoch": 0.5894186280426473, + "grad_norm": 3.920420130653373, + "learning_rate": 2.1622300602134824e-06, + "loss": 0.5559, + "step": 1465 + }, + { + "epoch": 0.5898209615771475, + "grad_norm": 4.574556709525141, + "learning_rate": 2.158751642749583e-06, + "loss": 0.5525, + "step": 1466 + }, + { + "epoch": 0.5902232951116475, + "grad_norm": 3.8284931733353513, + "learning_rate": 2.155273898321653e-06, + "loss": 0.5681, + "step": 1467 + }, + { + "epoch": 0.5906256286461476, + "grad_norm": 3.3784102201483677, + "learning_rate": 2.151796833788766e-06, + "loss": 0.4727, + "step": 1468 + }, + { + "epoch": 0.5910279621806478, + "grad_norm": 4.047681657794816, + "learning_rate": 2.1483204560086507e-06, + "loss": 0.5665, + "step": 1469 + }, + { + "epoch": 0.5914302957151478, + "grad_norm": 4.308357381109115, + "learning_rate": 2.144844771837685e-06, + "loss": 0.6112, + "step": 1470 + }, + { + "epoch": 0.591832629249648, + "grad_norm": 3.733873078989285, + "learning_rate": 2.1413697881308785e-06, + "loss": 0.494, + "step": 1471 + }, + { + "epoch": 0.5922349627841481, + "grad_norm": 4.130337043839704, + "learning_rate": 2.1378955117418567e-06, + "loss": 0.445, + "step": 1472 + }, + { + "epoch": 0.5926372963186481, + "grad_norm": 4.189878829812914, + "learning_rate": 2.1344219495228537e-06, + "loss": 0.569, + "step": 1473 + }, + { + "epoch": 0.5930396298531483, + "grad_norm": 3.926746164109752, + "learning_rate": 2.1309491083246915e-06, + "loss": 0.5696, + "step": 1474 + }, + { + "epoch": 0.5934419633876483, + "grad_norm": 3.688932516880435, + "learning_rate": 2.127476994996772e-06, + "loss": 0.5001, + "step": 1475 + }, + { + "epoch": 0.5938442969221485, + "grad_norm": 4.084824375606503, + "learning_rate": 2.1240056163870614e-06, + "loss": 0.613, + "step": 1476 + }, + { + "epoch": 0.5942466304566486, + "grad_norm": 4.614880824241214, + "learning_rate": 2.120534979342078e-06, + "loss": 0.5252, + "step": 1477 + }, + { + "epoch": 0.5946489639911486, + "grad_norm": 3.0888633468235014, + "learning_rate": 2.1170650907068742e-06, + "loss": 0.4531, + "step": 1478 + }, + { + "epoch": 0.5950512975256488, + "grad_norm": 4.213120019047452, + "learning_rate": 2.113595957325029e-06, + "loss": 0.6458, + "step": 1479 + }, + { + "epoch": 0.5954536310601488, + "grad_norm": 3.610258712936895, + "learning_rate": 2.1101275860386313e-06, + "loss": 0.5652, + "step": 1480 + }, + { + "epoch": 0.595855964594649, + "grad_norm": 3.901115632452379, + "learning_rate": 2.1066599836882667e-06, + "loss": 0.4731, + "step": 1481 + }, + { + "epoch": 0.5962582981291491, + "grad_norm": 4.576321666746087, + "learning_rate": 2.1031931571130053e-06, + "loss": 0.6799, + "step": 1482 + }, + { + "epoch": 0.5966606316636491, + "grad_norm": 4.188708168775149, + "learning_rate": 2.0997271131503847e-06, + "loss": 0.5172, + "step": 1483 + }, + { + "epoch": 0.5970629651981493, + "grad_norm": 3.993460741506799, + "learning_rate": 2.0962618586364005e-06, + "loss": 0.4828, + "step": 1484 + }, + { + "epoch": 0.5974652987326494, + "grad_norm": 3.977543353419007, + "learning_rate": 2.092797400405493e-06, + "loss": 0.5368, + "step": 1485 + }, + { + "epoch": 0.5978676322671495, + "grad_norm": 4.079669745022429, + "learning_rate": 2.0893337452905293e-06, + "loss": 0.4879, + "step": 1486 + }, + { + "epoch": 0.5982699658016496, + "grad_norm": 4.000645952876263, + "learning_rate": 2.0858709001227924e-06, + "loss": 0.5338, + "step": 1487 + }, + { + "epoch": 0.5986722993361496, + "grad_norm": 4.483257438703314, + "learning_rate": 2.0824088717319697e-06, + "loss": 0.5771, + "step": 1488 + }, + { + "epoch": 0.5990746328706498, + "grad_norm": 3.493611740579638, + "learning_rate": 2.0789476669461375e-06, + "loss": 0.4758, + "step": 1489 + }, + { + "epoch": 0.5994769664051499, + "grad_norm": 3.9869817394957385, + "learning_rate": 2.0754872925917465e-06, + "loss": 0.5482, + "step": 1490 + }, + { + "epoch": 0.59987929993965, + "grad_norm": 3.655771235240259, + "learning_rate": 2.072027755493612e-06, + "loss": 0.4924, + "step": 1491 + }, + { + "epoch": 0.6002816334741501, + "grad_norm": 3.959985484700295, + "learning_rate": 2.0685690624748943e-06, + "loss": 0.5653, + "step": 1492 + }, + { + "epoch": 0.6006839670086501, + "grad_norm": 3.4092513778358047, + "learning_rate": 2.0651112203570904e-06, + "loss": 0.4375, + "step": 1493 + }, + { + "epoch": 0.6010863005431503, + "grad_norm": 4.050380522574125, + "learning_rate": 2.061654235960021e-06, + "loss": 0.5627, + "step": 1494 + }, + { + "epoch": 0.6014886340776504, + "grad_norm": 4.329223167800351, + "learning_rate": 2.058198116101814e-06, + "loss": 0.5642, + "step": 1495 + }, + { + "epoch": 0.6018909676121504, + "grad_norm": 3.475966176009881, + "learning_rate": 2.0547428675988906e-06, + "loss": 0.4417, + "step": 1496 + }, + { + "epoch": 0.6022933011466506, + "grad_norm": 3.7228401404060443, + "learning_rate": 2.051288497265955e-06, + "loss": 0.4509, + "step": 1497 + }, + { + "epoch": 0.6026956346811507, + "grad_norm": 3.9298800993895284, + "learning_rate": 2.04783501191598e-06, + "loss": 0.4373, + "step": 1498 + }, + { + "epoch": 0.6030979682156508, + "grad_norm": 4.3431668642603105, + "learning_rate": 2.0443824183601917e-06, + "loss": 0.6361, + "step": 1499 + }, + { + "epoch": 0.6035003017501509, + "grad_norm": 3.759176356037184, + "learning_rate": 2.0409307234080577e-06, + "loss": 0.4214, + "step": 1500 + }, + { + "epoch": 0.6039026352846509, + "grad_norm": 3.881662137011941, + "learning_rate": 2.0374799338672723e-06, + "loss": 0.4868, + "step": 1501 + }, + { + "epoch": 0.6043049688191511, + "grad_norm": 4.786052455382573, + "learning_rate": 2.034030056543747e-06, + "loss": 0.6772, + "step": 1502 + }, + { + "epoch": 0.6047073023536512, + "grad_norm": 4.137240403825967, + "learning_rate": 2.0305810982415907e-06, + "loss": 0.4981, + "step": 1503 + }, + { + "epoch": 0.6051096358881513, + "grad_norm": 3.9245214475911947, + "learning_rate": 2.0271330657631034e-06, + "loss": 0.461, + "step": 1504 + }, + { + "epoch": 0.6055119694226514, + "grad_norm": 4.043701387123113, + "learning_rate": 2.023685965908755e-06, + "loss": 0.4861, + "step": 1505 + }, + { + "epoch": 0.6059143029571514, + "grad_norm": 4.578623188458431, + "learning_rate": 2.020239805477179e-06, + "loss": 0.5516, + "step": 1506 + }, + { + "epoch": 0.6063166364916516, + "grad_norm": 3.939503557149013, + "learning_rate": 2.0167945912651542e-06, + "loss": 0.5348, + "step": 1507 + }, + { + "epoch": 0.6067189700261517, + "grad_norm": 4.274904255142657, + "learning_rate": 2.0133503300675963e-06, + "loss": 0.5992, + "step": 1508 + }, + { + "epoch": 0.6071213035606517, + "grad_norm": 3.732143003199723, + "learning_rate": 2.0099070286775367e-06, + "loss": 0.4891, + "step": 1509 + }, + { + "epoch": 0.6075236370951519, + "grad_norm": 4.093400863247009, + "learning_rate": 2.0064646938861175e-06, + "loss": 0.4785, + "step": 1510 + }, + { + "epoch": 0.607925970629652, + "grad_norm": 3.850637017187709, + "learning_rate": 2.0030233324825734e-06, + "loss": 0.5371, + "step": 1511 + }, + { + "epoch": 0.6083283041641521, + "grad_norm": 3.726527595640385, + "learning_rate": 1.999582951254219e-06, + "loss": 0.4105, + "step": 1512 + }, + { + "epoch": 0.6087306376986522, + "grad_norm": 4.0679286236770045, + "learning_rate": 1.9961435569864366e-06, + "loss": 0.524, + "step": 1513 + }, + { + "epoch": 0.6091329712331522, + "grad_norm": 4.191017273712236, + "learning_rate": 1.99270515646266e-06, + "loss": 0.422, + "step": 1514 + }, + { + "epoch": 0.6095353047676524, + "grad_norm": 4.549567455680612, + "learning_rate": 1.9892677564643643e-06, + "loss": 0.5753, + "step": 1515 + }, + { + "epoch": 0.6099376383021525, + "grad_norm": 4.239380717200701, + "learning_rate": 1.985831363771052e-06, + "loss": 0.5251, + "step": 1516 + }, + { + "epoch": 0.6103399718366526, + "grad_norm": 3.8382650524932433, + "learning_rate": 1.9823959851602387e-06, + "loss": 0.6749, + "step": 1517 + }, + { + "epoch": 0.6107423053711527, + "grad_norm": 4.058628419689791, + "learning_rate": 1.9789616274074375e-06, + "loss": 0.5202, + "step": 1518 + }, + { + "epoch": 0.6111446389056527, + "grad_norm": 4.1466696163681815, + "learning_rate": 1.9755282972861515e-06, + "loss": 0.4595, + "step": 1519 + }, + { + "epoch": 0.6115469724401529, + "grad_norm": 4.193830162309673, + "learning_rate": 1.9720960015678555e-06, + "loss": 0.485, + "step": 1520 + }, + { + "epoch": 0.611949305974653, + "grad_norm": 4.239153001501294, + "learning_rate": 1.9686647470219835e-06, + "loss": 0.6782, + "step": 1521 + }, + { + "epoch": 0.6123516395091531, + "grad_norm": 4.138990574689833, + "learning_rate": 1.965234540415918e-06, + "loss": 0.6284, + "step": 1522 + }, + { + "epoch": 0.6127539730436532, + "grad_norm": 4.079634170381014, + "learning_rate": 1.9618053885149715e-06, + "loss": 0.4656, + "step": 1523 + }, + { + "epoch": 0.6131563065781533, + "grad_norm": 3.9398678794833204, + "learning_rate": 1.9583772980823797e-06, + "loss": 0.4345, + "step": 1524 + }, + { + "epoch": 0.6135586401126534, + "grad_norm": 3.8999859116475237, + "learning_rate": 1.9549502758792825e-06, + "loss": 0.4463, + "step": 1525 + }, + { + "epoch": 0.6139609736471535, + "grad_norm": 3.8236374179780652, + "learning_rate": 1.9515243286647155e-06, + "loss": 0.5032, + "step": 1526 + }, + { + "epoch": 0.6143633071816536, + "grad_norm": 3.9105069856639787, + "learning_rate": 1.9480994631955904e-06, + "loss": 0.5249, + "step": 1527 + }, + { + "epoch": 0.6147656407161537, + "grad_norm": 4.090697174474606, + "learning_rate": 1.944675686226688e-06, + "loss": 0.5986, + "step": 1528 + }, + { + "epoch": 0.6151679742506538, + "grad_norm": 4.026970204591501, + "learning_rate": 1.941253004510642e-06, + "loss": 0.5695, + "step": 1529 + }, + { + "epoch": 0.6155703077851539, + "grad_norm": 3.649099842703245, + "learning_rate": 1.937831424797926e-06, + "loss": 0.436, + "step": 1530 + }, + { + "epoch": 0.615972641319654, + "grad_norm": 3.8376284628790773, + "learning_rate": 1.9344109538368388e-06, + "loss": 0.4986, + "step": 1531 + }, + { + "epoch": 0.616374974854154, + "grad_norm": 4.048438236221887, + "learning_rate": 1.9309915983734933e-06, + "loss": 0.5608, + "step": 1532 + }, + { + "epoch": 0.6167773083886542, + "grad_norm": 4.036284880877361, + "learning_rate": 1.9275733651518038e-06, + "loss": 0.552, + "step": 1533 + }, + { + "epoch": 0.6171796419231543, + "grad_norm": 4.098386930913388, + "learning_rate": 1.9241562609134694e-06, + "loss": 0.5564, + "step": 1534 + }, + { + "epoch": 0.6175819754576544, + "grad_norm": 4.4349116098990065, + "learning_rate": 1.9207402923979633e-06, + "loss": 0.5937, + "step": 1535 + }, + { + "epoch": 0.6179843089921545, + "grad_norm": 3.9955554180753556, + "learning_rate": 1.9173254663425165e-06, + "loss": 0.4773, + "step": 1536 + }, + { + "epoch": 0.6183866425266547, + "grad_norm": 3.7895980004495553, + "learning_rate": 1.9139117894821107e-06, + "loss": 0.5117, + "step": 1537 + }, + { + "epoch": 0.6187889760611547, + "grad_norm": 4.503044474467485, + "learning_rate": 1.910499268549459e-06, + "loss": 0.5728, + "step": 1538 + }, + { + "epoch": 0.6191913095956548, + "grad_norm": 4.5048282169496865, + "learning_rate": 1.9070879102749944e-06, + "loss": 0.4952, + "step": 1539 + }, + { + "epoch": 0.6195936431301549, + "grad_norm": 3.6152226921367516, + "learning_rate": 1.9036777213868567e-06, + "loss": 0.4901, + "step": 1540 + }, + { + "epoch": 0.619995976664655, + "grad_norm": 4.56665728701204, + "learning_rate": 1.9002687086108808e-06, + "loss": 0.6217, + "step": 1541 + }, + { + "epoch": 0.6203983101991551, + "grad_norm": 3.8747261446432972, + "learning_rate": 1.8968608786705803e-06, + "loss": 0.4346, + "step": 1542 + }, + { + "epoch": 0.6208006437336552, + "grad_norm": 3.553549231891504, + "learning_rate": 1.8934542382871365e-06, + "loss": 0.4086, + "step": 1543 + }, + { + "epoch": 0.6212029772681553, + "grad_norm": 4.273007467466175, + "learning_rate": 1.8900487941793862e-06, + "loss": 0.5589, + "step": 1544 + }, + { + "epoch": 0.6216053108026554, + "grad_norm": 4.033041498302765, + "learning_rate": 1.8866445530638034e-06, + "loss": 0.5258, + "step": 1545 + }, + { + "epoch": 0.6220076443371555, + "grad_norm": 3.847764801262625, + "learning_rate": 1.883241521654492e-06, + "loss": 0.5286, + "step": 1546 + }, + { + "epoch": 0.6224099778716556, + "grad_norm": 4.141005318996514, + "learning_rate": 1.87983970666317e-06, + "loss": 0.5428, + "step": 1547 + }, + { + "epoch": 0.6228123114061557, + "grad_norm": 3.8545860722248286, + "learning_rate": 1.8764391147991556e-06, + "loss": 0.4504, + "step": 1548 + }, + { + "epoch": 0.6232146449406558, + "grad_norm": 3.9064270159769543, + "learning_rate": 1.8730397527693544e-06, + "loss": 0.49, + "step": 1549 + }, + { + "epoch": 0.623616978475156, + "grad_norm": 3.5305364890500903, + "learning_rate": 1.8696416272782467e-06, + "loss": 0.4747, + "step": 1550 + }, + { + "epoch": 0.624019312009656, + "grad_norm": 4.367606237537799, + "learning_rate": 1.866244745027874e-06, + "loss": 0.6066, + "step": 1551 + }, + { + "epoch": 0.6244216455441561, + "grad_norm": 3.778565429919078, + "learning_rate": 1.8628491127178272e-06, + "loss": 0.4401, + "step": 1552 + }, + { + "epoch": 0.6248239790786562, + "grad_norm": 4.4484086477744675, + "learning_rate": 1.8594547370452288e-06, + "loss": 0.5702, + "step": 1553 + }, + { + "epoch": 0.6252263126131563, + "grad_norm": 4.133384847499752, + "learning_rate": 1.8560616247047253e-06, + "loss": 0.4958, + "step": 1554 + }, + { + "epoch": 0.6256286461476565, + "grad_norm": 4.481750156438252, + "learning_rate": 1.8526697823884721e-06, + "loss": 0.5982, + "step": 1555 + }, + { + "epoch": 0.6260309796821565, + "grad_norm": 4.679844131200749, + "learning_rate": 1.849279216786117e-06, + "loss": 0.5699, + "step": 1556 + }, + { + "epoch": 0.6264333132166566, + "grad_norm": 4.035894316525195, + "learning_rate": 1.8458899345847936e-06, + "loss": 0.6036, + "step": 1557 + }, + { + "epoch": 0.6268356467511567, + "grad_norm": 3.3784184252716836, + "learning_rate": 1.8425019424690993e-06, + "loss": 0.3667, + "step": 1558 + }, + { + "epoch": 0.6272379802856568, + "grad_norm": 3.7155353324243325, + "learning_rate": 1.839115247121091e-06, + "loss": 0.4744, + "step": 1559 + }, + { + "epoch": 0.627640313820157, + "grad_norm": 4.062605549457995, + "learning_rate": 1.8357298552202676e-06, + "loss": 0.4846, + "step": 1560 + }, + { + "epoch": 0.628042647354657, + "grad_norm": 3.968579655925073, + "learning_rate": 1.832345773443557e-06, + "loss": 0.4563, + "step": 1561 + }, + { + "epoch": 0.6284449808891571, + "grad_norm": 3.859404877482336, + "learning_rate": 1.8289630084653009e-06, + "loss": 0.5339, + "step": 1562 + }, + { + "epoch": 0.6288473144236573, + "grad_norm": 4.32493582991546, + "learning_rate": 1.825581566957247e-06, + "loss": 0.5206, + "step": 1563 + }, + { + "epoch": 0.6292496479581573, + "grad_norm": 3.7620622738243474, + "learning_rate": 1.8222014555885303e-06, + "loss": 0.5187, + "step": 1564 + }, + { + "epoch": 0.6296519814926574, + "grad_norm": 3.9245871423313616, + "learning_rate": 1.8188226810256647e-06, + "loss": 0.5302, + "step": 1565 + }, + { + "epoch": 0.6300543150271575, + "grad_norm": 4.0730169054317305, + "learning_rate": 1.8154452499325273e-06, + "loss": 0.482, + "step": 1566 + }, + { + "epoch": 0.6304566485616576, + "grad_norm": 4.0887356034823314, + "learning_rate": 1.8120691689703423e-06, + "loss": 0.6323, + "step": 1567 + }, + { + "epoch": 0.6308589820961578, + "grad_norm": 4.134013360882183, + "learning_rate": 1.8086944447976745e-06, + "loss": 0.4215, + "step": 1568 + }, + { + "epoch": 0.6312613156306578, + "grad_norm": 4.267968482401992, + "learning_rate": 1.805321084070412e-06, + "loss": 0.6421, + "step": 1569 + }, + { + "epoch": 0.6316636491651579, + "grad_norm": 4.337432367486679, + "learning_rate": 1.8019490934417533e-06, + "loss": 0.614, + "step": 1570 + }, + { + "epoch": 0.632065982699658, + "grad_norm": 4.325604571245914, + "learning_rate": 1.7985784795621946e-06, + "loss": 0.6082, + "step": 1571 + }, + { + "epoch": 0.6324683162341581, + "grad_norm": 4.7255506567581005, + "learning_rate": 1.7952092490795165e-06, + "loss": 0.5848, + "step": 1572 + }, + { + "epoch": 0.6328706497686583, + "grad_norm": 3.9917358158112513, + "learning_rate": 1.7918414086387723e-06, + "loss": 0.51, + "step": 1573 + }, + { + "epoch": 0.6332729833031583, + "grad_norm": 3.9759190309543806, + "learning_rate": 1.788474964882273e-06, + "loss": 0.4348, + "step": 1574 + }, + { + "epoch": 0.6336753168376584, + "grad_norm": 3.9060781229407673, + "learning_rate": 1.7851099244495761e-06, + "loss": 0.5328, + "step": 1575 + }, + { + "epoch": 0.6340776503721586, + "grad_norm": 3.794629121755811, + "learning_rate": 1.7817462939774683e-06, + "loss": 0.5406, + "step": 1576 + }, + { + "epoch": 0.6344799839066586, + "grad_norm": 4.434204491241485, + "learning_rate": 1.778384080099959e-06, + "loss": 0.6955, + "step": 1577 + }, + { + "epoch": 0.6348823174411587, + "grad_norm": 4.243047298973574, + "learning_rate": 1.775023289448261e-06, + "loss": 0.6021, + "step": 1578 + }, + { + "epoch": 0.6352846509756588, + "grad_norm": 3.906621293593377, + "learning_rate": 1.7716639286507834e-06, + "loss": 0.5017, + "step": 1579 + }, + { + "epoch": 0.6356869845101589, + "grad_norm": 4.433075604901836, + "learning_rate": 1.7683060043331102e-06, + "loss": 0.6304, + "step": 1580 + }, + { + "epoch": 0.6360893180446591, + "grad_norm": 4.0324680233039425, + "learning_rate": 1.764949523117997e-06, + "loss": 0.4593, + "step": 1581 + }, + { + "epoch": 0.6364916515791591, + "grad_norm": 4.861200946299266, + "learning_rate": 1.7615944916253511e-06, + "loss": 0.5755, + "step": 1582 + }, + { + "epoch": 0.6368939851136592, + "grad_norm": 3.4859779578870462, + "learning_rate": 1.7582409164722219e-06, + "loss": 0.4494, + "step": 1583 + }, + { + "epoch": 0.6372963186481593, + "grad_norm": 4.394364956739334, + "learning_rate": 1.7548888042727835e-06, + "loss": 0.5756, + "step": 1584 + }, + { + "epoch": 0.6376986521826594, + "grad_norm": 4.474860851761582, + "learning_rate": 1.751538161638327e-06, + "loss": 0.5217, + "step": 1585 + }, + { + "epoch": 0.6381009857171596, + "grad_norm": 4.665357328027418, + "learning_rate": 1.748188995177245e-06, + "loss": 0.5599, + "step": 1586 + }, + { + "epoch": 0.6385033192516596, + "grad_norm": 3.186769209343777, + "learning_rate": 1.744841311495019e-06, + "loss": 0.3409, + "step": 1587 + }, + { + "epoch": 0.6389056527861597, + "grad_norm": 3.3506177658392744, + "learning_rate": 1.7414951171942057e-06, + "loss": 0.5176, + "step": 1588 + }, + { + "epoch": 0.6393079863206599, + "grad_norm": 3.4136910306135935, + "learning_rate": 1.7381504188744218e-06, + "loss": 0.4464, + "step": 1589 + }, + { + "epoch": 0.6397103198551599, + "grad_norm": 3.4917018490110467, + "learning_rate": 1.7348072231323382e-06, + "loss": 0.438, + "step": 1590 + }, + { + "epoch": 0.6401126533896601, + "grad_norm": 3.421803275946972, + "learning_rate": 1.7314655365616584e-06, + "loss": 0.4189, + "step": 1591 + }, + { + "epoch": 0.6405149869241601, + "grad_norm": 4.433987296079437, + "learning_rate": 1.7281253657531123e-06, + "loss": 0.606, + "step": 1592 + }, + { + "epoch": 0.6409173204586602, + "grad_norm": 3.6737703307949983, + "learning_rate": 1.7247867172944367e-06, + "loss": 0.4349, + "step": 1593 + }, + { + "epoch": 0.6413196539931604, + "grad_norm": 3.8221357123925914, + "learning_rate": 1.721449597770369e-06, + "loss": 0.484, + "step": 1594 + }, + { + "epoch": 0.6417219875276604, + "grad_norm": 4.069693321778052, + "learning_rate": 1.718114013762631e-06, + "loss": 0.6507, + "step": 1595 + }, + { + "epoch": 0.6421243210621606, + "grad_norm": 4.268934323506783, + "learning_rate": 1.714779971849915e-06, + "loss": 0.498, + "step": 1596 + }, + { + "epoch": 0.6425266545966606, + "grad_norm": 4.426082622301671, + "learning_rate": 1.7114474786078716e-06, + "loss": 0.6755, + "step": 1597 + }, + { + "epoch": 0.6429289881311607, + "grad_norm": 3.6655062613985288, + "learning_rate": 1.708116540609098e-06, + "loss": 0.5078, + "step": 1598 + }, + { + "epoch": 0.6433313216656609, + "grad_norm": 4.096742196755438, + "learning_rate": 1.7047871644231226e-06, + "loss": 0.5085, + "step": 1599 + }, + { + "epoch": 0.6437336552001609, + "grad_norm": 4.341802218643631, + "learning_rate": 1.7014593566163949e-06, + "loss": 0.6022, + "step": 1600 + }, + { + "epoch": 0.644135988734661, + "grad_norm": 4.167303639683251, + "learning_rate": 1.6981331237522716e-06, + "loss": 0.595, + "step": 1601 + }, + { + "epoch": 0.6445383222691611, + "grad_norm": 3.860753980307849, + "learning_rate": 1.6948084723910002e-06, + "loss": 0.4189, + "step": 1602 + }, + { + "epoch": 0.6449406558036612, + "grad_norm": 3.801435117916446, + "learning_rate": 1.6914854090897124e-06, + "loss": 0.4691, + "step": 1603 + }, + { + "epoch": 0.6453429893381614, + "grad_norm": 4.309844638484278, + "learning_rate": 1.6881639404024063e-06, + "loss": 0.4802, + "step": 1604 + }, + { + "epoch": 0.6457453228726614, + "grad_norm": 3.8218645060402285, + "learning_rate": 1.6848440728799346e-06, + "loss": 0.6087, + "step": 1605 + }, + { + "epoch": 0.6461476564071615, + "grad_norm": 4.318515117319728, + "learning_rate": 1.6815258130699938e-06, + "loss": 0.6644, + "step": 1606 + }, + { + "epoch": 0.6465499899416617, + "grad_norm": 3.8706029461346114, + "learning_rate": 1.6782091675171064e-06, + "loss": 0.5314, + "step": 1607 + }, + { + "epoch": 0.6469523234761617, + "grad_norm": 4.588786654868146, + "learning_rate": 1.6748941427626142e-06, + "loss": 0.5713, + "step": 1608 + }, + { + "epoch": 0.6473546570106619, + "grad_norm": 4.206226833574013, + "learning_rate": 1.671580745344661e-06, + "loss": 0.5207, + "step": 1609 + }, + { + "epoch": 0.6477569905451619, + "grad_norm": 4.1241193689590565, + "learning_rate": 1.6682689817981827e-06, + "loss": 0.5325, + "step": 1610 + }, + { + "epoch": 0.648159324079662, + "grad_norm": 4.013331656155903, + "learning_rate": 1.6649588586548888e-06, + "loss": 0.5593, + "step": 1611 + }, + { + "epoch": 0.6485616576141622, + "grad_norm": 4.701503684635364, + "learning_rate": 1.6616503824432573e-06, + "loss": 0.6878, + "step": 1612 + }, + { + "epoch": 0.6489639911486622, + "grad_norm": 4.047494584818948, + "learning_rate": 1.658343559688516e-06, + "loss": 0.5131, + "step": 1613 + }, + { + "epoch": 0.6493663246831624, + "grad_norm": 4.0317139401208575, + "learning_rate": 1.6550383969126341e-06, + "loss": 0.512, + "step": 1614 + }, + { + "epoch": 0.6497686582176624, + "grad_norm": 3.9682317404566034, + "learning_rate": 1.6517349006343025e-06, + "loss": 0.5921, + "step": 1615 + }, + { + "epoch": 0.6501709917521625, + "grad_norm": 3.824331386002604, + "learning_rate": 1.6484330773689292e-06, + "loss": 0.4712, + "step": 1616 + }, + { + "epoch": 0.6505733252866627, + "grad_norm": 3.7019715035054204, + "learning_rate": 1.645132933628621e-06, + "loss": 0.4426, + "step": 1617 + }, + { + "epoch": 0.6509756588211627, + "grad_norm": 4.0459494986124245, + "learning_rate": 1.6418344759221734e-06, + "loss": 0.6728, + "step": 1618 + }, + { + "epoch": 0.6513779923556628, + "grad_norm": 4.154812270453969, + "learning_rate": 1.6385377107550543e-06, + "loss": 0.559, + "step": 1619 + }, + { + "epoch": 0.651780325890163, + "grad_norm": 4.703387446268752, + "learning_rate": 1.6352426446293952e-06, + "loss": 0.5548, + "step": 1620 + }, + { + "epoch": 0.652182659424663, + "grad_norm": 3.894555312354832, + "learning_rate": 1.631949284043976e-06, + "loss": 0.4987, + "step": 1621 + }, + { + "epoch": 0.6525849929591632, + "grad_norm": 4.281016820029128, + "learning_rate": 1.628657635494213e-06, + "loss": 0.6007, + "step": 1622 + }, + { + "epoch": 0.6529873264936632, + "grad_norm": 3.708021853167261, + "learning_rate": 1.625367705472147e-06, + "loss": 0.5041, + "step": 1623 + }, + { + "epoch": 0.6533896600281633, + "grad_norm": 4.184667272519736, + "learning_rate": 1.6220795004664264e-06, + "loss": 0.563, + "step": 1624 + }, + { + "epoch": 0.6537919935626635, + "grad_norm": 4.386544791097414, + "learning_rate": 1.6187930269623001e-06, + "loss": 0.6591, + "step": 1625 + }, + { + "epoch": 0.6541943270971635, + "grad_norm": 4.095997806253382, + "learning_rate": 1.6155082914416018e-06, + "loss": 0.5122, + "step": 1626 + }, + { + "epoch": 0.6545966606316637, + "grad_norm": 3.97582656513418, + "learning_rate": 1.6122253003827353e-06, + "loss": 0.5805, + "step": 1627 + }, + { + "epoch": 0.6549989941661637, + "grad_norm": 3.990901166079429, + "learning_rate": 1.6089440602606675e-06, + "loss": 0.569, + "step": 1628 + }, + { + "epoch": 0.6554013277006638, + "grad_norm": 3.8624368116988403, + "learning_rate": 1.6056645775469077e-06, + "loss": 0.4729, + "step": 1629 + }, + { + "epoch": 0.655803661235164, + "grad_norm": 4.529362920268054, + "learning_rate": 1.6023868587095016e-06, + "loss": 0.5482, + "step": 1630 + }, + { + "epoch": 0.656205994769664, + "grad_norm": 4.017565552354934, + "learning_rate": 1.599110910213016e-06, + "loss": 0.4916, + "step": 1631 + }, + { + "epoch": 0.6566083283041642, + "grad_norm": 4.4182156313212575, + "learning_rate": 1.595836738518527e-06, + "loss": 0.5792, + "step": 1632 + }, + { + "epoch": 0.6570106618386643, + "grad_norm": 3.5677962985553524, + "learning_rate": 1.592564350083603e-06, + "loss": 0.4697, + "step": 1633 + }, + { + "epoch": 0.6574129953731643, + "grad_norm": 3.909711067921425, + "learning_rate": 1.5892937513622977e-06, + "loss": 0.4383, + "step": 1634 + }, + { + "epoch": 0.6578153289076645, + "grad_norm": 4.098825788465001, + "learning_rate": 1.5860249488051355e-06, + "loss": 0.4745, + "step": 1635 + }, + { + "epoch": 0.6582176624421645, + "grad_norm": 3.2004894026603616, + "learning_rate": 1.5827579488590977e-06, + "loss": 0.4775, + "step": 1636 + }, + { + "epoch": 0.6586199959766647, + "grad_norm": 3.970341598297501, + "learning_rate": 1.5794927579676089e-06, + "loss": 0.5471, + "step": 1637 + }, + { + "epoch": 0.6590223295111648, + "grad_norm": 4.241718162619275, + "learning_rate": 1.5762293825705276e-06, + "loss": 0.6544, + "step": 1638 + }, + { + "epoch": 0.6594246630456648, + "grad_norm": 3.924819583311759, + "learning_rate": 1.5729678291041318e-06, + "loss": 0.5368, + "step": 1639 + }, + { + "epoch": 0.659826996580165, + "grad_norm": 3.8382551527997397, + "learning_rate": 1.5697081040011043e-06, + "loss": 0.5121, + "step": 1640 + }, + { + "epoch": 0.660229330114665, + "grad_norm": 3.6682266397110337, + "learning_rate": 1.5664502136905236e-06, + "loss": 0.4353, + "step": 1641 + }, + { + "epoch": 0.6606316636491651, + "grad_norm": 3.9060845453518143, + "learning_rate": 1.563194164597848e-06, + "loss": 0.4814, + "step": 1642 + }, + { + "epoch": 0.6610339971836653, + "grad_norm": 3.985541657912861, + "learning_rate": 1.5599399631449052e-06, + "loss": 0.4643, + "step": 1643 + }, + { + "epoch": 0.6614363307181653, + "grad_norm": 4.246558106503365, + "learning_rate": 1.5566876157498797e-06, + "loss": 0.5457, + "step": 1644 + }, + { + "epoch": 0.6618386642526655, + "grad_norm": 4.538356504223212, + "learning_rate": 1.5534371288272992e-06, + "loss": 0.5586, + "step": 1645 + }, + { + "epoch": 0.6622409977871656, + "grad_norm": 3.512817123486096, + "learning_rate": 1.5501885087880187e-06, + "loss": 0.3777, + "step": 1646 + }, + { + "epoch": 0.6626433313216656, + "grad_norm": 3.611315735282059, + "learning_rate": 1.546941762039216e-06, + "loss": 0.4559, + "step": 1647 + }, + { + "epoch": 0.6630456648561658, + "grad_norm": 3.804900233889627, + "learning_rate": 1.5436968949843705e-06, + "loss": 0.5593, + "step": 1648 + }, + { + "epoch": 0.6634479983906658, + "grad_norm": 4.290281020900778, + "learning_rate": 1.5404539140232572e-06, + "loss": 0.5433, + "step": 1649 + }, + { + "epoch": 0.663850331925166, + "grad_norm": 4.261399800624298, + "learning_rate": 1.5372128255519298e-06, + "loss": 0.5669, + "step": 1650 + }, + { + "epoch": 0.6642526654596661, + "grad_norm": 4.8206169191275094, + "learning_rate": 1.533973635962708e-06, + "loss": 0.535, + "step": 1651 + }, + { + "epoch": 0.6646549989941661, + "grad_norm": 3.5705866846183882, + "learning_rate": 1.5307363516441693e-06, + "loss": 0.4436, + "step": 1652 + }, + { + "epoch": 0.6650573325286663, + "grad_norm": 4.015284104793611, + "learning_rate": 1.5275009789811317e-06, + "loss": 0.4606, + "step": 1653 + }, + { + "epoch": 0.6654596660631663, + "grad_norm": 3.746888831379427, + "learning_rate": 1.5242675243546435e-06, + "loss": 0.5344, + "step": 1654 + }, + { + "epoch": 0.6658619995976665, + "grad_norm": 5.350374969998737, + "learning_rate": 1.5210359941419698e-06, + "loss": 0.7187, + "step": 1655 + }, + { + "epoch": 0.6662643331321666, + "grad_norm": 4.992414873463243, + "learning_rate": 1.5178063947165795e-06, + "loss": 0.5597, + "step": 1656 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 3.9536138723501413, + "learning_rate": 1.514578732448135e-06, + "loss": 0.5345, + "step": 1657 + }, + { + "epoch": 0.6670690002011668, + "grad_norm": 4.023274067891729, + "learning_rate": 1.5113530137024774e-06, + "loss": 0.4893, + "step": 1658 + }, + { + "epoch": 0.6674713337356669, + "grad_norm": 4.108018429855622, + "learning_rate": 1.5081292448416158e-06, + "loss": 0.645, + "step": 1659 + }, + { + "epoch": 0.667873667270167, + "grad_norm": 3.817031046521698, + "learning_rate": 1.5049074322237097e-06, + "loss": 0.4838, + "step": 1660 + }, + { + "epoch": 0.6682760008046671, + "grad_norm": 4.10734795999481, + "learning_rate": 1.5016875822030651e-06, + "loss": 0.5853, + "step": 1661 + }, + { + "epoch": 0.6686783343391671, + "grad_norm": 4.104736058188134, + "learning_rate": 1.4984697011301142e-06, + "loss": 0.5032, + "step": 1662 + }, + { + "epoch": 0.6690806678736673, + "grad_norm": 4.213736489738209, + "learning_rate": 1.4952537953514078e-06, + "loss": 0.4234, + "step": 1663 + }, + { + "epoch": 0.6694830014081674, + "grad_norm": 4.592368631770386, + "learning_rate": 1.4920398712095985e-06, + "loss": 0.5323, + "step": 1664 + }, + { + "epoch": 0.6698853349426674, + "grad_norm": 4.593839464020305, + "learning_rate": 1.4888279350434325e-06, + "loss": 0.5821, + "step": 1665 + }, + { + "epoch": 0.6702876684771676, + "grad_norm": 3.881133737659202, + "learning_rate": 1.485617993187735e-06, + "loss": 0.5224, + "step": 1666 + }, + { + "epoch": 0.6706900020116676, + "grad_norm": 4.088068946795046, + "learning_rate": 1.4824100519733985e-06, + "loss": 0.556, + "step": 1667 + }, + { + "epoch": 0.6710923355461678, + "grad_norm": 4.328495324161391, + "learning_rate": 1.4792041177273668e-06, + "loss": 0.6097, + "step": 1668 + }, + { + "epoch": 0.6714946690806679, + "grad_norm": 4.220316709489982, + "learning_rate": 1.4760001967726283e-06, + "loss": 0.6015, + "step": 1669 + }, + { + "epoch": 0.6718970026151679, + "grad_norm": 3.788396547796328, + "learning_rate": 1.472798295428199e-06, + "loss": 0.5863, + "step": 1670 + }, + { + "epoch": 0.6722993361496681, + "grad_norm": 3.8086553750558494, + "learning_rate": 1.4695984200091129e-06, + "loss": 0.5155, + "step": 1671 + }, + { + "epoch": 0.6727016696841682, + "grad_norm": 3.792467737324882, + "learning_rate": 1.4664005768264084e-06, + "loss": 0.4517, + "step": 1672 + }, + { + "epoch": 0.6731040032186683, + "grad_norm": 4.528886592192105, + "learning_rate": 1.4632047721871134e-06, + "loss": 0.5514, + "step": 1673 + }, + { + "epoch": 0.6735063367531684, + "grad_norm": 3.687134247150264, + "learning_rate": 1.4600110123942376e-06, + "loss": 0.5765, + "step": 1674 + }, + { + "epoch": 0.6739086702876684, + "grad_norm": 3.651752055706122, + "learning_rate": 1.4568193037467574e-06, + "loss": 0.4434, + "step": 1675 + }, + { + "epoch": 0.6743110038221686, + "grad_norm": 4.196555677667497, + "learning_rate": 1.453629652539604e-06, + "loss": 0.5832, + "step": 1676 + }, + { + "epoch": 0.6747133373566687, + "grad_norm": 3.3768494467370047, + "learning_rate": 1.4504420650636485e-06, + "loss": 0.4094, + "step": 1677 + }, + { + "epoch": 0.6751156708911688, + "grad_norm": 3.9350013310511582, + "learning_rate": 1.4472565476056952e-06, + "loss": 0.4875, + "step": 1678 + }, + { + "epoch": 0.6755180044256689, + "grad_norm": 3.9364895166754383, + "learning_rate": 1.444073106448462e-06, + "loss": 0.626, + "step": 1679 + }, + { + "epoch": 0.6759203379601689, + "grad_norm": 4.218774138421791, + "learning_rate": 1.4408917478705748e-06, + "loss": 0.5318, + "step": 1680 + }, + { + "epoch": 0.6763226714946691, + "grad_norm": 3.557852927029882, + "learning_rate": 1.4377124781465517e-06, + "loss": 0.4564, + "step": 1681 + }, + { + "epoch": 0.6767250050291692, + "grad_norm": 4.131292301591179, + "learning_rate": 1.4345353035467888e-06, + "loss": 0.5458, + "step": 1682 + }, + { + "epoch": 0.6771273385636692, + "grad_norm": 4.261120314757306, + "learning_rate": 1.4313602303375523e-06, + "loss": 0.6402, + "step": 1683 + }, + { + "epoch": 0.6775296720981694, + "grad_norm": 4.027439558487913, + "learning_rate": 1.4281872647809628e-06, + "loss": 0.5762, + "step": 1684 + }, + { + "epoch": 0.6779320056326695, + "grad_norm": 3.9491401859678605, + "learning_rate": 1.4250164131349858e-06, + "loss": 0.4807, + "step": 1685 + }, + { + "epoch": 0.6783343391671696, + "grad_norm": 3.60940986670999, + "learning_rate": 1.421847681653414e-06, + "loss": 0.3659, + "step": 1686 + }, + { + "epoch": 0.6787366727016697, + "grad_norm": 3.5822407819650515, + "learning_rate": 1.4186810765858616e-06, + "loss": 0.5165, + "step": 1687 + }, + { + "epoch": 0.6791390062361697, + "grad_norm": 4.161182758555724, + "learning_rate": 1.4155166041777483e-06, + "loss": 0.495, + "step": 1688 + }, + { + "epoch": 0.6795413397706699, + "grad_norm": 4.076431248542256, + "learning_rate": 1.4123542706702868e-06, + "loss": 0.4569, + "step": 1689 + }, + { + "epoch": 0.67994367330517, + "grad_norm": 3.401944784975411, + "learning_rate": 1.409194082300473e-06, + "loss": 0.4661, + "step": 1690 + }, + { + "epoch": 0.6803460068396701, + "grad_norm": 4.053896920545759, + "learning_rate": 1.4060360453010691e-06, + "loss": 0.6945, + "step": 1691 + }, + { + "epoch": 0.6807483403741702, + "grad_norm": 3.6315876209619726, + "learning_rate": 1.4028801659005975e-06, + "loss": 0.5088, + "step": 1692 + }, + { + "epoch": 0.6811506739086702, + "grad_norm": 3.9033263695876617, + "learning_rate": 1.3997264503233222e-06, + "loss": 0.5535, + "step": 1693 + }, + { + "epoch": 0.6815530074431704, + "grad_norm": 3.6520508454132568, + "learning_rate": 1.3965749047892421e-06, + "loss": 0.4627, + "step": 1694 + }, + { + "epoch": 0.6819553409776705, + "grad_norm": 4.253145100591827, + "learning_rate": 1.3934255355140735e-06, + "loss": 0.6059, + "step": 1695 + }, + { + "epoch": 0.6823576745121706, + "grad_norm": 3.6442609114551168, + "learning_rate": 1.3902783487092436e-06, + "loss": 0.437, + "step": 1696 + }, + { + "epoch": 0.6827600080466707, + "grad_norm": 3.6843000132081083, + "learning_rate": 1.387133350581873e-06, + "loss": 0.4741, + "step": 1697 + }, + { + "epoch": 0.6831623415811708, + "grad_norm": 4.082924701084028, + "learning_rate": 1.383990547334767e-06, + "loss": 0.5523, + "step": 1698 + }, + { + "epoch": 0.6835646751156709, + "grad_norm": 4.525193517476986, + "learning_rate": 1.3808499451663998e-06, + "loss": 0.5733, + "step": 1699 + }, + { + "epoch": 0.683967008650171, + "grad_norm": 3.8434172838289142, + "learning_rate": 1.377711550270907e-06, + "loss": 0.4042, + "step": 1700 + }, + { + "epoch": 0.684369342184671, + "grad_norm": 3.475844585621674, + "learning_rate": 1.3745753688380692e-06, + "loss": 0.4492, + "step": 1701 + }, + { + "epoch": 0.6847716757191712, + "grad_norm": 4.106466165584271, + "learning_rate": 1.3714414070533022e-06, + "loss": 0.4795, + "step": 1702 + }, + { + "epoch": 0.6851740092536713, + "grad_norm": 3.9408796988863477, + "learning_rate": 1.368309671097645e-06, + "loss": 0.4586, + "step": 1703 + }, + { + "epoch": 0.6855763427881714, + "grad_norm": 3.892878635048642, + "learning_rate": 1.3651801671477435e-06, + "loss": 0.5256, + "step": 1704 + }, + { + "epoch": 0.6859786763226715, + "grad_norm": 3.602589673252341, + "learning_rate": 1.3620529013758444e-06, + "loss": 0.4474, + "step": 1705 + }, + { + "epoch": 0.6863810098571715, + "grad_norm": 4.29050518808401, + "learning_rate": 1.3589278799497802e-06, + "loss": 0.4799, + "step": 1706 + }, + { + "epoch": 0.6867833433916717, + "grad_norm": 4.530188596691838, + "learning_rate": 1.3558051090329536e-06, + "loss": 0.6044, + "step": 1707 + }, + { + "epoch": 0.6871856769261718, + "grad_norm": 3.8842771157381932, + "learning_rate": 1.352684594784333e-06, + "loss": 0.4759, + "step": 1708 + }, + { + "epoch": 0.6875880104606719, + "grad_norm": 4.206785571596244, + "learning_rate": 1.3495663433584327e-06, + "loss": 0.58, + "step": 1709 + }, + { + "epoch": 0.687990343995172, + "grad_norm": 4.648975138578508, + "learning_rate": 1.3464503609053053e-06, + "loss": 0.5508, + "step": 1710 + }, + { + "epoch": 0.6883926775296721, + "grad_norm": 4.513492173355037, + "learning_rate": 1.3433366535705292e-06, + "loss": 0.6606, + "step": 1711 + }, + { + "epoch": 0.6887950110641722, + "grad_norm": 4.149085872177257, + "learning_rate": 1.3402252274951954e-06, + "loss": 0.5438, + "step": 1712 + }, + { + "epoch": 0.6891973445986723, + "grad_norm": 4.3080582119504935, + "learning_rate": 1.3371160888158935e-06, + "loss": 0.7277, + "step": 1713 + }, + { + "epoch": 0.6895996781331724, + "grad_norm": 4.195944255893043, + "learning_rate": 1.3340092436647045e-06, + "loss": 0.5682, + "step": 1714 + }, + { + "epoch": 0.6900020116676725, + "grad_norm": 4.22717459587227, + "learning_rate": 1.330904698169184e-06, + "loss": 0.5394, + "step": 1715 + }, + { + "epoch": 0.6904043452021726, + "grad_norm": 3.8328998433470334, + "learning_rate": 1.3278024584523542e-06, + "loss": 0.519, + "step": 1716 + }, + { + "epoch": 0.6908066787366727, + "grad_norm": 3.809809815071977, + "learning_rate": 1.324702530632686e-06, + "loss": 0.4609, + "step": 1717 + }, + { + "epoch": 0.6912090122711728, + "grad_norm": 4.408644772842878, + "learning_rate": 1.3216049208240945e-06, + "loss": 0.5259, + "step": 1718 + }, + { + "epoch": 0.6916113458056729, + "grad_norm": 4.351040955206057, + "learning_rate": 1.3185096351359211e-06, + "loss": 0.531, + "step": 1719 + }, + { + "epoch": 0.692013679340173, + "grad_norm": 3.8862458710646135, + "learning_rate": 1.3154166796729245e-06, + "loss": 0.5521, + "step": 1720 + }, + { + "epoch": 0.6924160128746731, + "grad_norm": 3.8732946833809283, + "learning_rate": 1.3123260605352647e-06, + "loss": 0.4054, + "step": 1721 + }, + { + "epoch": 0.6928183464091732, + "grad_norm": 4.014263742357025, + "learning_rate": 1.3092377838184984e-06, + "loss": 0.5096, + "step": 1722 + }, + { + "epoch": 0.6932206799436733, + "grad_norm": 3.8457014699738306, + "learning_rate": 1.3061518556135572e-06, + "loss": 0.5361, + "step": 1723 + }, + { + "epoch": 0.6936230134781735, + "grad_norm": 4.700764158732767, + "learning_rate": 1.3030682820067447e-06, + "loss": 0.6363, + "step": 1724 + }, + { + "epoch": 0.6940253470126735, + "grad_norm": 4.056861341126563, + "learning_rate": 1.2999870690797207e-06, + "loss": 0.5499, + "step": 1725 + }, + { + "epoch": 0.6944276805471736, + "grad_norm": 3.824223111444158, + "learning_rate": 1.2969082229094853e-06, + "loss": 0.5049, + "step": 1726 + }, + { + "epoch": 0.6948300140816737, + "grad_norm": 3.8066989110426332, + "learning_rate": 1.2938317495683744e-06, + "loss": 0.487, + "step": 1727 + }, + { + "epoch": 0.6952323476161738, + "grad_norm": 4.382851529690001, + "learning_rate": 1.2907576551240425e-06, + "loss": 0.5016, + "step": 1728 + }, + { + "epoch": 0.695634681150674, + "grad_norm": 4.2709658357684415, + "learning_rate": 1.2876859456394537e-06, + "loss": 0.5287, + "step": 1729 + }, + { + "epoch": 0.696037014685174, + "grad_norm": 4.10428141385855, + "learning_rate": 1.284616627172865e-06, + "loss": 0.5048, + "step": 1730 + }, + { + "epoch": 0.6964393482196741, + "grad_norm": 3.9240197132726133, + "learning_rate": 1.281549705777821e-06, + "loss": 0.5834, + "step": 1731 + }, + { + "epoch": 0.6968416817541742, + "grad_norm": 3.917832710676123, + "learning_rate": 1.278485187503137e-06, + "loss": 0.4978, + "step": 1732 + }, + { + "epoch": 0.6972440152886743, + "grad_norm": 3.9886241574028123, + "learning_rate": 1.2754230783928907e-06, + "loss": 0.5046, + "step": 1733 + }, + { + "epoch": 0.6976463488231744, + "grad_norm": 3.8252585006207744, + "learning_rate": 1.2723633844864036e-06, + "loss": 0.4337, + "step": 1734 + }, + { + "epoch": 0.6980486823576745, + "grad_norm": 4.327909530843214, + "learning_rate": 1.2693061118182399e-06, + "loss": 0.5663, + "step": 1735 + }, + { + "epoch": 0.6984510158921746, + "grad_norm": 3.9729308104472225, + "learning_rate": 1.2662512664181826e-06, + "loss": 0.4753, + "step": 1736 + }, + { + "epoch": 0.6988533494266748, + "grad_norm": 3.4193536839822434, + "learning_rate": 1.2631988543112314e-06, + "loss": 0.3984, + "step": 1737 + }, + { + "epoch": 0.6992556829611748, + "grad_norm": 3.199663843957577, + "learning_rate": 1.2601488815175867e-06, + "loss": 0.3615, + "step": 1738 + }, + { + "epoch": 0.6996580164956749, + "grad_norm": 4.275095481044542, + "learning_rate": 1.257101354052635e-06, + "loss": 0.635, + "step": 1739 + }, + { + "epoch": 0.700060350030175, + "grad_norm": 4.154894699295693, + "learning_rate": 1.2540562779269421e-06, + "loss": 0.468, + "step": 1740 + }, + { + "epoch": 0.7004626835646751, + "grad_norm": 4.21271106717443, + "learning_rate": 1.2510136591462399e-06, + "loss": 0.557, + "step": 1741 + }, + { + "epoch": 0.7008650170991753, + "grad_norm": 3.9488032178250276, + "learning_rate": 1.2479735037114118e-06, + "loss": 0.4743, + "step": 1742 + }, + { + "epoch": 0.7012673506336753, + "grad_norm": 3.6437407466020453, + "learning_rate": 1.2449358176184848e-06, + "loss": 0.4958, + "step": 1743 + }, + { + "epoch": 0.7016696841681754, + "grad_norm": 4.011291060246906, + "learning_rate": 1.2419006068586125e-06, + "loss": 0.5109, + "step": 1744 + }, + { + "epoch": 0.7020720177026755, + "grad_norm": 3.7545390991024026, + "learning_rate": 1.2388678774180698e-06, + "loss": 0.5269, + "step": 1745 + }, + { + "epoch": 0.7024743512371756, + "grad_norm": 3.821504931037996, + "learning_rate": 1.2358376352782358e-06, + "loss": 0.4905, + "step": 1746 + }, + { + "epoch": 0.7028766847716758, + "grad_norm": 4.3580981207011655, + "learning_rate": 1.232809886415586e-06, + "loss": 0.4904, + "step": 1747 + }, + { + "epoch": 0.7032790183061758, + "grad_norm": 3.539055820090707, + "learning_rate": 1.2297846368016748e-06, + "loss": 0.4124, + "step": 1748 + }, + { + "epoch": 0.7036813518406759, + "grad_norm": 3.9844890147713254, + "learning_rate": 1.2267618924031317e-06, + "loss": 0.5321, + "step": 1749 + }, + { + "epoch": 0.7040836853751761, + "grad_norm": 4.178112023828336, + "learning_rate": 1.2237416591816412e-06, + "loss": 0.6362, + "step": 1750 + }, + { + "epoch": 0.7044860189096761, + "grad_norm": 4.0472275399599695, + "learning_rate": 1.220723943093939e-06, + "loss": 0.5101, + "step": 1751 + }, + { + "epoch": 0.7048883524441762, + "grad_norm": 4.312523214514084, + "learning_rate": 1.2177087500917925e-06, + "loss": 0.5277, + "step": 1752 + }, + { + "epoch": 0.7052906859786763, + "grad_norm": 4.133768938615848, + "learning_rate": 1.2146960861219954e-06, + "loss": 0.5271, + "step": 1753 + }, + { + "epoch": 0.7056930195131764, + "grad_norm": 4.430552507921185, + "learning_rate": 1.211685957126353e-06, + "loss": 0.711, + "step": 1754 + }, + { + "epoch": 0.7060953530476766, + "grad_norm": 4.119451270646705, + "learning_rate": 1.2086783690416704e-06, + "loss": 0.4358, + "step": 1755 + }, + { + "epoch": 0.7064976865821766, + "grad_norm": 4.295693476571627, + "learning_rate": 1.205673327799743e-06, + "loss": 0.5466, + "step": 1756 + }, + { + "epoch": 0.7069000201166767, + "grad_norm": 4.312446284251359, + "learning_rate": 1.2026708393273392e-06, + "loss": 0.5861, + "step": 1757 + }, + { + "epoch": 0.7073023536511768, + "grad_norm": 3.9109516304734733, + "learning_rate": 1.1996709095461967e-06, + "loss": 0.4782, + "step": 1758 + }, + { + "epoch": 0.7077046871856769, + "grad_norm": 4.425022693894618, + "learning_rate": 1.1966735443730052e-06, + "loss": 0.5627, + "step": 1759 + }, + { + "epoch": 0.7081070207201771, + "grad_norm": 4.259706435608531, + "learning_rate": 1.193678749719397e-06, + "loss": 0.5274, + "step": 1760 + }, + { + "epoch": 0.7085093542546771, + "grad_norm": 4.219902699982978, + "learning_rate": 1.190686531491932e-06, + "loss": 0.5397, + "step": 1761 + }, + { + "epoch": 0.7089116877891772, + "grad_norm": 4.105043115699033, + "learning_rate": 1.1876968955920916e-06, + "loss": 0.4845, + "step": 1762 + }, + { + "epoch": 0.7093140213236774, + "grad_norm": 3.905684008250733, + "learning_rate": 1.1847098479162644e-06, + "loss": 0.5769, + "step": 1763 + }, + { + "epoch": 0.7097163548581774, + "grad_norm": 4.0434341189775616, + "learning_rate": 1.1817253943557308e-06, + "loss": 0.4338, + "step": 1764 + }, + { + "epoch": 0.7101186883926776, + "grad_norm": 4.331648953621162, + "learning_rate": 1.1787435407966593e-06, + "loss": 0.6231, + "step": 1765 + }, + { + "epoch": 0.7105210219271776, + "grad_norm": 4.225800580764582, + "learning_rate": 1.1757642931200864e-06, + "loss": 0.4946, + "step": 1766 + }, + { + "epoch": 0.7109233554616777, + "grad_norm": 3.832934508222693, + "learning_rate": 1.1727876572019117e-06, + "loss": 0.5319, + "step": 1767 + }, + { + "epoch": 0.7113256889961779, + "grad_norm": 3.994089631167334, + "learning_rate": 1.1698136389128833e-06, + "loss": 0.5005, + "step": 1768 + }, + { + "epoch": 0.7117280225306779, + "grad_norm": 3.8399445750919425, + "learning_rate": 1.166842244118587e-06, + "loss": 0.4726, + "step": 1769 + }, + { + "epoch": 0.712130356065178, + "grad_norm": 4.043691249667768, + "learning_rate": 1.163873478679432e-06, + "loss": 0.4885, + "step": 1770 + }, + { + "epoch": 0.7125326895996781, + "grad_norm": 4.26586065507147, + "learning_rate": 1.1609073484506441e-06, + "loss": 0.5955, + "step": 1771 + }, + { + "epoch": 0.7129350231341782, + "grad_norm": 3.2723853825243006, + "learning_rate": 1.157943859282251e-06, + "loss": 0.393, + "step": 1772 + }, + { + "epoch": 0.7133373566686784, + "grad_norm": 4.228399556429077, + "learning_rate": 1.1549830170190714e-06, + "loss": 0.6138, + "step": 1773 + }, + { + "epoch": 0.7137396902031784, + "grad_norm": 3.703790105757895, + "learning_rate": 1.152024827500705e-06, + "loss": 0.4952, + "step": 1774 + }, + { + "epoch": 0.7141420237376785, + "grad_norm": 4.037603384099717, + "learning_rate": 1.149069296561516e-06, + "loss": 0.5791, + "step": 1775 + }, + { + "epoch": 0.7145443572721787, + "grad_norm": 3.958021889837016, + "learning_rate": 1.1461164300306292e-06, + "loss": 0.495, + "step": 1776 + }, + { + "epoch": 0.7149466908066787, + "grad_norm": 4.142430978673458, + "learning_rate": 1.1431662337319107e-06, + "loss": 0.6534, + "step": 1777 + }, + { + "epoch": 0.7153490243411789, + "grad_norm": 4.484413621549814, + "learning_rate": 1.1402187134839643e-06, + "loss": 0.4885, + "step": 1778 + }, + { + "epoch": 0.7157513578756789, + "grad_norm": 3.6695545389300395, + "learning_rate": 1.1372738751001111e-06, + "loss": 0.4446, + "step": 1779 + }, + { + "epoch": 0.716153691410179, + "grad_norm": 3.492829736247777, + "learning_rate": 1.134331724388387e-06, + "loss": 0.5245, + "step": 1780 + }, + { + "epoch": 0.7165560249446792, + "grad_norm": 3.8848835038152107, + "learning_rate": 1.1313922671515252e-06, + "loss": 0.5042, + "step": 1781 + }, + { + "epoch": 0.7169583584791792, + "grad_norm": 4.4510516838349705, + "learning_rate": 1.128455509186948e-06, + "loss": 0.6898, + "step": 1782 + }, + { + "epoch": 0.7173606920136794, + "grad_norm": 4.381017376532981, + "learning_rate": 1.1255214562867503e-06, + "loss": 0.5289, + "step": 1783 + }, + { + "epoch": 0.7177630255481794, + "grad_norm": 3.5079718266262416, + "learning_rate": 1.122590114237696e-06, + "loss": 0.4724, + "step": 1784 + }, + { + "epoch": 0.7181653590826795, + "grad_norm": 4.292226353007704, + "learning_rate": 1.1196614888212007e-06, + "loss": 0.5796, + "step": 1785 + }, + { + "epoch": 0.7185676926171797, + "grad_norm": 4.052087753788445, + "learning_rate": 1.1167355858133226e-06, + "loss": 0.4938, + "step": 1786 + }, + { + "epoch": 0.7189700261516797, + "grad_norm": 5.157980356295375, + "learning_rate": 1.1138124109847503e-06, + "loss": 0.5949, + "step": 1787 + }, + { + "epoch": 0.7193723596861799, + "grad_norm": 3.285940613285488, + "learning_rate": 1.1108919701007907e-06, + "loss": 0.3908, + "step": 1788 + }, + { + "epoch": 0.71977469322068, + "grad_norm": 3.5335717637010937, + "learning_rate": 1.1079742689213593e-06, + "loss": 0.5013, + "step": 1789 + }, + { + "epoch": 0.72017702675518, + "grad_norm": 4.450180240936036, + "learning_rate": 1.1050593132009703e-06, + "loss": 0.6329, + "step": 1790 + }, + { + "epoch": 0.7205793602896802, + "grad_norm": 4.404576812133769, + "learning_rate": 1.1021471086887187e-06, + "loss": 0.5781, + "step": 1791 + }, + { + "epoch": 0.7209816938241802, + "grad_norm": 3.3602741770117177, + "learning_rate": 1.0992376611282776e-06, + "loss": 0.3621, + "step": 1792 + }, + { + "epoch": 0.7213840273586803, + "grad_norm": 3.381272077924684, + "learning_rate": 1.0963309762578795e-06, + "loss": 0.384, + "step": 1793 + }, + { + "epoch": 0.7217863608931805, + "grad_norm": 3.5780173251972025, + "learning_rate": 1.0934270598103094e-06, + "loss": 0.5372, + "step": 1794 + }, + { + "epoch": 0.7221886944276805, + "grad_norm": 3.7738716419210814, + "learning_rate": 1.0905259175128931e-06, + "loss": 0.4336, + "step": 1795 + }, + { + "epoch": 0.7225910279621807, + "grad_norm": 3.820342940812247, + "learning_rate": 1.0876275550874846e-06, + "loss": 0.5335, + "step": 1796 + }, + { + "epoch": 0.7229933614966807, + "grad_norm": 4.176570559423123, + "learning_rate": 1.0847319782504532e-06, + "loss": 0.5038, + "step": 1797 + }, + { + "epoch": 0.7233956950311808, + "grad_norm": 3.67077010562448, + "learning_rate": 1.0818391927126764e-06, + "loss": 0.4342, + "step": 1798 + }, + { + "epoch": 0.723798028565681, + "grad_norm": 3.9880069700930423, + "learning_rate": 1.078949204179526e-06, + "loss": 0.4953, + "step": 1799 + }, + { + "epoch": 0.724200362100181, + "grad_norm": 4.161334696726195, + "learning_rate": 1.0760620183508582e-06, + "loss": 0.5799, + "step": 1800 + }, + { + "epoch": 0.7246026956346812, + "grad_norm": 4.126464363948578, + "learning_rate": 1.073177640920999e-06, + "loss": 0.5138, + "step": 1801 + }, + { + "epoch": 0.7250050291691813, + "grad_norm": 4.1895057334300505, + "learning_rate": 1.0702960775787374e-06, + "loss": 0.5121, + "step": 1802 + }, + { + "epoch": 0.7254073627036813, + "grad_norm": 3.884515999748438, + "learning_rate": 1.0674173340073121e-06, + "loss": 0.4891, + "step": 1803 + }, + { + "epoch": 0.7258096962381815, + "grad_norm": 4.286867137892461, + "learning_rate": 1.0645414158844012e-06, + "loss": 0.6118, + "step": 1804 + }, + { + "epoch": 0.7262120297726815, + "grad_norm": 3.8959234040057806, + "learning_rate": 1.0616683288821067e-06, + "loss": 0.5234, + "step": 1805 + }, + { + "epoch": 0.7266143633071817, + "grad_norm": 4.239978882514847, + "learning_rate": 1.058798078666952e-06, + "loss": 0.5183, + "step": 1806 + }, + { + "epoch": 0.7270166968416818, + "grad_norm": 3.7122065670739235, + "learning_rate": 1.0559306708998607e-06, + "loss": 0.501, + "step": 1807 + }, + { + "epoch": 0.7274190303761818, + "grad_norm": 4.556831231079228, + "learning_rate": 1.0530661112361533e-06, + "loss": 0.5257, + "step": 1808 + }, + { + "epoch": 0.727821363910682, + "grad_norm": 4.308685532929913, + "learning_rate": 1.050204405325534e-06, + "loss": 0.5534, + "step": 1809 + }, + { + "epoch": 0.728223697445182, + "grad_norm": 3.9297951909048368, + "learning_rate": 1.0473455588120738e-06, + "loss": 0.4681, + "step": 1810 + }, + { + "epoch": 0.7286260309796821, + "grad_norm": 3.98667448491781, + "learning_rate": 1.0444895773342092e-06, + "loss": 0.4785, + "step": 1811 + }, + { + "epoch": 0.7290283645141823, + "grad_norm": 4.2934600586771845, + "learning_rate": 1.0416364665247236e-06, + "loss": 0.5408, + "step": 1812 + }, + { + "epoch": 0.7294306980486823, + "grad_norm": 3.9347493779534, + "learning_rate": 1.0387862320107403e-06, + "loss": 0.5895, + "step": 1813 + }, + { + "epoch": 0.7298330315831825, + "grad_norm": 4.338941092246621, + "learning_rate": 1.0359388794137068e-06, + "loss": 0.5648, + "step": 1814 + }, + { + "epoch": 0.7302353651176826, + "grad_norm": 3.792647800273631, + "learning_rate": 1.0330944143493892e-06, + "loss": 0.5449, + "step": 1815 + }, + { + "epoch": 0.7306376986521826, + "grad_norm": 3.821766624740813, + "learning_rate": 1.0302528424278585e-06, + "loss": 0.4544, + "step": 1816 + }, + { + "epoch": 0.7310400321866828, + "grad_norm": 4.176558757203708, + "learning_rate": 1.0274141692534784e-06, + "loss": 0.4802, + "step": 1817 + }, + { + "epoch": 0.7314423657211828, + "grad_norm": 3.5626608894002803, + "learning_rate": 1.0245784004248968e-06, + "loss": 0.4273, + "step": 1818 + }, + { + "epoch": 0.731844699255683, + "grad_norm": 3.8719939712343936, + "learning_rate": 1.0217455415350316e-06, + "loss": 0.5409, + "step": 1819 + }, + { + "epoch": 0.7322470327901831, + "grad_norm": 4.537877117506395, + "learning_rate": 1.0189155981710642e-06, + "loss": 0.5705, + "step": 1820 + }, + { + "epoch": 0.7326493663246831, + "grad_norm": 3.932184586794533, + "learning_rate": 1.0160885759144217e-06, + "loss": 0.4904, + "step": 1821 + }, + { + "epoch": 0.7330516998591833, + "grad_norm": 4.266661778185627, + "learning_rate": 1.0132644803407753e-06, + "loss": 0.4769, + "step": 1822 + }, + { + "epoch": 0.7334540333936833, + "grad_norm": 3.7589693532265827, + "learning_rate": 1.0104433170200188e-06, + "loss": 0.4891, + "step": 1823 + }, + { + "epoch": 0.7338563669281835, + "grad_norm": 3.8562553041075263, + "learning_rate": 1.007625091516266e-06, + "loss": 0.417, + "step": 1824 + }, + { + "epoch": 0.7342587004626836, + "grad_norm": 4.130360177598847, + "learning_rate": 1.0048098093878352e-06, + "loss": 0.5207, + "step": 1825 + }, + { + "epoch": 0.7346610339971836, + "grad_norm": 4.16513812007776, + "learning_rate": 1.001997476187241e-06, + "loss": 0.5213, + "step": 1826 + }, + { + "epoch": 0.7350633675316838, + "grad_norm": 4.4208377164885055, + "learning_rate": 9.991880974611809e-07, + "loss": 0.6269, + "step": 1827 + }, + { + "epoch": 0.7354657010661839, + "grad_norm": 4.867032713419264, + "learning_rate": 9.96381678750524e-07, + "loss": 0.4808, + "step": 1828 + }, + { + "epoch": 0.735868034600684, + "grad_norm": 3.982545780740484, + "learning_rate": 9.935782255903034e-07, + "loss": 0.5066, + "step": 1829 + }, + { + "epoch": 0.7362703681351841, + "grad_norm": 4.286351965336647, + "learning_rate": 9.907777435097029e-07, + "loss": 0.5151, + "step": 1830 + }, + { + "epoch": 0.7366727016696841, + "grad_norm": 4.362229906545009, + "learning_rate": 9.879802380320473e-07, + "loss": 0.5936, + "step": 1831 + }, + { + "epoch": 0.7370750352041843, + "grad_norm": 4.660993080328005, + "learning_rate": 9.851857146747873e-07, + "loss": 0.6829, + "step": 1832 + }, + { + "epoch": 0.7374773687386844, + "grad_norm": 4.222997821431192, + "learning_rate": 9.82394178949497e-07, + "loss": 0.514, + "step": 1833 + }, + { + "epoch": 0.7378797022731844, + "grad_norm": 4.369792251301722, + "learning_rate": 9.796056363618533e-07, + "loss": 0.5742, + "step": 1834 + }, + { + "epoch": 0.7382820358076846, + "grad_norm": 4.422672878513299, + "learning_rate": 9.768200924116338e-07, + "loss": 0.568, + "step": 1835 + }, + { + "epoch": 0.7386843693421846, + "grad_norm": 4.507594407110887, + "learning_rate": 9.740375525926988e-07, + "loss": 0.6017, + "step": 1836 + }, + { + "epoch": 0.7390867028766848, + "grad_norm": 3.8125967279749835, + "learning_rate": 9.71258022392985e-07, + "loss": 0.5959, + "step": 1837 + }, + { + "epoch": 0.7394890364111849, + "grad_norm": 3.7273585439154586, + "learning_rate": 9.684815072944946e-07, + "loss": 0.3825, + "step": 1838 + }, + { + "epoch": 0.7398913699456849, + "grad_norm": 3.738805120372397, + "learning_rate": 9.657080127732807e-07, + "loss": 0.4931, + "step": 1839 + }, + { + "epoch": 0.7402937034801851, + "grad_norm": 3.9828190911127526, + "learning_rate": 9.629375442994418e-07, + "loss": 0.5182, + "step": 1840 + }, + { + "epoch": 0.7406960370146852, + "grad_norm": 3.9798516908695882, + "learning_rate": 9.601701073371045e-07, + "loss": 0.548, + "step": 1841 + }, + { + "epoch": 0.7410983705491853, + "grad_norm": 3.7199861649210715, + "learning_rate": 9.574057073444195e-07, + "loss": 0.468, + "step": 1842 + }, + { + "epoch": 0.7415007040836854, + "grad_norm": 4.0680484984054335, + "learning_rate": 9.546443497735467e-07, + "loss": 0.5782, + "step": 1843 + }, + { + "epoch": 0.7419030376181854, + "grad_norm": 3.4109362486846115, + "learning_rate": 9.518860400706465e-07, + "loss": 0.4245, + "step": 1844 + }, + { + "epoch": 0.7423053711526856, + "grad_norm": 3.931799967484987, + "learning_rate": 9.491307836758651e-07, + "loss": 0.4863, + "step": 1845 + }, + { + "epoch": 0.7427077046871857, + "grad_norm": 4.211928044347116, + "learning_rate": 9.463785860233301e-07, + "loss": 0.5459, + "step": 1846 + }, + { + "epoch": 0.7431100382216858, + "grad_norm": 4.0902800868050715, + "learning_rate": 9.436294525411357e-07, + "loss": 0.6375, + "step": 1847 + }, + { + "epoch": 0.7435123717561859, + "grad_norm": 3.524293359481542, + "learning_rate": 9.408833886513303e-07, + "loss": 0.444, + "step": 1848 + }, + { + "epoch": 0.7439147052906859, + "grad_norm": 3.6061388537955827, + "learning_rate": 9.381403997699117e-07, + "loss": 0.5217, + "step": 1849 + }, + { + "epoch": 0.7443170388251861, + "grad_norm": 3.6547603411472838, + "learning_rate": 9.354004913068096e-07, + "loss": 0.5229, + "step": 1850 + }, + { + "epoch": 0.7447193723596862, + "grad_norm": 4.340731245011316, + "learning_rate": 9.326636686658808e-07, + "loss": 0.5924, + "step": 1851 + }, + { + "epoch": 0.7451217058941862, + "grad_norm": 3.4273270658083166, + "learning_rate": 9.299299372448953e-07, + "loss": 0.4941, + "step": 1852 + }, + { + "epoch": 0.7455240394286864, + "grad_norm": 3.9330296000490668, + "learning_rate": 9.271993024355263e-07, + "loss": 0.5988, + "step": 1853 + }, + { + "epoch": 0.7459263729631865, + "grad_norm": 4.79047981984662, + "learning_rate": 9.244717696233388e-07, + "loss": 0.6193, + "step": 1854 + }, + { + "epoch": 0.7463287064976866, + "grad_norm": 3.6277035412847924, + "learning_rate": 9.217473441877805e-07, + "loss": 0.4905, + "step": 1855 + }, + { + "epoch": 0.7467310400321867, + "grad_norm": 4.652255454626382, + "learning_rate": 9.19026031502171e-07, + "loss": 0.6231, + "step": 1856 + }, + { + "epoch": 0.7471333735666867, + "grad_norm": 4.139619345656057, + "learning_rate": 9.1630783693369e-07, + "loss": 0.4981, + "step": 1857 + }, + { + "epoch": 0.7475357071011869, + "grad_norm": 5.059383612132635, + "learning_rate": 9.135927658433685e-07, + "loss": 0.7292, + "step": 1858 + }, + { + "epoch": 0.747938040635687, + "grad_norm": 4.664959181536374, + "learning_rate": 9.108808235860744e-07, + "loss": 0.6118, + "step": 1859 + }, + { + "epoch": 0.7483403741701871, + "grad_norm": 4.243108848129753, + "learning_rate": 9.081720155105076e-07, + "loss": 0.5616, + "step": 1860 + }, + { + "epoch": 0.7487427077046872, + "grad_norm": 3.4693576707426548, + "learning_rate": 9.054663469591862e-07, + "loss": 0.4451, + "step": 1861 + }, + { + "epoch": 0.7491450412391872, + "grad_norm": 4.037761942838199, + "learning_rate": 9.027638232684339e-07, + "loss": 0.5354, + "step": 1862 + }, + { + "epoch": 0.7495473747736874, + "grad_norm": 4.037140323875402, + "learning_rate": 9.00064449768375e-07, + "loss": 0.3946, + "step": 1863 + }, + { + "epoch": 0.7499497083081875, + "grad_norm": 4.420498117292749, + "learning_rate": 8.97368231782918e-07, + "loss": 0.6091, + "step": 1864 + }, + { + "epoch": 0.7503520418426876, + "grad_norm": 4.098542987137017, + "learning_rate": 8.946751746297494e-07, + "loss": 0.6347, + "step": 1865 + }, + { + "epoch": 0.7507543753771877, + "grad_norm": 3.718783314907544, + "learning_rate": 8.919852836203224e-07, + "loss": 0.5523, + "step": 1866 + }, + { + "epoch": 0.7511567089116878, + "grad_norm": 4.238567081312867, + "learning_rate": 8.892985640598434e-07, + "loss": 0.5716, + "step": 1867 + }, + { + "epoch": 0.7515590424461879, + "grad_norm": 3.878960051531824, + "learning_rate": 8.866150212472657e-07, + "loss": 0.5552, + "step": 1868 + }, + { + "epoch": 0.751961375980688, + "grad_norm": 4.0405823701369155, + "learning_rate": 8.839346604752763e-07, + "loss": 0.5286, + "step": 1869 + }, + { + "epoch": 0.752363709515188, + "grad_norm": 4.251404967808422, + "learning_rate": 8.81257487030287e-07, + "loss": 0.4872, + "step": 1870 + }, + { + "epoch": 0.7527660430496882, + "grad_norm": 3.6336762556932163, + "learning_rate": 8.785835061924234e-07, + "loss": 0.4301, + "step": 1871 + }, + { + "epoch": 0.7531683765841883, + "grad_norm": 4.324873818573403, + "learning_rate": 8.759127232355128e-07, + "loss": 0.5405, + "step": 1872 + }, + { + "epoch": 0.7535707101186884, + "grad_norm": 4.42447965555661, + "learning_rate": 8.732451434270767e-07, + "loss": 0.5209, + "step": 1873 + }, + { + "epoch": 0.7539730436531885, + "grad_norm": 4.251290292687751, + "learning_rate": 8.705807720283199e-07, + "loss": 0.5951, + "step": 1874 + }, + { + "epoch": 0.7543753771876885, + "grad_norm": 4.171584555966765, + "learning_rate": 8.679196142941173e-07, + "loss": 0.6086, + "step": 1875 + }, + { + "epoch": 0.7547777107221887, + "grad_norm": 4.0958958819445614, + "learning_rate": 8.652616754730075e-07, + "loss": 0.5693, + "step": 1876 + }, + { + "epoch": 0.7551800442566888, + "grad_norm": 3.683062028222347, + "learning_rate": 8.626069608071785e-07, + "loss": 0.4584, + "step": 1877 + }, + { + "epoch": 0.7555823777911889, + "grad_norm": 3.9012710111822924, + "learning_rate": 8.599554755324613e-07, + "loss": 0.4611, + "step": 1878 + }, + { + "epoch": 0.755984711325689, + "grad_norm": 3.5489182194802704, + "learning_rate": 8.573072248783165e-07, + "loss": 0.5083, + "step": 1879 + }, + { + "epoch": 0.7563870448601892, + "grad_norm": 3.9884722155489225, + "learning_rate": 8.546622140678265e-07, + "loss": 0.5094, + "step": 1880 + }, + { + "epoch": 0.7567893783946892, + "grad_norm": 4.328049459143733, + "learning_rate": 8.520204483176814e-07, + "loss": 0.3656, + "step": 1881 + }, + { + "epoch": 0.7571917119291893, + "grad_norm": 3.9339309591347087, + "learning_rate": 8.493819328381731e-07, + "loss": 0.5317, + "step": 1882 + }, + { + "epoch": 0.7575940454636894, + "grad_norm": 4.2270154261756385, + "learning_rate": 8.467466728331828e-07, + "loss": 0.4588, + "step": 1883 + }, + { + "epoch": 0.7579963789981895, + "grad_norm": 3.303945186981914, + "learning_rate": 8.441146735001715e-07, + "loss": 0.4649, + "step": 1884 + }, + { + "epoch": 0.7583987125326896, + "grad_norm": 4.054188026080457, + "learning_rate": 8.414859400301667e-07, + "loss": 0.5645, + "step": 1885 + }, + { + "epoch": 0.7588010460671897, + "grad_norm": 4.17992148808867, + "learning_rate": 8.388604776077575e-07, + "loss": 0.5204, + "step": 1886 + }, + { + "epoch": 0.7592033796016898, + "grad_norm": 4.428173356861454, + "learning_rate": 8.362382914110804e-07, + "loss": 0.6007, + "step": 1887 + }, + { + "epoch": 0.7596057131361899, + "grad_norm": 4.255468940795954, + "learning_rate": 8.336193866118117e-07, + "loss": 0.5153, + "step": 1888 + }, + { + "epoch": 0.76000804667069, + "grad_norm": 4.025791541260584, + "learning_rate": 8.310037683751527e-07, + "loss": 0.4859, + "step": 1889 + }, + { + "epoch": 0.7604103802051901, + "grad_norm": 3.8573782279099533, + "learning_rate": 8.283914418598263e-07, + "loss": 0.5058, + "step": 1890 + }, + { + "epoch": 0.7608127137396902, + "grad_norm": 3.332235385366874, + "learning_rate": 8.257824122180602e-07, + "loss": 0.3737, + "step": 1891 + }, + { + "epoch": 0.7612150472741903, + "grad_norm": 3.394561752123056, + "learning_rate": 8.231766845955818e-07, + "loss": 0.4197, + "step": 1892 + }, + { + "epoch": 0.7616173808086905, + "grad_norm": 3.3171785759316443, + "learning_rate": 8.205742641316064e-07, + "loss": 0.417, + "step": 1893 + }, + { + "epoch": 0.7620197143431905, + "grad_norm": 4.487532016958629, + "learning_rate": 8.179751559588234e-07, + "loss": 0.5663, + "step": 1894 + }, + { + "epoch": 0.7624220478776906, + "grad_norm": 4.546271967177341, + "learning_rate": 8.153793652033931e-07, + "loss": 0.5597, + "step": 1895 + }, + { + "epoch": 0.7628243814121907, + "grad_norm": 3.914164846084329, + "learning_rate": 8.12786896984931e-07, + "loss": 0.4677, + "step": 1896 + }, + { + "epoch": 0.7632267149466908, + "grad_norm": 3.852819616304247, + "learning_rate": 8.101977564165011e-07, + "loss": 0.503, + "step": 1897 + }, + { + "epoch": 0.763629048481191, + "grad_norm": 4.067929177534588, + "learning_rate": 8.076119486046019e-07, + "loss": 0.5396, + "step": 1898 + }, + { + "epoch": 0.764031382015691, + "grad_norm": 4.010571344025329, + "learning_rate": 8.050294786491611e-07, + "loss": 0.4829, + "step": 1899 + }, + { + "epoch": 0.7644337155501911, + "grad_norm": 4.199207618578668, + "learning_rate": 8.024503516435222e-07, + "loss": 0.6123, + "step": 1900 + }, + { + "epoch": 0.7648360490846912, + "grad_norm": 4.313823267969129, + "learning_rate": 7.998745726744358e-07, + "loss": 0.5597, + "step": 1901 + }, + { + "epoch": 0.7652383826191913, + "grad_norm": 3.7661528466618956, + "learning_rate": 7.973021468220502e-07, + "loss": 0.5036, + "step": 1902 + }, + { + "epoch": 0.7656407161536914, + "grad_norm": 4.017310363498917, + "learning_rate": 7.947330791598973e-07, + "loss": 0.5435, + "step": 1903 + }, + { + "epoch": 0.7660430496881915, + "grad_norm": 4.447099011053462, + "learning_rate": 7.921673747548895e-07, + "loss": 0.4892, + "step": 1904 + }, + { + "epoch": 0.7664453832226916, + "grad_norm": 4.1491367299305555, + "learning_rate": 7.896050386673026e-07, + "loss": 0.5234, + "step": 1905 + }, + { + "epoch": 0.7668477167571918, + "grad_norm": 4.640959967611087, + "learning_rate": 7.870460759507722e-07, + "loss": 0.5619, + "step": 1906 + }, + { + "epoch": 0.7672500502916918, + "grad_norm": 4.207681163367058, + "learning_rate": 7.844904916522784e-07, + "loss": 0.6313, + "step": 1907 + }, + { + "epoch": 0.7676523838261919, + "grad_norm": 4.624725241802207, + "learning_rate": 7.819382908121386e-07, + "loss": 0.6169, + "step": 1908 + }, + { + "epoch": 0.768054717360692, + "grad_norm": 3.891844454436063, + "learning_rate": 7.793894784639986e-07, + "loss": 0.4932, + "step": 1909 + }, + { + "epoch": 0.7684570508951921, + "grad_norm": 3.6005349724480573, + "learning_rate": 7.768440596348192e-07, + "loss": 0.5835, + "step": 1910 + }, + { + "epoch": 0.7688593844296923, + "grad_norm": 4.073804566700674, + "learning_rate": 7.743020393448702e-07, + "loss": 0.4882, + "step": 1911 + }, + { + "epoch": 0.7692617179641923, + "grad_norm": 3.959969867245135, + "learning_rate": 7.717634226077156e-07, + "loss": 0.4945, + "step": 1912 + }, + { + "epoch": 0.7696640514986924, + "grad_norm": 3.599166244946097, + "learning_rate": 7.692282144302094e-07, + "loss": 0.3889, + "step": 1913 + }, + { + "epoch": 0.7700663850331925, + "grad_norm": 3.539375774753211, + "learning_rate": 7.666964198124819e-07, + "loss": 0.523, + "step": 1914 + }, + { + "epoch": 0.7704687185676926, + "grad_norm": 3.912187073942566, + "learning_rate": 7.64168043747932e-07, + "loss": 0.5567, + "step": 1915 + }, + { + "epoch": 0.7708710521021928, + "grad_norm": 4.034008046236846, + "learning_rate": 7.616430912232137e-07, + "loss": 0.5219, + "step": 1916 + }, + { + "epoch": 0.7712733856366928, + "grad_norm": 4.074005253663992, + "learning_rate": 7.59121567218232e-07, + "loss": 0.5302, + "step": 1917 + }, + { + "epoch": 0.7716757191711929, + "grad_norm": 3.4776790939949347, + "learning_rate": 7.566034767061265e-07, + "loss": 0.2992, + "step": 1918 + }, + { + "epoch": 0.7720780527056931, + "grad_norm": 3.755741404325393, + "learning_rate": 7.540888246532688e-07, + "loss": 0.4337, + "step": 1919 + }, + { + "epoch": 0.7724803862401931, + "grad_norm": 3.854464881913325, + "learning_rate": 7.515776160192454e-07, + "loss": 0.5286, + "step": 1920 + }, + { + "epoch": 0.7728827197746932, + "grad_norm": 3.74441109291879, + "learning_rate": 7.490698557568535e-07, + "loss": 0.5449, + "step": 1921 + }, + { + "epoch": 0.7732850533091933, + "grad_norm": 4.570844263947877, + "learning_rate": 7.465655488120891e-07, + "loss": 0.6153, + "step": 1922 + }, + { + "epoch": 0.7736873868436934, + "grad_norm": 3.7650085288496613, + "learning_rate": 7.440647001241363e-07, + "loss": 0.5166, + "step": 1923 + }, + { + "epoch": 0.7740897203781936, + "grad_norm": 3.6719483984817236, + "learning_rate": 7.415673146253604e-07, + "loss": 0.4842, + "step": 1924 + }, + { + "epoch": 0.7744920539126936, + "grad_norm": 4.057471839423865, + "learning_rate": 7.390733972412933e-07, + "loss": 0.6075, + "step": 1925 + }, + { + "epoch": 0.7748943874471937, + "grad_norm": 4.4626847789976924, + "learning_rate": 7.365829528906293e-07, + "loss": 0.5292, + "step": 1926 + }, + { + "epoch": 0.7752967209816938, + "grad_norm": 3.76677011259477, + "learning_rate": 7.340959864852124e-07, + "loss": 0.405, + "step": 1927 + }, + { + "epoch": 0.7756990545161939, + "grad_norm": 3.8469846124186082, + "learning_rate": 7.316125029300275e-07, + "loss": 0.4827, + "step": 1928 + }, + { + "epoch": 0.7761013880506941, + "grad_norm": 3.8371445392089827, + "learning_rate": 7.291325071231883e-07, + "loss": 0.5134, + "step": 1929 + }, + { + "epoch": 0.7765037215851941, + "grad_norm": 3.8383451450589887, + "learning_rate": 7.26656003955932e-07, + "loss": 0.4061, + "step": 1930 + }, + { + "epoch": 0.7769060551196942, + "grad_norm": 4.037666657124557, + "learning_rate": 7.241829983126075e-07, + "loss": 0.4976, + "step": 1931 + }, + { + "epoch": 0.7773083886541944, + "grad_norm": 3.980302342869356, + "learning_rate": 7.217134950706625e-07, + "loss": 0.5149, + "step": 1932 + }, + { + "epoch": 0.7777107221886944, + "grad_norm": 3.7463100803001277, + "learning_rate": 7.192474991006416e-07, + "loss": 0.459, + "step": 1933 + }, + { + "epoch": 0.7781130557231946, + "grad_norm": 3.9126255555142597, + "learning_rate": 7.167850152661673e-07, + "loss": 0.5537, + "step": 1934 + }, + { + "epoch": 0.7785153892576946, + "grad_norm": 4.306777769767276, + "learning_rate": 7.143260484239386e-07, + "loss": 0.4997, + "step": 1935 + }, + { + "epoch": 0.7789177227921947, + "grad_norm": 4.472035583740821, + "learning_rate": 7.118706034237172e-07, + "loss": 0.6373, + "step": 1936 + }, + { + "epoch": 0.7793200563266949, + "grad_norm": 3.8082808763272356, + "learning_rate": 7.094186851083187e-07, + "loss": 0.5713, + "step": 1937 + }, + { + "epoch": 0.7797223898611949, + "grad_norm": 3.907355998411772, + "learning_rate": 7.069702983136023e-07, + "loss": 0.4916, + "step": 1938 + }, + { + "epoch": 0.780124723395695, + "grad_norm": 3.4404115391335623, + "learning_rate": 7.045254478684626e-07, + "loss": 0.4552, + "step": 1939 + }, + { + "epoch": 0.7805270569301951, + "grad_norm": 4.072479874224982, + "learning_rate": 7.020841385948199e-07, + "loss": 0.5061, + "step": 1940 + }, + { + "epoch": 0.7809293904646952, + "grad_norm": 5.014496338999608, + "learning_rate": 6.996463753076102e-07, + "loss": 0.5401, + "step": 1941 + }, + { + "epoch": 0.7813317239991954, + "grad_norm": 4.5982221422710134, + "learning_rate": 6.972121628147766e-07, + "loss": 0.6385, + "step": 1942 + }, + { + "epoch": 0.7817340575336954, + "grad_norm": 3.866309336941733, + "learning_rate": 6.94781505917256e-07, + "loss": 0.5007, + "step": 1943 + }, + { + "epoch": 0.7821363910681955, + "grad_norm": 4.000641087223532, + "learning_rate": 6.923544094089762e-07, + "loss": 0.4097, + "step": 1944 + }, + { + "epoch": 0.7825387246026957, + "grad_norm": 4.183627371809607, + "learning_rate": 6.899308780768423e-07, + "loss": 0.4986, + "step": 1945 + }, + { + "epoch": 0.7829410581371957, + "grad_norm": 3.834784873841407, + "learning_rate": 6.875109167007255e-07, + "loss": 0.4719, + "step": 1946 + }, + { + "epoch": 0.7833433916716959, + "grad_norm": 4.624045837237235, + "learning_rate": 6.85094530053459e-07, + "loss": 0.5742, + "step": 1947 + }, + { + "epoch": 0.7837457252061959, + "grad_norm": 4.968557042198568, + "learning_rate": 6.826817229008232e-07, + "loss": 0.6943, + "step": 1948 + }, + { + "epoch": 0.784148058740696, + "grad_norm": 3.798467897259186, + "learning_rate": 6.802725000015406e-07, + "loss": 0.5297, + "step": 1949 + }, + { + "epoch": 0.7845503922751962, + "grad_norm": 3.974644659610978, + "learning_rate": 6.778668661072643e-07, + "loss": 0.4803, + "step": 1950 + }, + { + "epoch": 0.7849527258096962, + "grad_norm": 4.048333500653427, + "learning_rate": 6.754648259625673e-07, + "loss": 0.5941, + "step": 1951 + }, + { + "epoch": 0.7853550593441964, + "grad_norm": 4.3035958811807875, + "learning_rate": 6.730663843049367e-07, + "loss": 0.5955, + "step": 1952 + }, + { + "epoch": 0.7857573928786964, + "grad_norm": 3.6544635394961764, + "learning_rate": 6.706715458647615e-07, + "loss": 0.4851, + "step": 1953 + }, + { + "epoch": 0.7861597264131965, + "grad_norm": 4.218917868283341, + "learning_rate": 6.682803153653241e-07, + "loss": 0.6081, + "step": 1954 + }, + { + "epoch": 0.7865620599476967, + "grad_norm": 3.617134711107577, + "learning_rate": 6.658926975227923e-07, + "loss": 0.4414, + "step": 1955 + }, + { + "epoch": 0.7869643934821967, + "grad_norm": 4.041925903109421, + "learning_rate": 6.635086970462057e-07, + "loss": 0.4589, + "step": 1956 + }, + { + "epoch": 0.7873667270166969, + "grad_norm": 3.9272206421591944, + "learning_rate": 6.611283186374723e-07, + "loss": 0.4697, + "step": 1957 + }, + { + "epoch": 0.7877690605511969, + "grad_norm": 4.242979124153208, + "learning_rate": 6.587515669913558e-07, + "loss": 0.4185, + "step": 1958 + }, + { + "epoch": 0.788171394085697, + "grad_norm": 3.38359591341345, + "learning_rate": 6.563784467954668e-07, + "loss": 0.3828, + "step": 1959 + }, + { + "epoch": 0.7885737276201972, + "grad_norm": 4.348954859955942, + "learning_rate": 6.540089627302523e-07, + "loss": 0.5046, + "step": 1960 + }, + { + "epoch": 0.7889760611546972, + "grad_norm": 4.120739882383607, + "learning_rate": 6.516431194689907e-07, + "loss": 0.5227, + "step": 1961 + }, + { + "epoch": 0.7893783946891973, + "grad_norm": 3.6740741354898305, + "learning_rate": 6.492809216777762e-07, + "loss": 0.4134, + "step": 1962 + }, + { + "epoch": 0.7897807282236975, + "grad_norm": 3.690318810926642, + "learning_rate": 6.469223740155159e-07, + "loss": 0.4901, + "step": 1963 + }, + { + "epoch": 0.7901830617581975, + "grad_norm": 4.2381781760809485, + "learning_rate": 6.445674811339175e-07, + "loss": 0.4225, + "step": 1964 + }, + { + "epoch": 0.7905853952926977, + "grad_norm": 4.106368765680081, + "learning_rate": 6.422162476774788e-07, + "loss": 0.5645, + "step": 1965 + }, + { + "epoch": 0.7909877288271977, + "grad_norm": 3.706517698008605, + "learning_rate": 6.398686782834815e-07, + "loss": 0.4995, + "step": 1966 + }, + { + "epoch": 0.7913900623616978, + "grad_norm": 4.432403407342872, + "learning_rate": 6.375247775819809e-07, + "loss": 0.5127, + "step": 1967 + }, + { + "epoch": 0.791792395896198, + "grad_norm": 4.022769337353194, + "learning_rate": 6.351845501957971e-07, + "loss": 0.4525, + "step": 1968 + }, + { + "epoch": 0.792194729430698, + "grad_norm": 4.43806071523494, + "learning_rate": 6.328480007405027e-07, + "loss": 0.5067, + "step": 1969 + }, + { + "epoch": 0.7925970629651982, + "grad_norm": 3.7265320273813027, + "learning_rate": 6.305151338244192e-07, + "loss": 0.4277, + "step": 1970 + }, + { + "epoch": 0.7929993964996982, + "grad_norm": 4.06221248413846, + "learning_rate": 6.281859540486043e-07, + "loss": 0.6297, + "step": 1971 + }, + { + "epoch": 0.7934017300341983, + "grad_norm": 4.391919257148749, + "learning_rate": 6.258604660068443e-07, + "loss": 0.4994, + "step": 1972 + }, + { + "epoch": 0.7938040635686985, + "grad_norm": 4.252838693305146, + "learning_rate": 6.23538674285642e-07, + "loss": 0.5949, + "step": 1973 + }, + { + "epoch": 0.7942063971031985, + "grad_norm": 4.325947652020019, + "learning_rate": 6.212205834642132e-07, + "loss": 0.5428, + "step": 1974 + }, + { + "epoch": 0.7946087306376987, + "grad_norm": 4.185176650642697, + "learning_rate": 6.189061981144715e-07, + "loss": 0.5732, + "step": 1975 + }, + { + "epoch": 0.7950110641721988, + "grad_norm": 3.2994931408989627, + "learning_rate": 6.165955228010242e-07, + "loss": 0.37, + "step": 1976 + }, + { + "epoch": 0.7954133977066988, + "grad_norm": 3.9581400842405277, + "learning_rate": 6.14288562081162e-07, + "loss": 0.6003, + "step": 1977 + }, + { + "epoch": 0.795815731241199, + "grad_norm": 3.855839195313672, + "learning_rate": 6.11985320504847e-07, + "loss": 0.5755, + "step": 1978 + }, + { + "epoch": 0.796218064775699, + "grad_norm": 4.28483670710571, + "learning_rate": 6.096858026147079e-07, + "loss": 0.5674, + "step": 1979 + }, + { + "epoch": 0.7966203983101992, + "grad_norm": 3.946092310322622, + "learning_rate": 6.073900129460286e-07, + "loss": 0.4429, + "step": 1980 + }, + { + "epoch": 0.7970227318446993, + "grad_norm": 3.9442626812628725, + "learning_rate": 6.050979560267414e-07, + "loss": 0.4953, + "step": 1981 + }, + { + "epoch": 0.7974250653791993, + "grad_norm": 4.023203847345498, + "learning_rate": 6.028096363774136e-07, + "loss": 0.5437, + "step": 1982 + }, + { + "epoch": 0.7978273989136995, + "grad_norm": 4.5536613551126335, + "learning_rate": 6.005250585112437e-07, + "loss": 0.6177, + "step": 1983 + }, + { + "epoch": 0.7982297324481995, + "grad_norm": 3.8312030174786313, + "learning_rate": 5.982442269340505e-07, + "loss": 0.4801, + "step": 1984 + }, + { + "epoch": 0.7986320659826996, + "grad_norm": 3.8614656531277896, + "learning_rate": 5.959671461442634e-07, + "loss": 0.4395, + "step": 1985 + }, + { + "epoch": 0.7990343995171998, + "grad_norm": 4.191209006767766, + "learning_rate": 5.936938206329148e-07, + "loss": 0.4913, + "step": 1986 + }, + { + "epoch": 0.7994367330516998, + "grad_norm": 4.201209209604027, + "learning_rate": 5.914242548836291e-07, + "loss": 0.567, + "step": 1987 + }, + { + "epoch": 0.7998390665862, + "grad_norm": 4.711191490796992, + "learning_rate": 5.891584533726177e-07, + "loss": 0.6445, + "step": 1988 + }, + { + "epoch": 0.8002414001207001, + "grad_norm": 4.050282830317415, + "learning_rate": 5.868964205686653e-07, + "loss": 0.4854, + "step": 1989 + }, + { + "epoch": 0.8006437336552001, + "grad_norm": 3.4477078683288838, + "learning_rate": 5.84638160933127e-07, + "loss": 0.4578, + "step": 1990 + }, + { + "epoch": 0.8010460671897003, + "grad_norm": 3.977635083967608, + "learning_rate": 5.82383678919912e-07, + "loss": 0.4966, + "step": 1991 + }, + { + "epoch": 0.8014484007242003, + "grad_norm": 4.254984216161458, + "learning_rate": 5.801329789754826e-07, + "loss": 0.5101, + "step": 1992 + }, + { + "epoch": 0.8018507342587005, + "grad_norm": 4.1472621686996005, + "learning_rate": 5.778860655388399e-07, + "loss": 0.6366, + "step": 1993 + }, + { + "epoch": 0.8022530677932006, + "grad_norm": 4.0712688803226715, + "learning_rate": 5.75642943041518e-07, + "loss": 0.4908, + "step": 1994 + }, + { + "epoch": 0.8026554013277006, + "grad_norm": 4.69998065888643, + "learning_rate": 5.734036159075743e-07, + "loss": 0.6769, + "step": 1995 + }, + { + "epoch": 0.8030577348622008, + "grad_norm": 3.5455182945935144, + "learning_rate": 5.711680885535784e-07, + "loss": 0.4177, + "step": 1996 + }, + { + "epoch": 0.8034600683967008, + "grad_norm": 3.853566513480482, + "learning_rate": 5.689363653886085e-07, + "loss": 0.4434, + "step": 1997 + }, + { + "epoch": 0.803862401931201, + "grad_norm": 4.050333441948282, + "learning_rate": 5.667084508142386e-07, + "loss": 0.4822, + "step": 1998 + }, + { + "epoch": 0.8042647354657011, + "grad_norm": 4.442980318720399, + "learning_rate": 5.644843492245322e-07, + "loss": 0.4872, + "step": 1999 + }, + { + "epoch": 0.8046670690002011, + "grad_norm": 4.457653531733048, + "learning_rate": 5.622640650060301e-07, + "loss": 0.6444, + "step": 2000 + }, + { + "epoch": 0.8050694025347013, + "grad_norm": 3.921778521793536, + "learning_rate": 5.600476025377463e-07, + "loss": 0.5391, + "step": 2001 + }, + { + "epoch": 0.8054717360692014, + "grad_norm": 3.8081912666035755, + "learning_rate": 5.578349661911578e-07, + "loss": 0.4802, + "step": 2002 + }, + { + "epoch": 0.8058740696037014, + "grad_norm": 4.001868301091059, + "learning_rate": 5.556261603301924e-07, + "loss": 0.5459, + "step": 2003 + }, + { + "epoch": 0.8062764031382016, + "grad_norm": 3.847308434546458, + "learning_rate": 5.534211893112268e-07, + "loss": 0.4606, + "step": 2004 + }, + { + "epoch": 0.8066787366727016, + "grad_norm": 4.15027221954597, + "learning_rate": 5.512200574830712e-07, + "loss": 0.5705, + "step": 2005 + }, + { + "epoch": 0.8070810702072018, + "grad_norm": 3.8098780638264027, + "learning_rate": 5.490227691869656e-07, + "loss": 0.491, + "step": 2006 + }, + { + "epoch": 0.8074834037417019, + "grad_norm": 3.9362205164209616, + "learning_rate": 5.468293287565696e-07, + "loss": 0.4342, + "step": 2007 + }, + { + "epoch": 0.8078857372762019, + "grad_norm": 3.621695424225381, + "learning_rate": 5.446397405179534e-07, + "loss": 0.4943, + "step": 2008 + }, + { + "epoch": 0.8082880708107021, + "grad_norm": 4.894139218776488, + "learning_rate": 5.424540087895888e-07, + "loss": 0.7108, + "step": 2009 + }, + { + "epoch": 0.8086904043452021, + "grad_norm": 4.43306730596569, + "learning_rate": 5.402721378823422e-07, + "loss": 0.5759, + "step": 2010 + }, + { + "epoch": 0.8090927378797023, + "grad_norm": 4.303012456014276, + "learning_rate": 5.380941320994659e-07, + "loss": 0.575, + "step": 2011 + }, + { + "epoch": 0.8094950714142024, + "grad_norm": 4.313613214512251, + "learning_rate": 5.35919995736589e-07, + "loss": 0.4761, + "step": 2012 + }, + { + "epoch": 0.8098974049487024, + "grad_norm": 3.8521996380430394, + "learning_rate": 5.337497330817074e-07, + "loss": 0.5352, + "step": 2013 + }, + { + "epoch": 0.8102997384832026, + "grad_norm": 4.528703112399277, + "learning_rate": 5.315833484151786e-07, + "loss": 0.5959, + "step": 2014 + }, + { + "epoch": 0.8107020720177027, + "grad_norm": 3.878303004429545, + "learning_rate": 5.294208460097125e-07, + "loss": 0.4815, + "step": 2015 + }, + { + "epoch": 0.8111044055522028, + "grad_norm": 4.490283907824032, + "learning_rate": 5.272622301303587e-07, + "loss": 0.6368, + "step": 2016 + }, + { + "epoch": 0.8115067390867029, + "grad_norm": 4.279104111581364, + "learning_rate": 5.251075050345056e-07, + "loss": 0.5216, + "step": 2017 + }, + { + "epoch": 0.8119090726212029, + "grad_norm": 3.6608651815923086, + "learning_rate": 5.229566749718645e-07, + "loss": 0.4809, + "step": 2018 + }, + { + "epoch": 0.8123114061557031, + "grad_norm": 4.666782295826111, + "learning_rate": 5.208097441844668e-07, + "loss": 0.504, + "step": 2019 + }, + { + "epoch": 0.8127137396902032, + "grad_norm": 3.84925143665575, + "learning_rate": 5.186667169066523e-07, + "loss": 0.5423, + "step": 2020 + }, + { + "epoch": 0.8131160732247033, + "grad_norm": 3.578156563304788, + "learning_rate": 5.165275973650633e-07, + "loss": 0.3936, + "step": 2021 + }, + { + "epoch": 0.8135184067592034, + "grad_norm": 4.074315611637044, + "learning_rate": 5.143923897786329e-07, + "loss": 0.6302, + "step": 2022 + }, + { + "epoch": 0.8139207402937034, + "grad_norm": 4.085845386699454, + "learning_rate": 5.122610983585808e-07, + "loss": 0.5789, + "step": 2023 + }, + { + "epoch": 0.8143230738282036, + "grad_norm": 4.000394354287652, + "learning_rate": 5.101337273084017e-07, + "loss": 0.4229, + "step": 2024 + }, + { + "epoch": 0.8147254073627037, + "grad_norm": 3.835079915243062, + "learning_rate": 5.080102808238585e-07, + "loss": 0.5005, + "step": 2025 + }, + { + "epoch": 0.8151277408972037, + "grad_norm": 3.590556700281588, + "learning_rate": 5.058907630929749e-07, + "loss": 0.4452, + "step": 2026 + }, + { + "epoch": 0.8155300744317039, + "grad_norm": 4.254640813185372, + "learning_rate": 5.037751782960234e-07, + "loss": 0.6328, + "step": 2027 + }, + { + "epoch": 0.815932407966204, + "grad_norm": 4.290085466017056, + "learning_rate": 5.016635306055223e-07, + "loss": 0.5785, + "step": 2028 + }, + { + "epoch": 0.8163347415007041, + "grad_norm": 4.159416054122212, + "learning_rate": 4.995558241862242e-07, + "loss": 0.5129, + "step": 2029 + }, + { + "epoch": 0.8167370750352042, + "grad_norm": 3.7061504996754153, + "learning_rate": 4.974520631951069e-07, + "loss": 0.5305, + "step": 2030 + }, + { + "epoch": 0.8171394085697042, + "grad_norm": 4.178397491082031, + "learning_rate": 4.953522517813689e-07, + "loss": 0.5855, + "step": 2031 + }, + { + "epoch": 0.8175417421042044, + "grad_norm": 4.154495627484645, + "learning_rate": 4.932563940864168e-07, + "loss": 0.4164, + "step": 2032 + }, + { + "epoch": 0.8179440756387045, + "grad_norm": 4.283562955009001, + "learning_rate": 4.911644942438614e-07, + "loss": 0.4942, + "step": 2033 + }, + { + "epoch": 0.8183464091732046, + "grad_norm": 3.307871102947408, + "learning_rate": 4.890765563795072e-07, + "loss": 0.4552, + "step": 2034 + }, + { + "epoch": 0.8187487427077047, + "grad_norm": 4.15735416688211, + "learning_rate": 4.86992584611343e-07, + "loss": 0.5797, + "step": 2035 + }, + { + "epoch": 0.8191510762422047, + "grad_norm": 4.111111195260256, + "learning_rate": 4.849125830495366e-07, + "loss": 0.5776, + "step": 2036 + }, + { + "epoch": 0.8195534097767049, + "grad_norm": 3.534887324822181, + "learning_rate": 4.828365557964257e-07, + "loss": 0.4913, + "step": 2037 + }, + { + "epoch": 0.819955743311205, + "grad_norm": 3.8702017938683606, + "learning_rate": 4.807645069465095e-07, + "loss": 0.4638, + "step": 2038 + }, + { + "epoch": 0.820358076845705, + "grad_norm": 4.105810885704208, + "learning_rate": 4.786964405864403e-07, + "loss": 0.51, + "step": 2039 + }, + { + "epoch": 0.8207604103802052, + "grad_norm": 3.7721413153025183, + "learning_rate": 4.766323607950149e-07, + "loss": 0.5327, + "step": 2040 + }, + { + "epoch": 0.8211627439147053, + "grad_norm": 4.604383899796593, + "learning_rate": 4.7457227164316914e-07, + "loss": 0.6425, + "step": 2041 + }, + { + "epoch": 0.8215650774492054, + "grad_norm": 4.7452724983070595, + "learning_rate": 4.725161771939679e-07, + "loss": 0.5968, + "step": 2042 + }, + { + "epoch": 0.8219674109837055, + "grad_norm": 4.13939509952924, + "learning_rate": 4.704640815025971e-07, + "loss": 0.5454, + "step": 2043 + }, + { + "epoch": 0.8223697445182055, + "grad_norm": 4.07811911040908, + "learning_rate": 4.684159886163553e-07, + "loss": 0.4846, + "step": 2044 + }, + { + "epoch": 0.8227720780527057, + "grad_norm": 4.223857527886817, + "learning_rate": 4.663719025746477e-07, + "loss": 0.548, + "step": 2045 + }, + { + "epoch": 0.8231744115872058, + "grad_norm": 4.055398221449843, + "learning_rate": 4.643318274089756e-07, + "loss": 0.4762, + "step": 2046 + }, + { + "epoch": 0.8235767451217059, + "grad_norm": 3.93292252326897, + "learning_rate": 4.6229576714293067e-07, + "loss": 0.5506, + "step": 2047 + }, + { + "epoch": 0.823979078656206, + "grad_norm": 3.5755501548250357, + "learning_rate": 4.6026372579218686e-07, + "loss": 0.5277, + "step": 2048 + }, + { + "epoch": 0.824381412190706, + "grad_norm": 3.900846576287911, + "learning_rate": 4.5823570736448913e-07, + "loss": 0.4289, + "step": 2049 + }, + { + "epoch": 0.8247837457252062, + "grad_norm": 4.422345897561269, + "learning_rate": 4.5621171585965046e-07, + "loss": 0.5439, + "step": 2050 + }, + { + "epoch": 0.8251860792597063, + "grad_norm": 3.86188238287213, + "learning_rate": 4.5419175526954074e-07, + "loss": 0.4971, + "step": 2051 + }, + { + "epoch": 0.8255884127942064, + "grad_norm": 4.149840183197456, + "learning_rate": 4.5217582957808064e-07, + "loss": 0.4913, + "step": 2052 + }, + { + "epoch": 0.8259907463287065, + "grad_norm": 4.14442670188834, + "learning_rate": 4.501639427612309e-07, + "loss": 0.593, + "step": 2053 + }, + { + "epoch": 0.8263930798632066, + "grad_norm": 4.160435006991155, + "learning_rate": 4.4815609878698847e-07, + "loss": 0.5702, + "step": 2054 + }, + { + "epoch": 0.8267954133977067, + "grad_norm": 4.699994325668033, + "learning_rate": 4.46152301615376e-07, + "loss": 0.6052, + "step": 2055 + }, + { + "epoch": 0.8271977469322068, + "grad_norm": 4.158216057441291, + "learning_rate": 4.441525551984349e-07, + "loss": 0.6375, + "step": 2056 + }, + { + "epoch": 0.8276000804667069, + "grad_norm": 5.052807884367178, + "learning_rate": 4.421568634802165e-07, + "loss": 0.687, + "step": 2057 + }, + { + "epoch": 0.828002414001207, + "grad_norm": 3.9353061661541133, + "learning_rate": 4.40165230396776e-07, + "loss": 0.4774, + "step": 2058 + }, + { + "epoch": 0.8284047475357071, + "grad_norm": 4.01193456780117, + "learning_rate": 4.3817765987616474e-07, + "loss": 0.4932, + "step": 2059 + }, + { + "epoch": 0.8288070810702072, + "grad_norm": 4.976729164334829, + "learning_rate": 4.361941558384189e-07, + "loss": 0.6807, + "step": 2060 + }, + { + "epoch": 0.8292094146047073, + "grad_norm": 4.339682582742787, + "learning_rate": 4.3421472219555725e-07, + "loss": 0.4727, + "step": 2061 + }, + { + "epoch": 0.8296117481392074, + "grad_norm": 4.503522087170257, + "learning_rate": 4.3223936285156836e-07, + "loss": 0.6022, + "step": 2062 + }, + { + "epoch": 0.8300140816737075, + "grad_norm": 4.009166626990359, + "learning_rate": 4.3026808170240686e-07, + "loss": 0.6318, + "step": 2063 + }, + { + "epoch": 0.8304164152082076, + "grad_norm": 4.433472277211533, + "learning_rate": 4.283008826359833e-07, + "loss": 0.6059, + "step": 2064 + }, + { + "epoch": 0.8308187487427077, + "grad_norm": 4.239654566860423, + "learning_rate": 4.263377695321577e-07, + "loss": 0.5319, + "step": 2065 + }, + { + "epoch": 0.8312210822772078, + "grad_norm": 4.33872236149242, + "learning_rate": 4.2437874626273024e-07, + "loss": 0.5192, + "step": 2066 + }, + { + "epoch": 0.831623415811708, + "grad_norm": 4.396517305178651, + "learning_rate": 4.2242381669143603e-07, + "loss": 0.6266, + "step": 2067 + }, + { + "epoch": 0.832025749346208, + "grad_norm": 4.213477790744039, + "learning_rate": 4.204729846739358e-07, + "loss": 0.6738, + "step": 2068 + }, + { + "epoch": 0.8324280828807081, + "grad_norm": 3.312219624247552, + "learning_rate": 4.1852625405780914e-07, + "loss": 0.4833, + "step": 2069 + }, + { + "epoch": 0.8328304164152082, + "grad_norm": 3.9392383731854257, + "learning_rate": 4.165836286825464e-07, + "loss": 0.4513, + "step": 2070 + }, + { + "epoch": 0.8332327499497083, + "grad_norm": 3.8224909723307734, + "learning_rate": 4.1464511237954056e-07, + "loss": 0.5628, + "step": 2071 + }, + { + "epoch": 0.8336350834842085, + "grad_norm": 4.5465506777870885, + "learning_rate": 4.1271070897208166e-07, + "loss": 0.5761, + "step": 2072 + }, + { + "epoch": 0.8340374170187085, + "grad_norm": 3.87942801566727, + "learning_rate": 4.107804222753464e-07, + "loss": 0.4364, + "step": 2073 + }, + { + "epoch": 0.8344397505532086, + "grad_norm": 3.901824805945848, + "learning_rate": 4.088542560963937e-07, + "loss": 0.4897, + "step": 2074 + }, + { + "epoch": 0.8348420840877087, + "grad_norm": 4.2369387169530155, + "learning_rate": 4.0693221423415447e-07, + "loss": 0.6186, + "step": 2075 + }, + { + "epoch": 0.8352444176222088, + "grad_norm": 4.1329987533853005, + "learning_rate": 4.0501430047942603e-07, + "loss": 0.5358, + "step": 2076 + }, + { + "epoch": 0.835646751156709, + "grad_norm": 4.533598883351018, + "learning_rate": 4.03100518614864e-07, + "loss": 0.5522, + "step": 2077 + }, + { + "epoch": 0.836049084691209, + "grad_norm": 3.998009974659786, + "learning_rate": 4.011908724149746e-07, + "loss": 0.3743, + "step": 2078 + }, + { + "epoch": 0.8364514182257091, + "grad_norm": 3.417865078970055, + "learning_rate": 3.9928536564610767e-07, + "loss": 0.4035, + "step": 2079 + }, + { + "epoch": 0.8368537517602093, + "grad_norm": 3.9749987307006345, + "learning_rate": 3.9738400206644766e-07, + "loss": 0.6477, + "step": 2080 + }, + { + "epoch": 0.8372560852947093, + "grad_norm": 4.037697255647064, + "learning_rate": 3.9548678542600886e-07, + "loss": 0.5661, + "step": 2081 + }, + { + "epoch": 0.8376584188292094, + "grad_norm": 4.431095797135675, + "learning_rate": 3.935937194666267e-07, + "loss": 0.4818, + "step": 2082 + }, + { + "epoch": 0.8380607523637095, + "grad_norm": 4.260045700942229, + "learning_rate": 3.9170480792194985e-07, + "loss": 0.4263, + "step": 2083 + }, + { + "epoch": 0.8384630858982096, + "grad_norm": 4.051159338313962, + "learning_rate": 3.8982005451743243e-07, + "loss": 0.487, + "step": 2084 + }, + { + "epoch": 0.8388654194327098, + "grad_norm": 3.505632578812268, + "learning_rate": 3.8793946297032926e-07, + "loss": 0.5173, + "step": 2085 + }, + { + "epoch": 0.8392677529672098, + "grad_norm": 3.4621570841739917, + "learning_rate": 3.8606303698968643e-07, + "loss": 0.3609, + "step": 2086 + }, + { + "epoch": 0.8396700865017099, + "grad_norm": 3.708654324006204, + "learning_rate": 3.841907802763331e-07, + "loss": 0.4437, + "step": 2087 + }, + { + "epoch": 0.84007242003621, + "grad_norm": 3.3369961693311447, + "learning_rate": 3.8232269652287713e-07, + "loss": 0.4295, + "step": 2088 + }, + { + "epoch": 0.8404747535707101, + "grad_norm": 3.8475631558462884, + "learning_rate": 3.8045878941369434e-07, + "loss": 0.5628, + "step": 2089 + }, + { + "epoch": 0.8408770871052103, + "grad_norm": 3.827602289260064, + "learning_rate": 3.7859906262492503e-07, + "loss": 0.5158, + "step": 2090 + }, + { + "epoch": 0.8412794206397103, + "grad_norm": 4.093693428637088, + "learning_rate": 3.7674351982446347e-07, + "loss": 0.4742, + "step": 2091 + }, + { + "epoch": 0.8416817541742104, + "grad_norm": 4.310046930338328, + "learning_rate": 3.7489216467195304e-07, + "loss": 0.6097, + "step": 2092 + }, + { + "epoch": 0.8420840877087106, + "grad_norm": 4.017527345482499, + "learning_rate": 3.730450008187761e-07, + "loss": 0.5075, + "step": 2093 + }, + { + "epoch": 0.8424864212432106, + "grad_norm": 3.946920016020563, + "learning_rate": 3.7120203190805034e-07, + "loss": 0.4275, + "step": 2094 + }, + { + "epoch": 0.8428887547777107, + "grad_norm": 4.145255268119006, + "learning_rate": 3.6936326157461925e-07, + "loss": 0.5914, + "step": 2095 + }, + { + "epoch": 0.8432910883122108, + "grad_norm": 4.2460232246744525, + "learning_rate": 3.6752869344504586e-07, + "loss": 0.5152, + "step": 2096 + }, + { + "epoch": 0.8436934218467109, + "grad_norm": 3.95033426326261, + "learning_rate": 3.656983311376042e-07, + "loss": 0.4448, + "step": 2097 + }, + { + "epoch": 0.8440957553812111, + "grad_norm": 3.8931938435117286, + "learning_rate": 3.638721782622745e-07, + "loss": 0.499, + "step": 2098 + }, + { + "epoch": 0.8444980889157111, + "grad_norm": 4.3118732957405985, + "learning_rate": 3.620502384207342e-07, + "loss": 0.49, + "step": 2099 + }, + { + "epoch": 0.8449004224502112, + "grad_norm": 4.005112629756995, + "learning_rate": 3.602325152063524e-07, + "loss": 0.4125, + "step": 2100 + }, + { + "epoch": 0.8453027559847113, + "grad_norm": 3.9812849673226114, + "learning_rate": 3.5841901220418e-07, + "loss": 0.5451, + "step": 2101 + }, + { + "epoch": 0.8457050895192114, + "grad_norm": 3.948178072792975, + "learning_rate": 3.566097329909468e-07, + "loss": 0.6427, + "step": 2102 + }, + { + "epoch": 0.8461074230537116, + "grad_norm": 3.7188722401214016, + "learning_rate": 3.5480468113504946e-07, + "loss": 0.405, + "step": 2103 + }, + { + "epoch": 0.8465097565882116, + "grad_norm": 3.913596983806647, + "learning_rate": 3.530038601965491e-07, + "loss": 0.4635, + "step": 2104 + }, + { + "epoch": 0.8469120901227117, + "grad_norm": 4.411578338116122, + "learning_rate": 3.5120727372716295e-07, + "loss": 0.5511, + "step": 2105 + }, + { + "epoch": 0.8473144236572119, + "grad_norm": 3.282083239679109, + "learning_rate": 3.494149252702536e-07, + "loss": 0.4441, + "step": 2106 + }, + { + "epoch": 0.8477167571917119, + "grad_norm": 4.011319326831284, + "learning_rate": 3.476268183608281e-07, + "loss": 0.5018, + "step": 2107 + }, + { + "epoch": 0.848119090726212, + "grad_norm": 4.284870919327636, + "learning_rate": 3.4584295652552667e-07, + "loss": 0.4846, + "step": 2108 + }, + { + "epoch": 0.8485214242607121, + "grad_norm": 4.276416896965254, + "learning_rate": 3.4406334328261765e-07, + "loss": 0.5376, + "step": 2109 + }, + { + "epoch": 0.8489237577952122, + "grad_norm": 3.834125175179826, + "learning_rate": 3.4228798214198974e-07, + "loss": 0.5651, + "step": 2110 + }, + { + "epoch": 0.8493260913297124, + "grad_norm": 3.9930841906813517, + "learning_rate": 3.405168766051445e-07, + "loss": 0.3851, + "step": 2111 + }, + { + "epoch": 0.8497284248642124, + "grad_norm": 4.2112351410317626, + "learning_rate": 3.3875003016519137e-07, + "loss": 0.5431, + "step": 2112 + }, + { + "epoch": 0.8501307583987125, + "grad_norm": 4.52788654812861, + "learning_rate": 3.369874463068398e-07, + "loss": 0.5849, + "step": 2113 + }, + { + "epoch": 0.8505330919332126, + "grad_norm": 3.8509825855331163, + "learning_rate": 3.3522912850639106e-07, + "loss": 0.5748, + "step": 2114 + }, + { + "epoch": 0.8509354254677127, + "grad_norm": 3.510509250668585, + "learning_rate": 3.334750802317338e-07, + "loss": 0.4397, + "step": 2115 + }, + { + "epoch": 0.8513377590022129, + "grad_norm": 3.8349389092945114, + "learning_rate": 3.3172530494233467e-07, + "loss": 0.4769, + "step": 2116 + }, + { + "epoch": 0.8517400925367129, + "grad_norm": 3.8288751826393197, + "learning_rate": 3.2997980608923426e-07, + "loss": 0.5434, + "step": 2117 + }, + { + "epoch": 0.852142426071213, + "grad_norm": 3.9001403159636374, + "learning_rate": 3.282385871150387e-07, + "loss": 0.4523, + "step": 2118 + }, + { + "epoch": 0.8525447596057132, + "grad_norm": 3.6933669568597978, + "learning_rate": 3.26501651453911e-07, + "loss": 0.3905, + "step": 2119 + }, + { + "epoch": 0.8529470931402132, + "grad_norm": 3.811615026103731, + "learning_rate": 3.247690025315689e-07, + "loss": 0.4305, + "step": 2120 + }, + { + "epoch": 0.8533494266747134, + "grad_norm": 3.6210800698038925, + "learning_rate": 3.2304064376527417e-07, + "loss": 0.5332, + "step": 2121 + }, + { + "epoch": 0.8537517602092134, + "grad_norm": 4.231885438858511, + "learning_rate": 3.2131657856382753e-07, + "loss": 0.5911, + "step": 2122 + }, + { + "epoch": 0.8541540937437135, + "grad_norm": 4.464935656077254, + "learning_rate": 3.195968103275618e-07, + "loss": 0.5702, + "step": 2123 + }, + { + "epoch": 0.8545564272782137, + "grad_norm": 4.3078684853827465, + "learning_rate": 3.17881342448334e-07, + "loss": 0.5691, + "step": 2124 + }, + { + "epoch": 0.8549587608127137, + "grad_norm": 3.8241264163503255, + "learning_rate": 3.1617017830952084e-07, + "loss": 0.5375, + "step": 2125 + }, + { + "epoch": 0.8553610943472139, + "grad_norm": 3.519525197941784, + "learning_rate": 3.1446332128601015e-07, + "loss": 0.4416, + "step": 2126 + }, + { + "epoch": 0.8557634278817139, + "grad_norm": 3.4697912989841524, + "learning_rate": 3.12760774744196e-07, + "loss": 0.4459, + "step": 2127 + }, + { + "epoch": 0.856165761416214, + "grad_norm": 3.7851796176468455, + "learning_rate": 3.1106254204196883e-07, + "loss": 0.4682, + "step": 2128 + }, + { + "epoch": 0.8565680949507142, + "grad_norm": 3.936606771811409, + "learning_rate": 3.093686265287138e-07, + "loss": 0.5729, + "step": 2129 + }, + { + "epoch": 0.8569704284852142, + "grad_norm": 4.190455136191216, + "learning_rate": 3.076790315452988e-07, + "loss": 0.5277, + "step": 2130 + }, + { + "epoch": 0.8573727620197144, + "grad_norm": 4.108892268475147, + "learning_rate": 3.05993760424072e-07, + "loss": 0.4329, + "step": 2131 + }, + { + "epoch": 0.8577750955542145, + "grad_norm": 4.390564938079403, + "learning_rate": 3.043128164888537e-07, + "loss": 0.5011, + "step": 2132 + }, + { + "epoch": 0.8581774290887145, + "grad_norm": 4.367475352759402, + "learning_rate": 3.0263620305492877e-07, + "loss": 0.4797, + "step": 2133 + }, + { + "epoch": 0.8585797626232147, + "grad_norm": 3.8444843612579924, + "learning_rate": 3.009639234290418e-07, + "loss": 0.5327, + "step": 2134 + }, + { + "epoch": 0.8589820961577147, + "grad_norm": 3.6090606638950318, + "learning_rate": 2.992959809093901e-07, + "loss": 0.4413, + "step": 2135 + }, + { + "epoch": 0.8593844296922148, + "grad_norm": 4.388230282282369, + "learning_rate": 2.976323787856175e-07, + "loss": 0.5717, + "step": 2136 + }, + { + "epoch": 0.859786763226715, + "grad_norm": 4.333473286328121, + "learning_rate": 2.9597312033880557e-07, + "loss": 0.6314, + "step": 2137 + }, + { + "epoch": 0.860189096761215, + "grad_norm": 3.890510795148687, + "learning_rate": 2.943182088414701e-07, + "loss": 0.433, + "step": 2138 + }, + { + "epoch": 0.8605914302957152, + "grad_norm": 4.234653691147445, + "learning_rate": 2.9266764755755406e-07, + "loss": 0.54, + "step": 2139 + }, + { + "epoch": 0.8609937638302152, + "grad_norm": 4.183758834087082, + "learning_rate": 2.910214397424202e-07, + "loss": 0.4845, + "step": 2140 + }, + { + "epoch": 0.8613960973647153, + "grad_norm": 3.799668423286151, + "learning_rate": 2.8937958864284353e-07, + "loss": 0.4333, + "step": 2141 + }, + { + "epoch": 0.8617984308992155, + "grad_norm": 4.133997388279442, + "learning_rate": 2.877420974970088e-07, + "loss": 0.4396, + "step": 2142 + }, + { + "epoch": 0.8622007644337155, + "grad_norm": 4.0525547445688614, + "learning_rate": 2.861089695345007e-07, + "loss": 0.4857, + "step": 2143 + }, + { + "epoch": 0.8626030979682157, + "grad_norm": 3.7880628956269904, + "learning_rate": 2.8448020797629735e-07, + "loss": 0.5046, + "step": 2144 + }, + { + "epoch": 0.8630054315027158, + "grad_norm": 4.143267859487349, + "learning_rate": 2.8285581603476747e-07, + "loss": 0.5291, + "step": 2145 + }, + { + "epoch": 0.8634077650372158, + "grad_norm": 3.8047765913147993, + "learning_rate": 2.8123579691365923e-07, + "loss": 0.5513, + "step": 2146 + }, + { + "epoch": 0.863810098571716, + "grad_norm": 4.041711864076377, + "learning_rate": 2.796201538080981e-07, + "loss": 0.448, + "step": 2147 + }, + { + "epoch": 0.864212432106216, + "grad_norm": 4.193683351365294, + "learning_rate": 2.7800888990457855e-07, + "loss": 0.5598, + "step": 2148 + }, + { + "epoch": 0.8646147656407162, + "grad_norm": 3.6613935569874583, + "learning_rate": 2.7640200838095793e-07, + "loss": 0.4067, + "step": 2149 + }, + { + "epoch": 0.8650170991752163, + "grad_norm": 4.104620267298875, + "learning_rate": 2.7479951240644953e-07, + "loss": 0.5648, + "step": 2150 + }, + { + "epoch": 0.8654194327097163, + "grad_norm": 3.7600215255952256, + "learning_rate": 2.7320140514161846e-07, + "loss": 0.3946, + "step": 2151 + }, + { + "epoch": 0.8658217662442165, + "grad_norm": 3.9525358112645965, + "learning_rate": 2.7160768973837293e-07, + "loss": 0.53, + "step": 2152 + }, + { + "epoch": 0.8662240997787165, + "grad_norm": 3.5450533427675226, + "learning_rate": 2.7001836933995995e-07, + "loss": 0.5298, + "step": 2153 + }, + { + "epoch": 0.8666264333132166, + "grad_norm": 3.6668666878628815, + "learning_rate": 2.684334470809588e-07, + "loss": 0.5056, + "step": 2154 + }, + { + "epoch": 0.8670287668477168, + "grad_norm": 4.224237505147509, + "learning_rate": 2.6685292608727236e-07, + "loss": 0.5218, + "step": 2155 + }, + { + "epoch": 0.8674311003822168, + "grad_norm": 3.69557439406065, + "learning_rate": 2.652768094761257e-07, + "loss": 0.3756, + "step": 2156 + }, + { + "epoch": 0.867833433916717, + "grad_norm": 4.281090167368473, + "learning_rate": 2.637051003560548e-07, + "loss": 0.6402, + "step": 2157 + }, + { + "epoch": 0.8682357674512171, + "grad_norm": 3.9633212522861054, + "learning_rate": 2.621378018269047e-07, + "loss": 0.5075, + "step": 2158 + }, + { + "epoch": 0.8686381009857171, + "grad_norm": 3.8583945548771825, + "learning_rate": 2.6057491697982e-07, + "loss": 0.6033, + "step": 2159 + }, + { + "epoch": 0.8690404345202173, + "grad_norm": 3.3558856738663487, + "learning_rate": 2.5901644889724186e-07, + "loss": 0.4015, + "step": 2160 + }, + { + "epoch": 0.8694427680547173, + "grad_norm": 3.5749358804434483, + "learning_rate": 2.574624006528992e-07, + "loss": 0.4944, + "step": 2161 + }, + { + "epoch": 0.8698451015892175, + "grad_norm": 3.86587826976141, + "learning_rate": 2.55912775311804e-07, + "loss": 0.4891, + "step": 2162 + }, + { + "epoch": 0.8702474351237176, + "grad_norm": 3.9917581972004785, + "learning_rate": 2.543675759302464e-07, + "loss": 0.5079, + "step": 2163 + }, + { + "epoch": 0.8706497686582176, + "grad_norm": 4.393976999836958, + "learning_rate": 2.5282680555578477e-07, + "loss": 0.6145, + "step": 2164 + }, + { + "epoch": 0.8710521021927178, + "grad_norm": 4.04228191167184, + "learning_rate": 2.51290467227244e-07, + "loss": 0.5429, + "step": 2165 + }, + { + "epoch": 0.8714544357272178, + "grad_norm": 4.4282484060475324, + "learning_rate": 2.497585639747077e-07, + "loss": 0.6325, + "step": 2166 + }, + { + "epoch": 0.871856769261718, + "grad_norm": 3.816275160612967, + "learning_rate": 2.482310988195125e-07, + "loss": 0.5543, + "step": 2167 + }, + { + "epoch": 0.8722591027962181, + "grad_norm": 4.1145602142866675, + "learning_rate": 2.4670807477424007e-07, + "loss": 0.5527, + "step": 2168 + }, + { + "epoch": 0.8726614363307181, + "grad_norm": 4.231321602541556, + "learning_rate": 2.4518949484271525e-07, + "loss": 0.6496, + "step": 2169 + }, + { + "epoch": 0.8730637698652183, + "grad_norm": 4.220523454555819, + "learning_rate": 2.4367536201999695e-07, + "loss": 0.5652, + "step": 2170 + }, + { + "epoch": 0.8734661033997184, + "grad_norm": 3.938889256449136, + "learning_rate": 2.421656792923724e-07, + "loss": 0.456, + "step": 2171 + }, + { + "epoch": 0.8738684369342185, + "grad_norm": 4.5217908297671485, + "learning_rate": 2.406604496373535e-07, + "loss": 0.6334, + "step": 2172 + }, + { + "epoch": 0.8742707704687186, + "grad_norm": 3.7526929266944773, + "learning_rate": 2.391596760236681e-07, + "loss": 0.4743, + "step": 2173 + }, + { + "epoch": 0.8746731040032186, + "grad_norm": 3.710705344410006, + "learning_rate": 2.376633614112561e-07, + "loss": 0.5661, + "step": 2174 + }, + { + "epoch": 0.8750754375377188, + "grad_norm": 3.2609967694053243, + "learning_rate": 2.361715087512631e-07, + "loss": 0.3589, + "step": 2175 + }, + { + "epoch": 0.8754777710722189, + "grad_norm": 4.108101136746485, + "learning_rate": 2.346841209860351e-07, + "loss": 0.4368, + "step": 2176 + }, + { + "epoch": 0.875880104606719, + "grad_norm": 3.763267262463342, + "learning_rate": 2.3320120104911053e-07, + "loss": 0.4953, + "step": 2177 + }, + { + "epoch": 0.8762824381412191, + "grad_norm": 5.025294007587999, + "learning_rate": 2.317227518652168e-07, + "loss": 0.6524, + "step": 2178 + }, + { + "epoch": 0.8766847716757191, + "grad_norm": 3.9367012271057584, + "learning_rate": 2.302487763502645e-07, + "loss": 0.4149, + "step": 2179 + }, + { + "epoch": 0.8770871052102193, + "grad_norm": 4.236893564539156, + "learning_rate": 2.2877927741134036e-07, + "loss": 0.5532, + "step": 2180 + }, + { + "epoch": 0.8774894387447194, + "grad_norm": 3.5424218920413373, + "learning_rate": 2.2731425794670135e-07, + "loss": 0.4691, + "step": 2181 + }, + { + "epoch": 0.8778917722792194, + "grad_norm": 4.04407474745495, + "learning_rate": 2.2585372084577033e-07, + "loss": 0.5125, + "step": 2182 + }, + { + "epoch": 0.8782941058137196, + "grad_norm": 3.8612028242647147, + "learning_rate": 2.2439766898913e-07, + "loss": 0.5256, + "step": 2183 + }, + { + "epoch": 0.8786964393482197, + "grad_norm": 4.271299959657279, + "learning_rate": 2.2294610524851685e-07, + "loss": 0.4944, + "step": 2184 + }, + { + "epoch": 0.8790987728827198, + "grad_norm": 4.465135814281888, + "learning_rate": 2.2149903248681464e-07, + "loss": 0.6834, + "step": 2185 + }, + { + "epoch": 0.8795011064172199, + "grad_norm": 3.9117028793633173, + "learning_rate": 2.2005645355805127e-07, + "loss": 0.4605, + "step": 2186 + }, + { + "epoch": 0.8799034399517199, + "grad_norm": 3.8402220541360292, + "learning_rate": 2.1861837130738945e-07, + "loss": 0.4467, + "step": 2187 + }, + { + "epoch": 0.8803057734862201, + "grad_norm": 4.223182944719613, + "learning_rate": 2.1718478857112506e-07, + "loss": 0.5347, + "step": 2188 + }, + { + "epoch": 0.8807081070207202, + "grad_norm": 3.8811613151194213, + "learning_rate": 2.1575570817667952e-07, + "loss": 0.5265, + "step": 2189 + }, + { + "epoch": 0.8811104405552203, + "grad_norm": 3.8705546986459956, + "learning_rate": 2.143311329425926e-07, + "loss": 0.4427, + "step": 2190 + }, + { + "epoch": 0.8815127740897204, + "grad_norm": 4.5414937373063955, + "learning_rate": 2.1291106567852105e-07, + "loss": 0.6442, + "step": 2191 + }, + { + "epoch": 0.8819151076242204, + "grad_norm": 3.588299387450043, + "learning_rate": 2.1149550918522887e-07, + "loss": 0.4173, + "step": 2192 + }, + { + "epoch": 0.8823174411587206, + "grad_norm": 4.165225372551161, + "learning_rate": 2.100844662545848e-07, + "loss": 0.5496, + "step": 2193 + }, + { + "epoch": 0.8827197746932207, + "grad_norm": 4.217297273907552, + "learning_rate": 2.086779396695554e-07, + "loss": 0.5329, + "step": 2194 + }, + { + "epoch": 0.8831221082277207, + "grad_norm": 3.980212414160413, + "learning_rate": 2.0727593220419812e-07, + "loss": 0.4682, + "step": 2195 + }, + { + "epoch": 0.8835244417622209, + "grad_norm": 4.015069610064507, + "learning_rate": 2.0587844662365963e-07, + "loss": 0.4975, + "step": 2196 + }, + { + "epoch": 0.883926775296721, + "grad_norm": 3.675033120744641, + "learning_rate": 2.0448548568416715e-07, + "loss": 0.4932, + "step": 2197 + }, + { + "epoch": 0.8843291088312211, + "grad_norm": 3.7917456823069178, + "learning_rate": 2.030970521330247e-07, + "loss": 0.4523, + "step": 2198 + }, + { + "epoch": 0.8847314423657212, + "grad_norm": 4.2084887531102195, + "learning_rate": 2.0171314870860548e-07, + "loss": 0.5797, + "step": 2199 + }, + { + "epoch": 0.8851337759002212, + "grad_norm": 4.1920678247066245, + "learning_rate": 2.0033377814034999e-07, + "loss": 0.5458, + "step": 2200 + }, + { + "epoch": 0.8855361094347214, + "grad_norm": 4.341033030777012, + "learning_rate": 1.9895894314875707e-07, + "loss": 0.5069, + "step": 2201 + }, + { + "epoch": 0.8859384429692215, + "grad_norm": 4.251489033634488, + "learning_rate": 1.9758864644538188e-07, + "loss": 0.5658, + "step": 2202 + }, + { + "epoch": 0.8863407765037216, + "grad_norm": 4.118913583140747, + "learning_rate": 1.9622289073282675e-07, + "loss": 0.5186, + "step": 2203 + }, + { + "epoch": 0.8867431100382217, + "grad_norm": 4.17198274381597, + "learning_rate": 1.9486167870473915e-07, + "loss": 0.5394, + "step": 2204 + }, + { + "epoch": 0.8871454435727217, + "grad_norm": 4.1140466186765865, + "learning_rate": 1.9350501304580577e-07, + "loss": 0.6302, + "step": 2205 + }, + { + "epoch": 0.8875477771072219, + "grad_norm": 4.252952260852059, + "learning_rate": 1.9215289643174485e-07, + "loss": 0.5715, + "step": 2206 + }, + { + "epoch": 0.887950110641722, + "grad_norm": 3.9053915680472673, + "learning_rate": 1.9080533152930486e-07, + "loss": 0.502, + "step": 2207 + }, + { + "epoch": 0.8883524441762221, + "grad_norm": 3.702225135993101, + "learning_rate": 1.8946232099625482e-07, + "loss": 0.3917, + "step": 2208 + }, + { + "epoch": 0.8887547777107222, + "grad_norm": 3.5875962058050863, + "learning_rate": 1.8812386748138234e-07, + "loss": 0.3413, + "step": 2209 + }, + { + "epoch": 0.8891571112452223, + "grad_norm": 3.6894845944247723, + "learning_rate": 1.8678997362448743e-07, + "loss": 0.4334, + "step": 2210 + }, + { + "epoch": 0.8895594447797224, + "grad_norm": 3.977873097022255, + "learning_rate": 1.8546064205637758e-07, + "loss": 0.4746, + "step": 2211 + }, + { + "epoch": 0.8899617783142225, + "grad_norm": 3.4865669801340147, + "learning_rate": 1.8413587539886064e-07, + "loss": 0.4173, + "step": 2212 + }, + { + "epoch": 0.8903641118487226, + "grad_norm": 4.35862373407535, + "learning_rate": 1.828156762647429e-07, + "loss": 0.5653, + "step": 2213 + }, + { + "epoch": 0.8907664453832227, + "grad_norm": 4.2652495550914775, + "learning_rate": 1.81500047257821e-07, + "loss": 0.5071, + "step": 2214 + }, + { + "epoch": 0.8911687789177228, + "grad_norm": 3.285190042692873, + "learning_rate": 1.8018899097287906e-07, + "loss": 0.3731, + "step": 2215 + }, + { + "epoch": 0.8915711124522229, + "grad_norm": 3.782205241606253, + "learning_rate": 1.78882509995682e-07, + "loss": 0.4907, + "step": 2216 + }, + { + "epoch": 0.891973445986723, + "grad_norm": 4.391188417961942, + "learning_rate": 1.7758060690297035e-07, + "loss": 0.3594, + "step": 2217 + }, + { + "epoch": 0.892375779521223, + "grad_norm": 3.172977998916911, + "learning_rate": 1.7628328426245677e-07, + "loss": 0.3866, + "step": 2218 + }, + { + "epoch": 0.8927781130557232, + "grad_norm": 4.109857114119739, + "learning_rate": 1.7499054463281985e-07, + "loss": 0.5489, + "step": 2219 + }, + { + "epoch": 0.8931804465902233, + "grad_norm": 3.6387926247562214, + "learning_rate": 1.7370239056369946e-07, + "loss": 0.3892, + "step": 2220 + }, + { + "epoch": 0.8935827801247234, + "grad_norm": 3.897940749417438, + "learning_rate": 1.7241882459569008e-07, + "loss": 0.6287, + "step": 2221 + }, + { + "epoch": 0.8939851136592235, + "grad_norm": 3.968476012728493, + "learning_rate": 1.7113984926033832e-07, + "loss": 0.459, + "step": 2222 + }, + { + "epoch": 0.8943874471937237, + "grad_norm": 4.209938898766111, + "learning_rate": 1.6986546708013722e-07, + "loss": 0.5426, + "step": 2223 + }, + { + "epoch": 0.8947897807282237, + "grad_norm": 3.791580194250541, + "learning_rate": 1.6859568056852016e-07, + "loss": 0.498, + "step": 2224 + }, + { + "epoch": 0.8951921142627238, + "grad_norm": 3.8639160637014665, + "learning_rate": 1.673304922298563e-07, + "loss": 0.4049, + "step": 2225 + }, + { + "epoch": 0.8955944477972239, + "grad_norm": 4.147228062957371, + "learning_rate": 1.6606990455944634e-07, + "loss": 0.5258, + "step": 2226 + }, + { + "epoch": 0.895996781331724, + "grad_norm": 4.247308433616599, + "learning_rate": 1.648139200435178e-07, + "loss": 0.5252, + "step": 2227 + }, + { + "epoch": 0.8963991148662241, + "grad_norm": 4.057558052673904, + "learning_rate": 1.635625411592179e-07, + "loss": 0.4851, + "step": 2228 + }, + { + "epoch": 0.8968014484007242, + "grad_norm": 4.529371570037193, + "learning_rate": 1.6231577037461228e-07, + "loss": 0.6091, + "step": 2229 + }, + { + "epoch": 0.8972037819352243, + "grad_norm": 4.092889891063284, + "learning_rate": 1.6107361014867622e-07, + "loss": 0.5051, + "step": 2230 + }, + { + "epoch": 0.8976061154697244, + "grad_norm": 4.057510211081129, + "learning_rate": 1.598360629312931e-07, + "loss": 0.5379, + "step": 2231 + }, + { + "epoch": 0.8980084490042245, + "grad_norm": 4.08275289824886, + "learning_rate": 1.586031311632477e-07, + "loss": 0.5542, + "step": 2232 + }, + { + "epoch": 0.8984107825387246, + "grad_norm": 5.7116466876850875, + "learning_rate": 1.5737481727622212e-07, + "loss": 0.5021, + "step": 2233 + }, + { + "epoch": 0.8988131160732247, + "grad_norm": 3.507507523132951, + "learning_rate": 1.5615112369278957e-07, + "loss": 0.391, + "step": 2234 + }, + { + "epoch": 0.8992154496077248, + "grad_norm": 3.7113726990024194, + "learning_rate": 1.5493205282641228e-07, + "loss": 0.4515, + "step": 2235 + }, + { + "epoch": 0.899617783142225, + "grad_norm": 3.512898207115907, + "learning_rate": 1.5371760708143447e-07, + "loss": 0.4935, + "step": 2236 + }, + { + "epoch": 0.900020116676725, + "grad_norm": 4.213491522925021, + "learning_rate": 1.525077888530782e-07, + "loss": 0.538, + "step": 2237 + }, + { + "epoch": 0.9004224502112251, + "grad_norm": 4.482285849278304, + "learning_rate": 1.5130260052743955e-07, + "loss": 0.6545, + "step": 2238 + }, + { + "epoch": 0.9008247837457252, + "grad_norm": 3.7492718480617384, + "learning_rate": 1.5010204448148214e-07, + "loss": 0.4442, + "step": 2239 + }, + { + "epoch": 0.9012271172802253, + "grad_norm": 3.9838020107249195, + "learning_rate": 1.489061230830338e-07, + "loss": 0.4953, + "step": 2240 + }, + { + "epoch": 0.9016294508147255, + "grad_norm": 3.9003308620272796, + "learning_rate": 1.4771483869078228e-07, + "loss": 0.5208, + "step": 2241 + }, + { + "epoch": 0.9020317843492255, + "grad_norm": 3.788631062888711, + "learning_rate": 1.4652819365426867e-07, + "loss": 0.5613, + "step": 2242 + }, + { + "epoch": 0.9024341178837256, + "grad_norm": 4.676276663949487, + "learning_rate": 1.453461903138853e-07, + "loss": 0.5352, + "step": 2243 + }, + { + "epoch": 0.9028364514182257, + "grad_norm": 4.525346271987768, + "learning_rate": 1.441688310008682e-07, + "loss": 0.5099, + "step": 2244 + }, + { + "epoch": 0.9032387849527258, + "grad_norm": 3.625199458865597, + "learning_rate": 1.4299611803729547e-07, + "loss": 0.4766, + "step": 2245 + }, + { + "epoch": 0.903641118487226, + "grad_norm": 3.8753089109297787, + "learning_rate": 1.4182805373608055e-07, + "loss": 0.5326, + "step": 2246 + }, + { + "epoch": 0.904043452021726, + "grad_norm": 4.021228834417677, + "learning_rate": 1.4066464040096956e-07, + "loss": 0.5691, + "step": 2247 + }, + { + "epoch": 0.9044457855562261, + "grad_norm": 4.6752863808538105, + "learning_rate": 1.395058803265334e-07, + "loss": 0.7598, + "step": 2248 + }, + { + "epoch": 0.9048481190907263, + "grad_norm": 3.7179471187703514, + "learning_rate": 1.3835177579816783e-07, + "loss": 0.3955, + "step": 2249 + }, + { + "epoch": 0.9052504526252263, + "grad_norm": 5.00701237752423, + "learning_rate": 1.3720232909208541e-07, + "loss": 0.6319, + "step": 2250 + }, + { + "epoch": 0.9056527861597264, + "grad_norm": 3.7834417036951895, + "learning_rate": 1.3605754247531245e-07, + "loss": 0.5174, + "step": 2251 + }, + { + "epoch": 0.9060551196942265, + "grad_norm": 3.680557056193272, + "learning_rate": 1.3491741820568366e-07, + "loss": 0.4133, + "step": 2252 + }, + { + "epoch": 0.9064574532287266, + "grad_norm": 4.046295696144512, + "learning_rate": 1.337819585318395e-07, + "loss": 0.6443, + "step": 2253 + }, + { + "epoch": 0.9068597867632268, + "grad_norm": 4.363328152667618, + "learning_rate": 1.3265116569322e-07, + "loss": 0.5437, + "step": 2254 + }, + { + "epoch": 0.9072621202977268, + "grad_norm": 4.437499912355556, + "learning_rate": 1.3152504192006005e-07, + "loss": 0.5439, + "step": 2255 + }, + { + "epoch": 0.9076644538322269, + "grad_norm": 3.904533797833089, + "learning_rate": 1.3040358943338742e-07, + "loss": 0.4058, + "step": 2256 + }, + { + "epoch": 0.908066787366727, + "grad_norm": 3.990363526582056, + "learning_rate": 1.2928681044501513e-07, + "loss": 0.4204, + "step": 2257 + }, + { + "epoch": 0.9084691209012271, + "grad_norm": 4.0760801231394606, + "learning_rate": 1.281747071575404e-07, + "loss": 0.6104, + "step": 2258 + }, + { + "epoch": 0.9088714544357273, + "grad_norm": 4.256580027345346, + "learning_rate": 1.270672817643376e-07, + "loss": 0.6544, + "step": 2259 + }, + { + "epoch": 0.9092737879702273, + "grad_norm": 4.71579332313011, + "learning_rate": 1.2596453644955597e-07, + "loss": 0.6273, + "step": 2260 + }, + { + "epoch": 0.9096761215047274, + "grad_norm": 3.483958223270458, + "learning_rate": 1.2486647338811248e-07, + "loss": 0.4546, + "step": 2261 + }, + { + "epoch": 0.9100784550392276, + "grad_norm": 3.9093619147070946, + "learning_rate": 1.2377309474569165e-07, + "loss": 0.4985, + "step": 2262 + }, + { + "epoch": 0.9104807885737276, + "grad_norm": 4.057320495855987, + "learning_rate": 1.226844026787377e-07, + "loss": 0.5448, + "step": 2263 + }, + { + "epoch": 0.9108831221082277, + "grad_norm": 3.5360903799800503, + "learning_rate": 1.2160039933445222e-07, + "loss": 0.4164, + "step": 2264 + }, + { + "epoch": 0.9112854556427278, + "grad_norm": 4.087420290794404, + "learning_rate": 1.2052108685078834e-07, + "loss": 0.594, + "step": 2265 + }, + { + "epoch": 0.9116877891772279, + "grad_norm": 3.9832710853164763, + "learning_rate": 1.1944646735644882e-07, + "loss": 0.452, + "step": 2266 + }, + { + "epoch": 0.9120901227117281, + "grad_norm": 4.282566514995135, + "learning_rate": 1.1837654297087997e-07, + "loss": 0.5473, + "step": 2267 + }, + { + "epoch": 0.9124924562462281, + "grad_norm": 3.9435273915235656, + "learning_rate": 1.1731131580426796e-07, + "loss": 0.6224, + "step": 2268 + }, + { + "epoch": 0.9128947897807282, + "grad_norm": 4.3103292605724155, + "learning_rate": 1.1625078795753476e-07, + "loss": 0.5467, + "step": 2269 + }, + { + "epoch": 0.9132971233152283, + "grad_norm": 4.22573396631058, + "learning_rate": 1.1519496152233439e-07, + "loss": 0.5037, + "step": 2270 + }, + { + "epoch": 0.9136994568497284, + "grad_norm": 3.7886525775152338, + "learning_rate": 1.1414383858104756e-07, + "loss": 0.4498, + "step": 2271 + }, + { + "epoch": 0.9141017903842286, + "grad_norm": 3.2224197154073617, + "learning_rate": 1.1309742120677897e-07, + "loss": 0.4794, + "step": 2272 + }, + { + "epoch": 0.9145041239187286, + "grad_norm": 4.705013349470272, + "learning_rate": 1.1205571146335303e-07, + "loss": 0.6041, + "step": 2273 + }, + { + "epoch": 0.9149064574532287, + "grad_norm": 4.6403427413148535, + "learning_rate": 1.1101871140530851e-07, + "loss": 0.6308, + "step": 2274 + }, + { + "epoch": 0.9153087909877289, + "grad_norm": 3.819381857552772, + "learning_rate": 1.0998642307789576e-07, + "loss": 0.5461, + "step": 2275 + }, + { + "epoch": 0.9157111245222289, + "grad_norm": 3.997200472540085, + "learning_rate": 1.0895884851707255e-07, + "loss": 0.4926, + "step": 2276 + }, + { + "epoch": 0.9161134580567291, + "grad_norm": 4.162425768576248, + "learning_rate": 1.079359897494997e-07, + "loss": 0.489, + "step": 2277 + }, + { + "epoch": 0.9165157915912291, + "grad_norm": 3.8422971328140454, + "learning_rate": 1.0691784879253736e-07, + "loss": 0.5291, + "step": 2278 + }, + { + "epoch": 0.9169181251257292, + "grad_norm": 4.112928250644688, + "learning_rate": 1.0590442765423953e-07, + "loss": 0.4926, + "step": 2279 + }, + { + "epoch": 0.9173204586602294, + "grad_norm": 3.9378155981778664, + "learning_rate": 1.0489572833335349e-07, + "loss": 0.4722, + "step": 2280 + }, + { + "epoch": 0.9177227921947294, + "grad_norm": 4.631927118939465, + "learning_rate": 1.038917528193123e-07, + "loss": 0.5274, + "step": 2281 + }, + { + "epoch": 0.9181251257292296, + "grad_norm": 4.035377760065258, + "learning_rate": 1.0289250309223314e-07, + "loss": 0.5788, + "step": 2282 + }, + { + "epoch": 0.9185274592637296, + "grad_norm": 3.868825647605808, + "learning_rate": 1.0189798112291177e-07, + "loss": 0.5046, + "step": 2283 + }, + { + "epoch": 0.9189297927982297, + "grad_norm": 4.194215283322997, + "learning_rate": 1.0090818887282112e-07, + "loss": 0.6029, + "step": 2284 + }, + { + "epoch": 0.9193321263327299, + "grad_norm": 4.215786203028234, + "learning_rate": 9.99231282941035e-08, + "loss": 0.4987, + "step": 2285 + }, + { + "epoch": 0.9197344598672299, + "grad_norm": 4.911243421308879, + "learning_rate": 9.894280132957124e-08, + "loss": 0.6475, + "step": 2286 + }, + { + "epoch": 0.92013679340173, + "grad_norm": 4.551350447477857, + "learning_rate": 9.79672099126991e-08, + "loss": 0.6439, + "step": 2287 + }, + { + "epoch": 0.9205391269362302, + "grad_norm": 4.0194710398308615, + "learning_rate": 9.69963559676232e-08, + "loss": 0.5376, + "step": 2288 + }, + { + "epoch": 0.9209414604707302, + "grad_norm": 3.83021269408099, + "learning_rate": 9.60302414091352e-08, + "loss": 0.5093, + "step": 2289 + }, + { + "epoch": 0.9213437940052304, + "grad_norm": 4.1940419822947534, + "learning_rate": 9.506886814268007e-08, + "loss": 0.5702, + "step": 2290 + }, + { + "epoch": 0.9217461275397304, + "grad_norm": 4.341103613511488, + "learning_rate": 9.41122380643511e-08, + "loss": 0.5271, + "step": 2291 + }, + { + "epoch": 0.9221484610742305, + "grad_norm": 4.421018842084841, + "learning_rate": 9.316035306088684e-08, + "loss": 0.5923, + "step": 2292 + }, + { + "epoch": 0.9225507946087307, + "grad_norm": 4.192150302264493, + "learning_rate": 9.221321500966746e-08, + "loss": 0.5475, + "step": 2293 + }, + { + "epoch": 0.9229531281432307, + "grad_norm": 4.56023548088152, + "learning_rate": 9.127082577871038e-08, + "loss": 0.6055, + "step": 2294 + }, + { + "epoch": 0.9233554616777309, + "grad_norm": 3.727375694339219, + "learning_rate": 9.033318722666801e-08, + "loss": 0.3706, + "step": 2295 + }, + { + "epoch": 0.9237577952122309, + "grad_norm": 3.7651015398148897, + "learning_rate": 8.940030120282162e-08, + "loss": 0.5118, + "step": 2296 + }, + { + "epoch": 0.924160128746731, + "grad_norm": 4.305229814421829, + "learning_rate": 8.847216954708055e-08, + "loss": 0.5453, + "step": 2297 + }, + { + "epoch": 0.9245624622812312, + "grad_norm": 4.1747564223004066, + "learning_rate": 8.754879408997636e-08, + "loss": 0.5156, + "step": 2298 + }, + { + "epoch": 0.9249647958157312, + "grad_norm": 3.649399704340268, + "learning_rate": 8.663017665266032e-08, + "loss": 0.3701, + "step": 2299 + }, + { + "epoch": 0.9253671293502314, + "grad_norm": 3.7596577501273734, + "learning_rate": 8.571631904690014e-08, + "loss": 0.4399, + "step": 2300 + }, + { + "epoch": 0.9257694628847315, + "grad_norm": 4.230082435346815, + "learning_rate": 8.480722307507461e-08, + "loss": 0.4874, + "step": 2301 + }, + { + "epoch": 0.9261717964192315, + "grad_norm": 3.9612627542935197, + "learning_rate": 8.390289053017226e-08, + "loss": 0.5485, + "step": 2302 + }, + { + "epoch": 0.9265741299537317, + "grad_norm": 3.5438909670467003, + "learning_rate": 8.300332319578664e-08, + "loss": 0.3951, + "step": 2303 + }, + { + "epoch": 0.9269764634882317, + "grad_norm": 3.2833308636054217, + "learning_rate": 8.210852284611299e-08, + "loss": 0.407, + "step": 2304 + }, + { + "epoch": 0.9273787970227318, + "grad_norm": 3.4444311785022865, + "learning_rate": 8.121849124594405e-08, + "loss": 0.3925, + "step": 2305 + }, + { + "epoch": 0.927781130557232, + "grad_norm": 4.548198093339736, + "learning_rate": 8.033323015066818e-08, + "loss": 0.719, + "step": 2306 + }, + { + "epoch": 0.928183464091732, + "grad_norm": 3.6123982321182284, + "learning_rate": 7.945274130626484e-08, + "loss": 0.3828, + "step": 2307 + }, + { + "epoch": 0.9285857976262322, + "grad_norm": 3.177723392355348, + "learning_rate": 7.857702644930133e-08, + "loss": 0.353, + "step": 2308 + }, + { + "epoch": 0.9289881311607322, + "grad_norm": 4.133100828896777, + "learning_rate": 7.770608730692852e-08, + "loss": 0.4977, + "step": 2309 + }, + { + "epoch": 0.9293904646952323, + "grad_norm": 3.562342202197354, + "learning_rate": 7.683992559687881e-08, + "loss": 0.3892, + "step": 2310 + }, + { + "epoch": 0.9297927982297325, + "grad_norm": 3.898366854418511, + "learning_rate": 7.597854302746288e-08, + "loss": 0.4583, + "step": 2311 + }, + { + "epoch": 0.9301951317642325, + "grad_norm": 4.612414420760638, + "learning_rate": 7.512194129756428e-08, + "loss": 0.5581, + "step": 2312 + }, + { + "epoch": 0.9305974652987327, + "grad_norm": 3.988029769560431, + "learning_rate": 7.42701220966388e-08, + "loss": 0.4744, + "step": 2313 + }, + { + "epoch": 0.9309997988332327, + "grad_norm": 4.481476524243926, + "learning_rate": 7.342308710470814e-08, + "loss": 0.59, + "step": 2314 + }, + { + "epoch": 0.9314021323677328, + "grad_norm": 4.562506876878105, + "learning_rate": 7.258083799235982e-08, + "loss": 0.6742, + "step": 2315 + }, + { + "epoch": 0.931804465902233, + "grad_norm": 4.628598132831451, + "learning_rate": 7.174337642074147e-08, + "loss": 0.6008, + "step": 2316 + }, + { + "epoch": 0.932206799436733, + "grad_norm": 4.283379928862786, + "learning_rate": 7.091070404155854e-08, + "loss": 0.5053, + "step": 2317 + }, + { + "epoch": 0.9326091329712332, + "grad_norm": 3.9776126177559155, + "learning_rate": 7.008282249707065e-08, + "loss": 0.4839, + "step": 2318 + }, + { + "epoch": 0.9330114665057333, + "grad_norm": 4.069433607383545, + "learning_rate": 6.925973342008897e-08, + "loss": 0.453, + "step": 2319 + }, + { + "epoch": 0.9334138000402333, + "grad_norm": 4.037972743715134, + "learning_rate": 6.844143843397238e-08, + "loss": 0.5427, + "step": 2320 + }, + { + "epoch": 0.9338161335747335, + "grad_norm": 3.547048708842676, + "learning_rate": 6.76279391526241e-08, + "loss": 0.4561, + "step": 2321 + }, + { + "epoch": 0.9342184671092335, + "grad_norm": 3.7831575419487256, + "learning_rate": 6.681923718049016e-08, + "loss": 0.5184, + "step": 2322 + }, + { + "epoch": 0.9346208006437337, + "grad_norm": 3.7960605831677428, + "learning_rate": 6.601533411255328e-08, + "loss": 0.428, + "step": 2323 + }, + { + "epoch": 0.9350231341782338, + "grad_norm": 3.699019611927255, + "learning_rate": 6.52162315343327e-08, + "loss": 0.4711, + "step": 2324 + }, + { + "epoch": 0.9354254677127338, + "grad_norm": 3.7499929742301465, + "learning_rate": 6.44219310218791e-08, + "loss": 0.494, + "step": 2325 + }, + { + "epoch": 0.935827801247234, + "grad_norm": 3.982088440613016, + "learning_rate": 6.363243414177239e-08, + "loss": 0.5013, + "step": 2326 + }, + { + "epoch": 0.936230134781734, + "grad_norm": 3.5405567172339034, + "learning_rate": 6.284774245111868e-08, + "loss": 0.4151, + "step": 2327 + }, + { + "epoch": 0.9366324683162341, + "grad_norm": 3.9069783277866894, + "learning_rate": 6.206785749754613e-08, + "loss": 0.5257, + "step": 2328 + }, + { + "epoch": 0.9370348018507343, + "grad_norm": 4.560908800971963, + "learning_rate": 6.12927808192032e-08, + "loss": 0.5827, + "step": 2329 + }, + { + "epoch": 0.9374371353852343, + "grad_norm": 3.6067893534884967, + "learning_rate": 6.052251394475544e-08, + "loss": 0.4533, + "step": 2330 + }, + { + "epoch": 0.9378394689197345, + "grad_norm": 4.228506402107548, + "learning_rate": 5.975705839338208e-08, + "loss": 0.5367, + "step": 2331 + }, + { + "epoch": 0.9382418024542346, + "grad_norm": 4.377928096595712, + "learning_rate": 5.89964156747716e-08, + "loss": 0.487, + "step": 2332 + }, + { + "epoch": 0.9386441359887346, + "grad_norm": 4.001917632135577, + "learning_rate": 5.8240587289122576e-08, + "loss": 0.5991, + "step": 2333 + }, + { + "epoch": 0.9390464695232348, + "grad_norm": 3.89102054624204, + "learning_rate": 5.7489574727136734e-08, + "loss": 0.5289, + "step": 2334 + }, + { + "epoch": 0.9394488030577348, + "grad_norm": 3.9529925607216314, + "learning_rate": 5.67433794700184e-08, + "loss": 0.5516, + "step": 2335 + }, + { + "epoch": 0.939851136592235, + "grad_norm": 3.9965015035833114, + "learning_rate": 5.600200298947034e-08, + "loss": 0.5758, + "step": 2336 + }, + { + "epoch": 0.9402534701267351, + "grad_norm": 4.354462501990094, + "learning_rate": 5.52654467476918e-08, + "loss": 0.5793, + "step": 2337 + }, + { + "epoch": 0.9406558036612351, + "grad_norm": 4.199766571714081, + "learning_rate": 5.453371219737491e-08, + "loss": 0.5688, + "step": 2338 + }, + { + "epoch": 0.9410581371957353, + "grad_norm": 3.3551684650390454, + "learning_rate": 5.380680078170275e-08, + "loss": 0.3524, + "step": 2339 + }, + { + "epoch": 0.9414604707302353, + "grad_norm": 4.408645627917106, + "learning_rate": 5.3084713934344613e-08, + "loss": 0.5442, + "step": 2340 + }, + { + "epoch": 0.9418628042647355, + "grad_norm": 4.4356672674608895, + "learning_rate": 5.236745307945545e-08, + "loss": 0.5028, + "step": 2341 + }, + { + "epoch": 0.9422651377992356, + "grad_norm": 3.46236072019421, + "learning_rate": 5.165501963167174e-08, + "loss": 0.3921, + "step": 2342 + }, + { + "epoch": 0.9426674713337356, + "grad_norm": 4.4269078668736075, + "learning_rate": 5.094741499610839e-08, + "loss": 0.5983, + "step": 2343 + }, + { + "epoch": 0.9430698048682358, + "grad_norm": 4.503787561410546, + "learning_rate": 5.02446405683582e-08, + "loss": 0.585, + "step": 2344 + }, + { + "epoch": 0.9434721384027359, + "grad_norm": 3.658547311389223, + "learning_rate": 4.9546697734485224e-08, + "loss": 0.4762, + "step": 2345 + }, + { + "epoch": 0.943874471937236, + "grad_norm": 4.353601396348682, + "learning_rate": 4.885358787102584e-08, + "loss": 0.6102, + "step": 2346 + }, + { + "epoch": 0.9442768054717361, + "grad_norm": 4.143732790296102, + "learning_rate": 4.816531234498406e-08, + "loss": 0.6124, + "step": 2347 + }, + { + "epoch": 0.9446791390062361, + "grad_norm": 4.3975559021632575, + "learning_rate": 4.7481872513829575e-08, + "loss": 0.5633, + "step": 2348 + }, + { + "epoch": 0.9450814725407363, + "grad_norm": 4.066972538920918, + "learning_rate": 4.680326972549387e-08, + "loss": 0.5118, + "step": 2349 + }, + { + "epoch": 0.9454838060752364, + "grad_norm": 4.03878765773838, + "learning_rate": 4.6129505318369127e-08, + "loss": 0.4778, + "step": 2350 + }, + { + "epoch": 0.9458861396097364, + "grad_norm": 4.201321170733065, + "learning_rate": 4.546058062130487e-08, + "loss": 0.5128, + "step": 2351 + }, + { + "epoch": 0.9462884731442366, + "grad_norm": 3.8094183727880395, + "learning_rate": 4.4796496953605493e-08, + "loss": 0.5441, + "step": 2352 + }, + { + "epoch": 0.9466908066787366, + "grad_norm": 3.5299699606250834, + "learning_rate": 4.413725562502691e-08, + "loss": 0.4678, + "step": 2353 + }, + { + "epoch": 0.9470931402132368, + "grad_norm": 3.9087226991955717, + "learning_rate": 4.348285793577517e-08, + "loss": 0.5329, + "step": 2354 + }, + { + "epoch": 0.9474954737477369, + "grad_norm": 4.0121109992844755, + "learning_rate": 4.2833305176503136e-08, + "loss": 0.5837, + "step": 2355 + }, + { + "epoch": 0.9478978072822369, + "grad_norm": 4.544145628732853, + "learning_rate": 4.218859862830826e-08, + "loss": 0.3829, + "step": 2356 + }, + { + "epoch": 0.9483001408167371, + "grad_norm": 3.89076415048523, + "learning_rate": 4.154873956272981e-08, + "loss": 0.4691, + "step": 2357 + }, + { + "epoch": 0.9487024743512372, + "grad_norm": 4.1725175451999, + "learning_rate": 4.091372924174636e-08, + "loss": 0.4885, + "step": 2358 + }, + { + "epoch": 0.9491048078857373, + "grad_norm": 4.3450167781713285, + "learning_rate": 4.028356891777385e-08, + "loss": 0.5124, + "step": 2359 + }, + { + "epoch": 0.9495071414202374, + "grad_norm": 4.4322372412295605, + "learning_rate": 3.9658259833662284e-08, + "loss": 0.5234, + "step": 2360 + }, + { + "epoch": 0.9499094749547374, + "grad_norm": 3.707524532117256, + "learning_rate": 3.903780322269374e-08, + "loss": 0.4149, + "step": 2361 + }, + { + "epoch": 0.9503118084892376, + "grad_norm": 5.065679790929374, + "learning_rate": 3.842220030858018e-08, + "loss": 0.688, + "step": 2362 + }, + { + "epoch": 0.9507141420237377, + "grad_norm": 3.786912616101777, + "learning_rate": 3.781145230546013e-08, + "loss": 0.43, + "step": 2363 + }, + { + "epoch": 0.9511164755582378, + "grad_norm": 3.6227665097152224, + "learning_rate": 3.720556041789752e-08, + "loss": 0.4382, + "step": 2364 + }, + { + "epoch": 0.9515188090927379, + "grad_norm": 3.999125768684484, + "learning_rate": 3.660452584087815e-08, + "loss": 0.5716, + "step": 2365 + }, + { + "epoch": 0.9519211426272379, + "grad_norm": 3.409457081881056, + "learning_rate": 3.600834975980877e-08, + "loss": 0.5407, + "step": 2366 + }, + { + "epoch": 0.9523234761617381, + "grad_norm": 3.892658686650397, + "learning_rate": 3.5417033350512444e-08, + "loss": 0.4589, + "step": 2367 + }, + { + "epoch": 0.9527258096962382, + "grad_norm": 4.000437062384583, + "learning_rate": 3.483057777922905e-08, + "loss": 0.4875, + "step": 2368 + }, + { + "epoch": 0.9531281432307382, + "grad_norm": 4.046666707315621, + "learning_rate": 3.42489842026103e-08, + "loss": 0.5162, + "step": 2369 + }, + { + "epoch": 0.9535304767652384, + "grad_norm": 4.290796216817974, + "learning_rate": 3.3672253767719485e-08, + "loss": 0.6233, + "step": 2370 + }, + { + "epoch": 0.9539328102997385, + "grad_norm": 4.023014608954607, + "learning_rate": 3.310038761202839e-08, + "loss": 0.5062, + "step": 2371 + }, + { + "epoch": 0.9543351438342386, + "grad_norm": 4.43210170452369, + "learning_rate": 3.2533386863414784e-08, + "loss": 0.5594, + "step": 2372 + }, + { + "epoch": 0.9547374773687387, + "grad_norm": 3.8979482975499624, + "learning_rate": 3.1971252640160545e-08, + "loss": 0.5128, + "step": 2373 + }, + { + "epoch": 0.9551398109032387, + "grad_norm": 4.162386221402383, + "learning_rate": 3.141398605094992e-08, + "loss": 0.5848, + "step": 2374 + }, + { + "epoch": 0.9555421444377389, + "grad_norm": 4.022603602017269, + "learning_rate": 3.086158819486651e-08, + "loss": 0.5771, + "step": 2375 + }, + { + "epoch": 0.955944477972239, + "grad_norm": 3.8920410241916863, + "learning_rate": 3.031406016139077e-08, + "loss": 0.5498, + "step": 2376 + }, + { + "epoch": 0.9563468115067391, + "grad_norm": 4.084788242660989, + "learning_rate": 2.9771403030399713e-08, + "loss": 0.598, + "step": 2377 + }, + { + "epoch": 0.9567491450412392, + "grad_norm": 4.048190861072185, + "learning_rate": 2.923361787216278e-08, + "loss": 0.4405, + "step": 2378 + }, + { + "epoch": 0.9571514785757392, + "grad_norm": 4.502700834332798, + "learning_rate": 2.8700705747340963e-08, + "loss": 0.5169, + "step": 2379 + }, + { + "epoch": 0.9575538121102394, + "grad_norm": 4.054509620797073, + "learning_rate": 2.8172667706983792e-08, + "loss": 0.53, + "step": 2380 + }, + { + "epoch": 0.9579561456447395, + "grad_norm": 3.939484097293469, + "learning_rate": 2.7649504792528204e-08, + "loss": 0.5566, + "step": 2381 + }, + { + "epoch": 0.9583584791792396, + "grad_norm": 4.299563624727568, + "learning_rate": 2.7131218035796324e-08, + "loss": 0.5603, + "step": 2382 + }, + { + "epoch": 0.9587608127137397, + "grad_norm": 3.723520114625019, + "learning_rate": 2.6617808458991867e-08, + "loss": 0.4346, + "step": 2383 + }, + { + "epoch": 0.9591631462482398, + "grad_norm": 3.667012045051573, + "learning_rate": 2.6109277074700955e-08, + "loss": 0.3638, + "step": 2384 + }, + { + "epoch": 0.9595654797827399, + "grad_norm": 3.725825844080702, + "learning_rate": 2.5605624885887414e-08, + "loss": 0.4541, + "step": 2385 + }, + { + "epoch": 0.95996781331724, + "grad_norm": 3.7094680890029776, + "learning_rate": 2.5106852885892764e-08, + "loss": 0.4693, + "step": 2386 + }, + { + "epoch": 0.96037014685174, + "grad_norm": 3.7229547246513937, + "learning_rate": 2.461296205843261e-08, + "loss": 0.5144, + "step": 2387 + }, + { + "epoch": 0.9607724803862402, + "grad_norm": 4.159612295768075, + "learning_rate": 2.412395337759693e-08, + "loss": 0.522, + "step": 2388 + }, + { + "epoch": 0.9611748139207403, + "grad_norm": 3.8986140359372286, + "learning_rate": 2.3639827807844796e-08, + "loss": 0.5237, + "step": 2389 + }, + { + "epoch": 0.9615771474552404, + "grad_norm": 4.121244015419092, + "learning_rate": 2.3160586304006026e-08, + "loss": 0.538, + "step": 2390 + }, + { + "epoch": 0.9619794809897405, + "grad_norm": 4.166559231330322, + "learning_rate": 2.268622981127705e-08, + "loss": 0.5822, + "step": 2391 + }, + { + "epoch": 0.9623818145242405, + "grad_norm": 3.9211071919172813, + "learning_rate": 2.2216759265220044e-08, + "loss": 0.4897, + "step": 2392 + }, + { + "epoch": 0.9627841480587407, + "grad_norm": 4.349097825226008, + "learning_rate": 2.1752175591760184e-08, + "loss": 0.5211, + "step": 2393 + }, + { + "epoch": 0.9631864815932408, + "grad_norm": 3.9282047251137677, + "learning_rate": 2.1292479707184787e-08, + "loss": 0.5741, + "step": 2394 + }, + { + "epoch": 0.9635888151277409, + "grad_norm": 4.159989263744697, + "learning_rate": 2.0837672518141395e-08, + "loss": 0.5238, + "step": 2395 + }, + { + "epoch": 0.963991148662241, + "grad_norm": 3.8198728228381293, + "learning_rate": 2.0387754921634972e-08, + "loss": 0.5144, + "step": 2396 + }, + { + "epoch": 0.9643934821967411, + "grad_norm": 3.346255200741844, + "learning_rate": 1.9942727805027373e-08, + "loss": 0.4384, + "step": 2397 + }, + { + "epoch": 0.9647958157312412, + "grad_norm": 3.8469019196516845, + "learning_rate": 1.9502592046035107e-08, + "loss": 0.486, + "step": 2398 + }, + { + "epoch": 0.9651981492657413, + "grad_norm": 3.5776773638982893, + "learning_rate": 1.9067348512726847e-08, + "loss": 0.4414, + "step": 2399 + }, + { + "epoch": 0.9656004828002414, + "grad_norm": 4.031486537261451, + "learning_rate": 1.8636998063523705e-08, + "loss": 0.5328, + "step": 2400 + }, + { + "epoch": 0.9660028163347415, + "grad_norm": 4.260664576753229, + "learning_rate": 1.8211541547195067e-08, + "loss": 0.5136, + "step": 2401 + }, + { + "epoch": 0.9664051498692416, + "grad_norm": 4.152186034058617, + "learning_rate": 1.7790979802858876e-08, + "loss": 0.6316, + "step": 2402 + }, + { + "epoch": 0.9668074834037417, + "grad_norm": 3.995357070807909, + "learning_rate": 1.7375313659978855e-08, + "loss": 0.5377, + "step": 2403 + }, + { + "epoch": 0.9672098169382418, + "grad_norm": 3.4651763373212474, + "learning_rate": 1.6964543938363388e-08, + "loss": 0.4157, + "step": 2404 + }, + { + "epoch": 0.9676121504727418, + "grad_norm": 4.360198459376172, + "learning_rate": 1.6558671448163866e-08, + "loss": 0.6273, + "step": 2405 + }, + { + "epoch": 0.968014484007242, + "grad_norm": 4.006498747343065, + "learning_rate": 1.6157696989873017e-08, + "loss": 0.4914, + "step": 2406 + }, + { + "epoch": 0.9684168175417421, + "grad_norm": 3.7398733702546822, + "learning_rate": 1.5761621354322965e-08, + "loss": 0.4022, + "step": 2407 + }, + { + "epoch": 0.9688191510762422, + "grad_norm": 4.1110217996882294, + "learning_rate": 1.5370445322684392e-08, + "loss": 0.4698, + "step": 2408 + }, + { + "epoch": 0.9692214846107423, + "grad_norm": 4.362254470529691, + "learning_rate": 1.498416966646432e-08, + "loss": 0.5598, + "step": 2409 + }, + { + "epoch": 0.9696238181452425, + "grad_norm": 3.8880330136972328, + "learning_rate": 1.4602795147504734e-08, + "loss": 0.473, + "step": 2410 + }, + { + "epoch": 0.9700261516797425, + "grad_norm": 4.110874787245324, + "learning_rate": 1.4226322517981728e-08, + "loss": 0.4901, + "step": 2411 + }, + { + "epoch": 0.9704284852142426, + "grad_norm": 4.138814644855194, + "learning_rate": 1.3854752520403303e-08, + "loss": 0.5618, + "step": 2412 + }, + { + "epoch": 0.9708308187487427, + "grad_norm": 4.481371755440245, + "learning_rate": 1.3488085887607972e-08, + "loss": 0.6152, + "step": 2413 + }, + { + "epoch": 0.9712331522832428, + "grad_norm": 3.6050373310270176, + "learning_rate": 1.3126323342763648e-08, + "loss": 0.5525, + "step": 2414 + }, + { + "epoch": 0.971635485817743, + "grad_norm": 4.061084112826385, + "learning_rate": 1.2769465599365982e-08, + "loss": 0.4619, + "step": 2415 + }, + { + "epoch": 0.972037819352243, + "grad_norm": 3.8920028399978097, + "learning_rate": 1.24175133612367e-08, + "loss": 0.4833, + "step": 2416 + }, + { + "epoch": 0.9724401528867431, + "grad_norm": 3.909035344776143, + "learning_rate": 1.2070467322522762e-08, + "loss": 0.5188, + "step": 2417 + }, + { + "epoch": 0.9728424864212432, + "grad_norm": 4.225663392112197, + "learning_rate": 1.1728328167695258e-08, + "loss": 0.6696, + "step": 2418 + }, + { + "epoch": 0.9732448199557433, + "grad_norm": 4.6741890030042725, + "learning_rate": 1.1391096571546633e-08, + "loss": 0.5815, + "step": 2419 + }, + { + "epoch": 0.9736471534902434, + "grad_norm": 4.51608876386972, + "learning_rate": 1.1058773199190965e-08, + "loss": 0.5173, + "step": 2420 + }, + { + "epoch": 0.9740494870247435, + "grad_norm": 4.0654982873308665, + "learning_rate": 1.0731358706061456e-08, + "loss": 0.5749, + "step": 2421 + }, + { + "epoch": 0.9744518205592436, + "grad_norm": 4.2304161552160835, + "learning_rate": 1.0408853737909896e-08, + "loss": 0.5813, + "step": 2422 + }, + { + "epoch": 0.9748541540937438, + "grad_norm": 3.4742799346124493, + "learning_rate": 1.0091258930805259e-08, + "loss": 0.4705, + "step": 2423 + }, + { + "epoch": 0.9752564876282438, + "grad_norm": 4.08672880848899, + "learning_rate": 9.778574911132044e-09, + "loss": 0.516, + "step": 2424 + }, + { + "epoch": 0.9756588211627439, + "grad_norm": 3.669930620914309, + "learning_rate": 9.470802295589442e-09, + "loss": 0.4138, + "step": 2425 + }, + { + "epoch": 0.976061154697244, + "grad_norm": 3.5860479619434673, + "learning_rate": 9.16794169119023e-09, + "loss": 0.4224, + "step": 2426 + }, + { + "epoch": 0.9764634882317441, + "grad_norm": 4.642054584241187, + "learning_rate": 8.869993695258816e-09, + "loss": 0.611, + "step": 2427 + }, + { + "epoch": 0.9768658217662443, + "grad_norm": 3.977155407637389, + "learning_rate": 8.576958895431531e-09, + "loss": 0.5681, + "step": 2428 + }, + { + "epoch": 0.9772681553007443, + "grad_norm": 3.6884328747774857, + "learning_rate": 8.28883786965301e-09, + "loss": 0.4945, + "step": 2429 + }, + { + "epoch": 0.9776704888352444, + "grad_norm": 4.023502331465119, + "learning_rate": 8.005631186177864e-09, + "loss": 0.4201, + "step": 2430 + }, + { + "epoch": 0.9780728223697445, + "grad_norm": 3.901800973146699, + "learning_rate": 7.727339403568179e-09, + "loss": 0.4977, + "step": 2431 + }, + { + "epoch": 0.9784751559042446, + "grad_norm": 3.9967038572140856, + "learning_rate": 7.45396307069185e-09, + "loss": 0.6418, + "step": 2432 + }, + { + "epoch": 0.9788774894387448, + "grad_norm": 4.291073567899429, + "learning_rate": 7.185502726722305e-09, + "loss": 0.6601, + "step": 2433 + }, + { + "epoch": 0.9792798229732448, + "grad_norm": 3.974019553234897, + "learning_rate": 6.921958901137948e-09, + "loss": 0.4781, + "step": 2434 + }, + { + "epoch": 0.9796821565077449, + "grad_norm": 3.833029352112665, + "learning_rate": 6.66333211371939e-09, + "loss": 0.4373, + "step": 2435 + }, + { + "epoch": 0.9800844900422451, + "grad_norm": 4.398810707509668, + "learning_rate": 6.409622874550547e-09, + "loss": 0.5795, + "step": 2436 + }, + { + "epoch": 0.9804868235767451, + "grad_norm": 3.5697882170733557, + "learning_rate": 6.160831684015878e-09, + "loss": 0.398, + "step": 2437 + }, + { + "epoch": 0.9808891571112452, + "grad_norm": 3.904433602202901, + "learning_rate": 5.9169590328003735e-09, + "loss": 0.5029, + "step": 2438 + }, + { + "epoch": 0.9812914906457453, + "grad_norm": 3.7353800882065715, + "learning_rate": 5.678005401888176e-09, + "loss": 0.4911, + "step": 2439 + }, + { + "epoch": 0.9816938241802454, + "grad_norm": 3.795949925181833, + "learning_rate": 5.443971262561465e-09, + "loss": 0.4405, + "step": 2440 + }, + { + "epoch": 0.9820961577147456, + "grad_norm": 3.617609885039579, + "learning_rate": 5.214857076400736e-09, + "loss": 0.4329, + "step": 2441 + }, + { + "epoch": 0.9824984912492456, + "grad_norm": 3.5249358490195637, + "learning_rate": 4.990663295282305e-09, + "loss": 0.4285, + "step": 2442 + }, + { + "epoch": 0.9829008247837457, + "grad_norm": 4.142775577116706, + "learning_rate": 4.771390361377748e-09, + "loss": 0.4777, + "step": 2443 + }, + { + "epoch": 0.9833031583182458, + "grad_norm": 4.312344250176257, + "learning_rate": 4.5570387071544595e-09, + "loss": 0.6627, + "step": 2444 + }, + { + "epoch": 0.9837054918527459, + "grad_norm": 3.67386973137855, + "learning_rate": 4.347608755372601e-09, + "loss": 0.4431, + "step": 2445 + }, + { + "epoch": 0.9841078253872461, + "grad_norm": 4.191733748165821, + "learning_rate": 4.143100919086207e-09, + "loss": 0.4884, + "step": 2446 + }, + { + "epoch": 0.9845101589217461, + "grad_norm": 4.38016174030762, + "learning_rate": 3.943515601640968e-09, + "loss": 0.5822, + "step": 2447 + }, + { + "epoch": 0.9849124924562462, + "grad_norm": 4.227081529941771, + "learning_rate": 3.748853196673952e-09, + "loss": 0.49, + "step": 2448 + }, + { + "epoch": 0.9853148259907464, + "grad_norm": 3.6169704926991124, + "learning_rate": 3.5591140881138797e-09, + "loss": 0.5642, + "step": 2449 + }, + { + "epoch": 0.9857171595252464, + "grad_norm": 3.228283450712886, + "learning_rate": 3.374298650178076e-09, + "loss": 0.3926, + "step": 2450 + }, + { + "epoch": 0.9861194930597466, + "grad_norm": 3.996805720599477, + "learning_rate": 3.194407247373854e-09, + "loss": 0.5175, + "step": 2451 + }, + { + "epoch": 0.9865218265942466, + "grad_norm": 3.786750548322366, + "learning_rate": 3.0194402344965733e-09, + "loss": 0.4496, + "step": 2452 + }, + { + "epoch": 0.9869241601287467, + "grad_norm": 3.730366038034417, + "learning_rate": 2.8493979566296403e-09, + "loss": 0.4701, + "step": 2453 + }, + { + "epoch": 0.9873264936632469, + "grad_norm": 4.616677722561612, + "learning_rate": 2.6842807491433974e-09, + "loss": 0.6207, + "step": 2454 + }, + { + "epoch": 0.9877288271977469, + "grad_norm": 3.9983804243694645, + "learning_rate": 2.524088937694291e-09, + "loss": 0.4734, + "step": 2455 + }, + { + "epoch": 0.988131160732247, + "grad_norm": 4.364062652259154, + "learning_rate": 2.3688228382251487e-09, + "loss": 0.6254, + "step": 2456 + }, + { + "epoch": 0.9885334942667471, + "grad_norm": 3.97492888774164, + "learning_rate": 2.218482756963236e-09, + "loss": 0.5645, + "step": 2457 + }, + { + "epoch": 0.9889358278012472, + "grad_norm": 3.9671569540217493, + "learning_rate": 2.07306899042109e-09, + "loss": 0.4821, + "step": 2458 + }, + { + "epoch": 0.9893381613357474, + "grad_norm": 3.808293451273474, + "learning_rate": 1.9325818253945748e-09, + "loss": 0.4502, + "step": 2459 + }, + { + "epoch": 0.9897404948702474, + "grad_norm": 4.467378355634954, + "learning_rate": 1.7970215389628842e-09, + "loss": 0.5963, + "step": 2460 + }, + { + "epoch": 0.9901428284047475, + "grad_norm": 4.625199727103312, + "learning_rate": 1.6663883984888163e-09, + "loss": 0.6419, + "step": 2461 + }, + { + "epoch": 0.9905451619392477, + "grad_norm": 3.9326948253065184, + "learning_rate": 1.5406826616168324e-09, + "loss": 0.5575, + "step": 2462 + }, + { + "epoch": 0.9909474954737477, + "grad_norm": 4.4568921518738485, + "learning_rate": 1.4199045762730569e-09, + "loss": 0.6483, + "step": 2463 + }, + { + "epoch": 0.9913498290082479, + "grad_norm": 4.351427492104435, + "learning_rate": 1.3040543806658313e-09, + "loss": 0.5948, + "step": 2464 + }, + { + "epoch": 0.9917521625427479, + "grad_norm": 3.4818450271582857, + "learning_rate": 1.1931323032832177e-09, + "loss": 0.3535, + "step": 2465 + }, + { + "epoch": 0.992154496077248, + "grad_norm": 4.11227816177031, + "learning_rate": 1.0871385628941077e-09, + "loss": 0.5317, + "step": 2466 + }, + { + "epoch": 0.9925568296117482, + "grad_norm": 4.418916055074932, + "learning_rate": 9.860733685479463e-10, + "loss": 0.6433, + "step": 2467 + }, + { + "epoch": 0.9929591631462482, + "grad_norm": 3.6128250845259724, + "learning_rate": 8.899369195727869e-10, + "loss": 0.5447, + "step": 2468 + }, + { + "epoch": 0.9933614966807484, + "grad_norm": 4.364090118207612, + "learning_rate": 7.987294055761264e-10, + "loss": 0.6099, + "step": 2469 + }, + { + "epoch": 0.9937638302152484, + "grad_norm": 3.739315819157321, + "learning_rate": 7.124510064446255e-10, + "loss": 0.4998, + "step": 2470 + }, + { + "epoch": 0.9941661637497485, + "grad_norm": 4.035011038094296, + "learning_rate": 6.311018923432776e-10, + "loss": 0.4558, + "step": 2471 + }, + { + "epoch": 0.9945684972842487, + "grad_norm": 3.577550936704116, + "learning_rate": 5.546822237145755e-10, + "loss": 0.4408, + "step": 2472 + }, + { + "epoch": 0.9949708308187487, + "grad_norm": 3.7748701163123584, + "learning_rate": 4.831921512796211e-10, + "loss": 0.4403, + "step": 2473 + }, + { + "epoch": 0.9953731643532489, + "grad_norm": 3.7403701320267686, + "learning_rate": 4.166318160361837e-10, + "loss": 0.4989, + "step": 2474 + }, + { + "epoch": 0.995775497887749, + "grad_norm": 3.8787870872673857, + "learning_rate": 3.5500134925980925e-10, + "loss": 0.5066, + "step": 2475 + }, + { + "epoch": 0.996177831422249, + "grad_norm": 4.321954937291901, + "learning_rate": 2.983008725029879e-10, + "loss": 0.498, + "step": 2476 + }, + { + "epoch": 0.9965801649567492, + "grad_norm": 3.574339539948393, + "learning_rate": 2.4653049759432167e-10, + "loss": 0.4164, + "step": 2477 + }, + { + "epoch": 0.9969824984912492, + "grad_norm": 3.7965194640316478, + "learning_rate": 1.9969032663935683e-10, + "loss": 0.4769, + "step": 2478 + }, + { + "epoch": 0.9973848320257493, + "grad_norm": 4.23022008583803, + "learning_rate": 1.5778045202002878e-10, + "loss": 0.4481, + "step": 2479 + }, + { + "epoch": 0.9977871655602495, + "grad_norm": 4.071535851310751, + "learning_rate": 1.2080095639410704e-10, + "loss": 0.535, + "step": 2480 + }, + { + "epoch": 0.9981894990947495, + "grad_norm": 3.9007108087469646, + "learning_rate": 8.875191269547279e-11, + "loss": 0.542, + "step": 2481 + }, + { + "epoch": 0.9985918326292497, + "grad_norm": 4.05416067842318, + "learning_rate": 6.163338413328612e-11, + "loss": 0.5738, + "step": 2482 + }, + { + "epoch": 0.9989941661637497, + "grad_norm": 3.9020981788420386, + "learning_rate": 3.944542419337394e-11, + "loss": 0.4897, + "step": 2483 + }, + { + "epoch": 0.9993964996982498, + "grad_norm": 3.6120331011996085, + "learning_rate": 2.2188076636009415e-11, + "loss": 0.5195, + "step": 2484 + }, + { + "epoch": 0.99979883323275, + "grad_norm": 3.4028120769775385, + "learning_rate": 9.861375497577375e-12, + "loss": 0.3763, + "step": 2485 + }, + { + "epoch": 1.0, + "grad_norm": 3.4028120769775385, + "learning_rate": 2.4653450900191666e-12, + "loss": 0.4064, + "step": 2486 + }, + { + "epoch": 1.0, + "step": 2486, + "total_flos": 882515131957248.0, + "train_loss": 0.5852410430066162, + "train_runtime": 194010.5784, + "train_samples_per_second": 1.64, + "train_steps_per_second": 0.013 + } + ], + "logging_steps": 1.0, + "max_steps": 2486, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 882515131957248.0, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}