Proxy3D-8B / trainer_state.json
Spacewanderer8263's picture
Add files using upload-large-folder tool
2bd21fb verified
raw
history blame
432 kB
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 2486,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0004023335345001006,
"grad_norm": 27.82010551409744,
"learning_rate": 0.0,
"loss": 1.7886,
"step": 1
},
{
"epoch": 0.0008046670690002012,
"grad_norm": 19.872529474831897,
"learning_rate": 2.008032128514056e-08,
"loss": 1.8136,
"step": 2
},
{
"epoch": 0.0012070006035003018,
"grad_norm": 24.180123303383308,
"learning_rate": 4.016064257028112e-08,
"loss": 1.9299,
"step": 3
},
{
"epoch": 0.0016093341380004024,
"grad_norm": 21.489391381684396,
"learning_rate": 6.02409638554217e-08,
"loss": 1.92,
"step": 4
},
{
"epoch": 0.0020116676725005027,
"grad_norm": 17.558745295130834,
"learning_rate": 8.032128514056224e-08,
"loss": 1.8227,
"step": 5
},
{
"epoch": 0.0024140012070006035,
"grad_norm": 21.46699607887353,
"learning_rate": 1.0040160642570281e-07,
"loss": 1.8824,
"step": 6
},
{
"epoch": 0.002816334741500704,
"grad_norm": 21.298475273313784,
"learning_rate": 1.204819277108434e-07,
"loss": 1.8179,
"step": 7
},
{
"epoch": 0.0032186682760008047,
"grad_norm": 21.372007205869878,
"learning_rate": 1.4056224899598394e-07,
"loss": 1.6859,
"step": 8
},
{
"epoch": 0.003621001810500905,
"grad_norm": 23.292357598058903,
"learning_rate": 1.6064257028112448e-07,
"loss": 2.0856,
"step": 9
},
{
"epoch": 0.0040233353450010055,
"grad_norm": 23.233045584039576,
"learning_rate": 1.8072289156626505e-07,
"loss": 2.1395,
"step": 10
},
{
"epoch": 0.004425668879501106,
"grad_norm": 19.4337570491022,
"learning_rate": 2.0080321285140563e-07,
"loss": 1.8688,
"step": 11
},
{
"epoch": 0.004828002414001207,
"grad_norm": 24.82725637763247,
"learning_rate": 2.208835341365462e-07,
"loss": 1.8054,
"step": 12
},
{
"epoch": 0.005230335948501308,
"grad_norm": 27.369177322875704,
"learning_rate": 2.409638554216868e-07,
"loss": 1.8794,
"step": 13
},
{
"epoch": 0.005632669483001408,
"grad_norm": 20.854134960964704,
"learning_rate": 2.610441767068273e-07,
"loss": 1.8873,
"step": 14
},
{
"epoch": 0.006035003017501509,
"grad_norm": 25.685971689153487,
"learning_rate": 2.811244979919679e-07,
"loss": 2.0743,
"step": 15
},
{
"epoch": 0.0064373365520016095,
"grad_norm": 21.501547984115398,
"learning_rate": 3.0120481927710845e-07,
"loss": 1.8415,
"step": 16
},
{
"epoch": 0.00683967008650171,
"grad_norm": 22.75991867094887,
"learning_rate": 3.2128514056224897e-07,
"loss": 1.9089,
"step": 17
},
{
"epoch": 0.00724200362100181,
"grad_norm": 22.23601729043514,
"learning_rate": 3.413654618473896e-07,
"loss": 1.9641,
"step": 18
},
{
"epoch": 0.007644337155501911,
"grad_norm": 26.35207563502954,
"learning_rate": 3.614457831325301e-07,
"loss": 2.0458,
"step": 19
},
{
"epoch": 0.008046670690002011,
"grad_norm": 20.57786492934256,
"learning_rate": 3.8152610441767073e-07,
"loss": 1.7877,
"step": 20
},
{
"epoch": 0.008449004224502113,
"grad_norm": 25.563694935501676,
"learning_rate": 4.0160642570281125e-07,
"loss": 2.1414,
"step": 21
},
{
"epoch": 0.008851337759002213,
"grad_norm": 18.98280746175437,
"learning_rate": 4.216867469879518e-07,
"loss": 1.7995,
"step": 22
},
{
"epoch": 0.009253671293502314,
"grad_norm": 22.278384224532818,
"learning_rate": 4.417670682730924e-07,
"loss": 2.0412,
"step": 23
},
{
"epoch": 0.009656004828002414,
"grad_norm": 18.256806226009903,
"learning_rate": 4.6184738955823296e-07,
"loss": 1.7538,
"step": 24
},
{
"epoch": 0.010058338362502514,
"grad_norm": 19.087838303092653,
"learning_rate": 4.819277108433736e-07,
"loss": 1.708,
"step": 25
},
{
"epoch": 0.010460671897002616,
"grad_norm": 18.961034656693933,
"learning_rate": 5.020080321285141e-07,
"loss": 1.6661,
"step": 26
},
{
"epoch": 0.010863005431502716,
"grad_norm": 18.11814523029599,
"learning_rate": 5.220883534136546e-07,
"loss": 1.753,
"step": 27
},
{
"epoch": 0.011265338966002816,
"grad_norm": 19.63691340845534,
"learning_rate": 5.421686746987952e-07,
"loss": 1.7005,
"step": 28
},
{
"epoch": 0.011667672500502917,
"grad_norm": 16.293048438980087,
"learning_rate": 5.622489959839358e-07,
"loss": 1.7269,
"step": 29
},
{
"epoch": 0.012070006035003017,
"grad_norm": 15.86827584036101,
"learning_rate": 5.823293172690764e-07,
"loss": 1.5401,
"step": 30
},
{
"epoch": 0.012472339569503117,
"grad_norm": 17.38302283506762,
"learning_rate": 6.024096385542169e-07,
"loss": 1.5943,
"step": 31
},
{
"epoch": 0.012874673104003219,
"grad_norm": 15.374566189246599,
"learning_rate": 6.224899598393574e-07,
"loss": 1.6221,
"step": 32
},
{
"epoch": 0.013277006638503319,
"grad_norm": 14.904500940732934,
"learning_rate": 6.425702811244979e-07,
"loss": 1.5452,
"step": 33
},
{
"epoch": 0.01367934017300342,
"grad_norm": 16.29526851000047,
"learning_rate": 6.626506024096387e-07,
"loss": 1.2262,
"step": 34
},
{
"epoch": 0.01408167370750352,
"grad_norm": 12.819998198911712,
"learning_rate": 6.827309236947792e-07,
"loss": 1.3476,
"step": 35
},
{
"epoch": 0.01448400724200362,
"grad_norm": 13.25182925491938,
"learning_rate": 7.028112449799197e-07,
"loss": 1.4371,
"step": 36
},
{
"epoch": 0.014886340776503722,
"grad_norm": 14.477871111934059,
"learning_rate": 7.228915662650602e-07,
"loss": 1.3476,
"step": 37
},
{
"epoch": 0.015288674311003822,
"grad_norm": 14.6196272151239,
"learning_rate": 7.429718875502008e-07,
"loss": 1.5309,
"step": 38
},
{
"epoch": 0.015691007845503924,
"grad_norm": 11.90353169885529,
"learning_rate": 7.630522088353415e-07,
"loss": 1.3026,
"step": 39
},
{
"epoch": 0.016093341380004022,
"grad_norm": 13.271008324770724,
"learning_rate": 7.83132530120482e-07,
"loss": 1.4655,
"step": 40
},
{
"epoch": 0.016495674914504124,
"grad_norm": 11.2349199407351,
"learning_rate": 8.032128514056225e-07,
"loss": 1.2782,
"step": 41
},
{
"epoch": 0.016898008449004225,
"grad_norm": 9.663111453936613,
"learning_rate": 8.232931726907631e-07,
"loss": 1.0628,
"step": 42
},
{
"epoch": 0.017300341983504323,
"grad_norm": 10.290302552273262,
"learning_rate": 8.433734939759036e-07,
"loss": 1.2398,
"step": 43
},
{
"epoch": 0.017702675518004425,
"grad_norm": 9.437964794637868,
"learning_rate": 8.634538152610443e-07,
"loss": 1.1087,
"step": 44
},
{
"epoch": 0.018105009052504527,
"grad_norm": 9.928972956598479,
"learning_rate": 8.835341365461848e-07,
"loss": 1.1231,
"step": 45
},
{
"epoch": 0.01850734258700463,
"grad_norm": 7.193942700746645,
"learning_rate": 9.036144578313254e-07,
"loss": 0.9053,
"step": 46
},
{
"epoch": 0.018909676121504727,
"grad_norm": 8.545887733782381,
"learning_rate": 9.236947791164659e-07,
"loss": 1.0188,
"step": 47
},
{
"epoch": 0.01931200965600483,
"grad_norm": 9.964310833592453,
"learning_rate": 9.437751004016064e-07,
"loss": 1.2336,
"step": 48
},
{
"epoch": 0.01971434319050493,
"grad_norm": 8.465102281501581,
"learning_rate": 9.638554216867472e-07,
"loss": 0.978,
"step": 49
},
{
"epoch": 0.020116676725005028,
"grad_norm": 8.387761909932218,
"learning_rate": 9.839357429718876e-07,
"loss": 1.0664,
"step": 50
},
{
"epoch": 0.02051901025950513,
"grad_norm": 8.853387211734997,
"learning_rate": 1.0040160642570282e-06,
"loss": 1.1316,
"step": 51
},
{
"epoch": 0.02092134379400523,
"grad_norm": 8.36811488486395,
"learning_rate": 1.0240963855421688e-06,
"loss": 0.9175,
"step": 52
},
{
"epoch": 0.02132367732850533,
"grad_norm": 8.394630361238342,
"learning_rate": 1.0441767068273092e-06,
"loss": 0.9371,
"step": 53
},
{
"epoch": 0.02172601086300543,
"grad_norm": 7.411500707300452,
"learning_rate": 1.0642570281124499e-06,
"loss": 0.9567,
"step": 54
},
{
"epoch": 0.022128344397505533,
"grad_norm": 8.194667792436356,
"learning_rate": 1.0843373493975905e-06,
"loss": 0.9994,
"step": 55
},
{
"epoch": 0.02253067793200563,
"grad_norm": 7.141458407534788,
"learning_rate": 1.1044176706827311e-06,
"loss": 0.8714,
"step": 56
},
{
"epoch": 0.022933011466505733,
"grad_norm": 5.765708921674402,
"learning_rate": 1.1244979919678715e-06,
"loss": 0.9238,
"step": 57
},
{
"epoch": 0.023335345001005835,
"grad_norm": 7.924253998059311,
"learning_rate": 1.1445783132530121e-06,
"loss": 0.9884,
"step": 58
},
{
"epoch": 0.023737678535505933,
"grad_norm": 6.597672306780267,
"learning_rate": 1.1646586345381528e-06,
"loss": 0.9904,
"step": 59
},
{
"epoch": 0.024140012070006035,
"grad_norm": 6.937688487718752,
"learning_rate": 1.1847389558232934e-06,
"loss": 0.971,
"step": 60
},
{
"epoch": 0.024542345604506136,
"grad_norm": 6.148469030855221,
"learning_rate": 1.2048192771084338e-06,
"loss": 0.7868,
"step": 61
},
{
"epoch": 0.024944679139006234,
"grad_norm": 6.388105517333996,
"learning_rate": 1.2248995983935744e-06,
"loss": 0.8567,
"step": 62
},
{
"epoch": 0.025347012673506336,
"grad_norm": 6.105324559387419,
"learning_rate": 1.2449799196787148e-06,
"loss": 0.9989,
"step": 63
},
{
"epoch": 0.025749346208006438,
"grad_norm": 6.565241933640869,
"learning_rate": 1.2650602409638555e-06,
"loss": 0.8641,
"step": 64
},
{
"epoch": 0.02615167974250654,
"grad_norm": 5.869907162685047,
"learning_rate": 1.2851405622489959e-06,
"loss": 0.9526,
"step": 65
},
{
"epoch": 0.026554013277006638,
"grad_norm": 5.851371649159624,
"learning_rate": 1.3052208835341367e-06,
"loss": 0.8387,
"step": 66
},
{
"epoch": 0.02695634681150674,
"grad_norm": 6.3493989766408765,
"learning_rate": 1.3253012048192773e-06,
"loss": 1.0458,
"step": 67
},
{
"epoch": 0.02735868034600684,
"grad_norm": 5.9283912382403114,
"learning_rate": 1.345381526104418e-06,
"loss": 1.0158,
"step": 68
},
{
"epoch": 0.02776101388050694,
"grad_norm": 6.184126721660312,
"learning_rate": 1.3654618473895584e-06,
"loss": 0.8218,
"step": 69
},
{
"epoch": 0.02816334741500704,
"grad_norm": 6.0169493477739975,
"learning_rate": 1.385542168674699e-06,
"loss": 0.7966,
"step": 70
},
{
"epoch": 0.028565680949507143,
"grad_norm": 6.084096016864222,
"learning_rate": 1.4056224899598394e-06,
"loss": 0.8221,
"step": 71
},
{
"epoch": 0.02896801448400724,
"grad_norm": 6.441697966234361,
"learning_rate": 1.42570281124498e-06,
"loss": 0.7189,
"step": 72
},
{
"epoch": 0.029370348018507342,
"grad_norm": 5.684758471405106,
"learning_rate": 1.4457831325301204e-06,
"loss": 0.811,
"step": 73
},
{
"epoch": 0.029772681553007444,
"grad_norm": 6.066554340342954,
"learning_rate": 1.465863453815261e-06,
"loss": 0.8128,
"step": 74
},
{
"epoch": 0.030175015087507542,
"grad_norm": 5.398586406654046,
"learning_rate": 1.4859437751004017e-06,
"loss": 0.8521,
"step": 75
},
{
"epoch": 0.030577348622007644,
"grad_norm": 5.72164636854203,
"learning_rate": 1.5060240963855425e-06,
"loss": 0.8733,
"step": 76
},
{
"epoch": 0.030979682156507746,
"grad_norm": 5.317662321058184,
"learning_rate": 1.526104417670683e-06,
"loss": 0.6758,
"step": 77
},
{
"epoch": 0.03138201569100785,
"grad_norm": 5.008046012529329,
"learning_rate": 1.5461847389558236e-06,
"loss": 0.77,
"step": 78
},
{
"epoch": 0.03178434922550795,
"grad_norm": 5.806938383916589,
"learning_rate": 1.566265060240964e-06,
"loss": 0.9218,
"step": 79
},
{
"epoch": 0.032186682760008044,
"grad_norm": 5.687427641764964,
"learning_rate": 1.5863453815261046e-06,
"loss": 0.873,
"step": 80
},
{
"epoch": 0.032589016294508145,
"grad_norm": 5.877177229539529,
"learning_rate": 1.606425702811245e-06,
"loss": 0.8347,
"step": 81
},
{
"epoch": 0.03299134982900825,
"grad_norm": 5.661277204998088,
"learning_rate": 1.6265060240963856e-06,
"loss": 0.7537,
"step": 82
},
{
"epoch": 0.03339368336350835,
"grad_norm": 6.0004706518847675,
"learning_rate": 1.6465863453815263e-06,
"loss": 0.9341,
"step": 83
},
{
"epoch": 0.03379601689800845,
"grad_norm": 5.383372715468371,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.716,
"step": 84
},
{
"epoch": 0.03419835043250855,
"grad_norm": 5.870492937673813,
"learning_rate": 1.6867469879518073e-06,
"loss": 0.8515,
"step": 85
},
{
"epoch": 0.03460068396700865,
"grad_norm": 5.40282247472889,
"learning_rate": 1.7068273092369481e-06,
"loss": 0.7575,
"step": 86
},
{
"epoch": 0.03500301750150875,
"grad_norm": 5.244298542730251,
"learning_rate": 1.7269076305220885e-06,
"loss": 0.7342,
"step": 87
},
{
"epoch": 0.03540535103600885,
"grad_norm": 5.118324863903489,
"learning_rate": 1.7469879518072292e-06,
"loss": 0.7633,
"step": 88
},
{
"epoch": 0.03580768457050895,
"grad_norm": 5.559590746485852,
"learning_rate": 1.7670682730923696e-06,
"loss": 0.915,
"step": 89
},
{
"epoch": 0.036210018105009054,
"grad_norm": 5.093580932305843,
"learning_rate": 1.7871485943775102e-06,
"loss": 0.7234,
"step": 90
},
{
"epoch": 0.036612351639509155,
"grad_norm": 5.949076037097858,
"learning_rate": 1.8072289156626508e-06,
"loss": 0.9392,
"step": 91
},
{
"epoch": 0.03701468517400926,
"grad_norm": 6.126962098905458,
"learning_rate": 1.8273092369477912e-06,
"loss": 0.7973,
"step": 92
},
{
"epoch": 0.03741701870850935,
"grad_norm": 5.418925049178253,
"learning_rate": 1.8473895582329318e-06,
"loss": 0.7386,
"step": 93
},
{
"epoch": 0.03781935224300945,
"grad_norm": 5.30622581323819,
"learning_rate": 1.8674698795180723e-06,
"loss": 0.7712,
"step": 94
},
{
"epoch": 0.038221685777509555,
"grad_norm": 5.203905538553713,
"learning_rate": 1.8875502008032129e-06,
"loss": 0.7403,
"step": 95
},
{
"epoch": 0.03862401931200966,
"grad_norm": 5.497752744670633,
"learning_rate": 1.9076305220883537e-06,
"loss": 0.8343,
"step": 96
},
{
"epoch": 0.03902635284650976,
"grad_norm": 5.792666990034754,
"learning_rate": 1.9277108433734943e-06,
"loss": 0.9855,
"step": 97
},
{
"epoch": 0.03942868638100986,
"grad_norm": 4.95103734609948,
"learning_rate": 1.947791164658635e-06,
"loss": 0.7676,
"step": 98
},
{
"epoch": 0.039831019915509955,
"grad_norm": 5.313239350915902,
"learning_rate": 1.967871485943775e-06,
"loss": 0.8112,
"step": 99
},
{
"epoch": 0.040233353450010056,
"grad_norm": 5.262581677945773,
"learning_rate": 1.987951807228916e-06,
"loss": 0.7726,
"step": 100
},
{
"epoch": 0.04063568698451016,
"grad_norm": 5.7284063119812965,
"learning_rate": 2.0080321285140564e-06,
"loss": 0.7143,
"step": 101
},
{
"epoch": 0.04103802051901026,
"grad_norm": 5.61074467493987,
"learning_rate": 2.028112449799197e-06,
"loss": 0.7821,
"step": 102
},
{
"epoch": 0.04144035405351036,
"grad_norm": 5.668818940572837,
"learning_rate": 2.0481927710843377e-06,
"loss": 0.6217,
"step": 103
},
{
"epoch": 0.04184268758801046,
"grad_norm": 6.102827563222031,
"learning_rate": 2.068273092369478e-06,
"loss": 0.7094,
"step": 104
},
{
"epoch": 0.04224502112251056,
"grad_norm": 5.106901986680839,
"learning_rate": 2.0883534136546185e-06,
"loss": 0.7452,
"step": 105
},
{
"epoch": 0.04264735465701066,
"grad_norm": 5.320689458556449,
"learning_rate": 2.1084337349397595e-06,
"loss": 0.8121,
"step": 106
},
{
"epoch": 0.04304968819151076,
"grad_norm": 4.851575241509713,
"learning_rate": 2.1285140562248997e-06,
"loss": 0.7695,
"step": 107
},
{
"epoch": 0.04345202172601086,
"grad_norm": 5.7294050539399315,
"learning_rate": 2.1485943775100404e-06,
"loss": 0.8343,
"step": 108
},
{
"epoch": 0.043854355260510965,
"grad_norm": 6.2102432153860265,
"learning_rate": 2.168674698795181e-06,
"loss": 0.8478,
"step": 109
},
{
"epoch": 0.044256688795011066,
"grad_norm": 5.8055227032618415,
"learning_rate": 2.1887550200803216e-06,
"loss": 0.7647,
"step": 110
},
{
"epoch": 0.04465902232951117,
"grad_norm": 5.59165747057277,
"learning_rate": 2.2088353413654622e-06,
"loss": 0.6905,
"step": 111
},
{
"epoch": 0.04506135586401126,
"grad_norm": 4.682897930842293,
"learning_rate": 2.2289156626506024e-06,
"loss": 0.6875,
"step": 112
},
{
"epoch": 0.045463689398511364,
"grad_norm": 4.320466219380915,
"learning_rate": 2.248995983935743e-06,
"loss": 0.696,
"step": 113
},
{
"epoch": 0.045866022933011466,
"grad_norm": 5.050663115560423,
"learning_rate": 2.2690763052208837e-06,
"loss": 0.6224,
"step": 114
},
{
"epoch": 0.04626835646751157,
"grad_norm": 4.681780822596801,
"learning_rate": 2.2891566265060243e-06,
"loss": 0.6462,
"step": 115
},
{
"epoch": 0.04667069000201167,
"grad_norm": 5.954135299109462,
"learning_rate": 2.309236947791165e-06,
"loss": 0.8848,
"step": 116
},
{
"epoch": 0.04707302353651177,
"grad_norm": 5.573998592787985,
"learning_rate": 2.3293172690763055e-06,
"loss": 0.8088,
"step": 117
},
{
"epoch": 0.047475357071011866,
"grad_norm": 5.070007694615057,
"learning_rate": 2.349397590361446e-06,
"loss": 0.7173,
"step": 118
},
{
"epoch": 0.04787769060551197,
"grad_norm": 5.671554086319649,
"learning_rate": 2.3694779116465868e-06,
"loss": 0.8136,
"step": 119
},
{
"epoch": 0.04828002414001207,
"grad_norm": 5.769697685018851,
"learning_rate": 2.389558232931727e-06,
"loss": 0.7733,
"step": 120
},
{
"epoch": 0.04868235767451217,
"grad_norm": 4.920298621449247,
"learning_rate": 2.4096385542168676e-06,
"loss": 0.5931,
"step": 121
},
{
"epoch": 0.04908469120901227,
"grad_norm": 5.49690667652572,
"learning_rate": 2.4297188755020082e-06,
"loss": 0.8079,
"step": 122
},
{
"epoch": 0.049487024743512374,
"grad_norm": 5.369187249008034,
"learning_rate": 2.449799196787149e-06,
"loss": 0.726,
"step": 123
},
{
"epoch": 0.04988935827801247,
"grad_norm": 4.8515152911462796,
"learning_rate": 2.469879518072289e-06,
"loss": 0.7691,
"step": 124
},
{
"epoch": 0.05029169181251257,
"grad_norm": 5.099204794104859,
"learning_rate": 2.4899598393574297e-06,
"loss": 0.8361,
"step": 125
},
{
"epoch": 0.05069402534701267,
"grad_norm": 5.390609640239935,
"learning_rate": 2.5100401606425707e-06,
"loss": 0.6527,
"step": 126
},
{
"epoch": 0.051096358881512774,
"grad_norm": 5.053936099287885,
"learning_rate": 2.530120481927711e-06,
"loss": 0.7326,
"step": 127
},
{
"epoch": 0.051498692416012876,
"grad_norm": 5.876719841742455,
"learning_rate": 2.5502008032128516e-06,
"loss": 0.8571,
"step": 128
},
{
"epoch": 0.05190102595051298,
"grad_norm": 5.291120710287588,
"learning_rate": 2.5702811244979918e-06,
"loss": 0.7641,
"step": 129
},
{
"epoch": 0.05230335948501308,
"grad_norm": 5.2648612797895336,
"learning_rate": 2.590361445783133e-06,
"loss": 0.768,
"step": 130
},
{
"epoch": 0.052705693019513174,
"grad_norm": 5.146667666976231,
"learning_rate": 2.6104417670682734e-06,
"loss": 0.729,
"step": 131
},
{
"epoch": 0.053108026554013275,
"grad_norm": 5.607251668752912,
"learning_rate": 2.6305220883534136e-06,
"loss": 0.7812,
"step": 132
},
{
"epoch": 0.05351036008851338,
"grad_norm": 4.774953294338218,
"learning_rate": 2.6506024096385547e-06,
"loss": 0.6996,
"step": 133
},
{
"epoch": 0.05391269362301348,
"grad_norm": 5.187652607031923,
"learning_rate": 2.670682730923695e-06,
"loss": 0.7029,
"step": 134
},
{
"epoch": 0.05431502715751358,
"grad_norm": 5.631076339839678,
"learning_rate": 2.690763052208836e-06,
"loss": 0.7618,
"step": 135
},
{
"epoch": 0.05471736069201368,
"grad_norm": 5.1080370644891735,
"learning_rate": 2.710843373493976e-06,
"loss": 0.6746,
"step": 136
},
{
"epoch": 0.05511969422651378,
"grad_norm": 5.177774314908037,
"learning_rate": 2.7309236947791167e-06,
"loss": 0.6576,
"step": 137
},
{
"epoch": 0.05552202776101388,
"grad_norm": 5.4428731816372355,
"learning_rate": 2.751004016064257e-06,
"loss": 0.8666,
"step": 138
},
{
"epoch": 0.05592436129551398,
"grad_norm": 4.737577892176393,
"learning_rate": 2.771084337349398e-06,
"loss": 0.6303,
"step": 139
},
{
"epoch": 0.05632669483001408,
"grad_norm": 5.958074686629733,
"learning_rate": 2.791164658634538e-06,
"loss": 0.7375,
"step": 140
},
{
"epoch": 0.056729028364514184,
"grad_norm": 5.347638262573924,
"learning_rate": 2.811244979919679e-06,
"loss": 0.7496,
"step": 141
},
{
"epoch": 0.057131361899014285,
"grad_norm": 5.25633214854894,
"learning_rate": 2.83132530120482e-06,
"loss": 0.7703,
"step": 142
},
{
"epoch": 0.05753369543351439,
"grad_norm": 6.3493191578507355,
"learning_rate": 2.85140562248996e-06,
"loss": 0.7297,
"step": 143
},
{
"epoch": 0.05793602896801448,
"grad_norm": 5.291536986822124,
"learning_rate": 2.8714859437751007e-06,
"loss": 0.7848,
"step": 144
},
{
"epoch": 0.05833836250251458,
"grad_norm": 4.327776932092466,
"learning_rate": 2.891566265060241e-06,
"loss": 0.5537,
"step": 145
},
{
"epoch": 0.058740696037014685,
"grad_norm": 4.827406966450388,
"learning_rate": 2.911646586345382e-06,
"loss": 0.6614,
"step": 146
},
{
"epoch": 0.05914302957151479,
"grad_norm": 5.4512595989034605,
"learning_rate": 2.931726907630522e-06,
"loss": 0.5105,
"step": 147
},
{
"epoch": 0.05954536310601489,
"grad_norm": 5.349806231832712,
"learning_rate": 2.9518072289156627e-06,
"loss": 0.7398,
"step": 148
},
{
"epoch": 0.05994769664051499,
"grad_norm": 5.376464953972205,
"learning_rate": 2.9718875502008034e-06,
"loss": 0.832,
"step": 149
},
{
"epoch": 0.060350030175015085,
"grad_norm": 4.4973784287811,
"learning_rate": 2.991967871485944e-06,
"loss": 0.6631,
"step": 150
},
{
"epoch": 0.060752363709515186,
"grad_norm": 4.1574904576313925,
"learning_rate": 3.012048192771085e-06,
"loss": 0.5736,
"step": 151
},
{
"epoch": 0.06115469724401529,
"grad_norm": 4.932454920014607,
"learning_rate": 3.0321285140562252e-06,
"loss": 0.6843,
"step": 152
},
{
"epoch": 0.06155703077851539,
"grad_norm": 4.920072821237323,
"learning_rate": 3.052208835341366e-06,
"loss": 0.6115,
"step": 153
},
{
"epoch": 0.06195936431301549,
"grad_norm": 5.069151288790082,
"learning_rate": 3.072289156626506e-06,
"loss": 0.6592,
"step": 154
},
{
"epoch": 0.06236169784751559,
"grad_norm": 4.940734045454529,
"learning_rate": 3.092369477911647e-06,
"loss": 0.6823,
"step": 155
},
{
"epoch": 0.0627640313820157,
"grad_norm": 4.7329970403721635,
"learning_rate": 3.1124497991967873e-06,
"loss": 0.6183,
"step": 156
},
{
"epoch": 0.0631663649165158,
"grad_norm": 5.584324231160161,
"learning_rate": 3.132530120481928e-06,
"loss": 0.8407,
"step": 157
},
{
"epoch": 0.0635686984510159,
"grad_norm": 4.9494988025514095,
"learning_rate": 3.152610441767068e-06,
"loss": 0.7362,
"step": 158
},
{
"epoch": 0.06397103198551599,
"grad_norm": 4.826336881024485,
"learning_rate": 3.172690763052209e-06,
"loss": 0.6756,
"step": 159
},
{
"epoch": 0.06437336552001609,
"grad_norm": 4.663584106854257,
"learning_rate": 3.1927710843373494e-06,
"loss": 0.6618,
"step": 160
},
{
"epoch": 0.06477569905451619,
"grad_norm": 5.358113823802763,
"learning_rate": 3.21285140562249e-06,
"loss": 0.7845,
"step": 161
},
{
"epoch": 0.06517803258901629,
"grad_norm": 4.393966562734061,
"learning_rate": 3.232931726907631e-06,
"loss": 0.6643,
"step": 162
},
{
"epoch": 0.06558036612351639,
"grad_norm": 5.269005574739929,
"learning_rate": 3.2530120481927713e-06,
"loss": 0.7892,
"step": 163
},
{
"epoch": 0.0659826996580165,
"grad_norm": 5.191377211307637,
"learning_rate": 3.2730923694779123e-06,
"loss": 0.6834,
"step": 164
},
{
"epoch": 0.0663850331925166,
"grad_norm": 5.9411717854682875,
"learning_rate": 3.2931726907630525e-06,
"loss": 0.682,
"step": 165
},
{
"epoch": 0.0667873667270167,
"grad_norm": 4.657895689681778,
"learning_rate": 3.313253012048193e-06,
"loss": 0.6533,
"step": 166
},
{
"epoch": 0.0671897002615168,
"grad_norm": 4.689805393308291,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.5851,
"step": 167
},
{
"epoch": 0.0675920337960169,
"grad_norm": 4.934046837424603,
"learning_rate": 3.3534136546184744e-06,
"loss": 0.7834,
"step": 168
},
{
"epoch": 0.067994367330517,
"grad_norm": 5.295304511873193,
"learning_rate": 3.3734939759036146e-06,
"loss": 0.6978,
"step": 169
},
{
"epoch": 0.0683967008650171,
"grad_norm": 5.8805644086916935,
"learning_rate": 3.393574297188755e-06,
"loss": 0.7022,
"step": 170
},
{
"epoch": 0.0687990343995172,
"grad_norm": 4.986410631963467,
"learning_rate": 3.4136546184738962e-06,
"loss": 0.8589,
"step": 171
},
{
"epoch": 0.0692013679340173,
"grad_norm": 4.587536055949948,
"learning_rate": 3.4337349397590364e-06,
"loss": 0.5962,
"step": 172
},
{
"epoch": 0.0696037014685174,
"grad_norm": 4.542521557186459,
"learning_rate": 3.453815261044177e-06,
"loss": 0.5778,
"step": 173
},
{
"epoch": 0.0700060350030175,
"grad_norm": 5.006237220238328,
"learning_rate": 3.4738955823293173e-06,
"loss": 0.7066,
"step": 174
},
{
"epoch": 0.0704083685375176,
"grad_norm": 5.03127022104005,
"learning_rate": 3.4939759036144583e-06,
"loss": 0.7143,
"step": 175
},
{
"epoch": 0.0708107020720177,
"grad_norm": 4.978976609084576,
"learning_rate": 3.5140562248995985e-06,
"loss": 0.5806,
"step": 176
},
{
"epoch": 0.0712130356065178,
"grad_norm": 4.953415145134539,
"learning_rate": 3.534136546184739e-06,
"loss": 0.6169,
"step": 177
},
{
"epoch": 0.0716153691410179,
"grad_norm": 5.224274445585009,
"learning_rate": 3.5542168674698798e-06,
"loss": 0.7055,
"step": 178
},
{
"epoch": 0.072017702675518,
"grad_norm": 4.793231833957601,
"learning_rate": 3.5742971887550204e-06,
"loss": 0.6589,
"step": 179
},
{
"epoch": 0.07242003621001811,
"grad_norm": 4.897332398174339,
"learning_rate": 3.5943775100401606e-06,
"loss": 0.5835,
"step": 180
},
{
"epoch": 0.07282236974451821,
"grad_norm": 5.461164108231627,
"learning_rate": 3.6144578313253016e-06,
"loss": 0.6153,
"step": 181
},
{
"epoch": 0.07322470327901831,
"grad_norm": 4.750713371821981,
"learning_rate": 3.6345381526104423e-06,
"loss": 0.5576,
"step": 182
},
{
"epoch": 0.07362703681351841,
"grad_norm": 4.612815212986164,
"learning_rate": 3.6546184738955825e-06,
"loss": 0.6695,
"step": 183
},
{
"epoch": 0.07402937034801851,
"grad_norm": 5.035137988160493,
"learning_rate": 3.6746987951807235e-06,
"loss": 0.6451,
"step": 184
},
{
"epoch": 0.0744317038825186,
"grad_norm": 5.189253752749726,
"learning_rate": 3.6947791164658637e-06,
"loss": 0.6897,
"step": 185
},
{
"epoch": 0.0748340374170187,
"grad_norm": 5.061038391462879,
"learning_rate": 3.7148594377510043e-06,
"loss": 0.7363,
"step": 186
},
{
"epoch": 0.0752363709515188,
"grad_norm": 5.12457801565824,
"learning_rate": 3.7349397590361445e-06,
"loss": 0.6662,
"step": 187
},
{
"epoch": 0.0756387044860189,
"grad_norm": 5.468566358022457,
"learning_rate": 3.7550200803212856e-06,
"loss": 0.7333,
"step": 188
},
{
"epoch": 0.07604103802051901,
"grad_norm": 4.83323388980342,
"learning_rate": 3.7751004016064258e-06,
"loss": 0.6179,
"step": 189
},
{
"epoch": 0.07644337155501911,
"grad_norm": 5.618759075616741,
"learning_rate": 3.7951807228915664e-06,
"loss": 0.7861,
"step": 190
},
{
"epoch": 0.07684570508951921,
"grad_norm": 5.571721878536531,
"learning_rate": 3.8152610441767074e-06,
"loss": 0.6481,
"step": 191
},
{
"epoch": 0.07724803862401931,
"grad_norm": 5.067103682358002,
"learning_rate": 3.835341365461848e-06,
"loss": 0.7295,
"step": 192
},
{
"epoch": 0.07765037215851942,
"grad_norm": 4.945134722999282,
"learning_rate": 3.855421686746989e-06,
"loss": 0.6275,
"step": 193
},
{
"epoch": 0.07805270569301952,
"grad_norm": 4.995120702157204,
"learning_rate": 3.875502008032129e-06,
"loss": 0.6156,
"step": 194
},
{
"epoch": 0.07845503922751962,
"grad_norm": 4.936376571967402,
"learning_rate": 3.89558232931727e-06,
"loss": 0.6141,
"step": 195
},
{
"epoch": 0.07885737276201972,
"grad_norm": 5.014965363677265,
"learning_rate": 3.91566265060241e-06,
"loss": 0.6706,
"step": 196
},
{
"epoch": 0.07925970629651982,
"grad_norm": 4.379077775855287,
"learning_rate": 3.93574297188755e-06,
"loss": 0.625,
"step": 197
},
{
"epoch": 0.07966203983101991,
"grad_norm": 5.213065143955326,
"learning_rate": 3.9558232931726905e-06,
"loss": 0.6259,
"step": 198
},
{
"epoch": 0.08006437336552001,
"grad_norm": 4.943005393732397,
"learning_rate": 3.975903614457832e-06,
"loss": 0.6867,
"step": 199
},
{
"epoch": 0.08046670690002011,
"grad_norm": 6.1486973779873795,
"learning_rate": 3.995983935742972e-06,
"loss": 0.8347,
"step": 200
},
{
"epoch": 0.08086904043452021,
"grad_norm": 4.733552314219692,
"learning_rate": 4.016064257028113e-06,
"loss": 0.6611,
"step": 201
},
{
"epoch": 0.08127137396902032,
"grad_norm": 4.8269187112270195,
"learning_rate": 4.036144578313254e-06,
"loss": 0.6939,
"step": 202
},
{
"epoch": 0.08167370750352042,
"grad_norm": 5.3439700963806125,
"learning_rate": 4.056224899598394e-06,
"loss": 0.7133,
"step": 203
},
{
"epoch": 0.08207604103802052,
"grad_norm": 5.29998702305394,
"learning_rate": 4.076305220883534e-06,
"loss": 0.7792,
"step": 204
},
{
"epoch": 0.08247837457252062,
"grad_norm": 5.408216152846704,
"learning_rate": 4.096385542168675e-06,
"loss": 0.6411,
"step": 205
},
{
"epoch": 0.08288070810702072,
"grad_norm": 4.419630999712613,
"learning_rate": 4.1164658634538155e-06,
"loss": 0.615,
"step": 206
},
{
"epoch": 0.08328304164152082,
"grad_norm": 4.543564292966689,
"learning_rate": 4.136546184738956e-06,
"loss": 0.5462,
"step": 207
},
{
"epoch": 0.08368537517602093,
"grad_norm": 4.814136944508732,
"learning_rate": 4.156626506024097e-06,
"loss": 0.6407,
"step": 208
},
{
"epoch": 0.08408770871052103,
"grad_norm": 5.118958059055448,
"learning_rate": 4.176706827309237e-06,
"loss": 0.6922,
"step": 209
},
{
"epoch": 0.08449004224502112,
"grad_norm": 5.761968321529592,
"learning_rate": 4.196787148594378e-06,
"loss": 0.7785,
"step": 210
},
{
"epoch": 0.08489237577952122,
"grad_norm": 4.7973887668760185,
"learning_rate": 4.216867469879519e-06,
"loss": 0.5855,
"step": 211
},
{
"epoch": 0.08529470931402132,
"grad_norm": 5.063119851943765,
"learning_rate": 4.236947791164659e-06,
"loss": 0.5061,
"step": 212
},
{
"epoch": 0.08569704284852142,
"grad_norm": 5.185457141038217,
"learning_rate": 4.2570281124497995e-06,
"loss": 0.6547,
"step": 213
},
{
"epoch": 0.08609937638302152,
"grad_norm": 4.908063923355144,
"learning_rate": 4.27710843373494e-06,
"loss": 0.6325,
"step": 214
},
{
"epoch": 0.08650170991752162,
"grad_norm": 4.902236509892525,
"learning_rate": 4.297188755020081e-06,
"loss": 0.771,
"step": 215
},
{
"epoch": 0.08690404345202173,
"grad_norm": 4.939421470208826,
"learning_rate": 4.317269076305221e-06,
"loss": 0.6009,
"step": 216
},
{
"epoch": 0.08730637698652183,
"grad_norm": 5.585627020639388,
"learning_rate": 4.337349397590362e-06,
"loss": 0.75,
"step": 217
},
{
"epoch": 0.08770871052102193,
"grad_norm": 5.768733340993033,
"learning_rate": 4.357429718875502e-06,
"loss": 0.8889,
"step": 218
},
{
"epoch": 0.08811104405552203,
"grad_norm": 4.756073859984964,
"learning_rate": 4.377510040160643e-06,
"loss": 0.5845,
"step": 219
},
{
"epoch": 0.08851337759002213,
"grad_norm": 5.602957070495176,
"learning_rate": 4.397590361445783e-06,
"loss": 0.5766,
"step": 220
},
{
"epoch": 0.08891571112452223,
"grad_norm": 5.51373108651177,
"learning_rate": 4.4176706827309244e-06,
"loss": 0.667,
"step": 221
},
{
"epoch": 0.08931804465902234,
"grad_norm": 5.2818421493270895,
"learning_rate": 4.437751004016065e-06,
"loss": 0.6132,
"step": 222
},
{
"epoch": 0.08972037819352242,
"grad_norm": 4.687481374933495,
"learning_rate": 4.457831325301205e-06,
"loss": 0.69,
"step": 223
},
{
"epoch": 0.09012271172802253,
"grad_norm": 5.045402036516882,
"learning_rate": 4.477911646586346e-06,
"loss": 0.5496,
"step": 224
},
{
"epoch": 0.09052504526252263,
"grad_norm": 4.949655367217769,
"learning_rate": 4.497991967871486e-06,
"loss": 0.6777,
"step": 225
},
{
"epoch": 0.09092737879702273,
"grad_norm": 4.725446380096972,
"learning_rate": 4.518072289156627e-06,
"loss": 0.5211,
"step": 226
},
{
"epoch": 0.09132971233152283,
"grad_norm": 5.627747083860127,
"learning_rate": 4.538152610441767e-06,
"loss": 0.7335,
"step": 227
},
{
"epoch": 0.09173204586602293,
"grad_norm": 4.658915831199445,
"learning_rate": 4.558232931726908e-06,
"loss": 0.618,
"step": 228
},
{
"epoch": 0.09213437940052303,
"grad_norm": 5.169380307261355,
"learning_rate": 4.578313253012049e-06,
"loss": 0.8083,
"step": 229
},
{
"epoch": 0.09253671293502314,
"grad_norm": 4.462280637097748,
"learning_rate": 4.598393574297189e-06,
"loss": 0.5873,
"step": 230
},
{
"epoch": 0.09293904646952324,
"grad_norm": 5.088180514844276,
"learning_rate": 4.61847389558233e-06,
"loss": 0.6948,
"step": 231
},
{
"epoch": 0.09334138000402334,
"grad_norm": 4.911645340461795,
"learning_rate": 4.63855421686747e-06,
"loss": 0.5014,
"step": 232
},
{
"epoch": 0.09374371353852344,
"grad_norm": 5.183650867648958,
"learning_rate": 4.658634538152611e-06,
"loss": 0.6731,
"step": 233
},
{
"epoch": 0.09414604707302354,
"grad_norm": 4.942647346081943,
"learning_rate": 4.678714859437751e-06,
"loss": 0.796,
"step": 234
},
{
"epoch": 0.09454838060752364,
"grad_norm": 5.413754367435916,
"learning_rate": 4.698795180722892e-06,
"loss": 0.7029,
"step": 235
},
{
"epoch": 0.09495071414202373,
"grad_norm": 5.394996509930947,
"learning_rate": 4.7188755020080325e-06,
"loss": 0.7448,
"step": 236
},
{
"epoch": 0.09535304767652383,
"grad_norm": 4.807099705141704,
"learning_rate": 4.7389558232931736e-06,
"loss": 0.6608,
"step": 237
},
{
"epoch": 0.09575538121102393,
"grad_norm": 5.204336077501797,
"learning_rate": 4.759036144578314e-06,
"loss": 0.585,
"step": 238
},
{
"epoch": 0.09615771474552404,
"grad_norm": 4.845053158029385,
"learning_rate": 4.779116465863454e-06,
"loss": 0.6299,
"step": 239
},
{
"epoch": 0.09656004828002414,
"grad_norm": 5.037832345926416,
"learning_rate": 4.799196787148594e-06,
"loss": 0.6598,
"step": 240
},
{
"epoch": 0.09696238181452424,
"grad_norm": 4.930571962391263,
"learning_rate": 4.819277108433735e-06,
"loss": 0.5918,
"step": 241
},
{
"epoch": 0.09736471534902434,
"grad_norm": 4.749158619882368,
"learning_rate": 4.839357429718876e-06,
"loss": 0.5785,
"step": 242
},
{
"epoch": 0.09776704888352444,
"grad_norm": 5.774299138017609,
"learning_rate": 4.8594377510040165e-06,
"loss": 0.7465,
"step": 243
},
{
"epoch": 0.09816938241802455,
"grad_norm": 4.582929546310413,
"learning_rate": 4.8795180722891575e-06,
"loss": 0.5464,
"step": 244
},
{
"epoch": 0.09857171595252465,
"grad_norm": 4.604283671045787,
"learning_rate": 4.899598393574298e-06,
"loss": 0.6183,
"step": 245
},
{
"epoch": 0.09897404948702475,
"grad_norm": 5.126092092509994,
"learning_rate": 4.919678714859438e-06,
"loss": 0.7267,
"step": 246
},
{
"epoch": 0.09937638302152485,
"grad_norm": 4.6087281951772345,
"learning_rate": 4.939759036144578e-06,
"loss": 0.5911,
"step": 247
},
{
"epoch": 0.09977871655602494,
"grad_norm": 4.4308954136706165,
"learning_rate": 4.959839357429719e-06,
"loss": 0.6525,
"step": 248
},
{
"epoch": 0.10018105009052504,
"grad_norm": 4.704767612087696,
"learning_rate": 4.979919678714859e-06,
"loss": 0.7537,
"step": 249
},
{
"epoch": 0.10058338362502514,
"grad_norm": 5.171569719861641,
"learning_rate": 5e-06,
"loss": 0.5941,
"step": 250
},
{
"epoch": 0.10098571715952524,
"grad_norm": 5.7449160884679245,
"learning_rate": 4.99999753465491e-06,
"loss": 0.6686,
"step": 251
},
{
"epoch": 0.10138805069402534,
"grad_norm": 4.648308764039487,
"learning_rate": 4.999990138624503e-06,
"loss": 0.6264,
"step": 252
},
{
"epoch": 0.10179038422852545,
"grad_norm": 5.064339952308565,
"learning_rate": 4.999977811923365e-06,
"loss": 0.6921,
"step": 253
},
{
"epoch": 0.10219271776302555,
"grad_norm": 4.810326779178174,
"learning_rate": 4.9999605545758065e-06,
"loss": 0.6119,
"step": 254
},
{
"epoch": 0.10259505129752565,
"grad_norm": 4.805314994215823,
"learning_rate": 4.999938366615867e-06,
"loss": 0.6364,
"step": 255
},
{
"epoch": 0.10299738483202575,
"grad_norm": 4.473992853756316,
"learning_rate": 4.999911248087305e-06,
"loss": 0.5508,
"step": 256
},
{
"epoch": 0.10339971836652585,
"grad_norm": 5.452559581429947,
"learning_rate": 4.999879199043606e-06,
"loss": 0.7831,
"step": 257
},
{
"epoch": 0.10380205190102595,
"grad_norm": 5.061524793864594,
"learning_rate": 4.99984221954798e-06,
"loss": 0.5644,
"step": 258
},
{
"epoch": 0.10420438543552606,
"grad_norm": 4.673975167806026,
"learning_rate": 4.999800309673361e-06,
"loss": 0.649,
"step": 259
},
{
"epoch": 0.10460671897002616,
"grad_norm": 4.8181768852216855,
"learning_rate": 4.999753469502406e-06,
"loss": 0.6528,
"step": 260
},
{
"epoch": 0.10500905250452625,
"grad_norm": 5.3922298074237345,
"learning_rate": 4.999701699127497e-06,
"loss": 0.8216,
"step": 261
},
{
"epoch": 0.10541138603902635,
"grad_norm": 4.463138014998754,
"learning_rate": 4.99964499865074e-06,
"loss": 0.6824,
"step": 262
},
{
"epoch": 0.10581371957352645,
"grad_norm": 4.937246753810877,
"learning_rate": 4.999583368183965e-06,
"loss": 0.5256,
"step": 263
},
{
"epoch": 0.10621605310802655,
"grad_norm": 5.02716535200141,
"learning_rate": 4.999516807848721e-06,
"loss": 0.6658,
"step": 264
},
{
"epoch": 0.10661838664252665,
"grad_norm": 4.719903057858775,
"learning_rate": 4.999445317776286e-06,
"loss": 0.5943,
"step": 265
},
{
"epoch": 0.10702072017702675,
"grad_norm": 4.327891289019763,
"learning_rate": 4.9993688981076574e-06,
"loss": 0.6267,
"step": 266
},
{
"epoch": 0.10742305371152686,
"grad_norm": 5.04792717283726,
"learning_rate": 4.999287548993555e-06,
"loss": 0.7649,
"step": 267
},
{
"epoch": 0.10782538724602696,
"grad_norm": 4.398384896034094,
"learning_rate": 4.999201270594424e-06,
"loss": 0.6237,
"step": 268
},
{
"epoch": 0.10822772078052706,
"grad_norm": 4.513204332172601,
"learning_rate": 4.999110063080428e-06,
"loss": 0.685,
"step": 269
},
{
"epoch": 0.10863005431502716,
"grad_norm": 4.910504259364484,
"learning_rate": 4.999013926631453e-06,
"loss": 0.6393,
"step": 270
},
{
"epoch": 0.10903238784952726,
"grad_norm": 3.8051932823856927,
"learning_rate": 4.998912861437106e-06,
"loss": 0.557,
"step": 271
},
{
"epoch": 0.10943472138402736,
"grad_norm": 4.261798228362777,
"learning_rate": 4.9988068676967176e-06,
"loss": 0.6685,
"step": 272
},
{
"epoch": 0.10983705491852747,
"grad_norm": 5.199991754679721,
"learning_rate": 4.9986959456193345e-06,
"loss": 0.6926,
"step": 273
},
{
"epoch": 0.11023938845302755,
"grad_norm": 4.800143443642004,
"learning_rate": 4.998580095423728e-06,
"loss": 0.6262,
"step": 274
},
{
"epoch": 0.11064172198752766,
"grad_norm": 5.703676474195908,
"learning_rate": 4.998459317338383e-06,
"loss": 0.8102,
"step": 275
},
{
"epoch": 0.11104405552202776,
"grad_norm": 5.554521472398849,
"learning_rate": 4.998333611601511e-06,
"loss": 0.8502,
"step": 276
},
{
"epoch": 0.11144638905652786,
"grad_norm": 5.223628882218702,
"learning_rate": 4.998202978461038e-06,
"loss": 0.6968,
"step": 277
},
{
"epoch": 0.11184872259102796,
"grad_norm": 5.028503778823592,
"learning_rate": 4.9980674181746055e-06,
"loss": 0.6237,
"step": 278
},
{
"epoch": 0.11225105612552806,
"grad_norm": 4.810062581233454,
"learning_rate": 4.997926931009579e-06,
"loss": 0.6566,
"step": 279
},
{
"epoch": 0.11265338966002816,
"grad_norm": 4.895795240240787,
"learning_rate": 4.997781517243037e-06,
"loss": 0.6358,
"step": 280
},
{
"epoch": 0.11305572319452827,
"grad_norm": 4.678689813671165,
"learning_rate": 4.997631177161775e-06,
"loss": 0.7098,
"step": 281
},
{
"epoch": 0.11345805672902837,
"grad_norm": 4.8363206866892465,
"learning_rate": 4.9974759110623065e-06,
"loss": 0.7234,
"step": 282
},
{
"epoch": 0.11386039026352847,
"grad_norm": 4.872936830415585,
"learning_rate": 4.997315719250857e-06,
"loss": 0.5937,
"step": 283
},
{
"epoch": 0.11426272379802857,
"grad_norm": 4.28046131366056,
"learning_rate": 4.997150602043371e-06,
"loss": 0.6211,
"step": 284
},
{
"epoch": 0.11466505733252867,
"grad_norm": 4.837073628599811,
"learning_rate": 4.996980559765504e-06,
"loss": 0.5828,
"step": 285
},
{
"epoch": 0.11506739086702877,
"grad_norm": 4.889286615585772,
"learning_rate": 4.996805592752626e-06,
"loss": 0.7301,
"step": 286
},
{
"epoch": 0.11546972440152886,
"grad_norm": 3.885336153971243,
"learning_rate": 4.996625701349822e-06,
"loss": 0.5091,
"step": 287
},
{
"epoch": 0.11587205793602896,
"grad_norm": 4.738737155243486,
"learning_rate": 4.996440885911886e-06,
"loss": 0.5519,
"step": 288
},
{
"epoch": 0.11627439147052906,
"grad_norm": 4.600896924297193,
"learning_rate": 4.996251146803326e-06,
"loss": 0.5357,
"step": 289
},
{
"epoch": 0.11667672500502917,
"grad_norm": 4.348185410245171,
"learning_rate": 4.996056484398359e-06,
"loss": 0.606,
"step": 290
},
{
"epoch": 0.11707905853952927,
"grad_norm": 4.609344088685658,
"learning_rate": 4.995856899080914e-06,
"loss": 0.7079,
"step": 291
},
{
"epoch": 0.11748139207402937,
"grad_norm": 4.982538761307109,
"learning_rate": 4.9956523912446274e-06,
"loss": 0.6626,
"step": 292
},
{
"epoch": 0.11788372560852947,
"grad_norm": 4.958292330122082,
"learning_rate": 4.9954429612928455e-06,
"loss": 0.5589,
"step": 293
},
{
"epoch": 0.11828605914302957,
"grad_norm": 5.3063992889234015,
"learning_rate": 4.995228609638623e-06,
"loss": 0.6761,
"step": 294
},
{
"epoch": 0.11868839267752967,
"grad_norm": 4.787491132159558,
"learning_rate": 4.995009336704718e-06,
"loss": 0.6907,
"step": 295
},
{
"epoch": 0.11909072621202978,
"grad_norm": 5.042532894966667,
"learning_rate": 4.994785142923599e-06,
"loss": 0.7259,
"step": 296
},
{
"epoch": 0.11949305974652988,
"grad_norm": 5.24654264942669,
"learning_rate": 4.994556028737439e-06,
"loss": 0.7376,
"step": 297
},
{
"epoch": 0.11989539328102998,
"grad_norm": 4.7381264053196475,
"learning_rate": 4.9943219945981126e-06,
"loss": 0.6419,
"step": 298
},
{
"epoch": 0.12029772681553007,
"grad_norm": 4.174895491105691,
"learning_rate": 4.9940830409672e-06,
"loss": 0.5846,
"step": 299
},
{
"epoch": 0.12070006035003017,
"grad_norm": 5.265350639824991,
"learning_rate": 4.993839168315985e-06,
"loss": 0.7269,
"step": 300
},
{
"epoch": 0.12110239388453027,
"grad_norm": 5.293670654130999,
"learning_rate": 4.99359037712545e-06,
"loss": 0.7539,
"step": 301
},
{
"epoch": 0.12150472741903037,
"grad_norm": 4.832726818218027,
"learning_rate": 4.993336667886281e-06,
"loss": 0.6617,
"step": 302
},
{
"epoch": 0.12190706095353047,
"grad_norm": 4.982756198499471,
"learning_rate": 4.993078041098862e-06,
"loss": 0.6517,
"step": 303
},
{
"epoch": 0.12230939448803058,
"grad_norm": 4.787146336142149,
"learning_rate": 4.9928144972732785e-06,
"loss": 0.5943,
"step": 304
},
{
"epoch": 0.12271172802253068,
"grad_norm": 5.725699716607812,
"learning_rate": 4.992546036929309e-06,
"loss": 0.7385,
"step": 305
},
{
"epoch": 0.12311406155703078,
"grad_norm": 4.388246603675133,
"learning_rate": 4.992272660596432e-06,
"loss": 0.5268,
"step": 306
},
{
"epoch": 0.12351639509153088,
"grad_norm": 4.323516644235707,
"learning_rate": 4.991994368813823e-06,
"loss": 0.5069,
"step": 307
},
{
"epoch": 0.12391872862603098,
"grad_norm": 4.968139714274168,
"learning_rate": 4.991711162130347e-06,
"loss": 0.6089,
"step": 308
},
{
"epoch": 0.12432106216053108,
"grad_norm": 5.393689160165955,
"learning_rate": 4.99142304110457e-06,
"loss": 0.5749,
"step": 309
},
{
"epoch": 0.12472339569503119,
"grad_norm": 5.010889684729688,
"learning_rate": 4.991130006304742e-06,
"loss": 0.6868,
"step": 310
},
{
"epoch": 0.12512572922953127,
"grad_norm": 4.77584804740267,
"learning_rate": 4.99083205830881e-06,
"loss": 0.7153,
"step": 311
},
{
"epoch": 0.1255280627640314,
"grad_norm": 3.8671259573453343,
"learning_rate": 4.990529197704411e-06,
"loss": 0.4938,
"step": 312
},
{
"epoch": 0.12593039629853148,
"grad_norm": 4.878545701273808,
"learning_rate": 4.990221425088868e-06,
"loss": 0.5954,
"step": 313
},
{
"epoch": 0.1263327298330316,
"grad_norm": 4.481594062259529,
"learning_rate": 4.989908741069195e-06,
"loss": 0.6397,
"step": 314
},
{
"epoch": 0.12673506336753168,
"grad_norm": 4.317626894827104,
"learning_rate": 4.98959114626209e-06,
"loss": 0.609,
"step": 315
},
{
"epoch": 0.1271373969020318,
"grad_norm": 4.415809282581258,
"learning_rate": 4.989268641293939e-06,
"loss": 0.5092,
"step": 316
},
{
"epoch": 0.12753973043653188,
"grad_norm": 4.4424914565481375,
"learning_rate": 4.9889412268008096e-06,
"loss": 0.6316,
"step": 317
},
{
"epoch": 0.12794206397103197,
"grad_norm": 4.431277408902586,
"learning_rate": 4.988608903428454e-06,
"loss": 0.6015,
"step": 318
},
{
"epoch": 0.1283443975055321,
"grad_norm": 4.618861534462686,
"learning_rate": 4.988271671832305e-06,
"loss": 0.5562,
"step": 319
},
{
"epoch": 0.12874673104003218,
"grad_norm": 5.078731601531793,
"learning_rate": 4.987929532677478e-06,
"loss": 0.7308,
"step": 320
},
{
"epoch": 0.1291490645745323,
"grad_norm": 4.814095780713649,
"learning_rate": 4.987582486638763e-06,
"loss": 0.6856,
"step": 321
},
{
"epoch": 0.12955139810903238,
"grad_norm": 5.171953766801144,
"learning_rate": 4.987230534400634e-06,
"loss": 0.7413,
"step": 322
},
{
"epoch": 0.1299537316435325,
"grad_norm": 4.527962518284751,
"learning_rate": 4.986873676657237e-06,
"loss": 0.5227,
"step": 323
},
{
"epoch": 0.13035606517803258,
"grad_norm": 4.9194193843235405,
"learning_rate": 4.986511914112392e-06,
"loss": 0.6038,
"step": 324
},
{
"epoch": 0.1307583987125327,
"grad_norm": 4.403379209157847,
"learning_rate": 4.986145247479597e-06,
"loss": 0.6624,
"step": 325
},
{
"epoch": 0.13116073224703279,
"grad_norm": 4.784759384702467,
"learning_rate": 4.985773677482018e-06,
"loss": 0.7109,
"step": 326
},
{
"epoch": 0.1315630657815329,
"grad_norm": 4.928746115672455,
"learning_rate": 4.985397204852496e-06,
"loss": 0.6241,
"step": 327
},
{
"epoch": 0.131965399316033,
"grad_norm": 5.6685753103245755,
"learning_rate": 4.985015830333536e-06,
"loss": 0.7121,
"step": 328
},
{
"epoch": 0.1323677328505331,
"grad_norm": 4.723811520101061,
"learning_rate": 4.984629554677316e-06,
"loss": 0.5683,
"step": 329
},
{
"epoch": 0.1327700663850332,
"grad_norm": 4.106210405752034,
"learning_rate": 4.984238378645677e-06,
"loss": 0.6128,
"step": 330
},
{
"epoch": 0.13317239991953328,
"grad_norm": 5.439987884236508,
"learning_rate": 4.983842303010127e-06,
"loss": 0.7294,
"step": 331
},
{
"epoch": 0.1335747334540334,
"grad_norm": 4.76281391948892,
"learning_rate": 4.9834413285518365e-06,
"loss": 0.6116,
"step": 332
},
{
"epoch": 0.13397706698853348,
"grad_norm": 4.594103270680001,
"learning_rate": 4.983035456061637e-06,
"loss": 0.5509,
"step": 333
},
{
"epoch": 0.1343794005230336,
"grad_norm": 4.8229418474364785,
"learning_rate": 4.982624686340022e-06,
"loss": 0.6565,
"step": 334
},
{
"epoch": 0.1347817340575337,
"grad_norm": 4.572125613909981,
"learning_rate": 4.9822090201971414e-06,
"loss": 0.5739,
"step": 335
},
{
"epoch": 0.1351840675920338,
"grad_norm": 4.657722588184461,
"learning_rate": 4.981788458452806e-06,
"loss": 0.716,
"step": 336
},
{
"epoch": 0.1355864011265339,
"grad_norm": 4.485714819813409,
"learning_rate": 4.9813630019364765e-06,
"loss": 0.5809,
"step": 337
},
{
"epoch": 0.135988734661034,
"grad_norm": 4.730520981461843,
"learning_rate": 4.9809326514872735e-06,
"loss": 0.6006,
"step": 338
},
{
"epoch": 0.1363910681955341,
"grad_norm": 4.407751381335739,
"learning_rate": 4.980497407953966e-06,
"loss": 0.6979,
"step": 339
},
{
"epoch": 0.1367934017300342,
"grad_norm": 5.036488553382197,
"learning_rate": 4.980057272194973e-06,
"loss": 0.6719,
"step": 340
},
{
"epoch": 0.1371957352645343,
"grad_norm": 4.131020183485464,
"learning_rate": 4.979612245078366e-06,
"loss": 0.656,
"step": 341
},
{
"epoch": 0.1375980687990344,
"grad_norm": 4.494385344326802,
"learning_rate": 4.979162327481859e-06,
"loss": 0.5824,
"step": 342
},
{
"epoch": 0.1380004023335345,
"grad_norm": 4.887767115382313,
"learning_rate": 4.9787075202928155e-06,
"loss": 0.7065,
"step": 343
},
{
"epoch": 0.1384027358680346,
"grad_norm": 4.929080555596624,
"learning_rate": 4.97824782440824e-06,
"loss": 0.6462,
"step": 344
},
{
"epoch": 0.1388050694025347,
"grad_norm": 4.768200287000246,
"learning_rate": 4.97778324073478e-06,
"loss": 0.7052,
"step": 345
},
{
"epoch": 0.1392074029370348,
"grad_norm": 4.824866676754677,
"learning_rate": 4.977313770188723e-06,
"loss": 0.6169,
"step": 346
},
{
"epoch": 0.1396097364715349,
"grad_norm": 5.0730067342381515,
"learning_rate": 4.976839413695994e-06,
"loss": 0.6907,
"step": 347
},
{
"epoch": 0.140012070006035,
"grad_norm": 4.400048281159567,
"learning_rate": 4.976360172192156e-06,
"loss": 0.6843,
"step": 348
},
{
"epoch": 0.1404144035405351,
"grad_norm": 4.826940942258961,
"learning_rate": 4.975876046622404e-06,
"loss": 0.6425,
"step": 349
},
{
"epoch": 0.1408167370750352,
"grad_norm": 4.1980516292894645,
"learning_rate": 4.975387037941568e-06,
"loss": 0.513,
"step": 350
},
{
"epoch": 0.1412190706095353,
"grad_norm": 4.599176289887721,
"learning_rate": 4.974893147114108e-06,
"loss": 0.6138,
"step": 351
},
{
"epoch": 0.1416214041440354,
"grad_norm": 5.224094683564219,
"learning_rate": 4.974394375114113e-06,
"loss": 0.7688,
"step": 352
},
{
"epoch": 0.14202373767853552,
"grad_norm": 4.886795665659055,
"learning_rate": 4.9738907229253e-06,
"loss": 0.6447,
"step": 353
},
{
"epoch": 0.1424260712130356,
"grad_norm": 4.686162997153853,
"learning_rate": 4.973382191541008e-06,
"loss": 0.6489,
"step": 354
},
{
"epoch": 0.14282840474753572,
"grad_norm": 4.633889069385792,
"learning_rate": 4.972868781964204e-06,
"loss": 0.6841,
"step": 355
},
{
"epoch": 0.1432307382820358,
"grad_norm": 4.714123893258144,
"learning_rate": 4.972350495207472e-06,
"loss": 0.7282,
"step": 356
},
{
"epoch": 0.1436330718165359,
"grad_norm": 4.620909154154889,
"learning_rate": 4.971827332293017e-06,
"loss": 0.6267,
"step": 357
},
{
"epoch": 0.144035405351036,
"grad_norm": 5.112805770554144,
"learning_rate": 4.971299294252659e-06,
"loss": 0.7187,
"step": 358
},
{
"epoch": 0.1444377388855361,
"grad_norm": 4.491917871078975,
"learning_rate": 4.970766382127838e-06,
"loss": 0.6243,
"step": 359
},
{
"epoch": 0.14484007242003621,
"grad_norm": 4.91569774829319,
"learning_rate": 4.9702285969696006e-06,
"loss": 0.6044,
"step": 360
},
{
"epoch": 0.1452424059545363,
"grad_norm": 4.347791813311231,
"learning_rate": 4.96968593983861e-06,
"loss": 0.6496,
"step": 361
},
{
"epoch": 0.14564473948903642,
"grad_norm": 4.6534811500518645,
"learning_rate": 4.9691384118051346e-06,
"loss": 0.7414,
"step": 362
},
{
"epoch": 0.1460470730235365,
"grad_norm": 4.5918592667861455,
"learning_rate": 4.968586013949051e-06,
"loss": 0.5802,
"step": 363
},
{
"epoch": 0.14644940655803662,
"grad_norm": 4.435482175102624,
"learning_rate": 4.96802874735984e-06,
"loss": 0.5162,
"step": 364
},
{
"epoch": 0.1468517400925367,
"grad_norm": 4.407001515831722,
"learning_rate": 4.967466613136586e-06,
"loss": 0.5361,
"step": 365
},
{
"epoch": 0.14725407362703682,
"grad_norm": 4.369142207445966,
"learning_rate": 4.966899612387972e-06,
"loss": 0.664,
"step": 366
},
{
"epoch": 0.1476564071615369,
"grad_norm": 4.607767758241303,
"learning_rate": 4.966327746232281e-06,
"loss": 0.5882,
"step": 367
},
{
"epoch": 0.14805874069603703,
"grad_norm": 4.938226519089643,
"learning_rate": 4.96575101579739e-06,
"loss": 0.5841,
"step": 368
},
{
"epoch": 0.14846107423053712,
"grad_norm": 4.096479564143721,
"learning_rate": 4.965169422220771e-06,
"loss": 0.5691,
"step": 369
},
{
"epoch": 0.1488634077650372,
"grad_norm": 4.775093512712085,
"learning_rate": 4.964582966649488e-06,
"loss": 0.6063,
"step": 370
},
{
"epoch": 0.14926574129953732,
"grad_norm": 4.211658294897708,
"learning_rate": 4.963991650240192e-06,
"loss": 0.5215,
"step": 371
},
{
"epoch": 0.1496680748340374,
"grad_norm": 4.364777540143213,
"learning_rate": 4.963395474159122e-06,
"loss": 0.4866,
"step": 372
},
{
"epoch": 0.15007040836853752,
"grad_norm": 4.681087632312734,
"learning_rate": 4.962794439582102e-06,
"loss": 0.6122,
"step": 373
},
{
"epoch": 0.1504727419030376,
"grad_norm": 5.077820479708522,
"learning_rate": 4.9621885476945405e-06,
"loss": 0.7555,
"step": 374
},
{
"epoch": 0.15087507543753773,
"grad_norm": 5.978600203961224,
"learning_rate": 4.961577799691421e-06,
"loss": 0.7113,
"step": 375
},
{
"epoch": 0.1512774089720378,
"grad_norm": 5.376960173895862,
"learning_rate": 4.960962196777307e-06,
"loss": 0.577,
"step": 376
},
{
"epoch": 0.15167974250653793,
"grad_norm": 5.18055770132837,
"learning_rate": 4.960341740166338e-06,
"loss": 0.6625,
"step": 377
},
{
"epoch": 0.15208207604103802,
"grad_norm": 4.460009104960063,
"learning_rate": 4.959716431082227e-06,
"loss": 0.5306,
"step": 378
},
{
"epoch": 0.15248440957553813,
"grad_norm": 4.303191111673615,
"learning_rate": 4.959086270758255e-06,
"loss": 0.5604,
"step": 379
},
{
"epoch": 0.15288674311003822,
"grad_norm": 4.8364861459301185,
"learning_rate": 4.9584512604372704e-06,
"loss": 0.6154,
"step": 380
},
{
"epoch": 0.15328907664453834,
"grad_norm": 4.638041078223287,
"learning_rate": 4.957811401371692e-06,
"loss": 0.5852,
"step": 381
},
{
"epoch": 0.15369141017903842,
"grad_norm": 4.8155087041398295,
"learning_rate": 4.9571666948234975e-06,
"loss": 0.591,
"step": 382
},
{
"epoch": 0.1540937437135385,
"grad_norm": 4.365724907978,
"learning_rate": 4.956517142064226e-06,
"loss": 0.5645,
"step": 383
},
{
"epoch": 0.15449607724803863,
"grad_norm": 4.800750875035706,
"learning_rate": 4.955862744374974e-06,
"loss": 0.6958,
"step": 384
},
{
"epoch": 0.15489841078253871,
"grad_norm": 4.218913109898912,
"learning_rate": 4.9552035030463955e-06,
"loss": 0.5981,
"step": 385
},
{
"epoch": 0.15530074431703883,
"grad_norm": 3.8608056472996775,
"learning_rate": 4.954539419378695e-06,
"loss": 0.5516,
"step": 386
},
{
"epoch": 0.15570307785153892,
"grad_norm": 4.422539247772956,
"learning_rate": 4.953870494681631e-06,
"loss": 0.466,
"step": 387
},
{
"epoch": 0.15610541138603903,
"grad_norm": 5.389694964679017,
"learning_rate": 4.9531967302745065e-06,
"loss": 0.7187,
"step": 388
},
{
"epoch": 0.15650774492053912,
"grad_norm": 4.641539036731872,
"learning_rate": 4.952518127486171e-06,
"loss": 0.6118,
"step": 389
},
{
"epoch": 0.15691007845503924,
"grad_norm": 4.38792439545412,
"learning_rate": 4.951834687655016e-06,
"loss": 0.6125,
"step": 390
},
{
"epoch": 0.15731241198953932,
"grad_norm": 4.859276712096591,
"learning_rate": 4.9511464121289745e-06,
"loss": 0.789,
"step": 391
},
{
"epoch": 0.15771474552403944,
"grad_norm": 4.123781252290954,
"learning_rate": 4.950453302265516e-06,
"loss": 0.5189,
"step": 392
},
{
"epoch": 0.15811707905853953,
"grad_norm": 4.5665024165035915,
"learning_rate": 4.949755359431643e-06,
"loss": 0.5843,
"step": 393
},
{
"epoch": 0.15851941259303964,
"grad_norm": 4.554644389766188,
"learning_rate": 4.949052585003892e-06,
"loss": 0.5165,
"step": 394
},
{
"epoch": 0.15892174612753973,
"grad_norm": 4.652445776343059,
"learning_rate": 4.9483449803683295e-06,
"loss": 0.6496,
"step": 395
},
{
"epoch": 0.15932407966203982,
"grad_norm": 4.155340040403773,
"learning_rate": 4.947632546920545e-06,
"loss": 0.5823,
"step": 396
},
{
"epoch": 0.15972641319653993,
"grad_norm": 5.685373741993176,
"learning_rate": 4.946915286065656e-06,
"loss": 0.7257,
"step": 397
},
{
"epoch": 0.16012874673104002,
"grad_norm": 3.794107938623536,
"learning_rate": 4.946193199218298e-06,
"loss": 0.5539,
"step": 398
},
{
"epoch": 0.16053108026554014,
"grad_norm": 3.345825002647911,
"learning_rate": 4.945466287802625e-06,
"loss": 0.4522,
"step": 399
},
{
"epoch": 0.16093341380004023,
"grad_norm": 4.499068362412442,
"learning_rate": 4.944734553252308e-06,
"loss": 0.7294,
"step": 400
},
{
"epoch": 0.16133574733454034,
"grad_norm": 4.710035321209727,
"learning_rate": 4.94399799701053e-06,
"loss": 0.6495,
"step": 401
},
{
"epoch": 0.16173808086904043,
"grad_norm": 4.665060195725978,
"learning_rate": 4.943256620529983e-06,
"loss": 0.5549,
"step": 402
},
{
"epoch": 0.16214041440354054,
"grad_norm": 4.526568640183799,
"learning_rate": 4.942510425272864e-06,
"loss": 0.5342,
"step": 403
},
{
"epoch": 0.16254274793804063,
"grad_norm": 4.500912340574114,
"learning_rate": 4.941759412710878e-06,
"loss": 0.5154,
"step": 404
},
{
"epoch": 0.16294508147254075,
"grad_norm": 4.871395743184042,
"learning_rate": 4.941003584325229e-06,
"loss": 0.7273,
"step": 405
},
{
"epoch": 0.16334741500704084,
"grad_norm": 5.226443625458406,
"learning_rate": 4.940242941606619e-06,
"loss": 0.5695,
"step": 406
},
{
"epoch": 0.16374974854154092,
"grad_norm": 4.3424277680190215,
"learning_rate": 4.939477486055245e-06,
"loss": 0.6903,
"step": 407
},
{
"epoch": 0.16415208207604104,
"grad_norm": 4.63055317361875,
"learning_rate": 4.938707219180797e-06,
"loss": 0.7724,
"step": 408
},
{
"epoch": 0.16455441561054113,
"grad_norm": 4.632025607743007,
"learning_rate": 4.9379321425024545e-06,
"loss": 0.5579,
"step": 409
},
{
"epoch": 0.16495674914504124,
"grad_norm": 4.949408477476733,
"learning_rate": 4.9371522575488814e-06,
"loss": 0.5585,
"step": 410
},
{
"epoch": 0.16535908267954133,
"grad_norm": 4.82833151193623,
"learning_rate": 4.936367565858228e-06,
"loss": 0.6281,
"step": 411
},
{
"epoch": 0.16576141621404145,
"grad_norm": 4.135170899195267,
"learning_rate": 4.935578068978121e-06,
"loss": 0.5236,
"step": 412
},
{
"epoch": 0.16616374974854153,
"grad_norm": 4.848865820547502,
"learning_rate": 4.9347837684656675e-06,
"loss": 0.7572,
"step": 413
},
{
"epoch": 0.16656608328304165,
"grad_norm": 4.717705787081838,
"learning_rate": 4.933984665887447e-06,
"loss": 0.7282,
"step": 414
},
{
"epoch": 0.16696841681754174,
"grad_norm": 4.528215697412672,
"learning_rate": 4.93318076281951e-06,
"loss": 0.5974,
"step": 415
},
{
"epoch": 0.16737075035204185,
"grad_norm": 5.165943035201723,
"learning_rate": 4.9323720608473754e-06,
"loss": 0.6622,
"step": 416
},
{
"epoch": 0.16777308388654194,
"grad_norm": 4.244718678666535,
"learning_rate": 4.9315585615660275e-06,
"loss": 0.5827,
"step": 417
},
{
"epoch": 0.16817541742104206,
"grad_norm": 5.152052375848677,
"learning_rate": 4.930740266579911e-06,
"loss": 0.6625,
"step": 418
},
{
"epoch": 0.16857775095554214,
"grad_norm": 4.55196067766976,
"learning_rate": 4.92991717750293e-06,
"loss": 0.5734,
"step": 419
},
{
"epoch": 0.16898008449004223,
"grad_norm": 5.130317188995626,
"learning_rate": 4.929089295958442e-06,
"loss": 0.6448,
"step": 420
},
{
"epoch": 0.16938241802454235,
"grad_norm": 4.221455100939359,
"learning_rate": 4.928256623579259e-06,
"loss": 0.5736,
"step": 421
},
{
"epoch": 0.16978475155904243,
"grad_norm": 5.042585074077865,
"learning_rate": 4.9274191620076405e-06,
"loss": 0.6948,
"step": 422
},
{
"epoch": 0.17018708509354255,
"grad_norm": 4.475716088239217,
"learning_rate": 4.926576912895292e-06,
"loss": 0.5762,
"step": 423
},
{
"epoch": 0.17058941862804264,
"grad_norm": 4.629681308629434,
"learning_rate": 4.9257298779033615e-06,
"loss": 0.7584,
"step": 424
},
{
"epoch": 0.17099175216254275,
"grad_norm": 4.624861162695507,
"learning_rate": 4.924878058702436e-06,
"loss": 0.6725,
"step": 425
},
{
"epoch": 0.17139408569704284,
"grad_norm": 4.729048209294549,
"learning_rate": 4.9240214569725375e-06,
"loss": 0.7722,
"step": 426
},
{
"epoch": 0.17179641923154296,
"grad_norm": 4.21954665313362,
"learning_rate": 4.923160074403122e-06,
"loss": 0.6253,
"step": 427
},
{
"epoch": 0.17219875276604305,
"grad_norm": 3.755869045635417,
"learning_rate": 4.922293912693072e-06,
"loss": 0.4626,
"step": 428
},
{
"epoch": 0.17260108630054316,
"grad_norm": 4.263780577574945,
"learning_rate": 4.921422973550699e-06,
"loss": 0.695,
"step": 429
},
{
"epoch": 0.17300341983504325,
"grad_norm": 5.108523767367576,
"learning_rate": 4.920547258693735e-06,
"loss": 0.7092,
"step": 430
},
{
"epoch": 0.17340575336954336,
"grad_norm": 4.317384116136861,
"learning_rate": 4.919666769849332e-06,
"loss": 0.4564,
"step": 431
},
{
"epoch": 0.17380808690404345,
"grad_norm": 5.268931042465572,
"learning_rate": 4.918781508754057e-06,
"loss": 0.7064,
"step": 432
},
{
"epoch": 0.17421042043854354,
"grad_norm": 4.295948103968784,
"learning_rate": 4.9178914771538875e-06,
"loss": 0.6892,
"step": 433
},
{
"epoch": 0.17461275397304366,
"grad_norm": 4.214776194153066,
"learning_rate": 4.916996676804214e-06,
"loss": 0.5877,
"step": 434
},
{
"epoch": 0.17501508750754374,
"grad_norm": 4.187051473349894,
"learning_rate": 4.916097109469829e-06,
"loss": 0.5542,
"step": 435
},
{
"epoch": 0.17541742104204386,
"grad_norm": 4.7370251117595235,
"learning_rate": 4.915192776924926e-06,
"loss": 0.6131,
"step": 436
},
{
"epoch": 0.17581975457654395,
"grad_norm": 4.790572516699059,
"learning_rate": 4.9142836809531e-06,
"loss": 0.6847,
"step": 437
},
{
"epoch": 0.17622208811104406,
"grad_norm": 4.024460804862858,
"learning_rate": 4.91336982334734e-06,
"loss": 0.4475,
"step": 438
},
{
"epoch": 0.17662442164554415,
"grad_norm": 4.675530435693111,
"learning_rate": 4.912451205910024e-06,
"loss": 0.641,
"step": 439
},
{
"epoch": 0.17702675518004427,
"grad_norm": 4.070435859439258,
"learning_rate": 4.91152783045292e-06,
"loss": 0.5715,
"step": 440
},
{
"epoch": 0.17742908871454435,
"grad_norm": 5.014878172443035,
"learning_rate": 4.910599698797179e-06,
"loss": 0.6227,
"step": 441
},
{
"epoch": 0.17783142224904447,
"grad_norm": 4.471540054012383,
"learning_rate": 4.909666812773333e-06,
"loss": 0.5145,
"step": 442
},
{
"epoch": 0.17823375578354456,
"grad_norm": 4.442414191534445,
"learning_rate": 4.908729174221289e-06,
"loss": 0.6688,
"step": 443
},
{
"epoch": 0.17863608931804467,
"grad_norm": 4.624231827213014,
"learning_rate": 4.9077867849903325e-06,
"loss": 0.6327,
"step": 444
},
{
"epoch": 0.17903842285254476,
"grad_norm": 4.743489707184984,
"learning_rate": 4.906839646939113e-06,
"loss": 0.7075,
"step": 445
},
{
"epoch": 0.17944075638704485,
"grad_norm": 4.200250704275894,
"learning_rate": 4.905887761935649e-06,
"loss": 0.5171,
"step": 446
},
{
"epoch": 0.17984308992154496,
"grad_norm": 4.0039914177118545,
"learning_rate": 4.90493113185732e-06,
"loss": 0.5334,
"step": 447
},
{
"epoch": 0.18024542345604505,
"grad_norm": 4.437438466205263,
"learning_rate": 4.903969758590865e-06,
"loss": 0.584,
"step": 448
},
{
"epoch": 0.18064775699054517,
"grad_norm": 4.668639275979892,
"learning_rate": 4.9030036440323766e-06,
"loss": 0.669,
"step": 449
},
{
"epoch": 0.18105009052504525,
"grad_norm": 4.262713251256841,
"learning_rate": 4.902032790087301e-06,
"loss": 0.518,
"step": 450
},
{
"epoch": 0.18145242405954537,
"grad_norm": 4.58562990715973,
"learning_rate": 4.9010571986704295e-06,
"loss": 0.5627,
"step": 451
},
{
"epoch": 0.18185475759404546,
"grad_norm": 4.2511574164879615,
"learning_rate": 4.900076871705897e-06,
"loss": 0.6189,
"step": 452
},
{
"epoch": 0.18225709112854557,
"grad_norm": 3.879096119034437,
"learning_rate": 4.89909181112718e-06,
"loss": 0.4688,
"step": 453
},
{
"epoch": 0.18265942466304566,
"grad_norm": 4.4698856213971645,
"learning_rate": 4.898102018877088e-06,
"loss": 0.6272,
"step": 454
},
{
"epoch": 0.18306175819754578,
"grad_norm": 4.25349966912043,
"learning_rate": 4.897107496907767e-06,
"loss": 0.4844,
"step": 455
},
{
"epoch": 0.18346409173204586,
"grad_norm": 5.180942020863308,
"learning_rate": 4.896108247180688e-06,
"loss": 0.7363,
"step": 456
},
{
"epoch": 0.18386642526654598,
"grad_norm": 4.464543366415725,
"learning_rate": 4.895104271666647e-06,
"loss": 0.507,
"step": 457
},
{
"epoch": 0.18426875880104607,
"grad_norm": 4.35656943725317,
"learning_rate": 4.8940955723457604e-06,
"loss": 0.5626,
"step": 458
},
{
"epoch": 0.18467109233554616,
"grad_norm": 4.46990880672421,
"learning_rate": 4.893082151207464e-06,
"loss": 0.5986,
"step": 459
},
{
"epoch": 0.18507342587004627,
"grad_norm": 4.315333956337568,
"learning_rate": 4.8920640102505005e-06,
"loss": 0.5399,
"step": 460
},
{
"epoch": 0.18547575940454636,
"grad_norm": 4.414262914713829,
"learning_rate": 4.891041151482928e-06,
"loss": 0.6866,
"step": 461
},
{
"epoch": 0.18587809293904647,
"grad_norm": 4.68767919716049,
"learning_rate": 4.8900135769221045e-06,
"loss": 0.6205,
"step": 462
},
{
"epoch": 0.18628042647354656,
"grad_norm": 4.55495109723687,
"learning_rate": 4.888981288594692e-06,
"loss": 0.6232,
"step": 463
},
{
"epoch": 0.18668276000804668,
"grad_norm": 5.072112114859536,
"learning_rate": 4.887944288536648e-06,
"loss": 0.6246,
"step": 464
},
{
"epoch": 0.18708509354254677,
"grad_norm": 4.598620261784645,
"learning_rate": 4.886902578793221e-06,
"loss": 0.6688,
"step": 465
},
{
"epoch": 0.18748742707704688,
"grad_norm": 4.156182800147685,
"learning_rate": 4.885856161418953e-06,
"loss": 0.6185,
"step": 466
},
{
"epoch": 0.18788976061154697,
"grad_norm": 5.123876789008787,
"learning_rate": 4.8848050384776665e-06,
"loss": 0.6485,
"step": 467
},
{
"epoch": 0.18829209414604708,
"grad_norm": 4.7877986741742555,
"learning_rate": 4.883749212042466e-06,
"loss": 0.6336,
"step": 468
},
{
"epoch": 0.18869442768054717,
"grad_norm": 4.170455407715718,
"learning_rate": 4.882688684195732e-06,
"loss": 0.5708,
"step": 469
},
{
"epoch": 0.1890967612150473,
"grad_norm": 4.394557070936156,
"learning_rate": 4.881623457029121e-06,
"loss": 0.6584,
"step": 470
},
{
"epoch": 0.18949909474954738,
"grad_norm": 4.4866621521500205,
"learning_rate": 4.880553532643552e-06,
"loss": 0.5816,
"step": 471
},
{
"epoch": 0.18990142828404746,
"grad_norm": 4.263089971509693,
"learning_rate": 4.879478913149212e-06,
"loss": 0.6563,
"step": 472
},
{
"epoch": 0.19030376181854758,
"grad_norm": 3.897184614846541,
"learning_rate": 4.8783996006655485e-06,
"loss": 0.5816,
"step": 473
},
{
"epoch": 0.19070609535304767,
"grad_norm": 4.759719975025391,
"learning_rate": 4.877315597321263e-06,
"loss": 0.7219,
"step": 474
},
{
"epoch": 0.19110842888754778,
"grad_norm": 4.561889671156211,
"learning_rate": 4.876226905254309e-06,
"loss": 0.6307,
"step": 475
},
{
"epoch": 0.19151076242204787,
"grad_norm": 4.622879218120913,
"learning_rate": 4.875133526611888e-06,
"loss": 0.5739,
"step": 476
},
{
"epoch": 0.19191309595654799,
"grad_norm": 4.183836330008031,
"learning_rate": 4.874035463550445e-06,
"loss": 0.4165,
"step": 477
},
{
"epoch": 0.19231542949104807,
"grad_norm": 4.431780943086593,
"learning_rate": 4.872932718235663e-06,
"loss": 0.5768,
"step": 478
},
{
"epoch": 0.1927177630255482,
"grad_norm": 4.700772397419956,
"learning_rate": 4.87182529284246e-06,
"loss": 0.6635,
"step": 479
},
{
"epoch": 0.19312009656004828,
"grad_norm": 4.0384190687435915,
"learning_rate": 4.870713189554985e-06,
"loss": 0.4994,
"step": 480
},
{
"epoch": 0.1935224300945484,
"grad_norm": 4.16066989214431,
"learning_rate": 4.869596410566614e-06,
"loss": 0.523,
"step": 481
},
{
"epoch": 0.19392476362904848,
"grad_norm": 3.7758221004395565,
"learning_rate": 4.8684749580799405e-06,
"loss": 0.5106,
"step": 482
},
{
"epoch": 0.1943270971635486,
"grad_norm": 4.792256321316127,
"learning_rate": 4.867348834306781e-06,
"loss": 0.5955,
"step": 483
},
{
"epoch": 0.19472943069804868,
"grad_norm": 4.408528934091589,
"learning_rate": 4.866218041468161e-06,
"loss": 0.6032,
"step": 484
},
{
"epoch": 0.19513176423254877,
"grad_norm": 4.425894822386934,
"learning_rate": 4.865082581794317e-06,
"loss": 0.5298,
"step": 485
},
{
"epoch": 0.1955340977670489,
"grad_norm": 4.356368057153947,
"learning_rate": 4.863942457524689e-06,
"loss": 0.5381,
"step": 486
},
{
"epoch": 0.19593643130154897,
"grad_norm": 3.990737447917161,
"learning_rate": 4.862797670907915e-06,
"loss": 0.5049,
"step": 487
},
{
"epoch": 0.1963387648360491,
"grad_norm": 4.229237889406895,
"learning_rate": 4.8616482242018325e-06,
"loss": 0.5467,
"step": 488
},
{
"epoch": 0.19674109837054918,
"grad_norm": 4.126506367371427,
"learning_rate": 4.860494119673466e-06,
"loss": 0.5455,
"step": 489
},
{
"epoch": 0.1971434319050493,
"grad_norm": 5.231977983505604,
"learning_rate": 4.859335359599031e-06,
"loss": 0.7528,
"step": 490
},
{
"epoch": 0.19754576543954938,
"grad_norm": 5.1291616835145115,
"learning_rate": 4.85817194626392e-06,
"loss": 0.679,
"step": 491
},
{
"epoch": 0.1979480989740495,
"grad_norm": 3.737681461647685,
"learning_rate": 4.857003881962705e-06,
"loss": 0.4926,
"step": 492
},
{
"epoch": 0.19835043250854958,
"grad_norm": 4.703925829656442,
"learning_rate": 4.855831168999132e-06,
"loss": 0.6439,
"step": 493
},
{
"epoch": 0.1987527660430497,
"grad_norm": 4.627856862526262,
"learning_rate": 4.854653809686115e-06,
"loss": 0.6698,
"step": 494
},
{
"epoch": 0.1991550995775498,
"grad_norm": 4.364999176797817,
"learning_rate": 4.853471806345732e-06,
"loss": 0.6015,
"step": 495
},
{
"epoch": 0.19955743311204988,
"grad_norm": 4.602684651491329,
"learning_rate": 4.852285161309218e-06,
"loss": 0.5809,
"step": 496
},
{
"epoch": 0.19995976664655,
"grad_norm": 4.325677574896615,
"learning_rate": 4.851093876916967e-06,
"loss": 0.526,
"step": 497
},
{
"epoch": 0.20036210018105008,
"grad_norm": 4.81776828160967,
"learning_rate": 4.849897955518518e-06,
"loss": 0.6654,
"step": 498
},
{
"epoch": 0.2007644337155502,
"grad_norm": 5.49681992539213,
"learning_rate": 4.848697399472561e-06,
"loss": 0.8468,
"step": 499
},
{
"epoch": 0.20116676725005028,
"grad_norm": 4.782602085058219,
"learning_rate": 4.8474922111469225e-06,
"loss": 0.6339,
"step": 500
},
{
"epoch": 0.2015691007845504,
"grad_norm": 4.705465090280715,
"learning_rate": 4.846282392918566e-06,
"loss": 0.5111,
"step": 501
},
{
"epoch": 0.20197143431905049,
"grad_norm": 4.044280178788859,
"learning_rate": 4.845067947173589e-06,
"loss": 0.4453,
"step": 502
},
{
"epoch": 0.2023737678535506,
"grad_norm": 5.299453770363048,
"learning_rate": 4.843848876307211e-06,
"loss": 0.7287,
"step": 503
},
{
"epoch": 0.2027761013880507,
"grad_norm": 3.9060416143674,
"learning_rate": 4.842625182723779e-06,
"loss": 0.4903,
"step": 504
},
{
"epoch": 0.2031784349225508,
"grad_norm": 4.320966869452432,
"learning_rate": 4.841396868836753e-06,
"loss": 0.5312,
"step": 505
},
{
"epoch": 0.2035807684570509,
"grad_norm": 4.059590568863906,
"learning_rate": 4.840163937068707e-06,
"loss": 0.5953,
"step": 506
},
{
"epoch": 0.203983101991551,
"grad_norm": 4.544560777009811,
"learning_rate": 4.838926389851324e-06,
"loss": 0.5274,
"step": 507
},
{
"epoch": 0.2043854355260511,
"grad_norm": 4.643239965625325,
"learning_rate": 4.837684229625389e-06,
"loss": 0.6284,
"step": 508
},
{
"epoch": 0.20478776906055118,
"grad_norm": 3.848260216211789,
"learning_rate": 4.836437458840783e-06,
"loss": 0.5052,
"step": 509
},
{
"epoch": 0.2051901025950513,
"grad_norm": 4.65542258301039,
"learning_rate": 4.835186079956483e-06,
"loss": 0.5641,
"step": 510
},
{
"epoch": 0.2055924361295514,
"grad_norm": 3.7048624001601738,
"learning_rate": 4.8339300954405545e-06,
"loss": 0.5507,
"step": 511
},
{
"epoch": 0.2059947696640515,
"grad_norm": 4.721724963544143,
"learning_rate": 4.832669507770144e-06,
"loss": 0.6945,
"step": 512
},
{
"epoch": 0.2063971031985516,
"grad_norm": 3.9937965024761892,
"learning_rate": 4.83140431943148e-06,
"loss": 0.5444,
"step": 513
},
{
"epoch": 0.2067994367330517,
"grad_norm": 4.150283320138307,
"learning_rate": 4.830134532919863e-06,
"loss": 0.5835,
"step": 514
},
{
"epoch": 0.2072017702675518,
"grad_norm": 5.2365671694488904,
"learning_rate": 4.828860150739662e-06,
"loss": 0.697,
"step": 515
},
{
"epoch": 0.2076041038020519,
"grad_norm": 3.98085660862504,
"learning_rate": 4.827581175404311e-06,
"loss": 0.4872,
"step": 516
},
{
"epoch": 0.208006437336552,
"grad_norm": 4.574435055815776,
"learning_rate": 4.8262976094363016e-06,
"loss": 0.5925,
"step": 517
},
{
"epoch": 0.2084087708710521,
"grad_norm": 4.459834553682444,
"learning_rate": 4.825009455367181e-06,
"loss": 0.6426,
"step": 518
},
{
"epoch": 0.2088111044055522,
"grad_norm": 3.8557681901771037,
"learning_rate": 4.823716715737544e-06,
"loss": 0.5719,
"step": 519
},
{
"epoch": 0.20921343794005232,
"grad_norm": 4.776729631302629,
"learning_rate": 4.82241939309703e-06,
"loss": 0.6118,
"step": 520
},
{
"epoch": 0.2096157714745524,
"grad_norm": 4.701717061474551,
"learning_rate": 4.821117490004319e-06,
"loss": 0.4766,
"step": 521
},
{
"epoch": 0.2100181050090525,
"grad_norm": 4.486472816066567,
"learning_rate": 4.819811009027122e-06,
"loss": 0.6134,
"step": 522
},
{
"epoch": 0.2104204385435526,
"grad_norm": 4.784573013198364,
"learning_rate": 4.818499952742179e-06,
"loss": 0.5982,
"step": 523
},
{
"epoch": 0.2108227720780527,
"grad_norm": 4.953599282635229,
"learning_rate": 4.817184323735258e-06,
"loss": 0.5732,
"step": 524
},
{
"epoch": 0.2112251056125528,
"grad_norm": 4.808483380881421,
"learning_rate": 4.815864124601139e-06,
"loss": 0.6771,
"step": 525
},
{
"epoch": 0.2116274391470529,
"grad_norm": 4.024625165836511,
"learning_rate": 4.814539357943622e-06,
"loss": 0.6355,
"step": 526
},
{
"epoch": 0.212029772681553,
"grad_norm": 3.943315026550887,
"learning_rate": 4.813210026375513e-06,
"loss": 0.521,
"step": 527
},
{
"epoch": 0.2124321062160531,
"grad_norm": 3.463894710975418,
"learning_rate": 4.811876132518618e-06,
"loss": 0.4542,
"step": 528
},
{
"epoch": 0.21283443975055322,
"grad_norm": 4.513349087640872,
"learning_rate": 4.810537679003746e-06,
"loss": 0.6093,
"step": 529
},
{
"epoch": 0.2132367732850533,
"grad_norm": 4.649205790356446,
"learning_rate": 4.8091946684706956e-06,
"loss": 0.6726,
"step": 530
},
{
"epoch": 0.21363910681955342,
"grad_norm": 4.2037442337203315,
"learning_rate": 4.807847103568255e-06,
"loss": 0.6265,
"step": 531
},
{
"epoch": 0.2140414403540535,
"grad_norm": 4.893988522032171,
"learning_rate": 4.806494986954195e-06,
"loss": 0.7373,
"step": 532
},
{
"epoch": 0.21444377388855362,
"grad_norm": 4.455715486715626,
"learning_rate": 4.805138321295262e-06,
"loss": 0.5457,
"step": 533
},
{
"epoch": 0.2148461074230537,
"grad_norm": 4.649691607815862,
"learning_rate": 4.803777109267174e-06,
"loss": 0.5872,
"step": 534
},
{
"epoch": 0.2152484409575538,
"grad_norm": 4.606972921739361,
"learning_rate": 4.802411353554619e-06,
"loss": 0.6556,
"step": 535
},
{
"epoch": 0.21565077449205391,
"grad_norm": 4.262679120822924,
"learning_rate": 4.801041056851244e-06,
"loss": 0.5814,
"step": 536
},
{
"epoch": 0.216053108026554,
"grad_norm": 4.734408703763602,
"learning_rate": 4.7996662218596505e-06,
"loss": 0.5975,
"step": 537
},
{
"epoch": 0.21645544156105412,
"grad_norm": 4.0966775674750515,
"learning_rate": 4.798286851291395e-06,
"loss": 0.5373,
"step": 538
},
{
"epoch": 0.2168577750955542,
"grad_norm": 4.794723705583196,
"learning_rate": 4.796902947866976e-06,
"loss": 0.6411,
"step": 539
},
{
"epoch": 0.21726010863005432,
"grad_norm": 5.177706885118282,
"learning_rate": 4.795514514315833e-06,
"loss": 0.7129,
"step": 540
},
{
"epoch": 0.2176624421645544,
"grad_norm": 4.235334038988407,
"learning_rate": 4.794121553376341e-06,
"loss": 0.6368,
"step": 541
},
{
"epoch": 0.21806477569905452,
"grad_norm": 4.250039615427904,
"learning_rate": 4.792724067795802e-06,
"loss": 0.5897,
"step": 542
},
{
"epoch": 0.2184671092335546,
"grad_norm": 4.248095551159398,
"learning_rate": 4.7913220603304455e-06,
"loss": 0.6119,
"step": 543
},
{
"epoch": 0.21886944276805473,
"grad_norm": 4.7784528939759126,
"learning_rate": 4.789915533745415e-06,
"loss": 0.6428,
"step": 544
},
{
"epoch": 0.21927177630255482,
"grad_norm": 3.7433210513797857,
"learning_rate": 4.788504490814771e-06,
"loss": 0.4616,
"step": 545
},
{
"epoch": 0.21967410983705493,
"grad_norm": 4.477510756273104,
"learning_rate": 4.78708893432148e-06,
"loss": 0.5564,
"step": 546
},
{
"epoch": 0.22007644337155502,
"grad_norm": 4.159189423897508,
"learning_rate": 4.785668867057408e-06,
"loss": 0.5005,
"step": 547
},
{
"epoch": 0.2204787769060551,
"grad_norm": 4.801923291900065,
"learning_rate": 4.784244291823321e-06,
"loss": 0.6703,
"step": 548
},
{
"epoch": 0.22088111044055522,
"grad_norm": 4.9705885588356065,
"learning_rate": 4.782815211428875e-06,
"loss": 0.6892,
"step": 549
},
{
"epoch": 0.2212834439750553,
"grad_norm": 4.689425225044057,
"learning_rate": 4.7813816286926116e-06,
"loss": 0.6438,
"step": 550
},
{
"epoch": 0.22168577750955543,
"grad_norm": 3.9964370496914836,
"learning_rate": 4.7799435464419496e-06,
"loss": 0.5283,
"step": 551
},
{
"epoch": 0.22208811104405551,
"grad_norm": 3.9063906383285523,
"learning_rate": 4.778500967513186e-06,
"loss": 0.5477,
"step": 552
},
{
"epoch": 0.22249044457855563,
"grad_norm": 4.972674347366186,
"learning_rate": 4.777053894751484e-06,
"loss": 0.6762,
"step": 553
},
{
"epoch": 0.22289277811305572,
"grad_norm": 4.605207302352324,
"learning_rate": 4.77560233101087e-06,
"loss": 0.5823,
"step": 554
},
{
"epoch": 0.22329511164755583,
"grad_norm": 5.023738846112498,
"learning_rate": 4.774146279154231e-06,
"loss": 0.645,
"step": 555
},
{
"epoch": 0.22369744518205592,
"grad_norm": 4.21478405630217,
"learning_rate": 4.772685742053299e-06,
"loss": 0.6361,
"step": 556
},
{
"epoch": 0.22409977871655604,
"grad_norm": 4.443333059344436,
"learning_rate": 4.7712207225886605e-06,
"loss": 0.5808,
"step": 557
},
{
"epoch": 0.22450211225105612,
"grad_norm": 4.136335273179504,
"learning_rate": 4.7697512236497355e-06,
"loss": 0.4543,
"step": 558
},
{
"epoch": 0.22490444578555624,
"grad_norm": 5.228702785570883,
"learning_rate": 4.7682772481347835e-06,
"loss": 0.6677,
"step": 559
},
{
"epoch": 0.22530677932005633,
"grad_norm": 4.607023738968648,
"learning_rate": 4.76679879895089e-06,
"loss": 0.6972,
"step": 560
},
{
"epoch": 0.22570911285455642,
"grad_norm": 3.9623979182055646,
"learning_rate": 4.7653158790139655e-06,
"loss": 0.562,
"step": 561
},
{
"epoch": 0.22611144638905653,
"grad_norm": 4.785540664618103,
"learning_rate": 4.763828491248737e-06,
"loss": 0.6667,
"step": 562
},
{
"epoch": 0.22651377992355662,
"grad_norm": 4.626975611158776,
"learning_rate": 4.762336638588745e-06,
"loss": 0.5222,
"step": 563
},
{
"epoch": 0.22691611345805673,
"grad_norm": 4.148716543735861,
"learning_rate": 4.760840323976333e-06,
"loss": 0.5792,
"step": 564
},
{
"epoch": 0.22731844699255682,
"grad_norm": 3.7576481851339785,
"learning_rate": 4.759339550362647e-06,
"loss": 0.4171,
"step": 565
},
{
"epoch": 0.22772078052705694,
"grad_norm": 4.056873981914918,
"learning_rate": 4.757834320707629e-06,
"loss": 0.5823,
"step": 566
},
{
"epoch": 0.22812311406155703,
"grad_norm": 4.485956120607346,
"learning_rate": 4.7563246379800036e-06,
"loss": 0.6038,
"step": 567
},
{
"epoch": 0.22852544759605714,
"grad_norm": 4.095512060377993,
"learning_rate": 4.754810505157285e-06,
"loss": 0.5807,
"step": 568
},
{
"epoch": 0.22892778113055723,
"grad_norm": 4.316692633895088,
"learning_rate": 4.753291925225761e-06,
"loss": 0.5689,
"step": 569
},
{
"epoch": 0.22933011466505734,
"grad_norm": 4.0964431049127095,
"learning_rate": 4.751768901180488e-06,
"loss": 0.4724,
"step": 570
},
{
"epoch": 0.22973244819955743,
"grad_norm": 3.8505086953071683,
"learning_rate": 4.750241436025292e-06,
"loss": 0.5599,
"step": 571
},
{
"epoch": 0.23013478173405755,
"grad_norm": 4.97127913813276,
"learning_rate": 4.748709532772756e-06,
"loss": 0.7189,
"step": 572
},
{
"epoch": 0.23053711526855764,
"grad_norm": 4.376792824386442,
"learning_rate": 4.7471731944442154e-06,
"loss": 0.6149,
"step": 573
},
{
"epoch": 0.23093944880305772,
"grad_norm": 4.633754764886852,
"learning_rate": 4.745632424069755e-06,
"loss": 0.6118,
"step": 574
},
{
"epoch": 0.23134178233755784,
"grad_norm": 4.682951429282118,
"learning_rate": 4.744087224688197e-06,
"loss": 0.6089,
"step": 575
},
{
"epoch": 0.23174411587205793,
"grad_norm": 4.504821455509545,
"learning_rate": 4.742537599347101e-06,
"loss": 0.5747,
"step": 576
},
{
"epoch": 0.23214644940655804,
"grad_norm": 4.922801626757808,
"learning_rate": 4.740983551102759e-06,
"loss": 0.7267,
"step": 577
},
{
"epoch": 0.23254878294105813,
"grad_norm": 4.128104128218885,
"learning_rate": 4.73942508302018e-06,
"loss": 0.5344,
"step": 578
},
{
"epoch": 0.23295111647555825,
"grad_norm": 4.502862129743137,
"learning_rate": 4.7378621981730966e-06,
"loss": 0.4465,
"step": 579
},
{
"epoch": 0.23335345001005833,
"grad_norm": 4.476573377559871,
"learning_rate": 4.736294899643946e-06,
"loss": 0.5327,
"step": 580
},
{
"epoch": 0.23375578354455845,
"grad_norm": 4.525722183672701,
"learning_rate": 4.734723190523875e-06,
"loss": 0.6182,
"step": 581
},
{
"epoch": 0.23415811707905854,
"grad_norm": 4.4174724143850455,
"learning_rate": 4.7331470739127284e-06,
"loss": 0.582,
"step": 582
},
{
"epoch": 0.23456045061355865,
"grad_norm": 4.087588153474757,
"learning_rate": 4.731566552919042e-06,
"loss": 0.5626,
"step": 583
},
{
"epoch": 0.23496278414805874,
"grad_norm": 4.793168855698109,
"learning_rate": 4.7299816306600395e-06,
"loss": 0.712,
"step": 584
},
{
"epoch": 0.23536511768255883,
"grad_norm": 4.088862969423477,
"learning_rate": 4.728392310261628e-06,
"loss": 0.5867,
"step": 585
},
{
"epoch": 0.23576745121705894,
"grad_norm": 4.31442875545057,
"learning_rate": 4.726798594858382e-06,
"loss": 0.5255,
"step": 586
},
{
"epoch": 0.23616978475155903,
"grad_norm": 4.2694537336500495,
"learning_rate": 4.7252004875935506e-06,
"loss": 0.6822,
"step": 587
},
{
"epoch": 0.23657211828605915,
"grad_norm": 4.179846752454285,
"learning_rate": 4.723597991619043e-06,
"loss": 0.5433,
"step": 588
},
{
"epoch": 0.23697445182055923,
"grad_norm": 4.1764050880208,
"learning_rate": 4.721991110095422e-06,
"loss": 0.586,
"step": 589
},
{
"epoch": 0.23737678535505935,
"grad_norm": 3.699033328030597,
"learning_rate": 4.720379846191903e-06,
"loss": 0.4942,
"step": 590
},
{
"epoch": 0.23777911888955944,
"grad_norm": 4.798015125682991,
"learning_rate": 4.718764203086342e-06,
"loss": 0.7269,
"step": 591
},
{
"epoch": 0.23818145242405955,
"grad_norm": 4.974938507339555,
"learning_rate": 4.717144183965233e-06,
"loss": 0.6753,
"step": 592
},
{
"epoch": 0.23858378595855964,
"grad_norm": 5.08563513422606,
"learning_rate": 4.715519792023703e-06,
"loss": 0.6613,
"step": 593
},
{
"epoch": 0.23898611949305976,
"grad_norm": 4.6691045707182655,
"learning_rate": 4.7138910304655e-06,
"loss": 0.5635,
"step": 594
},
{
"epoch": 0.23938845302755984,
"grad_norm": 4.366016787794075,
"learning_rate": 4.712257902502992e-06,
"loss": 0.6603,
"step": 595
},
{
"epoch": 0.23979078656205996,
"grad_norm": 4.98238528037948,
"learning_rate": 4.710620411357157e-06,
"loss": 0.6241,
"step": 596
},
{
"epoch": 0.24019312009656005,
"grad_norm": 4.023463566077988,
"learning_rate": 4.7089785602575805e-06,
"loss": 0.5202,
"step": 597
},
{
"epoch": 0.24059545363106014,
"grad_norm": 4.040098710840788,
"learning_rate": 4.707332352442446e-06,
"loss": 0.5843,
"step": 598
},
{
"epoch": 0.24099778716556025,
"grad_norm": 4.151014624312927,
"learning_rate": 4.705681791158531e-06,
"loss": 0.6206,
"step": 599
},
{
"epoch": 0.24140012070006034,
"grad_norm": 4.0875213661566665,
"learning_rate": 4.704026879661196e-06,
"loss": 0.5614,
"step": 600
},
{
"epoch": 0.24180245423456045,
"grad_norm": 4.553877858262055,
"learning_rate": 4.7023676212143834e-06,
"loss": 0.6649,
"step": 601
},
{
"epoch": 0.24220478776906054,
"grad_norm": 4.674650194014434,
"learning_rate": 4.70070401909061e-06,
"loss": 0.606,
"step": 602
},
{
"epoch": 0.24260712130356066,
"grad_norm": 4.63353917939836,
"learning_rate": 4.699036076570959e-06,
"loss": 0.6668,
"step": 603
},
{
"epoch": 0.24300945483806075,
"grad_norm": 4.723870743698028,
"learning_rate": 4.697363796945072e-06,
"loss": 0.6728,
"step": 604
},
{
"epoch": 0.24341178837256086,
"grad_norm": 4.386624516452244,
"learning_rate": 4.6956871835111475e-06,
"loss": 0.5139,
"step": 605
},
{
"epoch": 0.24381412190706095,
"grad_norm": 4.244611479893601,
"learning_rate": 4.694006239575929e-06,
"loss": 0.609,
"step": 606
},
{
"epoch": 0.24421645544156106,
"grad_norm": 4.072589096296677,
"learning_rate": 4.692320968454702e-06,
"loss": 0.5657,
"step": 607
},
{
"epoch": 0.24461878897606115,
"grad_norm": 4.88910796894103,
"learning_rate": 4.690631373471287e-06,
"loss": 0.701,
"step": 608
},
{
"epoch": 0.24502112251056127,
"grad_norm": 4.686738823395271,
"learning_rate": 4.6889374579580315e-06,
"loss": 0.637,
"step": 609
},
{
"epoch": 0.24542345604506136,
"grad_norm": 4.905698840133413,
"learning_rate": 4.687239225255805e-06,
"loss": 0.6003,
"step": 610
},
{
"epoch": 0.24582578957956144,
"grad_norm": 4.338740241210239,
"learning_rate": 4.68553667871399e-06,
"loss": 0.5906,
"step": 611
},
{
"epoch": 0.24622812311406156,
"grad_norm": 4.119688719772878,
"learning_rate": 4.68382982169048e-06,
"loss": 0.4954,
"step": 612
},
{
"epoch": 0.24663045664856165,
"grad_norm": 3.96691281411716,
"learning_rate": 4.6821186575516665e-06,
"loss": 0.6126,
"step": 613
},
{
"epoch": 0.24703279018306176,
"grad_norm": 4.760926273551105,
"learning_rate": 4.680403189672439e-06,
"loss": 0.6613,
"step": 614
},
{
"epoch": 0.24743512371756185,
"grad_norm": 4.4384316824035075,
"learning_rate": 4.678683421436173e-06,
"loss": 0.5687,
"step": 615
},
{
"epoch": 0.24783745725206197,
"grad_norm": 3.931664070509481,
"learning_rate": 4.676959356234726e-06,
"loss": 0.5857,
"step": 616
},
{
"epoch": 0.24823979078656205,
"grad_norm": 5.119336252733733,
"learning_rate": 4.6752309974684315e-06,
"loss": 0.6312,
"step": 617
},
{
"epoch": 0.24864212432106217,
"grad_norm": 4.229793646214086,
"learning_rate": 4.6734983485460895e-06,
"loss": 0.5093,
"step": 618
},
{
"epoch": 0.24904445785556226,
"grad_norm": 4.9794376894703944,
"learning_rate": 4.671761412884962e-06,
"loss": 0.7092,
"step": 619
},
{
"epoch": 0.24944679139006237,
"grad_norm": 4.36204998657697,
"learning_rate": 4.670020193910766e-06,
"loss": 0.559,
"step": 620
},
{
"epoch": 0.24984912492456246,
"grad_norm": 4.297208174823204,
"learning_rate": 4.668274695057666e-06,
"loss": 0.4965,
"step": 621
},
{
"epoch": 0.25025145845906255,
"grad_norm": 4.235510447453734,
"learning_rate": 4.666524919768267e-06,
"loss": 0.5434,
"step": 622
},
{
"epoch": 0.25065379199356264,
"grad_norm": 4.297045599708538,
"learning_rate": 4.66477087149361e-06,
"loss": 0.5559,
"step": 623
},
{
"epoch": 0.2510561255280628,
"grad_norm": 5.638228122643803,
"learning_rate": 4.663012553693161e-06,
"loss": 0.6393,
"step": 624
},
{
"epoch": 0.25145845906256287,
"grad_norm": 4.307369262356312,
"learning_rate": 4.661249969834809e-06,
"loss": 0.5516,
"step": 625
},
{
"epoch": 0.25186079259706295,
"grad_norm": 4.348008968073508,
"learning_rate": 4.659483123394855e-06,
"loss": 0.5858,
"step": 626
},
{
"epoch": 0.25226312613156304,
"grad_norm": 4.245775793874246,
"learning_rate": 4.657712017858011e-06,
"loss": 0.5985,
"step": 627
},
{
"epoch": 0.2526654596660632,
"grad_norm": 4.27241102613026,
"learning_rate": 4.6559366567173824e-06,
"loss": 0.5247,
"step": 628
},
{
"epoch": 0.2530677932005633,
"grad_norm": 3.9143192748019384,
"learning_rate": 4.6541570434744735e-06,
"loss": 0.4936,
"step": 629
},
{
"epoch": 0.25347012673506336,
"grad_norm": 4.550429412487192,
"learning_rate": 4.6523731816391725e-06,
"loss": 0.7499,
"step": 630
},
{
"epoch": 0.25387246026956345,
"grad_norm": 4.202697728155226,
"learning_rate": 4.650585074729747e-06,
"loss": 0.5184,
"step": 631
},
{
"epoch": 0.2542747938040636,
"grad_norm": 4.083559252407811,
"learning_rate": 4.648792726272838e-06,
"loss": 0.5242,
"step": 632
},
{
"epoch": 0.2546771273385637,
"grad_norm": 3.9861255665329267,
"learning_rate": 4.646996139803452e-06,
"loss": 0.5269,
"step": 633
},
{
"epoch": 0.25507946087306377,
"grad_norm": 4.559490860669659,
"learning_rate": 4.645195318864951e-06,
"loss": 0.6372,
"step": 634
},
{
"epoch": 0.25548179440756386,
"grad_norm": 4.394457869705605,
"learning_rate": 4.643390267009054e-06,
"loss": 0.5105,
"step": 635
},
{
"epoch": 0.25588412794206394,
"grad_norm": 5.018376167864149,
"learning_rate": 4.641580987795821e-06,
"loss": 0.6628,
"step": 636
},
{
"epoch": 0.2562864614765641,
"grad_norm": 3.7893948824739465,
"learning_rate": 4.639767484793648e-06,
"loss": 0.508,
"step": 637
},
{
"epoch": 0.2566887950110642,
"grad_norm": 4.596240270523842,
"learning_rate": 4.637949761579266e-06,
"loss": 0.6529,
"step": 638
},
{
"epoch": 0.25709112854556426,
"grad_norm": 5.140583964722019,
"learning_rate": 4.636127821737726e-06,
"loss": 0.6489,
"step": 639
},
{
"epoch": 0.25749346208006435,
"grad_norm": 3.8394758085160823,
"learning_rate": 4.634301668862397e-06,
"loss": 0.4736,
"step": 640
},
{
"epoch": 0.2578957956145645,
"grad_norm": 4.38356346084736,
"learning_rate": 4.632471306554955e-06,
"loss": 0.5353,
"step": 641
},
{
"epoch": 0.2582981291490646,
"grad_norm": 3.8889360472386185,
"learning_rate": 4.630636738425381e-06,
"loss": 0.5511,
"step": 642
},
{
"epoch": 0.25870046268356467,
"grad_norm": 4.817925477512709,
"learning_rate": 4.62879796809195e-06,
"loss": 0.6963,
"step": 643
},
{
"epoch": 0.25910279621806476,
"grad_norm": 4.742191877421797,
"learning_rate": 4.626954999181224e-06,
"loss": 0.7752,
"step": 644
},
{
"epoch": 0.2595051297525649,
"grad_norm": 4.245951623762571,
"learning_rate": 4.625107835328048e-06,
"loss": 0.5754,
"step": 645
},
{
"epoch": 0.259907463287065,
"grad_norm": 4.836575633325676,
"learning_rate": 4.623256480175537e-06,
"loss": 0.6639,
"step": 646
},
{
"epoch": 0.2603097968215651,
"grad_norm": 3.806176284035789,
"learning_rate": 4.621400937375075e-06,
"loss": 0.5292,
"step": 647
},
{
"epoch": 0.26071213035606516,
"grad_norm": 4.025115976057339,
"learning_rate": 4.619541210586307e-06,
"loss": 0.4815,
"step": 648
},
{
"epoch": 0.26111446389056525,
"grad_norm": 4.1347392050793825,
"learning_rate": 4.617677303477124e-06,
"loss": 0.5681,
"step": 649
},
{
"epoch": 0.2615167974250654,
"grad_norm": 4.371554358928095,
"learning_rate": 4.615809219723667e-06,
"loss": 0.592,
"step": 650
},
{
"epoch": 0.2619191309595655,
"grad_norm": 4.976927834726334,
"learning_rate": 4.613936963010314e-06,
"loss": 0.7898,
"step": 651
},
{
"epoch": 0.26232146449406557,
"grad_norm": 4.80195477309382,
"learning_rate": 4.612060537029671e-06,
"loss": 0.6565,
"step": 652
},
{
"epoch": 0.26272379802856566,
"grad_norm": 4.2203256747490805,
"learning_rate": 4.610179945482568e-06,
"loss": 0.5597,
"step": 653
},
{
"epoch": 0.2631261315630658,
"grad_norm": 4.3119037789418035,
"learning_rate": 4.608295192078051e-06,
"loss": 0.5773,
"step": 654
},
{
"epoch": 0.2635284650975659,
"grad_norm": 3.938767904065124,
"learning_rate": 4.606406280533373e-06,
"loss": 0.4691,
"step": 655
},
{
"epoch": 0.263930798632066,
"grad_norm": 3.873313542196617,
"learning_rate": 4.6045132145739914e-06,
"loss": 0.5446,
"step": 656
},
{
"epoch": 0.26433313216656606,
"grad_norm": 4.720535373828625,
"learning_rate": 4.602615997933552e-06,
"loss": 0.6451,
"step": 657
},
{
"epoch": 0.2647354657010662,
"grad_norm": 5.319338434241136,
"learning_rate": 4.600714634353893e-06,
"loss": 0.6786,
"step": 658
},
{
"epoch": 0.2651377992355663,
"grad_norm": 3.986441538874652,
"learning_rate": 4.598809127585026e-06,
"loss": 0.5815,
"step": 659
},
{
"epoch": 0.2655401327700664,
"grad_norm": 4.190229942704305,
"learning_rate": 4.596899481385137e-06,
"loss": 0.5933,
"step": 660
},
{
"epoch": 0.26594246630456647,
"grad_norm": 4.211284711325467,
"learning_rate": 4.5949856995205745e-06,
"loss": 0.6062,
"step": 661
},
{
"epoch": 0.26634479983906656,
"grad_norm": 5.022372015405014,
"learning_rate": 4.593067785765846e-06,
"loss": 0.6632,
"step": 662
},
{
"epoch": 0.2667471333735667,
"grad_norm": 3.9365356585193148,
"learning_rate": 4.5911457439036075e-06,
"loss": 0.5324,
"step": 663
},
{
"epoch": 0.2671494669080668,
"grad_norm": 4.891303145874928,
"learning_rate": 4.589219577724654e-06,
"loss": 0.7504,
"step": 664
},
{
"epoch": 0.2675518004425669,
"grad_norm": 3.855940902006332,
"learning_rate": 4.5872892910279185e-06,
"loss": 0.5476,
"step": 665
},
{
"epoch": 0.26795413397706697,
"grad_norm": 4.37750980680622,
"learning_rate": 4.58535488762046e-06,
"loss": 0.5902,
"step": 666
},
{
"epoch": 0.2683564675115671,
"grad_norm": 4.450877429194335,
"learning_rate": 4.583416371317454e-06,
"loss": 0.6345,
"step": 667
},
{
"epoch": 0.2687588010460672,
"grad_norm": 4.162519879691083,
"learning_rate": 4.581473745942191e-06,
"loss": 0.444,
"step": 668
},
{
"epoch": 0.2691611345805673,
"grad_norm": 4.2236499578643025,
"learning_rate": 4.579527015326065e-06,
"loss": 0.492,
"step": 669
},
{
"epoch": 0.2695634681150674,
"grad_norm": 4.482147273633291,
"learning_rate": 4.5775761833085645e-06,
"loss": 0.5648,
"step": 670
},
{
"epoch": 0.2699658016495675,
"grad_norm": 4.760060597047131,
"learning_rate": 4.57562125373727e-06,
"loss": 0.6674,
"step": 671
},
{
"epoch": 0.2703681351840676,
"grad_norm": 4.362389446472747,
"learning_rate": 4.573662230467844e-06,
"loss": 0.577,
"step": 672
},
{
"epoch": 0.2707704687185677,
"grad_norm": 3.8159558241349987,
"learning_rate": 4.5716991173640165e-06,
"loss": 0.5185,
"step": 673
},
{
"epoch": 0.2711728022530678,
"grad_norm": 4.005051687655319,
"learning_rate": 4.5697319182975944e-06,
"loss": 0.5922,
"step": 674
},
{
"epoch": 0.27157513578756787,
"grad_norm": 4.667139467518304,
"learning_rate": 4.567760637148432e-06,
"loss": 0.6032,
"step": 675
},
{
"epoch": 0.271977469322068,
"grad_norm": 3.9606776473436405,
"learning_rate": 4.5657852778044435e-06,
"loss": 0.5174,
"step": 676
},
{
"epoch": 0.2723798028565681,
"grad_norm": 4.322780347285095,
"learning_rate": 4.5638058441615815e-06,
"loss": 0.6089,
"step": 677
},
{
"epoch": 0.2727821363910682,
"grad_norm": 4.433317399280547,
"learning_rate": 4.561822340123836e-06,
"loss": 0.6047,
"step": 678
},
{
"epoch": 0.2731844699255683,
"grad_norm": 4.552805703827102,
"learning_rate": 4.559834769603224e-06,
"loss": 0.6281,
"step": 679
},
{
"epoch": 0.2735868034600684,
"grad_norm": 5.128182512733822,
"learning_rate": 4.557843136519784e-06,
"loss": 0.6697,
"step": 680
},
{
"epoch": 0.2739891369945685,
"grad_norm": 4.063920951512369,
"learning_rate": 4.555847444801565e-06,
"loss": 0.5885,
"step": 681
},
{
"epoch": 0.2743914705290686,
"grad_norm": 4.908841427141714,
"learning_rate": 4.5538476983846245e-06,
"loss": 0.6771,
"step": 682
},
{
"epoch": 0.2747938040635687,
"grad_norm": 4.405724720337232,
"learning_rate": 4.551843901213012e-06,
"loss": 0.6479,
"step": 683
},
{
"epoch": 0.2751961375980688,
"grad_norm": 3.945129587052982,
"learning_rate": 4.549836057238769e-06,
"loss": 0.6039,
"step": 684
},
{
"epoch": 0.2755984711325689,
"grad_norm": 4.96376428039534,
"learning_rate": 4.547824170421921e-06,
"loss": 0.7191,
"step": 685
},
{
"epoch": 0.276000804667069,
"grad_norm": 4.378595779287015,
"learning_rate": 4.54580824473046e-06,
"loss": 0.5549,
"step": 686
},
{
"epoch": 0.2764031382015691,
"grad_norm": 4.262426880188478,
"learning_rate": 4.54378828414035e-06,
"loss": 0.5024,
"step": 687
},
{
"epoch": 0.2768054717360692,
"grad_norm": 4.426695285396216,
"learning_rate": 4.541764292635512e-06,
"loss": 0.7052,
"step": 688
},
{
"epoch": 0.2772078052705693,
"grad_norm": 4.473404410086587,
"learning_rate": 4.5397362742078145e-06,
"loss": 0.5557,
"step": 689
},
{
"epoch": 0.2776101388050694,
"grad_norm": 4.052844689460026,
"learning_rate": 4.537704232857069e-06,
"loss": 0.4541,
"step": 690
},
{
"epoch": 0.2780124723395695,
"grad_norm": 3.89050386632389,
"learning_rate": 4.5356681725910255e-06,
"loss": 0.549,
"step": 691
},
{
"epoch": 0.2784148058740696,
"grad_norm": 4.349538143222671,
"learning_rate": 4.5336280974253535e-06,
"loss": 0.6076,
"step": 692
},
{
"epoch": 0.2788171394085697,
"grad_norm": 4.361849160433631,
"learning_rate": 4.5315840113836454e-06,
"loss": 0.6524,
"step": 693
},
{
"epoch": 0.2792194729430698,
"grad_norm": 4.617234754063093,
"learning_rate": 4.529535918497403e-06,
"loss": 0.5616,
"step": 694
},
{
"epoch": 0.2796218064775699,
"grad_norm": 4.09023083083644,
"learning_rate": 4.5274838228060326e-06,
"loss": 0.5407,
"step": 695
},
{
"epoch": 0.28002414001207,
"grad_norm": 3.7157295921866718,
"learning_rate": 4.525427728356832e-06,
"loss": 0.5361,
"step": 696
},
{
"epoch": 0.28042647354657013,
"grad_norm": 4.457227461962157,
"learning_rate": 4.523367639204985e-06,
"loss": 0.5593,
"step": 697
},
{
"epoch": 0.2808288070810702,
"grad_norm": 4.348610370546784,
"learning_rate": 4.52130355941356e-06,
"loss": 0.5121,
"step": 698
},
{
"epoch": 0.2812311406155703,
"grad_norm": 3.9173059552199194,
"learning_rate": 4.519235493053491e-06,
"loss": 0.4927,
"step": 699
},
{
"epoch": 0.2816334741500704,
"grad_norm": 4.208431911956715,
"learning_rate": 4.517163444203575e-06,
"loss": 0.5504,
"step": 700
},
{
"epoch": 0.2820358076845705,
"grad_norm": 4.404532536673193,
"learning_rate": 4.515087416950464e-06,
"loss": 0.6057,
"step": 701
},
{
"epoch": 0.2824381412190706,
"grad_norm": 4.163697909062566,
"learning_rate": 4.513007415388659e-06,
"loss": 0.4682,
"step": 702
},
{
"epoch": 0.2828404747535707,
"grad_norm": 5.152927142247097,
"learning_rate": 4.510923443620494e-06,
"loss": 0.6601,
"step": 703
},
{
"epoch": 0.2832428082880708,
"grad_norm": 4.544361331270338,
"learning_rate": 4.50883550575614e-06,
"loss": 0.5968,
"step": 704
},
{
"epoch": 0.2836451418225709,
"grad_norm": 4.134394917706537,
"learning_rate": 4.5067436059135836e-06,
"loss": 0.6647,
"step": 705
},
{
"epoch": 0.28404747535707103,
"grad_norm": 4.734085611831448,
"learning_rate": 4.504647748218633e-06,
"loss": 0.601,
"step": 706
},
{
"epoch": 0.2844498088915711,
"grad_norm": 4.681267518422373,
"learning_rate": 4.502547936804894e-06,
"loss": 0.5539,
"step": 707
},
{
"epoch": 0.2848521424260712,
"grad_norm": 4.325088947732379,
"learning_rate": 4.500444175813776e-06,
"loss": 0.5872,
"step": 708
},
{
"epoch": 0.2852544759605713,
"grad_norm": 3.850534498836246,
"learning_rate": 4.4983364693944775e-06,
"loss": 0.5061,
"step": 709
},
{
"epoch": 0.28565680949507144,
"grad_norm": 4.373584617561936,
"learning_rate": 4.496224821703977e-06,
"loss": 0.6693,
"step": 710
},
{
"epoch": 0.2860591430295715,
"grad_norm": 4.377477636040013,
"learning_rate": 4.494109236907026e-06,
"loss": 0.5966,
"step": 711
},
{
"epoch": 0.2864614765640716,
"grad_norm": 4.271381874115667,
"learning_rate": 4.491989719176142e-06,
"loss": 0.6221,
"step": 712
},
{
"epoch": 0.2868638100985717,
"grad_norm": 3.7259627257148202,
"learning_rate": 4.489866272691599e-06,
"loss": 0.4313,
"step": 713
},
{
"epoch": 0.2872661436330718,
"grad_norm": 4.0336320558123155,
"learning_rate": 4.48773890164142e-06,
"loss": 0.5305,
"step": 714
},
{
"epoch": 0.28766847716757193,
"grad_norm": 4.585992205034134,
"learning_rate": 4.485607610221367e-06,
"loss": 0.7636,
"step": 715
},
{
"epoch": 0.288070810702072,
"grad_norm": 4.67820441036977,
"learning_rate": 4.4834724026349376e-06,
"loss": 0.5866,
"step": 716
},
{
"epoch": 0.2884731442365721,
"grad_norm": 5.012556667522281,
"learning_rate": 4.4813332830933484e-06,
"loss": 0.6498,
"step": 717
},
{
"epoch": 0.2888754777710722,
"grad_norm": 4.19440277086636,
"learning_rate": 4.4791902558155345e-06,
"loss": 0.5434,
"step": 718
},
{
"epoch": 0.28927781130557234,
"grad_norm": 4.330716534353544,
"learning_rate": 4.477043325028137e-06,
"loss": 0.5507,
"step": 719
},
{
"epoch": 0.28968014484007243,
"grad_norm": 4.097976140286037,
"learning_rate": 4.474892494965495e-06,
"loss": 0.4591,
"step": 720
},
{
"epoch": 0.2900824783745725,
"grad_norm": 4.132010732456553,
"learning_rate": 4.472737769869642e-06,
"loss": 0.5217,
"step": 721
},
{
"epoch": 0.2904848119090726,
"grad_norm": 3.7815121295908343,
"learning_rate": 4.470579153990288e-06,
"loss": 0.4932,
"step": 722
},
{
"epoch": 0.29088714544357275,
"grad_norm": 3.724905186650606,
"learning_rate": 4.468416651584822e-06,
"loss": 0.5448,
"step": 723
},
{
"epoch": 0.29128947897807284,
"grad_norm": 3.82443773314746,
"learning_rate": 4.4662502669182935e-06,
"loss": 0.5298,
"step": 724
},
{
"epoch": 0.2916918125125729,
"grad_norm": 4.343371987708737,
"learning_rate": 4.464080004263411e-06,
"loss": 0.606,
"step": 725
},
{
"epoch": 0.292094146047073,
"grad_norm": 4.009570342579401,
"learning_rate": 4.461905867900535e-06,
"loss": 0.5916,
"step": 726
},
{
"epoch": 0.2924964795815731,
"grad_norm": 4.148496022814315,
"learning_rate": 4.459727862117658e-06,
"loss": 0.5155,
"step": 727
},
{
"epoch": 0.29289881311607324,
"grad_norm": 4.568599931304069,
"learning_rate": 4.457545991210412e-06,
"loss": 0.555,
"step": 728
},
{
"epoch": 0.29330114665057333,
"grad_norm": 4.026645287125634,
"learning_rate": 4.455360259482047e-06,
"loss": 0.5265,
"step": 729
},
{
"epoch": 0.2937034801850734,
"grad_norm": 4.4466778657917265,
"learning_rate": 4.4531706712434305e-06,
"loss": 0.6318,
"step": 730
},
{
"epoch": 0.2941058137195735,
"grad_norm": 4.444852619715758,
"learning_rate": 4.450977230813035e-06,
"loss": 0.671,
"step": 731
},
{
"epoch": 0.29450814725407365,
"grad_norm": 5.136550532204147,
"learning_rate": 4.44877994251693e-06,
"loss": 0.611,
"step": 732
},
{
"epoch": 0.29491048078857374,
"grad_norm": 4.472272090985854,
"learning_rate": 4.446578810688774e-06,
"loss": 0.595,
"step": 733
},
{
"epoch": 0.2953128143230738,
"grad_norm": 4.755288406312762,
"learning_rate": 4.444373839669808e-06,
"loss": 0.801,
"step": 734
},
{
"epoch": 0.2957151478575739,
"grad_norm": 4.964978523542175,
"learning_rate": 4.442165033808843e-06,
"loss": 0.6712,
"step": 735
},
{
"epoch": 0.29611748139207406,
"grad_norm": 4.590633696444609,
"learning_rate": 4.439952397462254e-06,
"loss": 0.6214,
"step": 736
},
{
"epoch": 0.29651981492657414,
"grad_norm": 4.05753322012328,
"learning_rate": 4.43773593499397e-06,
"loss": 0.4477,
"step": 737
},
{
"epoch": 0.29692214846107423,
"grad_norm": 4.354711360779599,
"learning_rate": 4.435515650775468e-06,
"loss": 0.5618,
"step": 738
},
{
"epoch": 0.2973244819955743,
"grad_norm": 4.543128550216277,
"learning_rate": 4.433291549185761e-06,
"loss": 0.4807,
"step": 739
},
{
"epoch": 0.2977268155300744,
"grad_norm": 3.9444140759236737,
"learning_rate": 4.431063634611392e-06,
"loss": 0.5393,
"step": 740
},
{
"epoch": 0.29812914906457455,
"grad_norm": 4.516456335270593,
"learning_rate": 4.428831911446422e-06,
"loss": 0.5839,
"step": 741
},
{
"epoch": 0.29853148259907464,
"grad_norm": 4.4607761036583815,
"learning_rate": 4.426596384092426e-06,
"loss": 0.6538,
"step": 742
},
{
"epoch": 0.2989338161335747,
"grad_norm": 6.196021389574422,
"learning_rate": 4.424357056958483e-06,
"loss": 0.6484,
"step": 743
},
{
"epoch": 0.2993361496680748,
"grad_norm": 4.251042887568661,
"learning_rate": 4.422113934461161e-06,
"loss": 0.5523,
"step": 744
},
{
"epoch": 0.29973848320257496,
"grad_norm": 4.697763472346534,
"learning_rate": 4.4198670210245186e-06,
"loss": 0.6053,
"step": 745
},
{
"epoch": 0.30014081673707504,
"grad_norm": 4.665389980976664,
"learning_rate": 4.41761632108009e-06,
"loss": 0.6618,
"step": 746
},
{
"epoch": 0.30054315027157513,
"grad_norm": 4.2980125151698525,
"learning_rate": 4.415361839066874e-06,
"loss": 0.4863,
"step": 747
},
{
"epoch": 0.3009454838060752,
"grad_norm": 4.25422933800907,
"learning_rate": 4.413103579431335e-06,
"loss": 0.6285,
"step": 748
},
{
"epoch": 0.30134781734057536,
"grad_norm": 4.408647829220997,
"learning_rate": 4.410841546627383e-06,
"loss": 0.6214,
"step": 749
},
{
"epoch": 0.30175015087507545,
"grad_norm": 3.5964308822983626,
"learning_rate": 4.408575745116371e-06,
"loss": 0.4709,
"step": 750
},
{
"epoch": 0.30215248440957554,
"grad_norm": 4.177190728366978,
"learning_rate": 4.406306179367086e-06,
"loss": 0.6584,
"step": 751
},
{
"epoch": 0.3025548179440756,
"grad_norm": 3.9436659867851,
"learning_rate": 4.404032853855736e-06,
"loss": 0.5804,
"step": 752
},
{
"epoch": 0.3029571514785757,
"grad_norm": 4.5139405502647465,
"learning_rate": 4.40175577306595e-06,
"loss": 0.6131,
"step": 753
},
{
"epoch": 0.30335948501307586,
"grad_norm": 4.656972033870911,
"learning_rate": 4.399474941488756e-06,
"loss": 0.5378,
"step": 754
},
{
"epoch": 0.30376181854757595,
"grad_norm": 4.2754391925078785,
"learning_rate": 4.397190363622588e-06,
"loss": 0.5117,
"step": 755
},
{
"epoch": 0.30416415208207603,
"grad_norm": 4.057091286795055,
"learning_rate": 4.3949020439732594e-06,
"loss": 0.4969,
"step": 756
},
{
"epoch": 0.3045664856165761,
"grad_norm": 3.914038182134245,
"learning_rate": 4.392609987053972e-06,
"loss": 0.4484,
"step": 757
},
{
"epoch": 0.30496881915107626,
"grad_norm": 4.364733148673897,
"learning_rate": 4.390314197385292e-06,
"loss": 0.662,
"step": 758
},
{
"epoch": 0.30537115268557635,
"grad_norm": 4.866819672714999,
"learning_rate": 4.388014679495154e-06,
"loss": 0.6832,
"step": 759
},
{
"epoch": 0.30577348622007644,
"grad_norm": 4.563239673265389,
"learning_rate": 4.385711437918839e-06,
"loss": 0.633,
"step": 760
},
{
"epoch": 0.30617581975457653,
"grad_norm": 3.952293727810083,
"learning_rate": 4.383404477198976e-06,
"loss": 0.4319,
"step": 761
},
{
"epoch": 0.30657815328907667,
"grad_norm": 3.8982970475676866,
"learning_rate": 4.38109380188553e-06,
"loss": 0.5412,
"step": 762
},
{
"epoch": 0.30698048682357676,
"grad_norm": 4.767355411844817,
"learning_rate": 4.3787794165357875e-06,
"loss": 0.6688,
"step": 763
},
{
"epoch": 0.30738282035807685,
"grad_norm": 5.490594878274815,
"learning_rate": 4.3764613257143585e-06,
"loss": 0.8315,
"step": 764
},
{
"epoch": 0.30778515389257693,
"grad_norm": 4.4753569929038965,
"learning_rate": 4.374139533993157e-06,
"loss": 0.5577,
"step": 765
},
{
"epoch": 0.308187487427077,
"grad_norm": 4.234373998218113,
"learning_rate": 4.371814045951396e-06,
"loss": 0.6195,
"step": 766
},
{
"epoch": 0.30858982096157717,
"grad_norm": 3.8789979804624277,
"learning_rate": 4.369484866175581e-06,
"loss": 0.5812,
"step": 767
},
{
"epoch": 0.30899215449607725,
"grad_norm": 4.3774657256617635,
"learning_rate": 4.367151999259498e-06,
"loss": 0.5334,
"step": 768
},
{
"epoch": 0.30939448803057734,
"grad_norm": 4.413468965671534,
"learning_rate": 4.364815449804204e-06,
"loss": 0.6043,
"step": 769
},
{
"epoch": 0.30979682156507743,
"grad_norm": 4.977248443532888,
"learning_rate": 4.362475222418019e-06,
"loss": 0.7427,
"step": 770
},
{
"epoch": 0.3101991550995776,
"grad_norm": 4.449181903975735,
"learning_rate": 4.360131321716518e-06,
"loss": 0.5336,
"step": 771
},
{
"epoch": 0.31060148863407766,
"grad_norm": 4.203059807958007,
"learning_rate": 4.357783752322522e-06,
"loss": 0.5362,
"step": 772
},
{
"epoch": 0.31100382216857775,
"grad_norm": 3.8919997316821004,
"learning_rate": 4.355432518866084e-06,
"loss": 0.5202,
"step": 773
},
{
"epoch": 0.31140615570307784,
"grad_norm": 4.125644261476737,
"learning_rate": 4.353077625984484e-06,
"loss": 0.5136,
"step": 774
},
{
"epoch": 0.311808489237578,
"grad_norm": 3.7230845852814687,
"learning_rate": 4.3507190783222245e-06,
"loss": 0.4731,
"step": 775
},
{
"epoch": 0.31221082277207807,
"grad_norm": 4.619811557245368,
"learning_rate": 4.348356880531011e-06,
"loss": 0.698,
"step": 776
},
{
"epoch": 0.31261315630657815,
"grad_norm": 4.064918048601741,
"learning_rate": 4.345991037269748e-06,
"loss": 0.5594,
"step": 777
},
{
"epoch": 0.31301548984107824,
"grad_norm": 4.28030910141664,
"learning_rate": 4.343621553204533e-06,
"loss": 0.6272,
"step": 778
},
{
"epoch": 0.31341782337557833,
"grad_norm": 4.3485146730819,
"learning_rate": 4.341248433008645e-06,
"loss": 0.6078,
"step": 779
},
{
"epoch": 0.3138201569100785,
"grad_norm": 4.211683329853448,
"learning_rate": 4.338871681362528e-06,
"loss": 0.5247,
"step": 780
},
{
"epoch": 0.31422249044457856,
"grad_norm": 3.8752448583148453,
"learning_rate": 4.3364913029537946e-06,
"loss": 0.5274,
"step": 781
},
{
"epoch": 0.31462482397907865,
"grad_norm": 3.999853155987491,
"learning_rate": 4.334107302477208e-06,
"loss": 0.5421,
"step": 782
},
{
"epoch": 0.31502715751357874,
"grad_norm": 4.961501923709214,
"learning_rate": 4.331719684634676e-06,
"loss": 0.7151,
"step": 783
},
{
"epoch": 0.3154294910480789,
"grad_norm": 4.148580969978077,
"learning_rate": 4.3293284541352384e-06,
"loss": 0.5084,
"step": 784
},
{
"epoch": 0.31583182458257897,
"grad_norm": 3.623114769478068,
"learning_rate": 4.326933615695064e-06,
"loss": 0.4912,
"step": 785
},
{
"epoch": 0.31623415811707906,
"grad_norm": 4.179951168808134,
"learning_rate": 4.324535174037433e-06,
"loss": 0.4943,
"step": 786
},
{
"epoch": 0.31663649165157914,
"grad_norm": 4.206035486231768,
"learning_rate": 4.322133133892737e-06,
"loss": 0.6175,
"step": 787
},
{
"epoch": 0.3170388251860793,
"grad_norm": 4.029670988995613,
"learning_rate": 4.31972749999846e-06,
"loss": 0.5329,
"step": 788
},
{
"epoch": 0.3174411587205794,
"grad_norm": 4.024639871547482,
"learning_rate": 4.317318277099178e-06,
"loss": 0.5121,
"step": 789
},
{
"epoch": 0.31784349225507946,
"grad_norm": 4.04730827392964,
"learning_rate": 4.314905469946542e-06,
"loss": 0.4979,
"step": 790
},
{
"epoch": 0.31824582578957955,
"grad_norm": 4.440133410808506,
"learning_rate": 4.312489083299275e-06,
"loss": 0.6333,
"step": 791
},
{
"epoch": 0.31864815932407964,
"grad_norm": 4.079663513764819,
"learning_rate": 4.310069121923158e-06,
"loss": 0.5987,
"step": 792
},
{
"epoch": 0.3190504928585798,
"grad_norm": 4.17692958034715,
"learning_rate": 4.307645590591024e-06,
"loss": 0.4798,
"step": 793
},
{
"epoch": 0.31945282639307987,
"grad_norm": 4.417314019145798,
"learning_rate": 4.305218494082744e-06,
"loss": 0.4832,
"step": 794
},
{
"epoch": 0.31985515992757996,
"grad_norm": 4.920563992053252,
"learning_rate": 4.302787837185224e-06,
"loss": 0.6288,
"step": 795
},
{
"epoch": 0.32025749346208005,
"grad_norm": 4.067969775601857,
"learning_rate": 4.300353624692389e-06,
"loss": 0.52,
"step": 796
},
{
"epoch": 0.3206598269965802,
"grad_norm": 4.232889302389981,
"learning_rate": 4.2979158614051805e-06,
"loss": 0.6205,
"step": 797
},
{
"epoch": 0.3210621605310803,
"grad_norm": 4.335177482388414,
"learning_rate": 4.295474552131538e-06,
"loss": 0.6153,
"step": 798
},
{
"epoch": 0.32146449406558036,
"grad_norm": 4.698659170570413,
"learning_rate": 4.2930297016863985e-06,
"loss": 0.6404,
"step": 799
},
{
"epoch": 0.32186682760008045,
"grad_norm": 3.931383036386479,
"learning_rate": 4.2905813148916816e-06,
"loss": 0.599,
"step": 800
},
{
"epoch": 0.32226916113458054,
"grad_norm": 5.048444025824699,
"learning_rate": 4.288129396576284e-06,
"loss": 0.7259,
"step": 801
},
{
"epoch": 0.3226714946690807,
"grad_norm": 4.851397090992265,
"learning_rate": 4.285673951576062e-06,
"loss": 0.642,
"step": 802
},
{
"epoch": 0.32307382820358077,
"grad_norm": 4.28645665558046,
"learning_rate": 4.283214984733833e-06,
"loss": 0.5315,
"step": 803
},
{
"epoch": 0.32347616173808086,
"grad_norm": 4.265366711136778,
"learning_rate": 4.28075250089936e-06,
"loss": 0.5529,
"step": 804
},
{
"epoch": 0.32387849527258095,
"grad_norm": 3.848835311590686,
"learning_rate": 4.278286504929338e-06,
"loss": 0.5323,
"step": 805
},
{
"epoch": 0.3242808288070811,
"grad_norm": 4.627843495385828,
"learning_rate": 4.2758170016873934e-06,
"loss": 0.5561,
"step": 806
},
{
"epoch": 0.3246831623415812,
"grad_norm": 4.363493591898399,
"learning_rate": 4.273343996044068e-06,
"loss": 0.4805,
"step": 807
},
{
"epoch": 0.32508549587608127,
"grad_norm": 3.9805290188441043,
"learning_rate": 4.270867492876812e-06,
"loss": 0.5204,
"step": 808
},
{
"epoch": 0.32548782941058135,
"grad_norm": 3.9644340799634588,
"learning_rate": 4.268387497069974e-06,
"loss": 0.4642,
"step": 809
},
{
"epoch": 0.3258901629450815,
"grad_norm": 5.269711583475643,
"learning_rate": 4.265904013514788e-06,
"loss": 0.6183,
"step": 810
},
{
"epoch": 0.3262924964795816,
"grad_norm": 4.6945618708498165,
"learning_rate": 4.263417047109371e-06,
"loss": 0.6567,
"step": 811
},
{
"epoch": 0.32669483001408167,
"grad_norm": 4.413878483629632,
"learning_rate": 4.260926602758707e-06,
"loss": 0.5518,
"step": 812
},
{
"epoch": 0.32709716354858176,
"grad_norm": 4.355371114130468,
"learning_rate": 4.258432685374641e-06,
"loss": 0.478,
"step": 813
},
{
"epoch": 0.32749949708308185,
"grad_norm": 4.625144876838143,
"learning_rate": 4.255935299875864e-06,
"loss": 0.5918,
"step": 814
},
{
"epoch": 0.327901830617582,
"grad_norm": 3.8722185488332994,
"learning_rate": 4.253434451187911e-06,
"loss": 0.4698,
"step": 815
},
{
"epoch": 0.3283041641520821,
"grad_norm": 3.8155166378719403,
"learning_rate": 4.250930144243147e-06,
"loss": 0.43,
"step": 816
},
{
"epoch": 0.32870649768658217,
"grad_norm": 4.373317134651141,
"learning_rate": 4.248422383980756e-06,
"loss": 0.581,
"step": 817
},
{
"epoch": 0.32910883122108225,
"grad_norm": 4.316760308023842,
"learning_rate": 4.245911175346733e-06,
"loss": 0.525,
"step": 818
},
{
"epoch": 0.3295111647555824,
"grad_norm": 4.1806833930239895,
"learning_rate": 4.243396523293875e-06,
"loss": 0.6027,
"step": 819
},
{
"epoch": 0.3299134982900825,
"grad_norm": 4.011770765526375,
"learning_rate": 4.240878432781769e-06,
"loss": 0.5642,
"step": 820
},
{
"epoch": 0.3303158318245826,
"grad_norm": 4.106735076414742,
"learning_rate": 4.238356908776786e-06,
"loss": 0.5576,
"step": 821
},
{
"epoch": 0.33071816535908266,
"grad_norm": 4.7608485825878955,
"learning_rate": 4.235831956252069e-06,
"loss": 0.7088,
"step": 822
},
{
"epoch": 0.3311204988935828,
"grad_norm": 4.8786607685170305,
"learning_rate": 4.2333035801875175e-06,
"loss": 0.6513,
"step": 823
},
{
"epoch": 0.3315228324280829,
"grad_norm": 4.288963783688971,
"learning_rate": 4.230771785569791e-06,
"loss": 0.5073,
"step": 824
},
{
"epoch": 0.331925165962583,
"grad_norm": 4.014903400829248,
"learning_rate": 4.228236577392285e-06,
"loss": 0.5901,
"step": 825
},
{
"epoch": 0.33232749949708307,
"grad_norm": 4.154812074980563,
"learning_rate": 4.225697960655131e-06,
"loss": 0.5568,
"step": 826
},
{
"epoch": 0.33272983303158316,
"grad_norm": 3.827755329803438,
"learning_rate": 4.223155940365181e-06,
"loss": 0.4738,
"step": 827
},
{
"epoch": 0.3331321665660833,
"grad_norm": 3.856077516315061,
"learning_rate": 4.220610521536002e-06,
"loss": 0.4963,
"step": 828
},
{
"epoch": 0.3335345001005834,
"grad_norm": 4.615466783911023,
"learning_rate": 4.218061709187862e-06,
"loss": 0.5656,
"step": 829
},
{
"epoch": 0.3339368336350835,
"grad_norm": 4.176966772191371,
"learning_rate": 4.215509508347723e-06,
"loss": 0.5398,
"step": 830
},
{
"epoch": 0.33433916716958356,
"grad_norm": 4.49327690032652,
"learning_rate": 4.212953924049229e-06,
"loss": 0.6772,
"step": 831
},
{
"epoch": 0.3347415007040837,
"grad_norm": 4.067266904294674,
"learning_rate": 4.210394961332698e-06,
"loss": 0.5311,
"step": 832
},
{
"epoch": 0.3351438342385838,
"grad_norm": 3.7375221543142128,
"learning_rate": 4.207832625245112e-06,
"loss": 0.4573,
"step": 833
},
{
"epoch": 0.3355461677730839,
"grad_norm": 4.056435688935082,
"learning_rate": 4.205266920840103e-06,
"loss": 0.543,
"step": 834
},
{
"epoch": 0.33594850130758397,
"grad_norm": 3.9887060748088254,
"learning_rate": 4.202697853177951e-06,
"loss": 0.4613,
"step": 835
},
{
"epoch": 0.3363508348420841,
"grad_norm": 3.5896338953258473,
"learning_rate": 4.2001254273255646e-06,
"loss": 0.4031,
"step": 836
},
{
"epoch": 0.3367531683765842,
"grad_norm": 4.094188015061212,
"learning_rate": 4.197549648356478e-06,
"loss": 0.5933,
"step": 837
},
{
"epoch": 0.3371555019110843,
"grad_norm": 3.6605198066957216,
"learning_rate": 4.19497052135084e-06,
"loss": 0.417,
"step": 838
},
{
"epoch": 0.3375578354455844,
"grad_norm": 5.538220960178179,
"learning_rate": 4.192388051395398e-06,
"loss": 0.7984,
"step": 839
},
{
"epoch": 0.33796016898008446,
"grad_norm": 4.543985654558515,
"learning_rate": 4.1898022435835e-06,
"loss": 0.5526,
"step": 840
},
{
"epoch": 0.3383625025145846,
"grad_norm": 3.719487495093868,
"learning_rate": 4.187213103015069e-06,
"loss": 0.4399,
"step": 841
},
{
"epoch": 0.3387648360490847,
"grad_norm": 3.958292027050775,
"learning_rate": 4.184620634796608e-06,
"loss": 0.4891,
"step": 842
},
{
"epoch": 0.3391671695835848,
"grad_norm": 4.219443448956219,
"learning_rate": 4.182024844041177e-06,
"loss": 0.6417,
"step": 843
},
{
"epoch": 0.33956950311808487,
"grad_norm": 4.637518704467857,
"learning_rate": 4.179425735868395e-06,
"loss": 0.5912,
"step": 844
},
{
"epoch": 0.339971836652585,
"grad_norm": 4.725084492132898,
"learning_rate": 4.176823315404419e-06,
"loss": 0.705,
"step": 845
},
{
"epoch": 0.3403741701870851,
"grad_norm": 4.081067015807228,
"learning_rate": 4.17421758778194e-06,
"loss": 0.5006,
"step": 846
},
{
"epoch": 0.3407765037215852,
"grad_norm": 4.431343567792434,
"learning_rate": 4.1716085581401746e-06,
"loss": 0.5081,
"step": 847
},
{
"epoch": 0.3411788372560853,
"grad_norm": 4.173696495740008,
"learning_rate": 4.1689962316248475e-06,
"loss": 0.497,
"step": 848
},
{
"epoch": 0.3415811707905854,
"grad_norm": 4.125681881515539,
"learning_rate": 4.166380613388189e-06,
"loss": 0.5498,
"step": 849
},
{
"epoch": 0.3419835043250855,
"grad_norm": 4.333217037647524,
"learning_rate": 4.163761708588919e-06,
"loss": 0.5597,
"step": 850
},
{
"epoch": 0.3423858378595856,
"grad_norm": 4.191848923537687,
"learning_rate": 4.161139522392243e-06,
"loss": 0.7259,
"step": 851
},
{
"epoch": 0.3427881713940857,
"grad_norm": 3.1772920188403164,
"learning_rate": 4.158514059969834e-06,
"loss": 0.4225,
"step": 852
},
{
"epoch": 0.34319050492858577,
"grad_norm": 4.166291911511059,
"learning_rate": 4.15588532649983e-06,
"loss": 0.6061,
"step": 853
},
{
"epoch": 0.3435928384630859,
"grad_norm": 3.5991395036430447,
"learning_rate": 4.1532533271668175e-06,
"loss": 0.461,
"step": 854
},
{
"epoch": 0.343995171997586,
"grad_norm": 4.29116994161306,
"learning_rate": 4.150618067161828e-06,
"loss": 0.567,
"step": 855
},
{
"epoch": 0.3443975055320861,
"grad_norm": 4.403897532608296,
"learning_rate": 4.14797955168232e-06,
"loss": 0.6118,
"step": 856
},
{
"epoch": 0.3447998390665862,
"grad_norm": 4.145689255622535,
"learning_rate": 4.145337785932174e-06,
"loss": 0.5709,
"step": 857
},
{
"epoch": 0.3452021726010863,
"grad_norm": 4.540600097537597,
"learning_rate": 4.142692775121684e-06,
"loss": 0.5506,
"step": 858
},
{
"epoch": 0.3456045061355864,
"grad_norm": 4.730037786980354,
"learning_rate": 4.14004452446754e-06,
"loss": 0.6088,
"step": 859
},
{
"epoch": 0.3460068396700865,
"grad_norm": 3.9409862740254487,
"learning_rate": 4.137393039192822e-06,
"loss": 0.5775,
"step": 860
},
{
"epoch": 0.3464091732045866,
"grad_norm": 4.113320460512886,
"learning_rate": 4.1347383245269935e-06,
"loss": 0.5695,
"step": 861
},
{
"epoch": 0.34681150673908673,
"grad_norm": 4.6445735426137364,
"learning_rate": 4.1320803857058835e-06,
"loss": 0.6158,
"step": 862
},
{
"epoch": 0.3472138402735868,
"grad_norm": 4.634455536740089,
"learning_rate": 4.129419227971681e-06,
"loss": 0.6799,
"step": 863
},
{
"epoch": 0.3476161738080869,
"grad_norm": 4.767706536948578,
"learning_rate": 4.1267548565729235e-06,
"loss": 0.5473,
"step": 864
},
{
"epoch": 0.348018507342587,
"grad_norm": 4.4478746327105405,
"learning_rate": 4.124087276764488e-06,
"loss": 0.6564,
"step": 865
},
{
"epoch": 0.3484208408770871,
"grad_norm": 5.1282545461378195,
"learning_rate": 4.121416493807577e-06,
"loss": 0.617,
"step": 866
},
{
"epoch": 0.3488231744115872,
"grad_norm": 3.850370414902996,
"learning_rate": 4.118742512969713e-06,
"loss": 0.5775,
"step": 867
},
{
"epoch": 0.3492255079460873,
"grad_norm": 4.66805214245803,
"learning_rate": 4.116065339524724e-06,
"loss": 0.6724,
"step": 868
},
{
"epoch": 0.3496278414805874,
"grad_norm": 4.7651508619199765,
"learning_rate": 4.113384978752734e-06,
"loss": 0.6151,
"step": 869
},
{
"epoch": 0.3500301750150875,
"grad_norm": 4.207593613220341,
"learning_rate": 4.110701435940157e-06,
"loss": 0.641,
"step": 870
},
{
"epoch": 0.35043250854958763,
"grad_norm": 4.3034195107515965,
"learning_rate": 4.108014716379679e-06,
"loss": 0.6542,
"step": 871
},
{
"epoch": 0.3508348420840877,
"grad_norm": 3.987719667248183,
"learning_rate": 4.105324825370251e-06,
"loss": 0.4996,
"step": 872
},
{
"epoch": 0.3512371756185878,
"grad_norm": 3.9071320373239824,
"learning_rate": 4.102631768217083e-06,
"loss": 0.5168,
"step": 873
},
{
"epoch": 0.3516395091530879,
"grad_norm": 4.2277706570506295,
"learning_rate": 4.099935550231626e-06,
"loss": 0.5379,
"step": 874
},
{
"epoch": 0.35204184268758804,
"grad_norm": 3.9154962483786373,
"learning_rate": 4.097236176731567e-06,
"loss": 0.5523,
"step": 875
},
{
"epoch": 0.3524441762220881,
"grad_norm": 3.465039425796171,
"learning_rate": 4.094533653040814e-06,
"loss": 0.4001,
"step": 876
},
{
"epoch": 0.3528465097565882,
"grad_norm": 3.4706674601181167,
"learning_rate": 4.091827984489493e-06,
"loss": 0.4629,
"step": 877
},
{
"epoch": 0.3532488432910883,
"grad_norm": 4.340590023749821,
"learning_rate": 4.089119176413926e-06,
"loss": 0.6815,
"step": 878
},
{
"epoch": 0.3536511768255884,
"grad_norm": 4.471466388864396,
"learning_rate": 4.086407234156633e-06,
"loss": 0.6459,
"step": 879
},
{
"epoch": 0.35405351036008853,
"grad_norm": 4.449019271914241,
"learning_rate": 4.08369216306631e-06,
"loss": 0.5592,
"step": 880
},
{
"epoch": 0.3544558438945886,
"grad_norm": 4.125369265434845,
"learning_rate": 4.080973968497829e-06,
"loss": 0.5138,
"step": 881
},
{
"epoch": 0.3548581774290887,
"grad_norm": 4.523305588736027,
"learning_rate": 4.07825265581222e-06,
"loss": 0.5929,
"step": 882
},
{
"epoch": 0.3552605109635888,
"grad_norm": 4.077469504247697,
"learning_rate": 4.075528230376662e-06,
"loss": 0.5288,
"step": 883
},
{
"epoch": 0.35566284449808894,
"grad_norm": 5.092003038039258,
"learning_rate": 4.072800697564474e-06,
"loss": 0.6624,
"step": 884
},
{
"epoch": 0.356065178032589,
"grad_norm": 3.7756634610021935,
"learning_rate": 4.070070062755105e-06,
"loss": 0.5265,
"step": 885
},
{
"epoch": 0.3564675115670891,
"grad_norm": 4.106660405022582,
"learning_rate": 4.0673363313341195e-06,
"loss": 0.4853,
"step": 886
},
{
"epoch": 0.3568698451015892,
"grad_norm": 4.215215996153993,
"learning_rate": 4.064599508693191e-06,
"loss": 0.5721,
"step": 887
},
{
"epoch": 0.35727217863608934,
"grad_norm": 4.171756065223583,
"learning_rate": 4.06185960023009e-06,
"loss": 0.4649,
"step": 888
},
{
"epoch": 0.35767451217058943,
"grad_norm": 5.138546885500691,
"learning_rate": 4.05911661134867e-06,
"loss": 0.5972,
"step": 889
},
{
"epoch": 0.3580768457050895,
"grad_norm": 4.136963664338033,
"learning_rate": 4.0563705474588655e-06,
"loss": 0.5027,
"step": 890
},
{
"epoch": 0.3584791792395896,
"grad_norm": 3.940446094705443,
"learning_rate": 4.05362141397667e-06,
"loss": 0.5627,
"step": 891
},
{
"epoch": 0.3588815127740897,
"grad_norm": 4.024692229149252,
"learning_rate": 4.0508692163241356e-06,
"loss": 0.5521,
"step": 892
},
{
"epoch": 0.35928384630858984,
"grad_norm": 3.866043251441755,
"learning_rate": 4.048113959929354e-06,
"loss": 0.5125,
"step": 893
},
{
"epoch": 0.3596861798430899,
"grad_norm": 4.786779622248551,
"learning_rate": 4.0453556502264535e-06,
"loss": 0.6475,
"step": 894
},
{
"epoch": 0.36008851337759,
"grad_norm": 4.127247200504385,
"learning_rate": 4.042594292655581e-06,
"loss": 0.4877,
"step": 895
},
{
"epoch": 0.3604908469120901,
"grad_norm": 3.3408217784719634,
"learning_rate": 4.039829892662897e-06,
"loss": 0.4337,
"step": 896
},
{
"epoch": 0.36089318044659024,
"grad_norm": 4.010232394860985,
"learning_rate": 4.037062455700559e-06,
"loss": 0.493,
"step": 897
},
{
"epoch": 0.36129551398109033,
"grad_norm": 4.3695516591433785,
"learning_rate": 4.03429198722672e-06,
"loss": 0.6051,
"step": 898
},
{
"epoch": 0.3616978475155904,
"grad_norm": 4.766410636581824,
"learning_rate": 4.0315184927055065e-06,
"loss": 0.6602,
"step": 899
},
{
"epoch": 0.3621001810500905,
"grad_norm": 4.1149436048907955,
"learning_rate": 4.028741977607016e-06,
"loss": 0.599,
"step": 900
},
{
"epoch": 0.36250251458459065,
"grad_norm": 3.4871089312720045,
"learning_rate": 4.0259624474073025e-06,
"loss": 0.4779,
"step": 901
},
{
"epoch": 0.36290484811909074,
"grad_norm": 3.6367752409912266,
"learning_rate": 4.023179907588367e-06,
"loss": 0.4541,
"step": 902
},
{
"epoch": 0.3633071816535908,
"grad_norm": 3.3961433962608685,
"learning_rate": 4.020394363638147e-06,
"loss": 0.4265,
"step": 903
},
{
"epoch": 0.3637095151880909,
"grad_norm": 3.8352646417492524,
"learning_rate": 4.0176058210505045e-06,
"loss": 0.5351,
"step": 904
},
{
"epoch": 0.364111848722591,
"grad_norm": 4.242220553627046,
"learning_rate": 4.0148142853252125e-06,
"loss": 0.5792,
"step": 905
},
{
"epoch": 0.36451418225709115,
"grad_norm": 4.096283695840529,
"learning_rate": 4.0120197619679536e-06,
"loss": 0.5412,
"step": 906
},
{
"epoch": 0.36491651579159123,
"grad_norm": 4.729251903541468,
"learning_rate": 4.009222256490297e-06,
"loss": 0.6285,
"step": 907
},
{
"epoch": 0.3653188493260913,
"grad_norm": 4.084180186714877,
"learning_rate": 4.006421774409697e-06,
"loss": 0.6304,
"step": 908
},
{
"epoch": 0.3657211828605914,
"grad_norm": 4.280731987602122,
"learning_rate": 4.003618321249476e-06,
"loss": 0.5632,
"step": 909
},
{
"epoch": 0.36612351639509155,
"grad_norm": 3.868764716611033,
"learning_rate": 4.000811902538821e-06,
"loss": 0.4715,
"step": 910
},
{
"epoch": 0.36652584992959164,
"grad_norm": 3.6445778686982155,
"learning_rate": 3.99800252381276e-06,
"loss": 0.4084,
"step": 911
},
{
"epoch": 0.36692818346409173,
"grad_norm": 4.600601862430362,
"learning_rate": 3.995190190612165e-06,
"loss": 0.7218,
"step": 912
},
{
"epoch": 0.3673305169985918,
"grad_norm": 4.09418361850337,
"learning_rate": 3.992374908483735e-06,
"loss": 0.5913,
"step": 913
},
{
"epoch": 0.36773285053309196,
"grad_norm": 4.495329343052196,
"learning_rate": 3.9895566829799825e-06,
"loss": 0.6735,
"step": 914
},
{
"epoch": 0.36813518406759205,
"grad_norm": 4.408452152812246,
"learning_rate": 3.986735519659226e-06,
"loss": 0.6896,
"step": 915
},
{
"epoch": 0.36853751760209214,
"grad_norm": 3.7084757585295276,
"learning_rate": 3.983911424085578e-06,
"loss": 0.552,
"step": 916
},
{
"epoch": 0.3689398511365922,
"grad_norm": 3.8368900456075985,
"learning_rate": 3.981084401828937e-06,
"loss": 0.4944,
"step": 917
},
{
"epoch": 0.3693421846710923,
"grad_norm": 5.696133375208026,
"learning_rate": 3.978254458464969e-06,
"loss": 0.6338,
"step": 918
},
{
"epoch": 0.36974451820559245,
"grad_norm": 4.144365741395642,
"learning_rate": 3.975421599575103e-06,
"loss": 0.6463,
"step": 919
},
{
"epoch": 0.37014685174009254,
"grad_norm": 3.7665583063095682,
"learning_rate": 3.972585830746522e-06,
"loss": 0.4442,
"step": 920
},
{
"epoch": 0.37054918527459263,
"grad_norm": 4.341604001350467,
"learning_rate": 3.969747157572142e-06,
"loss": 0.7191,
"step": 921
},
{
"epoch": 0.3709515188090927,
"grad_norm": 4.724527719836307,
"learning_rate": 3.966905585650611e-06,
"loss": 0.7508,
"step": 922
},
{
"epoch": 0.37135385234359286,
"grad_norm": 6.112644943798556,
"learning_rate": 3.964061120586294e-06,
"loss": 0.6744,
"step": 923
},
{
"epoch": 0.37175618587809295,
"grad_norm": 4.630742134072846,
"learning_rate": 3.961213767989261e-06,
"loss": 0.5703,
"step": 924
},
{
"epoch": 0.37215851941259304,
"grad_norm": 3.5499854666112802,
"learning_rate": 3.958363533475277e-06,
"loss": 0.4563,
"step": 925
},
{
"epoch": 0.3725608529470931,
"grad_norm": 3.926859330076321,
"learning_rate": 3.955510422665791e-06,
"loss": 0.5159,
"step": 926
},
{
"epoch": 0.37296318648159327,
"grad_norm": 4.029967110899967,
"learning_rate": 3.952654441187927e-06,
"loss": 0.5267,
"step": 927
},
{
"epoch": 0.37336552001609336,
"grad_norm": 3.4483709639926414,
"learning_rate": 3.9497955946744675e-06,
"loss": 0.4584,
"step": 928
},
{
"epoch": 0.37376785355059344,
"grad_norm": 4.752285945278346,
"learning_rate": 3.946933888763847e-06,
"loss": 0.5743,
"step": 929
},
{
"epoch": 0.37417018708509353,
"grad_norm": 3.985801543832948,
"learning_rate": 3.9440693291001406e-06,
"loss": 0.5319,
"step": 930
},
{
"epoch": 0.3745725206195936,
"grad_norm": 3.5946794819227903,
"learning_rate": 3.941201921333048e-06,
"loss": 0.4217,
"step": 931
},
{
"epoch": 0.37497485415409376,
"grad_norm": 4.035357970865375,
"learning_rate": 3.9383316711178946e-06,
"loss": 0.503,
"step": 932
},
{
"epoch": 0.37537718768859385,
"grad_norm": 4.034839036975055,
"learning_rate": 3.935458584115599e-06,
"loss": 0.4698,
"step": 933
},
{
"epoch": 0.37577952122309394,
"grad_norm": 4.286963186422666,
"learning_rate": 3.932582665992688e-06,
"loss": 0.5901,
"step": 934
},
{
"epoch": 0.376181854757594,
"grad_norm": 4.021909925329209,
"learning_rate": 3.929703922421263e-06,
"loss": 0.5111,
"step": 935
},
{
"epoch": 0.37658418829209417,
"grad_norm": 4.0942098752342035,
"learning_rate": 3.926822359079001e-06,
"loss": 0.5472,
"step": 936
},
{
"epoch": 0.37698652182659426,
"grad_norm": 4.261504012145757,
"learning_rate": 3.923937981649143e-06,
"loss": 0.6453,
"step": 937
},
{
"epoch": 0.37738885536109434,
"grad_norm": 4.014117574115026,
"learning_rate": 3.921050795820474e-06,
"loss": 0.5056,
"step": 938
},
{
"epoch": 0.37779118889559443,
"grad_norm": 3.904726379027022,
"learning_rate": 3.9181608072873244e-06,
"loss": 0.5304,
"step": 939
},
{
"epoch": 0.3781935224300946,
"grad_norm": 3.881019901522325,
"learning_rate": 3.9152680217495475e-06,
"loss": 0.5298,
"step": 940
},
{
"epoch": 0.37859585596459466,
"grad_norm": 4.058485666284913,
"learning_rate": 3.912372444912517e-06,
"loss": 0.574,
"step": 941
},
{
"epoch": 0.37899818949909475,
"grad_norm": 3.9630383581476205,
"learning_rate": 3.909474082487107e-06,
"loss": 0.5804,
"step": 942
},
{
"epoch": 0.37940052303359484,
"grad_norm": 3.8676063595588683,
"learning_rate": 3.906572940189691e-06,
"loss": 0.4804,
"step": 943
},
{
"epoch": 0.3798028565680949,
"grad_norm": 4.32834774261345,
"learning_rate": 3.9036690237421215e-06,
"loss": 0.6161,
"step": 944
},
{
"epoch": 0.38020519010259507,
"grad_norm": 4.726002477400136,
"learning_rate": 3.900762338871723e-06,
"loss": 0.6898,
"step": 945
},
{
"epoch": 0.38060752363709516,
"grad_norm": 4.577558255257823,
"learning_rate": 3.897852891311282e-06,
"loss": 0.5504,
"step": 946
},
{
"epoch": 0.38100985717159525,
"grad_norm": 4.300878005170684,
"learning_rate": 3.89494068679903e-06,
"loss": 0.6734,
"step": 947
},
{
"epoch": 0.38141219070609533,
"grad_norm": 4.519778162808581,
"learning_rate": 3.892025731078641e-06,
"loss": 0.5694,
"step": 948
},
{
"epoch": 0.3818145242405955,
"grad_norm": 3.8176275411987683,
"learning_rate": 3.88910802989921e-06,
"loss": 0.4444,
"step": 949
},
{
"epoch": 0.38221685777509556,
"grad_norm": 4.014909553832162,
"learning_rate": 3.886187589015251e-06,
"loss": 0.6054,
"step": 950
},
{
"epoch": 0.38261919130959565,
"grad_norm": 4.267781377287406,
"learning_rate": 3.883264414186677e-06,
"loss": 0.6257,
"step": 951
},
{
"epoch": 0.38302152484409574,
"grad_norm": 4.345044782459552,
"learning_rate": 3.8803385111788e-06,
"loss": 0.5543,
"step": 952
},
{
"epoch": 0.3834238583785959,
"grad_norm": 4.221128544535545,
"learning_rate": 3.877409885762305e-06,
"loss": 0.5544,
"step": 953
},
{
"epoch": 0.38382619191309597,
"grad_norm": 4.000060887768021,
"learning_rate": 3.8744785437132506e-06,
"loss": 0.5878,
"step": 954
},
{
"epoch": 0.38422852544759606,
"grad_norm": 3.8372993108948488,
"learning_rate": 3.871544490813054e-06,
"loss": 0.5076,
"step": 955
},
{
"epoch": 0.38463085898209615,
"grad_norm": 4.059706126497868,
"learning_rate": 3.868607732848475e-06,
"loss": 0.5353,
"step": 956
},
{
"epoch": 0.38503319251659623,
"grad_norm": 4.408924474844376,
"learning_rate": 3.865668275611614e-06,
"loss": 0.519,
"step": 957
},
{
"epoch": 0.3854355260510964,
"grad_norm": 4.010089079135815,
"learning_rate": 3.862726124899889e-06,
"loss": 0.5952,
"step": 958
},
{
"epoch": 0.38583785958559647,
"grad_norm": 5.459071377748317,
"learning_rate": 3.859781286516037e-06,
"loss": 0.7175,
"step": 959
},
{
"epoch": 0.38624019312009655,
"grad_norm": 4.542741280511702,
"learning_rate": 3.85683376626809e-06,
"loss": 0.5667,
"step": 960
},
{
"epoch": 0.38664252665459664,
"grad_norm": 3.99184341470565,
"learning_rate": 3.853883569969372e-06,
"loss": 0.6148,
"step": 961
},
{
"epoch": 0.3870448601890968,
"grad_norm": 4.078873510656604,
"learning_rate": 3.850930703438484e-06,
"loss": 0.5383,
"step": 962
},
{
"epoch": 0.38744719372359687,
"grad_norm": 3.896065226196885,
"learning_rate": 3.847975172499295e-06,
"loss": 0.6785,
"step": 963
},
{
"epoch": 0.38784952725809696,
"grad_norm": 4.328062559714918,
"learning_rate": 3.845016982980929e-06,
"loss": 0.5543,
"step": 964
},
{
"epoch": 0.38825186079259705,
"grad_norm": 4.2981177827507935,
"learning_rate": 3.842056140717749e-06,
"loss": 0.6935,
"step": 965
},
{
"epoch": 0.3886541943270972,
"grad_norm": 4.489409385825919,
"learning_rate": 3.839092651549357e-06,
"loss": 0.6412,
"step": 966
},
{
"epoch": 0.3890565278615973,
"grad_norm": 4.162311148181462,
"learning_rate": 3.836126521320569e-06,
"loss": 0.5133,
"step": 967
},
{
"epoch": 0.38945886139609737,
"grad_norm": 3.7564756341509464,
"learning_rate": 3.833157755881414e-06,
"loss": 0.5198,
"step": 968
},
{
"epoch": 0.38986119493059745,
"grad_norm": 4.2353823981014385,
"learning_rate": 3.830186361087117e-06,
"loss": 0.4271,
"step": 969
},
{
"epoch": 0.39026352846509754,
"grad_norm": 3.4858668216855437,
"learning_rate": 3.827212342798089e-06,
"loss": 0.4654,
"step": 970
},
{
"epoch": 0.3906658619995977,
"grad_norm": 3.920259549785761,
"learning_rate": 3.824235706879915e-06,
"loss": 0.5822,
"step": 971
},
{
"epoch": 0.3910681955340978,
"grad_norm": 4.239371514940848,
"learning_rate": 3.821256459203343e-06,
"loss": 0.5093,
"step": 972
},
{
"epoch": 0.39147052906859786,
"grad_norm": 4.037277711642943,
"learning_rate": 3.81827460564427e-06,
"loss": 0.5222,
"step": 973
},
{
"epoch": 0.39187286260309795,
"grad_norm": 4.456827976048942,
"learning_rate": 3.815290152083737e-06,
"loss": 0.6783,
"step": 974
},
{
"epoch": 0.3922751961375981,
"grad_norm": 4.246220187755789,
"learning_rate": 3.8123031044079084e-06,
"loss": 0.5888,
"step": 975
},
{
"epoch": 0.3926775296720982,
"grad_norm": 4.3203808668287635,
"learning_rate": 3.809313468508068e-06,
"loss": 0.5569,
"step": 976
},
{
"epoch": 0.39307986320659827,
"grad_norm": 4.860143257226567,
"learning_rate": 3.8063212502806035e-06,
"loss": 0.5953,
"step": 977
},
{
"epoch": 0.39348219674109836,
"grad_norm": 4.233378915892834,
"learning_rate": 3.8033264556269954e-06,
"loss": 0.6167,
"step": 978
},
{
"epoch": 0.39388453027559844,
"grad_norm": 4.3534789683754305,
"learning_rate": 3.8003290904538035e-06,
"loss": 0.6133,
"step": 979
},
{
"epoch": 0.3942868638100986,
"grad_norm": 3.9519882332598257,
"learning_rate": 3.7973291606726614e-06,
"loss": 0.4515,
"step": 980
},
{
"epoch": 0.3946891973445987,
"grad_norm": 4.138408229253922,
"learning_rate": 3.794326672200258e-06,
"loss": 0.5607,
"step": 981
},
{
"epoch": 0.39509153087909876,
"grad_norm": 4.3761985221008635,
"learning_rate": 3.79132163095833e-06,
"loss": 0.5507,
"step": 982
},
{
"epoch": 0.39549386441359885,
"grad_norm": 4.023180334056979,
"learning_rate": 3.7883140428736477e-06,
"loss": 0.5314,
"step": 983
},
{
"epoch": 0.395896197948099,
"grad_norm": 4.151232898341051,
"learning_rate": 3.7853039138780057e-06,
"loss": 0.5392,
"step": 984
},
{
"epoch": 0.3962985314825991,
"grad_norm": 4.539638233125882,
"learning_rate": 3.7822912499082087e-06,
"loss": 0.5791,
"step": 985
},
{
"epoch": 0.39670086501709917,
"grad_norm": 4.335330132423181,
"learning_rate": 3.7792760569060626e-06,
"loss": 0.5178,
"step": 986
},
{
"epoch": 0.39710319855159926,
"grad_norm": 4.143033413053681,
"learning_rate": 3.7762583408183594e-06,
"loss": 0.5815,
"step": 987
},
{
"epoch": 0.3975055320860994,
"grad_norm": 4.915299915525361,
"learning_rate": 3.7732381075968694e-06,
"loss": 0.6269,
"step": 988
},
{
"epoch": 0.3979078656205995,
"grad_norm": 3.7212476195736506,
"learning_rate": 3.770215363198325e-06,
"loss": 0.4599,
"step": 989
},
{
"epoch": 0.3983101991550996,
"grad_norm": 3.9977745943956884,
"learning_rate": 3.7671901135844148e-06,
"loss": 0.5973,
"step": 990
},
{
"epoch": 0.39871253268959966,
"grad_norm": 4.3088468446821615,
"learning_rate": 3.764162364721764e-06,
"loss": 0.5963,
"step": 991
},
{
"epoch": 0.39911486622409975,
"grad_norm": 4.2991674025596955,
"learning_rate": 3.761132122581931e-06,
"loss": 0.6923,
"step": 992
},
{
"epoch": 0.3995171997585999,
"grad_norm": 3.952746214125807,
"learning_rate": 3.758099393141388e-06,
"loss": 0.5384,
"step": 993
},
{
"epoch": 0.3999195332931,
"grad_norm": 4.327315800519712,
"learning_rate": 3.755064182381516e-06,
"loss": 0.5144,
"step": 994
},
{
"epoch": 0.40032186682760007,
"grad_norm": 4.713412610694241,
"learning_rate": 3.752026496288588e-06,
"loss": 0.641,
"step": 995
},
{
"epoch": 0.40072420036210016,
"grad_norm": 3.9912240405323507,
"learning_rate": 3.74898634085376e-06,
"loss": 0.5646,
"step": 996
},
{
"epoch": 0.4011265338966003,
"grad_norm": 4.530973641905437,
"learning_rate": 3.7459437220730583e-06,
"loss": 0.5614,
"step": 997
},
{
"epoch": 0.4015288674311004,
"grad_norm": 6.864543852090396,
"learning_rate": 3.742898645947366e-06,
"loss": 0.5336,
"step": 998
},
{
"epoch": 0.4019312009656005,
"grad_norm": 4.429956976671279,
"learning_rate": 3.7398511184824144e-06,
"loss": 0.7004,
"step": 999
},
{
"epoch": 0.40233353450010056,
"grad_norm": 3.164463326468864,
"learning_rate": 3.7368011456887693e-06,
"loss": 0.4163,
"step": 1000
},
{
"epoch": 0.4027358680346007,
"grad_norm": 3.7984416296897456,
"learning_rate": 3.7337487335818185e-06,
"loss": 0.5342,
"step": 1001
},
{
"epoch": 0.4031382015691008,
"grad_norm": 4.186189188641626,
"learning_rate": 3.730693888181761e-06,
"loss": 0.6125,
"step": 1002
},
{
"epoch": 0.4035405351036009,
"grad_norm": 3.8782636909404506,
"learning_rate": 3.7276366155135968e-06,
"loss": 0.5669,
"step": 1003
},
{
"epoch": 0.40394286863810097,
"grad_norm": 4.222464351010827,
"learning_rate": 3.7245769216071104e-06,
"loss": 0.5574,
"step": 1004
},
{
"epoch": 0.40434520217260106,
"grad_norm": 4.0429197887930135,
"learning_rate": 3.721514812496863e-06,
"loss": 0.5031,
"step": 1005
},
{
"epoch": 0.4047475357071012,
"grad_norm": 4.015964628315664,
"learning_rate": 3.718450294222179e-06,
"loss": 0.5994,
"step": 1006
},
{
"epoch": 0.4051498692416013,
"grad_norm": 3.846521589103569,
"learning_rate": 3.7153833728271356e-06,
"loss": 0.4928,
"step": 1007
},
{
"epoch": 0.4055522027761014,
"grad_norm": 4.138445980201307,
"learning_rate": 3.712314054360547e-06,
"loss": 0.5722,
"step": 1008
},
{
"epoch": 0.40595453631060147,
"grad_norm": 4.476195845673172,
"learning_rate": 3.7092423448759577e-06,
"loss": 0.563,
"step": 1009
},
{
"epoch": 0.4063568698451016,
"grad_norm": 3.786638148470317,
"learning_rate": 3.7061682504316264e-06,
"loss": 0.5181,
"step": 1010
},
{
"epoch": 0.4067592033796017,
"grad_norm": 4.563364096910093,
"learning_rate": 3.7030917770905155e-06,
"loss": 0.5247,
"step": 1011
},
{
"epoch": 0.4071615369141018,
"grad_norm": 4.276417493626447,
"learning_rate": 3.7000129309202804e-06,
"loss": 0.5815,
"step": 1012
},
{
"epoch": 0.4075638704486019,
"grad_norm": 4.495474983317033,
"learning_rate": 3.696931717993256e-06,
"loss": 0.7269,
"step": 1013
},
{
"epoch": 0.407966203983102,
"grad_norm": 3.6238415974427527,
"learning_rate": 3.693848144386444e-06,
"loss": 0.5462,
"step": 1014
},
{
"epoch": 0.4083685375176021,
"grad_norm": 4.262996632479673,
"learning_rate": 3.6907622161815027e-06,
"loss": 0.64,
"step": 1015
},
{
"epoch": 0.4087708710521022,
"grad_norm": 3.8744140465414745,
"learning_rate": 3.687673939464736e-06,
"loss": 0.5281,
"step": 1016
},
{
"epoch": 0.4091732045866023,
"grad_norm": 4.021193749407921,
"learning_rate": 3.6845833203270766e-06,
"loss": 0.5238,
"step": 1017
},
{
"epoch": 0.40957553812110237,
"grad_norm": 3.7942462225606084,
"learning_rate": 3.6814903648640787e-06,
"loss": 0.5274,
"step": 1018
},
{
"epoch": 0.4099778716556025,
"grad_norm": 3.8584763272892193,
"learning_rate": 3.6783950791759053e-06,
"loss": 0.6652,
"step": 1019
},
{
"epoch": 0.4103802051901026,
"grad_norm": 4.224914273954211,
"learning_rate": 3.6752974693673144e-06,
"loss": 0.6418,
"step": 1020
},
{
"epoch": 0.4107825387246027,
"grad_norm": 3.611777243329413,
"learning_rate": 3.6721975415476475e-06,
"loss": 0.4478,
"step": 1021
},
{
"epoch": 0.4111848722591028,
"grad_norm": 4.492308490256186,
"learning_rate": 3.669095301830816e-06,
"loss": 0.6383,
"step": 1022
},
{
"epoch": 0.4115872057936029,
"grad_norm": 3.966328790708119,
"learning_rate": 3.6659907563352964e-06,
"loss": 0.5496,
"step": 1023
},
{
"epoch": 0.411989539328103,
"grad_norm": 4.331090241917638,
"learning_rate": 3.6628839111841067e-06,
"loss": 0.6138,
"step": 1024
},
{
"epoch": 0.4123918728626031,
"grad_norm": 4.039380276956274,
"learning_rate": 3.6597747725048054e-06,
"loss": 0.514,
"step": 1025
},
{
"epoch": 0.4127942063971032,
"grad_norm": 4.2678553912438515,
"learning_rate": 3.656663346429471e-06,
"loss": 0.5304,
"step": 1026
},
{
"epoch": 0.4131965399316033,
"grad_norm": 3.6311274537700444,
"learning_rate": 3.6535496390946958e-06,
"loss": 0.5324,
"step": 1027
},
{
"epoch": 0.4135988734661034,
"grad_norm": 4.179608207864191,
"learning_rate": 3.650433656641569e-06,
"loss": 0.5913,
"step": 1028
},
{
"epoch": 0.4140012070006035,
"grad_norm": 3.8042489733382308,
"learning_rate": 3.647315405215668e-06,
"loss": 0.5007,
"step": 1029
},
{
"epoch": 0.4144035405351036,
"grad_norm": 3.4526666984193404,
"learning_rate": 3.644194890967047e-06,
"loss": 0.4141,
"step": 1030
},
{
"epoch": 0.4148058740696037,
"grad_norm": 4.657500083163389,
"learning_rate": 3.641072120050221e-06,
"loss": 0.6175,
"step": 1031
},
{
"epoch": 0.4152082076041038,
"grad_norm": 3.5394692246958934,
"learning_rate": 3.637947098624156e-06,
"loss": 0.493,
"step": 1032
},
{
"epoch": 0.4156105411386039,
"grad_norm": 3.6382852897066003,
"learning_rate": 3.6348198328522565e-06,
"loss": 0.4733,
"step": 1033
},
{
"epoch": 0.416012874673104,
"grad_norm": 4.701593231983809,
"learning_rate": 3.631690328902355e-06,
"loss": 0.6759,
"step": 1034
},
{
"epoch": 0.4164152082076041,
"grad_norm": 4.126519542614592,
"learning_rate": 3.6285585929466982e-06,
"loss": 0.6065,
"step": 1035
},
{
"epoch": 0.4168175417421042,
"grad_norm": 4.174303064525279,
"learning_rate": 3.6254246311619317e-06,
"loss": 0.6019,
"step": 1036
},
{
"epoch": 0.4172198752766043,
"grad_norm": 4.369071505943041,
"learning_rate": 3.6222884497290937e-06,
"loss": 0.47,
"step": 1037
},
{
"epoch": 0.4176222088111044,
"grad_norm": 4.547283701461554,
"learning_rate": 3.619150054833601e-06,
"loss": 0.5666,
"step": 1038
},
{
"epoch": 0.4180245423456045,
"grad_norm": 6.593645962994248,
"learning_rate": 3.616009452665234e-06,
"loss": 0.4854,
"step": 1039
},
{
"epoch": 0.41842687588010463,
"grad_norm": 4.3350300541777145,
"learning_rate": 3.6128666494181274e-06,
"loss": 0.6522,
"step": 1040
},
{
"epoch": 0.4188292094146047,
"grad_norm": 3.4916298319626233,
"learning_rate": 3.6097216512907574e-06,
"loss": 0.5449,
"step": 1041
},
{
"epoch": 0.4192315429491048,
"grad_norm": 3.721480691979447,
"learning_rate": 3.6065744644859276e-06,
"loss": 0.4152,
"step": 1042
},
{
"epoch": 0.4196338764836049,
"grad_norm": 4.419075622535455,
"learning_rate": 3.6034250952107598e-06,
"loss": 0.5998,
"step": 1043
},
{
"epoch": 0.420036210018105,
"grad_norm": 4.066379360835346,
"learning_rate": 3.6002735496766787e-06,
"loss": 0.4443,
"step": 1044
},
{
"epoch": 0.4204385435526051,
"grad_norm": 5.277931637633872,
"learning_rate": 3.5971198340994035e-06,
"loss": 0.6578,
"step": 1045
},
{
"epoch": 0.4208408770871052,
"grad_norm": 4.171317951485179,
"learning_rate": 3.5939639546989315e-06,
"loss": 0.5362,
"step": 1046
},
{
"epoch": 0.4212432106216053,
"grad_norm": 3.6581061080480786,
"learning_rate": 3.5908059176995274e-06,
"loss": 0.4868,
"step": 1047
},
{
"epoch": 0.4216455441561054,
"grad_norm": 4.089304060454962,
"learning_rate": 3.587645729329713e-06,
"loss": 0.4962,
"step": 1048
},
{
"epoch": 0.42204787769060553,
"grad_norm": 4.193929975116655,
"learning_rate": 3.584483395822252e-06,
"loss": 0.585,
"step": 1049
},
{
"epoch": 0.4224502112251056,
"grad_norm": 4.4833727752476324,
"learning_rate": 3.5813189234141386e-06,
"loss": 0.5295,
"step": 1050
},
{
"epoch": 0.4228525447596057,
"grad_norm": 4.131740789977084,
"learning_rate": 3.5781523183465868e-06,
"loss": 0.5408,
"step": 1051
},
{
"epoch": 0.4232548782941058,
"grad_norm": 3.972093255632944,
"learning_rate": 3.574983586865015e-06,
"loss": 0.5256,
"step": 1052
},
{
"epoch": 0.42365721182860594,
"grad_norm": 4.619320510735413,
"learning_rate": 3.571812735219037e-06,
"loss": 0.6365,
"step": 1053
},
{
"epoch": 0.424059545363106,
"grad_norm": 4.543758819054478,
"learning_rate": 3.5686397696624486e-06,
"loss": 0.5358,
"step": 1054
},
{
"epoch": 0.4244618788976061,
"grad_norm": 3.9488894694458305,
"learning_rate": 3.565464696453212e-06,
"loss": 0.5101,
"step": 1055
},
{
"epoch": 0.4248642124321062,
"grad_norm": 3.9008766737779217,
"learning_rate": 3.5622875218534493e-06,
"loss": 0.6157,
"step": 1056
},
{
"epoch": 0.4252665459666063,
"grad_norm": 4.072195384389267,
"learning_rate": 3.5591082521294264e-06,
"loss": 0.572,
"step": 1057
},
{
"epoch": 0.42566887950110643,
"grad_norm": 3.7407260188563622,
"learning_rate": 3.555926893551539e-06,
"loss": 0.4571,
"step": 1058
},
{
"epoch": 0.4260712130356065,
"grad_norm": 4.030808733877317,
"learning_rate": 3.552743452394306e-06,
"loss": 0.5615,
"step": 1059
},
{
"epoch": 0.4264735465701066,
"grad_norm": 3.833659387273303,
"learning_rate": 3.5495579349363517e-06,
"loss": 0.4978,
"step": 1060
},
{
"epoch": 0.4268758801046067,
"grad_norm": 4.304988237070821,
"learning_rate": 3.5463703474603964e-06,
"loss": 0.6441,
"step": 1061
},
{
"epoch": 0.42727821363910684,
"grad_norm": 4.68914366889499,
"learning_rate": 3.5431806962532426e-06,
"loss": 0.5009,
"step": 1062
},
{
"epoch": 0.42768054717360693,
"grad_norm": 4.208417431491987,
"learning_rate": 3.539988987605763e-06,
"loss": 0.5556,
"step": 1063
},
{
"epoch": 0.428082880708107,
"grad_norm": 3.913021062753838,
"learning_rate": 3.5367952278128877e-06,
"loss": 0.3997,
"step": 1064
},
{
"epoch": 0.4284852142426071,
"grad_norm": 3.975035687293758,
"learning_rate": 3.5335994231735927e-06,
"loss": 0.5394,
"step": 1065
},
{
"epoch": 0.42888754777710725,
"grad_norm": 3.969954958131227,
"learning_rate": 3.5304015799908875e-06,
"loss": 0.579,
"step": 1066
},
{
"epoch": 0.42928988131160734,
"grad_norm": 4.942462188783348,
"learning_rate": 3.5272017045718016e-06,
"loss": 0.6723,
"step": 1067
},
{
"epoch": 0.4296922148461074,
"grad_norm": 4.354492013681113,
"learning_rate": 3.5239998032273727e-06,
"loss": 0.5724,
"step": 1068
},
{
"epoch": 0.4300945483806075,
"grad_norm": 4.275789917585218,
"learning_rate": 3.520795882272634e-06,
"loss": 0.5592,
"step": 1069
},
{
"epoch": 0.4304968819151076,
"grad_norm": 4.0220453706013535,
"learning_rate": 3.5175899480266023e-06,
"loss": 0.6004,
"step": 1070
},
{
"epoch": 0.43089921544960774,
"grad_norm": 3.956820759857449,
"learning_rate": 3.5143820068122647e-06,
"loss": 0.4667,
"step": 1071
},
{
"epoch": 0.43130154898410783,
"grad_norm": 3.9359268522062996,
"learning_rate": 3.5111720649565685e-06,
"loss": 0.4766,
"step": 1072
},
{
"epoch": 0.4317038825186079,
"grad_norm": 3.9665074015294555,
"learning_rate": 3.507960128790402e-06,
"loss": 0.5496,
"step": 1073
},
{
"epoch": 0.432106216053108,
"grad_norm": 4.086635346156316,
"learning_rate": 3.5047462046485934e-06,
"loss": 0.4863,
"step": 1074
},
{
"epoch": 0.43250854958760815,
"grad_norm": 4.234886651798155,
"learning_rate": 3.501530298869886e-06,
"loss": 0.5373,
"step": 1075
},
{
"epoch": 0.43291088312210824,
"grad_norm": 4.578190826029045,
"learning_rate": 3.4983124177969353e-06,
"loss": 0.6952,
"step": 1076
},
{
"epoch": 0.4333132166566083,
"grad_norm": 4.198352580671215,
"learning_rate": 3.495092567776291e-06,
"loss": 0.6753,
"step": 1077
},
{
"epoch": 0.4337155501911084,
"grad_norm": 4.25314346947828,
"learning_rate": 3.4918707551583853e-06,
"loss": 0.6139,
"step": 1078
},
{
"epoch": 0.43411788372560856,
"grad_norm": 3.7809774120139084,
"learning_rate": 3.4886469862975224e-06,
"loss": 0.5401,
"step": 1079
},
{
"epoch": 0.43452021726010864,
"grad_norm": 3.861747118430867,
"learning_rate": 3.485421267551865e-06,
"loss": 0.4343,
"step": 1080
},
{
"epoch": 0.43492255079460873,
"grad_norm": 4.1971689225974655,
"learning_rate": 3.482193605283421e-06,
"loss": 0.4651,
"step": 1081
},
{
"epoch": 0.4353248843291088,
"grad_norm": 4.719733540576249,
"learning_rate": 3.478964005858031e-06,
"loss": 0.6265,
"step": 1082
},
{
"epoch": 0.4357272178636089,
"grad_norm": 4.199121895188385,
"learning_rate": 3.475732475645357e-06,
"loss": 0.5915,
"step": 1083
},
{
"epoch": 0.43612955139810905,
"grad_norm": 4.117901413415391,
"learning_rate": 3.4724990210188693e-06,
"loss": 0.5859,
"step": 1084
},
{
"epoch": 0.43653188493260914,
"grad_norm": 4.063314519073365,
"learning_rate": 3.4692636483558316e-06,
"loss": 0.5799,
"step": 1085
},
{
"epoch": 0.4369342184671092,
"grad_norm": 4.291259902841789,
"learning_rate": 3.4660263640372926e-06,
"loss": 0.5972,
"step": 1086
},
{
"epoch": 0.4373365520016093,
"grad_norm": 4.317671430467203,
"learning_rate": 3.462787174448071e-06,
"loss": 0.5818,
"step": 1087
},
{
"epoch": 0.43773888553610946,
"grad_norm": 4.041367856828536,
"learning_rate": 3.459546085976743e-06,
"loss": 0.5635,
"step": 1088
},
{
"epoch": 0.43814121907060954,
"grad_norm": 5.295724094479603,
"learning_rate": 3.45630310501563e-06,
"loss": 0.6712,
"step": 1089
},
{
"epoch": 0.43854355260510963,
"grad_norm": 4.415113705137869,
"learning_rate": 3.453058237960785e-06,
"loss": 0.578,
"step": 1090
},
{
"epoch": 0.4389458861396097,
"grad_norm": 4.107191353105501,
"learning_rate": 3.4498114912119817e-06,
"loss": 0.5199,
"step": 1091
},
{
"epoch": 0.43934821967410986,
"grad_norm": 4.1779462968909336,
"learning_rate": 3.4465628711727022e-06,
"loss": 0.577,
"step": 1092
},
{
"epoch": 0.43975055320860995,
"grad_norm": 3.779499327289251,
"learning_rate": 3.4433123842501205e-06,
"loss": 0.4508,
"step": 1093
},
{
"epoch": 0.44015288674311004,
"grad_norm": 3.9872787051249077,
"learning_rate": 3.440060036855095e-06,
"loss": 0.5583,
"step": 1094
},
{
"epoch": 0.4405552202776101,
"grad_norm": 4.317372927534622,
"learning_rate": 3.4368058354021526e-06,
"loss": 0.5136,
"step": 1095
},
{
"epoch": 0.4409575538121102,
"grad_norm": 3.7148972800031053,
"learning_rate": 3.4335497863094773e-06,
"loss": 0.514,
"step": 1096
},
{
"epoch": 0.44135988734661036,
"grad_norm": 4.342233656787418,
"learning_rate": 3.430291895998896e-06,
"loss": 0.6742,
"step": 1097
},
{
"epoch": 0.44176222088111045,
"grad_norm": 4.799719138290185,
"learning_rate": 3.4270321708958686e-06,
"loss": 0.6306,
"step": 1098
},
{
"epoch": 0.44216455441561053,
"grad_norm": 3.9209860725209382,
"learning_rate": 3.4237706174294726e-06,
"loss": 0.5836,
"step": 1099
},
{
"epoch": 0.4425668879501106,
"grad_norm": 4.127404724187905,
"learning_rate": 3.420507242032392e-06,
"loss": 0.5628,
"step": 1100
},
{
"epoch": 0.44296922148461076,
"grad_norm": 4.956274683013535,
"learning_rate": 3.4172420511409033e-06,
"loss": 0.5366,
"step": 1101
},
{
"epoch": 0.44337155501911085,
"grad_norm": 3.9527811079754285,
"learning_rate": 3.413975051194865e-06,
"loss": 0.4531,
"step": 1102
},
{
"epoch": 0.44377388855361094,
"grad_norm": 4.364141896413706,
"learning_rate": 3.4107062486377028e-06,
"loss": 0.4914,
"step": 1103
},
{
"epoch": 0.44417622208811103,
"grad_norm": 4.349108898092637,
"learning_rate": 3.4074356499163976e-06,
"loss": 0.7239,
"step": 1104
},
{
"epoch": 0.44457855562261117,
"grad_norm": 4.038974541606429,
"learning_rate": 3.4041632614814734e-06,
"loss": 0.6375,
"step": 1105
},
{
"epoch": 0.44498088915711126,
"grad_norm": 4.811612078810641,
"learning_rate": 3.400889089786984e-06,
"loss": 0.4963,
"step": 1106
},
{
"epoch": 0.44538322269161135,
"grad_norm": 4.07473806896573,
"learning_rate": 3.397613141290499e-06,
"loss": 0.575,
"step": 1107
},
{
"epoch": 0.44578555622611143,
"grad_norm": 4.160302353185681,
"learning_rate": 3.3943354224530934e-06,
"loss": 0.4766,
"step": 1108
},
{
"epoch": 0.4461878897606115,
"grad_norm": 3.6900007830747934,
"learning_rate": 3.3910559397393335e-06,
"loss": 0.5544,
"step": 1109
},
{
"epoch": 0.44659022329511167,
"grad_norm": 4.333978444807443,
"learning_rate": 3.3877746996172645e-06,
"loss": 0.5637,
"step": 1110
},
{
"epoch": 0.44699255682961175,
"grad_norm": 3.768061403793603,
"learning_rate": 3.384491708558399e-06,
"loss": 0.4659,
"step": 1111
},
{
"epoch": 0.44739489036411184,
"grad_norm": 3.9713683928936208,
"learning_rate": 3.3812069730377007e-06,
"loss": 0.539,
"step": 1112
},
{
"epoch": 0.44779722389861193,
"grad_norm": 4.95129840549997,
"learning_rate": 3.3779204995335747e-06,
"loss": 0.6069,
"step": 1113
},
{
"epoch": 0.4481995574331121,
"grad_norm": 4.316287155035261,
"learning_rate": 3.3746322945278543e-06,
"loss": 0.66,
"step": 1114
},
{
"epoch": 0.44860189096761216,
"grad_norm": 4.697629623928124,
"learning_rate": 3.3713423645057873e-06,
"loss": 0.4885,
"step": 1115
},
{
"epoch": 0.44900422450211225,
"grad_norm": 3.920242841924085,
"learning_rate": 3.368050715956025e-06,
"loss": 0.5213,
"step": 1116
},
{
"epoch": 0.44940655803661234,
"grad_norm": 4.521566910606868,
"learning_rate": 3.3647573553706052e-06,
"loss": 0.5743,
"step": 1117
},
{
"epoch": 0.4498088915711125,
"grad_norm": 4.082210637179168,
"learning_rate": 3.3614622892449465e-06,
"loss": 0.5361,
"step": 1118
},
{
"epoch": 0.45021122510561257,
"grad_norm": 4.017855379880608,
"learning_rate": 3.3581655240778277e-06,
"loss": 0.5492,
"step": 1119
},
{
"epoch": 0.45061355864011265,
"grad_norm": 3.3069156613524577,
"learning_rate": 3.35486706637138e-06,
"loss": 0.4142,
"step": 1120
},
{
"epoch": 0.45101589217461274,
"grad_norm": 4.230766496789279,
"learning_rate": 3.3515669226310716e-06,
"loss": 0.5883,
"step": 1121
},
{
"epoch": 0.45141822570911283,
"grad_norm": 3.932193721134766,
"learning_rate": 3.3482650993656983e-06,
"loss": 0.4732,
"step": 1122
},
{
"epoch": 0.451820559243613,
"grad_norm": 3.9297564952360577,
"learning_rate": 3.344961603087367e-06,
"loss": 0.5852,
"step": 1123
},
{
"epoch": 0.45222289277811306,
"grad_norm": 4.909401223605511,
"learning_rate": 3.341656440311484e-06,
"loss": 0.5709,
"step": 1124
},
{
"epoch": 0.45262522631261315,
"grad_norm": 3.6889862040375525,
"learning_rate": 3.3383496175567435e-06,
"loss": 0.5966,
"step": 1125
},
{
"epoch": 0.45302755984711324,
"grad_norm": 3.4699341807873694,
"learning_rate": 3.3350411413451125e-06,
"loss": 0.4692,
"step": 1126
},
{
"epoch": 0.4534298933816134,
"grad_norm": 4.481843121210536,
"learning_rate": 3.3317310182018186e-06,
"loss": 0.6374,
"step": 1127
},
{
"epoch": 0.45383222691611347,
"grad_norm": 4.674277827776204,
"learning_rate": 3.328419254655339e-06,
"loss": 0.7174,
"step": 1128
},
{
"epoch": 0.45423456045061356,
"grad_norm": 3.8262305960434575,
"learning_rate": 3.325105857237386e-06,
"loss": 0.4618,
"step": 1129
},
{
"epoch": 0.45463689398511364,
"grad_norm": 3.9899678016482305,
"learning_rate": 3.3217908324828942e-06,
"loss": 0.5839,
"step": 1130
},
{
"epoch": 0.4550392275196138,
"grad_norm": 3.39198458807276,
"learning_rate": 3.3184741869300073e-06,
"loss": 0.5191,
"step": 1131
},
{
"epoch": 0.4554415610541139,
"grad_norm": 4.032286065452944,
"learning_rate": 3.3151559271200656e-06,
"loss": 0.4835,
"step": 1132
},
{
"epoch": 0.45584389458861396,
"grad_norm": 4.381591989574539,
"learning_rate": 3.3118360595975947e-06,
"loss": 0.5592,
"step": 1133
},
{
"epoch": 0.45624622812311405,
"grad_norm": 4.022213508423109,
"learning_rate": 3.3085145909102885e-06,
"loss": 0.5304,
"step": 1134
},
{
"epoch": 0.45664856165761414,
"grad_norm": 3.834367329162151,
"learning_rate": 3.305191527609e-06,
"loss": 0.4678,
"step": 1135
},
{
"epoch": 0.4570508951921143,
"grad_norm": 4.448203020700801,
"learning_rate": 3.3018668762477294e-06,
"loss": 0.5172,
"step": 1136
},
{
"epoch": 0.45745322872661437,
"grad_norm": 4.128965375687235,
"learning_rate": 3.2985406433836055e-06,
"loss": 0.4742,
"step": 1137
},
{
"epoch": 0.45785556226111446,
"grad_norm": 4.129504251009841,
"learning_rate": 3.295212835576878e-06,
"loss": 0.5873,
"step": 1138
},
{
"epoch": 0.45825789579561454,
"grad_norm": 4.422224105540513,
"learning_rate": 3.2918834593909028e-06,
"loss": 0.6049,
"step": 1139
},
{
"epoch": 0.4586602293301147,
"grad_norm": 4.434937349809798,
"learning_rate": 3.2885525213921286e-06,
"loss": 0.5708,
"step": 1140
},
{
"epoch": 0.4590625628646148,
"grad_norm": 3.668465429492197,
"learning_rate": 3.2852200281500855e-06,
"loss": 0.4559,
"step": 1141
},
{
"epoch": 0.45946489639911486,
"grad_norm": 4.232532968956556,
"learning_rate": 3.2818859862373696e-06,
"loss": 0.5721,
"step": 1142
},
{
"epoch": 0.45986722993361495,
"grad_norm": 4.7338258325238085,
"learning_rate": 3.2785504022296316e-06,
"loss": 0.6157,
"step": 1143
},
{
"epoch": 0.4602695634681151,
"grad_norm": 3.7791426452155723,
"learning_rate": 3.2752132827055646e-06,
"loss": 0.5458,
"step": 1144
},
{
"epoch": 0.4606718970026152,
"grad_norm": 4.1019637445172314,
"learning_rate": 3.2718746342468894e-06,
"loss": 0.5972,
"step": 1145
},
{
"epoch": 0.46107423053711527,
"grad_norm": 5.277495798462077,
"learning_rate": 3.268534463438342e-06,
"loss": 0.8266,
"step": 1146
},
{
"epoch": 0.46147656407161536,
"grad_norm": 4.184453273932426,
"learning_rate": 3.2651927768676626e-06,
"loss": 0.4562,
"step": 1147
},
{
"epoch": 0.46187889760611545,
"grad_norm": 4.45907273350243,
"learning_rate": 3.261849581125579e-06,
"loss": 0.6532,
"step": 1148
},
{
"epoch": 0.4622812311406156,
"grad_norm": 3.965076229787784,
"learning_rate": 3.258504882805796e-06,
"loss": 0.5247,
"step": 1149
},
{
"epoch": 0.4626835646751157,
"grad_norm": 4.461055711184354,
"learning_rate": 3.2551586885049814e-06,
"loss": 0.4588,
"step": 1150
},
{
"epoch": 0.46308589820961576,
"grad_norm": 4.3124706302870255,
"learning_rate": 3.2518110048227553e-06,
"loss": 0.7004,
"step": 1151
},
{
"epoch": 0.46348823174411585,
"grad_norm": 4.326890541788323,
"learning_rate": 3.2484618383616735e-06,
"loss": 0.5989,
"step": 1152
},
{
"epoch": 0.463890565278616,
"grad_norm": 4.323105374837098,
"learning_rate": 3.2451111957272173e-06,
"loss": 0.5451,
"step": 1153
},
{
"epoch": 0.4642928988131161,
"grad_norm": 4.134824677878268,
"learning_rate": 3.241759083527779e-06,
"loss": 0.5589,
"step": 1154
},
{
"epoch": 0.46469523234761617,
"grad_norm": 4.0981640141993365,
"learning_rate": 3.238405508374649e-06,
"loss": 0.5771,
"step": 1155
},
{
"epoch": 0.46509756588211626,
"grad_norm": 3.7455178072585644,
"learning_rate": 3.2350504768820036e-06,
"loss": 0.5125,
"step": 1156
},
{
"epoch": 0.46549989941661635,
"grad_norm": 3.72359460364635,
"learning_rate": 3.23169399566689e-06,
"loss": 0.4213,
"step": 1157
},
{
"epoch": 0.4659022329511165,
"grad_norm": 3.509972924609966,
"learning_rate": 3.228336071349218e-06,
"loss": 0.4757,
"step": 1158
},
{
"epoch": 0.4663045664856166,
"grad_norm": 4.028400275439829,
"learning_rate": 3.2249767105517395e-06,
"loss": 0.5022,
"step": 1159
},
{
"epoch": 0.46670690002011667,
"grad_norm": 4.150972255373417,
"learning_rate": 3.221615919900042e-06,
"loss": 0.6257,
"step": 1160
},
{
"epoch": 0.46710923355461675,
"grad_norm": 4.494189407683215,
"learning_rate": 3.218253706022533e-06,
"loss": 0.5914,
"step": 1161
},
{
"epoch": 0.4675115670891169,
"grad_norm": 3.558180537366331,
"learning_rate": 3.2148900755504253e-06,
"loss": 0.5102,
"step": 1162
},
{
"epoch": 0.467913900623617,
"grad_norm": 3.438406963531958,
"learning_rate": 3.2115250351177274e-06,
"loss": 0.515,
"step": 1163
},
{
"epoch": 0.4683162341581171,
"grad_norm": 4.323925078750027,
"learning_rate": 3.208158591361228e-06,
"loss": 0.5109,
"step": 1164
},
{
"epoch": 0.46871856769261716,
"grad_norm": 4.1534707963201285,
"learning_rate": 3.204790750920484e-06,
"loss": 0.5736,
"step": 1165
},
{
"epoch": 0.4691209012271173,
"grad_norm": 3.6871720181954406,
"learning_rate": 3.2014215204378064e-06,
"loss": 0.5765,
"step": 1166
},
{
"epoch": 0.4695232347616174,
"grad_norm": 3.9220478852176264,
"learning_rate": 3.1980509065582476e-06,
"loss": 0.5545,
"step": 1167
},
{
"epoch": 0.4699255682961175,
"grad_norm": 4.603512689826395,
"learning_rate": 3.194678915929589e-06,
"loss": 0.5587,
"step": 1168
},
{
"epoch": 0.47032790183061757,
"grad_norm": 4.0356910482849075,
"learning_rate": 3.1913055552023263e-06,
"loss": 0.4642,
"step": 1169
},
{
"epoch": 0.47073023536511766,
"grad_norm": 4.251943286279792,
"learning_rate": 3.1879308310296586e-06,
"loss": 0.5496,
"step": 1170
},
{
"epoch": 0.4711325688996178,
"grad_norm": 4.189042957707438,
"learning_rate": 3.1845547500674737e-06,
"loss": 0.5646,
"step": 1171
},
{
"epoch": 0.4715349024341179,
"grad_norm": 3.3326084801089353,
"learning_rate": 3.1811773189743355e-06,
"loss": 0.4837,
"step": 1172
},
{
"epoch": 0.471937235968618,
"grad_norm": 4.523273563587766,
"learning_rate": 3.17779854441147e-06,
"loss": 0.6807,
"step": 1173
},
{
"epoch": 0.47233956950311806,
"grad_norm": 3.7232216855517315,
"learning_rate": 3.1744184330427543e-06,
"loss": 0.4492,
"step": 1174
},
{
"epoch": 0.4727419030376182,
"grad_norm": 3.8162014027986437,
"learning_rate": 3.1710369915347e-06,
"loss": 0.5626,
"step": 1175
},
{
"epoch": 0.4731442365721183,
"grad_norm": 4.575379469711398,
"learning_rate": 3.1676542265564443e-06,
"loss": 0.6358,
"step": 1176
},
{
"epoch": 0.4735465701066184,
"grad_norm": 4.563047737373793,
"learning_rate": 3.1642701447797324e-06,
"loss": 0.6282,
"step": 1177
},
{
"epoch": 0.47394890364111847,
"grad_norm": 4.254463762469985,
"learning_rate": 3.1608847528789096e-06,
"loss": 0.4974,
"step": 1178
},
{
"epoch": 0.4743512371756186,
"grad_norm": 4.290561801295637,
"learning_rate": 3.1574980575309015e-06,
"loss": 0.6116,
"step": 1179
},
{
"epoch": 0.4747535707101187,
"grad_norm": 4.276362615412023,
"learning_rate": 3.154110065415208e-06,
"loss": 0.5274,
"step": 1180
},
{
"epoch": 0.4751559042446188,
"grad_norm": 3.8389354369931374,
"learning_rate": 3.150720783213883e-06,
"loss": 0.5338,
"step": 1181
},
{
"epoch": 0.4755582377791189,
"grad_norm": 4.448041403696648,
"learning_rate": 3.1473302176115283e-06,
"loss": 0.5265,
"step": 1182
},
{
"epoch": 0.47596057131361896,
"grad_norm": 4.46050378901124,
"learning_rate": 3.1439383752952753e-06,
"loss": 0.6299,
"step": 1183
},
{
"epoch": 0.4763629048481191,
"grad_norm": 4.111077775261664,
"learning_rate": 3.140545262954772e-06,
"loss": 0.5723,
"step": 1184
},
{
"epoch": 0.4767652383826192,
"grad_norm": 3.678492409328739,
"learning_rate": 3.137150887282174e-06,
"loss": 0.4178,
"step": 1185
},
{
"epoch": 0.4771675719171193,
"grad_norm": 3.661689522078919,
"learning_rate": 3.1337552549721262e-06,
"loss": 0.4658,
"step": 1186
},
{
"epoch": 0.47756990545161937,
"grad_norm": 4.250351645694971,
"learning_rate": 3.130358372721754e-06,
"loss": 0.5614,
"step": 1187
},
{
"epoch": 0.4779722389861195,
"grad_norm": 3.599583256588181,
"learning_rate": 3.126960247230646e-06,
"loss": 0.4569,
"step": 1188
},
{
"epoch": 0.4783745725206196,
"grad_norm": 3.821896657809243,
"learning_rate": 3.1235608852008446e-06,
"loss": 0.4834,
"step": 1189
},
{
"epoch": 0.4787769060551197,
"grad_norm": 3.9335780159969906,
"learning_rate": 3.1201602933368308e-06,
"loss": 0.5152,
"step": 1190
},
{
"epoch": 0.4791792395896198,
"grad_norm": 4.770540317196732,
"learning_rate": 3.116758478345509e-06,
"loss": 0.5948,
"step": 1191
},
{
"epoch": 0.4795815731241199,
"grad_norm": 3.998481220511283,
"learning_rate": 3.1133554469361976e-06,
"loss": 0.4881,
"step": 1192
},
{
"epoch": 0.47998390665862,
"grad_norm": 4.208648289802026,
"learning_rate": 3.109951205820615e-06,
"loss": 0.4933,
"step": 1193
},
{
"epoch": 0.4803862401931201,
"grad_norm": 3.752607346638518,
"learning_rate": 3.1065457617128635e-06,
"loss": 0.547,
"step": 1194
},
{
"epoch": 0.4807885737276202,
"grad_norm": 4.151251812002077,
"learning_rate": 3.1031391213294203e-06,
"loss": 0.5015,
"step": 1195
},
{
"epoch": 0.48119090726212027,
"grad_norm": 4.7037289072786095,
"learning_rate": 3.0997312913891196e-06,
"loss": 0.6795,
"step": 1196
},
{
"epoch": 0.4815932407966204,
"grad_norm": 4.364147422750159,
"learning_rate": 3.0963222786131435e-06,
"loss": 0.5422,
"step": 1197
},
{
"epoch": 0.4819955743311205,
"grad_norm": 3.301076213262608,
"learning_rate": 3.0929120897250066e-06,
"loss": 0.4646,
"step": 1198
},
{
"epoch": 0.4823979078656206,
"grad_norm": 3.803770968997955,
"learning_rate": 3.0895007314505415e-06,
"loss": 0.4579,
"step": 1199
},
{
"epoch": 0.4828002414001207,
"grad_norm": 3.6319529462360816,
"learning_rate": 3.0860882105178897e-06,
"loss": 0.5152,
"step": 1200
},
{
"epoch": 0.4832025749346208,
"grad_norm": 3.7950010562493453,
"learning_rate": 3.082674533657484e-06,
"loss": 0.588,
"step": 1201
},
{
"epoch": 0.4836049084691209,
"grad_norm": 3.9284108228666796,
"learning_rate": 3.079259707602038e-06,
"loss": 0.5565,
"step": 1202
},
{
"epoch": 0.484007242003621,
"grad_norm": 4.764319789283065,
"learning_rate": 3.0758437390865314e-06,
"loss": 0.6115,
"step": 1203
},
{
"epoch": 0.4844095755381211,
"grad_norm": 3.781006793375389,
"learning_rate": 3.0724266348481966e-06,
"loss": 0.5239,
"step": 1204
},
{
"epoch": 0.4848119090726212,
"grad_norm": 4.259150542468107,
"learning_rate": 3.069008401626507e-06,
"loss": 0.5505,
"step": 1205
},
{
"epoch": 0.4852142426071213,
"grad_norm": 4.222901377467788,
"learning_rate": 3.065589046163162e-06,
"loss": 0.5425,
"step": 1206
},
{
"epoch": 0.4856165761416214,
"grad_norm": 3.783266465549031,
"learning_rate": 3.062168575202075e-06,
"loss": 0.542,
"step": 1207
},
{
"epoch": 0.4860189096761215,
"grad_norm": 4.428497795054238,
"learning_rate": 3.0587469954893584e-06,
"loss": 0.6567,
"step": 1208
},
{
"epoch": 0.4864212432106216,
"grad_norm": 3.691837740063112,
"learning_rate": 3.0553243137733125e-06,
"loss": 0.4558,
"step": 1209
},
{
"epoch": 0.4868235767451217,
"grad_norm": 4.073296743584824,
"learning_rate": 3.051900536804411e-06,
"loss": 0.5664,
"step": 1210
},
{
"epoch": 0.4872259102796218,
"grad_norm": 3.7192683213034234,
"learning_rate": 3.0484756713352858e-06,
"loss": 0.494,
"step": 1211
},
{
"epoch": 0.4876282438141219,
"grad_norm": 3.6377595381026557,
"learning_rate": 3.045049724120718e-06,
"loss": 0.4885,
"step": 1212
},
{
"epoch": 0.488030577348622,
"grad_norm": 4.338683717827157,
"learning_rate": 3.0416227019176208e-06,
"loss": 0.6165,
"step": 1213
},
{
"epoch": 0.48843291088312213,
"grad_norm": 4.374868800699678,
"learning_rate": 3.0381946114850293e-06,
"loss": 0.5755,
"step": 1214
},
{
"epoch": 0.4888352444176222,
"grad_norm": 4.54077234147531,
"learning_rate": 3.034765459584083e-06,
"loss": 0.5954,
"step": 1215
},
{
"epoch": 0.4892375779521223,
"grad_norm": 4.166268412885333,
"learning_rate": 3.031335252978017e-06,
"loss": 0.6657,
"step": 1216
},
{
"epoch": 0.4896399114866224,
"grad_norm": 3.675310094808539,
"learning_rate": 3.027903998432145e-06,
"loss": 0.4676,
"step": 1217
},
{
"epoch": 0.49004224502112254,
"grad_norm": 4.091571307318168,
"learning_rate": 3.0244717027138494e-06,
"loss": 0.579,
"step": 1218
},
{
"epoch": 0.4904445785556226,
"grad_norm": 4.2620047470347595,
"learning_rate": 3.021038372592563e-06,
"loss": 0.577,
"step": 1219
},
{
"epoch": 0.4908469120901227,
"grad_norm": 3.687909147915886,
"learning_rate": 3.017604014839762e-06,
"loss": 0.557,
"step": 1220
},
{
"epoch": 0.4912492456246228,
"grad_norm": 4.003092692480048,
"learning_rate": 3.014168636228948e-06,
"loss": 0.5517,
"step": 1221
},
{
"epoch": 0.4916515791591229,
"grad_norm": 4.349298145062102,
"learning_rate": 3.010732243535636e-06,
"loss": 0.5991,
"step": 1222
},
{
"epoch": 0.49205391269362303,
"grad_norm": 3.9645619754575963,
"learning_rate": 3.0072948435373408e-06,
"loss": 0.5145,
"step": 1223
},
{
"epoch": 0.4924562462281231,
"grad_norm": 4.216377733109406,
"learning_rate": 3.003856443013564e-06,
"loss": 0.6017,
"step": 1224
},
{
"epoch": 0.4928585797626232,
"grad_norm": 3.9206543304610584,
"learning_rate": 3.0004170487457813e-06,
"loss": 0.5468,
"step": 1225
},
{
"epoch": 0.4932609132971233,
"grad_norm": 4.196664240626691,
"learning_rate": 2.996976667517427e-06,
"loss": 0.5191,
"step": 1226
},
{
"epoch": 0.49366324683162344,
"grad_norm": 3.7775234111981493,
"learning_rate": 2.9935353061138833e-06,
"loss": 0.5094,
"step": 1227
},
{
"epoch": 0.4940655803661235,
"grad_norm": 4.412452469732304,
"learning_rate": 2.990092971322464e-06,
"loss": 0.566,
"step": 1228
},
{
"epoch": 0.4944679139006236,
"grad_norm": 3.967665512410053,
"learning_rate": 2.986649669932405e-06,
"loss": 0.5005,
"step": 1229
},
{
"epoch": 0.4948702474351237,
"grad_norm": 3.8061725506913353,
"learning_rate": 2.983205408734846e-06,
"loss": 0.5214,
"step": 1230
},
{
"epoch": 0.49527258096962384,
"grad_norm": 3.615684737615807,
"learning_rate": 2.9797601945228214e-06,
"loss": 0.4647,
"step": 1231
},
{
"epoch": 0.49567491450412393,
"grad_norm": 4.27866020604673,
"learning_rate": 2.9763140340912463e-06,
"loss": 0.6409,
"step": 1232
},
{
"epoch": 0.496077248038624,
"grad_norm": 3.596076423790147,
"learning_rate": 2.972866934236898e-06,
"loss": 0.412,
"step": 1233
},
{
"epoch": 0.4964795815731241,
"grad_norm": 4.0355620490752955,
"learning_rate": 2.9694189017584097e-06,
"loss": 0.5596,
"step": 1234
},
{
"epoch": 0.4968819151076242,
"grad_norm": 4.148051125186017,
"learning_rate": 2.965969943456254e-06,
"loss": 0.5191,
"step": 1235
},
{
"epoch": 0.49728424864212434,
"grad_norm": 4.2692044767988,
"learning_rate": 2.962520066132728e-06,
"loss": 0.5066,
"step": 1236
},
{
"epoch": 0.4976865821766244,
"grad_norm": 4.28381423937242,
"learning_rate": 2.9590692765919436e-06,
"loss": 0.598,
"step": 1237
},
{
"epoch": 0.4980889157111245,
"grad_norm": 4.175358078657294,
"learning_rate": 2.955617581639809e-06,
"loss": 0.5953,
"step": 1238
},
{
"epoch": 0.4984912492456246,
"grad_norm": 4.5191335184254955,
"learning_rate": 2.952164988084021e-06,
"loss": 0.6236,
"step": 1239
},
{
"epoch": 0.49889358278012474,
"grad_norm": 4.126996114146587,
"learning_rate": 2.9487115027340458e-06,
"loss": 0.5455,
"step": 1240
},
{
"epoch": 0.49929591631462483,
"grad_norm": 3.978032878939077,
"learning_rate": 2.9452571324011103e-06,
"loss": 0.5658,
"step": 1241
},
{
"epoch": 0.4996982498491249,
"grad_norm": 4.494778767380183,
"learning_rate": 2.941801883898187e-06,
"loss": 0.6639,
"step": 1242
},
{
"epoch": 0.5001005833836251,
"grad_norm": 3.7291179386624376,
"learning_rate": 2.9383457640399794e-06,
"loss": 0.4724,
"step": 1243
},
{
"epoch": 0.5005029169181251,
"grad_norm": 4.0158904670389175,
"learning_rate": 2.9348887796429104e-06,
"loss": 0.5611,
"step": 1244
},
{
"epoch": 0.5009052504526252,
"grad_norm": 3.93870398764487,
"learning_rate": 2.9314309375251065e-06,
"loss": 0.5596,
"step": 1245
},
{
"epoch": 0.5013075839871253,
"grad_norm": 4.111802544399985,
"learning_rate": 2.9279722445063883e-06,
"loss": 0.5872,
"step": 1246
},
{
"epoch": 0.5017099175216254,
"grad_norm": 3.948101448350297,
"learning_rate": 2.9245127074082535e-06,
"loss": 0.5131,
"step": 1247
},
{
"epoch": 0.5021122510561256,
"grad_norm": 4.345129844652909,
"learning_rate": 2.9210523330538625e-06,
"loss": 0.6389,
"step": 1248
},
{
"epoch": 0.5025145845906256,
"grad_norm": 3.6367326824064112,
"learning_rate": 2.9175911282680307e-06,
"loss": 0.4518,
"step": 1249
},
{
"epoch": 0.5029169181251257,
"grad_norm": 4.027476673290246,
"learning_rate": 2.914129099877208e-06,
"loss": 0.5234,
"step": 1250
},
{
"epoch": 0.5033192516596259,
"grad_norm": 3.450546244138055,
"learning_rate": 2.9106662547094724e-06,
"loss": 0.3807,
"step": 1251
},
{
"epoch": 0.5037215851941259,
"grad_norm": 3.696528336339893,
"learning_rate": 2.907202599594508e-06,
"loss": 0.3884,
"step": 1252
},
{
"epoch": 0.504123918728626,
"grad_norm": 4.252344685184106,
"learning_rate": 2.9037381413636e-06,
"loss": 0.6935,
"step": 1253
},
{
"epoch": 0.5045262522631261,
"grad_norm": 4.336362311318111,
"learning_rate": 2.900272886849616e-06,
"loss": 0.5331,
"step": 1254
},
{
"epoch": 0.5049285857976262,
"grad_norm": 3.6842567382976688,
"learning_rate": 2.8968068428869955e-06,
"loss": 0.5545,
"step": 1255
},
{
"epoch": 0.5053309193321264,
"grad_norm": 3.699409904804457,
"learning_rate": 2.8933400163117337e-06,
"loss": 0.3754,
"step": 1256
},
{
"epoch": 0.5057332528666264,
"grad_norm": 4.205702275086939,
"learning_rate": 2.889872413961369e-06,
"loss": 0.6633,
"step": 1257
},
{
"epoch": 0.5061355864011265,
"grad_norm": 3.6639853190462626,
"learning_rate": 2.886404042674972e-06,
"loss": 0.3892,
"step": 1258
},
{
"epoch": 0.5065379199356266,
"grad_norm": 3.5640332071999765,
"learning_rate": 2.882934909293127e-06,
"loss": 0.4084,
"step": 1259
},
{
"epoch": 0.5069402534701267,
"grad_norm": 4.0471975082058025,
"learning_rate": 2.8794650206579234e-06,
"loss": 0.5812,
"step": 1260
},
{
"epoch": 0.5073425870046269,
"grad_norm": 4.0654643808596616,
"learning_rate": 2.875994383612939e-06,
"loss": 0.568,
"step": 1261
},
{
"epoch": 0.5077449205391269,
"grad_norm": 4.330240433028303,
"learning_rate": 2.872523005003229e-06,
"loss": 0.5165,
"step": 1262
},
{
"epoch": 0.508147254073627,
"grad_norm": 4.015050113056623,
"learning_rate": 2.869050891675309e-06,
"loss": 0.5002,
"step": 1263
},
{
"epoch": 0.5085495876081272,
"grad_norm": 4.393983632365189,
"learning_rate": 2.8655780504771476e-06,
"loss": 0.573,
"step": 1264
},
{
"epoch": 0.5089519211426272,
"grad_norm": 3.193554396038774,
"learning_rate": 2.8621044882581433e-06,
"loss": 0.3709,
"step": 1265
},
{
"epoch": 0.5093542546771274,
"grad_norm": 3.3841882285013813,
"learning_rate": 2.8586302118691223e-06,
"loss": 0.4387,
"step": 1266
},
{
"epoch": 0.5097565882116274,
"grad_norm": 3.9970258932238742,
"learning_rate": 2.8551552281623157e-06,
"loss": 0.5225,
"step": 1267
},
{
"epoch": 0.5101589217461275,
"grad_norm": 4.513606814044003,
"learning_rate": 2.85167954399135e-06,
"loss": 0.5528,
"step": 1268
},
{
"epoch": 0.5105612552806277,
"grad_norm": 4.684727392976221,
"learning_rate": 2.848203166211235e-06,
"loss": 0.6793,
"step": 1269
},
{
"epoch": 0.5109635888151277,
"grad_norm": 4.222583621631339,
"learning_rate": 2.8447261016783474e-06,
"loss": 0.6053,
"step": 1270
},
{
"epoch": 0.5113659223496279,
"grad_norm": 3.8724976223748744,
"learning_rate": 2.841248357250418e-06,
"loss": 0.4797,
"step": 1271
},
{
"epoch": 0.5117682558841279,
"grad_norm": 3.5925982833624652,
"learning_rate": 2.8377699397865184e-06,
"loss": 0.6524,
"step": 1272
},
{
"epoch": 0.512170589418628,
"grad_norm": 4.4527497576787285,
"learning_rate": 2.8342908561470496e-06,
"loss": 0.6426,
"step": 1273
},
{
"epoch": 0.5125729229531282,
"grad_norm": 3.758694762638942,
"learning_rate": 2.8308111131937237e-06,
"loss": 0.4863,
"step": 1274
},
{
"epoch": 0.5129752564876282,
"grad_norm": 3.432356742958401,
"learning_rate": 2.827330717789555e-06,
"loss": 0.5223,
"step": 1275
},
{
"epoch": 0.5133775900221283,
"grad_norm": 4.33004996309828,
"learning_rate": 2.8238496767988433e-06,
"loss": 0.6004,
"step": 1276
},
{
"epoch": 0.5137799235566285,
"grad_norm": 4.200608717420535,
"learning_rate": 2.820367997087164e-06,
"loss": 0.6228,
"step": 1277
},
{
"epoch": 0.5141822570911285,
"grad_norm": 4.0453448815375035,
"learning_rate": 2.8168856855213493e-06,
"loss": 0.6242,
"step": 1278
},
{
"epoch": 0.5145845906256287,
"grad_norm": 4.227997150733838,
"learning_rate": 2.81340274896948e-06,
"loss": 0.6175,
"step": 1279
},
{
"epoch": 0.5149869241601287,
"grad_norm": 3.68997651052952,
"learning_rate": 2.8099191943008673e-06,
"loss": 0.3891,
"step": 1280
},
{
"epoch": 0.5153892576946288,
"grad_norm": 3.983876769382806,
"learning_rate": 2.806435028386044e-06,
"loss": 0.4676,
"step": 1281
},
{
"epoch": 0.515791591229129,
"grad_norm": 3.7821657559918362,
"learning_rate": 2.8029502580967467e-06,
"loss": 0.5712,
"step": 1282
},
{
"epoch": 0.516193924763629,
"grad_norm": 4.613891160258615,
"learning_rate": 2.7994648903059034e-06,
"loss": 0.6892,
"step": 1283
},
{
"epoch": 0.5165962582981292,
"grad_norm": 4.260048115855587,
"learning_rate": 2.7959789318876235e-06,
"loss": 0.587,
"step": 1284
},
{
"epoch": 0.5169985918326292,
"grad_norm": 3.2651956443139905,
"learning_rate": 2.792492389717178e-06,
"loss": 0.4202,
"step": 1285
},
{
"epoch": 0.5174009253671293,
"grad_norm": 3.8441567338972136,
"learning_rate": 2.7890052706709913e-06,
"loss": 0.5274,
"step": 1286
},
{
"epoch": 0.5178032589016295,
"grad_norm": 4.244436141220556,
"learning_rate": 2.785517581626624e-06,
"loss": 0.5557,
"step": 1287
},
{
"epoch": 0.5182055924361295,
"grad_norm": 3.777185766227343,
"learning_rate": 2.782029329462763e-06,
"loss": 0.5508,
"step": 1288
},
{
"epoch": 0.5186079259706297,
"grad_norm": 4.024624316501224,
"learning_rate": 2.778540521059204e-06,
"loss": 0.4091,
"step": 1289
},
{
"epoch": 0.5190102595051298,
"grad_norm": 3.429656579938711,
"learning_rate": 2.775051163296839e-06,
"loss": 0.373,
"step": 1290
},
{
"epoch": 0.5194125930396298,
"grad_norm": 4.115516308278128,
"learning_rate": 2.771561263057647e-06,
"loss": 0.6296,
"step": 1291
},
{
"epoch": 0.51981492657413,
"grad_norm": 4.5340536303593,
"learning_rate": 2.7680708272246738e-06,
"loss": 0.5478,
"step": 1292
},
{
"epoch": 0.52021726010863,
"grad_norm": 4.1573055283514675,
"learning_rate": 2.7645798626820226e-06,
"loss": 0.5928,
"step": 1293
},
{
"epoch": 0.5206195936431302,
"grad_norm": 3.5540381627057926,
"learning_rate": 2.7610883763148382e-06,
"loss": 0.4204,
"step": 1294
},
{
"epoch": 0.5210219271776303,
"grad_norm": 3.7082383118644406,
"learning_rate": 2.7575963750092976e-06,
"loss": 0.5897,
"step": 1295
},
{
"epoch": 0.5214242607121303,
"grad_norm": 3.670768517938044,
"learning_rate": 2.7541038656525906e-06,
"loss": 0.4524,
"step": 1296
},
{
"epoch": 0.5218265942466305,
"grad_norm": 3.9599618494551336,
"learning_rate": 2.7506108551329084e-06,
"loss": 0.5334,
"step": 1297
},
{
"epoch": 0.5222289277811305,
"grad_norm": 3.8332760152691425,
"learning_rate": 2.747117350339434e-06,
"loss": 0.5027,
"step": 1298
},
{
"epoch": 0.5226312613156306,
"grad_norm": 3.5387447785248463,
"learning_rate": 2.743623358162322e-06,
"loss": 0.4453,
"step": 1299
},
{
"epoch": 0.5230335948501308,
"grad_norm": 3.665534245478194,
"learning_rate": 2.74012888549269e-06,
"loss": 0.4541,
"step": 1300
},
{
"epoch": 0.5234359283846308,
"grad_norm": 4.118363799618076,
"learning_rate": 2.736633939222604e-06,
"loss": 0.4863,
"step": 1301
},
{
"epoch": 0.523838261919131,
"grad_norm": 4.07159865432201,
"learning_rate": 2.733138526245061e-06,
"loss": 0.5836,
"step": 1302
},
{
"epoch": 0.5242405954536311,
"grad_norm": 4.9413620392110955,
"learning_rate": 2.729642653453981e-06,
"loss": 0.6331,
"step": 1303
},
{
"epoch": 0.5246429289881311,
"grad_norm": 4.349366316003809,
"learning_rate": 2.7261463277441913e-06,
"loss": 0.6584,
"step": 1304
},
{
"epoch": 0.5250452625226313,
"grad_norm": 4.013818084372327,
"learning_rate": 2.7226495560114097e-06,
"loss": 0.6444,
"step": 1305
},
{
"epoch": 0.5254475960571313,
"grad_norm": 3.9731448823583158,
"learning_rate": 2.719152345152237e-06,
"loss": 0.4717,
"step": 1306
},
{
"epoch": 0.5258499295916315,
"grad_norm": 3.836192951376196,
"learning_rate": 2.7156547020641377e-06,
"loss": 0.4813,
"step": 1307
},
{
"epoch": 0.5262522631261316,
"grad_norm": 4.526603353793837,
"learning_rate": 2.7121566336454295e-06,
"loss": 0.6223,
"step": 1308
},
{
"epoch": 0.5266545966606316,
"grad_norm": 4.542236066581728,
"learning_rate": 2.70865814679527e-06,
"loss": 0.541,
"step": 1309
},
{
"epoch": 0.5270569301951318,
"grad_norm": 4.477320709723044,
"learning_rate": 2.70515924841364e-06,
"loss": 0.4898,
"step": 1310
},
{
"epoch": 0.5274592637296318,
"grad_norm": 4.188969544607867,
"learning_rate": 2.7016599454013324e-06,
"loss": 0.478,
"step": 1311
},
{
"epoch": 0.527861597264132,
"grad_norm": 3.57921084312417,
"learning_rate": 2.6981602446599408e-06,
"loss": 0.4797,
"step": 1312
},
{
"epoch": 0.5282639307986321,
"grad_norm": 3.4828294543575784,
"learning_rate": 2.69466015309184e-06,
"loss": 0.4464,
"step": 1313
},
{
"epoch": 0.5286662643331321,
"grad_norm": 3.6542821945407002,
"learning_rate": 2.691159677600177e-06,
"loss": 0.5241,
"step": 1314
},
{
"epoch": 0.5290685978676323,
"grad_norm": 4.138582178147402,
"learning_rate": 2.6876588250888556e-06,
"loss": 0.4976,
"step": 1315
},
{
"epoch": 0.5294709314021324,
"grad_norm": 4.156252344667509,
"learning_rate": 2.6841576024625247e-06,
"loss": 0.5346,
"step": 1316
},
{
"epoch": 0.5298732649366324,
"grad_norm": 3.9783539180873118,
"learning_rate": 2.680656016626561e-06,
"loss": 0.5054,
"step": 1317
},
{
"epoch": 0.5302755984711326,
"grad_norm": 4.167749289113631,
"learning_rate": 2.677154074487057e-06,
"loss": 0.5308,
"step": 1318
},
{
"epoch": 0.5306779320056326,
"grad_norm": 3.6243864679157536,
"learning_rate": 2.6736517829508124e-06,
"loss": 0.4969,
"step": 1319
},
{
"epoch": 0.5310802655401328,
"grad_norm": 4.080652153118834,
"learning_rate": 2.6701491489253107e-06,
"loss": 0.5166,
"step": 1320
},
{
"epoch": 0.5314825990746329,
"grad_norm": 4.386897362383017,
"learning_rate": 2.6666461793187143e-06,
"loss": 0.4953,
"step": 1321
},
{
"epoch": 0.5318849326091329,
"grad_norm": 4.286991486124671,
"learning_rate": 2.6631428810398453e-06,
"loss": 0.4804,
"step": 1322
},
{
"epoch": 0.5322872661436331,
"grad_norm": 3.996180824655324,
"learning_rate": 2.659639260998176e-06,
"loss": 0.5182,
"step": 1323
},
{
"epoch": 0.5326895996781331,
"grad_norm": 4.305112080018121,
"learning_rate": 2.6561353261038127e-06,
"loss": 0.6548,
"step": 1324
},
{
"epoch": 0.5330919332126333,
"grad_norm": 3.6190226327616557,
"learning_rate": 2.6526310832674807e-06,
"loss": 0.5383,
"step": 1325
},
{
"epoch": 0.5334942667471334,
"grad_norm": 4.316996697367821,
"learning_rate": 2.6491265394005157e-06,
"loss": 0.5516,
"step": 1326
},
{
"epoch": 0.5338966002816334,
"grad_norm": 4.0648368660213325,
"learning_rate": 2.645621701414845e-06,
"loss": 0.6482,
"step": 1327
},
{
"epoch": 0.5342989338161336,
"grad_norm": 3.7739876249793047,
"learning_rate": 2.6421165762229777e-06,
"loss": 0.4783,
"step": 1328
},
{
"epoch": 0.5347012673506337,
"grad_norm": 3.789987275308021,
"learning_rate": 2.6386111707379873e-06,
"loss": 0.4525,
"step": 1329
},
{
"epoch": 0.5351036008851338,
"grad_norm": 3.630575939542573,
"learning_rate": 2.635105491873502e-06,
"loss": 0.5329,
"step": 1330
},
{
"epoch": 0.5355059344196339,
"grad_norm": 3.961557081061209,
"learning_rate": 2.6315995465436878e-06,
"loss": 0.5432,
"step": 1331
},
{
"epoch": 0.5359082679541339,
"grad_norm": 3.8780055120857204,
"learning_rate": 2.628093341663236e-06,
"loss": 0.454,
"step": 1332
},
{
"epoch": 0.5363106014886341,
"grad_norm": 4.031126634163913,
"learning_rate": 2.6245868841473525e-06,
"loss": 0.5452,
"step": 1333
},
{
"epoch": 0.5367129350231342,
"grad_norm": 3.3788818500452997,
"learning_rate": 2.6210801809117377e-06,
"loss": 0.4364,
"step": 1334
},
{
"epoch": 0.5371152685576343,
"grad_norm": 4.566291916332284,
"learning_rate": 2.61757323887258e-06,
"loss": 0.6427,
"step": 1335
},
{
"epoch": 0.5375176020921344,
"grad_norm": 4.46318346856033,
"learning_rate": 2.614066064946536e-06,
"loss": 0.6142,
"step": 1336
},
{
"epoch": 0.5379199356266344,
"grad_norm": 3.8724650951472377,
"learning_rate": 2.6105586660507214e-06,
"loss": 0.4865,
"step": 1337
},
{
"epoch": 0.5383222691611346,
"grad_norm": 4.107266394665508,
"learning_rate": 2.607051049102696e-06,
"loss": 0.5618,
"step": 1338
},
{
"epoch": 0.5387246026956347,
"grad_norm": 4.131578269950126,
"learning_rate": 2.6035432210204475e-06,
"loss": 0.6571,
"step": 1339
},
{
"epoch": 0.5391269362301347,
"grad_norm": 4.135237490425349,
"learning_rate": 2.6000351887223817e-06,
"loss": 0.506,
"step": 1340
},
{
"epoch": 0.5395292697646349,
"grad_norm": 4.2799950491642695,
"learning_rate": 2.5965269591273075e-06,
"loss": 0.5054,
"step": 1341
},
{
"epoch": 0.539931603299135,
"grad_norm": 4.111277509443795,
"learning_rate": 2.59301853915442e-06,
"loss": 0.5896,
"step": 1342
},
{
"epoch": 0.5403339368336351,
"grad_norm": 3.8507472387773416,
"learning_rate": 2.5895099357232946e-06,
"loss": 0.4784,
"step": 1343
},
{
"epoch": 0.5407362703681352,
"grad_norm": 4.553167071691355,
"learning_rate": 2.5860011557538644e-06,
"loss": 0.5889,
"step": 1344
},
{
"epoch": 0.5411386039026352,
"grad_norm": 4.8377815361815175,
"learning_rate": 2.5824922061664125e-06,
"loss": 0.7153,
"step": 1345
},
{
"epoch": 0.5415409374371354,
"grad_norm": 4.026163246451428,
"learning_rate": 2.578983093881557e-06,
"loss": 0.5578,
"step": 1346
},
{
"epoch": 0.5419432709716355,
"grad_norm": 4.396870764657521,
"learning_rate": 2.5754738258202345e-06,
"loss": 0.7799,
"step": 1347
},
{
"epoch": 0.5423456045061356,
"grad_norm": 4.349518153263542,
"learning_rate": 2.5719644089036916e-06,
"loss": 0.5919,
"step": 1348
},
{
"epoch": 0.5427479380406357,
"grad_norm": 4.355976508524607,
"learning_rate": 2.568454850053467e-06,
"loss": 0.5903,
"step": 1349
},
{
"epoch": 0.5431502715751357,
"grad_norm": 3.951612074252993,
"learning_rate": 2.5649451561913796e-06,
"loss": 0.5038,
"step": 1350
},
{
"epoch": 0.5435526051096359,
"grad_norm": 4.030293462511148,
"learning_rate": 2.561435334239515e-06,
"loss": 0.5308,
"step": 1351
},
{
"epoch": 0.543954938644136,
"grad_norm": 4.530833691231785,
"learning_rate": 2.55792539112021e-06,
"loss": 0.5977,
"step": 1352
},
{
"epoch": 0.544357272178636,
"grad_norm": 4.1989735383766185,
"learning_rate": 2.5544153337560422e-06,
"loss": 0.5686,
"step": 1353
},
{
"epoch": 0.5447596057131362,
"grad_norm": 4.32105387076282,
"learning_rate": 2.5509051690698137e-06,
"loss": 0.6257,
"step": 1354
},
{
"epoch": 0.5451619392476363,
"grad_norm": 3.34919768311592,
"learning_rate": 2.5473949039845388e-06,
"loss": 0.4131,
"step": 1355
},
{
"epoch": 0.5455642727821364,
"grad_norm": 4.128988896294745,
"learning_rate": 2.5438845454234284e-06,
"loss": 0.6007,
"step": 1356
},
{
"epoch": 0.5459666063166365,
"grad_norm": 4.122788652813818,
"learning_rate": 2.5403741003098793e-06,
"loss": 0.5223,
"step": 1357
},
{
"epoch": 0.5463689398511365,
"grad_norm": 4.420118909526267,
"learning_rate": 2.5368635755674586e-06,
"loss": 0.5506,
"step": 1358
},
{
"epoch": 0.5467712733856367,
"grad_norm": 4.2195230085645345,
"learning_rate": 2.5333529781198895e-06,
"loss": 0.5989,
"step": 1359
},
{
"epoch": 0.5471736069201368,
"grad_norm": 4.20795164503954,
"learning_rate": 2.5298423148910396e-06,
"loss": 0.4462,
"step": 1360
},
{
"epoch": 0.5475759404546369,
"grad_norm": 4.010223353178494,
"learning_rate": 2.5263315928049063e-06,
"loss": 0.5192,
"step": 1361
},
{
"epoch": 0.547978273989137,
"grad_norm": 4.816701841105469,
"learning_rate": 2.5228208187856024e-06,
"loss": 0.6599,
"step": 1362
},
{
"epoch": 0.548380607523637,
"grad_norm": 4.023413458632839,
"learning_rate": 2.5193099997573437e-06,
"loss": 0.6185,
"step": 1363
},
{
"epoch": 0.5487829410581372,
"grad_norm": 3.6516400727809115,
"learning_rate": 2.5157991426444343e-06,
"loss": 0.5859,
"step": 1364
},
{
"epoch": 0.5491852745926373,
"grad_norm": 4.119751084462459,
"learning_rate": 2.5122882543712546e-06,
"loss": 0.5255,
"step": 1365
},
{
"epoch": 0.5495876081271374,
"grad_norm": 3.957835467170672,
"learning_rate": 2.5087773418622447e-06,
"loss": 0.4659,
"step": 1366
},
{
"epoch": 0.5499899416616375,
"grad_norm": 3.500258182406765,
"learning_rate": 2.505266412041893e-06,
"loss": 0.4385,
"step": 1367
},
{
"epoch": 0.5503922751961376,
"grad_norm": 4.041408070624076,
"learning_rate": 2.5017554718347233e-06,
"loss": 0.5367,
"step": 1368
},
{
"epoch": 0.5507946087306377,
"grad_norm": 4.091765389792879,
"learning_rate": 2.4982445281652775e-06,
"loss": 0.5744,
"step": 1369
},
{
"epoch": 0.5511969422651378,
"grad_norm": 4.469717920616285,
"learning_rate": 2.4947335879581073e-06,
"loss": 0.5818,
"step": 1370
},
{
"epoch": 0.5515992757996379,
"grad_norm": 3.9779088164073833,
"learning_rate": 2.4912226581377566e-06,
"loss": 0.5893,
"step": 1371
},
{
"epoch": 0.552001609334138,
"grad_norm": 3.627034776419984,
"learning_rate": 2.4877117456287463e-06,
"loss": 0.4902,
"step": 1372
},
{
"epoch": 0.5524039428686381,
"grad_norm": 3.8562269767576405,
"learning_rate": 2.484200857355566e-06,
"loss": 0.6268,
"step": 1373
},
{
"epoch": 0.5528062764031382,
"grad_norm": 4.192164362531014,
"learning_rate": 2.4806900002426567e-06,
"loss": 0.6121,
"step": 1374
},
{
"epoch": 0.5532086099376383,
"grad_norm": 3.916801705190097,
"learning_rate": 2.477179181214398e-06,
"loss": 0.4926,
"step": 1375
},
{
"epoch": 0.5536109434721384,
"grad_norm": 3.693356946438789,
"learning_rate": 2.473668407195095e-06,
"loss": 0.5189,
"step": 1376
},
{
"epoch": 0.5540132770066385,
"grad_norm": 4.229576715855437,
"learning_rate": 2.4701576851089613e-06,
"loss": 0.6398,
"step": 1377
},
{
"epoch": 0.5544156105411386,
"grad_norm": 3.6167470608574197,
"learning_rate": 2.4666470218801113e-06,
"loss": 0.5175,
"step": 1378
},
{
"epoch": 0.5548179440756387,
"grad_norm": 5.070171032981664,
"learning_rate": 2.463136424432542e-06,
"loss": 0.4988,
"step": 1379
},
{
"epoch": 0.5552202776101388,
"grad_norm": 4.058368062649386,
"learning_rate": 2.4596258996901215e-06,
"loss": 0.4313,
"step": 1380
},
{
"epoch": 0.555622611144639,
"grad_norm": 3.9184186874819464,
"learning_rate": 2.4561154545765724e-06,
"loss": 0.5338,
"step": 1381
},
{
"epoch": 0.556024944679139,
"grad_norm": 3.1244549306950784,
"learning_rate": 2.4526050960154616e-06,
"loss": 0.4289,
"step": 1382
},
{
"epoch": 0.5564272782136391,
"grad_norm": 4.04400253855949,
"learning_rate": 2.4490948309301867e-06,
"loss": 0.5257,
"step": 1383
},
{
"epoch": 0.5568296117481392,
"grad_norm": 3.9303715740181513,
"learning_rate": 2.445584666243959e-06,
"loss": 0.5149,
"step": 1384
},
{
"epoch": 0.5572319452826393,
"grad_norm": 4.395655391522824,
"learning_rate": 2.442074608879791e-06,
"loss": 0.6249,
"step": 1385
},
{
"epoch": 0.5576342788171395,
"grad_norm": 3.8268439749597576,
"learning_rate": 2.438564665760486e-06,
"loss": 0.5613,
"step": 1386
},
{
"epoch": 0.5580366123516395,
"grad_norm": 4.255793020118025,
"learning_rate": 2.4350548438086212e-06,
"loss": 0.6155,
"step": 1387
},
{
"epoch": 0.5584389458861396,
"grad_norm": 3.979739013550483,
"learning_rate": 2.4315451499465336e-06,
"loss": 0.5394,
"step": 1388
},
{
"epoch": 0.5588412794206397,
"grad_norm": 3.781411432288367,
"learning_rate": 2.4280355910963097e-06,
"loss": 0.4507,
"step": 1389
},
{
"epoch": 0.5592436129551398,
"grad_norm": 4.371781717934846,
"learning_rate": 2.4245261741797663e-06,
"loss": 0.5305,
"step": 1390
},
{
"epoch": 0.55964594648964,
"grad_norm": 3.52895283713103,
"learning_rate": 2.421016906118444e-06,
"loss": 0.5194,
"step": 1391
},
{
"epoch": 0.56004828002414,
"grad_norm": 3.872042926144624,
"learning_rate": 2.417507793833587e-06,
"loss": 0.4998,
"step": 1392
},
{
"epoch": 0.5604506135586401,
"grad_norm": 3.391234256928363,
"learning_rate": 2.413998844246136e-06,
"loss": 0.4093,
"step": 1393
},
{
"epoch": 0.5608529470931403,
"grad_norm": 3.9654243758806174,
"learning_rate": 2.4104900642767066e-06,
"loss": 0.4515,
"step": 1394
},
{
"epoch": 0.5612552806276403,
"grad_norm": 3.5879075344069378,
"learning_rate": 2.4069814608455804e-06,
"loss": 0.5207,
"step": 1395
},
{
"epoch": 0.5616576141621404,
"grad_norm": 4.104094820028554,
"learning_rate": 2.4034730408726938e-06,
"loss": 0.5625,
"step": 1396
},
{
"epoch": 0.5620599476966405,
"grad_norm": 3.772298699595217,
"learning_rate": 2.3999648112776183e-06,
"loss": 0.5068,
"step": 1397
},
{
"epoch": 0.5624622812311406,
"grad_norm": 3.688052093132043,
"learning_rate": 2.3964567789795533e-06,
"loss": 0.5042,
"step": 1398
},
{
"epoch": 0.5628646147656408,
"grad_norm": 3.8931716258044498,
"learning_rate": 2.3929489508973047e-06,
"loss": 0.5485,
"step": 1399
},
{
"epoch": 0.5632669483001408,
"grad_norm": 4.224728616738161,
"learning_rate": 2.389441333949279e-06,
"loss": 0.6582,
"step": 1400
},
{
"epoch": 0.5636692818346409,
"grad_norm": 3.8494878180062897,
"learning_rate": 2.3859339350534646e-06,
"loss": 0.5379,
"step": 1401
},
{
"epoch": 0.564071615369141,
"grad_norm": 4.200325419085908,
"learning_rate": 2.382426761127421e-06,
"loss": 0.5997,
"step": 1402
},
{
"epoch": 0.5644739489036411,
"grad_norm": 4.143252972394019,
"learning_rate": 2.378919819088263e-06,
"loss": 0.6195,
"step": 1403
},
{
"epoch": 0.5648762824381413,
"grad_norm": 3.7174533902571145,
"learning_rate": 2.3754131158526483e-06,
"loss": 0.5317,
"step": 1404
},
{
"epoch": 0.5652786159726413,
"grad_norm": 4.0056931921159356,
"learning_rate": 2.371906658336764e-06,
"loss": 0.5232,
"step": 1405
},
{
"epoch": 0.5656809495071414,
"grad_norm": 3.830247393182765,
"learning_rate": 2.368400453456313e-06,
"loss": 0.4898,
"step": 1406
},
{
"epoch": 0.5660832830416416,
"grad_norm": 3.7820185066794787,
"learning_rate": 2.364894508126499e-06,
"loss": 0.5438,
"step": 1407
},
{
"epoch": 0.5664856165761416,
"grad_norm": 4.4898443665627,
"learning_rate": 2.3613888292620135e-06,
"loss": 0.6322,
"step": 1408
},
{
"epoch": 0.5668879501106417,
"grad_norm": 4.157236186991902,
"learning_rate": 2.3578834237770227e-06,
"loss": 0.5978,
"step": 1409
},
{
"epoch": 0.5672902836451418,
"grad_norm": 4.576188931404494,
"learning_rate": 2.354378298585155e-06,
"loss": 0.5774,
"step": 1410
},
{
"epoch": 0.5676926171796419,
"grad_norm": 3.714487248907507,
"learning_rate": 2.3508734605994855e-06,
"loss": 0.4305,
"step": 1411
},
{
"epoch": 0.5680949507141421,
"grad_norm": 3.9270791753173313,
"learning_rate": 2.3473689167325205e-06,
"loss": 0.4912,
"step": 1412
},
{
"epoch": 0.5684972842486421,
"grad_norm": 3.655781362606588,
"learning_rate": 2.3438646738961886e-06,
"loss": 0.5125,
"step": 1413
},
{
"epoch": 0.5688996177831422,
"grad_norm": 4.2097153394323845,
"learning_rate": 2.3403607390018246e-06,
"loss": 0.4273,
"step": 1414
},
{
"epoch": 0.5693019513176423,
"grad_norm": 3.7418997500983,
"learning_rate": 2.3368571189601556e-06,
"loss": 0.4534,
"step": 1415
},
{
"epoch": 0.5697042848521424,
"grad_norm": 3.7385152001909154,
"learning_rate": 2.3333538206812874e-06,
"loss": 0.4461,
"step": 1416
},
{
"epoch": 0.5701066183866426,
"grad_norm": 4.243485694858217,
"learning_rate": 2.32985085107469e-06,
"loss": 0.4943,
"step": 1417
},
{
"epoch": 0.5705089519211426,
"grad_norm": 4.064181274521891,
"learning_rate": 2.3263482170491884e-06,
"loss": 0.6696,
"step": 1418
},
{
"epoch": 0.5709112854556427,
"grad_norm": 4.168819688893492,
"learning_rate": 2.3228459255129425e-06,
"loss": 0.5914,
"step": 1419
},
{
"epoch": 0.5713136189901429,
"grad_norm": 3.299064883781313,
"learning_rate": 2.3193439833734403e-06,
"loss": 0.3987,
"step": 1420
},
{
"epoch": 0.5717159525246429,
"grad_norm": 4.887635853989916,
"learning_rate": 2.3158423975374757e-06,
"loss": 0.4842,
"step": 1421
},
{
"epoch": 0.572118286059143,
"grad_norm": 3.4644648967520806,
"learning_rate": 2.312341174911145e-06,
"loss": 0.4643,
"step": 1422
},
{
"epoch": 0.5725206195936431,
"grad_norm": 3.75941250297541,
"learning_rate": 2.3088403223998235e-06,
"loss": 0.4772,
"step": 1423
},
{
"epoch": 0.5729229531281432,
"grad_norm": 3.4615485090255578,
"learning_rate": 2.3053398469081613e-06,
"loss": 0.5075,
"step": 1424
},
{
"epoch": 0.5733252866626434,
"grad_norm": 4.2227564973481995,
"learning_rate": 2.30183975534006e-06,
"loss": 0.5125,
"step": 1425
},
{
"epoch": 0.5737276201971434,
"grad_norm": 4.090888673139273,
"learning_rate": 2.298340054598668e-06,
"loss": 0.5183,
"step": 1426
},
{
"epoch": 0.5741299537316435,
"grad_norm": 4.548845256935699,
"learning_rate": 2.294840751586361e-06,
"loss": 0.639,
"step": 1427
},
{
"epoch": 0.5745322872661436,
"grad_norm": 3.901263423192965,
"learning_rate": 2.291341853204731e-06,
"loss": 0.623,
"step": 1428
},
{
"epoch": 0.5749346208006437,
"grad_norm": 3.347357658304439,
"learning_rate": 2.287843366354571e-06,
"loss": 0.4009,
"step": 1429
},
{
"epoch": 0.5753369543351439,
"grad_norm": 4.512229941763247,
"learning_rate": 2.284345297935863e-06,
"loss": 0.6336,
"step": 1430
},
{
"epoch": 0.5757392878696439,
"grad_norm": 4.132485335550154,
"learning_rate": 2.2808476548477636e-06,
"loss": 0.5015,
"step": 1431
},
{
"epoch": 0.576141621404144,
"grad_norm": 4.226000440762799,
"learning_rate": 2.2773504439885903e-06,
"loss": 0.5063,
"step": 1432
},
{
"epoch": 0.5765439549386442,
"grad_norm": 4.010604789353783,
"learning_rate": 2.27385367225581e-06,
"loss": 0.5085,
"step": 1433
},
{
"epoch": 0.5769462884731442,
"grad_norm": 3.71838877160847,
"learning_rate": 2.2703573465460194e-06,
"loss": 0.483,
"step": 1434
},
{
"epoch": 0.5773486220076444,
"grad_norm": 3.715720870136391,
"learning_rate": 2.2668614737549395e-06,
"loss": 0.5292,
"step": 1435
},
{
"epoch": 0.5777509555421444,
"grad_norm": 3.7875519229722756,
"learning_rate": 2.263366060777397e-06,
"loss": 0.4504,
"step": 1436
},
{
"epoch": 0.5781532890766445,
"grad_norm": 4.235567100677654,
"learning_rate": 2.25987111450731e-06,
"loss": 0.6075,
"step": 1437
},
{
"epoch": 0.5785556226111447,
"grad_norm": 4.23015897289991,
"learning_rate": 2.256376641837679e-06,
"loss": 0.6036,
"step": 1438
},
{
"epoch": 0.5789579561456447,
"grad_norm": 4.397376552533781,
"learning_rate": 2.2528826496605672e-06,
"loss": 0.6584,
"step": 1439
},
{
"epoch": 0.5793602896801449,
"grad_norm": 3.987430088200103,
"learning_rate": 2.249389144867092e-06,
"loss": 0.5312,
"step": 1440
},
{
"epoch": 0.5797626232146449,
"grad_norm": 4.349402058381357,
"learning_rate": 2.24589613434741e-06,
"loss": 0.4806,
"step": 1441
},
{
"epoch": 0.580164956749145,
"grad_norm": 4.389627692568862,
"learning_rate": 2.2424036249907032e-06,
"loss": 0.649,
"step": 1442
},
{
"epoch": 0.5805672902836452,
"grad_norm": 3.889422512107434,
"learning_rate": 2.238911623685162e-06,
"loss": 0.5553,
"step": 1443
},
{
"epoch": 0.5809696238181452,
"grad_norm": 4.1471977764513674,
"learning_rate": 2.235420137317978e-06,
"loss": 0.5527,
"step": 1444
},
{
"epoch": 0.5813719573526454,
"grad_norm": 3.642246289302764,
"learning_rate": 2.2319291727753266e-06,
"loss": 0.4669,
"step": 1445
},
{
"epoch": 0.5817742908871455,
"grad_norm": 3.7186046973608406,
"learning_rate": 2.228438736942354e-06,
"loss": 0.3984,
"step": 1446
},
{
"epoch": 0.5821766244216455,
"grad_norm": 4.208948818822306,
"learning_rate": 2.2249488367031612e-06,
"loss": 0.574,
"step": 1447
},
{
"epoch": 0.5825789579561457,
"grad_norm": 3.4944920810310633,
"learning_rate": 2.221459478940797e-06,
"loss": 0.4565,
"step": 1448
},
{
"epoch": 0.5829812914906457,
"grad_norm": 3.8209272903130325,
"learning_rate": 2.2179706705372377e-06,
"loss": 0.4116,
"step": 1449
},
{
"epoch": 0.5833836250251458,
"grad_norm": 4.562727204828236,
"learning_rate": 2.214482418373376e-06,
"loss": 0.6272,
"step": 1450
},
{
"epoch": 0.583785958559646,
"grad_norm": 3.986021082183279,
"learning_rate": 2.21099472932901e-06,
"loss": 0.4946,
"step": 1451
},
{
"epoch": 0.584188292094146,
"grad_norm": 3.619941584357637,
"learning_rate": 2.207507610282823e-06,
"loss": 0.5552,
"step": 1452
},
{
"epoch": 0.5845906256286462,
"grad_norm": 4.265275522318885,
"learning_rate": 2.2040210681123773e-06,
"loss": 0.5387,
"step": 1453
},
{
"epoch": 0.5849929591631462,
"grad_norm": 3.592671313761321,
"learning_rate": 2.2005351096940966e-06,
"loss": 0.4221,
"step": 1454
},
{
"epoch": 0.5853952926976463,
"grad_norm": 4.0275300106962755,
"learning_rate": 2.1970497419032546e-06,
"loss": 0.5781,
"step": 1455
},
{
"epoch": 0.5857976262321465,
"grad_norm": 3.8884576976533425,
"learning_rate": 2.1935649716139567e-06,
"loss": 0.5121,
"step": 1456
},
{
"epoch": 0.5861999597666465,
"grad_norm": 4.481785017500937,
"learning_rate": 2.1900808056991335e-06,
"loss": 0.6321,
"step": 1457
},
{
"epoch": 0.5866022933011467,
"grad_norm": 3.74252328127819,
"learning_rate": 2.186597251030521e-06,
"loss": 0.4935,
"step": 1458
},
{
"epoch": 0.5870046268356468,
"grad_norm": 3.9659714356438442,
"learning_rate": 2.183114314478651e-06,
"loss": 0.5503,
"step": 1459
},
{
"epoch": 0.5874069603701468,
"grad_norm": 3.435746642131932,
"learning_rate": 2.179632002912837e-06,
"loss": 0.3508,
"step": 1460
},
{
"epoch": 0.587809293904647,
"grad_norm": 4.32896011694038,
"learning_rate": 2.176150323201157e-06,
"loss": 0.5728,
"step": 1461
},
{
"epoch": 0.588211627439147,
"grad_norm": 4.1710482850436215,
"learning_rate": 2.1726692822104455e-06,
"loss": 0.5115,
"step": 1462
},
{
"epoch": 0.5886139609736472,
"grad_norm": 4.1244090573644145,
"learning_rate": 2.169188886806277e-06,
"loss": 0.3078,
"step": 1463
},
{
"epoch": 0.5890162945081473,
"grad_norm": 3.462366030557245,
"learning_rate": 2.1657091438529517e-06,
"loss": 0.472,
"step": 1464
},
{
"epoch": 0.5894186280426473,
"grad_norm": 3.920420130653373,
"learning_rate": 2.1622300602134824e-06,
"loss": 0.5559,
"step": 1465
},
{
"epoch": 0.5898209615771475,
"grad_norm": 4.574556709525141,
"learning_rate": 2.158751642749583e-06,
"loss": 0.5525,
"step": 1466
},
{
"epoch": 0.5902232951116475,
"grad_norm": 3.8284931733353513,
"learning_rate": 2.155273898321653e-06,
"loss": 0.5681,
"step": 1467
},
{
"epoch": 0.5906256286461476,
"grad_norm": 3.3784102201483677,
"learning_rate": 2.151796833788766e-06,
"loss": 0.4727,
"step": 1468
},
{
"epoch": 0.5910279621806478,
"grad_norm": 4.047681657794816,
"learning_rate": 2.1483204560086507e-06,
"loss": 0.5665,
"step": 1469
},
{
"epoch": 0.5914302957151478,
"grad_norm": 4.308357381109115,
"learning_rate": 2.144844771837685e-06,
"loss": 0.6112,
"step": 1470
},
{
"epoch": 0.591832629249648,
"grad_norm": 3.733873078989285,
"learning_rate": 2.1413697881308785e-06,
"loss": 0.494,
"step": 1471
},
{
"epoch": 0.5922349627841481,
"grad_norm": 4.130337043839704,
"learning_rate": 2.1378955117418567e-06,
"loss": 0.445,
"step": 1472
},
{
"epoch": 0.5926372963186481,
"grad_norm": 4.189878829812914,
"learning_rate": 2.1344219495228537e-06,
"loss": 0.569,
"step": 1473
},
{
"epoch": 0.5930396298531483,
"grad_norm": 3.926746164109752,
"learning_rate": 2.1309491083246915e-06,
"loss": 0.5696,
"step": 1474
},
{
"epoch": 0.5934419633876483,
"grad_norm": 3.688932516880435,
"learning_rate": 2.127476994996772e-06,
"loss": 0.5001,
"step": 1475
},
{
"epoch": 0.5938442969221485,
"grad_norm": 4.084824375606503,
"learning_rate": 2.1240056163870614e-06,
"loss": 0.613,
"step": 1476
},
{
"epoch": 0.5942466304566486,
"grad_norm": 4.614880824241214,
"learning_rate": 2.120534979342078e-06,
"loss": 0.5252,
"step": 1477
},
{
"epoch": 0.5946489639911486,
"grad_norm": 3.0888633468235014,
"learning_rate": 2.1170650907068742e-06,
"loss": 0.4531,
"step": 1478
},
{
"epoch": 0.5950512975256488,
"grad_norm": 4.213120019047452,
"learning_rate": 2.113595957325029e-06,
"loss": 0.6458,
"step": 1479
},
{
"epoch": 0.5954536310601488,
"grad_norm": 3.610258712936895,
"learning_rate": 2.1101275860386313e-06,
"loss": 0.5652,
"step": 1480
},
{
"epoch": 0.595855964594649,
"grad_norm": 3.901115632452379,
"learning_rate": 2.1066599836882667e-06,
"loss": 0.4731,
"step": 1481
},
{
"epoch": 0.5962582981291491,
"grad_norm": 4.576321666746087,
"learning_rate": 2.1031931571130053e-06,
"loss": 0.6799,
"step": 1482
},
{
"epoch": 0.5966606316636491,
"grad_norm": 4.188708168775149,
"learning_rate": 2.0997271131503847e-06,
"loss": 0.5172,
"step": 1483
},
{
"epoch": 0.5970629651981493,
"grad_norm": 3.993460741506799,
"learning_rate": 2.0962618586364005e-06,
"loss": 0.4828,
"step": 1484
},
{
"epoch": 0.5974652987326494,
"grad_norm": 3.977543353419007,
"learning_rate": 2.092797400405493e-06,
"loss": 0.5368,
"step": 1485
},
{
"epoch": 0.5978676322671495,
"grad_norm": 4.079669745022429,
"learning_rate": 2.0893337452905293e-06,
"loss": 0.4879,
"step": 1486
},
{
"epoch": 0.5982699658016496,
"grad_norm": 4.000645952876263,
"learning_rate": 2.0858709001227924e-06,
"loss": 0.5338,
"step": 1487
},
{
"epoch": 0.5986722993361496,
"grad_norm": 4.483257438703314,
"learning_rate": 2.0824088717319697e-06,
"loss": 0.5771,
"step": 1488
},
{
"epoch": 0.5990746328706498,
"grad_norm": 3.493611740579638,
"learning_rate": 2.0789476669461375e-06,
"loss": 0.4758,
"step": 1489
},
{
"epoch": 0.5994769664051499,
"grad_norm": 3.9869817394957385,
"learning_rate": 2.0754872925917465e-06,
"loss": 0.5482,
"step": 1490
},
{
"epoch": 0.59987929993965,
"grad_norm": 3.655771235240259,
"learning_rate": 2.072027755493612e-06,
"loss": 0.4924,
"step": 1491
},
{
"epoch": 0.6002816334741501,
"grad_norm": 3.959985484700295,
"learning_rate": 2.0685690624748943e-06,
"loss": 0.5653,
"step": 1492
},
{
"epoch": 0.6006839670086501,
"grad_norm": 3.4092513778358047,
"learning_rate": 2.0651112203570904e-06,
"loss": 0.4375,
"step": 1493
},
{
"epoch": 0.6010863005431503,
"grad_norm": 4.050380522574125,
"learning_rate": 2.061654235960021e-06,
"loss": 0.5627,
"step": 1494
},
{
"epoch": 0.6014886340776504,
"grad_norm": 4.329223167800351,
"learning_rate": 2.058198116101814e-06,
"loss": 0.5642,
"step": 1495
},
{
"epoch": 0.6018909676121504,
"grad_norm": 3.475966176009881,
"learning_rate": 2.0547428675988906e-06,
"loss": 0.4417,
"step": 1496
},
{
"epoch": 0.6022933011466506,
"grad_norm": 3.7228401404060443,
"learning_rate": 2.051288497265955e-06,
"loss": 0.4509,
"step": 1497
},
{
"epoch": 0.6026956346811507,
"grad_norm": 3.9298800993895284,
"learning_rate": 2.04783501191598e-06,
"loss": 0.4373,
"step": 1498
},
{
"epoch": 0.6030979682156508,
"grad_norm": 4.3431668642603105,
"learning_rate": 2.0443824183601917e-06,
"loss": 0.6361,
"step": 1499
},
{
"epoch": 0.6035003017501509,
"grad_norm": 3.759176356037184,
"learning_rate": 2.0409307234080577e-06,
"loss": 0.4214,
"step": 1500
},
{
"epoch": 0.6039026352846509,
"grad_norm": 3.881662137011941,
"learning_rate": 2.0374799338672723e-06,
"loss": 0.4868,
"step": 1501
},
{
"epoch": 0.6043049688191511,
"grad_norm": 4.786052455382573,
"learning_rate": 2.034030056543747e-06,
"loss": 0.6772,
"step": 1502
},
{
"epoch": 0.6047073023536512,
"grad_norm": 4.137240403825967,
"learning_rate": 2.0305810982415907e-06,
"loss": 0.4981,
"step": 1503
},
{
"epoch": 0.6051096358881513,
"grad_norm": 3.9245214475911947,
"learning_rate": 2.0271330657631034e-06,
"loss": 0.461,
"step": 1504
},
{
"epoch": 0.6055119694226514,
"grad_norm": 4.043701387123113,
"learning_rate": 2.023685965908755e-06,
"loss": 0.4861,
"step": 1505
},
{
"epoch": 0.6059143029571514,
"grad_norm": 4.578623188458431,
"learning_rate": 2.020239805477179e-06,
"loss": 0.5516,
"step": 1506
},
{
"epoch": 0.6063166364916516,
"grad_norm": 3.939503557149013,
"learning_rate": 2.0167945912651542e-06,
"loss": 0.5348,
"step": 1507
},
{
"epoch": 0.6067189700261517,
"grad_norm": 4.274904255142657,
"learning_rate": 2.0133503300675963e-06,
"loss": 0.5992,
"step": 1508
},
{
"epoch": 0.6071213035606517,
"grad_norm": 3.732143003199723,
"learning_rate": 2.0099070286775367e-06,
"loss": 0.4891,
"step": 1509
},
{
"epoch": 0.6075236370951519,
"grad_norm": 4.093400863247009,
"learning_rate": 2.0064646938861175e-06,
"loss": 0.4785,
"step": 1510
},
{
"epoch": 0.607925970629652,
"grad_norm": 3.850637017187709,
"learning_rate": 2.0030233324825734e-06,
"loss": 0.5371,
"step": 1511
},
{
"epoch": 0.6083283041641521,
"grad_norm": 3.726527595640385,
"learning_rate": 1.999582951254219e-06,
"loss": 0.4105,
"step": 1512
},
{
"epoch": 0.6087306376986522,
"grad_norm": 4.0679286236770045,
"learning_rate": 1.9961435569864366e-06,
"loss": 0.524,
"step": 1513
},
{
"epoch": 0.6091329712331522,
"grad_norm": 4.191017273712236,
"learning_rate": 1.99270515646266e-06,
"loss": 0.422,
"step": 1514
},
{
"epoch": 0.6095353047676524,
"grad_norm": 4.549567455680612,
"learning_rate": 1.9892677564643643e-06,
"loss": 0.5753,
"step": 1515
},
{
"epoch": 0.6099376383021525,
"grad_norm": 4.239380717200701,
"learning_rate": 1.985831363771052e-06,
"loss": 0.5251,
"step": 1516
},
{
"epoch": 0.6103399718366526,
"grad_norm": 3.8382650524932433,
"learning_rate": 1.9823959851602387e-06,
"loss": 0.6749,
"step": 1517
},
{
"epoch": 0.6107423053711527,
"grad_norm": 4.058628419689791,
"learning_rate": 1.9789616274074375e-06,
"loss": 0.5202,
"step": 1518
},
{
"epoch": 0.6111446389056527,
"grad_norm": 4.1466696163681815,
"learning_rate": 1.9755282972861515e-06,
"loss": 0.4595,
"step": 1519
},
{
"epoch": 0.6115469724401529,
"grad_norm": 4.193830162309673,
"learning_rate": 1.9720960015678555e-06,
"loss": 0.485,
"step": 1520
},
{
"epoch": 0.611949305974653,
"grad_norm": 4.239153001501294,
"learning_rate": 1.9686647470219835e-06,
"loss": 0.6782,
"step": 1521
},
{
"epoch": 0.6123516395091531,
"grad_norm": 4.138990574689833,
"learning_rate": 1.965234540415918e-06,
"loss": 0.6284,
"step": 1522
},
{
"epoch": 0.6127539730436532,
"grad_norm": 4.079634170381014,
"learning_rate": 1.9618053885149715e-06,
"loss": 0.4656,
"step": 1523
},
{
"epoch": 0.6131563065781533,
"grad_norm": 3.9398678794833204,
"learning_rate": 1.9583772980823797e-06,
"loss": 0.4345,
"step": 1524
},
{
"epoch": 0.6135586401126534,
"grad_norm": 3.8999859116475237,
"learning_rate": 1.9549502758792825e-06,
"loss": 0.4463,
"step": 1525
},
{
"epoch": 0.6139609736471535,
"grad_norm": 3.8236374179780652,
"learning_rate": 1.9515243286647155e-06,
"loss": 0.5032,
"step": 1526
},
{
"epoch": 0.6143633071816536,
"grad_norm": 3.9105069856639787,
"learning_rate": 1.9480994631955904e-06,
"loss": 0.5249,
"step": 1527
},
{
"epoch": 0.6147656407161537,
"grad_norm": 4.090697174474606,
"learning_rate": 1.944675686226688e-06,
"loss": 0.5986,
"step": 1528
},
{
"epoch": 0.6151679742506538,
"grad_norm": 4.026970204591501,
"learning_rate": 1.941253004510642e-06,
"loss": 0.5695,
"step": 1529
},
{
"epoch": 0.6155703077851539,
"grad_norm": 3.649099842703245,
"learning_rate": 1.937831424797926e-06,
"loss": 0.436,
"step": 1530
},
{
"epoch": 0.615972641319654,
"grad_norm": 3.8376284628790773,
"learning_rate": 1.9344109538368388e-06,
"loss": 0.4986,
"step": 1531
},
{
"epoch": 0.616374974854154,
"grad_norm": 4.048438236221887,
"learning_rate": 1.9309915983734933e-06,
"loss": 0.5608,
"step": 1532
},
{
"epoch": 0.6167773083886542,
"grad_norm": 4.036284880877361,
"learning_rate": 1.9275733651518038e-06,
"loss": 0.552,
"step": 1533
},
{
"epoch": 0.6171796419231543,
"grad_norm": 4.098386930913388,
"learning_rate": 1.9241562609134694e-06,
"loss": 0.5564,
"step": 1534
},
{
"epoch": 0.6175819754576544,
"grad_norm": 4.4349116098990065,
"learning_rate": 1.9207402923979633e-06,
"loss": 0.5937,
"step": 1535
},
{
"epoch": 0.6179843089921545,
"grad_norm": 3.9955554180753556,
"learning_rate": 1.9173254663425165e-06,
"loss": 0.4773,
"step": 1536
},
{
"epoch": 0.6183866425266547,
"grad_norm": 3.7895980004495553,
"learning_rate": 1.9139117894821107e-06,
"loss": 0.5117,
"step": 1537
},
{
"epoch": 0.6187889760611547,
"grad_norm": 4.503044474467485,
"learning_rate": 1.910499268549459e-06,
"loss": 0.5728,
"step": 1538
},
{
"epoch": 0.6191913095956548,
"grad_norm": 4.5048282169496865,
"learning_rate": 1.9070879102749944e-06,
"loss": 0.4952,
"step": 1539
},
{
"epoch": 0.6195936431301549,
"grad_norm": 3.6152226921367516,
"learning_rate": 1.9036777213868567e-06,
"loss": 0.4901,
"step": 1540
},
{
"epoch": 0.619995976664655,
"grad_norm": 4.56665728701204,
"learning_rate": 1.9002687086108808e-06,
"loss": 0.6217,
"step": 1541
},
{
"epoch": 0.6203983101991551,
"grad_norm": 3.8747261446432972,
"learning_rate": 1.8968608786705803e-06,
"loss": 0.4346,
"step": 1542
},
{
"epoch": 0.6208006437336552,
"grad_norm": 3.553549231891504,
"learning_rate": 1.8934542382871365e-06,
"loss": 0.4086,
"step": 1543
},
{
"epoch": 0.6212029772681553,
"grad_norm": 4.273007467466175,
"learning_rate": 1.8900487941793862e-06,
"loss": 0.5589,
"step": 1544
},
{
"epoch": 0.6216053108026554,
"grad_norm": 4.033041498302765,
"learning_rate": 1.8866445530638034e-06,
"loss": 0.5258,
"step": 1545
},
{
"epoch": 0.6220076443371555,
"grad_norm": 3.847764801262625,
"learning_rate": 1.883241521654492e-06,
"loss": 0.5286,
"step": 1546
},
{
"epoch": 0.6224099778716556,
"grad_norm": 4.141005318996514,
"learning_rate": 1.87983970666317e-06,
"loss": 0.5428,
"step": 1547
},
{
"epoch": 0.6228123114061557,
"grad_norm": 3.8545860722248286,
"learning_rate": 1.8764391147991556e-06,
"loss": 0.4504,
"step": 1548
},
{
"epoch": 0.6232146449406558,
"grad_norm": 3.9064270159769543,
"learning_rate": 1.8730397527693544e-06,
"loss": 0.49,
"step": 1549
},
{
"epoch": 0.623616978475156,
"grad_norm": 3.5305364890500903,
"learning_rate": 1.8696416272782467e-06,
"loss": 0.4747,
"step": 1550
},
{
"epoch": 0.624019312009656,
"grad_norm": 4.367606237537799,
"learning_rate": 1.866244745027874e-06,
"loss": 0.6066,
"step": 1551
},
{
"epoch": 0.6244216455441561,
"grad_norm": 3.778565429919078,
"learning_rate": 1.8628491127178272e-06,
"loss": 0.4401,
"step": 1552
},
{
"epoch": 0.6248239790786562,
"grad_norm": 4.4484086477744675,
"learning_rate": 1.8594547370452288e-06,
"loss": 0.5702,
"step": 1553
},
{
"epoch": 0.6252263126131563,
"grad_norm": 4.133384847499752,
"learning_rate": 1.8560616247047253e-06,
"loss": 0.4958,
"step": 1554
},
{
"epoch": 0.6256286461476565,
"grad_norm": 4.481750156438252,
"learning_rate": 1.8526697823884721e-06,
"loss": 0.5982,
"step": 1555
},
{
"epoch": 0.6260309796821565,
"grad_norm": 4.679844131200749,
"learning_rate": 1.849279216786117e-06,
"loss": 0.5699,
"step": 1556
},
{
"epoch": 0.6264333132166566,
"grad_norm": 4.035894316525195,
"learning_rate": 1.8458899345847936e-06,
"loss": 0.6036,
"step": 1557
},
{
"epoch": 0.6268356467511567,
"grad_norm": 3.3784184252716836,
"learning_rate": 1.8425019424690993e-06,
"loss": 0.3667,
"step": 1558
},
{
"epoch": 0.6272379802856568,
"grad_norm": 3.7155353324243325,
"learning_rate": 1.839115247121091e-06,
"loss": 0.4744,
"step": 1559
},
{
"epoch": 0.627640313820157,
"grad_norm": 4.062605549457995,
"learning_rate": 1.8357298552202676e-06,
"loss": 0.4846,
"step": 1560
},
{
"epoch": 0.628042647354657,
"grad_norm": 3.968579655925073,
"learning_rate": 1.832345773443557e-06,
"loss": 0.4563,
"step": 1561
},
{
"epoch": 0.6284449808891571,
"grad_norm": 3.859404877482336,
"learning_rate": 1.8289630084653009e-06,
"loss": 0.5339,
"step": 1562
},
{
"epoch": 0.6288473144236573,
"grad_norm": 4.32493582991546,
"learning_rate": 1.825581566957247e-06,
"loss": 0.5206,
"step": 1563
},
{
"epoch": 0.6292496479581573,
"grad_norm": 3.7620622738243474,
"learning_rate": 1.8222014555885303e-06,
"loss": 0.5187,
"step": 1564
},
{
"epoch": 0.6296519814926574,
"grad_norm": 3.9245871423313616,
"learning_rate": 1.8188226810256647e-06,
"loss": 0.5302,
"step": 1565
},
{
"epoch": 0.6300543150271575,
"grad_norm": 4.0730169054317305,
"learning_rate": 1.8154452499325273e-06,
"loss": 0.482,
"step": 1566
},
{
"epoch": 0.6304566485616576,
"grad_norm": 4.0887356034823314,
"learning_rate": 1.8120691689703423e-06,
"loss": 0.6323,
"step": 1567
},
{
"epoch": 0.6308589820961578,
"grad_norm": 4.134013360882183,
"learning_rate": 1.8086944447976745e-06,
"loss": 0.4215,
"step": 1568
},
{
"epoch": 0.6312613156306578,
"grad_norm": 4.267968482401992,
"learning_rate": 1.805321084070412e-06,
"loss": 0.6421,
"step": 1569
},
{
"epoch": 0.6316636491651579,
"grad_norm": 4.337432367486679,
"learning_rate": 1.8019490934417533e-06,
"loss": 0.614,
"step": 1570
},
{
"epoch": 0.632065982699658,
"grad_norm": 4.325604571245914,
"learning_rate": 1.7985784795621946e-06,
"loss": 0.6082,
"step": 1571
},
{
"epoch": 0.6324683162341581,
"grad_norm": 4.7255506567581005,
"learning_rate": 1.7952092490795165e-06,
"loss": 0.5848,
"step": 1572
},
{
"epoch": 0.6328706497686583,
"grad_norm": 3.9917358158112513,
"learning_rate": 1.7918414086387723e-06,
"loss": 0.51,
"step": 1573
},
{
"epoch": 0.6332729833031583,
"grad_norm": 3.9759190309543806,
"learning_rate": 1.788474964882273e-06,
"loss": 0.4348,
"step": 1574
},
{
"epoch": 0.6336753168376584,
"grad_norm": 3.9060781229407673,
"learning_rate": 1.7851099244495761e-06,
"loss": 0.5328,
"step": 1575
},
{
"epoch": 0.6340776503721586,
"grad_norm": 3.794629121755811,
"learning_rate": 1.7817462939774683e-06,
"loss": 0.5406,
"step": 1576
},
{
"epoch": 0.6344799839066586,
"grad_norm": 4.434204491241485,
"learning_rate": 1.778384080099959e-06,
"loss": 0.6955,
"step": 1577
},
{
"epoch": 0.6348823174411587,
"grad_norm": 4.243047298973574,
"learning_rate": 1.775023289448261e-06,
"loss": 0.6021,
"step": 1578
},
{
"epoch": 0.6352846509756588,
"grad_norm": 3.906621293593377,
"learning_rate": 1.7716639286507834e-06,
"loss": 0.5017,
"step": 1579
},
{
"epoch": 0.6356869845101589,
"grad_norm": 4.433075604901836,
"learning_rate": 1.7683060043331102e-06,
"loss": 0.6304,
"step": 1580
},
{
"epoch": 0.6360893180446591,
"grad_norm": 4.0324680233039425,
"learning_rate": 1.764949523117997e-06,
"loss": 0.4593,
"step": 1581
},
{
"epoch": 0.6364916515791591,
"grad_norm": 4.861200946299266,
"learning_rate": 1.7615944916253511e-06,
"loss": 0.5755,
"step": 1582
},
{
"epoch": 0.6368939851136592,
"grad_norm": 3.4859779578870462,
"learning_rate": 1.7582409164722219e-06,
"loss": 0.4494,
"step": 1583
},
{
"epoch": 0.6372963186481593,
"grad_norm": 4.394364956739334,
"learning_rate": 1.7548888042727835e-06,
"loss": 0.5756,
"step": 1584
},
{
"epoch": 0.6376986521826594,
"grad_norm": 4.474860851761582,
"learning_rate": 1.751538161638327e-06,
"loss": 0.5217,
"step": 1585
},
{
"epoch": 0.6381009857171596,
"grad_norm": 4.665357328027418,
"learning_rate": 1.748188995177245e-06,
"loss": 0.5599,
"step": 1586
},
{
"epoch": 0.6385033192516596,
"grad_norm": 3.186769209343777,
"learning_rate": 1.744841311495019e-06,
"loss": 0.3409,
"step": 1587
},
{
"epoch": 0.6389056527861597,
"grad_norm": 3.3506177658392744,
"learning_rate": 1.7414951171942057e-06,
"loss": 0.5176,
"step": 1588
},
{
"epoch": 0.6393079863206599,
"grad_norm": 3.4136910306135935,
"learning_rate": 1.7381504188744218e-06,
"loss": 0.4464,
"step": 1589
},
{
"epoch": 0.6397103198551599,
"grad_norm": 3.4917018490110467,
"learning_rate": 1.7348072231323382e-06,
"loss": 0.438,
"step": 1590
},
{
"epoch": 0.6401126533896601,
"grad_norm": 3.421803275946972,
"learning_rate": 1.7314655365616584e-06,
"loss": 0.4189,
"step": 1591
},
{
"epoch": 0.6405149869241601,
"grad_norm": 4.433987296079437,
"learning_rate": 1.7281253657531123e-06,
"loss": 0.606,
"step": 1592
},
{
"epoch": 0.6409173204586602,
"grad_norm": 3.6737703307949983,
"learning_rate": 1.7247867172944367e-06,
"loss": 0.4349,
"step": 1593
},
{
"epoch": 0.6413196539931604,
"grad_norm": 3.8221357123925914,
"learning_rate": 1.721449597770369e-06,
"loss": 0.484,
"step": 1594
},
{
"epoch": 0.6417219875276604,
"grad_norm": 4.069693321778052,
"learning_rate": 1.718114013762631e-06,
"loss": 0.6507,
"step": 1595
},
{
"epoch": 0.6421243210621606,
"grad_norm": 4.268934323506783,
"learning_rate": 1.714779971849915e-06,
"loss": 0.498,
"step": 1596
},
{
"epoch": 0.6425266545966606,
"grad_norm": 4.426082622301671,
"learning_rate": 1.7114474786078716e-06,
"loss": 0.6755,
"step": 1597
},
{
"epoch": 0.6429289881311607,
"grad_norm": 3.6655062613985288,
"learning_rate": 1.708116540609098e-06,
"loss": 0.5078,
"step": 1598
},
{
"epoch": 0.6433313216656609,
"grad_norm": 4.096742196755438,
"learning_rate": 1.7047871644231226e-06,
"loss": 0.5085,
"step": 1599
},
{
"epoch": 0.6437336552001609,
"grad_norm": 4.341802218643631,
"learning_rate": 1.7014593566163949e-06,
"loss": 0.6022,
"step": 1600
},
{
"epoch": 0.644135988734661,
"grad_norm": 4.167303639683251,
"learning_rate": 1.6981331237522716e-06,
"loss": 0.595,
"step": 1601
},
{
"epoch": 0.6445383222691611,
"grad_norm": 3.860753980307849,
"learning_rate": 1.6948084723910002e-06,
"loss": 0.4189,
"step": 1602
},
{
"epoch": 0.6449406558036612,
"grad_norm": 3.801435117916446,
"learning_rate": 1.6914854090897124e-06,
"loss": 0.4691,
"step": 1603
},
{
"epoch": 0.6453429893381614,
"grad_norm": 4.309844638484278,
"learning_rate": 1.6881639404024063e-06,
"loss": 0.4802,
"step": 1604
},
{
"epoch": 0.6457453228726614,
"grad_norm": 3.8218645060402285,
"learning_rate": 1.6848440728799346e-06,
"loss": 0.6087,
"step": 1605
},
{
"epoch": 0.6461476564071615,
"grad_norm": 4.318515117319728,
"learning_rate": 1.6815258130699938e-06,
"loss": 0.6644,
"step": 1606
},
{
"epoch": 0.6465499899416617,
"grad_norm": 3.8706029461346114,
"learning_rate": 1.6782091675171064e-06,
"loss": 0.5314,
"step": 1607
},
{
"epoch": 0.6469523234761617,
"grad_norm": 4.588786654868146,
"learning_rate": 1.6748941427626142e-06,
"loss": 0.5713,
"step": 1608
},
{
"epoch": 0.6473546570106619,
"grad_norm": 4.206226833574013,
"learning_rate": 1.671580745344661e-06,
"loss": 0.5207,
"step": 1609
},
{
"epoch": 0.6477569905451619,
"grad_norm": 4.1241193689590565,
"learning_rate": 1.6682689817981827e-06,
"loss": 0.5325,
"step": 1610
},
{
"epoch": 0.648159324079662,
"grad_norm": 4.013331656155903,
"learning_rate": 1.6649588586548888e-06,
"loss": 0.5593,
"step": 1611
},
{
"epoch": 0.6485616576141622,
"grad_norm": 4.701503684635364,
"learning_rate": 1.6616503824432573e-06,
"loss": 0.6878,
"step": 1612
},
{
"epoch": 0.6489639911486622,
"grad_norm": 4.047494584818948,
"learning_rate": 1.658343559688516e-06,
"loss": 0.5131,
"step": 1613
},
{
"epoch": 0.6493663246831624,
"grad_norm": 4.0317139401208575,
"learning_rate": 1.6550383969126341e-06,
"loss": 0.512,
"step": 1614
},
{
"epoch": 0.6497686582176624,
"grad_norm": 3.9682317404566034,
"learning_rate": 1.6517349006343025e-06,
"loss": 0.5921,
"step": 1615
},
{
"epoch": 0.6501709917521625,
"grad_norm": 3.824331386002604,
"learning_rate": 1.6484330773689292e-06,
"loss": 0.4712,
"step": 1616
},
{
"epoch": 0.6505733252866627,
"grad_norm": 3.7019715035054204,
"learning_rate": 1.645132933628621e-06,
"loss": 0.4426,
"step": 1617
},
{
"epoch": 0.6509756588211627,
"grad_norm": 4.0459494986124245,
"learning_rate": 1.6418344759221734e-06,
"loss": 0.6728,
"step": 1618
},
{
"epoch": 0.6513779923556628,
"grad_norm": 4.154812270453969,
"learning_rate": 1.6385377107550543e-06,
"loss": 0.559,
"step": 1619
},
{
"epoch": 0.651780325890163,
"grad_norm": 4.703387446268752,
"learning_rate": 1.6352426446293952e-06,
"loss": 0.5548,
"step": 1620
},
{
"epoch": 0.652182659424663,
"grad_norm": 3.894555312354832,
"learning_rate": 1.631949284043976e-06,
"loss": 0.4987,
"step": 1621
},
{
"epoch": 0.6525849929591632,
"grad_norm": 4.281016820029128,
"learning_rate": 1.628657635494213e-06,
"loss": 0.6007,
"step": 1622
},
{
"epoch": 0.6529873264936632,
"grad_norm": 3.708021853167261,
"learning_rate": 1.625367705472147e-06,
"loss": 0.5041,
"step": 1623
},
{
"epoch": 0.6533896600281633,
"grad_norm": 4.184667272519736,
"learning_rate": 1.6220795004664264e-06,
"loss": 0.563,
"step": 1624
},
{
"epoch": 0.6537919935626635,
"grad_norm": 4.386544791097414,
"learning_rate": 1.6187930269623001e-06,
"loss": 0.6591,
"step": 1625
},
{
"epoch": 0.6541943270971635,
"grad_norm": 4.095997806253382,
"learning_rate": 1.6155082914416018e-06,
"loss": 0.5122,
"step": 1626
},
{
"epoch": 0.6545966606316637,
"grad_norm": 3.97582656513418,
"learning_rate": 1.6122253003827353e-06,
"loss": 0.5805,
"step": 1627
},
{
"epoch": 0.6549989941661637,
"grad_norm": 3.990901166079429,
"learning_rate": 1.6089440602606675e-06,
"loss": 0.569,
"step": 1628
},
{
"epoch": 0.6554013277006638,
"grad_norm": 3.8624368116988403,
"learning_rate": 1.6056645775469077e-06,
"loss": 0.4729,
"step": 1629
},
{
"epoch": 0.655803661235164,
"grad_norm": 4.529362920268054,
"learning_rate": 1.6023868587095016e-06,
"loss": 0.5482,
"step": 1630
},
{
"epoch": 0.656205994769664,
"grad_norm": 4.017565552354934,
"learning_rate": 1.599110910213016e-06,
"loss": 0.4916,
"step": 1631
},
{
"epoch": 0.6566083283041642,
"grad_norm": 4.4182156313212575,
"learning_rate": 1.595836738518527e-06,
"loss": 0.5792,
"step": 1632
},
{
"epoch": 0.6570106618386643,
"grad_norm": 3.5677962985553524,
"learning_rate": 1.592564350083603e-06,
"loss": 0.4697,
"step": 1633
},
{
"epoch": 0.6574129953731643,
"grad_norm": 3.909711067921425,
"learning_rate": 1.5892937513622977e-06,
"loss": 0.4383,
"step": 1634
},
{
"epoch": 0.6578153289076645,
"grad_norm": 4.098825788465001,
"learning_rate": 1.5860249488051355e-06,
"loss": 0.4745,
"step": 1635
},
{
"epoch": 0.6582176624421645,
"grad_norm": 3.2004894026603616,
"learning_rate": 1.5827579488590977e-06,
"loss": 0.4775,
"step": 1636
},
{
"epoch": 0.6586199959766647,
"grad_norm": 3.970341598297501,
"learning_rate": 1.5794927579676089e-06,
"loss": 0.5471,
"step": 1637
},
{
"epoch": 0.6590223295111648,
"grad_norm": 4.241718162619275,
"learning_rate": 1.5762293825705276e-06,
"loss": 0.6544,
"step": 1638
},
{
"epoch": 0.6594246630456648,
"grad_norm": 3.924819583311759,
"learning_rate": 1.5729678291041318e-06,
"loss": 0.5368,
"step": 1639
},
{
"epoch": 0.659826996580165,
"grad_norm": 3.8382551527997397,
"learning_rate": 1.5697081040011043e-06,
"loss": 0.5121,
"step": 1640
},
{
"epoch": 0.660229330114665,
"grad_norm": 3.6682266397110337,
"learning_rate": 1.5664502136905236e-06,
"loss": 0.4353,
"step": 1641
},
{
"epoch": 0.6606316636491651,
"grad_norm": 3.9060845453518143,
"learning_rate": 1.563194164597848e-06,
"loss": 0.4814,
"step": 1642
},
{
"epoch": 0.6610339971836653,
"grad_norm": 3.985541657912861,
"learning_rate": 1.5599399631449052e-06,
"loss": 0.4643,
"step": 1643
},
{
"epoch": 0.6614363307181653,
"grad_norm": 4.246558106503365,
"learning_rate": 1.5566876157498797e-06,
"loss": 0.5457,
"step": 1644
},
{
"epoch": 0.6618386642526655,
"grad_norm": 4.538356504223212,
"learning_rate": 1.5534371288272992e-06,
"loss": 0.5586,
"step": 1645
},
{
"epoch": 0.6622409977871656,
"grad_norm": 3.512817123486096,
"learning_rate": 1.5501885087880187e-06,
"loss": 0.3777,
"step": 1646
},
{
"epoch": 0.6626433313216656,
"grad_norm": 3.611315735282059,
"learning_rate": 1.546941762039216e-06,
"loss": 0.4559,
"step": 1647
},
{
"epoch": 0.6630456648561658,
"grad_norm": 3.804900233889627,
"learning_rate": 1.5436968949843705e-06,
"loss": 0.5593,
"step": 1648
},
{
"epoch": 0.6634479983906658,
"grad_norm": 4.290281020900778,
"learning_rate": 1.5404539140232572e-06,
"loss": 0.5433,
"step": 1649
},
{
"epoch": 0.663850331925166,
"grad_norm": 4.261399800624298,
"learning_rate": 1.5372128255519298e-06,
"loss": 0.5669,
"step": 1650
},
{
"epoch": 0.6642526654596661,
"grad_norm": 4.8206169191275094,
"learning_rate": 1.533973635962708e-06,
"loss": 0.535,
"step": 1651
},
{
"epoch": 0.6646549989941661,
"grad_norm": 3.5705866846183882,
"learning_rate": 1.5307363516441693e-06,
"loss": 0.4436,
"step": 1652
},
{
"epoch": 0.6650573325286663,
"grad_norm": 4.015284104793611,
"learning_rate": 1.5275009789811317e-06,
"loss": 0.4606,
"step": 1653
},
{
"epoch": 0.6654596660631663,
"grad_norm": 3.746888831379427,
"learning_rate": 1.5242675243546435e-06,
"loss": 0.5344,
"step": 1654
},
{
"epoch": 0.6658619995976665,
"grad_norm": 5.350374969998737,
"learning_rate": 1.5210359941419698e-06,
"loss": 0.7187,
"step": 1655
},
{
"epoch": 0.6662643331321666,
"grad_norm": 4.992414873463243,
"learning_rate": 1.5178063947165795e-06,
"loss": 0.5597,
"step": 1656
},
{
"epoch": 0.6666666666666666,
"grad_norm": 3.9536138723501413,
"learning_rate": 1.514578732448135e-06,
"loss": 0.5345,
"step": 1657
},
{
"epoch": 0.6670690002011668,
"grad_norm": 4.023274067891729,
"learning_rate": 1.5113530137024774e-06,
"loss": 0.4893,
"step": 1658
},
{
"epoch": 0.6674713337356669,
"grad_norm": 4.108018429855622,
"learning_rate": 1.5081292448416158e-06,
"loss": 0.645,
"step": 1659
},
{
"epoch": 0.667873667270167,
"grad_norm": 3.817031046521698,
"learning_rate": 1.5049074322237097e-06,
"loss": 0.4838,
"step": 1660
},
{
"epoch": 0.6682760008046671,
"grad_norm": 4.10734795999481,
"learning_rate": 1.5016875822030651e-06,
"loss": 0.5853,
"step": 1661
},
{
"epoch": 0.6686783343391671,
"grad_norm": 4.104736058188134,
"learning_rate": 1.4984697011301142e-06,
"loss": 0.5032,
"step": 1662
},
{
"epoch": 0.6690806678736673,
"grad_norm": 4.213736489738209,
"learning_rate": 1.4952537953514078e-06,
"loss": 0.4234,
"step": 1663
},
{
"epoch": 0.6694830014081674,
"grad_norm": 4.592368631770386,
"learning_rate": 1.4920398712095985e-06,
"loss": 0.5323,
"step": 1664
},
{
"epoch": 0.6698853349426674,
"grad_norm": 4.593839464020305,
"learning_rate": 1.4888279350434325e-06,
"loss": 0.5821,
"step": 1665
},
{
"epoch": 0.6702876684771676,
"grad_norm": 3.881133737659202,
"learning_rate": 1.485617993187735e-06,
"loss": 0.5224,
"step": 1666
},
{
"epoch": 0.6706900020116676,
"grad_norm": 4.088068946795046,
"learning_rate": 1.4824100519733985e-06,
"loss": 0.556,
"step": 1667
},
{
"epoch": 0.6710923355461678,
"grad_norm": 4.328495324161391,
"learning_rate": 1.4792041177273668e-06,
"loss": 0.6097,
"step": 1668
},
{
"epoch": 0.6714946690806679,
"grad_norm": 4.220316709489982,
"learning_rate": 1.4760001967726283e-06,
"loss": 0.6015,
"step": 1669
},
{
"epoch": 0.6718970026151679,
"grad_norm": 3.788396547796328,
"learning_rate": 1.472798295428199e-06,
"loss": 0.5863,
"step": 1670
},
{
"epoch": 0.6722993361496681,
"grad_norm": 3.8086553750558494,
"learning_rate": 1.4695984200091129e-06,
"loss": 0.5155,
"step": 1671
},
{
"epoch": 0.6727016696841682,
"grad_norm": 3.792467737324882,
"learning_rate": 1.4664005768264084e-06,
"loss": 0.4517,
"step": 1672
},
{
"epoch": 0.6731040032186683,
"grad_norm": 4.528886592192105,
"learning_rate": 1.4632047721871134e-06,
"loss": 0.5514,
"step": 1673
},
{
"epoch": 0.6735063367531684,
"grad_norm": 3.687134247150264,
"learning_rate": 1.4600110123942376e-06,
"loss": 0.5765,
"step": 1674
},
{
"epoch": 0.6739086702876684,
"grad_norm": 3.651752055706122,
"learning_rate": 1.4568193037467574e-06,
"loss": 0.4434,
"step": 1675
},
{
"epoch": 0.6743110038221686,
"grad_norm": 4.196555677667497,
"learning_rate": 1.453629652539604e-06,
"loss": 0.5832,
"step": 1676
},
{
"epoch": 0.6747133373566687,
"grad_norm": 3.3768494467370047,
"learning_rate": 1.4504420650636485e-06,
"loss": 0.4094,
"step": 1677
},
{
"epoch": 0.6751156708911688,
"grad_norm": 3.9350013310511582,
"learning_rate": 1.4472565476056952e-06,
"loss": 0.4875,
"step": 1678
},
{
"epoch": 0.6755180044256689,
"grad_norm": 3.9364895166754383,
"learning_rate": 1.444073106448462e-06,
"loss": 0.626,
"step": 1679
},
{
"epoch": 0.6759203379601689,
"grad_norm": 4.218774138421791,
"learning_rate": 1.4408917478705748e-06,
"loss": 0.5318,
"step": 1680
},
{
"epoch": 0.6763226714946691,
"grad_norm": 3.557852927029882,
"learning_rate": 1.4377124781465517e-06,
"loss": 0.4564,
"step": 1681
},
{
"epoch": 0.6767250050291692,
"grad_norm": 4.131292301591179,
"learning_rate": 1.4345353035467888e-06,
"loss": 0.5458,
"step": 1682
},
{
"epoch": 0.6771273385636692,
"grad_norm": 4.261120314757306,
"learning_rate": 1.4313602303375523e-06,
"loss": 0.6402,
"step": 1683
},
{
"epoch": 0.6775296720981694,
"grad_norm": 4.027439558487913,
"learning_rate": 1.4281872647809628e-06,
"loss": 0.5762,
"step": 1684
},
{
"epoch": 0.6779320056326695,
"grad_norm": 3.9491401859678605,
"learning_rate": 1.4250164131349858e-06,
"loss": 0.4807,
"step": 1685
},
{
"epoch": 0.6783343391671696,
"grad_norm": 3.60940986670999,
"learning_rate": 1.421847681653414e-06,
"loss": 0.3659,
"step": 1686
},
{
"epoch": 0.6787366727016697,
"grad_norm": 3.5822407819650515,
"learning_rate": 1.4186810765858616e-06,
"loss": 0.5165,
"step": 1687
},
{
"epoch": 0.6791390062361697,
"grad_norm": 4.161182758555724,
"learning_rate": 1.4155166041777483e-06,
"loss": 0.495,
"step": 1688
},
{
"epoch": 0.6795413397706699,
"grad_norm": 4.076431248542256,
"learning_rate": 1.4123542706702868e-06,
"loss": 0.4569,
"step": 1689
},
{
"epoch": 0.67994367330517,
"grad_norm": 3.401944784975411,
"learning_rate": 1.409194082300473e-06,
"loss": 0.4661,
"step": 1690
},
{
"epoch": 0.6803460068396701,
"grad_norm": 4.053896920545759,
"learning_rate": 1.4060360453010691e-06,
"loss": 0.6945,
"step": 1691
},
{
"epoch": 0.6807483403741702,
"grad_norm": 3.6315876209619726,
"learning_rate": 1.4028801659005975e-06,
"loss": 0.5088,
"step": 1692
},
{
"epoch": 0.6811506739086702,
"grad_norm": 3.9033263695876617,
"learning_rate": 1.3997264503233222e-06,
"loss": 0.5535,
"step": 1693
},
{
"epoch": 0.6815530074431704,
"grad_norm": 3.6520508454132568,
"learning_rate": 1.3965749047892421e-06,
"loss": 0.4627,
"step": 1694
},
{
"epoch": 0.6819553409776705,
"grad_norm": 4.253145100591827,
"learning_rate": 1.3934255355140735e-06,
"loss": 0.6059,
"step": 1695
},
{
"epoch": 0.6823576745121706,
"grad_norm": 3.6442609114551168,
"learning_rate": 1.3902783487092436e-06,
"loss": 0.437,
"step": 1696
},
{
"epoch": 0.6827600080466707,
"grad_norm": 3.6843000132081083,
"learning_rate": 1.387133350581873e-06,
"loss": 0.4741,
"step": 1697
},
{
"epoch": 0.6831623415811708,
"grad_norm": 4.082924701084028,
"learning_rate": 1.383990547334767e-06,
"loss": 0.5523,
"step": 1698
},
{
"epoch": 0.6835646751156709,
"grad_norm": 4.525193517476986,
"learning_rate": 1.3808499451663998e-06,
"loss": 0.5733,
"step": 1699
},
{
"epoch": 0.683967008650171,
"grad_norm": 3.8434172838289142,
"learning_rate": 1.377711550270907e-06,
"loss": 0.4042,
"step": 1700
},
{
"epoch": 0.684369342184671,
"grad_norm": 3.475844585621674,
"learning_rate": 1.3745753688380692e-06,
"loss": 0.4492,
"step": 1701
},
{
"epoch": 0.6847716757191712,
"grad_norm": 4.106466165584271,
"learning_rate": 1.3714414070533022e-06,
"loss": 0.4795,
"step": 1702
},
{
"epoch": 0.6851740092536713,
"grad_norm": 3.9408796988863477,
"learning_rate": 1.368309671097645e-06,
"loss": 0.4586,
"step": 1703
},
{
"epoch": 0.6855763427881714,
"grad_norm": 3.892878635048642,
"learning_rate": 1.3651801671477435e-06,
"loss": 0.5256,
"step": 1704
},
{
"epoch": 0.6859786763226715,
"grad_norm": 3.602589673252341,
"learning_rate": 1.3620529013758444e-06,
"loss": 0.4474,
"step": 1705
},
{
"epoch": 0.6863810098571715,
"grad_norm": 4.29050518808401,
"learning_rate": 1.3589278799497802e-06,
"loss": 0.4799,
"step": 1706
},
{
"epoch": 0.6867833433916717,
"grad_norm": 4.530188596691838,
"learning_rate": 1.3558051090329536e-06,
"loss": 0.6044,
"step": 1707
},
{
"epoch": 0.6871856769261718,
"grad_norm": 3.8842771157381932,
"learning_rate": 1.352684594784333e-06,
"loss": 0.4759,
"step": 1708
},
{
"epoch": 0.6875880104606719,
"grad_norm": 4.206785571596244,
"learning_rate": 1.3495663433584327e-06,
"loss": 0.58,
"step": 1709
},
{
"epoch": 0.687990343995172,
"grad_norm": 4.648975138578508,
"learning_rate": 1.3464503609053053e-06,
"loss": 0.5508,
"step": 1710
},
{
"epoch": 0.6883926775296721,
"grad_norm": 4.513492173355037,
"learning_rate": 1.3433366535705292e-06,
"loss": 0.6606,
"step": 1711
},
{
"epoch": 0.6887950110641722,
"grad_norm": 4.149085872177257,
"learning_rate": 1.3402252274951954e-06,
"loss": 0.5438,
"step": 1712
},
{
"epoch": 0.6891973445986723,
"grad_norm": 4.3080582119504935,
"learning_rate": 1.3371160888158935e-06,
"loss": 0.7277,
"step": 1713
},
{
"epoch": 0.6895996781331724,
"grad_norm": 4.195944255893043,
"learning_rate": 1.3340092436647045e-06,
"loss": 0.5682,
"step": 1714
},
{
"epoch": 0.6900020116676725,
"grad_norm": 4.22717459587227,
"learning_rate": 1.330904698169184e-06,
"loss": 0.5394,
"step": 1715
},
{
"epoch": 0.6904043452021726,
"grad_norm": 3.8328998433470334,
"learning_rate": 1.3278024584523542e-06,
"loss": 0.519,
"step": 1716
},
{
"epoch": 0.6908066787366727,
"grad_norm": 3.809809815071977,
"learning_rate": 1.324702530632686e-06,
"loss": 0.4609,
"step": 1717
},
{
"epoch": 0.6912090122711728,
"grad_norm": 4.408644772842878,
"learning_rate": 1.3216049208240945e-06,
"loss": 0.5259,
"step": 1718
},
{
"epoch": 0.6916113458056729,
"grad_norm": 4.351040955206057,
"learning_rate": 1.3185096351359211e-06,
"loss": 0.531,
"step": 1719
},
{
"epoch": 0.692013679340173,
"grad_norm": 3.8862458710646135,
"learning_rate": 1.3154166796729245e-06,
"loss": 0.5521,
"step": 1720
},
{
"epoch": 0.6924160128746731,
"grad_norm": 3.8732946833809283,
"learning_rate": 1.3123260605352647e-06,
"loss": 0.4054,
"step": 1721
},
{
"epoch": 0.6928183464091732,
"grad_norm": 4.014263742357025,
"learning_rate": 1.3092377838184984e-06,
"loss": 0.5096,
"step": 1722
},
{
"epoch": 0.6932206799436733,
"grad_norm": 3.8457014699738306,
"learning_rate": 1.3061518556135572e-06,
"loss": 0.5361,
"step": 1723
},
{
"epoch": 0.6936230134781735,
"grad_norm": 4.700764158732767,
"learning_rate": 1.3030682820067447e-06,
"loss": 0.6363,
"step": 1724
},
{
"epoch": 0.6940253470126735,
"grad_norm": 4.056861341126563,
"learning_rate": 1.2999870690797207e-06,
"loss": 0.5499,
"step": 1725
},
{
"epoch": 0.6944276805471736,
"grad_norm": 3.824223111444158,
"learning_rate": 1.2969082229094853e-06,
"loss": 0.5049,
"step": 1726
},
{
"epoch": 0.6948300140816737,
"grad_norm": 3.8066989110426332,
"learning_rate": 1.2938317495683744e-06,
"loss": 0.487,
"step": 1727
},
{
"epoch": 0.6952323476161738,
"grad_norm": 4.382851529690001,
"learning_rate": 1.2907576551240425e-06,
"loss": 0.5016,
"step": 1728
},
{
"epoch": 0.695634681150674,
"grad_norm": 4.2709658357684415,
"learning_rate": 1.2876859456394537e-06,
"loss": 0.5287,
"step": 1729
},
{
"epoch": 0.696037014685174,
"grad_norm": 4.10428141385855,
"learning_rate": 1.284616627172865e-06,
"loss": 0.5048,
"step": 1730
},
{
"epoch": 0.6964393482196741,
"grad_norm": 3.9240197132726133,
"learning_rate": 1.281549705777821e-06,
"loss": 0.5834,
"step": 1731
},
{
"epoch": 0.6968416817541742,
"grad_norm": 3.917832710676123,
"learning_rate": 1.278485187503137e-06,
"loss": 0.4978,
"step": 1732
},
{
"epoch": 0.6972440152886743,
"grad_norm": 3.9886241574028123,
"learning_rate": 1.2754230783928907e-06,
"loss": 0.5046,
"step": 1733
},
{
"epoch": 0.6976463488231744,
"grad_norm": 3.8252585006207744,
"learning_rate": 1.2723633844864036e-06,
"loss": 0.4337,
"step": 1734
},
{
"epoch": 0.6980486823576745,
"grad_norm": 4.327909530843214,
"learning_rate": 1.2693061118182399e-06,
"loss": 0.5663,
"step": 1735
},
{
"epoch": 0.6984510158921746,
"grad_norm": 3.9729308104472225,
"learning_rate": 1.2662512664181826e-06,
"loss": 0.4753,
"step": 1736
},
{
"epoch": 0.6988533494266748,
"grad_norm": 3.4193536839822434,
"learning_rate": 1.2631988543112314e-06,
"loss": 0.3984,
"step": 1737
},
{
"epoch": 0.6992556829611748,
"grad_norm": 3.199663843957577,
"learning_rate": 1.2601488815175867e-06,
"loss": 0.3615,
"step": 1738
},
{
"epoch": 0.6996580164956749,
"grad_norm": 4.275095481044542,
"learning_rate": 1.257101354052635e-06,
"loss": 0.635,
"step": 1739
},
{
"epoch": 0.700060350030175,
"grad_norm": 4.154894699295693,
"learning_rate": 1.2540562779269421e-06,
"loss": 0.468,
"step": 1740
},
{
"epoch": 0.7004626835646751,
"grad_norm": 4.21271106717443,
"learning_rate": 1.2510136591462399e-06,
"loss": 0.557,
"step": 1741
},
{
"epoch": 0.7008650170991753,
"grad_norm": 3.9488032178250276,
"learning_rate": 1.2479735037114118e-06,
"loss": 0.4743,
"step": 1742
},
{
"epoch": 0.7012673506336753,
"grad_norm": 3.6437407466020453,
"learning_rate": 1.2449358176184848e-06,
"loss": 0.4958,
"step": 1743
},
{
"epoch": 0.7016696841681754,
"grad_norm": 4.011291060246906,
"learning_rate": 1.2419006068586125e-06,
"loss": 0.5109,
"step": 1744
},
{
"epoch": 0.7020720177026755,
"grad_norm": 3.7545390991024026,
"learning_rate": 1.2388678774180698e-06,
"loss": 0.5269,
"step": 1745
},
{
"epoch": 0.7024743512371756,
"grad_norm": 3.821504931037996,
"learning_rate": 1.2358376352782358e-06,
"loss": 0.4905,
"step": 1746
},
{
"epoch": 0.7028766847716758,
"grad_norm": 4.3580981207011655,
"learning_rate": 1.232809886415586e-06,
"loss": 0.4904,
"step": 1747
},
{
"epoch": 0.7032790183061758,
"grad_norm": 3.539055820090707,
"learning_rate": 1.2297846368016748e-06,
"loss": 0.4124,
"step": 1748
},
{
"epoch": 0.7036813518406759,
"grad_norm": 3.9844890147713254,
"learning_rate": 1.2267618924031317e-06,
"loss": 0.5321,
"step": 1749
},
{
"epoch": 0.7040836853751761,
"grad_norm": 4.178112023828336,
"learning_rate": 1.2237416591816412e-06,
"loss": 0.6362,
"step": 1750
},
{
"epoch": 0.7044860189096761,
"grad_norm": 4.0472275399599695,
"learning_rate": 1.220723943093939e-06,
"loss": 0.5101,
"step": 1751
},
{
"epoch": 0.7048883524441762,
"grad_norm": 4.312523214514084,
"learning_rate": 1.2177087500917925e-06,
"loss": 0.5277,
"step": 1752
},
{
"epoch": 0.7052906859786763,
"grad_norm": 4.133768938615848,
"learning_rate": 1.2146960861219954e-06,
"loss": 0.5271,
"step": 1753
},
{
"epoch": 0.7056930195131764,
"grad_norm": 4.430552507921185,
"learning_rate": 1.211685957126353e-06,
"loss": 0.711,
"step": 1754
},
{
"epoch": 0.7060953530476766,
"grad_norm": 4.119451270646705,
"learning_rate": 1.2086783690416704e-06,
"loss": 0.4358,
"step": 1755
},
{
"epoch": 0.7064976865821766,
"grad_norm": 4.295693476571627,
"learning_rate": 1.205673327799743e-06,
"loss": 0.5466,
"step": 1756
},
{
"epoch": 0.7069000201166767,
"grad_norm": 4.312446284251359,
"learning_rate": 1.2026708393273392e-06,
"loss": 0.5861,
"step": 1757
},
{
"epoch": 0.7073023536511768,
"grad_norm": 3.9109516304734733,
"learning_rate": 1.1996709095461967e-06,
"loss": 0.4782,
"step": 1758
},
{
"epoch": 0.7077046871856769,
"grad_norm": 4.425022693894618,
"learning_rate": 1.1966735443730052e-06,
"loss": 0.5627,
"step": 1759
},
{
"epoch": 0.7081070207201771,
"grad_norm": 4.259706435608531,
"learning_rate": 1.193678749719397e-06,
"loss": 0.5274,
"step": 1760
},
{
"epoch": 0.7085093542546771,
"grad_norm": 4.219902699982978,
"learning_rate": 1.190686531491932e-06,
"loss": 0.5397,
"step": 1761
},
{
"epoch": 0.7089116877891772,
"grad_norm": 4.105043115699033,
"learning_rate": 1.1876968955920916e-06,
"loss": 0.4845,
"step": 1762
},
{
"epoch": 0.7093140213236774,
"grad_norm": 3.905684008250733,
"learning_rate": 1.1847098479162644e-06,
"loss": 0.5769,
"step": 1763
},
{
"epoch": 0.7097163548581774,
"grad_norm": 4.0434341189775616,
"learning_rate": 1.1817253943557308e-06,
"loss": 0.4338,
"step": 1764
},
{
"epoch": 0.7101186883926776,
"grad_norm": 4.331648953621162,
"learning_rate": 1.1787435407966593e-06,
"loss": 0.6231,
"step": 1765
},
{
"epoch": 0.7105210219271776,
"grad_norm": 4.225800580764582,
"learning_rate": 1.1757642931200864e-06,
"loss": 0.4946,
"step": 1766
},
{
"epoch": 0.7109233554616777,
"grad_norm": 3.832934508222693,
"learning_rate": 1.1727876572019117e-06,
"loss": 0.5319,
"step": 1767
},
{
"epoch": 0.7113256889961779,
"grad_norm": 3.994089631167334,
"learning_rate": 1.1698136389128833e-06,
"loss": 0.5005,
"step": 1768
},
{
"epoch": 0.7117280225306779,
"grad_norm": 3.8399445750919425,
"learning_rate": 1.166842244118587e-06,
"loss": 0.4726,
"step": 1769
},
{
"epoch": 0.712130356065178,
"grad_norm": 4.043691249667768,
"learning_rate": 1.163873478679432e-06,
"loss": 0.4885,
"step": 1770
},
{
"epoch": 0.7125326895996781,
"grad_norm": 4.26586065507147,
"learning_rate": 1.1609073484506441e-06,
"loss": 0.5955,
"step": 1771
},
{
"epoch": 0.7129350231341782,
"grad_norm": 3.2723853825243006,
"learning_rate": 1.157943859282251e-06,
"loss": 0.393,
"step": 1772
},
{
"epoch": 0.7133373566686784,
"grad_norm": 4.228399556429077,
"learning_rate": 1.1549830170190714e-06,
"loss": 0.6138,
"step": 1773
},
{
"epoch": 0.7137396902031784,
"grad_norm": 3.703790105757895,
"learning_rate": 1.152024827500705e-06,
"loss": 0.4952,
"step": 1774
},
{
"epoch": 0.7141420237376785,
"grad_norm": 4.037603384099717,
"learning_rate": 1.149069296561516e-06,
"loss": 0.5791,
"step": 1775
},
{
"epoch": 0.7145443572721787,
"grad_norm": 3.958021889837016,
"learning_rate": 1.1461164300306292e-06,
"loss": 0.495,
"step": 1776
},
{
"epoch": 0.7149466908066787,
"grad_norm": 4.142430978673458,
"learning_rate": 1.1431662337319107e-06,
"loss": 0.6534,
"step": 1777
},
{
"epoch": 0.7153490243411789,
"grad_norm": 4.484413621549814,
"learning_rate": 1.1402187134839643e-06,
"loss": 0.4885,
"step": 1778
},
{
"epoch": 0.7157513578756789,
"grad_norm": 3.6695545389300395,
"learning_rate": 1.1372738751001111e-06,
"loss": 0.4446,
"step": 1779
},
{
"epoch": 0.716153691410179,
"grad_norm": 3.492829736247777,
"learning_rate": 1.134331724388387e-06,
"loss": 0.5245,
"step": 1780
},
{
"epoch": 0.7165560249446792,
"grad_norm": 3.8848835038152107,
"learning_rate": 1.1313922671515252e-06,
"loss": 0.5042,
"step": 1781
},
{
"epoch": 0.7169583584791792,
"grad_norm": 4.4510516838349705,
"learning_rate": 1.128455509186948e-06,
"loss": 0.6898,
"step": 1782
},
{
"epoch": 0.7173606920136794,
"grad_norm": 4.381017376532981,
"learning_rate": 1.1255214562867503e-06,
"loss": 0.5289,
"step": 1783
},
{
"epoch": 0.7177630255481794,
"grad_norm": 3.5079718266262416,
"learning_rate": 1.122590114237696e-06,
"loss": 0.4724,
"step": 1784
},
{
"epoch": 0.7181653590826795,
"grad_norm": 4.292226353007704,
"learning_rate": 1.1196614888212007e-06,
"loss": 0.5796,
"step": 1785
},
{
"epoch": 0.7185676926171797,
"grad_norm": 4.052087753788445,
"learning_rate": 1.1167355858133226e-06,
"loss": 0.4938,
"step": 1786
},
{
"epoch": 0.7189700261516797,
"grad_norm": 5.157980356295375,
"learning_rate": 1.1138124109847503e-06,
"loss": 0.5949,
"step": 1787
},
{
"epoch": 0.7193723596861799,
"grad_norm": 3.285940613285488,
"learning_rate": 1.1108919701007907e-06,
"loss": 0.3908,
"step": 1788
},
{
"epoch": 0.71977469322068,
"grad_norm": 3.5335717637010937,
"learning_rate": 1.1079742689213593e-06,
"loss": 0.5013,
"step": 1789
},
{
"epoch": 0.72017702675518,
"grad_norm": 4.450180240936036,
"learning_rate": 1.1050593132009703e-06,
"loss": 0.6329,
"step": 1790
},
{
"epoch": 0.7205793602896802,
"grad_norm": 4.404576812133769,
"learning_rate": 1.1021471086887187e-06,
"loss": 0.5781,
"step": 1791
},
{
"epoch": 0.7209816938241802,
"grad_norm": 3.3602741770117177,
"learning_rate": 1.0992376611282776e-06,
"loss": 0.3621,
"step": 1792
},
{
"epoch": 0.7213840273586803,
"grad_norm": 3.381272077924684,
"learning_rate": 1.0963309762578795e-06,
"loss": 0.384,
"step": 1793
},
{
"epoch": 0.7217863608931805,
"grad_norm": 3.5780173251972025,
"learning_rate": 1.0934270598103094e-06,
"loss": 0.5372,
"step": 1794
},
{
"epoch": 0.7221886944276805,
"grad_norm": 3.7738716419210814,
"learning_rate": 1.0905259175128931e-06,
"loss": 0.4336,
"step": 1795
},
{
"epoch": 0.7225910279621807,
"grad_norm": 3.820342940812247,
"learning_rate": 1.0876275550874846e-06,
"loss": 0.5335,
"step": 1796
},
{
"epoch": 0.7229933614966807,
"grad_norm": 4.176570559423123,
"learning_rate": 1.0847319782504532e-06,
"loss": 0.5038,
"step": 1797
},
{
"epoch": 0.7233956950311808,
"grad_norm": 3.67077010562448,
"learning_rate": 1.0818391927126764e-06,
"loss": 0.4342,
"step": 1798
},
{
"epoch": 0.723798028565681,
"grad_norm": 3.9880069700930423,
"learning_rate": 1.078949204179526e-06,
"loss": 0.4953,
"step": 1799
},
{
"epoch": 0.724200362100181,
"grad_norm": 4.161334696726195,
"learning_rate": 1.0760620183508582e-06,
"loss": 0.5799,
"step": 1800
},
{
"epoch": 0.7246026956346812,
"grad_norm": 4.126464363948578,
"learning_rate": 1.073177640920999e-06,
"loss": 0.5138,
"step": 1801
},
{
"epoch": 0.7250050291691813,
"grad_norm": 4.1895057334300505,
"learning_rate": 1.0702960775787374e-06,
"loss": 0.5121,
"step": 1802
},
{
"epoch": 0.7254073627036813,
"grad_norm": 3.884515999748438,
"learning_rate": 1.0674173340073121e-06,
"loss": 0.4891,
"step": 1803
},
{
"epoch": 0.7258096962381815,
"grad_norm": 4.286867137892461,
"learning_rate": 1.0645414158844012e-06,
"loss": 0.6118,
"step": 1804
},
{
"epoch": 0.7262120297726815,
"grad_norm": 3.8959234040057806,
"learning_rate": 1.0616683288821067e-06,
"loss": 0.5234,
"step": 1805
},
{
"epoch": 0.7266143633071817,
"grad_norm": 4.239978882514847,
"learning_rate": 1.058798078666952e-06,
"loss": 0.5183,
"step": 1806
},
{
"epoch": 0.7270166968416818,
"grad_norm": 3.7122065670739235,
"learning_rate": 1.0559306708998607e-06,
"loss": 0.501,
"step": 1807
},
{
"epoch": 0.7274190303761818,
"grad_norm": 4.556831231079228,
"learning_rate": 1.0530661112361533e-06,
"loss": 0.5257,
"step": 1808
},
{
"epoch": 0.727821363910682,
"grad_norm": 4.308685532929913,
"learning_rate": 1.050204405325534e-06,
"loss": 0.5534,
"step": 1809
},
{
"epoch": 0.728223697445182,
"grad_norm": 3.9297951909048368,
"learning_rate": 1.0473455588120738e-06,
"loss": 0.4681,
"step": 1810
},
{
"epoch": 0.7286260309796821,
"grad_norm": 3.98667448491781,
"learning_rate": 1.0444895773342092e-06,
"loss": 0.4785,
"step": 1811
},
{
"epoch": 0.7290283645141823,
"grad_norm": 4.2934600586771845,
"learning_rate": 1.0416364665247236e-06,
"loss": 0.5408,
"step": 1812
},
{
"epoch": 0.7294306980486823,
"grad_norm": 3.9347493779534,
"learning_rate": 1.0387862320107403e-06,
"loss": 0.5895,
"step": 1813
},
{
"epoch": 0.7298330315831825,
"grad_norm": 4.338941092246621,
"learning_rate": 1.0359388794137068e-06,
"loss": 0.5648,
"step": 1814
},
{
"epoch": 0.7302353651176826,
"grad_norm": 3.792647800273631,
"learning_rate": 1.0330944143493892e-06,
"loss": 0.5449,
"step": 1815
},
{
"epoch": 0.7306376986521826,
"grad_norm": 3.821766624740813,
"learning_rate": 1.0302528424278585e-06,
"loss": 0.4544,
"step": 1816
},
{
"epoch": 0.7310400321866828,
"grad_norm": 4.176558757203708,
"learning_rate": 1.0274141692534784e-06,
"loss": 0.4802,
"step": 1817
},
{
"epoch": 0.7314423657211828,
"grad_norm": 3.5626608894002803,
"learning_rate": 1.0245784004248968e-06,
"loss": 0.4273,
"step": 1818
},
{
"epoch": 0.731844699255683,
"grad_norm": 3.8719939712343936,
"learning_rate": 1.0217455415350316e-06,
"loss": 0.5409,
"step": 1819
},
{
"epoch": 0.7322470327901831,
"grad_norm": 4.537877117506395,
"learning_rate": 1.0189155981710642e-06,
"loss": 0.5705,
"step": 1820
},
{
"epoch": 0.7326493663246831,
"grad_norm": 3.932184586794533,
"learning_rate": 1.0160885759144217e-06,
"loss": 0.4904,
"step": 1821
},
{
"epoch": 0.7330516998591833,
"grad_norm": 4.266661778185627,
"learning_rate": 1.0132644803407753e-06,
"loss": 0.4769,
"step": 1822
},
{
"epoch": 0.7334540333936833,
"grad_norm": 3.7589693532265827,
"learning_rate": 1.0104433170200188e-06,
"loss": 0.4891,
"step": 1823
},
{
"epoch": 0.7338563669281835,
"grad_norm": 3.8562553041075263,
"learning_rate": 1.007625091516266e-06,
"loss": 0.417,
"step": 1824
},
{
"epoch": 0.7342587004626836,
"grad_norm": 4.130360177598847,
"learning_rate": 1.0048098093878352e-06,
"loss": 0.5207,
"step": 1825
},
{
"epoch": 0.7346610339971836,
"grad_norm": 4.16513812007776,
"learning_rate": 1.001997476187241e-06,
"loss": 0.5213,
"step": 1826
},
{
"epoch": 0.7350633675316838,
"grad_norm": 4.4208377164885055,
"learning_rate": 9.991880974611809e-07,
"loss": 0.6269,
"step": 1827
},
{
"epoch": 0.7354657010661839,
"grad_norm": 4.867032713419264,
"learning_rate": 9.96381678750524e-07,
"loss": 0.4808,
"step": 1828
},
{
"epoch": 0.735868034600684,
"grad_norm": 3.982545780740484,
"learning_rate": 9.935782255903034e-07,
"loss": 0.5066,
"step": 1829
},
{
"epoch": 0.7362703681351841,
"grad_norm": 4.286351965336647,
"learning_rate": 9.907777435097029e-07,
"loss": 0.5151,
"step": 1830
},
{
"epoch": 0.7366727016696841,
"grad_norm": 4.362229906545009,
"learning_rate": 9.879802380320473e-07,
"loss": 0.5936,
"step": 1831
},
{
"epoch": 0.7370750352041843,
"grad_norm": 4.660993080328005,
"learning_rate": 9.851857146747873e-07,
"loss": 0.6829,
"step": 1832
},
{
"epoch": 0.7374773687386844,
"grad_norm": 4.222997821431192,
"learning_rate": 9.82394178949497e-07,
"loss": 0.514,
"step": 1833
},
{
"epoch": 0.7378797022731844,
"grad_norm": 4.369792251301722,
"learning_rate": 9.796056363618533e-07,
"loss": 0.5742,
"step": 1834
},
{
"epoch": 0.7382820358076846,
"grad_norm": 4.422672878513299,
"learning_rate": 9.768200924116338e-07,
"loss": 0.568,
"step": 1835
},
{
"epoch": 0.7386843693421846,
"grad_norm": 4.507594407110887,
"learning_rate": 9.740375525926988e-07,
"loss": 0.6017,
"step": 1836
},
{
"epoch": 0.7390867028766848,
"grad_norm": 3.8125967279749835,
"learning_rate": 9.71258022392985e-07,
"loss": 0.5959,
"step": 1837
},
{
"epoch": 0.7394890364111849,
"grad_norm": 3.7273585439154586,
"learning_rate": 9.684815072944946e-07,
"loss": 0.3825,
"step": 1838
},
{
"epoch": 0.7398913699456849,
"grad_norm": 3.738805120372397,
"learning_rate": 9.657080127732807e-07,
"loss": 0.4931,
"step": 1839
},
{
"epoch": 0.7402937034801851,
"grad_norm": 3.9828190911127526,
"learning_rate": 9.629375442994418e-07,
"loss": 0.5182,
"step": 1840
},
{
"epoch": 0.7406960370146852,
"grad_norm": 3.9798516908695882,
"learning_rate": 9.601701073371045e-07,
"loss": 0.548,
"step": 1841
},
{
"epoch": 0.7410983705491853,
"grad_norm": 3.7199861649210715,
"learning_rate": 9.574057073444195e-07,
"loss": 0.468,
"step": 1842
},
{
"epoch": 0.7415007040836854,
"grad_norm": 4.0680484984054335,
"learning_rate": 9.546443497735467e-07,
"loss": 0.5782,
"step": 1843
},
{
"epoch": 0.7419030376181854,
"grad_norm": 3.4109362486846115,
"learning_rate": 9.518860400706465e-07,
"loss": 0.4245,
"step": 1844
},
{
"epoch": 0.7423053711526856,
"grad_norm": 3.931799967484987,
"learning_rate": 9.491307836758651e-07,
"loss": 0.4863,
"step": 1845
},
{
"epoch": 0.7427077046871857,
"grad_norm": 4.211928044347116,
"learning_rate": 9.463785860233301e-07,
"loss": 0.5459,
"step": 1846
},
{
"epoch": 0.7431100382216858,
"grad_norm": 4.0902800868050715,
"learning_rate": 9.436294525411357e-07,
"loss": 0.6375,
"step": 1847
},
{
"epoch": 0.7435123717561859,
"grad_norm": 3.524293359481542,
"learning_rate": 9.408833886513303e-07,
"loss": 0.444,
"step": 1848
},
{
"epoch": 0.7439147052906859,
"grad_norm": 3.6061388537955827,
"learning_rate": 9.381403997699117e-07,
"loss": 0.5217,
"step": 1849
},
{
"epoch": 0.7443170388251861,
"grad_norm": 3.6547603411472838,
"learning_rate": 9.354004913068096e-07,
"loss": 0.5229,
"step": 1850
},
{
"epoch": 0.7447193723596862,
"grad_norm": 4.340731245011316,
"learning_rate": 9.326636686658808e-07,
"loss": 0.5924,
"step": 1851
},
{
"epoch": 0.7451217058941862,
"grad_norm": 3.4273270658083166,
"learning_rate": 9.299299372448953e-07,
"loss": 0.4941,
"step": 1852
},
{
"epoch": 0.7455240394286864,
"grad_norm": 3.9330296000490668,
"learning_rate": 9.271993024355263e-07,
"loss": 0.5988,
"step": 1853
},
{
"epoch": 0.7459263729631865,
"grad_norm": 4.79047981984662,
"learning_rate": 9.244717696233388e-07,
"loss": 0.6193,
"step": 1854
},
{
"epoch": 0.7463287064976866,
"grad_norm": 3.6277035412847924,
"learning_rate": 9.217473441877805e-07,
"loss": 0.4905,
"step": 1855
},
{
"epoch": 0.7467310400321867,
"grad_norm": 4.652255454626382,
"learning_rate": 9.19026031502171e-07,
"loss": 0.6231,
"step": 1856
},
{
"epoch": 0.7471333735666867,
"grad_norm": 4.139619345656057,
"learning_rate": 9.1630783693369e-07,
"loss": 0.4981,
"step": 1857
},
{
"epoch": 0.7475357071011869,
"grad_norm": 5.059383612132635,
"learning_rate": 9.135927658433685e-07,
"loss": 0.7292,
"step": 1858
},
{
"epoch": 0.747938040635687,
"grad_norm": 4.664959181536374,
"learning_rate": 9.108808235860744e-07,
"loss": 0.6118,
"step": 1859
},
{
"epoch": 0.7483403741701871,
"grad_norm": 4.243108848129753,
"learning_rate": 9.081720155105076e-07,
"loss": 0.5616,
"step": 1860
},
{
"epoch": 0.7487427077046872,
"grad_norm": 3.4693576707426548,
"learning_rate": 9.054663469591862e-07,
"loss": 0.4451,
"step": 1861
},
{
"epoch": 0.7491450412391872,
"grad_norm": 4.037761942838199,
"learning_rate": 9.027638232684339e-07,
"loss": 0.5354,
"step": 1862
},
{
"epoch": 0.7495473747736874,
"grad_norm": 4.037140323875402,
"learning_rate": 9.00064449768375e-07,
"loss": 0.3946,
"step": 1863
},
{
"epoch": 0.7499497083081875,
"grad_norm": 4.420498117292749,
"learning_rate": 8.97368231782918e-07,
"loss": 0.6091,
"step": 1864
},
{
"epoch": 0.7503520418426876,
"grad_norm": 4.098542987137017,
"learning_rate": 8.946751746297494e-07,
"loss": 0.6347,
"step": 1865
},
{
"epoch": 0.7507543753771877,
"grad_norm": 3.718783314907544,
"learning_rate": 8.919852836203224e-07,
"loss": 0.5523,
"step": 1866
},
{
"epoch": 0.7511567089116878,
"grad_norm": 4.238567081312867,
"learning_rate": 8.892985640598434e-07,
"loss": 0.5716,
"step": 1867
},
{
"epoch": 0.7515590424461879,
"grad_norm": 3.878960051531824,
"learning_rate": 8.866150212472657e-07,
"loss": 0.5552,
"step": 1868
},
{
"epoch": 0.751961375980688,
"grad_norm": 4.0405823701369155,
"learning_rate": 8.839346604752763e-07,
"loss": 0.5286,
"step": 1869
},
{
"epoch": 0.752363709515188,
"grad_norm": 4.251404967808422,
"learning_rate": 8.81257487030287e-07,
"loss": 0.4872,
"step": 1870
},
{
"epoch": 0.7527660430496882,
"grad_norm": 3.6336762556932163,
"learning_rate": 8.785835061924234e-07,
"loss": 0.4301,
"step": 1871
},
{
"epoch": 0.7531683765841883,
"grad_norm": 4.324873818573403,
"learning_rate": 8.759127232355128e-07,
"loss": 0.5405,
"step": 1872
},
{
"epoch": 0.7535707101186884,
"grad_norm": 4.42447965555661,
"learning_rate": 8.732451434270767e-07,
"loss": 0.5209,
"step": 1873
},
{
"epoch": 0.7539730436531885,
"grad_norm": 4.251290292687751,
"learning_rate": 8.705807720283199e-07,
"loss": 0.5951,
"step": 1874
},
{
"epoch": 0.7543753771876885,
"grad_norm": 4.171584555966765,
"learning_rate": 8.679196142941173e-07,
"loss": 0.6086,
"step": 1875
},
{
"epoch": 0.7547777107221887,
"grad_norm": 4.0958958819445614,
"learning_rate": 8.652616754730075e-07,
"loss": 0.5693,
"step": 1876
},
{
"epoch": 0.7551800442566888,
"grad_norm": 3.683062028222347,
"learning_rate": 8.626069608071785e-07,
"loss": 0.4584,
"step": 1877
},
{
"epoch": 0.7555823777911889,
"grad_norm": 3.9012710111822924,
"learning_rate": 8.599554755324613e-07,
"loss": 0.4611,
"step": 1878
},
{
"epoch": 0.755984711325689,
"grad_norm": 3.5489182194802704,
"learning_rate": 8.573072248783165e-07,
"loss": 0.5083,
"step": 1879
},
{
"epoch": 0.7563870448601892,
"grad_norm": 3.9884722155489225,
"learning_rate": 8.546622140678265e-07,
"loss": 0.5094,
"step": 1880
},
{
"epoch": 0.7567893783946892,
"grad_norm": 4.328049459143733,
"learning_rate": 8.520204483176814e-07,
"loss": 0.3656,
"step": 1881
},
{
"epoch": 0.7571917119291893,
"grad_norm": 3.9339309591347087,
"learning_rate": 8.493819328381731e-07,
"loss": 0.5317,
"step": 1882
},
{
"epoch": 0.7575940454636894,
"grad_norm": 4.2270154261756385,
"learning_rate": 8.467466728331828e-07,
"loss": 0.4588,
"step": 1883
},
{
"epoch": 0.7579963789981895,
"grad_norm": 3.303945186981914,
"learning_rate": 8.441146735001715e-07,
"loss": 0.4649,
"step": 1884
},
{
"epoch": 0.7583987125326896,
"grad_norm": 4.054188026080457,
"learning_rate": 8.414859400301667e-07,
"loss": 0.5645,
"step": 1885
},
{
"epoch": 0.7588010460671897,
"grad_norm": 4.17992148808867,
"learning_rate": 8.388604776077575e-07,
"loss": 0.5204,
"step": 1886
},
{
"epoch": 0.7592033796016898,
"grad_norm": 4.428173356861454,
"learning_rate": 8.362382914110804e-07,
"loss": 0.6007,
"step": 1887
},
{
"epoch": 0.7596057131361899,
"grad_norm": 4.255468940795954,
"learning_rate": 8.336193866118117e-07,
"loss": 0.5153,
"step": 1888
},
{
"epoch": 0.76000804667069,
"grad_norm": 4.025791541260584,
"learning_rate": 8.310037683751527e-07,
"loss": 0.4859,
"step": 1889
},
{
"epoch": 0.7604103802051901,
"grad_norm": 3.8573782279099533,
"learning_rate": 8.283914418598263e-07,
"loss": 0.5058,
"step": 1890
},
{
"epoch": 0.7608127137396902,
"grad_norm": 3.332235385366874,
"learning_rate": 8.257824122180602e-07,
"loss": 0.3737,
"step": 1891
},
{
"epoch": 0.7612150472741903,
"grad_norm": 3.394561752123056,
"learning_rate": 8.231766845955818e-07,
"loss": 0.4197,
"step": 1892
},
{
"epoch": 0.7616173808086905,
"grad_norm": 3.3171785759316443,
"learning_rate": 8.205742641316064e-07,
"loss": 0.417,
"step": 1893
},
{
"epoch": 0.7620197143431905,
"grad_norm": 4.487532016958629,
"learning_rate": 8.179751559588234e-07,
"loss": 0.5663,
"step": 1894
},
{
"epoch": 0.7624220478776906,
"grad_norm": 4.546271967177341,
"learning_rate": 8.153793652033931e-07,
"loss": 0.5597,
"step": 1895
},
{
"epoch": 0.7628243814121907,
"grad_norm": 3.914164846084329,
"learning_rate": 8.12786896984931e-07,
"loss": 0.4677,
"step": 1896
},
{
"epoch": 0.7632267149466908,
"grad_norm": 3.852819616304247,
"learning_rate": 8.101977564165011e-07,
"loss": 0.503,
"step": 1897
},
{
"epoch": 0.763629048481191,
"grad_norm": 4.067929177534588,
"learning_rate": 8.076119486046019e-07,
"loss": 0.5396,
"step": 1898
},
{
"epoch": 0.764031382015691,
"grad_norm": 4.010571344025329,
"learning_rate": 8.050294786491611e-07,
"loss": 0.4829,
"step": 1899
},
{
"epoch": 0.7644337155501911,
"grad_norm": 4.199207618578668,
"learning_rate": 8.024503516435222e-07,
"loss": 0.6123,
"step": 1900
},
{
"epoch": 0.7648360490846912,
"grad_norm": 4.313823267969129,
"learning_rate": 7.998745726744358e-07,
"loss": 0.5597,
"step": 1901
},
{
"epoch": 0.7652383826191913,
"grad_norm": 3.7661528466618956,
"learning_rate": 7.973021468220502e-07,
"loss": 0.5036,
"step": 1902
},
{
"epoch": 0.7656407161536914,
"grad_norm": 4.017310363498917,
"learning_rate": 7.947330791598973e-07,
"loss": 0.5435,
"step": 1903
},
{
"epoch": 0.7660430496881915,
"grad_norm": 4.447099011053462,
"learning_rate": 7.921673747548895e-07,
"loss": 0.4892,
"step": 1904
},
{
"epoch": 0.7664453832226916,
"grad_norm": 4.1491367299305555,
"learning_rate": 7.896050386673026e-07,
"loss": 0.5234,
"step": 1905
},
{
"epoch": 0.7668477167571918,
"grad_norm": 4.640959967611087,
"learning_rate": 7.870460759507722e-07,
"loss": 0.5619,
"step": 1906
},
{
"epoch": 0.7672500502916918,
"grad_norm": 4.207681163367058,
"learning_rate": 7.844904916522784e-07,
"loss": 0.6313,
"step": 1907
},
{
"epoch": 0.7676523838261919,
"grad_norm": 4.624725241802207,
"learning_rate": 7.819382908121386e-07,
"loss": 0.6169,
"step": 1908
},
{
"epoch": 0.768054717360692,
"grad_norm": 3.891844454436063,
"learning_rate": 7.793894784639986e-07,
"loss": 0.4932,
"step": 1909
},
{
"epoch": 0.7684570508951921,
"grad_norm": 3.6005349724480573,
"learning_rate": 7.768440596348192e-07,
"loss": 0.5835,
"step": 1910
},
{
"epoch": 0.7688593844296923,
"grad_norm": 4.073804566700674,
"learning_rate": 7.743020393448702e-07,
"loss": 0.4882,
"step": 1911
},
{
"epoch": 0.7692617179641923,
"grad_norm": 3.959969867245135,
"learning_rate": 7.717634226077156e-07,
"loss": 0.4945,
"step": 1912
},
{
"epoch": 0.7696640514986924,
"grad_norm": 3.599166244946097,
"learning_rate": 7.692282144302094e-07,
"loss": 0.3889,
"step": 1913
},
{
"epoch": 0.7700663850331925,
"grad_norm": 3.539375774753211,
"learning_rate": 7.666964198124819e-07,
"loss": 0.523,
"step": 1914
},
{
"epoch": 0.7704687185676926,
"grad_norm": 3.912187073942566,
"learning_rate": 7.64168043747932e-07,
"loss": 0.5567,
"step": 1915
},
{
"epoch": 0.7708710521021928,
"grad_norm": 4.034008046236846,
"learning_rate": 7.616430912232137e-07,
"loss": 0.5219,
"step": 1916
},
{
"epoch": 0.7712733856366928,
"grad_norm": 4.074005253663992,
"learning_rate": 7.59121567218232e-07,
"loss": 0.5302,
"step": 1917
},
{
"epoch": 0.7716757191711929,
"grad_norm": 3.4776790939949347,
"learning_rate": 7.566034767061265e-07,
"loss": 0.2992,
"step": 1918
},
{
"epoch": 0.7720780527056931,
"grad_norm": 3.755741404325393,
"learning_rate": 7.540888246532688e-07,
"loss": 0.4337,
"step": 1919
},
{
"epoch": 0.7724803862401931,
"grad_norm": 3.854464881913325,
"learning_rate": 7.515776160192454e-07,
"loss": 0.5286,
"step": 1920
},
{
"epoch": 0.7728827197746932,
"grad_norm": 3.74441109291879,
"learning_rate": 7.490698557568535e-07,
"loss": 0.5449,
"step": 1921
},
{
"epoch": 0.7732850533091933,
"grad_norm": 4.570844263947877,
"learning_rate": 7.465655488120891e-07,
"loss": 0.6153,
"step": 1922
},
{
"epoch": 0.7736873868436934,
"grad_norm": 3.7650085288496613,
"learning_rate": 7.440647001241363e-07,
"loss": 0.5166,
"step": 1923
},
{
"epoch": 0.7740897203781936,
"grad_norm": 3.6719483984817236,
"learning_rate": 7.415673146253604e-07,
"loss": 0.4842,
"step": 1924
},
{
"epoch": 0.7744920539126936,
"grad_norm": 4.057471839423865,
"learning_rate": 7.390733972412933e-07,
"loss": 0.6075,
"step": 1925
},
{
"epoch": 0.7748943874471937,
"grad_norm": 4.4626847789976924,
"learning_rate": 7.365829528906293e-07,
"loss": 0.5292,
"step": 1926
},
{
"epoch": 0.7752967209816938,
"grad_norm": 3.76677011259477,
"learning_rate": 7.340959864852124e-07,
"loss": 0.405,
"step": 1927
},
{
"epoch": 0.7756990545161939,
"grad_norm": 3.8469846124186082,
"learning_rate": 7.316125029300275e-07,
"loss": 0.4827,
"step": 1928
},
{
"epoch": 0.7761013880506941,
"grad_norm": 3.8371445392089827,
"learning_rate": 7.291325071231883e-07,
"loss": 0.5134,
"step": 1929
},
{
"epoch": 0.7765037215851941,
"grad_norm": 3.8383451450589887,
"learning_rate": 7.26656003955932e-07,
"loss": 0.4061,
"step": 1930
},
{
"epoch": 0.7769060551196942,
"grad_norm": 4.037666657124557,
"learning_rate": 7.241829983126075e-07,
"loss": 0.4976,
"step": 1931
},
{
"epoch": 0.7773083886541944,
"grad_norm": 3.980302342869356,
"learning_rate": 7.217134950706625e-07,
"loss": 0.5149,
"step": 1932
},
{
"epoch": 0.7777107221886944,
"grad_norm": 3.7463100803001277,
"learning_rate": 7.192474991006416e-07,
"loss": 0.459,
"step": 1933
},
{
"epoch": 0.7781130557231946,
"grad_norm": 3.9126255555142597,
"learning_rate": 7.167850152661673e-07,
"loss": 0.5537,
"step": 1934
},
{
"epoch": 0.7785153892576946,
"grad_norm": 4.306777769767276,
"learning_rate": 7.143260484239386e-07,
"loss": 0.4997,
"step": 1935
},
{
"epoch": 0.7789177227921947,
"grad_norm": 4.472035583740821,
"learning_rate": 7.118706034237172e-07,
"loss": 0.6373,
"step": 1936
},
{
"epoch": 0.7793200563266949,
"grad_norm": 3.8082808763272356,
"learning_rate": 7.094186851083187e-07,
"loss": 0.5713,
"step": 1937
},
{
"epoch": 0.7797223898611949,
"grad_norm": 3.907355998411772,
"learning_rate": 7.069702983136023e-07,
"loss": 0.4916,
"step": 1938
},
{
"epoch": 0.780124723395695,
"grad_norm": 3.4404115391335623,
"learning_rate": 7.045254478684626e-07,
"loss": 0.4552,
"step": 1939
},
{
"epoch": 0.7805270569301951,
"grad_norm": 4.072479874224982,
"learning_rate": 7.020841385948199e-07,
"loss": 0.5061,
"step": 1940
},
{
"epoch": 0.7809293904646952,
"grad_norm": 5.014496338999608,
"learning_rate": 6.996463753076102e-07,
"loss": 0.5401,
"step": 1941
},
{
"epoch": 0.7813317239991954,
"grad_norm": 4.5982221422710134,
"learning_rate": 6.972121628147766e-07,
"loss": 0.6385,
"step": 1942
},
{
"epoch": 0.7817340575336954,
"grad_norm": 3.866309336941733,
"learning_rate": 6.94781505917256e-07,
"loss": 0.5007,
"step": 1943
},
{
"epoch": 0.7821363910681955,
"grad_norm": 4.000641087223532,
"learning_rate": 6.923544094089762e-07,
"loss": 0.4097,
"step": 1944
},
{
"epoch": 0.7825387246026957,
"grad_norm": 4.183627371809607,
"learning_rate": 6.899308780768423e-07,
"loss": 0.4986,
"step": 1945
},
{
"epoch": 0.7829410581371957,
"grad_norm": 3.834784873841407,
"learning_rate": 6.875109167007255e-07,
"loss": 0.4719,
"step": 1946
},
{
"epoch": 0.7833433916716959,
"grad_norm": 4.624045837237235,
"learning_rate": 6.85094530053459e-07,
"loss": 0.5742,
"step": 1947
},
{
"epoch": 0.7837457252061959,
"grad_norm": 4.968557042198568,
"learning_rate": 6.826817229008232e-07,
"loss": 0.6943,
"step": 1948
},
{
"epoch": 0.784148058740696,
"grad_norm": 3.798467897259186,
"learning_rate": 6.802725000015406e-07,
"loss": 0.5297,
"step": 1949
},
{
"epoch": 0.7845503922751962,
"grad_norm": 3.974644659610978,
"learning_rate": 6.778668661072643e-07,
"loss": 0.4803,
"step": 1950
},
{
"epoch": 0.7849527258096962,
"grad_norm": 4.048333500653427,
"learning_rate": 6.754648259625673e-07,
"loss": 0.5941,
"step": 1951
},
{
"epoch": 0.7853550593441964,
"grad_norm": 4.3035958811807875,
"learning_rate": 6.730663843049367e-07,
"loss": 0.5955,
"step": 1952
},
{
"epoch": 0.7857573928786964,
"grad_norm": 3.6544635394961764,
"learning_rate": 6.706715458647615e-07,
"loss": 0.4851,
"step": 1953
},
{
"epoch": 0.7861597264131965,
"grad_norm": 4.218917868283341,
"learning_rate": 6.682803153653241e-07,
"loss": 0.6081,
"step": 1954
},
{
"epoch": 0.7865620599476967,
"grad_norm": 3.617134711107577,
"learning_rate": 6.658926975227923e-07,
"loss": 0.4414,
"step": 1955
},
{
"epoch": 0.7869643934821967,
"grad_norm": 4.041925903109421,
"learning_rate": 6.635086970462057e-07,
"loss": 0.4589,
"step": 1956
},
{
"epoch": 0.7873667270166969,
"grad_norm": 3.9272206421591944,
"learning_rate": 6.611283186374723e-07,
"loss": 0.4697,
"step": 1957
},
{
"epoch": 0.7877690605511969,
"grad_norm": 4.242979124153208,
"learning_rate": 6.587515669913558e-07,
"loss": 0.4185,
"step": 1958
},
{
"epoch": 0.788171394085697,
"grad_norm": 3.38359591341345,
"learning_rate": 6.563784467954668e-07,
"loss": 0.3828,
"step": 1959
},
{
"epoch": 0.7885737276201972,
"grad_norm": 4.348954859955942,
"learning_rate": 6.540089627302523e-07,
"loss": 0.5046,
"step": 1960
},
{
"epoch": 0.7889760611546972,
"grad_norm": 4.120739882383607,
"learning_rate": 6.516431194689907e-07,
"loss": 0.5227,
"step": 1961
},
{
"epoch": 0.7893783946891973,
"grad_norm": 3.6740741354898305,
"learning_rate": 6.492809216777762e-07,
"loss": 0.4134,
"step": 1962
},
{
"epoch": 0.7897807282236975,
"grad_norm": 3.690318810926642,
"learning_rate": 6.469223740155159e-07,
"loss": 0.4901,
"step": 1963
},
{
"epoch": 0.7901830617581975,
"grad_norm": 4.2381781760809485,
"learning_rate": 6.445674811339175e-07,
"loss": 0.4225,
"step": 1964
},
{
"epoch": 0.7905853952926977,
"grad_norm": 4.106368765680081,
"learning_rate": 6.422162476774788e-07,
"loss": 0.5645,
"step": 1965
},
{
"epoch": 0.7909877288271977,
"grad_norm": 3.706517698008605,
"learning_rate": 6.398686782834815e-07,
"loss": 0.4995,
"step": 1966
},
{
"epoch": 0.7913900623616978,
"grad_norm": 4.432403407342872,
"learning_rate": 6.375247775819809e-07,
"loss": 0.5127,
"step": 1967
},
{
"epoch": 0.791792395896198,
"grad_norm": 4.022769337353194,
"learning_rate": 6.351845501957971e-07,
"loss": 0.4525,
"step": 1968
},
{
"epoch": 0.792194729430698,
"grad_norm": 4.43806071523494,
"learning_rate": 6.328480007405027e-07,
"loss": 0.5067,
"step": 1969
},
{
"epoch": 0.7925970629651982,
"grad_norm": 3.7265320273813027,
"learning_rate": 6.305151338244192e-07,
"loss": 0.4277,
"step": 1970
},
{
"epoch": 0.7929993964996982,
"grad_norm": 4.06221248413846,
"learning_rate": 6.281859540486043e-07,
"loss": 0.6297,
"step": 1971
},
{
"epoch": 0.7934017300341983,
"grad_norm": 4.391919257148749,
"learning_rate": 6.258604660068443e-07,
"loss": 0.4994,
"step": 1972
},
{
"epoch": 0.7938040635686985,
"grad_norm": 4.252838693305146,
"learning_rate": 6.23538674285642e-07,
"loss": 0.5949,
"step": 1973
},
{
"epoch": 0.7942063971031985,
"grad_norm": 4.325947652020019,
"learning_rate": 6.212205834642132e-07,
"loss": 0.5428,
"step": 1974
},
{
"epoch": 0.7946087306376987,
"grad_norm": 4.185176650642697,
"learning_rate": 6.189061981144715e-07,
"loss": 0.5732,
"step": 1975
},
{
"epoch": 0.7950110641721988,
"grad_norm": 3.2994931408989627,
"learning_rate": 6.165955228010242e-07,
"loss": 0.37,
"step": 1976
},
{
"epoch": 0.7954133977066988,
"grad_norm": 3.9581400842405277,
"learning_rate": 6.14288562081162e-07,
"loss": 0.6003,
"step": 1977
},
{
"epoch": 0.795815731241199,
"grad_norm": 3.855839195313672,
"learning_rate": 6.11985320504847e-07,
"loss": 0.5755,
"step": 1978
},
{
"epoch": 0.796218064775699,
"grad_norm": 4.28483670710571,
"learning_rate": 6.096858026147079e-07,
"loss": 0.5674,
"step": 1979
},
{
"epoch": 0.7966203983101992,
"grad_norm": 3.946092310322622,
"learning_rate": 6.073900129460286e-07,
"loss": 0.4429,
"step": 1980
},
{
"epoch": 0.7970227318446993,
"grad_norm": 3.9442626812628725,
"learning_rate": 6.050979560267414e-07,
"loss": 0.4953,
"step": 1981
},
{
"epoch": 0.7974250653791993,
"grad_norm": 4.023203847345498,
"learning_rate": 6.028096363774136e-07,
"loss": 0.5437,
"step": 1982
},
{
"epoch": 0.7978273989136995,
"grad_norm": 4.5536613551126335,
"learning_rate": 6.005250585112437e-07,
"loss": 0.6177,
"step": 1983
},
{
"epoch": 0.7982297324481995,
"grad_norm": 3.8312030174786313,
"learning_rate": 5.982442269340505e-07,
"loss": 0.4801,
"step": 1984
},
{
"epoch": 0.7986320659826996,
"grad_norm": 3.8614656531277896,
"learning_rate": 5.959671461442634e-07,
"loss": 0.4395,
"step": 1985
},
{
"epoch": 0.7990343995171998,
"grad_norm": 4.191209006767766,
"learning_rate": 5.936938206329148e-07,
"loss": 0.4913,
"step": 1986
},
{
"epoch": 0.7994367330516998,
"grad_norm": 4.201209209604027,
"learning_rate": 5.914242548836291e-07,
"loss": 0.567,
"step": 1987
},
{
"epoch": 0.7998390665862,
"grad_norm": 4.711191490796992,
"learning_rate": 5.891584533726177e-07,
"loss": 0.6445,
"step": 1988
},
{
"epoch": 0.8002414001207001,
"grad_norm": 4.050282830317415,
"learning_rate": 5.868964205686653e-07,
"loss": 0.4854,
"step": 1989
},
{
"epoch": 0.8006437336552001,
"grad_norm": 3.4477078683288838,
"learning_rate": 5.84638160933127e-07,
"loss": 0.4578,
"step": 1990
},
{
"epoch": 0.8010460671897003,
"grad_norm": 3.977635083967608,
"learning_rate": 5.82383678919912e-07,
"loss": 0.4966,
"step": 1991
},
{
"epoch": 0.8014484007242003,
"grad_norm": 4.254984216161458,
"learning_rate": 5.801329789754826e-07,
"loss": 0.5101,
"step": 1992
},
{
"epoch": 0.8018507342587005,
"grad_norm": 4.1472621686996005,
"learning_rate": 5.778860655388399e-07,
"loss": 0.6366,
"step": 1993
},
{
"epoch": 0.8022530677932006,
"grad_norm": 4.0712688803226715,
"learning_rate": 5.75642943041518e-07,
"loss": 0.4908,
"step": 1994
},
{
"epoch": 0.8026554013277006,
"grad_norm": 4.69998065888643,
"learning_rate": 5.734036159075743e-07,
"loss": 0.6769,
"step": 1995
},
{
"epoch": 0.8030577348622008,
"grad_norm": 3.5455182945935144,
"learning_rate": 5.711680885535784e-07,
"loss": 0.4177,
"step": 1996
},
{
"epoch": 0.8034600683967008,
"grad_norm": 3.853566513480482,
"learning_rate": 5.689363653886085e-07,
"loss": 0.4434,
"step": 1997
},
{
"epoch": 0.803862401931201,
"grad_norm": 4.050333441948282,
"learning_rate": 5.667084508142386e-07,
"loss": 0.4822,
"step": 1998
},
{
"epoch": 0.8042647354657011,
"grad_norm": 4.442980318720399,
"learning_rate": 5.644843492245322e-07,
"loss": 0.4872,
"step": 1999
},
{
"epoch": 0.8046670690002011,
"grad_norm": 4.457653531733048,
"learning_rate": 5.622640650060301e-07,
"loss": 0.6444,
"step": 2000
},
{
"epoch": 0.8050694025347013,
"grad_norm": 3.921778521793536,
"learning_rate": 5.600476025377463e-07,
"loss": 0.5391,
"step": 2001
},
{
"epoch": 0.8054717360692014,
"grad_norm": 3.8081912666035755,
"learning_rate": 5.578349661911578e-07,
"loss": 0.4802,
"step": 2002
},
{
"epoch": 0.8058740696037014,
"grad_norm": 4.001868301091059,
"learning_rate": 5.556261603301924e-07,
"loss": 0.5459,
"step": 2003
},
{
"epoch": 0.8062764031382016,
"grad_norm": 3.847308434546458,
"learning_rate": 5.534211893112268e-07,
"loss": 0.4606,
"step": 2004
},
{
"epoch": 0.8066787366727016,
"grad_norm": 4.15027221954597,
"learning_rate": 5.512200574830712e-07,
"loss": 0.5705,
"step": 2005
},
{
"epoch": 0.8070810702072018,
"grad_norm": 3.8098780638264027,
"learning_rate": 5.490227691869656e-07,
"loss": 0.491,
"step": 2006
},
{
"epoch": 0.8074834037417019,
"grad_norm": 3.9362205164209616,
"learning_rate": 5.468293287565696e-07,
"loss": 0.4342,
"step": 2007
},
{
"epoch": 0.8078857372762019,
"grad_norm": 3.621695424225381,
"learning_rate": 5.446397405179534e-07,
"loss": 0.4943,
"step": 2008
},
{
"epoch": 0.8082880708107021,
"grad_norm": 4.894139218776488,
"learning_rate": 5.424540087895888e-07,
"loss": 0.7108,
"step": 2009
},
{
"epoch": 0.8086904043452021,
"grad_norm": 4.43306730596569,
"learning_rate": 5.402721378823422e-07,
"loss": 0.5759,
"step": 2010
},
{
"epoch": 0.8090927378797023,
"grad_norm": 4.303012456014276,
"learning_rate": 5.380941320994659e-07,
"loss": 0.575,
"step": 2011
},
{
"epoch": 0.8094950714142024,
"grad_norm": 4.313613214512251,
"learning_rate": 5.35919995736589e-07,
"loss": 0.4761,
"step": 2012
},
{
"epoch": 0.8098974049487024,
"grad_norm": 3.8521996380430394,
"learning_rate": 5.337497330817074e-07,
"loss": 0.5352,
"step": 2013
},
{
"epoch": 0.8102997384832026,
"grad_norm": 4.528703112399277,
"learning_rate": 5.315833484151786e-07,
"loss": 0.5959,
"step": 2014
},
{
"epoch": 0.8107020720177027,
"grad_norm": 3.878303004429545,
"learning_rate": 5.294208460097125e-07,
"loss": 0.4815,
"step": 2015
},
{
"epoch": 0.8111044055522028,
"grad_norm": 4.490283907824032,
"learning_rate": 5.272622301303587e-07,
"loss": 0.6368,
"step": 2016
},
{
"epoch": 0.8115067390867029,
"grad_norm": 4.279104111581364,
"learning_rate": 5.251075050345056e-07,
"loss": 0.5216,
"step": 2017
},
{
"epoch": 0.8119090726212029,
"grad_norm": 3.6608651815923086,
"learning_rate": 5.229566749718645e-07,
"loss": 0.4809,
"step": 2018
},
{
"epoch": 0.8123114061557031,
"grad_norm": 4.666782295826111,
"learning_rate": 5.208097441844668e-07,
"loss": 0.504,
"step": 2019
},
{
"epoch": 0.8127137396902032,
"grad_norm": 3.84925143665575,
"learning_rate": 5.186667169066523e-07,
"loss": 0.5423,
"step": 2020
},
{
"epoch": 0.8131160732247033,
"grad_norm": 3.578156563304788,
"learning_rate": 5.165275973650633e-07,
"loss": 0.3936,
"step": 2021
},
{
"epoch": 0.8135184067592034,
"grad_norm": 4.074315611637044,
"learning_rate": 5.143923897786329e-07,
"loss": 0.6302,
"step": 2022
},
{
"epoch": 0.8139207402937034,
"grad_norm": 4.085845386699454,
"learning_rate": 5.122610983585808e-07,
"loss": 0.5789,
"step": 2023
},
{
"epoch": 0.8143230738282036,
"grad_norm": 4.000394354287652,
"learning_rate": 5.101337273084017e-07,
"loss": 0.4229,
"step": 2024
},
{
"epoch": 0.8147254073627037,
"grad_norm": 3.835079915243062,
"learning_rate": 5.080102808238585e-07,
"loss": 0.5005,
"step": 2025
},
{
"epoch": 0.8151277408972037,
"grad_norm": 3.590556700281588,
"learning_rate": 5.058907630929749e-07,
"loss": 0.4452,
"step": 2026
},
{
"epoch": 0.8155300744317039,
"grad_norm": 4.254640813185372,
"learning_rate": 5.037751782960234e-07,
"loss": 0.6328,
"step": 2027
},
{
"epoch": 0.815932407966204,
"grad_norm": 4.290085466017056,
"learning_rate": 5.016635306055223e-07,
"loss": 0.5785,
"step": 2028
},
{
"epoch": 0.8163347415007041,
"grad_norm": 4.159416054122212,
"learning_rate": 4.995558241862242e-07,
"loss": 0.5129,
"step": 2029
},
{
"epoch": 0.8167370750352042,
"grad_norm": 3.7061504996754153,
"learning_rate": 4.974520631951069e-07,
"loss": 0.5305,
"step": 2030
},
{
"epoch": 0.8171394085697042,
"grad_norm": 4.178397491082031,
"learning_rate": 4.953522517813689e-07,
"loss": 0.5855,
"step": 2031
},
{
"epoch": 0.8175417421042044,
"grad_norm": 4.154495627484645,
"learning_rate": 4.932563940864168e-07,
"loss": 0.4164,
"step": 2032
},
{
"epoch": 0.8179440756387045,
"grad_norm": 4.283562955009001,
"learning_rate": 4.911644942438614e-07,
"loss": 0.4942,
"step": 2033
},
{
"epoch": 0.8183464091732046,
"grad_norm": 3.307871102947408,
"learning_rate": 4.890765563795072e-07,
"loss": 0.4552,
"step": 2034
},
{
"epoch": 0.8187487427077047,
"grad_norm": 4.15735416688211,
"learning_rate": 4.86992584611343e-07,
"loss": 0.5797,
"step": 2035
},
{
"epoch": 0.8191510762422047,
"grad_norm": 4.111111195260256,
"learning_rate": 4.849125830495366e-07,
"loss": 0.5776,
"step": 2036
},
{
"epoch": 0.8195534097767049,
"grad_norm": 3.534887324822181,
"learning_rate": 4.828365557964257e-07,
"loss": 0.4913,
"step": 2037
},
{
"epoch": 0.819955743311205,
"grad_norm": 3.8702017938683606,
"learning_rate": 4.807645069465095e-07,
"loss": 0.4638,
"step": 2038
},
{
"epoch": 0.820358076845705,
"grad_norm": 4.105810885704208,
"learning_rate": 4.786964405864403e-07,
"loss": 0.51,
"step": 2039
},
{
"epoch": 0.8207604103802052,
"grad_norm": 3.7721413153025183,
"learning_rate": 4.766323607950149e-07,
"loss": 0.5327,
"step": 2040
},
{
"epoch": 0.8211627439147053,
"grad_norm": 4.604383899796593,
"learning_rate": 4.7457227164316914e-07,
"loss": 0.6425,
"step": 2041
},
{
"epoch": 0.8215650774492054,
"grad_norm": 4.7452724983070595,
"learning_rate": 4.725161771939679e-07,
"loss": 0.5968,
"step": 2042
},
{
"epoch": 0.8219674109837055,
"grad_norm": 4.13939509952924,
"learning_rate": 4.704640815025971e-07,
"loss": 0.5454,
"step": 2043
},
{
"epoch": 0.8223697445182055,
"grad_norm": 4.07811911040908,
"learning_rate": 4.684159886163553e-07,
"loss": 0.4846,
"step": 2044
},
{
"epoch": 0.8227720780527057,
"grad_norm": 4.223857527886817,
"learning_rate": 4.663719025746477e-07,
"loss": 0.548,
"step": 2045
},
{
"epoch": 0.8231744115872058,
"grad_norm": 4.055398221449843,
"learning_rate": 4.643318274089756e-07,
"loss": 0.4762,
"step": 2046
},
{
"epoch": 0.8235767451217059,
"grad_norm": 3.93292252326897,
"learning_rate": 4.6229576714293067e-07,
"loss": 0.5506,
"step": 2047
},
{
"epoch": 0.823979078656206,
"grad_norm": 3.5755501548250357,
"learning_rate": 4.6026372579218686e-07,
"loss": 0.5277,
"step": 2048
},
{
"epoch": 0.824381412190706,
"grad_norm": 3.900846576287911,
"learning_rate": 4.5823570736448913e-07,
"loss": 0.4289,
"step": 2049
},
{
"epoch": 0.8247837457252062,
"grad_norm": 4.422345897561269,
"learning_rate": 4.5621171585965046e-07,
"loss": 0.5439,
"step": 2050
},
{
"epoch": 0.8251860792597063,
"grad_norm": 3.86188238287213,
"learning_rate": 4.5419175526954074e-07,
"loss": 0.4971,
"step": 2051
},
{
"epoch": 0.8255884127942064,
"grad_norm": 4.149840183197456,
"learning_rate": 4.5217582957808064e-07,
"loss": 0.4913,
"step": 2052
},
{
"epoch": 0.8259907463287065,
"grad_norm": 4.14442670188834,
"learning_rate": 4.501639427612309e-07,
"loss": 0.593,
"step": 2053
},
{
"epoch": 0.8263930798632066,
"grad_norm": 4.160435006991155,
"learning_rate": 4.4815609878698847e-07,
"loss": 0.5702,
"step": 2054
},
{
"epoch": 0.8267954133977067,
"grad_norm": 4.699994325668033,
"learning_rate": 4.46152301615376e-07,
"loss": 0.6052,
"step": 2055
},
{
"epoch": 0.8271977469322068,
"grad_norm": 4.158216057441291,
"learning_rate": 4.441525551984349e-07,
"loss": 0.6375,
"step": 2056
},
{
"epoch": 0.8276000804667069,
"grad_norm": 5.052807884367178,
"learning_rate": 4.421568634802165e-07,
"loss": 0.687,
"step": 2057
},
{
"epoch": 0.828002414001207,
"grad_norm": 3.9353061661541133,
"learning_rate": 4.40165230396776e-07,
"loss": 0.4774,
"step": 2058
},
{
"epoch": 0.8284047475357071,
"grad_norm": 4.01193456780117,
"learning_rate": 4.3817765987616474e-07,
"loss": 0.4932,
"step": 2059
},
{
"epoch": 0.8288070810702072,
"grad_norm": 4.976729164334829,
"learning_rate": 4.361941558384189e-07,
"loss": 0.6807,
"step": 2060
},
{
"epoch": 0.8292094146047073,
"grad_norm": 4.339682582742787,
"learning_rate": 4.3421472219555725e-07,
"loss": 0.4727,
"step": 2061
},
{
"epoch": 0.8296117481392074,
"grad_norm": 4.503522087170257,
"learning_rate": 4.3223936285156836e-07,
"loss": 0.6022,
"step": 2062
},
{
"epoch": 0.8300140816737075,
"grad_norm": 4.009166626990359,
"learning_rate": 4.3026808170240686e-07,
"loss": 0.6318,
"step": 2063
},
{
"epoch": 0.8304164152082076,
"grad_norm": 4.433472277211533,
"learning_rate": 4.283008826359833e-07,
"loss": 0.6059,
"step": 2064
},
{
"epoch": 0.8308187487427077,
"grad_norm": 4.239654566860423,
"learning_rate": 4.263377695321577e-07,
"loss": 0.5319,
"step": 2065
},
{
"epoch": 0.8312210822772078,
"grad_norm": 4.33872236149242,
"learning_rate": 4.2437874626273024e-07,
"loss": 0.5192,
"step": 2066
},
{
"epoch": 0.831623415811708,
"grad_norm": 4.396517305178651,
"learning_rate": 4.2242381669143603e-07,
"loss": 0.6266,
"step": 2067
},
{
"epoch": 0.832025749346208,
"grad_norm": 4.213477790744039,
"learning_rate": 4.204729846739358e-07,
"loss": 0.6738,
"step": 2068
},
{
"epoch": 0.8324280828807081,
"grad_norm": 3.312219624247552,
"learning_rate": 4.1852625405780914e-07,
"loss": 0.4833,
"step": 2069
},
{
"epoch": 0.8328304164152082,
"grad_norm": 3.9392383731854257,
"learning_rate": 4.165836286825464e-07,
"loss": 0.4513,
"step": 2070
},
{
"epoch": 0.8332327499497083,
"grad_norm": 3.8224909723307734,
"learning_rate": 4.1464511237954056e-07,
"loss": 0.5628,
"step": 2071
},
{
"epoch": 0.8336350834842085,
"grad_norm": 4.5465506777870885,
"learning_rate": 4.1271070897208166e-07,
"loss": 0.5761,
"step": 2072
},
{
"epoch": 0.8340374170187085,
"grad_norm": 3.87942801566727,
"learning_rate": 4.107804222753464e-07,
"loss": 0.4364,
"step": 2073
},
{
"epoch": 0.8344397505532086,
"grad_norm": 3.901824805945848,
"learning_rate": 4.088542560963937e-07,
"loss": 0.4897,
"step": 2074
},
{
"epoch": 0.8348420840877087,
"grad_norm": 4.2369387169530155,
"learning_rate": 4.0693221423415447e-07,
"loss": 0.6186,
"step": 2075
},
{
"epoch": 0.8352444176222088,
"grad_norm": 4.1329987533853005,
"learning_rate": 4.0501430047942603e-07,
"loss": 0.5358,
"step": 2076
},
{
"epoch": 0.835646751156709,
"grad_norm": 4.533598883351018,
"learning_rate": 4.03100518614864e-07,
"loss": 0.5522,
"step": 2077
},
{
"epoch": 0.836049084691209,
"grad_norm": 3.998009974659786,
"learning_rate": 4.011908724149746e-07,
"loss": 0.3743,
"step": 2078
},
{
"epoch": 0.8364514182257091,
"grad_norm": 3.417865078970055,
"learning_rate": 3.9928536564610767e-07,
"loss": 0.4035,
"step": 2079
},
{
"epoch": 0.8368537517602093,
"grad_norm": 3.9749987307006345,
"learning_rate": 3.9738400206644766e-07,
"loss": 0.6477,
"step": 2080
},
{
"epoch": 0.8372560852947093,
"grad_norm": 4.037697255647064,
"learning_rate": 3.9548678542600886e-07,
"loss": 0.5661,
"step": 2081
},
{
"epoch": 0.8376584188292094,
"grad_norm": 4.431095797135675,
"learning_rate": 3.935937194666267e-07,
"loss": 0.4818,
"step": 2082
},
{
"epoch": 0.8380607523637095,
"grad_norm": 4.260045700942229,
"learning_rate": 3.9170480792194985e-07,
"loss": 0.4263,
"step": 2083
},
{
"epoch": 0.8384630858982096,
"grad_norm": 4.051159338313962,
"learning_rate": 3.8982005451743243e-07,
"loss": 0.487,
"step": 2084
},
{
"epoch": 0.8388654194327098,
"grad_norm": 3.505632578812268,
"learning_rate": 3.8793946297032926e-07,
"loss": 0.5173,
"step": 2085
},
{
"epoch": 0.8392677529672098,
"grad_norm": 3.4621570841739917,
"learning_rate": 3.8606303698968643e-07,
"loss": 0.3609,
"step": 2086
},
{
"epoch": 0.8396700865017099,
"grad_norm": 3.708654324006204,
"learning_rate": 3.841907802763331e-07,
"loss": 0.4437,
"step": 2087
},
{
"epoch": 0.84007242003621,
"grad_norm": 3.3369961693311447,
"learning_rate": 3.8232269652287713e-07,
"loss": 0.4295,
"step": 2088
},
{
"epoch": 0.8404747535707101,
"grad_norm": 3.8475631558462884,
"learning_rate": 3.8045878941369434e-07,
"loss": 0.5628,
"step": 2089
},
{
"epoch": 0.8408770871052103,
"grad_norm": 3.827602289260064,
"learning_rate": 3.7859906262492503e-07,
"loss": 0.5158,
"step": 2090
},
{
"epoch": 0.8412794206397103,
"grad_norm": 4.093693428637088,
"learning_rate": 3.7674351982446347e-07,
"loss": 0.4742,
"step": 2091
},
{
"epoch": 0.8416817541742104,
"grad_norm": 4.310046930338328,
"learning_rate": 3.7489216467195304e-07,
"loss": 0.6097,
"step": 2092
},
{
"epoch": 0.8420840877087106,
"grad_norm": 4.017527345482499,
"learning_rate": 3.730450008187761e-07,
"loss": 0.5075,
"step": 2093
},
{
"epoch": 0.8424864212432106,
"grad_norm": 3.946920016020563,
"learning_rate": 3.7120203190805034e-07,
"loss": 0.4275,
"step": 2094
},
{
"epoch": 0.8428887547777107,
"grad_norm": 4.145255268119006,
"learning_rate": 3.6936326157461925e-07,
"loss": 0.5914,
"step": 2095
},
{
"epoch": 0.8432910883122108,
"grad_norm": 4.2460232246744525,
"learning_rate": 3.6752869344504586e-07,
"loss": 0.5152,
"step": 2096
},
{
"epoch": 0.8436934218467109,
"grad_norm": 3.95033426326261,
"learning_rate": 3.656983311376042e-07,
"loss": 0.4448,
"step": 2097
},
{
"epoch": 0.8440957553812111,
"grad_norm": 3.8931938435117286,
"learning_rate": 3.638721782622745e-07,
"loss": 0.499,
"step": 2098
},
{
"epoch": 0.8444980889157111,
"grad_norm": 4.3118732957405985,
"learning_rate": 3.620502384207342e-07,
"loss": 0.49,
"step": 2099
},
{
"epoch": 0.8449004224502112,
"grad_norm": 4.005112629756995,
"learning_rate": 3.602325152063524e-07,
"loss": 0.4125,
"step": 2100
},
{
"epoch": 0.8453027559847113,
"grad_norm": 3.9812849673226114,
"learning_rate": 3.5841901220418e-07,
"loss": 0.5451,
"step": 2101
},
{
"epoch": 0.8457050895192114,
"grad_norm": 3.948178072792975,
"learning_rate": 3.566097329909468e-07,
"loss": 0.6427,
"step": 2102
},
{
"epoch": 0.8461074230537116,
"grad_norm": 3.7188722401214016,
"learning_rate": 3.5480468113504946e-07,
"loss": 0.405,
"step": 2103
},
{
"epoch": 0.8465097565882116,
"grad_norm": 3.913596983806647,
"learning_rate": 3.530038601965491e-07,
"loss": 0.4635,
"step": 2104
},
{
"epoch": 0.8469120901227117,
"grad_norm": 4.411578338116122,
"learning_rate": 3.5120727372716295e-07,
"loss": 0.5511,
"step": 2105
},
{
"epoch": 0.8473144236572119,
"grad_norm": 3.282083239679109,
"learning_rate": 3.494149252702536e-07,
"loss": 0.4441,
"step": 2106
},
{
"epoch": 0.8477167571917119,
"grad_norm": 4.011319326831284,
"learning_rate": 3.476268183608281e-07,
"loss": 0.5018,
"step": 2107
},
{
"epoch": 0.848119090726212,
"grad_norm": 4.284870919327636,
"learning_rate": 3.4584295652552667e-07,
"loss": 0.4846,
"step": 2108
},
{
"epoch": 0.8485214242607121,
"grad_norm": 4.276416896965254,
"learning_rate": 3.4406334328261765e-07,
"loss": 0.5376,
"step": 2109
},
{
"epoch": 0.8489237577952122,
"grad_norm": 3.834125175179826,
"learning_rate": 3.4228798214198974e-07,
"loss": 0.5651,
"step": 2110
},
{
"epoch": 0.8493260913297124,
"grad_norm": 3.9930841906813517,
"learning_rate": 3.405168766051445e-07,
"loss": 0.3851,
"step": 2111
},
{
"epoch": 0.8497284248642124,
"grad_norm": 4.2112351410317626,
"learning_rate": 3.3875003016519137e-07,
"loss": 0.5431,
"step": 2112
},
{
"epoch": 0.8501307583987125,
"grad_norm": 4.52788654812861,
"learning_rate": 3.369874463068398e-07,
"loss": 0.5849,
"step": 2113
},
{
"epoch": 0.8505330919332126,
"grad_norm": 3.8509825855331163,
"learning_rate": 3.3522912850639106e-07,
"loss": 0.5748,
"step": 2114
},
{
"epoch": 0.8509354254677127,
"grad_norm": 3.510509250668585,
"learning_rate": 3.334750802317338e-07,
"loss": 0.4397,
"step": 2115
},
{
"epoch": 0.8513377590022129,
"grad_norm": 3.8349389092945114,
"learning_rate": 3.3172530494233467e-07,
"loss": 0.4769,
"step": 2116
},
{
"epoch": 0.8517400925367129,
"grad_norm": 3.8288751826393197,
"learning_rate": 3.2997980608923426e-07,
"loss": 0.5434,
"step": 2117
},
{
"epoch": 0.852142426071213,
"grad_norm": 3.9001403159636374,
"learning_rate": 3.282385871150387e-07,
"loss": 0.4523,
"step": 2118
},
{
"epoch": 0.8525447596057132,
"grad_norm": 3.6933669568597978,
"learning_rate": 3.26501651453911e-07,
"loss": 0.3905,
"step": 2119
},
{
"epoch": 0.8529470931402132,
"grad_norm": 3.811615026103731,
"learning_rate": 3.247690025315689e-07,
"loss": 0.4305,
"step": 2120
},
{
"epoch": 0.8533494266747134,
"grad_norm": 3.6210800698038925,
"learning_rate": 3.2304064376527417e-07,
"loss": 0.5332,
"step": 2121
},
{
"epoch": 0.8537517602092134,
"grad_norm": 4.231885438858511,
"learning_rate": 3.2131657856382753e-07,
"loss": 0.5911,
"step": 2122
},
{
"epoch": 0.8541540937437135,
"grad_norm": 4.464935656077254,
"learning_rate": 3.195968103275618e-07,
"loss": 0.5702,
"step": 2123
},
{
"epoch": 0.8545564272782137,
"grad_norm": 4.3078684853827465,
"learning_rate": 3.17881342448334e-07,
"loss": 0.5691,
"step": 2124
},
{
"epoch": 0.8549587608127137,
"grad_norm": 3.8241264163503255,
"learning_rate": 3.1617017830952084e-07,
"loss": 0.5375,
"step": 2125
},
{
"epoch": 0.8553610943472139,
"grad_norm": 3.519525197941784,
"learning_rate": 3.1446332128601015e-07,
"loss": 0.4416,
"step": 2126
},
{
"epoch": 0.8557634278817139,
"grad_norm": 3.4697912989841524,
"learning_rate": 3.12760774744196e-07,
"loss": 0.4459,
"step": 2127
},
{
"epoch": 0.856165761416214,
"grad_norm": 3.7851796176468455,
"learning_rate": 3.1106254204196883e-07,
"loss": 0.4682,
"step": 2128
},
{
"epoch": 0.8565680949507142,
"grad_norm": 3.936606771811409,
"learning_rate": 3.093686265287138e-07,
"loss": 0.5729,
"step": 2129
},
{
"epoch": 0.8569704284852142,
"grad_norm": 4.190455136191216,
"learning_rate": 3.076790315452988e-07,
"loss": 0.5277,
"step": 2130
},
{
"epoch": 0.8573727620197144,
"grad_norm": 4.108892268475147,
"learning_rate": 3.05993760424072e-07,
"loss": 0.4329,
"step": 2131
},
{
"epoch": 0.8577750955542145,
"grad_norm": 4.390564938079403,
"learning_rate": 3.043128164888537e-07,
"loss": 0.5011,
"step": 2132
},
{
"epoch": 0.8581774290887145,
"grad_norm": 4.367475352759402,
"learning_rate": 3.0263620305492877e-07,
"loss": 0.4797,
"step": 2133
},
{
"epoch": 0.8585797626232147,
"grad_norm": 3.8444843612579924,
"learning_rate": 3.009639234290418e-07,
"loss": 0.5327,
"step": 2134
},
{
"epoch": 0.8589820961577147,
"grad_norm": 3.6090606638950318,
"learning_rate": 2.992959809093901e-07,
"loss": 0.4413,
"step": 2135
},
{
"epoch": 0.8593844296922148,
"grad_norm": 4.388230282282369,
"learning_rate": 2.976323787856175e-07,
"loss": 0.5717,
"step": 2136
},
{
"epoch": 0.859786763226715,
"grad_norm": 4.333473286328121,
"learning_rate": 2.9597312033880557e-07,
"loss": 0.6314,
"step": 2137
},
{
"epoch": 0.860189096761215,
"grad_norm": 3.890510795148687,
"learning_rate": 2.943182088414701e-07,
"loss": 0.433,
"step": 2138
},
{
"epoch": 0.8605914302957152,
"grad_norm": 4.234653691147445,
"learning_rate": 2.9266764755755406e-07,
"loss": 0.54,
"step": 2139
},
{
"epoch": 0.8609937638302152,
"grad_norm": 4.183758834087082,
"learning_rate": 2.910214397424202e-07,
"loss": 0.4845,
"step": 2140
},
{
"epoch": 0.8613960973647153,
"grad_norm": 3.799668423286151,
"learning_rate": 2.8937958864284353e-07,
"loss": 0.4333,
"step": 2141
},
{
"epoch": 0.8617984308992155,
"grad_norm": 4.133997388279442,
"learning_rate": 2.877420974970088e-07,
"loss": 0.4396,
"step": 2142
},
{
"epoch": 0.8622007644337155,
"grad_norm": 4.0525547445688614,
"learning_rate": 2.861089695345007e-07,
"loss": 0.4857,
"step": 2143
},
{
"epoch": 0.8626030979682157,
"grad_norm": 3.7880628956269904,
"learning_rate": 2.8448020797629735e-07,
"loss": 0.5046,
"step": 2144
},
{
"epoch": 0.8630054315027158,
"grad_norm": 4.143267859487349,
"learning_rate": 2.8285581603476747e-07,
"loss": 0.5291,
"step": 2145
},
{
"epoch": 0.8634077650372158,
"grad_norm": 3.8047765913147993,
"learning_rate": 2.8123579691365923e-07,
"loss": 0.5513,
"step": 2146
},
{
"epoch": 0.863810098571716,
"grad_norm": 4.041711864076377,
"learning_rate": 2.796201538080981e-07,
"loss": 0.448,
"step": 2147
},
{
"epoch": 0.864212432106216,
"grad_norm": 4.193683351365294,
"learning_rate": 2.7800888990457855e-07,
"loss": 0.5598,
"step": 2148
},
{
"epoch": 0.8646147656407162,
"grad_norm": 3.6613935569874583,
"learning_rate": 2.7640200838095793e-07,
"loss": 0.4067,
"step": 2149
},
{
"epoch": 0.8650170991752163,
"grad_norm": 4.104620267298875,
"learning_rate": 2.7479951240644953e-07,
"loss": 0.5648,
"step": 2150
},
{
"epoch": 0.8654194327097163,
"grad_norm": 3.7600215255952256,
"learning_rate": 2.7320140514161846e-07,
"loss": 0.3946,
"step": 2151
},
{
"epoch": 0.8658217662442165,
"grad_norm": 3.9525358112645965,
"learning_rate": 2.7160768973837293e-07,
"loss": 0.53,
"step": 2152
},
{
"epoch": 0.8662240997787165,
"grad_norm": 3.5450533427675226,
"learning_rate": 2.7001836933995995e-07,
"loss": 0.5298,
"step": 2153
},
{
"epoch": 0.8666264333132166,
"grad_norm": 3.6668666878628815,
"learning_rate": 2.684334470809588e-07,
"loss": 0.5056,
"step": 2154
},
{
"epoch": 0.8670287668477168,
"grad_norm": 4.224237505147509,
"learning_rate": 2.6685292608727236e-07,
"loss": 0.5218,
"step": 2155
},
{
"epoch": 0.8674311003822168,
"grad_norm": 3.69557439406065,
"learning_rate": 2.652768094761257e-07,
"loss": 0.3756,
"step": 2156
},
{
"epoch": 0.867833433916717,
"grad_norm": 4.281090167368473,
"learning_rate": 2.637051003560548e-07,
"loss": 0.6402,
"step": 2157
},
{
"epoch": 0.8682357674512171,
"grad_norm": 3.9633212522861054,
"learning_rate": 2.621378018269047e-07,
"loss": 0.5075,
"step": 2158
},
{
"epoch": 0.8686381009857171,
"grad_norm": 3.8583945548771825,
"learning_rate": 2.6057491697982e-07,
"loss": 0.6033,
"step": 2159
},
{
"epoch": 0.8690404345202173,
"grad_norm": 3.3558856738663487,
"learning_rate": 2.5901644889724186e-07,
"loss": 0.4015,
"step": 2160
},
{
"epoch": 0.8694427680547173,
"grad_norm": 3.5749358804434483,
"learning_rate": 2.574624006528992e-07,
"loss": 0.4944,
"step": 2161
},
{
"epoch": 0.8698451015892175,
"grad_norm": 3.86587826976141,
"learning_rate": 2.55912775311804e-07,
"loss": 0.4891,
"step": 2162
},
{
"epoch": 0.8702474351237176,
"grad_norm": 3.9917581972004785,
"learning_rate": 2.543675759302464e-07,
"loss": 0.5079,
"step": 2163
},
{
"epoch": 0.8706497686582176,
"grad_norm": 4.393976999836958,
"learning_rate": 2.5282680555578477e-07,
"loss": 0.6145,
"step": 2164
},
{
"epoch": 0.8710521021927178,
"grad_norm": 4.04228191167184,
"learning_rate": 2.51290467227244e-07,
"loss": 0.5429,
"step": 2165
},
{
"epoch": 0.8714544357272178,
"grad_norm": 4.4282484060475324,
"learning_rate": 2.497585639747077e-07,
"loss": 0.6325,
"step": 2166
},
{
"epoch": 0.871856769261718,
"grad_norm": 3.816275160612967,
"learning_rate": 2.482310988195125e-07,
"loss": 0.5543,
"step": 2167
},
{
"epoch": 0.8722591027962181,
"grad_norm": 4.1145602142866675,
"learning_rate": 2.4670807477424007e-07,
"loss": 0.5527,
"step": 2168
},
{
"epoch": 0.8726614363307181,
"grad_norm": 4.231321602541556,
"learning_rate": 2.4518949484271525e-07,
"loss": 0.6496,
"step": 2169
},
{
"epoch": 0.8730637698652183,
"grad_norm": 4.220523454555819,
"learning_rate": 2.4367536201999695e-07,
"loss": 0.5652,
"step": 2170
},
{
"epoch": 0.8734661033997184,
"grad_norm": 3.938889256449136,
"learning_rate": 2.421656792923724e-07,
"loss": 0.456,
"step": 2171
},
{
"epoch": 0.8738684369342185,
"grad_norm": 4.5217908297671485,
"learning_rate": 2.406604496373535e-07,
"loss": 0.6334,
"step": 2172
},
{
"epoch": 0.8742707704687186,
"grad_norm": 3.7526929266944773,
"learning_rate": 2.391596760236681e-07,
"loss": 0.4743,
"step": 2173
},
{
"epoch": 0.8746731040032186,
"grad_norm": 3.710705344410006,
"learning_rate": 2.376633614112561e-07,
"loss": 0.5661,
"step": 2174
},
{
"epoch": 0.8750754375377188,
"grad_norm": 3.2609967694053243,
"learning_rate": 2.361715087512631e-07,
"loss": 0.3589,
"step": 2175
},
{
"epoch": 0.8754777710722189,
"grad_norm": 4.108101136746485,
"learning_rate": 2.346841209860351e-07,
"loss": 0.4368,
"step": 2176
},
{
"epoch": 0.875880104606719,
"grad_norm": 3.763267262463342,
"learning_rate": 2.3320120104911053e-07,
"loss": 0.4953,
"step": 2177
},
{
"epoch": 0.8762824381412191,
"grad_norm": 5.025294007587999,
"learning_rate": 2.317227518652168e-07,
"loss": 0.6524,
"step": 2178
},
{
"epoch": 0.8766847716757191,
"grad_norm": 3.9367012271057584,
"learning_rate": 2.302487763502645e-07,
"loss": 0.4149,
"step": 2179
},
{
"epoch": 0.8770871052102193,
"grad_norm": 4.236893564539156,
"learning_rate": 2.2877927741134036e-07,
"loss": 0.5532,
"step": 2180
},
{
"epoch": 0.8774894387447194,
"grad_norm": 3.5424218920413373,
"learning_rate": 2.2731425794670135e-07,
"loss": 0.4691,
"step": 2181
},
{
"epoch": 0.8778917722792194,
"grad_norm": 4.04407474745495,
"learning_rate": 2.2585372084577033e-07,
"loss": 0.5125,
"step": 2182
},
{
"epoch": 0.8782941058137196,
"grad_norm": 3.8612028242647147,
"learning_rate": 2.2439766898913e-07,
"loss": 0.5256,
"step": 2183
},
{
"epoch": 0.8786964393482197,
"grad_norm": 4.271299959657279,
"learning_rate": 2.2294610524851685e-07,
"loss": 0.4944,
"step": 2184
},
{
"epoch": 0.8790987728827198,
"grad_norm": 4.465135814281888,
"learning_rate": 2.2149903248681464e-07,
"loss": 0.6834,
"step": 2185
},
{
"epoch": 0.8795011064172199,
"grad_norm": 3.9117028793633173,
"learning_rate": 2.2005645355805127e-07,
"loss": 0.4605,
"step": 2186
},
{
"epoch": 0.8799034399517199,
"grad_norm": 3.8402220541360292,
"learning_rate": 2.1861837130738945e-07,
"loss": 0.4467,
"step": 2187
},
{
"epoch": 0.8803057734862201,
"grad_norm": 4.223182944719613,
"learning_rate": 2.1718478857112506e-07,
"loss": 0.5347,
"step": 2188
},
{
"epoch": 0.8807081070207202,
"grad_norm": 3.8811613151194213,
"learning_rate": 2.1575570817667952e-07,
"loss": 0.5265,
"step": 2189
},
{
"epoch": 0.8811104405552203,
"grad_norm": 3.8705546986459956,
"learning_rate": 2.143311329425926e-07,
"loss": 0.4427,
"step": 2190
},
{
"epoch": 0.8815127740897204,
"grad_norm": 4.5414937373063955,
"learning_rate": 2.1291106567852105e-07,
"loss": 0.6442,
"step": 2191
},
{
"epoch": 0.8819151076242204,
"grad_norm": 3.588299387450043,
"learning_rate": 2.1149550918522887e-07,
"loss": 0.4173,
"step": 2192
},
{
"epoch": 0.8823174411587206,
"grad_norm": 4.165225372551161,
"learning_rate": 2.100844662545848e-07,
"loss": 0.5496,
"step": 2193
},
{
"epoch": 0.8827197746932207,
"grad_norm": 4.217297273907552,
"learning_rate": 2.086779396695554e-07,
"loss": 0.5329,
"step": 2194
},
{
"epoch": 0.8831221082277207,
"grad_norm": 3.980212414160413,
"learning_rate": 2.0727593220419812e-07,
"loss": 0.4682,
"step": 2195
},
{
"epoch": 0.8835244417622209,
"grad_norm": 4.015069610064507,
"learning_rate": 2.0587844662365963e-07,
"loss": 0.4975,
"step": 2196
},
{
"epoch": 0.883926775296721,
"grad_norm": 3.675033120744641,
"learning_rate": 2.0448548568416715e-07,
"loss": 0.4932,
"step": 2197
},
{
"epoch": 0.8843291088312211,
"grad_norm": 3.7917456823069178,
"learning_rate": 2.030970521330247e-07,
"loss": 0.4523,
"step": 2198
},
{
"epoch": 0.8847314423657212,
"grad_norm": 4.2084887531102195,
"learning_rate": 2.0171314870860548e-07,
"loss": 0.5797,
"step": 2199
},
{
"epoch": 0.8851337759002212,
"grad_norm": 4.1920678247066245,
"learning_rate": 2.0033377814034999e-07,
"loss": 0.5458,
"step": 2200
},
{
"epoch": 0.8855361094347214,
"grad_norm": 4.341033030777012,
"learning_rate": 1.9895894314875707e-07,
"loss": 0.5069,
"step": 2201
},
{
"epoch": 0.8859384429692215,
"grad_norm": 4.251489033634488,
"learning_rate": 1.9758864644538188e-07,
"loss": 0.5658,
"step": 2202
},
{
"epoch": 0.8863407765037216,
"grad_norm": 4.118913583140747,
"learning_rate": 1.9622289073282675e-07,
"loss": 0.5186,
"step": 2203
},
{
"epoch": 0.8867431100382217,
"grad_norm": 4.17198274381597,
"learning_rate": 1.9486167870473915e-07,
"loss": 0.5394,
"step": 2204
},
{
"epoch": 0.8871454435727217,
"grad_norm": 4.1140466186765865,
"learning_rate": 1.9350501304580577e-07,
"loss": 0.6302,
"step": 2205
},
{
"epoch": 0.8875477771072219,
"grad_norm": 4.252952260852059,
"learning_rate": 1.9215289643174485e-07,
"loss": 0.5715,
"step": 2206
},
{
"epoch": 0.887950110641722,
"grad_norm": 3.9053915680472673,
"learning_rate": 1.9080533152930486e-07,
"loss": 0.502,
"step": 2207
},
{
"epoch": 0.8883524441762221,
"grad_norm": 3.702225135993101,
"learning_rate": 1.8946232099625482e-07,
"loss": 0.3917,
"step": 2208
},
{
"epoch": 0.8887547777107222,
"grad_norm": 3.5875962058050863,
"learning_rate": 1.8812386748138234e-07,
"loss": 0.3413,
"step": 2209
},
{
"epoch": 0.8891571112452223,
"grad_norm": 3.6894845944247723,
"learning_rate": 1.8678997362448743e-07,
"loss": 0.4334,
"step": 2210
},
{
"epoch": 0.8895594447797224,
"grad_norm": 3.977873097022255,
"learning_rate": 1.8546064205637758e-07,
"loss": 0.4746,
"step": 2211
},
{
"epoch": 0.8899617783142225,
"grad_norm": 3.4865669801340147,
"learning_rate": 1.8413587539886064e-07,
"loss": 0.4173,
"step": 2212
},
{
"epoch": 0.8903641118487226,
"grad_norm": 4.35862373407535,
"learning_rate": 1.828156762647429e-07,
"loss": 0.5653,
"step": 2213
},
{
"epoch": 0.8907664453832227,
"grad_norm": 4.2652495550914775,
"learning_rate": 1.81500047257821e-07,
"loss": 0.5071,
"step": 2214
},
{
"epoch": 0.8911687789177228,
"grad_norm": 3.285190042692873,
"learning_rate": 1.8018899097287906e-07,
"loss": 0.3731,
"step": 2215
},
{
"epoch": 0.8915711124522229,
"grad_norm": 3.782205241606253,
"learning_rate": 1.78882509995682e-07,
"loss": 0.4907,
"step": 2216
},
{
"epoch": 0.891973445986723,
"grad_norm": 4.391188417961942,
"learning_rate": 1.7758060690297035e-07,
"loss": 0.3594,
"step": 2217
},
{
"epoch": 0.892375779521223,
"grad_norm": 3.172977998916911,
"learning_rate": 1.7628328426245677e-07,
"loss": 0.3866,
"step": 2218
},
{
"epoch": 0.8927781130557232,
"grad_norm": 4.109857114119739,
"learning_rate": 1.7499054463281985e-07,
"loss": 0.5489,
"step": 2219
},
{
"epoch": 0.8931804465902233,
"grad_norm": 3.6387926247562214,
"learning_rate": 1.7370239056369946e-07,
"loss": 0.3892,
"step": 2220
},
{
"epoch": 0.8935827801247234,
"grad_norm": 3.897940749417438,
"learning_rate": 1.7241882459569008e-07,
"loss": 0.6287,
"step": 2221
},
{
"epoch": 0.8939851136592235,
"grad_norm": 3.968476012728493,
"learning_rate": 1.7113984926033832e-07,
"loss": 0.459,
"step": 2222
},
{
"epoch": 0.8943874471937237,
"grad_norm": 4.209938898766111,
"learning_rate": 1.6986546708013722e-07,
"loss": 0.5426,
"step": 2223
},
{
"epoch": 0.8947897807282237,
"grad_norm": 3.791580194250541,
"learning_rate": 1.6859568056852016e-07,
"loss": 0.498,
"step": 2224
},
{
"epoch": 0.8951921142627238,
"grad_norm": 3.8639160637014665,
"learning_rate": 1.673304922298563e-07,
"loss": 0.4049,
"step": 2225
},
{
"epoch": 0.8955944477972239,
"grad_norm": 4.147228062957371,
"learning_rate": 1.6606990455944634e-07,
"loss": 0.5258,
"step": 2226
},
{
"epoch": 0.895996781331724,
"grad_norm": 4.247308433616599,
"learning_rate": 1.648139200435178e-07,
"loss": 0.5252,
"step": 2227
},
{
"epoch": 0.8963991148662241,
"grad_norm": 4.057558052673904,
"learning_rate": 1.635625411592179e-07,
"loss": 0.4851,
"step": 2228
},
{
"epoch": 0.8968014484007242,
"grad_norm": 4.529371570037193,
"learning_rate": 1.6231577037461228e-07,
"loss": 0.6091,
"step": 2229
},
{
"epoch": 0.8972037819352243,
"grad_norm": 4.092889891063284,
"learning_rate": 1.6107361014867622e-07,
"loss": 0.5051,
"step": 2230
},
{
"epoch": 0.8976061154697244,
"grad_norm": 4.057510211081129,
"learning_rate": 1.598360629312931e-07,
"loss": 0.5379,
"step": 2231
},
{
"epoch": 0.8980084490042245,
"grad_norm": 4.08275289824886,
"learning_rate": 1.586031311632477e-07,
"loss": 0.5542,
"step": 2232
},
{
"epoch": 0.8984107825387246,
"grad_norm": 5.7116466876850875,
"learning_rate": 1.5737481727622212e-07,
"loss": 0.5021,
"step": 2233
},
{
"epoch": 0.8988131160732247,
"grad_norm": 3.507507523132951,
"learning_rate": 1.5615112369278957e-07,
"loss": 0.391,
"step": 2234
},
{
"epoch": 0.8992154496077248,
"grad_norm": 3.7113726990024194,
"learning_rate": 1.5493205282641228e-07,
"loss": 0.4515,
"step": 2235
},
{
"epoch": 0.899617783142225,
"grad_norm": 3.512898207115907,
"learning_rate": 1.5371760708143447e-07,
"loss": 0.4935,
"step": 2236
},
{
"epoch": 0.900020116676725,
"grad_norm": 4.213491522925021,
"learning_rate": 1.525077888530782e-07,
"loss": 0.538,
"step": 2237
},
{
"epoch": 0.9004224502112251,
"grad_norm": 4.482285849278304,
"learning_rate": 1.5130260052743955e-07,
"loss": 0.6545,
"step": 2238
},
{
"epoch": 0.9008247837457252,
"grad_norm": 3.7492718480617384,
"learning_rate": 1.5010204448148214e-07,
"loss": 0.4442,
"step": 2239
},
{
"epoch": 0.9012271172802253,
"grad_norm": 3.9838020107249195,
"learning_rate": 1.489061230830338e-07,
"loss": 0.4953,
"step": 2240
},
{
"epoch": 0.9016294508147255,
"grad_norm": 3.9003308620272796,
"learning_rate": 1.4771483869078228e-07,
"loss": 0.5208,
"step": 2241
},
{
"epoch": 0.9020317843492255,
"grad_norm": 3.788631062888711,
"learning_rate": 1.4652819365426867e-07,
"loss": 0.5613,
"step": 2242
},
{
"epoch": 0.9024341178837256,
"grad_norm": 4.676276663949487,
"learning_rate": 1.453461903138853e-07,
"loss": 0.5352,
"step": 2243
},
{
"epoch": 0.9028364514182257,
"grad_norm": 4.525346271987768,
"learning_rate": 1.441688310008682e-07,
"loss": 0.5099,
"step": 2244
},
{
"epoch": 0.9032387849527258,
"grad_norm": 3.625199458865597,
"learning_rate": 1.4299611803729547e-07,
"loss": 0.4766,
"step": 2245
},
{
"epoch": 0.903641118487226,
"grad_norm": 3.8753089109297787,
"learning_rate": 1.4182805373608055e-07,
"loss": 0.5326,
"step": 2246
},
{
"epoch": 0.904043452021726,
"grad_norm": 4.021228834417677,
"learning_rate": 1.4066464040096956e-07,
"loss": 0.5691,
"step": 2247
},
{
"epoch": 0.9044457855562261,
"grad_norm": 4.6752863808538105,
"learning_rate": 1.395058803265334e-07,
"loss": 0.7598,
"step": 2248
},
{
"epoch": 0.9048481190907263,
"grad_norm": 3.7179471187703514,
"learning_rate": 1.3835177579816783e-07,
"loss": 0.3955,
"step": 2249
},
{
"epoch": 0.9052504526252263,
"grad_norm": 5.00701237752423,
"learning_rate": 1.3720232909208541e-07,
"loss": 0.6319,
"step": 2250
},
{
"epoch": 0.9056527861597264,
"grad_norm": 3.7834417036951895,
"learning_rate": 1.3605754247531245e-07,
"loss": 0.5174,
"step": 2251
},
{
"epoch": 0.9060551196942265,
"grad_norm": 3.680557056193272,
"learning_rate": 1.3491741820568366e-07,
"loss": 0.4133,
"step": 2252
},
{
"epoch": 0.9064574532287266,
"grad_norm": 4.046295696144512,
"learning_rate": 1.337819585318395e-07,
"loss": 0.6443,
"step": 2253
},
{
"epoch": 0.9068597867632268,
"grad_norm": 4.363328152667618,
"learning_rate": 1.3265116569322e-07,
"loss": 0.5437,
"step": 2254
},
{
"epoch": 0.9072621202977268,
"grad_norm": 4.437499912355556,
"learning_rate": 1.3152504192006005e-07,
"loss": 0.5439,
"step": 2255
},
{
"epoch": 0.9076644538322269,
"grad_norm": 3.904533797833089,
"learning_rate": 1.3040358943338742e-07,
"loss": 0.4058,
"step": 2256
},
{
"epoch": 0.908066787366727,
"grad_norm": 3.990363526582056,
"learning_rate": 1.2928681044501513e-07,
"loss": 0.4204,
"step": 2257
},
{
"epoch": 0.9084691209012271,
"grad_norm": 4.0760801231394606,
"learning_rate": 1.281747071575404e-07,
"loss": 0.6104,
"step": 2258
},
{
"epoch": 0.9088714544357273,
"grad_norm": 4.256580027345346,
"learning_rate": 1.270672817643376e-07,
"loss": 0.6544,
"step": 2259
},
{
"epoch": 0.9092737879702273,
"grad_norm": 4.71579332313011,
"learning_rate": 1.2596453644955597e-07,
"loss": 0.6273,
"step": 2260
},
{
"epoch": 0.9096761215047274,
"grad_norm": 3.483958223270458,
"learning_rate": 1.2486647338811248e-07,
"loss": 0.4546,
"step": 2261
},
{
"epoch": 0.9100784550392276,
"grad_norm": 3.9093619147070946,
"learning_rate": 1.2377309474569165e-07,
"loss": 0.4985,
"step": 2262
},
{
"epoch": 0.9104807885737276,
"grad_norm": 4.057320495855987,
"learning_rate": 1.226844026787377e-07,
"loss": 0.5448,
"step": 2263
},
{
"epoch": 0.9108831221082277,
"grad_norm": 3.5360903799800503,
"learning_rate": 1.2160039933445222e-07,
"loss": 0.4164,
"step": 2264
},
{
"epoch": 0.9112854556427278,
"grad_norm": 4.087420290794404,
"learning_rate": 1.2052108685078834e-07,
"loss": 0.594,
"step": 2265
},
{
"epoch": 0.9116877891772279,
"grad_norm": 3.9832710853164763,
"learning_rate": 1.1944646735644882e-07,
"loss": 0.452,
"step": 2266
},
{
"epoch": 0.9120901227117281,
"grad_norm": 4.282566514995135,
"learning_rate": 1.1837654297087997e-07,
"loss": 0.5473,
"step": 2267
},
{
"epoch": 0.9124924562462281,
"grad_norm": 3.9435273915235656,
"learning_rate": 1.1731131580426796e-07,
"loss": 0.6224,
"step": 2268
},
{
"epoch": 0.9128947897807282,
"grad_norm": 4.3103292605724155,
"learning_rate": 1.1625078795753476e-07,
"loss": 0.5467,
"step": 2269
},
{
"epoch": 0.9132971233152283,
"grad_norm": 4.22573396631058,
"learning_rate": 1.1519496152233439e-07,
"loss": 0.5037,
"step": 2270
},
{
"epoch": 0.9136994568497284,
"grad_norm": 3.7886525775152338,
"learning_rate": 1.1414383858104756e-07,
"loss": 0.4498,
"step": 2271
},
{
"epoch": 0.9141017903842286,
"grad_norm": 3.2224197154073617,
"learning_rate": 1.1309742120677897e-07,
"loss": 0.4794,
"step": 2272
},
{
"epoch": 0.9145041239187286,
"grad_norm": 4.705013349470272,
"learning_rate": 1.1205571146335303e-07,
"loss": 0.6041,
"step": 2273
},
{
"epoch": 0.9149064574532287,
"grad_norm": 4.6403427413148535,
"learning_rate": 1.1101871140530851e-07,
"loss": 0.6308,
"step": 2274
},
{
"epoch": 0.9153087909877289,
"grad_norm": 3.819381857552772,
"learning_rate": 1.0998642307789576e-07,
"loss": 0.5461,
"step": 2275
},
{
"epoch": 0.9157111245222289,
"grad_norm": 3.997200472540085,
"learning_rate": 1.0895884851707255e-07,
"loss": 0.4926,
"step": 2276
},
{
"epoch": 0.9161134580567291,
"grad_norm": 4.162425768576248,
"learning_rate": 1.079359897494997e-07,
"loss": 0.489,
"step": 2277
},
{
"epoch": 0.9165157915912291,
"grad_norm": 3.8422971328140454,
"learning_rate": 1.0691784879253736e-07,
"loss": 0.5291,
"step": 2278
},
{
"epoch": 0.9169181251257292,
"grad_norm": 4.112928250644688,
"learning_rate": 1.0590442765423953e-07,
"loss": 0.4926,
"step": 2279
},
{
"epoch": 0.9173204586602294,
"grad_norm": 3.9378155981778664,
"learning_rate": 1.0489572833335349e-07,
"loss": 0.4722,
"step": 2280
},
{
"epoch": 0.9177227921947294,
"grad_norm": 4.631927118939465,
"learning_rate": 1.038917528193123e-07,
"loss": 0.5274,
"step": 2281
},
{
"epoch": 0.9181251257292296,
"grad_norm": 4.035377760065258,
"learning_rate": 1.0289250309223314e-07,
"loss": 0.5788,
"step": 2282
},
{
"epoch": 0.9185274592637296,
"grad_norm": 3.868825647605808,
"learning_rate": 1.0189798112291177e-07,
"loss": 0.5046,
"step": 2283
},
{
"epoch": 0.9189297927982297,
"grad_norm": 4.194215283322997,
"learning_rate": 1.0090818887282112e-07,
"loss": 0.6029,
"step": 2284
},
{
"epoch": 0.9193321263327299,
"grad_norm": 4.215786203028234,
"learning_rate": 9.99231282941035e-08,
"loss": 0.4987,
"step": 2285
},
{
"epoch": 0.9197344598672299,
"grad_norm": 4.911243421308879,
"learning_rate": 9.894280132957124e-08,
"loss": 0.6475,
"step": 2286
},
{
"epoch": 0.92013679340173,
"grad_norm": 4.551350447477857,
"learning_rate": 9.79672099126991e-08,
"loss": 0.6439,
"step": 2287
},
{
"epoch": 0.9205391269362302,
"grad_norm": 4.0194710398308615,
"learning_rate": 9.69963559676232e-08,
"loss": 0.5376,
"step": 2288
},
{
"epoch": 0.9209414604707302,
"grad_norm": 3.83021269408099,
"learning_rate": 9.60302414091352e-08,
"loss": 0.5093,
"step": 2289
},
{
"epoch": 0.9213437940052304,
"grad_norm": 4.1940419822947534,
"learning_rate": 9.506886814268007e-08,
"loss": 0.5702,
"step": 2290
},
{
"epoch": 0.9217461275397304,
"grad_norm": 4.341103613511488,
"learning_rate": 9.41122380643511e-08,
"loss": 0.5271,
"step": 2291
},
{
"epoch": 0.9221484610742305,
"grad_norm": 4.421018842084841,
"learning_rate": 9.316035306088684e-08,
"loss": 0.5923,
"step": 2292
},
{
"epoch": 0.9225507946087307,
"grad_norm": 4.192150302264493,
"learning_rate": 9.221321500966746e-08,
"loss": 0.5475,
"step": 2293
},
{
"epoch": 0.9229531281432307,
"grad_norm": 4.56023548088152,
"learning_rate": 9.127082577871038e-08,
"loss": 0.6055,
"step": 2294
},
{
"epoch": 0.9233554616777309,
"grad_norm": 3.727375694339219,
"learning_rate": 9.033318722666801e-08,
"loss": 0.3706,
"step": 2295
},
{
"epoch": 0.9237577952122309,
"grad_norm": 3.7651015398148897,
"learning_rate": 8.940030120282162e-08,
"loss": 0.5118,
"step": 2296
},
{
"epoch": 0.924160128746731,
"grad_norm": 4.305229814421829,
"learning_rate": 8.847216954708055e-08,
"loss": 0.5453,
"step": 2297
},
{
"epoch": 0.9245624622812312,
"grad_norm": 4.1747564223004066,
"learning_rate": 8.754879408997636e-08,
"loss": 0.5156,
"step": 2298
},
{
"epoch": 0.9249647958157312,
"grad_norm": 3.649399704340268,
"learning_rate": 8.663017665266032e-08,
"loss": 0.3701,
"step": 2299
},
{
"epoch": 0.9253671293502314,
"grad_norm": 3.7596577501273734,
"learning_rate": 8.571631904690014e-08,
"loss": 0.4399,
"step": 2300
},
{
"epoch": 0.9257694628847315,
"grad_norm": 4.230082435346815,
"learning_rate": 8.480722307507461e-08,
"loss": 0.4874,
"step": 2301
},
{
"epoch": 0.9261717964192315,
"grad_norm": 3.9612627542935197,
"learning_rate": 8.390289053017226e-08,
"loss": 0.5485,
"step": 2302
},
{
"epoch": 0.9265741299537317,
"grad_norm": 3.5438909670467003,
"learning_rate": 8.300332319578664e-08,
"loss": 0.3951,
"step": 2303
},
{
"epoch": 0.9269764634882317,
"grad_norm": 3.2833308636054217,
"learning_rate": 8.210852284611299e-08,
"loss": 0.407,
"step": 2304
},
{
"epoch": 0.9273787970227318,
"grad_norm": 3.4444311785022865,
"learning_rate": 8.121849124594405e-08,
"loss": 0.3925,
"step": 2305
},
{
"epoch": 0.927781130557232,
"grad_norm": 4.548198093339736,
"learning_rate": 8.033323015066818e-08,
"loss": 0.719,
"step": 2306
},
{
"epoch": 0.928183464091732,
"grad_norm": 3.6123982321182284,
"learning_rate": 7.945274130626484e-08,
"loss": 0.3828,
"step": 2307
},
{
"epoch": 0.9285857976262322,
"grad_norm": 3.177723392355348,
"learning_rate": 7.857702644930133e-08,
"loss": 0.353,
"step": 2308
},
{
"epoch": 0.9289881311607322,
"grad_norm": 4.133100828896777,
"learning_rate": 7.770608730692852e-08,
"loss": 0.4977,
"step": 2309
},
{
"epoch": 0.9293904646952323,
"grad_norm": 3.562342202197354,
"learning_rate": 7.683992559687881e-08,
"loss": 0.3892,
"step": 2310
},
{
"epoch": 0.9297927982297325,
"grad_norm": 3.898366854418511,
"learning_rate": 7.597854302746288e-08,
"loss": 0.4583,
"step": 2311
},
{
"epoch": 0.9301951317642325,
"grad_norm": 4.612414420760638,
"learning_rate": 7.512194129756428e-08,
"loss": 0.5581,
"step": 2312
},
{
"epoch": 0.9305974652987327,
"grad_norm": 3.988029769560431,
"learning_rate": 7.42701220966388e-08,
"loss": 0.4744,
"step": 2313
},
{
"epoch": 0.9309997988332327,
"grad_norm": 4.481476524243926,
"learning_rate": 7.342308710470814e-08,
"loss": 0.59,
"step": 2314
},
{
"epoch": 0.9314021323677328,
"grad_norm": 4.562506876878105,
"learning_rate": 7.258083799235982e-08,
"loss": 0.6742,
"step": 2315
},
{
"epoch": 0.931804465902233,
"grad_norm": 4.628598132831451,
"learning_rate": 7.174337642074147e-08,
"loss": 0.6008,
"step": 2316
},
{
"epoch": 0.932206799436733,
"grad_norm": 4.283379928862786,
"learning_rate": 7.091070404155854e-08,
"loss": 0.5053,
"step": 2317
},
{
"epoch": 0.9326091329712332,
"grad_norm": 3.9776126177559155,
"learning_rate": 7.008282249707065e-08,
"loss": 0.4839,
"step": 2318
},
{
"epoch": 0.9330114665057333,
"grad_norm": 4.069433607383545,
"learning_rate": 6.925973342008897e-08,
"loss": 0.453,
"step": 2319
},
{
"epoch": 0.9334138000402333,
"grad_norm": 4.037972743715134,
"learning_rate": 6.844143843397238e-08,
"loss": 0.5427,
"step": 2320
},
{
"epoch": 0.9338161335747335,
"grad_norm": 3.547048708842676,
"learning_rate": 6.76279391526241e-08,
"loss": 0.4561,
"step": 2321
},
{
"epoch": 0.9342184671092335,
"grad_norm": 3.7831575419487256,
"learning_rate": 6.681923718049016e-08,
"loss": 0.5184,
"step": 2322
},
{
"epoch": 0.9346208006437337,
"grad_norm": 3.7960605831677428,
"learning_rate": 6.601533411255328e-08,
"loss": 0.428,
"step": 2323
},
{
"epoch": 0.9350231341782338,
"grad_norm": 3.699019611927255,
"learning_rate": 6.52162315343327e-08,
"loss": 0.4711,
"step": 2324
},
{
"epoch": 0.9354254677127338,
"grad_norm": 3.7499929742301465,
"learning_rate": 6.44219310218791e-08,
"loss": 0.494,
"step": 2325
},
{
"epoch": 0.935827801247234,
"grad_norm": 3.982088440613016,
"learning_rate": 6.363243414177239e-08,
"loss": 0.5013,
"step": 2326
},
{
"epoch": 0.936230134781734,
"grad_norm": 3.5405567172339034,
"learning_rate": 6.284774245111868e-08,
"loss": 0.4151,
"step": 2327
},
{
"epoch": 0.9366324683162341,
"grad_norm": 3.9069783277866894,
"learning_rate": 6.206785749754613e-08,
"loss": 0.5257,
"step": 2328
},
{
"epoch": 0.9370348018507343,
"grad_norm": 4.560908800971963,
"learning_rate": 6.12927808192032e-08,
"loss": 0.5827,
"step": 2329
},
{
"epoch": 0.9374371353852343,
"grad_norm": 3.6067893534884967,
"learning_rate": 6.052251394475544e-08,
"loss": 0.4533,
"step": 2330
},
{
"epoch": 0.9378394689197345,
"grad_norm": 4.228506402107548,
"learning_rate": 5.975705839338208e-08,
"loss": 0.5367,
"step": 2331
},
{
"epoch": 0.9382418024542346,
"grad_norm": 4.377928096595712,
"learning_rate": 5.89964156747716e-08,
"loss": 0.487,
"step": 2332
},
{
"epoch": 0.9386441359887346,
"grad_norm": 4.001917632135577,
"learning_rate": 5.8240587289122576e-08,
"loss": 0.5991,
"step": 2333
},
{
"epoch": 0.9390464695232348,
"grad_norm": 3.89102054624204,
"learning_rate": 5.7489574727136734e-08,
"loss": 0.5289,
"step": 2334
},
{
"epoch": 0.9394488030577348,
"grad_norm": 3.9529925607216314,
"learning_rate": 5.67433794700184e-08,
"loss": 0.5516,
"step": 2335
},
{
"epoch": 0.939851136592235,
"grad_norm": 3.9965015035833114,
"learning_rate": 5.600200298947034e-08,
"loss": 0.5758,
"step": 2336
},
{
"epoch": 0.9402534701267351,
"grad_norm": 4.354462501990094,
"learning_rate": 5.52654467476918e-08,
"loss": 0.5793,
"step": 2337
},
{
"epoch": 0.9406558036612351,
"grad_norm": 4.199766571714081,
"learning_rate": 5.453371219737491e-08,
"loss": 0.5688,
"step": 2338
},
{
"epoch": 0.9410581371957353,
"grad_norm": 3.3551684650390454,
"learning_rate": 5.380680078170275e-08,
"loss": 0.3524,
"step": 2339
},
{
"epoch": 0.9414604707302353,
"grad_norm": 4.408645627917106,
"learning_rate": 5.3084713934344613e-08,
"loss": 0.5442,
"step": 2340
},
{
"epoch": 0.9418628042647355,
"grad_norm": 4.4356672674608895,
"learning_rate": 5.236745307945545e-08,
"loss": 0.5028,
"step": 2341
},
{
"epoch": 0.9422651377992356,
"grad_norm": 3.46236072019421,
"learning_rate": 5.165501963167174e-08,
"loss": 0.3921,
"step": 2342
},
{
"epoch": 0.9426674713337356,
"grad_norm": 4.4269078668736075,
"learning_rate": 5.094741499610839e-08,
"loss": 0.5983,
"step": 2343
},
{
"epoch": 0.9430698048682358,
"grad_norm": 4.503787561410546,
"learning_rate": 5.02446405683582e-08,
"loss": 0.585,
"step": 2344
},
{
"epoch": 0.9434721384027359,
"grad_norm": 3.658547311389223,
"learning_rate": 4.9546697734485224e-08,
"loss": 0.4762,
"step": 2345
},
{
"epoch": 0.943874471937236,
"grad_norm": 4.353601396348682,
"learning_rate": 4.885358787102584e-08,
"loss": 0.6102,
"step": 2346
},
{
"epoch": 0.9442768054717361,
"grad_norm": 4.143732790296102,
"learning_rate": 4.816531234498406e-08,
"loss": 0.6124,
"step": 2347
},
{
"epoch": 0.9446791390062361,
"grad_norm": 4.3975559021632575,
"learning_rate": 4.7481872513829575e-08,
"loss": 0.5633,
"step": 2348
},
{
"epoch": 0.9450814725407363,
"grad_norm": 4.066972538920918,
"learning_rate": 4.680326972549387e-08,
"loss": 0.5118,
"step": 2349
},
{
"epoch": 0.9454838060752364,
"grad_norm": 4.03878765773838,
"learning_rate": 4.6129505318369127e-08,
"loss": 0.4778,
"step": 2350
},
{
"epoch": 0.9458861396097364,
"grad_norm": 4.201321170733065,
"learning_rate": 4.546058062130487e-08,
"loss": 0.5128,
"step": 2351
},
{
"epoch": 0.9462884731442366,
"grad_norm": 3.8094183727880395,
"learning_rate": 4.4796496953605493e-08,
"loss": 0.5441,
"step": 2352
},
{
"epoch": 0.9466908066787366,
"grad_norm": 3.5299699606250834,
"learning_rate": 4.413725562502691e-08,
"loss": 0.4678,
"step": 2353
},
{
"epoch": 0.9470931402132368,
"grad_norm": 3.9087226991955717,
"learning_rate": 4.348285793577517e-08,
"loss": 0.5329,
"step": 2354
},
{
"epoch": 0.9474954737477369,
"grad_norm": 4.0121109992844755,
"learning_rate": 4.2833305176503136e-08,
"loss": 0.5837,
"step": 2355
},
{
"epoch": 0.9478978072822369,
"grad_norm": 4.544145628732853,
"learning_rate": 4.218859862830826e-08,
"loss": 0.3829,
"step": 2356
},
{
"epoch": 0.9483001408167371,
"grad_norm": 3.89076415048523,
"learning_rate": 4.154873956272981e-08,
"loss": 0.4691,
"step": 2357
},
{
"epoch": 0.9487024743512372,
"grad_norm": 4.1725175451999,
"learning_rate": 4.091372924174636e-08,
"loss": 0.4885,
"step": 2358
},
{
"epoch": 0.9491048078857373,
"grad_norm": 4.3450167781713285,
"learning_rate": 4.028356891777385e-08,
"loss": 0.5124,
"step": 2359
},
{
"epoch": 0.9495071414202374,
"grad_norm": 4.4322372412295605,
"learning_rate": 3.9658259833662284e-08,
"loss": 0.5234,
"step": 2360
},
{
"epoch": 0.9499094749547374,
"grad_norm": 3.707524532117256,
"learning_rate": 3.903780322269374e-08,
"loss": 0.4149,
"step": 2361
},
{
"epoch": 0.9503118084892376,
"grad_norm": 5.065679790929374,
"learning_rate": 3.842220030858018e-08,
"loss": 0.688,
"step": 2362
},
{
"epoch": 0.9507141420237377,
"grad_norm": 3.786912616101777,
"learning_rate": 3.781145230546013e-08,
"loss": 0.43,
"step": 2363
},
{
"epoch": 0.9511164755582378,
"grad_norm": 3.6227665097152224,
"learning_rate": 3.720556041789752e-08,
"loss": 0.4382,
"step": 2364
},
{
"epoch": 0.9515188090927379,
"grad_norm": 3.999125768684484,
"learning_rate": 3.660452584087815e-08,
"loss": 0.5716,
"step": 2365
},
{
"epoch": 0.9519211426272379,
"grad_norm": 3.409457081881056,
"learning_rate": 3.600834975980877e-08,
"loss": 0.5407,
"step": 2366
},
{
"epoch": 0.9523234761617381,
"grad_norm": 3.892658686650397,
"learning_rate": 3.5417033350512444e-08,
"loss": 0.4589,
"step": 2367
},
{
"epoch": 0.9527258096962382,
"grad_norm": 4.000437062384583,
"learning_rate": 3.483057777922905e-08,
"loss": 0.4875,
"step": 2368
},
{
"epoch": 0.9531281432307382,
"grad_norm": 4.046666707315621,
"learning_rate": 3.42489842026103e-08,
"loss": 0.5162,
"step": 2369
},
{
"epoch": 0.9535304767652384,
"grad_norm": 4.290796216817974,
"learning_rate": 3.3672253767719485e-08,
"loss": 0.6233,
"step": 2370
},
{
"epoch": 0.9539328102997385,
"grad_norm": 4.023014608954607,
"learning_rate": 3.310038761202839e-08,
"loss": 0.5062,
"step": 2371
},
{
"epoch": 0.9543351438342386,
"grad_norm": 4.43210170452369,
"learning_rate": 3.2533386863414784e-08,
"loss": 0.5594,
"step": 2372
},
{
"epoch": 0.9547374773687387,
"grad_norm": 3.8979482975499624,
"learning_rate": 3.1971252640160545e-08,
"loss": 0.5128,
"step": 2373
},
{
"epoch": 0.9551398109032387,
"grad_norm": 4.162386221402383,
"learning_rate": 3.141398605094992e-08,
"loss": 0.5848,
"step": 2374
},
{
"epoch": 0.9555421444377389,
"grad_norm": 4.022603602017269,
"learning_rate": 3.086158819486651e-08,
"loss": 0.5771,
"step": 2375
},
{
"epoch": 0.955944477972239,
"grad_norm": 3.8920410241916863,
"learning_rate": 3.031406016139077e-08,
"loss": 0.5498,
"step": 2376
},
{
"epoch": 0.9563468115067391,
"grad_norm": 4.084788242660989,
"learning_rate": 2.9771403030399713e-08,
"loss": 0.598,
"step": 2377
},
{
"epoch": 0.9567491450412392,
"grad_norm": 4.048190861072185,
"learning_rate": 2.923361787216278e-08,
"loss": 0.4405,
"step": 2378
},
{
"epoch": 0.9571514785757392,
"grad_norm": 4.502700834332798,
"learning_rate": 2.8700705747340963e-08,
"loss": 0.5169,
"step": 2379
},
{
"epoch": 0.9575538121102394,
"grad_norm": 4.054509620797073,
"learning_rate": 2.8172667706983792e-08,
"loss": 0.53,
"step": 2380
},
{
"epoch": 0.9579561456447395,
"grad_norm": 3.939484097293469,
"learning_rate": 2.7649504792528204e-08,
"loss": 0.5566,
"step": 2381
},
{
"epoch": 0.9583584791792396,
"grad_norm": 4.299563624727568,
"learning_rate": 2.7131218035796324e-08,
"loss": 0.5603,
"step": 2382
},
{
"epoch": 0.9587608127137397,
"grad_norm": 3.723520114625019,
"learning_rate": 2.6617808458991867e-08,
"loss": 0.4346,
"step": 2383
},
{
"epoch": 0.9591631462482398,
"grad_norm": 3.667012045051573,
"learning_rate": 2.6109277074700955e-08,
"loss": 0.3638,
"step": 2384
},
{
"epoch": 0.9595654797827399,
"grad_norm": 3.725825844080702,
"learning_rate": 2.5605624885887414e-08,
"loss": 0.4541,
"step": 2385
},
{
"epoch": 0.95996781331724,
"grad_norm": 3.7094680890029776,
"learning_rate": 2.5106852885892764e-08,
"loss": 0.4693,
"step": 2386
},
{
"epoch": 0.96037014685174,
"grad_norm": 3.7229547246513937,
"learning_rate": 2.461296205843261e-08,
"loss": 0.5144,
"step": 2387
},
{
"epoch": 0.9607724803862402,
"grad_norm": 4.159612295768075,
"learning_rate": 2.412395337759693e-08,
"loss": 0.522,
"step": 2388
},
{
"epoch": 0.9611748139207403,
"grad_norm": 3.8986140359372286,
"learning_rate": 2.3639827807844796e-08,
"loss": 0.5237,
"step": 2389
},
{
"epoch": 0.9615771474552404,
"grad_norm": 4.121244015419092,
"learning_rate": 2.3160586304006026e-08,
"loss": 0.538,
"step": 2390
},
{
"epoch": 0.9619794809897405,
"grad_norm": 4.166559231330322,
"learning_rate": 2.268622981127705e-08,
"loss": 0.5822,
"step": 2391
},
{
"epoch": 0.9623818145242405,
"grad_norm": 3.9211071919172813,
"learning_rate": 2.2216759265220044e-08,
"loss": 0.4897,
"step": 2392
},
{
"epoch": 0.9627841480587407,
"grad_norm": 4.349097825226008,
"learning_rate": 2.1752175591760184e-08,
"loss": 0.5211,
"step": 2393
},
{
"epoch": 0.9631864815932408,
"grad_norm": 3.9282047251137677,
"learning_rate": 2.1292479707184787e-08,
"loss": 0.5741,
"step": 2394
},
{
"epoch": 0.9635888151277409,
"grad_norm": 4.159989263744697,
"learning_rate": 2.0837672518141395e-08,
"loss": 0.5238,
"step": 2395
},
{
"epoch": 0.963991148662241,
"grad_norm": 3.8198728228381293,
"learning_rate": 2.0387754921634972e-08,
"loss": 0.5144,
"step": 2396
},
{
"epoch": 0.9643934821967411,
"grad_norm": 3.346255200741844,
"learning_rate": 1.9942727805027373e-08,
"loss": 0.4384,
"step": 2397
},
{
"epoch": 0.9647958157312412,
"grad_norm": 3.8469019196516845,
"learning_rate": 1.9502592046035107e-08,
"loss": 0.486,
"step": 2398
},
{
"epoch": 0.9651981492657413,
"grad_norm": 3.5776773638982893,
"learning_rate": 1.9067348512726847e-08,
"loss": 0.4414,
"step": 2399
},
{
"epoch": 0.9656004828002414,
"grad_norm": 4.031486537261451,
"learning_rate": 1.8636998063523705e-08,
"loss": 0.5328,
"step": 2400
},
{
"epoch": 0.9660028163347415,
"grad_norm": 4.260664576753229,
"learning_rate": 1.8211541547195067e-08,
"loss": 0.5136,
"step": 2401
},
{
"epoch": 0.9664051498692416,
"grad_norm": 4.152186034058617,
"learning_rate": 1.7790979802858876e-08,
"loss": 0.6316,
"step": 2402
},
{
"epoch": 0.9668074834037417,
"grad_norm": 3.995357070807909,
"learning_rate": 1.7375313659978855e-08,
"loss": 0.5377,
"step": 2403
},
{
"epoch": 0.9672098169382418,
"grad_norm": 3.4651763373212474,
"learning_rate": 1.6964543938363388e-08,
"loss": 0.4157,
"step": 2404
},
{
"epoch": 0.9676121504727418,
"grad_norm": 4.360198459376172,
"learning_rate": 1.6558671448163866e-08,
"loss": 0.6273,
"step": 2405
},
{
"epoch": 0.968014484007242,
"grad_norm": 4.006498747343065,
"learning_rate": 1.6157696989873017e-08,
"loss": 0.4914,
"step": 2406
},
{
"epoch": 0.9684168175417421,
"grad_norm": 3.7398733702546822,
"learning_rate": 1.5761621354322965e-08,
"loss": 0.4022,
"step": 2407
},
{
"epoch": 0.9688191510762422,
"grad_norm": 4.1110217996882294,
"learning_rate": 1.5370445322684392e-08,
"loss": 0.4698,
"step": 2408
},
{
"epoch": 0.9692214846107423,
"grad_norm": 4.362254470529691,
"learning_rate": 1.498416966646432e-08,
"loss": 0.5598,
"step": 2409
},
{
"epoch": 0.9696238181452425,
"grad_norm": 3.8880330136972328,
"learning_rate": 1.4602795147504734e-08,
"loss": 0.473,
"step": 2410
},
{
"epoch": 0.9700261516797425,
"grad_norm": 4.110874787245324,
"learning_rate": 1.4226322517981728e-08,
"loss": 0.4901,
"step": 2411
},
{
"epoch": 0.9704284852142426,
"grad_norm": 4.138814644855194,
"learning_rate": 1.3854752520403303e-08,
"loss": 0.5618,
"step": 2412
},
{
"epoch": 0.9708308187487427,
"grad_norm": 4.481371755440245,
"learning_rate": 1.3488085887607972e-08,
"loss": 0.6152,
"step": 2413
},
{
"epoch": 0.9712331522832428,
"grad_norm": 3.6050373310270176,
"learning_rate": 1.3126323342763648e-08,
"loss": 0.5525,
"step": 2414
},
{
"epoch": 0.971635485817743,
"grad_norm": 4.061084112826385,
"learning_rate": 1.2769465599365982e-08,
"loss": 0.4619,
"step": 2415
},
{
"epoch": 0.972037819352243,
"grad_norm": 3.8920028399978097,
"learning_rate": 1.24175133612367e-08,
"loss": 0.4833,
"step": 2416
},
{
"epoch": 0.9724401528867431,
"grad_norm": 3.909035344776143,
"learning_rate": 1.2070467322522762e-08,
"loss": 0.5188,
"step": 2417
},
{
"epoch": 0.9728424864212432,
"grad_norm": 4.225663392112197,
"learning_rate": 1.1728328167695258e-08,
"loss": 0.6696,
"step": 2418
},
{
"epoch": 0.9732448199557433,
"grad_norm": 4.6741890030042725,
"learning_rate": 1.1391096571546633e-08,
"loss": 0.5815,
"step": 2419
},
{
"epoch": 0.9736471534902434,
"grad_norm": 4.51608876386972,
"learning_rate": 1.1058773199190965e-08,
"loss": 0.5173,
"step": 2420
},
{
"epoch": 0.9740494870247435,
"grad_norm": 4.0654982873308665,
"learning_rate": 1.0731358706061456e-08,
"loss": 0.5749,
"step": 2421
},
{
"epoch": 0.9744518205592436,
"grad_norm": 4.2304161552160835,
"learning_rate": 1.0408853737909896e-08,
"loss": 0.5813,
"step": 2422
},
{
"epoch": 0.9748541540937438,
"grad_norm": 3.4742799346124493,
"learning_rate": 1.0091258930805259e-08,
"loss": 0.4705,
"step": 2423
},
{
"epoch": 0.9752564876282438,
"grad_norm": 4.08672880848899,
"learning_rate": 9.778574911132044e-09,
"loss": 0.516,
"step": 2424
},
{
"epoch": 0.9756588211627439,
"grad_norm": 3.669930620914309,
"learning_rate": 9.470802295589442e-09,
"loss": 0.4138,
"step": 2425
},
{
"epoch": 0.976061154697244,
"grad_norm": 3.5860479619434673,
"learning_rate": 9.16794169119023e-09,
"loss": 0.4224,
"step": 2426
},
{
"epoch": 0.9764634882317441,
"grad_norm": 4.642054584241187,
"learning_rate": 8.869993695258816e-09,
"loss": 0.611,
"step": 2427
},
{
"epoch": 0.9768658217662443,
"grad_norm": 3.977155407637389,
"learning_rate": 8.576958895431531e-09,
"loss": 0.5681,
"step": 2428
},
{
"epoch": 0.9772681553007443,
"grad_norm": 3.6884328747774857,
"learning_rate": 8.28883786965301e-09,
"loss": 0.4945,
"step": 2429
},
{
"epoch": 0.9776704888352444,
"grad_norm": 4.023502331465119,
"learning_rate": 8.005631186177864e-09,
"loss": 0.4201,
"step": 2430
},
{
"epoch": 0.9780728223697445,
"grad_norm": 3.901800973146699,
"learning_rate": 7.727339403568179e-09,
"loss": 0.4977,
"step": 2431
},
{
"epoch": 0.9784751559042446,
"grad_norm": 3.9967038572140856,
"learning_rate": 7.45396307069185e-09,
"loss": 0.6418,
"step": 2432
},
{
"epoch": 0.9788774894387448,
"grad_norm": 4.291073567899429,
"learning_rate": 7.185502726722305e-09,
"loss": 0.6601,
"step": 2433
},
{
"epoch": 0.9792798229732448,
"grad_norm": 3.974019553234897,
"learning_rate": 6.921958901137948e-09,
"loss": 0.4781,
"step": 2434
},
{
"epoch": 0.9796821565077449,
"grad_norm": 3.833029352112665,
"learning_rate": 6.66333211371939e-09,
"loss": 0.4373,
"step": 2435
},
{
"epoch": 0.9800844900422451,
"grad_norm": 4.398810707509668,
"learning_rate": 6.409622874550547e-09,
"loss": 0.5795,
"step": 2436
},
{
"epoch": 0.9804868235767451,
"grad_norm": 3.5697882170733557,
"learning_rate": 6.160831684015878e-09,
"loss": 0.398,
"step": 2437
},
{
"epoch": 0.9808891571112452,
"grad_norm": 3.904433602202901,
"learning_rate": 5.9169590328003735e-09,
"loss": 0.5029,
"step": 2438
},
{
"epoch": 0.9812914906457453,
"grad_norm": 3.7353800882065715,
"learning_rate": 5.678005401888176e-09,
"loss": 0.4911,
"step": 2439
},
{
"epoch": 0.9816938241802454,
"grad_norm": 3.795949925181833,
"learning_rate": 5.443971262561465e-09,
"loss": 0.4405,
"step": 2440
},
{
"epoch": 0.9820961577147456,
"grad_norm": 3.617609885039579,
"learning_rate": 5.214857076400736e-09,
"loss": 0.4329,
"step": 2441
},
{
"epoch": 0.9824984912492456,
"grad_norm": 3.5249358490195637,
"learning_rate": 4.990663295282305e-09,
"loss": 0.4285,
"step": 2442
},
{
"epoch": 0.9829008247837457,
"grad_norm": 4.142775577116706,
"learning_rate": 4.771390361377748e-09,
"loss": 0.4777,
"step": 2443
},
{
"epoch": 0.9833031583182458,
"grad_norm": 4.312344250176257,
"learning_rate": 4.5570387071544595e-09,
"loss": 0.6627,
"step": 2444
},
{
"epoch": 0.9837054918527459,
"grad_norm": 3.67386973137855,
"learning_rate": 4.347608755372601e-09,
"loss": 0.4431,
"step": 2445
},
{
"epoch": 0.9841078253872461,
"grad_norm": 4.191733748165821,
"learning_rate": 4.143100919086207e-09,
"loss": 0.4884,
"step": 2446
},
{
"epoch": 0.9845101589217461,
"grad_norm": 4.38016174030762,
"learning_rate": 3.943515601640968e-09,
"loss": 0.5822,
"step": 2447
},
{
"epoch": 0.9849124924562462,
"grad_norm": 4.227081529941771,
"learning_rate": 3.748853196673952e-09,
"loss": 0.49,
"step": 2448
},
{
"epoch": 0.9853148259907464,
"grad_norm": 3.6169704926991124,
"learning_rate": 3.5591140881138797e-09,
"loss": 0.5642,
"step": 2449
},
{
"epoch": 0.9857171595252464,
"grad_norm": 3.228283450712886,
"learning_rate": 3.374298650178076e-09,
"loss": 0.3926,
"step": 2450
},
{
"epoch": 0.9861194930597466,
"grad_norm": 3.996805720599477,
"learning_rate": 3.194407247373854e-09,
"loss": 0.5175,
"step": 2451
},
{
"epoch": 0.9865218265942466,
"grad_norm": 3.786750548322366,
"learning_rate": 3.0194402344965733e-09,
"loss": 0.4496,
"step": 2452
},
{
"epoch": 0.9869241601287467,
"grad_norm": 3.730366038034417,
"learning_rate": 2.8493979566296403e-09,
"loss": 0.4701,
"step": 2453
},
{
"epoch": 0.9873264936632469,
"grad_norm": 4.616677722561612,
"learning_rate": 2.6842807491433974e-09,
"loss": 0.6207,
"step": 2454
},
{
"epoch": 0.9877288271977469,
"grad_norm": 3.9983804243694645,
"learning_rate": 2.524088937694291e-09,
"loss": 0.4734,
"step": 2455
},
{
"epoch": 0.988131160732247,
"grad_norm": 4.364062652259154,
"learning_rate": 2.3688228382251487e-09,
"loss": 0.6254,
"step": 2456
},
{
"epoch": 0.9885334942667471,
"grad_norm": 3.97492888774164,
"learning_rate": 2.218482756963236e-09,
"loss": 0.5645,
"step": 2457
},
{
"epoch": 0.9889358278012472,
"grad_norm": 3.9671569540217493,
"learning_rate": 2.07306899042109e-09,
"loss": 0.4821,
"step": 2458
},
{
"epoch": 0.9893381613357474,
"grad_norm": 3.808293451273474,
"learning_rate": 1.9325818253945748e-09,
"loss": 0.4502,
"step": 2459
},
{
"epoch": 0.9897404948702474,
"grad_norm": 4.467378355634954,
"learning_rate": 1.7970215389628842e-09,
"loss": 0.5963,
"step": 2460
},
{
"epoch": 0.9901428284047475,
"grad_norm": 4.625199727103312,
"learning_rate": 1.6663883984888163e-09,
"loss": 0.6419,
"step": 2461
},
{
"epoch": 0.9905451619392477,
"grad_norm": 3.9326948253065184,
"learning_rate": 1.5406826616168324e-09,
"loss": 0.5575,
"step": 2462
},
{
"epoch": 0.9909474954737477,
"grad_norm": 4.4568921518738485,
"learning_rate": 1.4199045762730569e-09,
"loss": 0.6483,
"step": 2463
},
{
"epoch": 0.9913498290082479,
"grad_norm": 4.351427492104435,
"learning_rate": 1.3040543806658313e-09,
"loss": 0.5948,
"step": 2464
},
{
"epoch": 0.9917521625427479,
"grad_norm": 3.4818450271582857,
"learning_rate": 1.1931323032832177e-09,
"loss": 0.3535,
"step": 2465
},
{
"epoch": 0.992154496077248,
"grad_norm": 4.11227816177031,
"learning_rate": 1.0871385628941077e-09,
"loss": 0.5317,
"step": 2466
},
{
"epoch": 0.9925568296117482,
"grad_norm": 4.418916055074932,
"learning_rate": 9.860733685479463e-10,
"loss": 0.6433,
"step": 2467
},
{
"epoch": 0.9929591631462482,
"grad_norm": 3.6128250845259724,
"learning_rate": 8.899369195727869e-10,
"loss": 0.5447,
"step": 2468
},
{
"epoch": 0.9933614966807484,
"grad_norm": 4.364090118207612,
"learning_rate": 7.987294055761264e-10,
"loss": 0.6099,
"step": 2469
},
{
"epoch": 0.9937638302152484,
"grad_norm": 3.739315819157321,
"learning_rate": 7.124510064446255e-10,
"loss": 0.4998,
"step": 2470
},
{
"epoch": 0.9941661637497485,
"grad_norm": 4.035011038094296,
"learning_rate": 6.311018923432776e-10,
"loss": 0.4558,
"step": 2471
},
{
"epoch": 0.9945684972842487,
"grad_norm": 3.577550936704116,
"learning_rate": 5.546822237145755e-10,
"loss": 0.4408,
"step": 2472
},
{
"epoch": 0.9949708308187487,
"grad_norm": 3.7748701163123584,
"learning_rate": 4.831921512796211e-10,
"loss": 0.4403,
"step": 2473
},
{
"epoch": 0.9953731643532489,
"grad_norm": 3.7403701320267686,
"learning_rate": 4.166318160361837e-10,
"loss": 0.4989,
"step": 2474
},
{
"epoch": 0.995775497887749,
"grad_norm": 3.8787870872673857,
"learning_rate": 3.5500134925980925e-10,
"loss": 0.5066,
"step": 2475
},
{
"epoch": 0.996177831422249,
"grad_norm": 4.321954937291901,
"learning_rate": 2.983008725029879e-10,
"loss": 0.498,
"step": 2476
},
{
"epoch": 0.9965801649567492,
"grad_norm": 3.574339539948393,
"learning_rate": 2.4653049759432167e-10,
"loss": 0.4164,
"step": 2477
},
{
"epoch": 0.9969824984912492,
"grad_norm": 3.7965194640316478,
"learning_rate": 1.9969032663935683e-10,
"loss": 0.4769,
"step": 2478
},
{
"epoch": 0.9973848320257493,
"grad_norm": 4.23022008583803,
"learning_rate": 1.5778045202002878e-10,
"loss": 0.4481,
"step": 2479
},
{
"epoch": 0.9977871655602495,
"grad_norm": 4.071535851310751,
"learning_rate": 1.2080095639410704e-10,
"loss": 0.535,
"step": 2480
},
{
"epoch": 0.9981894990947495,
"grad_norm": 3.9007108087469646,
"learning_rate": 8.875191269547279e-11,
"loss": 0.542,
"step": 2481
},
{
"epoch": 0.9985918326292497,
"grad_norm": 4.05416067842318,
"learning_rate": 6.163338413328612e-11,
"loss": 0.5738,
"step": 2482
},
{
"epoch": 0.9989941661637497,
"grad_norm": 3.9020981788420386,
"learning_rate": 3.944542419337394e-11,
"loss": 0.4897,
"step": 2483
},
{
"epoch": 0.9993964996982498,
"grad_norm": 3.6120331011996085,
"learning_rate": 2.2188076636009415e-11,
"loss": 0.5195,
"step": 2484
},
{
"epoch": 0.99979883323275,
"grad_norm": 3.4028120769775385,
"learning_rate": 9.861375497577375e-12,
"loss": 0.3763,
"step": 2485
},
{
"epoch": 1.0,
"grad_norm": 3.4028120769775385,
"learning_rate": 2.4653450900191666e-12,
"loss": 0.4064,
"step": 2486
},
{
"epoch": 1.0,
"step": 2486,
"total_flos": 882515131957248.0,
"train_loss": 0.5852410430066162,
"train_runtime": 194010.5784,
"train_samples_per_second": 1.64,
"train_steps_per_second": 0.013
}
],
"logging_steps": 1.0,
"max_steps": 2486,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 882515131957248.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}